Experience
- Anthropic — Research Fellow (May 2026 – Present)
- Cisco — Research Intern (May 2025 – Dec 2025)
- HighArc — ML Research Intern (May 2024 – Aug 2024)
- Lawrence Livermore National Laboratory — Research Intern (May 2023 – Aug 2023)
|
Updates
[March 2026] I will be joining Anthropic as a Research Fellow this summer.
[January 2026] 1 paper accepted at SenSys 2026.
[January 2026] Two papers accepted at ICLR 2026.
[Sepetember 2025] ORAL paper accepted at EMNLP 2025 Findings.
[May 2025] Our preprint "Agents Under Siege: Breaking Pragmatic Multi-Agent LLM Systems with Optimized Prompt Attacks" have been accepted to ACL 2025's Main Track !
[April 2025] Excited to give a talk on our preprint "Agents Under Siege" at ResearchTrend.AI
[April 2025] Two new preprints on: (1) Conditional LoRA Generation and (2) Safety of Multi-agent Systems.
[March 2025] I will be joining Cisco for Summer 2025.
|
Research
I am broadly interested in efficient reasoning.
|
|
TMS: Trajectory-Mixed Supervision for Reward-Free, On-Policy SFT
Rana Muhammad Shahroz Khan, Zijie Liu, Zhen Tan, Charles Fleming, Tianlong Chen
Preprint, 2026
paper /
We present a simple, drop-in SFT replacement that recovers RL-like retention by training on trajectory-mixed, near-policy targets instead of static labels—eliminating reward models while preventing forgetting.
|
|
The Quest for Efficient Reasoning: A Data-Centric Benchmark to CoT Distillation
Ruichen Zhang*, Rana Muhammad Shahroz Khan*, Zhen Tan, Dawei Li, Song Wang, Tianlong Chen
International Conference on Learning Representations (ICLR), 2026
paper /
We present a comprehensive benchmark (DC-CoT) that systematically shows data-centric strategies, especially augmentation like reverse reasoning, are the most effective way to distill strong reasoning abilities into smaller LLMs.
|
|
CAR-LoRA: Training Compression-Aware and Robust LoRA Adapters for Evolving LLMs
Rana Muhammad Shahroz Khan, Zhen Tan, Ruichen Zhang, Hua Wei, Tianlong Chen, Charles Fleming
International Conference on Learning Representations (ICLR), 2026
paper /
We present a unified training framework (CAR-LoRA) that learns a single compression-aware and temporally robust LoRA adapter, eliminating the need to retrain separate adapters across different hardware compressions and evolving LLM versions.
|
|
ORAL: Prompting Your Large-Scale LoRAs via Conditional Recurrent Diffusion
Rana Muhammad Shahroz Khan, Dongwen Tang, Pingzhi Li, Kai Wang, Tianlong Chen
Conference on Empirical Methods in Natural Language Processing (EMNLP), Findings, 2025
paper /
ORAL leverages conditional recurrent diffusion to instantly craft task‑tuned LoRA adapters that scale to billion‑parameter models, remain compatible with future model updates, and rival—or beat—full fine‑tuning accuracy.
|
|
Agents Under Siege: Breaking Pragmatic Multi-Agent LLM Systems with Optimized Prompt Attacks
Rana Muhammad Shahroz Khan, Zhen Tan, Sukwon Yun, Charles Fleming, Tianlong Chen
Association for Computational Linguistics (ACL), ORAL, 2025
paper /
Agents Under Siege introduces a flow‑optimized, order‑agnostic prompt attack that stealthily threads limited‑bandwidth, high‑latency multi‑agent LLM networks to jailbreak target models, overwhelming modern safety guards and boosting attack success up to seven‑fold.
|
|
PortLLM: Personalizing Evolving Large Language Models with Training-Free and Portable Model Patches
Rana Muhammad Shahroz Khan, Pingzhi Li*, Sukwon Yun*, Zhenyu Wang, Shahriar Nirjon, Chau-Wai Wong, Tianlong Chen
International Conference on Learning Representations (ICLR), 2025
paper /
We present PORTLLM, a training-free framework that enables seamless knowledge transfer across evolving LLMs, achieving LoRA-level performance with significantly lower computational costs.
|
|
LLMGeo: Benchmarking Large Language Models on Image Geolocation In-the-wild
Zhiqiang Wang, Dejia Xu, Rana Muhammad Shahroz Khan, Yanbin Lin, Zhiwen Fan, Xingquan Zhu
Conference on Computer Vision and Pattern Recognition, Workshop on Computer Vision in the Wild (CVPRW), 2024
paper /
code /
We evaluate multimodal LLMs for image geolocation using a new dataset, showing that fine-tuning helps open-source models approach closed-source performance.
|
|
PRANC: Pseudo RAndom Networks for Compacting Deep Models
Parsa Nooralinejad, Ali Abbasi, Rana Muhammad Shahroz Khan*, Soroush Abbasi Koohpayegani*, Kossar Pourahmadi Meibodi*, Soheil Kolouri, Hamed Pirsiavash;
International Conference on Computer Vision (ICCV), 2023
paper /
code /
We propose PRANC, a framework that reparametrizes deep models as a linear combination of frozen random networks, enabling extreme compression, efficient storage, and memory-efficient inference.
|
|
Linear Optimal Partial Transport Embedding
Yikun Bai, Ivan Vladimir Medri, Rocio Diaz Martin, Rana Shahroz, Soheil Kolouri
International Conference on Machine Learning (ICML), 2023
paper /
code /
We propose the Linear Optimal Partial Transport (LOPT) embedding, enabling faster OPT distance computation and demonstrating its effectiveness in point-cloud interpolation and PCA analysis.
|
|