Rana Muhammad Shahroz Khan

I am a first second year PhD student at UNC Chapel Hill where I work on Efficient ML and NLP. I am grateful to be advised by Dr. Tianlong Chen. Before that I did my Bachelors in Computer Science and Mathematics from Vanderbilt University, where I was advised by Dr. Soheil Kolouri and worked in the intersection of Computer Vision and Optimal Transport.

During the summer of 2023, I was a Research Intern at Lawrence Livermore National Laboratory in the Machine Intelligence Group working under the guidance of Dr. Jay Thiagarajan, Dr. Shusen Liu and Dr. Rushil Anirudh on Bayesian Optimization for Uncertainty Quantification. In the Summer of 2024, I also had the pleasure of working as a ML Research Intern at HighArc under the supervision of Dr. Ardavan Bidgoli and Dr. Manuel Ladron de Guevara.

Email  /  GitHub  /  Google Scholar  /  LinkedIn  /  CV [03/2026]

profile photo

Experience

  • Anthropic — Research Fellow (May 2026 – Present)
  • Cisco — Research Intern (May 2025 – Dec 2025)
  • HighArc — ML Research Intern (May 2024 – Aug 2024)
  • Lawrence Livermore National Laboratory — Research Intern (May 2023 – Aug 2023)

Updates

  • [March 2026] I will be joining Anthropic as a Research Fellow this summer.
  • [January 2026] 1 paper accepted at SenSys 2026.
  • [January 2026] Two papers accepted at ICLR 2026.
  • [Sepetember 2025] ORAL paper accepted at EMNLP 2025 Findings.
  • [May 2025] Our preprint "Agents Under Siege: Breaking Pragmatic Multi-Agent LLM Systems with Optimized Prompt Attacks" have been accepted to ACL 2025's Main Track !
  • [April 2025] Excited to give a talk on our preprint "Agents Under Siege" at ResearchTrend.AI
  • [April 2025] Two new preprints on: (1) Conditional LoRA Generation and (2) Safety of Multi-agent Systems.
  • [March 2025] I will be joining Cisco for Summer 2025.
  • Research

    I am broadly interested in efficient reasoning.

    project image

    TMS: Trajectory-Mixed Supervision for Reward-Free, On-Policy SFT


    Rana Muhammad Shahroz Khan, Zijie Liu, Zhen Tan, Charles Fleming, Tianlong Chen
    Preprint, 2026
    paper /

    We present a simple, drop-in SFT replacement that recovers RL-like retention by training on trajectory-mixed, near-policy targets instead of static labels—eliminating reward models while preventing forgetting.

    project image

    The Quest for Efficient Reasoning: A Data-Centric Benchmark to CoT Distillation


    Ruichen Zhang*, Rana Muhammad Shahroz Khan*, Zhen Tan, Dawei Li, Song Wang, Tianlong Chen
    International Conference on Learning Representations (ICLR), 2026
    paper /

    We present a comprehensive benchmark (DC-CoT) that systematically shows data-centric strategies, especially augmentation like reverse reasoning, are the most effective way to distill strong reasoning abilities into smaller LLMs.

    project image

    CAR-LoRA: Training Compression-Aware and Robust LoRA Adapters for Evolving LLMs


    Rana Muhammad Shahroz Khan, Zhen Tan, Ruichen Zhang, Hua Wei, Tianlong Chen, Charles Fleming
    International Conference on Learning Representations (ICLR), 2026
    paper /

    We present a unified training framework (CAR-LoRA) that learns a single compression-aware and temporally robust LoRA adapter, eliminating the need to retrain separate adapters across different hardware compressions and evolving LLM versions.

    project image

    ORAL: Prompting Your Large-Scale LoRAs via Conditional Recurrent Diffusion


    Rana Muhammad Shahroz Khan, Dongwen Tang, Pingzhi Li, Kai Wang, Tianlong Chen
    Conference on Empirical Methods in Natural Language Processing (EMNLP), Findings, 2025
    paper /

    ORAL leverages conditional recurrent diffusion to instantly craft task‑tuned LoRA adapters that scale to billion‑parameter models, remain compatible with future model updates, and rival—or beat—full fine‑tuning accuracy.

    project image

    Agents Under Siege: Breaking Pragmatic Multi-Agent LLM Systems with Optimized Prompt Attacks


    Rana Muhammad Shahroz Khan, Zhen Tan, Sukwon Yun, Charles Fleming, Tianlong Chen
    Association for Computational Linguistics (ACL), ORAL, 2025
    paper /

    Agents Under Siege introduces a flow‑optimized, order‑agnostic prompt attack that stealthily threads limited‑bandwidth, high‑latency multi‑agent LLM networks to jailbreak target models, overwhelming modern safety guards and boosting attack success up to seven‑fold.

    project image

    PortLLM: Personalizing Evolving Large Language Models with Training-Free and Portable Model Patches


    Rana Muhammad Shahroz Khan, Pingzhi Li*, Sukwon Yun*, Zhenyu Wang, Shahriar Nirjon, Chau-Wai Wong, Tianlong Chen
    International Conference on Learning Representations (ICLR), 2025
    paper /

    We present PORTLLM, a training-free framework that enables seamless knowledge transfer across evolving LLMs, achieving LoRA-level performance with significantly lower computational costs.

    project image

    LLMGeo: Benchmarking Large Language Models on Image Geolocation In-the-wild


    Zhiqiang Wang, Dejia Xu, Rana Muhammad Shahroz Khan, Yanbin Lin, Zhiwen Fan, Xingquan Zhu
    Conference on Computer Vision and Pattern Recognition, Workshop on Computer Vision in the Wild (CVPRW), 2024
    paper / code /

    We evaluate multimodal LLMs for image geolocation using a new dataset, showing that fine-tuning helps open-source models approach closed-source performance.

    project image

    PRANC: Pseudo RAndom Networks for Compacting Deep Models


    Parsa Nooralinejad, Ali Abbasi, Rana Muhammad Shahroz Khan*, Soroush Abbasi Koohpayegani*, Kossar Pourahmadi Meibodi*, Soheil Kolouri, Hamed Pirsiavash;
    International Conference on Computer Vision (ICCV), 2023
    paper / code /

    We propose PRANC, a framework that reparametrizes deep models as a linear combination of frozen random networks, enabling extreme compression, efficient storage, and memory-efficient inference.

    project image

    Linear Optimal Partial Transport Embedding


    Yikun Bai, Ivan Vladimir Medri, Rocio Diaz Martin, Rana Shahroz, Soheil Kolouri
    International Conference on Machine Learning (ICML), 2023
    paper / code /

    We propose the Linear Optimal Partial Transport (LOPT) embedding, enabling faster OPT distance computation and demonstrating its effectiveness in point-cloud interpolation and PCA analysis.





    Design and source code from Jon Barron's website