Reza Esfandiarpoor

headshot_3.jpg

I am a Ph.D. candidate in Computer Science at Brown University, working with Stephen Bach. I am fortunate to have interned at Google (AdsAI team) and Microsoft (Office AI team). And, I am currently interning at Nvidia in the Nemo Retriever team, where I explore agentic information retrieval.

My research focuses on optimizing the interaction between large pretrained models to better understand their behavior and improve their performance across different domains. Here is a summary of my research:

  • AI Agents. I introduced a cheaper, faster, and more effective approach for building general-purpose agents using massive task-specific tool sets. I also created TheMCPCompany, a benchmark for evaluating enterprise agents with more than 18,000 tools and complex, realistic tasks, like managing Microsoft Azure resources.

  • Vision-Language Models. I introduced Extract and Explore, a novel analysis method that uses reinforcement learning to align an LLM with VLM preferences. By analyzing the aligned LLM outputs, I discovered that non-visual and even spurious information significantly impact VLM concept representations. I introduce Follow-up Differential Descriptions, a new method that adapts LLM-generated concept descriptions to each individual VLM to improve performance.

  • Information Retrieval. I introduced SyCL, a new synthetic data generation method that creates fine-grained training data with multi-level relevance labels for training dense retrievers. I also created Trove, an open-source toolkit for dense retrieval that simplifies IR experiments without sacrificing flexibility.

Before Ph.D., I completed my undergraduate studies with Shadrokh Samavi, where I worked on pruning and quantization methods to create efficient medical image analysis models for edge devices.

Open Source Software

  1. Trove Logo
    Trove: A Flexible Toolkit for Dense Retrieval
    Reza Esfandiarpoor, Max Zuo, and Stephen H. Bach
    2025

Selected Publications (see all)

  1. TheMCPCompany: Creating General-purpose Agents with Task-specific Tools
    Reza Esfandiarpoor, Vishwas Suryanarayanan, Stephen H. Bach, Vishal Chowdhary, and Anthony Aue
    arXiv preprint, 2025
  2. Beyond Contrastive Learning: Synthetic Data Enables List-wise Training with Multiple Levels of Relevance
    Reza Esfandiarpoor*, George Zerveas*, Ruochen Zhang, Macton Mgonzo, Carsten Eickhoff, and Stephen H. Bach
    Findings of the EMNLP, 2025
  3. If CLIP Could Talk: Understanding Vision-Language Model Representations Through Their Preferred Concept Descriptions
    Reza Esfandiarpoor, Cristina Menghini, and Stephen H. Bach
    EMNLP, 2024
  4. Follow-Up Differential Descriptions: Language Models Resolve Ambiguities for Image Classification
    Reza Esfandiarpoor and Stephen H. Bach
    ICLR, 2024
  5. Extended Few-Shot Learning: Exploiting Existing Resources for Novel Tasks
    Reza Esfandiarpoor, Amy Pu, Mohsen Hajabdollahi, and Stephen H. Bach
    arXiv, 2021