Xingchen Wan

Senior Research Scientist, Google

prof_pic.JPG

1600 Amphitheatre Parkway

Mountain View, CA 94043

xingchenw[at]google.com

I am a Senior Research Scientist at Google based in the San Francisco Bay Area.

Research Interests

My primary research drives innovations in large language models (LLMs), focusing on building systems that are more efficient, robust, and autonomous. My contributions span:

  • LLM Post-training (e.g., [1, 2, 3]);
  • Developing self-improving (multimodal) LLM agents (e.g., [4, 5, 6]);
  • Automated optimization techniques for LLMs and agents (e.g., [7, 8, 9]); and
  • Integrating GenAI with large-scale (unstructured) data systems (e.g., [10, 11]).

I have authored over 30 papers, including more than 20 published in top peer-reviewed conferences (e.g., NeurIPS, ICML, ICLR) and journals (e.g., JMLR, TACL) with more than 1200 citations as of Nov. 2025.

Previously, I did my PhD in the Machine Learning Research Group, Department of Engineering Science, University of Oxford where I worked on Bayesian optimization, AutoML, and machine learning on graphs.

Academic Services

Area chair/senior program committee member at NeurIPS (2024-25), ICML (2025), ACL ARR (2025-); Action editor at TMLR.

Reviewer/program committee member at ACL (2023-24), AutoML-Conf (2023-24), COLM (2024), CVPR (2024), ECCV (2024), EMNLP (2023-24), ICLR (2024-25), ICML (2023-24), JMLR, Machine Learning, NeurIPS (2022-23), WACV (2022-24), etc.

news

Oct 17, 2025 We present VISTA and Maestro, two self-improving multimodal generation agents for text-to-video and text-to-image generation, respectively.
May 19, 2025 We present Visual Planning, where we apply reinforcement learning post-training on pure-vision models to achieve state-of-the-art performance in visual reasoning tasks.

selected publications

  1. VISTA: A Test-Time Self-Improving Video Generation Agent
    Do Xuan Long, Xingchen Wan, Hootan Nakhost, Chen-Yu Lee, Tomas Pfister, and Sercan Ɩ. Arık
    arXiv preprint arXiv:2510.15831, 2025
  2. Maestro: Self-Improving Text-to-Image Generation via Agent Orchestration
    Xingchen Wan, Han Zhou, Ruoxi Sun, Hootan Nakhost, Ke Jiang, Rajarishi Sinha, and Sercan Ɩ. Arık
    arXiv preprint arXiv:2509.10704, 2025
  3. Visual Planning: Let’s Think Only with Images
    Yi Xu*, Chengzu Li*, Han Zhou*, Xingchen Wan, Caiqi Zhang, Anna Korhonen, and Ivan Vulić
    arXiv preprint arXiv:2505.11409. šŸ†šŸ„‰ #3 paper of the day at HuggingFace šŸ¤— , 2025
  4. Multi-Agent Design: Optimizing Agents with Better Prompts and Topologies
    Han Zhou, Xingchen Wan, Ruoxi Sun, Hamid Palangi, Shariq Iqbal, Ivan Vulić, Anna Korhonen, and Sercan Ɩ. Arık
    arXiv preprint arXiv:2502.02533, 2025
  5. ICLR 2025
    bridge.png
    From Few to Many: Self-Improving Many-Shot Reasoners Through Iterative Optimization and Generation
    Xingchen Wan, Han Zhou, Ruoxi Sun, Hootan Nakhost, Ke Jiang, and Sercan Ɩ. Arık
    In The Thirteenth International Conference on Learning Representations, 2025
  6. ACL 2025
    astuterag.png
    Astute RAG: Overcoming Imperfect Retrieval Augmentation and Knowledge Conflicts for Large Language Models
    Fei Wang, Xingchen Wan, Ruoxi Sun, Jiefeng Chen, and Sercan O Arik
    In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025
  7. NeurIPS 2024
    ioes.png
    Teach Better or Show Smarter? On Instructions and Exemplars in Automatic Prompt Optimization
    Xingchen Wan, Ruoxi Sun, Hootan Nakhost, and Sercan Ɩ. Arik
    In Advances in Neural Information Processing Systems 37. ā˜ļø Powers the Google Cloud Vertex AI Prompt Optimizer , 2024