Xingchen Wan

Senior Research Scientist, Google

prof_pic.JPG

1600 Amphitheatre Parkway

Mountain View, CA 94043

xingchenw[at]google.com

I am a Senior Research Scientist at Google based in the San Francisco Bay Area.

Research Interests

My primary research drives innovations in large language models (LLMs), focusing on building systems that are more efficient, robust, and autonomous. My contributions span:

  • LLM Post-training (e.g., [1, 2, 3, 4]);
  • Developing self-improving (multimodal) LLM agents (e.g., [5, 6, 7]);
  • Automating prompt and agent design (e.g., [8, 9, 10]); and
  • Integrating GenAI with large-scale (unstructured) data systems (e.g., [11, 12]).

Previously, I did my PhD in the Machine Learning Research Group, Department of Engineering Science, University of Oxford where I worked on Bayesian optimization, AutoML, and machine learning on graphs.

Academic Services

Area chair/senior program committee member at ICML (2025, 2026), NeurIPS (2024, 2025), ACL ARR (2025-); Action editor at TMLR.

Reviewer/program committee member at ACL (2023-24), AutoML-Conf (2023-24), COLM (2024), CVPR (2024), ECCV (2024), EMNLP (2023-24), ICLR (2024-25), ICML (2023-24), JMLR, Machine Learning, NeurIPS (2022-23), WACV (2022-24), etc.

news

Feb 06, 2026 Two papers have been accepted to ICLR 2026: MASS (Optimizing Agents with Better Prompts and Topologies) and Visual Planning with Reinforcement Learning.
Oct 17, 2025 We present VISTA and Maestro, two self-improving multimodal generation agents for text-to-video and text-to-image generation, respectively.
May 19, 2025 We present Visual Planning, where we apply reinforcement learning post-training on pure-vision models to achieve state-of-the-art performance in visual reasoning tasks.

selected publications

  1. VISTA: A Test-Time Self-Improving Video Generation Agent
    Do Xuan Long, Xingchen Wan, Hootan Nakhost, Chen-Yu Lee, Tomas Pfister, and Sercan Ɩ. Arık
    arXiv preprint arXiv:2510.15831, 2025
  2. Maestro: Self-Improving Text-to-Image Generation via Agent Orchestration
    Xingchen Wan, Han Zhou, Ruoxi Sun, Hootan Nakhost, Ke Jiang, Rajarishi Sinha, and Sercan Ɩ. Arık
    arXiv preprint arXiv:2509.10704, 2025
  3. ICLR 2026
    vprl.gif
    Visual Planning: Let’s Think Only with Images
    Yi Xu*, Chengzu Li*, Han Zhou*, Xingchen Wan, Caiqi Zhang, Anna Korhonen, and Ivan Vulić
    The Fourteenth International Conference on Learning Representations (to appear). šŸ†šŸ„‰ #3 paper of the day at HuggingFace šŸ¤— , 2026
  4. ICLR 2026
    mass.jpeg
    Multi-Agent Design: Optimizing Agents with Better Prompts and Topologies
    Han Zhou, Xingchen Wan, Ruoxi Sun, Hamid Palangi, Shariq Iqbal, Ivan Vulić, Anna Korhonen, and Sercan Ɩ. Arık
    The Fourteenth International Conference on Learning Representations (to appear), 2026
  5. ICLR 2025
    bridge.png
    From Few to Many: Self-Improving Many-Shot Reasoners Through Iterative Optimization and Generation
    Xingchen Wan, Han Zhou, Ruoxi Sun, Hootan Nakhost, Ke Jiang, and Sercan Ɩ. Arık
    In The Thirteenth International Conference on Learning Representations, 2025
  6. ACL 2025
    astuterag.png
    Astute RAG: Overcoming Imperfect Retrieval Augmentation and Knowledge Conflicts for Large Language Models
    Fei Wang, Xingchen Wan, Ruoxi Sun, Jiefeng Chen, and Sercan O Arik
    In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2025
  7. NeurIPS 2024
    ioes.png
    Teach Better or Show Smarter? On Instructions and Exemplars in Automatic Prompt Optimization
    Xingchen Wan, Ruoxi Sun, Hootan Nakhost, and Sercan Ɩ. Arik
    In Advances in Neural Information Processing Systems 37. ā˜ļø Powers the Google Cloud Vertex AI Prompt Optimizer , 2024