Xingchen Wan

Xingchen Wan

Research Scientist

Google

About me

I am a Research Scientist at Google based in the San Francisco Bay Area.

I did my DPhil (the Oxford way of saying PhD) in the Machine Learning Research Group, Department of Engineering Science, University of Oxford. I was also a Clarendon Scholar and a member of St John’s College, both at the University of Oxford. I previously interned at Google and Meta.

Academic Services

Reviewer/program committee member at ACL (2023-24), AutoML-Conf (2023-24), COLM (2024), CVPR (2024), ECCV (2024), EMNLP (2023-24), ICLR (2024-25), ICML (2023-24), JMLR, Machine Learning, NeurIPS (2022-23), WACV (2022-24), etc.

Area chair/senior program committee member at NeurIPS (2024), ICML (2025); Action editor at TMLR.

Interests
  • Large language models
  • Bayesian optimization
  • Automated machine learning (AutoML)
  • Machine learning on graphs
Education
  • DPhil (PhD), Machine Learning, 2019 - 2023

    University of Oxford

  • BA, MEng, Engineering Science. First-class Honours, 2015 - 2019

    University of Oxford

Recent News

💡 Preprint: Automating multi-agent designs
Our new work on automating prompt and topology design in multi-agent LLM systems is now available as an arXiv preprint!
📄 ICLR acceptance: Self-improving long-context reasoners
💡 Preprint: Astute RAG
Presenting AstuteRAG, our new work on overcoming imperfect retrieval and knowledge conflict.
📄 3 papers accepted at NeurIPS 2024 + 1 paper accepted at EMNLP 2024!

Publications

View all listed publications or view by tags. For a complete list including preprints & working papers, refer to my Google Scholar.
(2025). From Few to Many: Self-Improving Many-Shot Reasoners Through Iterative Optimization and Generation. International Conference on Learning Representations (ICLR).

PDF Cite Abstract

(2025). Multi-Agent Design: Optimizing Agents with Better Prompts and Topologies. arXiv preprint arXiv:2502.02533.

PDF Cite Abstract

(2024). Teach Better or Show Smarter? On Instructions and Exemplars in Automatic Prompt Optimization. Advances in Neural Information Processing Systems (NeurIPS).

PDF Cite Abstract OpenReview

(2024). UQE: A Query Engine for Unstructured Databases. Advances in Neural Information Processing Systems (NeurIPS).

PDF Cite Abstract OpenReview

(2024). Bayesian Optimization of Functions over Node Subsets in Graphs. Advances in Neural Information Processing Systems (NeurIPS).

PDF Cite Code Abstract OpenReview

(2024). Fairer Preferences Elicit Improved Human-Aligned Large Language Model Judgments. Empirical Methods in Natural Language Processing (EMNLP).

PDF Cite Code Abstract

(2024). Astute RAG: Overcoming Imperfect Retrieval Augmentation and Knowledge Conflicts for Large Language Models. arXiv preprint arXiv:2410.07176.

PDF Cite Abstract

(2024). Batch Calibration: Rethinking Calibration for In-Context Learning and Prompt Engineering. International Conference on Learning Representations (ICLR).

PDF Cite Google Research Blog Abstract OpenReview

(2024). Adaptive Batch Sizes for Active Learning: A Probabilistic Numerics Approach. International Conference on Artificial Intelligence and Statistics (AISTATS).

PDF Cite Code Abstract

(2024). Working Memory Capacity of ChatGPT: An Empirical Study. AAAI Conference on Artificial Intelligence (AAAI).

PDF Cite Code Abstract

Experience

 
 
 
 
 
Google Research, Cloud AI Team
Research Scientist
February 2024 – Present Sunnyvale, CA, US
 
 
 
 
 
Google Research, Cloud AI Team
Research Intern
October 2022 – June 2023 Sunnyvale, CA, US & London, UK
 
 
 
 
 
Meta Research
Research Intern
May 2022 – September 2022 London, UK
 
 
 
 
 
Oxford-Man Institute of Quantitative Finance, University of Oxford
Research Intern
August 2018 – September 2018 Oxford, UK
 
 
 
 
 
Morgan Stanley
Sales and Trading Summer Analyst
June 2018 – August 2018 London, UK

Accomplish­ments

University of Oxford
Clarendon Scholarship
Department of Engineering Science, University of Oxford
Maurice Lubbock Prize for Best Performance in the Honour School of Engineering Science
Deutsche Boerse Group
Deutsche Boerse Scholarship

Contact