me

Yixuan Su (苏熠暄)

I am an NLP researcher and engineer at Cohere working on retrieval-augmented language models. I obtained my Ph.D. at Language Technology Lab, University of Cambridge, advised by Prof. Nigel Collier. Previously, I recieved my M.Phil in Machine Learning, Speech and Language Technology (MLSALT) from Engineering Department of University of Cambridge. And I completed my Bachelor degree from Engineering Department of Beijing Institute of Technology.

Email: ys484 AT outlook.com

Work Experience

  • May 2023 - Present
    Research Scientist at Cohere

Education

I also collaborated with many awesome people.
News
  • [2023-Nov-06] Completed my Ph.D. Viva and now I am officially a Dr.! Huge thanks to my advisor, Prof. Nigel Collier, and my thesis committee, Prof. Andreas Vlachos and Prof. Mirella Lapata!
  • [2023-Oct] One paper accepted to TLLM 2023, one paper accepted to NeurIPS 2023, and two papers accepted to EMNLP 2023.
  • [2023-May] Start my journey at Cohere!
  • [2023/05/23] Release PandaGPT, the first foundation model capable of instruction-following data across six modalities.
  • [2023/05/04] Release OpenAlpaca, a fully open-source instruction-following model based on OpenLLaMA.
  • [2023/02/14] Our manuscript "Contrastive Search Is What You Need For Neural Text Generation" is accepted to TMLR 2023!
  • [2022/11/22] Released a technical report, "An Empirical Study On Contrastive Search And Contrastive Decoding For Open-ended Text Generation", that compares Contrastive Search with Meta's recently proposed Contrastive Decoding. [arxiv] [code]
  • [2022/10/26] Released our new preprint "Contrastive Search Is What You Need For Neural Text Generation". Check it out! [arxiv] [code]
  • [2022/09/14] Two papers (including SimCTG) accepted to NeurIPS 2022. See you in New Orleans!
  • [2022/08/02] One paper accepted to CIKM 2022. See you in Atlanta!
  • [2022/05/06] I am excited to release my latest work, MAGIC, the SOTA method on zero-shot multi-modal text generation tasks (e.g., zero-shot image captioning and visually grounded story generation). Check it out! [arxiv] [code]
  • [2022/04/08] One paper accepted to NAACL 2022. See you in Seattle!
  • [2022/02/24] Two papers accepted to ACL 2022. See you in Dublin!
  • [2022/02/15] I am excited to share my recent work, SimCTG, the SOTA method on open-ended text generation tasks. Check it out! [arxiv] [code]

Publications

(*: equal contribution)

Selected Publications
Full Publications (In Chronological Order)
Invited Talks
  • January 2023, IR Group, University of Glasgow
  • October 2022, NLG Student Webinar, Chinese Information Processing Society of China
  • August 2022, MLNLP Webinar
  • August 2022, NLP Group, Princeton University
  • June 2022, Language Technology Lab, University of Cambridge
  • April 2022, NLP Group, University of Washington
  • April 2022, Language Team, DeepMind, London, UK
  • March 2022, NLP Group, University of Oxford
  • March 2022, NLP Group, Nara Institute of Science and Technology
  • February 2022, NLP Group, Tencent AI Lab, Shenzhen, China
Feel free to drop me an email if you'd like me to give a talk at your event/seminar. Currently, I have two topics: (1) task-oriented dialogue systems [1]; (2) contrastive representation learning on pre-trained language model [2]; open-ended text generation [3]; and plug-and-play multi-modal text generation [4].
Students I Mentored
  • Yinhong Liu, PhD student at University of Cambridge. (Co-advise with Prof. Nigel Collier)
  • Tiancheng Hu, PhD student at University of Cambridge. (Co-advise with Prof. Nigel Collier)
  • Meiru Zhang, PhD student at University of Cambridge. (Co-advise with Prof. Nigel Collier)
  • Professional Service
      Program Committee Member/Reviewer:
      NeurIPS (2023-), ACL (2020-), EMNLP (2021-), NAACL (2021), AAAI (2021), EACL (2021), TASLP (2022-), ACL Rolling Review (2021-)