Location via proxy:   
[Report a bug]   [Manage cookies]                

Hi! I'm Pengcheng. I am a research scientist at Google Deepmind, working on Gemini Code and research projects in natural language to code generation. Before that, I was a Ph.D. student at Language Technologies Institute, Carnegie Mellon University.

Outside of work, I am a student pilot based at Palo Alto Airport. I also train Coke using RLHF :)

Research Papers

Please see my Google Scholar page for recent publications.

Past Industrial Experiences

  • Researcher Intern, Microsoft Semantic Machines
  • Part-time Research Collaborator, Facebook AI Research
  • Research Intern, Facebook AI Research London
  • Research Intern, Microsoft Research Cambridge, UK
  • Research Intern, Microsoft Research
  • Research Intern, Noah's Ark Lab, Huawei
  • Research Intern, Microsoft Research Asia

Professional Services

  • Area Chair: ICLR 2024, ICLR 2025, ACL 2025
  • Reviewer: ACL (outstanding reviewer @ ACL 2020), EMNLP, NAACL, NeurIPS, ICML (top 33% reviewer @ ICML '20), ICLR, etc.

Talks and Coding

  • Stanford CS224N DL with NLP: Code Generation (slides)
  • TranX: a general-purpose syntax-driven neural semantic parser
    • Strong results on six semantic parsing benchmarks
  • pytorch_basic_nmt: a basic implementation of attentional nerual seq2seq models
    • Used for instructional purposes in Stanford CS224N Nautral Language Processing with Deep Learning and CMU 11-731 Machine Translation and Sequence-to-Sequence Models.