Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

Zhihao Jia

Assistant Professor
Computer Science Department
Carnegie Mellon University
zhihao@cmu.edu

I am an assistant professor in the Computer Science Department at Carnegie Mellon University, where I work on computer systems and machine learning as part of CMU Catalyst Group and Parallel Data Lab. I received my PhD from the Computer Science Department at Stanford University in 2020, where I was co-advised by Alex Aiken and Matei Zaharia. Before coming to Stanford, I received my bachelor degree in Computer Science from the Special Pilot CS Class supervised by Andrew Yao at Tsinghua University.

Research interests: My research interests lie in the intersection of computer systems and machine learning (ML). In particular, my current research focuses on building efficient, scalable, and high-performance software systems for emerging ML applications, such as large language models and generative AI tasks.

I am actively looking for strong and self-motivated students interested in building systems for machine learning and quantum computing to join my group.
(1) Prospective students: if you are interested in working with me as a Ph.D. student, please apply through the CMU CS PhD program and mention me in your application.
(2) Current CMU students: if you are already a graduate, masters, or undergraduate student at CMU, please send me an email and we can find a time to chat.
(3) Prospective students not at CMU: if you are interested in joining my group for remote collaborations, please send me an email with your CV.

Teaching

Projects

  • Mirage is a superoptimizer that automatically generates highly-optimized GPU kernels for ML applications. Mirage can discover kernels up to 3.5x faster than the ones manually implemented by GPU experts.

  • FlexFlow Serve is a compiler and distributed runtime for low-latency, high-performance LLM serving by leveraging tree-based speculative inference and verification.

  • FlexFlow is a deep learning engine that accelerates distributed DNN training by automatically discovering fast parallelization strategies for a specific parallel machine.

  • TASO is a Tensor Algebra SuperOptimizer for deep learning. It optimizes DNN computation graphs using automatically generated graph transformations, achieving up to 3x speedup over existing DNN frameworks. PET further extends TASO by leveraging partially equivalent transformations and automated corrections.

  • Lux is a distributed multi-GPU system for high performance graph processing. Lux achieves fast graph processing by exploiting the aggregate memory bandwidth of multiple GPUs. Lux achieves up to 20x speedup over state-of-the-art graph processing systems.

  • Quartz is a quantum circuit superoptimizer that automatically generates and verifies circuit transformations for arbitrary quantum gate sets. By using these auto-generated transformations, Quartz can outperform existing quantum circuit optimizers on a diversity of gate sets.

  • Legion is a high performance programming system for heterogeneous, parallel machines with complex memory hierarchies.

Awards

Current Ph.D. and Postdoctoral Students

Alumni

Publications

2024

  • Optimal Kernel Orchestration for Tensor Programs with Korch
    Muyan Hu, Ashwin Venkatram, Shreyashri Biswas, Balamurugan Marimuthu, Bohan Hou, Gabriele Oliaro, Haojie Wang, Liyan Zheng, Xupeng Miao, Jidong Zhai, and Zhihao Jia
    In Proceedings of the ACM International Conference on Architectural Support for Programming Languages and Operating Systems (ASPLOS), April 2024.

2023

2022

  • Benchmarking Node Outlier Detection on Graphs
    Kay Liu, Yingtong Dou, Yue Zhao, Xueying Ding, Xiyang Hu, Ruitong Zhang, Kaize Ding, Canyu Chen, Hao Peng, Kai Shu, Lichao Sun, Jundong Li, George H. Chen, Zhihao Jia, Philip S. Yu
    In Proceedings of the Conference on Neural Information Processing Systems (NeurIPS), December 2022.

  • Quartz: Superoptimization of Quantum Circuits
    Mingkuan Xu, Zikun Li, Oded Padon, Sina Lin, Jessica Pointing, Auguste Hirth, Henry Ma, Jens Palsberg, Alex Aiken, Umut A. Acar, and Zhihao Jia
    In Proceedings of the Conference on Programming Language Design and Implementation (PLDI), June 2022.

2021

  • Scaling Implicit Parallelism via Dynamic Control Replication
    Michael Bauer, Wonchan Lee, Elliott Slaughter, Zhihao Jia, Mario Di Renzo, Manolis Papadakis, Galen Shipman, Pat McCormick, Michael Garland, and Alex Aiken.
    In Proceedings of the Principles and Practice of Parallel Programming (PPoPP), February 2021.

2020

2019

2018

2017 and earlier

  • Improving Integer Security for Systems with KINT
    Xi Wang, Haogang Chen, Zhihao Jia, Nickolai Zeldovich, and M. Frans Kaashoek.
    In Proceedings of the Symposium on Operating Systems Design and Implementation (OSDI), Hollywood, CA, October 2012.

  • Undefined Behavior: What Happened to My Code?
    Xi Wang, Haogang Chen, Alvin Cheung, Zhihao Jia, Nickolai Zeldovich, and M. Frans Kaashoek.
    In Proceedings of the Asia-Pacific Workshop on Systems (APSys), Seoul, South Korea, July 2012.

Education

  • 2013.9 - 2020.6: Ph.D. candidate at Stanford University

  • 2009.9 - 2013.6: B.Eng from Yao Class, Tsinghua University

  • 2012.1 - 2012.5: Exchange student at MIT

Experience

  • 2016.6 - 2016.9: Research intern at Los Alamos National Lab
    Working with Galen Shipman and Pat McCormick

  • 2014.6 - 2014.9: Research intern at Microsoft Research Silicon Valley
    Working with Yuan Yu

  • 2012.6 - 2013.6: Research intern at Microsoft Research Asia
    Working with Jiaxing Zhang and Lidong Zhou

  • 2011.2 - 2012.1: Research intern at Microsoft Research Asia
    Working with Ming Wu and Lidong Zhou