Catalyst is an interdisciplinary machine learning and systems research group exploring problems to automate learning systems. Our research spanning multiple layers of the machine learning and system stack. Our group is a collaboration between researchers from the Machine Learning Department, Computer Science Department and Electrical & Computer Engineering Department at the Carnegie Mellon University.


Machine Learning Compilation for Large Language Models

A universal solution that allows any language model to be deployed natively on a diverse set of hardware backends and native applications.

Read more »

FlexFlow Serve

Low-Latency, High-Performance LLM Serving

Read more »


End-to-end compilation of ML applications with dynamic and irregular control flow and data structure accesses

Read more »

Apache TVM Stack

TVM: An Automated End-to-End Optimizing Compiler for Deep Learning

Read more »

Hyperband and ASHA

Principled early-stopping approaches for hyperparameter optimization

Read more »


Automatically Discovering Fast Parallelization Strategies for DNN Training

Read more »


The Tensor Algebra SuperOptimizer for Deep Learning

Read more »


A Scalable Tree Boosting System

Read more »


The rapid advance in ML models and ML-specific hardware makes it increasingly challenging to build efficient and scalable learning systems that can take full advantage of the performance capability of modern hardware and runtime environments. Today's ML systems heavily rely on human effort to optimize the training and deployment of ML models on specific target platforms. Unlike conventional application domains, learning systems need to address a continuously growing complexity and diversity in machine learning models, hardware backends, and runtime time environments. Our response to this unique challenge in ML systems is Catalyst (CMU automated learning systems group), a joint research group across the area of machine learning, systems, programming languages, and computer architecture. Our mission is to build ML algorithms and learning systems that automate cross stack optimizations by leveraging mathematical and statistical properties of ML computations and by co-designing systems, hardware, and ML algorithms.



Tianqi Chen
Assistant Professor
Zhihao Jia
Assistant Professor
Ameet Talwalkar
Assistant Professor
Eric Xing


Yue Zhao
Ph.D., 2023. Assistant Professor at USC.
Pratik Fegade
Ph.D., 2023. Google.
Byungsoo Jeon
Ph.D., 2024. OctoAI.
Muyang Li
M.S., 2023. PhD student at MIT
Tony Xi
M.S., 2023. Now at Apple
Zeyu Wang
M.S., 2023. Now at Yahoo
Andrew Gu
M.S., 2022. Now at Meta
Zhongyu Chen
M.S., 2022. Now at LinkedIn
Bowen Chen
M.S., 2022. Now at Annapurna Labs
Neeraj Aggarwal
M.S., 2022. Now at Applied Intuition
Balamurugan Marimuthu
M.S., 2022. Now at SambaNova
Eric Zheng
M.S., 2022. Now at Citadel
Michelle Ma
M.S., 2022.
Sheng Xu
M.S, 2020. Now at AWS.
Peiyuan Liao
B.S, class of 2022
Poojan Palwai
B.S, class of 2023
Rae Wong
B.S, class of 2023