I lead the research and development of foundation models at Amazon AGI, where I focus on the co-evolution of algorithms and systems—the critical intersection that makes AI more capable, efficient, and accessible. My team built the models behind Amazon Nova, Amazon Q, Titan, and the distributed training infrastructure powering Amazon Bedrock and SageMaker HyperPod.
I built this team from zero, starting in 2018 with a focus on distributed training and shared representations. What started as a small group of tech leaders and hackers grew into the engine behind foundation models serving millions through AWS.
Before that, I shaped the open-source AI ecosystem as VP and PMC Chair of Apache MXNet, where I co-authored the Gluon interface. I founded GluonNLP—the first toolkit to reproduce BERT with record-setting training speeds. I served on the ONNX Steering Committee and co-founded the Python Data API Standards Consortium. I believe accessible tools and open standards are essential for an AI future that benefits everyone.
I hold an MS in Computer Science from the University of Maryland and a BS from Shanghai Jiao Tong University.