Qianli Ma

(+86)17855801919 Fazzie17855801919 fazzie@qq.com
370+ followers fazzie-key.cool QianliMa Google Scholar fazzie

I possess extensive engineering experience across both Machine Learning Systems (MLSys) and Large Language Model Algorithms. My goal is to advance next-generation AGI systems in order to create larger and better models. I am deeply passionate about the latest technologies and actively contribute to the open-source community as a core contributor to several popular open-source AI projects.

EDUCATION BACKGROUND


National University of Singapore
2022.8 - 2024.1
 Master of Computer Science
Singapore
Zhejiang University
2018.9 - 2022.7
 B.Eng in Electronic Science and Technology
Hangzhou, China

Other Honors: Second prize in the National High school Mathematics Competition(2017)

WORK EXPERIENCE


ByteDance  Seed
2023.12 - 2025.12
 Staff Research Engineer&&Leader of Multimodal Training Framework
Shanghai

As one of the earliest members of the Seed Team, I focused on AI infrastructure and large-scale training systems for LLMs and multimodal foundation models across pre-training and post-training. Starting from the first-generation Seed models, I supported large-scale training on 10,000-GPU clusters. I led a small team to build VeOmni, an open-source multimodal training system, and supported training at the scale of thousands of GPUs. I was deeply involved in the R&D of the core Seed 1.5 to Seed 2.0 model families, including reasoning and multimodal models, as well as the UI-TARS series of GUI agent models.

 Projects Highlights
  • VeOmni: led the development of a PyTorch-native multimodal training system for pre-training and post-training, supporting model initiatives including Seed core models and UI-TARS.
  • Core model R&D: participated in the development of the core Seed 1.5 to Seed 2.0 model families, covering major reasoning and multimodal model efforts.
  • UI-TARS series: contributed to the research and system infrastructure behind the UI-TARS family of native GUI agent models.
  • Open-source systems: contributed to veScale and verl for distributed LLM training and RL post-training.
ByteDance  AML
2023.6 - 2023.12
  LLMs Research Intern at Seed-Project
Shanghai

Worked on LLM post-training and agent research, with projects directly leading to publications on process reward modeling, SFT data selection, and data-analysis agents.

  • Process Reward Modeling: Built the full data-processing, training, and evaluation pipeline for step-level reward models, leading to the paper Let's Reward Step by Step: Step-Level Reward Model as the Navigators for Reasoning.
  • SFT Data Selection: Co-developed DavIR, a model-centric data selection method showing that 6% of Alpaca data can outperform full-dataset training, later published at ACL 2025.
  • Agent for Data Analysis: Built InfiAgent-DABench, including the benchmark, agent infrastructure, and evaluation pipeline for data-analysis tasks, later published at ICML 2024.
HPC-AI Technology
2022.7 - 2023.5
  Machine Learning System Engineeer
Singapore

Joined as Employee #15 and worked on large-model systems and open-source products from Seed to Series A.

SenseTime   Large model training
2021.12 - 2022.6
  AI Researcher Internship
Hangzhou

Participated in the development of SenseTime's early Megatron-style large-model training framework.

Huawei 2012 Lab   Distributed Parallel Lab
2021.7 - 2021.12
  AI Engineering Internship
Hangzhou

Contributed to MindSpore and MindSpore Lite on device-side GPU inference and runtime infrastructure.

PUBLICATION


KNOWLEDGE & SKILLS


CLUBS & ORGANISATIONAL EXPERIENCE


Zhejiang University Internet Society   Technology department  AI lab
2021.10 - 2022.8
String Program   Technology department   Member of the machine learning subdepartment
2020.7 - Present
Zhejiang University Electroacoustic Orchestra   Drummer of Six o'clock studio band
2018.11 - 2021.2