Yaqi Xia (夏亚奇)
PhD, School of Computer Science, Wuhan University
I am Yaqi Xia, I am currently working toward the Ph.D. degree in computer science with Wuhan University under the supervision of Prof. Dazhao Cheng. My research interests include distributed deep learning model training and high-performance computing system for AI/ML. Before pursuing my Ph.D., I obtained both my Bachelor's and Master's degrees from Xidian University, where I had the privilege of being mentored by Prof. Rui Song.
Wuhan University
Ph.D. in Artificial Intelligence Sep. 2021 -
Xidian University
M.S. in Electronics and Communication Engineering Sep. 2018 - Jul. 2021
Xidian University
B.S. in Communication Engineering Sep. 2014 - Jul. 2018
Research Center for Graph Computing, Zhejiang Lab
Research Intern Aug. 2023 - Dec. 2023
Yaqi Xia†, Weihu Wang†, Donglin Yang, Xiaobo Zhou, Dazhao Cheng(† equal contribution)
2025 USENIX Annual Technical Conference (ATC) 2025 ConferenceCCF-A
We introduce Voltrix-SpMM, a revolutionary GPU kernel design for sparse matrix-matrix multiplication.
Hulin Wang, Yaqi Xia, Donglin Yang, Xiaobo Zhou, Dazhao Cheng
ACM SIGPLAN Symposium on Principles and Practice of Parallel Programming (PPoPP) 2025 ConferenceCCF-A
We introduce CCFuser, a novel framework designed for efficient training of MoE models.
Yaqi Xia, Zheng Zhang, Donglin Yang, Chuang Hu, Xiaobo Zhou, Hongyang Chen, Qianlong Sang, Dazhao Cheng
IEEE Transactions on Parallel and Distributed (TPDS) 2024 JournalCCF-A
This work introduces Sven, a co-designed algorithm-system library aimed at accelerating TGNN training on a multi-GPU platform.
Weihu Wang, Yaqi Xia, Donglin Yang, Xiaobo Zhou, Dazhao Cheng
The International Conference for High Performance Computing, Networking, Storage, and Analysis (SC) 2024 ConferenceCCF-A
We introduce EcoRec, an advanced library that boosts DLRM training by integrating TT decomposition with distributed training.
Yaqi Xia, Donglin Yang, Xiaobo Zhou, Dazhao Cheng
The International Conference for High Performance Computing, Networking, Storage, and Analysis (SC) 2024 ConferenceCCF-A
In this paper, we introduced HyDRA, a pioneering framework for sampling-based GNN training on large-scale graphs.
Hulin Wang, Donglin Yang, Yaqi Xia, Zheng Zhang, Qigang Wang, Jianping Fan, Xiaobo Zhou, Dazhao Cheng
IEEE Transactions on Computers (TC) 2024 JournalCCF-A
We present Raptor-T, a cutting-edge transformer framework designed for handling long and variable-length sequences. Raptor-T harnesses the power of the sparse transformer to reduce resource requirements for processing long sequences while also implementing system-level optimizations to accelerate inference performance.
Zheng Zhang, Yaqi Xia, Hulin Wang, Donglin Yang, Chuang Hu, Xiaobo Zhou, Dazhao Cheng
IEEE Transactions on Parallel and Distributed (TPDS) 2024 JournalCCF-A
In this paper, we present the design and implementation of MPMoE, a high-performance library that accelerates MoE training with adaptive and memory-efficient pipeline parallelism.
Yaqi Xia, Zheng Zhang, Hulin Wang, Donglin Yang, Xiaobo Zhou, Dazhao Cheng
The 32nd International Symposium on High-Performance Parallel and Distributed Computing (ACM HPDC) 2023 ConferenceCCF-BBest Paper Nomination
This paper presents Sven, an algorithm and system co-designed TGNN training library for the end-to-end performance optimization on multi-node multi-GPU systems.
Yaqi Xia†, Yan Xia†, Wei Li, Rui Song, Kailang Cao, Uwe Stilla(† equal contribution)
Proceedings of the 29th ACM international conference on multimedia (ACM MM) 2021 ConferenceCCF-A
We tackle the problem of object completion from point clouds and propose a novel point cloud completion network employing an Asymmetrical Siamese Feature Matching strategy, termed as ASFM-Net.