I am Rongyi Yu (余荣毅), a Master student at Peking University, Center for Machine Learning Research. I am currently involved in research and engineering on large model training and alignment, with a focus on data attribution, dynamic data selection, and training efficiency optimization for Pretrain / SFT / RLHF (PPO) settings.

My recent work centers on data-centric AI and AI infrastructure, aiming to improve the alignment quality, robustness, and cost efficiency of large language models. I am comfortable with the full research pipeline from algorithm design to engineering implementation, including training, evaluation, reproduction, and performance diagnosis.

My current and previous research interests include:

  • Data Attribution
  • Data-Centric AI
  • LLM Training and Alignment
  • Dynamic Data Selection
  • AI Infrastructure
  • Diffusion Models

🔥 News

  • 2026:  🎉 I will join Peking University as a master’s student at the Center for Machine Learning Research.
  • 2026:  📝 Our paper “LongVidSearch: An Agentic Multi-hop Search Benchmark for Long Videos” is under submission to SIGIR 2026.
  • 2025:  🚀 Participated in research and engineering for DCAI-DataFlex, an open-source data-centric toolkit and dynamic training framework for LLMs.
  • 2025:  🏆 Received multiple honors including National First Prize in the Chinese Mathematics Competition and Top 10 Students at Harbin Institute of Technology.

📝 Publications

  • Rongyi Yu, Chenyuan Duan, et al.
    LongVidSearch: An Agentic Multi-hop Search Benchmark for Long Videos
    SIGIR 2026 Submission

🎖 Honors and Awards

  • 2025 National First Prize, Chinese Mathematics Competition (Mathematics Category)
    Highest-level national honor for mathematics undergraduates.
  • 2025 Top 10 Students, Harbin Institute of Technology
    Highest individual student honor at HIT; only 10 students selected across all campuses from freshman to senior year.
  • 2022, 2024 National Scholarship ×2
    Awarded to only about 2% of undergraduate students each year.
  • 2025 Bronze Medal, China International College Students’ Innovation Competition

📖 Education

  • 2026.09 - 2028.06 Peking University, Center for Machine Learning Research, M.S.
    I will join Peking University in Fall 2026 for my master’s study, co-supervised by Prof. Weinan E and Prof. Wentao Zhang.
  • 2022.09 - 2026.06 Harbin Institute of Technology, School of Mathematics, B.S.
    GPA: 99.53/100, Rank: 1/45
    CET-4: 585, CET-6: 571

🔬 Research Experience

  • 2025.09 - 2026.02 Research Assistant, Peking University, Center for Machine Learning Research
    Advisors: Prof. Wentao Zhang, Dr. Hao Liang
    • Worked on data-centric LLM training and parameter-efficient fine-tuning (PEFT).
    • Studied data quality, data selection, and training stability for supervised fine-tuning.
    • Participated in the development of the DataFlex dynamic training framework.
    • Supported the full pipeline of data scoring / sampling / training / evaluation to improve generation quality and training efficiency.
  • 2024.03 - 2025.08 Research Assistant, Harbin Institute of Technology & HKUST School of Mathematics
    • Collaborated closely with Prof. Yao Li and Prof. Yang Xiang.
    • Conducted research on diffusion-model-based theory and implementation for satellite image alignment and fusion.

💻 Selected Projects

DCAI-DataFlex

Open-source data-centric toolkit / dynamic training framework for large models

  • Problem: In SFT/RLHF training, noisy, redundant, and distribution-shifted data can lead to unstable generation quality and high training cost.
  • Method: As a major contributor, I helped build the dynamic training framework, designed multiple offline domain data selection operators, and developed domain-mixing strategies compatible with Qwen, Llama, and related training pipelines.
  • Result: Verified the framework’s effectiveness in SFT training: using only about 5% of the training data can achieve near full-data MMLU accuracy.
  • Impact: The method can be extended to pretraining, reinforcement learning, and hard-example mining, improving alignment quality and system stability.

🛠 Skills

  • Programming & ML: Python, PyTorch, Transformers, vLLM, DeepSpeed, LLaMA-Factory
  • Mathematics: Probability and Statistics, Stochastic Processes, Optimization, Numerical Analysis
  • Tools: Git, LaTeX, Markdown, Linux