Hanlin Zhang
@hlzhang109.bsky.social
π€ 23
π₯ 43
π 11
CS PhD student @Harvard
https://hanlin-zhang.com
Introducing EvoLM, a model suite with 100+ decoder-only LMs (1B/4B) trained from scratch, across four training stages β π¦ Pre-training π© Continued Pre-Training (CPT) π¨ Supervised Fine-Tuning (SFT) π₯ Reinforcement Learning (RL)
loading . . .
EvoLM: In Search of Lost Language Model Training Dynamics
Modern language model (LM) training has been divided into multiple stages, making it difficult for downstream developers to evaluate the impact of design choices made at each stage. We present EvoLM, ...
https://arxiv.org/abs/2506.16029
3 months ago
1
2
1
New work [JSKZ25] w/ Jikai, Vasilis,
@shamkakade.bsky.social
. We introduce new formulations and tools for evaluating LM capabilities, which help explain observations of post-training behaviors of Qwen-series models. More details: -
hanlin-zhang.com/causal-capab...
-
x.com/_hanlin_zhan...
3 months ago
0
0
0
Highlights from
#ICLR2025
β a brief thread π§΅
5 months ago
1
1
0
reposted by
Hanlin Zhang
Andreas Kirsch
6 months ago
I want to reshare
@brandfonbrener.bsky.social
's @NeurIPSConf 2024 paper on CoLoR-Filter: A simple yet powerful method for selecting high-quality data for language model pre-training! With
@hlzhang109.bsky.social
@schwarzjn.bsky.social
@shamkakade.bsky.social
2
17
9
reposted by
Hanlin Zhang
Sham Kakade
10 months ago
(1/n) π‘How can we speed up the serial runtime of long pre-training runs? Enter Critical Batch Size (CBS): the tipping point where the gains of data parallelism balance with diminishing efficiency. Doubling batch size halves the optimization stepsβuntil we hit CBS, beyond which returns diminish.
2
16
4
reposted by
Hanlin Zhang
Yuda Song
10 months ago
LLM self-improvement has critical implications in synthetic data, post-training and test-time inference. To understand LLMs' true capability of self-improvement, we perform large-scale experiments with multiple families of LLMs, tasks and mechanisms. Here is what we found: (1/9)
1
12
5
you reached the end!!
feeds!
log in