about me

Han Fang is an AI Research Scientist at Meta’s SuperIntelligence Labs, with research interest in RL and agents. In 2023-2024, he led post-training of Llama 2 & 3, debut Meta AI and grew it to 1 billion MAU. He drove the integrated training runs, core capabilities, tool use and orchestration, and data flywheel. In 2025, he focuses on teaching models to think, reason, and use tools. He is currently pushing the research frontier of Agents and RL, and has a deep interest in self-improvement.

Han holds a PhD in Applied Mathematics and has published in top-tier venues with 10K+ citations. He is a recipient of the President’s Award to Distinguished Doctoral Students, the Woo-Jong Kim Dissertation Award, and the Excellence in Research Award.

Google Scholar / CV / Linkedin / Twitter

blog

The Central Dogma of Artificial Intelligence

February 2026

Every mature science has its central dogma. Biology has DNA → RNA → Protein. What is ours? Intelligence is the compression of experience into generalization.

The RL Environment Field Guide

January 2026

A practical guide to RL environments using Pokemon Red as a case study. Covers the agent-environment loop, observation spaces, reward design, and credit assignment.

Post-training 101: A Hitchhiker's Guide

September 2025

A comprehensive guide to post-training techniques for LLMs, covering supervised fine-tuning, RLHF, reward models, and practical implementation details.

View all posts →

featured papers

Generalized Parallel Scaling with Interdependent Generations

Harry Dong, David Brandfonbrener, Eryk Helenowski, Yun He, Mrinal Kumar, Han Fang, Yuejie Chi, Karthik Abinav Sankararaman · ICML 2026

Think Smarter not Harder: Adaptive Reasoning with Inference Aware Optimization

Zishun Yu, Tengyu Xu, Di Jin, Karthik Abinav Sankararaman, Yun He, Wenxuan Zhou, Zhouhao Zeng, Eryk Helenowski, Chen Zhu, Sinong Wang, Hao Ma, Han Fang · ICML 2025

The Perfect Blend: Redefining RLHF with Mixture of Judges

Tengyu Xu, Eryk Helenowski, Karthik Abinav Sankararaman, Di Jin, Kaiyan Peng, Eric Han, Shaoliang Nie, Chen Zhu, Hejia Zhang, Wenxuan Zhou, Zhouhao Zeng, Yun He, Karishma Mandyam, Arya Talabzadeh, Madian Khabsa, Gabriel Cohen, Yuandong Tian, Hao Ma, Sinong Wang, Han Fang · arXiv 2024

Linformer: Self-Attention with Linear Complexity

Sinong Wang, Belinda Z. Li, Madian Khabsa, Han Fang, Hao Ma · arXiv 2020

View all publications →

news

Meta AI reached 1 billion MAU

Improved Meta AI's multilinguality, enabled roll-out to 12 languages and 40+ countries. Blog · News

2025

Launched voice mode and photo editing in Meta AI

Launched updated Llama 3 model for voice mode. Improved Planner for photo editing with multimodal inputs. Blog

Connect 2024

Launched Llama 3 on Meta AI

Developed Meta AI's online RL with Mixture of Judges, improving reasoning, instructions following, safety. Blog

2024

Launched Meta AI with Llama 2

Debut Meta AI into Family of Apps. Created Orchestrator for tool use. My talk at Connect

2023

Meta AI Few-Shot Learner (FSL)

Developed FSL for detecting new forms of harmful content across 100+ languages. Blog

2021

Training AI to detect hate speech

Built RL-based framework to E2E optimize hate speech classifiers. Blog

2020