How do we make LLMs reason better without burning the planet?
Working on: curriculum RL, concise chain-of-thought, long-context evaluation
Recent: FastCuRL accepted at EMNLP 2025 ✨
☑️ Hunyuan-MT — WMT 2025 · 2025.09
🥇 1st in 30/31 WMT25 language pairs · contributor
☑️ ConciseR — preprint · 2025.05
walk before you run — concise reasoning via GRPO++ → L-GRPO
☑️ FastCuRL — EMNLP 2025 🎉 · 2025.03
curriculum RL — SOTA with 50% fewer steps, 8 GPUs only
☑️ Counting-Stars — COLING 2025 · 2024.03
multi-evidence, position-aware benchmark for long-context LLMs

