Tag: Post-Training
All the articles with the tag "Post-Training".
-
The Mercor Breach: What 4TB of Stolen Data Reveals About How Frontier AI Labs Actually Train Models
· 22 min readA $10B AI data vendor was breached, exposing 84 Airtable workspaces of training data for OpenAI, Anthropic, Apple, Amazon, and Meta. This post analyzes what the public reporting reveals about each lab's evaluation methodology — rubric design, RLHF pipelines, and quality control — and what it means for the industry.
-
Instruction Following: What Models Get Wrong and How to Fix It with Better Post-Training Data
· 36 min readLLMs can write poetry and solve math, but ask them to 'respond in exactly 3 bullet points using only lowercase' and they stumble. This post dissects the taxonomy of instruction-following failures and provides a practical playbook for building post-training data that actually fixes them.