Empirical Science and Applied Research on Generative Models
Experiments on the open questions in generative models.
The most important questions about generative models are still open — what post-training actually changes inside the model, when alignment holds under pressure, whether RL improves reasoning or just its appearance. Applied Models runs focused experiments to find out.
The model landscape is moving faster than our understanding of it. That gap is the work. Seven directions, each with concrete questions worth pursuing precisely — and enough control over the experimental setup to actually answer them.
Models and datasets: Hugging Face organization
Artifact Index
All published work, in one place.
Every experiment, article, and notebook — indexed by type. Click any row to open the full record.
| Type | Title | Reference | Meta | Action |
|---|---|---|---|---|
| Notebooks |
0001 Agentic Evals Baseline Notebook
Python notebook |
A Python notebook for prompt fixtures, scoring checks, and baseline observations for the first experiment. | 2026-02-28 · Python notebook | |
| Experiments |
Experiment 0001: Agentic Evals for Small Models
Eval suite |
A compact evaluation suite for planning, tool choice, self-correction, and distractor resistance in smaller open models. | 2026-02-28 · In progress · Eval suite | |
| Articles |
Operating Principles for Applied Models
Working note |
How Applied Models decides what to publish, what counts as a valid result, and what the research record is for. | 2026-02-28 · Working note |