Research

We're not here to ride the AI hype bubble. We believe robust and fundamental technical breakthroughs applied to real-world systems is what really moves the needle.

November 11, 2025
Charles O'Neill
Jonathon Liu

BYO SWE-grep: automatically train blazing fast search sub-agents on your knowledge base (Pt. 1)

RL-trained search subagents that learn your knowledge base’s structure for fast, reliable retrieval

Research
October 30, 2025
Charles O'Neill
Harry Partridge
Max Kirkby
Jonathon Liu
Paras Stefanopoulos

Lumina: building self-improving evaluation through customer-in-the-loop refinement

Lumina: an adaptive evaluation engine that learns to judge like a subject matter expert.

Research
October 28, 2025
Charles O'Neill
Jonathon Liu
Kimbrian Canavan
Max Kirkby
Mudith Jayasekara

Attention-based attribution: what your model is actually looking at

Cosine similarity is cosplay. Attention is attribution.

Research
October 28, 2025
Charles O'Neill
Harry Partridge
October 27, 2025
Charles O'Neill
Max Kirkby

Training loss predicts evaluation performance, even for non-verifiable tasks

Loss: the cheapest evaluation you’ll ever run.

Research
October 27, 2025
Harry Partridge
Charles O'Neill

Robust, sample efficient SFT with prompt mutations

Low-KL divergence prompt mutations: better performance at a fraction of the cost.

Research
October 15, 2025
Charles O'Neill
Jonathon Liu
Harry Partridge
Max Kirkby
Mudith Jayasekara

Iterative SFT (iSFT): dense reward learning

Iterative SFT: dense, high-bandwidth learning

Research
October 12, 2025
Charles O'Neill
Max Kirkby
Harry Partridge
Jonathon Liu

Write small, learn forever: rank-1 LoRA for continual learning

Why rank-1 LoRA updates might be the missing link between static fine-tuning and truly continuous, live-on-GPU learning.

Research
October 10, 2025
Max Kirkby
Charles O'Neill

Practical LoRA Research

Fine-tuning at Scale: What LoRA Gets Right (and LoRA-XS Doesn’t).

Research
May 8, 2025
Mudith Jayasekara
Charles O'Neill
Max Kirkby

Do transformers notice their own mistakes? Finding a linear hallucination detector inside LLMs

A linear signal in LLMs reveals hallucinations, is detected by a frozen observer, and steered with a single vector.

Research
February 15, 2025
Charles O'Neill
Mudith Jayasekara
Max Kirkby

Resurrecting the salmon: seeing clearer inside LLMs with domain-specific SAEs

A powerful, efficient, and domain-robust strategy for safeguarding medical-text generation.

Research
January 13, 2025
Charles O'Neill
Mudith Jayasekara
Max Kirkby

Why mechanistic interpretability needs a paradigm inversion

The conventional scaling paradigm for language models themselves may be fundamentally misaligned with interp.

Research

Join the research team

See open roles