Open & Offline Models
Overview
ThinkNimble has a strong interest in open and offline-capable models. As AI moves from toy to infrastructure, the distinction between âopen weightâ (just the model) and âfully openâ (weights + training data + process) becomes security-critical. This note tracks the landscape.
OLMo 3: The Fully Open Benchmark
Ai2âs OLMo 3 releases what most âopenâ models donât: full training data (Dolma 3, ~9.3T tokens), training process documentation, and checkpoints - not just weights. Anthropic showed that just 250 poisoned documents can backdoor an LLM of any size, making auditable training data a security requirement. OLMo 3 trains on ~6x fewer tokens than competitors while reaching similar performance. Competitors in the âfully openâ space: Stanfordâs Marin, Swiss AIâs Apertus - but this remains a small category.
Connections
- Verification Complexity - Open training data is itself a verification tool: you can audit what the model learned and why.
- AI Agent Engineering - As agents become critical infrastructure, model auditability becomes a security requirement.