Encoderfile’s New Format: Why a “Dull” Design Wins
Encoder models don’t chat, and they don’t get much attention for it. But they’re behind most of the NLP that actually runs in production, powering…
Encoder models don’t chat, and they don’t get much attention for it. But they’re behind most of the NLP that actually runs in production, powering…
The AI revolution has triggered a massive shift in daily life for knowledge workers. Developers, writers, analysts, and designers have seen their output transform dramatically…
On March 24, 2026, LiteLLM, a Python package, with over 95 million monthly downloads, was compromised. Versions 1.82.7 and 1.82.8 on PyPI contained a credential-stealing…
Side A: Turtles all the way down / Side B: Mo’ tokens mo’ problems If you’ve been around long enough in anything you start to…
We are happy to announce the release of llamafile 0.10.0. Since our previous announcement, we’ve rebuilt llamafile from the ground up, following an approach that…
When the hardest part of building shifts, so does leadership We have gotten very good at building software. We have not gotten equally good at…
At Mozilla.ai, we believe useful machine intelligence shouldn’t require centralizing sensitive user data. Federated learning offers a practical path toward collective intelligence without surveillance-style data…
Software engineering is going through a shift that feels small on the surface but changes something fundamental: code is no longer scarce. For decades, writing…
Every AI model has blind spots. It might overlook context, lean toward certain patterns, or fill gaps with confident guesses. When you’re using an AI…
A core part of building any-llm is making sure it is present where developers already are. Over the past few months, we’ve integrated any-llm into…
Go where the models are When we released any-llm v1.0 last year, the goal was simple: one interface to use any model, cloud or local,…
Introduction In Evaluating Multilingual, Context-Aware Guardrails: Evidence from a Humanitarian LLM Use Case, we explored how guardrails responded to the same policies and prompts in…
Effective large language model (LLM) evaluation needs to be context-, language-, task-, and domain-specific. As developers gravitate towards custom performance benchmarks, they are also increasingly…
The recent State of AI report by OpenRouter and Andreessen Horowitz (a16z) offers compelling insights into the growing adoption of open-weight LLMs. It categorizes models…
Most teams don’t wake up asking for “more AI.” They just want less busywork and fewer tabs open. In practice, that usually means one thing:…