AI
Why open source models are the way forward for insurers

Adam Ben-David
Mar 20, 2026

The shift from proprietary to open source AI models gives insurers new freedom to build flexible, vendor-independent AI systems focused on business value rather than model dependency.
Open-weight models are production-viable. The economics are splitting by task type. Here's what that means for how insurers build.
Two things are true. Closed-weight frontier models from OpenAI, Anthropic, and Google are growing in capability and commercial adoption. Open-weight models are closing the inference performance gap across a widening range of tasks. Both are accelerating.
The interesting question isn't which side wins. It's how the task-type economics split, and what that means for insurance AI infrastructure decisions.
Scaling hit diminishing returns
Adding compute stopped yielding proportional capability gains. Training data at the quality and volume required for frontier improvement ran out.
All model developers shifted back toward foundational research: architecture efficiency, attention mechanisms, synthetic data generation, inference optimisation.
The research edge that large proprietary labs held has narrowed as open-source output volume and technical quality have grown. Closed-weight labs still lead on capability ceiling. The breadth of experimentation across the open-source ecosystem now outpaces what any single organisation can run internally.
The advantage once held by the frontier labs has compressed.
Matching models to task-type
For routine inference workloads, summarisation, document extraction, classification, triage, frontier models are becoming ROI-negative. A smaller open-weight model running on self-hosted infrastructure delivers equivalent output at a fraction of the cost. The production infrastructure for this has hardened: containerised deployment, optimised inference runtimes, model serving frameworks are no longer experimental.
For complex reasoning, multi-step agentic tasks, novel problem types, large models still lead. Competition at the frontier between closed-weight providers is intense. Performance differences are material. The trajectory is clear: capabilities that command a premium today commoditise over time, migrating from frontier models to open-weight alternatives as each generation matures.
Task-type economics for insurance AI
Task | Current direction |
|---|---|
Summarisation, extraction, classification, triage | Open-weight models reaching cost/performance parity |
Complex reasoning, multi-step agentic workflows | Closed-weight frontier models still lead |
Self-hosted inference infrastructure | Production-ready; self-hosting no longer experimental |
Research output breadth | Open Source community accelerating; frontier labs still lead on capability ceiling |
As models become more commoditised, the value for insurers becomes the context layer and orchestration architecture
As model capabilities commoditise, competitive advantage shifts to the context layer and orchestration architecture, not the model itself.
Decision context. Frontier and open-weight models alike lack embedded understanding of insurance-specific risk frameworks, appetite logic, portfolio constraints. The organisations that win will build agentic systems with access to complete decision context
Agentic orchestration. Coordinated agent workflows that handle complex, multi-step processes represent a different order of capability than isolated model calls. An agent pipeline that ingests submission data, routes it through appropriate pricing models, validates outputs against business rules, and flags exceptions for human review is where the durable value sits.
Function-specific opportunities
Underwriting: Agents trained on specific appetite guidelines and historical decisions reduce manual extraction and accelerate quote generation without requiring frontier model inference costs on every submission.
Actuarial: AI incorporating proprietary pricing models and portfolio characteristics allows actuaries to increase model sophistication and iteration speed without rebuilding the analytical layer from scratch.
Portfolio performance: Real-time analysis of risk concentrations and pricing adequacy against organisation-specific data generates insights that directly affect loss ratios at the portfolio level.
What are the AI-mature insurers doing?
The ones building well are testing multiple models, running workloads across open and closed ecosystems, and measuring outputs against consistent evaluation criteria.
The strategic question is whether your AI infrastructure is flexible enough to shift inference workloads as the model landscape evolves. The forward-looking carriers are focused on creating this infrastructure right now.



