INDEV Sovereign AI

We deploy, fine-tune, and operate domain-specific LLMs inside European banks — air-gapped, audited, and under SLA.

Production AI, operated inside your perimeter.

INDEV Sovereign AI deploys, fine-tunes, and operates domain-specific LLMs inside the regulatory perimeter of European banks. Your data stays in your environment, every inference is audited, and the model is trained on your corpus and operated under SLA. The same operational discipline that has kept INDEV's regulated software running in systemic banks for twenty years — now applied to the AI layer.

Why existing options fail

Hyperscaler AI ships your data out

Hyperscaler AI assumes data leaves the perimeter to be processed. Your supervisor does not. DORA, the EU AI Act, and national regulators draw a line that contracts alone cannot cross.

Open-source AI has no operations

Open-source LLM tooling is developer-grade. No SLA. No audit trail. No lifecycle management. No incident response. Running it in production requires an operating team most banks don't have.

Retrieval alone doesn't scale

Retrieval-only architectures don't scale to enterprise corpora. Decades of regulated documents don't fit in a context window — and don't want to be re-tokenized on every query.

You don't control the model lifecycle

Hosted-inference contracts tie your application to a vendor's model versioning, deprecation, and pricing. When their model changes, yours follows. DORA treats this as third-party concentration risk.

What we operate

Inference serving

vLLM-based serving with an OpenAI-compatible API, deployed on your GPU pool inside the cluster. No external endpoints. No internet egress.

Inference serving

Audit and observability

Self-hosted Langfuse captures every prompt, response, and trace. Full audit trail stays inside your perimeter.

Audit and observability

Evaluation

Per-field eval harnesses measure model and quantization performance on your data. Nothing ships without passing the eval.

Evaluation

Fine-tuning

Domain-corpus fine-tuning on dense and MoE models — LoRA, QLoRA, or full-parameter, selected by your eval results.

Fine-tuning

Model lifecycle

Versioning, A/B deployment, rollback, retirement — under change management, with the same release discipline as your regulated production systems.

Model lifecycle

Compliance

ISO 27001 and PCI DSS in place. DORA-aligned operations. EU AI Act human-in-the-loop by default.

Compliance

Models we run in production

Gemma 4 and Qwen 3.6, in dense and Mixture-of-Experts variants — open-weights only. Selected per workload by your eval set, not by a vendor relationship. As the open-weights frontier evolves, we evaluate and migrate under change management.

Considering Sovereign AI inside your perimeter?

Let's start with a discovery call.