← Blog ·

Building Sovereign AI Infrastructure: Why I Moved Everything Off the Cloud

After years of relying on cloud APIs, I built a fully self-hosted AI stack. Here's what I learned, what I use, and why sovereignty matters more than convenience.

Martin Fournier · March 01, 2026 · 3 min read

title: "Building Sovereign AI Infrastructure: Why I Moved Everything Off the Cloud" excerpt: "After years of relying on cloud APIs, I built a fully self-hosted AI stack. Here's what I learned, what I use, and why sovereignty matters more than convenience." category: "AI & Automation"

The Problem with Convenience

Every SaaS tool, every cloud API, every hosted service is a dependency. And dependencies have a cost — not just the monthly subscription, but the deeper cost of not owning your own data, your own workflows, your own intelligence.

I spent years building systems on top of OpenAI, AWS, and various SaaS layers. The systems worked. They were fast to build. But when a pricing change happened, or an API was deprecated, or a data policy changed — I felt it immediately.

That's when I started thinking seriously about sovereign tech.

What Sovereign Tech Means to Me

Sovereign tech isn't about rejecting all external services. It's about being intentional:

  • Own your data. If a service disappeared tomorrow, what would you lose?
  • Own your workflows. If an API raised prices 10x, could you survive?
  • Own your compute. Where possible, run things locally.

The goal isn't to build everything from scratch — it's to know exactly what you depend on, and to have a plan for each dependency.

My Current Stack

After 18 months of iteration, here's what my sovereign AI infrastructure looks like:

Local LLMs via Ollama

I run Ollama on my local machine and on my home server. Models like deepseek-coder, llama3, and mistral handle 80% of my daily AI tasks without a single token leaving my network.

ollama run deepseek-coder:6.7b

The quality is surprisingly close to cloud models for coding, writing, and reasoning tasks. Not identical — but for most workflows, the privacy and cost savings justify the delta.

Agent Orchestration via Docker

I built a custom orchestration layer in PHP (Laravel + Docker) that:

  1. Receives task descriptions
  2. Breaks them into subtasks using GTD methodology
  3. Assigns subtasks to the right model or tool
  4. Returns structured outputs

It's deployed on Fly.io for tasks that need internet access, and runs locally for sensitive work.

The Second Brain on NAS

All notes, documents, project files, and outputs flow into a self-hosted NAS running on my home network. Obsidian reads from it directly. Nothing is stored in Dropbox or Google Drive.

What I Gave Up

Let's be honest — there are real trade-offs:

  • Speed. Cloud GPUs are faster than local hardware.
  • State-of-the-art models. GPT-4o and Claude Opus still outperform local models for complex reasoning.
  • Convenience. Self-hosting requires maintenance.

I accept these trade-offs because the benefits — privacy, control, cost predictability, and the learning that comes from building — outweigh them for my use case.

The Takeaway

Sovereignty is a spectrum, not a binary. You don't have to host everything locally to benefit from sovereign thinking.

Start by mapping your dependencies. Ask: if this disappeared tomorrow, what breaks? Then decide, for each dependency, whether the risk is acceptable or worth mitigating.

That's the system. Build it iteratively. Own what matters.


If you're building a sovereign AI stack and want to compare notes, reach out.