View all articles
transparencyopen-sourcestacklocal-ai

The VORLUX AI Stack: Every Tool We Use, Nothing Hidden

JG
Jacobo Gonzalez Jaspe
|

The VORLUX AI Stack: Every Tool We Use, Nothing Hidden

When we tell clients their AI will run locally with no cloud dependency, the natural follow-up is: “Okay, but what exactly are you running?” Fair question. If we’re asking you to trust us with your infrastructure, you deserve to see everything under the hood.

This post is our full technology disclosure. Every component, every tool, every decision — and why we made it. No proprietary black boxes. No vague references to “our AI platform.” Just the actual stack.

The Core Components

Here’s everything that powers VORLUX AI, from inference to interface:

LayerTechnologyRoleWhy This One
InferenceOllamaLLM servingBest local inference server, 14 models loaded
APIFastAPI + PythonREST API & orchestrationFast, typed, async-native
DashboardNext.jsInternal operations dashboardReact ecosystem, SSR, real-time
DatabaseSQLiteAll persistenceZero config, zero network, battle-tested
Public SiteAstrovorluxai.comStatic-first, fast, SEO-optimized
Automationn8nWorkflow automationVisual workflows, self-hosted
SearchFAISS + BM25RAG retrievalVector + keyword hybrid search
SchedulingBackgroundSchedulerCron jobs58 scheduled tasks, Python-native
CacheRedisSession & task cacheIn-memory speed, Docker-hosted
HardwareMac M3 Pro 32GBPrimary serverApple Silicon = best performance/watt

Every single component either runs on our hardware or on the client’s hardware. Nothing phones home. Nothing sends telemetry. Nothing requires an internet connection to function.

How It All Fits Together

flowchart TB
    subgraph CLIENT["Client Layer"]
        SITE["Astro Site<br/>vorluxai.com"]
        DASH["Next.js Dashboard<br/>:3000"]
    end
    
    subgraph API_LAYER["API & Orchestration"]
        API["FastAPI API<br/>:8090"]
        ORCH["Orchestrator<br/>:8091"]
        N8N["n8n Workflows<br/>:5678"]
    end
    
    subgraph INFERENCE["Inference Layer"]
        OLLAMA["Ollama<br/>14 Models<br/>:11434"]
        RAG["FAISS + BM25<br/>RAG Search"]
    end
    
    subgraph DATA["Data Layer"]
        SQLITE[("SQLite<br/>All Persistence")]
        REDIS[("Redis<br/>Cache<br/>:6379")]
    end
    
    subgraph AUTOMATION["Automation Layer"]
        SCHED["BackgroundScheduler<br/>58 Cron Jobs"]
        LOOPS["36 Autonomous<br/>Loops"]
    end
    
    SITE --> API
    DASH --> API
    API --> OLLAMA
    API --> RAG
    API --> SQLITE
    API --> REDIS
    ORCH --> API
    ORCH --> N8N
    SCHED --> API
    LOOPS --> ORCH
    RAG --> SQLITE
    
    style CLIENT fill:#0B1628,color:#FAFAFA
    style INFERENCE fill:#059669,color:#fff
    style DATA fill:#F5A623,color:#0B1628

The 14 Models We Run

Not every task needs the same model. We run 14 models simultaneously, routing each request to the right one:

  • Gemma 2 9B — General-purpose reasoning and conversation
  • Llama 3.3 70B — Complex analysis and long-form generation
  • Mistral Small 24B — Fast, capable mid-range inference
  • Phi-4 — Lightweight tasks, fast turnaround
  • Qwen 2.5 72B — Multilingual tasks, excellent for Spanish
  • Qwen 2.5 Coder 7B — Code generation and review
  • DeepSeek V3 — Technical reasoning
  • Plus 7 specialized variants for embeddings, summarization, and classification

All running on a single Mac M3 Pro with 32GB of unified memory. No GPU cluster. No data center. One machine on a desk in Valencia.

36 Autonomous Loops, 58 Cron Jobs

The system doesn’t just respond to requests — it works autonomously. Here’s what runs around the clock:

  • Content loops: Research, draft, review, publish — fully automated content pipeline
  • Quality loops: Code review, test execution, knowledge base updates
  • Monitoring loops: Health checks every 60 seconds, auto-restart on failure
  • Business loops: Lead research, market analysis, competitive intelligence

The BackgroundScheduler manages 58 cron jobs that trigger these loops on precise schedules. The watchdog system ensures everything stays alive. If a service crashes at 3 AM, it restarts itself before anyone notices.

We detailed how this self-healing architecture works in our operations documentation.

Why Open-Source Matters

Every component in our stack is either open-source or built by us in-house. This isn’t ideological — it’s practical:

  1. No license fees — Our clients don’t pay software licenses. Hardware is the only cost.
  2. No vendor lock-in — If Ollama disappears tomorrow, we switch to llama.cpp or vLLM. Same models, different runtime.
  3. Full auditability — Regulated clients can inspect every line of code that touches their data. This directly satisfies GDPR Article 25 requirements for privacy by design.
  4. Community support — 50,000+ GitHub stars across our core dependencies. These aren’t experimental toys.

Compared to Cloud-Dependent Stacks

AspectVORLUX AI (Local)Typical Cloud Stack
Data locationYour hardwareAWS/Azure/GCP
Monthly costEUR 0 after hardwareEUR 500-5,000+/mo
Latency< 100ms first token200-800ms+
Internet requiredNoYes
GDPR complexityMinimalSignificant
Vendor lock-inNoneHigh
Model switchingMinutesDays-weeks
Uptime dependencyYour powerTheir SLA
Audit trailFull local logsProvider-dependent

The cloud stack isn’t wrong for everyone. But for businesses processing sensitive data under European regulation, local deployment eliminates entire categories of risk. We explored this tradeoff in depth in our cost analysis.

What This Means for You

When we deploy AI for your business, you get this exact stack — adapted to your hardware and your workloads. Not a watered-down version. Not a hosted service with a “local” label. The real thing, running on metal you own.

The Edge AI for SMEs service we’re launching in May uses this same architecture, scaled down to hardware that fits on a shelf and a budget that fits a small business.

See It in Action

We run live demos of this stack during our free assessment calls. No slides, no mockups — the actual system, running actual models, processing actual queries in real time.

Book your free 15-minute assessment and see for yourself what local AI looks like when it’s built properly.

Tomorrow, we reveal exactly what services we’re launching and what they cost. No surprises — just like the stack.


This is post 2 of our Launch Week series. Yesterday: Local AI Readiness Checklist. Tomorrow: Our Services and Pricing.

External references: Ollama | n8n Workflow Automation | Astro Web Framework | GDPR Article 25 & Local AI


Ready to Get Started?

VORLUX AI helps Spanish and European businesses deploy AI solutions that stay on your hardware, under your control. Whether you need edge AI deployment, LMS integration, or EU AI Act compliance consulting — we can help.

Book a free discovery call to discuss your AI strategy, or explore our services to see how we work.

Share: LinkedIn X
Newsletter

Access exclusive resources

Subscribe to unlock 230+ workflows, 43 agents, and 26 professional templates. Weekly insights, no spam.

Bonus: Free EU AI Act checklist when you subscribe
Once a week No spam Unsubscribe anytime
EU AI Act: 99 days to deadline

15 minutes to evaluate your case

No-commitment initial consultation. We analyze your infrastructure and recommend the optimal hybrid architecture.

No commitment 15 minutes Custom proposal

136 pages of free resources · 26 compliance templates · 22 certified devices