Technical Deep Dive

Routing vs. Computing: The Architecture That Isn't There

Central systems compute on your data. Federated learning computes on gradients. QIS doesn't compute at all—it routes. The packets coming back ARE the insights. Here's why that changes everything.

By Christopher Thomas Trevethan • January 16, 2026

When engineers hear "distributed AI" or "collective intelligence," they immediately think about computation. Where does the compute happen? How do you aggregate? What's the coordination overhead?

These are the wrong questions for QIS. They assume a paradigm that QIS doesn't use.

To understand QIS, you need to ask a simpler question:

Where does the computation happen?

The answer, for QIS, is: it doesn't.

Not centrally. Not in an aggregator. Not distributed across nodes. The routing infrastructure—whether DHT, vector database, or something else—isn't computing anything. It's just a phone book. The packets coming back aren't data to be processed. They ARE the insights. Local synthesis is trivial: count votes, average values, confidence-weight outcomes. That's it.

This isn't a minor architectural detail. It's the fundamental difference that makes everything else possible: real-time speed, cell phone deployment, quadratic scaling with logarithmic cost.

Let me show you exactly what I mean by walking through all three paradigms.

Paradigm 1: Central Data Pooling

This is what Climate FieldView, OpenAI, Google Health, and most enterprise AI systems do.

Central Data Pooling Architecture

Your Data → Their Servers → THEIR Computation → Results Back

You send your raw data (farm telemetry, medical records, sensor logs) to their cloud. Their servers—massive GPU clusters, data centers, infrastructure you don't see—run computation on that data. They train models, run analytics, generate predictions. You get results back.

There's a PLACE where the intelligence happens. Their data center. Their compute. Their models.

The computation is real and expensive. Training GPT-4 cost over $100 million. Running inference at scale costs millions per month. The more data, the more compute. The more users, the more infrastructure. Linear scaling at best.

And here's the critical limitation: you can't see the evidence. When FieldView tells you to plant at a certain rate, you can't ask "show me the 500 farms this is based on." The insights are locked in their models. You get predictions from a black box. Trust the algorithm or don't.

Paradigm 2: Federated Learning

Federated learning was designed to solve the privacy problem of central pooling. Your data stays local. Only model updates move.

Federated Learning Architecture

Your Data (local) → Train Local Model → Gradients/Weights → Central Aggregator → Combined Model Back

Your device trains a local model on your data. Instead of sending raw data, you send model gradients or weight updates to a central aggregator. The aggregator combines updates from all participants into a global model. That model gets pushed back to everyone.

There's STILL a place where computation happens: the aggregator. Aggregation IS computation.

Federated learning is a genuine improvement. Google's keyboard (Gboard) uses it—your typing patterns improve the model without Google seeing what you typed. Real privacy benefits.

But the aggregator is still central. Someone runs it. Someone coordinates the training rounds. And aggregation isn't free—algorithms like FedAvg, FedProx, and SCAFFOLD all require the aggregator to receive, process, and combine model updates from every participant, every round. That's computation. It scales with participants.

The communication overhead is real: O(model_size × clients × rounds). For large models and many participants, this gets expensive. And you still get a trained model as output—not evidence of what worked.

Paradigm 3: QIS Protocol

Now look at what QIS does:

QIS Architecture

Your Data (local) → Query Routes by Similarity → Outcome Packets Return → Local Synthesis

Your device generates a semantic fingerprint—a compact routing key derived from your situation. That fingerprint routes through the network (DHT, vector index, whatever) to find peers with matching situations. Those peers return outcome packets: "I tried this, here's what happened." Your device synthesizes locally.

There's NO central computation. The routing infrastructure doesn't compute. The packets ARE the insights—routed to you automatically because the routing itself is by exact similarity to your problem.

Read that again: the routing infrastructure doesn't compute anything.

A DHT is a phone book. You give it a key, it tells you where to find the matching entries. The DHT doesn't analyze, aggregate, or transform. It routes.

A vector database does approximate nearest neighbor lookup. You give it a query vector, it returns the closest matches with their metadata. The vector DB doesn't train models or generate insights. It finds similar things.

In QIS, the metadata attached to each entry IS the outcome itself: "Stage III colorectal, KRAS mutation, tried FOLFOX, progression-free 18 months." The insight isn't computed from the match—the insight IS the match. It was deposited there by a previous participant who faced the same situation and wanted to share what they learned. Because your semantic fingerprint routes by exact similarity, you land in the neighborhood of people who had your exact problem—and their deposited outcomes are waiting for you. (This is the "Three Elections" framework: experts define similarity, routing finds your cohort, outcomes speak.)

The Phone Book Analogy

Imagine you need a plumber. You look in the phone book under "plumbing." The phone book doesn't evaluate plumbers, train a model on plumber reviews, or compute which plumber is best. It just shows you the entries filed under "plumbing."

But here's the twist: in QIS, each entry already contains the outcome. "Called this plumber, fixed the leak in 2 hours, $150." The phone book didn't compute that—the previous customer deposited it when they filed their experience.

And here's the key: the category you look up IS your exact problem. You're not looking up "plumbing"—you're looking up "leaking pipe, copper, under kitchen sink, 1960s house." Everyone filed under that exact category had that exact problem. So the outcomes they deposited are directly relevant to you. The routing by similarity IS the matching. No analysis needed—just find your category and read what others learned.

You read 20 entries. You see the pattern. You make your decision. Local synthesis: trivial. Routing: O(log N). No central compute anywhere.

What Actually Moves

This is where the paradigms diverge most sharply:

Dimension Central Pooling Federated Learning QIS Protocol
What moves Raw data (your records, logs, telemetry) Model gradients/weights Semantic fingerprints + tiny outcome packets
Where compute happens Their data center Aggregator server Nowhere centrally
What the infra does Stores, processes, trains models Aggregates model updates Routes you to exactly the real-time insight needed
Output Predictions from black-box model Trained global model Actual outcomes from matched peers
Local synthesis cost N/A (all remote) Local training each round Trivial: vote/average tiny JSONs
Communication cost O(data size) O(model × clients × rounds) O(log N) per query
Hardware required Massive cloud infrastructure Aggregator + compute-capable clients Cell phones work fine

The Mental Model Everyone's Stuck On

Here's why this is hard to see.

For decades, computing has meant: data in → process → results out. When people hear "distributed AI" or "collective intelligence," they assume that process must be happening somewhere—just distributed across nodes, or aggregated in a server, or computed in the cloud.

What "Distributed AI" Usually Means

Most systems called "distributed AI" are really just distributed compute—splitting the same expensive processing across multiple machines. MapReduce, Spark, distributed training: find a partner node, share raw data or gradients, compute together. It's still two machines crunching numbers. Still linear scaling. Still latency. Still not solving YOUR specific problem in real-time—just parallelizing generic model training across a cluster.

QIS isn't distributed compute. It's distributed outcomes. No machine computes with another. Each node deposits what it learned. Routing finds the deposits relevant to you. The insight was already created—by someone who lived it.

What About Coordination?

I don't emphasize coordination much—it's only about a millionth as valuable as real-time scaling of intelligence. But yes, QIS enables precision coordination too, and it's trivially simple: networks that want to coordinate just add an API endpoint to the outcome packet metadata. That's it. No all-to-all messaging. No broadcast storms. You coordinate based on whatever parameters matter to you, informed by real-time insight on how to do so. Precision coordination—not coordination theater.

The mental model is:

The Assumed Model

Ingest raw data → Analyze/compute → Generate insights

This is what every database, every ML pipeline, every analytics platform does. So people assume QIS must do it too—they just can't see where.

QIS breaks this model completely. There's no "analyze" step because:

The QIS Model

Insights already exist (deposited by previous participants) → Route by exact similarity → Insights arrive (for your exact problem)

The insight isn't generated from data. The insight was RECORDED by someone who faced your situation and learned something. Routing by similarity finds them. That's it.

This is the gap people can't cross: they think "finding relevant information" requires computation. It doesn't. It requires routing by similarity. The similarity IS the relevance.

Think about it causally:

  1. Your semantic fingerprint is derived from your exact situation (disease profile, crop conditions, equipment failure mode)
  2. Routing by similarity lands you in a cohort of people who had that exact situation
  3. Those people already deposited what they learned when their outcomes resolved
  4. So the packets that return are insights for your exact problem—not because anyone computed "what's relevant to you," but because the routing topology itself creates relevance

The routing did the matching. No algorithm had to analyze what's relevant. No model had to learn your preferences. The similarity metric (defined by domain experts or embeddings) determines where you land. Everyone else who landed there had the same situation. Their deposited outcomes are your answers.

This is why QIS can be real-time, cheap, and run on cell phones. There's no "analyze" step to wait for. There's no computation to scale. Just routing—and the answers are already there, filed by people who came before you.

Why "No Central Compute" Changes Everything

When there's no central computation step, constraints that seem fundamental simply disappear:

Real-Time by Default

Central systems batch, train, deploy. Federated learning runs training rounds. QIS queries resolve in one round-trip. An outcome deposited now is queryable immediately. No waiting for the next model update. No training pipeline. Just routing.

Cell Phones Are Enough

When the "compute" is just routing queries and synthesizing tiny JSON packets, you don't need GPUs. You don't need cloud infrastructure. A smartphone has more than enough power. This isn't a degraded experience for resource-constrained devices—it's the native design.

Quadratic Intelligence, Logarithmic Cost

N agents create N(N-1)/2 synthesis opportunities—that's the quadratic intelligence. But each agent only pays O(log N) to route and O(K) to synthesize K matches—and "synthesize" just means reading tiny packets, not computing with those agents. No coordination. No shared processing. Just route to your cohort, grab the outcome packets they already deposited, tally locally. No central bottleneck scales with N². The opportunity is quadratic. The cost is not.

No Aggregation Bottleneck

Federated learning's aggregator has to process updates from every client every round. That's the scaling limit. QIS has no aggregator. Every query is independent. A million queries don't create a million-fold load on any single point—they distribute across the routing topology.

These inversions run deep. For the full list, see The 11 Flips—everything you think about AI is backwards.

What Each Agent Looks Like

In QIS, each agent is sovereign. It has its own data sources, its own local processing, its own decision-making. The network just helps it find relevant peers.

A QIS Agent (phone, server, device—anything)

📱 Local sensors
🔌 IoT devices
🗄️ Local database
🌐 APIs
📊 User input
🤖 Local AI/ML

The agent can pull from any data source it has access to: sensors, databases, APIs, user input, local AI models. None of this data leaves the device. The agent uses it to generate outcome packets when decisions resolve—the address for their problem is pre-determined by the experts building that QIS network.

When the agent queries, it gets back outcome packets from matched peers. It synthesizes locally—voting, averaging, confidence-weighting—and makes its decision. The whole process uses the same resources as loading a webpage.

This is why I say QIS is infrastructure-agnostic. It doesn't matter what data sources you have, what local AI you run, what platform you're on. If you can:

  1. Generate a semantic fingerprint from your situation
  2. Route that fingerprint to find similar peers
  3. Receive and synthesize outcome packets

...you can participate in QIS. The routing mechanism is pluggable—DHT, vector database, federated index, whatever works for your use case. The intelligence emerges from the topology, not from any specific technology.

The Outcome Packets

The packets that travel in QIS are tiny. They contain just enough information to be useful:

// Example outcome packet (healthcare) { "treatment": "FOLFOX + Bevacizumab", "outcome": "progression_free", "duration_months": 18, "confidence": 0.94, "timestamp": "2025-12-15" } // Example outcome packet (agriculture) { "intervention": "urea 50kg/ha + fungicide", "outcome": "yield_improvement", "delta_percent": 23, "season": "2024" }

No raw data. No identifying information. No CT scans, no genomes, no sensor logs. Just the structured outcome: what was tried, what happened, when.

Synthesizing these packets locally? Trivial—whether it's 50, 500, 5,000, or more. Count how many succeeded. Average the improvement percentages. Weight by recency or confidence. A cell phone can do this in milliseconds. The packets are tiny. The math is basic. The only limit is how many peers match your situation.

The packets coming back ARE the insights. They're not data to be processed—they're answers, pre-structured, ready to synthesize. That's why there's no central compute. The insight was created when the outcome was recorded. Routing just delivers it to whoever needs it.

Multiple Routing Options, Same Pattern

I keep saying "DHT or vector database or whatever" because the specific routing technology doesn't matter. What matters is the pattern:

All of these are just routing. None of them compute insights. The insight was deposited by a previous agent. The routing system just connects you to it.

Any System That Can...

Ingest data (APIs, IoT, databases—whatever the source), define similarity (expert templates, embeddings, hashes), route by similarity, propagate outcomes directly from a query, and synthesize packets locally...

...can enable QIS. Quadratic intelligence. Logarithmic cost. The architecture is flexible—the primitives are what matter.

For a deep dive on routing options, see Routing by Similarity—8 battle-tested methods that already work at planetary scale.

Why This Matters for Deployment

When people hear "global collective intelligence," they imagine massive infrastructure requirements. Data centers. Coordination servers. Expensive cloud contracts.

QIS inverts that assumption.

Because there's no central compute, deployment scales with participants, not infrastructure. Each new phone that joins is both a beneficiary and a contributor. It queries the network and deposits outcomes. The network gets smarter without any central resource scaling proportionally.

This is why QIS can realistically reach:

Not because we've optimized a centralized system to work in degraded conditions—but because the architecture never required central resources in the first place.

The Architecture That Isn't There

When engineers ask "where does the compute happen?" they're asking the wrong question for QIS. The question assumes something needs to be computed centrally. It doesn't.

The routing infrastructure is just a phone book—an addressing system that connects similar situations. The outcome packets are just structured answers deposited by previous participants. Local synthesis is just counting and averaging.

There's no black-box model. There's no aggregation step. There's no coordination server. There's no data center running inference.

The intelligence emerges from the topology itself—from the fact that every participant who faces a situation can find every other participant who faced something similar, and learn what they learned.

Central systems: "We compute on your data."
Federated learning: "We compute on your gradients."
QIS: "We don't compute. We route. The answers are already there."

That's the difference. That's why it works on cell phones. That's why it scales quadratically. That's why real-time is the default.

The architecture that seems like it should be there—the central server, the aggregator, the compute cluster—isn't there. And that's exactly why it works.

Subscribe on Substack One Round-Trip Every Component Exists QIS vs FL All Articles