Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Mainstream LLMs are centralized. What if they w...

Sponsored · Ship Features Fearlessly Turn features on and off without deploys. Used by thousands of Ruby developers.

Mainstream LLMs are centralized. What if they weren't?

Avatar for Favour Chukwuedo

Favour Chukwuedo

April 11, 2026

More Decks by Favour Chukwuedo

Other Decks in Technology

Transcript

  1. Mainstream LLMs are centralized. What if they weren't? Presented By:

    Favour Chukwuedo Codex Community Meetup: Vancouver April 10, 2026
  2. Running a 70B model requires $10K+ GPU hardware Every LLM

    call goes through OpenAI/Google/Anthropic etc servers They see your prompts, they control access, they set the price What if 10 people could each host a piece of one model and run it together? Presented By: Favour Chukwuedo Codex Community Meetup: Vancouver April 10, 2026
  3. Each person hosts a fraction of a model (layers 0–10

    of 28) Nodes find each other automatically (mDNS on LAN, DHT on internet) Inference running as a pipeline: your layers → next node's layers → response Privacy-first: onion-routed traffic, .llm cryptographic addresses llm-node: BitTorrent for LLM inference Presented By: Favour Chukwuedo Codex Community Meetup: Vancouver
  4. Run one command: llm-node setup --web Web wizard auto-detects resources

    you have and recommend (RAM, CPU, disk) Auto-discovers nearby nodes, connects and fills gaps automatically Shows which layers are missing and recommends what you should host Downloads only the layers you need (HTTP range requests for partial safetensors) Setup in under 60 seconds Presented By: Favour Chukwuedo Codex Community Meetup: Vancouver
  5. Dashboard shows: connected peers, layer coverage map, pipeline status Type

    a message → model responds via local inference or distributed pipeline Second node joins → dashboard updates in real-time, coverage map fills in Every node gets a .llm address (like a .onion address) Presented By: Favour Chukwuedo Codex Community Meetup: Vancouver
  6. Rust core (candle for inference, libp2p for networking) Supports Qwen3,

    Llama, Phi-3 — auto- detects architecture and dtype Onion routing with X25519 + AES-256-GCM (Tor-inspired) Next: Bootstrap nodes for internet-scale, NAT traversal, relay incentives Under the hood Presented By: Favour Chukwuedo Codex Community Meetup: Vancouver