From testnet to production. 9 versions charting the path from proof-of-concept to a fully decentralized AI inference network.
Whitepaper, P2P, CLI, API, Dashboard, BitNet, Benchmarks, Desktop App
CI/CD fixes, cleanup, test stabilization, Falcon3 prep, documentation — 6 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 1 | Fix CI/CD workflows across all platforms | Infra | Resolve build failures on Windows, macOS, Linux. Cross-platform test matrix. |
| 2 | Code cleanup and linting | Code | Remove dead code, fix lint warnings, standardize imports. |
| 3 | Test stabilization (196 tests passing) | Code | Fix flaky tests, ensure deterministic results, improve coverage. |
| 4 | Prepare Falcon3 1.58-bit integration | R&D | Verify GGUF availability, test bitnet.cpp compatibility, plan integration. |
| 5 | Documentation update (post-brainstorming Feb 2026) | Doc | Update README, roadmap, GitHub Pages with research-validated content. |
| 6 | GitHub Release v0.5.5 | Infra | Tag release, update version badges, sync pyproject.toml. |
Kademlia DHT, NAT traversal, desktop bridge, Falcon3/Edge models, 50+ community nodes — 10 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 7 | Kademlia DHT peer discovery | Code | Replace bootstrap servers with decentralized peer discovery via Kademlia DHT. |
| 8 | NAT traversal (STUN/TURN) | Code | Enable nodes behind routers to participate. STUN for discovery, TURN for relay. |
| 9 | Decentralized bootstrap mechanism | Infra | No central server dependency. Seed nodes + DHT for full decentralization. |
| 10 | Desktop ↔ Python backend bridge (live inference) | Code | Real-time bidirectional communication between desktop app and Python backend. Live inference metrics. |
| 11 | Real-time P2P network visualization | Web | Visual node map in desktop app showing connected peers, latency, throughput. |
| 12 | Integrate Falcon3 1.58-bit models (1B, 3B, 7B, 10B) | Code | Add TII instruction-tuned models. GGUF format, bitnet.cpp compatible. |
| 13 | Integrate Falcon-Edge (1B, 3B) + benchmark vs Microsoft BitNet | Code | Natively-trained 1-bit models from TII. Benchmark: 53.17% vs 51.54% avg. |
| 14 | Update bitnet.cpp to January 2026 parallel kernels | Code | Integrate ELUT parallel kernels for +1.15–2.1x additional CPU speedup. |
| 15 | Settings persistence + consent management | Code | Save user preferences, consent settings, and node configuration across sessions. |
| 16 | Public testnet with 50+ community nodes target | Com | Launch campaign: Show HN, r/LocalLLaMA. Target 50+ active nodes. |
Consensus Inference, Conversation Memory, reputation system, SAPO prototype — 10 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 17 | Consensus Inference: confidence-based routing (SLM-MUX style) | Code | Route queries based on model confidence scores. Avoids groupthink via orchestrated protocols. |
| 18 | Consensus Inference: cross-verification protocol | Code | Multiple models verify each other’s outputs. Majority voting + synthesis. |
| 19 | Consensus Inference: user-facing protocol selection | Code | Speed/Quality/Verify mode selection in UI. Energy budget controls. |
| 20 | Consensus Inference: provenance logging | Code | Record which nodes, which protocol, energy used. Full audit trail on ledger. |
| 21 | Conversation Memory: 3-tier architecture (Hot/Warm/Cold) | Code | Hot = current session, Warm = recent context, Cold = long-term storage. |
| 22 | Conversation Memory: Fact Store with heuristic + LLM extraction | Code | Extract and store facts from conversations. Heuristic + LLM-based extraction pipeline. |
| 23 | Conversation Memory: semantic recall via embeddings | Code | Retrieve relevant memories using semantic similarity. Local embedding model. |
| 24 | Node reputation system (reliability scoring) | Code | Score nodes on uptime, quality, energy efficiency. Published on ledger. |
| 25 | Anti-Sybil protections (IP + hardware fingerprint) | Code | Prevent fake nodes. IP diversity + hardware attestation. |
| 26 | SAPO integration prototype (reasoning rollout sharing) | R&D | Decentralized experience sharing for reasoning improvement. Based on Gensyn research. |
Whitepaper v2, PT-BitNet R&D, KV-Cache proof of concept, benchmarks — 6 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 27 | Whitepaper v2 | Doc | Complete WP v2: Consensus Inference, KV-Cache, ARIA-LM, reputation system, competitor analysis. |
| 28 | PT-BitNet R&D: ternarize Qwen3-14B | R&D | Post-training ternarization of Qwen3-14B (28 GB → ~3 GB). Quality benchmarks. |
| 29 | KV-Cache NVMe: proof of concept (save/restore between sessions) | R&D | Prototype SSD-based KV-cache persistence. Save and restore context across sessions. |
| 30 | Benchmark Falcon-Edge vs Microsoft BitNet (publication content) | R&D | Rigorous comparison. Publishable results for community credibility. |
| 31 | Reproduce SLM-MATRIX results on ARIA infrastructure | R&D | Validate multi-agent debate accuracy claims on ARIA’s consensus system. |
| 32 | Update threat model for distributed features | Doc | Address new attack vectors from DHT, consensus, memory, and reputation systems. |
KV-Cache NVMe paging, Knowledge Network, desktop UI — 10 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 33 | KV-Cache NVMe: attention-aware eviction + prefetch (SpeCache-style) | Code | Sparse attention identifies critical keys. Evict cold keys to SSD, prefetch hot keys. |
| 34 | KV-Cache NVMe: 2-bit quantization (KIVI) | Code | Compress KV-cache to 2-bit with <2% accuracy loss. Enables massive context on limited RAM. |
| 35 | KV-Cache NVMe: async I/O overlapped with compute | Code | Pipeline SSD reads with GPU/CPU compute to hide I/O latency. |
| 36 | Target: 500K+ tokens on 8GB RAM laptop | Code | Integration of eviction + quantization + async I/O to reach 500K+ token context. |
| 37 | Knowledge Network: P2P embedding sharing (Kademlia-based) | Code | Distribute document embeddings across the P2P network via DHT. |
| 38 | Knowledge Network: privacy-preserving embeddings (Eguard projection) | Code | Projection-based defense reduces inversion attacks from >95% to <5%. |
| 39 | Knowledge Network: standardized embedding model (community-governed) | Code | Community-selected embedding model. Community governance for model updates. |
| 40 | Knowledge Network: semantic search across distributed documents | Code | Query the network for relevant documents. Federated semantic search. |
| 41 | Infinite Context Mode UI in desktop app | Code | Desktop UI for managing extended context. Visual indicators for hot/cold cache. |
| 42 | Conversation Memory Manager UI in desktop app | Code | Browse, search, and manage stored memories. Fact store visualization. |
Community-evolving language model, SAPO, LoRA merging, distillation — 9 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 43 | ARIA-LM base model selection and distillation pipeline | Code | Select base 1-bit model. Build distillation pipeline from frontier models. |
| 44 | SAPO: decentralized reasoning rollout collection | Code | Collect reasoning traces from network nodes. Quality filtering via reputation. |
| 45 | SAPO: quality validation via reputation + cross-verification | Code | Only high-reputation nodes contribute training data. Cross-verify quality. |
| 46 | LoRA merging: community-contributed adapters (TIES-Merging/MergeKit) | Code | Community submits LoRA adapters. Merge using TIES-Merging or MergeKit. |
| 47 | LoRA anti-poisoning: benchmark testing before merge | Code | Automated benchmark suite validates adapter quality before community merge. |
| 48 | Meta-LoRA: CPU-only adapter generation from datasets | Code | Generate LoRA adapters on CPU from user-provided datasets. No GPU required. |
| 49 | RAG communautaire: document contribution and retrieval | Code | Community document sharing via Knowledge Network. Contribute and retrieve. |
| 50 | Consensus Inference → distillation pipeline | Code | Best debate outputs become training data for ARIA-LM. Self-improving loop. |
| 51 | ARIA-LM v0.1 release | Com | First community model release. Benchmarks, documentation, download. |
Stable production network, reputation system, hardened infrastructure — 6 tasks
| # | Task | Type | Detail |
|---|---|---|---|
| 52 | Distributed KV-Cache across P2P network | Code | Share KV-cache across nodes. Collaborative context extension. |
| 53 | Context Relay: map-reduce for long documents | Code | Split long documents across nodes. Map-reduce style processing. |
| 54 | Smart Inference Routing (local vs P2P vs cloud fallback) | Code | Intelligent routing: try local first, then P2P network, then cloud API fallback. |
| 55 | Universal API Bridge (OpenAI-compatible from any app) | Code | Drop-in replacement for OpenAI API. Works with any OpenAI-compatible application. |
| 56 | Production security hardening | Infra | P2P network audit. Bug bounty program. Third-party security review. |
| 57 | Production launch | Com | Stable production network for public use. Launch campaign: HN, Reddit, Twitter, Product Hunt. |
Mobile, mesh networking, automation, gamification, MoE+1-bit research