v0.5.5 in progress

ARIA Protocol — Roadmap v3.0

From testnet to production. 9 versions charting the path from proof-of-concept to a fully decentralized AI inference network.

9 Versions
54 Tasks
v0.5.5 Current

Version Timeline

v0.1 → v0.5.2 Genesis → Desktop Complete

Whitepaper, P2P, CLI, API, Dashboard, BitNet, Benchmarks, Desktop App

  • v0.1.0 Genesis — Whitepaper published, reference implementation
  • v0.2.0 Full Stack — P2P WebSocket + TLS, CLI, API, Dashboard, BitNet engine
  • v0.2.5 Hardening — Threat model, protocol spec, TLS support
  • v0.3.0 Benchmarks — Performance validation (89.65 t/s on 0.7B)
  • v0.4.0 Native BitNet — Python ctypes bindings to bitnet.cpp
  • v0.5.0 Desktop App — Tauri 2.0 + Electron, 12 languages, system tray
  • v0.5.1 Build Fix — Desktop build corrections
  • v0.5.2 Subprocess Backend — llama-cli bridge, multi-backend inference, CI/CD cross-platform builds
v0.5.5 Housekeeping & Foundations In Progress

CI/CD fixes, cleanup, test stabilization, Falcon3 prep, documentation — 6 tasks

View 6 tasks
#TaskTypeDetail
1Fix CI/CD workflows across all platformsInfraResolve build failures on Windows, macOS, Linux. Cross-platform test matrix.
2Code cleanup and lintingCodeRemove dead code, fix lint warnings, standardize imports.
3Test stabilization (196 tests passing)CodeFix flaky tests, ensure deterministic results, improve coverage.
4Prepare Falcon3 1.58-bit integrationR&DVerify GGUF availability, test bitnet.cpp compatibility, plan integration.
5Documentation update (post-brainstorming Feb 2026)DocUpdate README, roadmap, GitHub Pages with research-validated content.
6GitHub Release v0.5.5InfraTag release, update version badges, sync pyproject.toml.
v0.6.0 Testnet Alpha Next

Kademlia DHT, NAT traversal, desktop bridge, Falcon3/Edge models, 50+ community nodes — 10 tasks

View 10 tasks
#TaskTypeDetail
7Kademlia DHT peer discoveryCodeReplace bootstrap servers with decentralized peer discovery via Kademlia DHT.
8NAT traversal (STUN/TURN)CodeEnable nodes behind routers to participate. STUN for discovery, TURN for relay.
9Decentralized bootstrap mechanismInfraNo central server dependency. Seed nodes + DHT for full decentralization.
10Desktop ↔ Python backend bridge (live inference)CodeReal-time bidirectional communication between desktop app and Python backend. Live inference metrics.
11Real-time P2P network visualizationWebVisual node map in desktop app showing connected peers, latency, throughput.
12Integrate Falcon3 1.58-bit models (1B, 3B, 7B, 10B)CodeAdd TII instruction-tuned models. GGUF format, bitnet.cpp compatible.
13Integrate Falcon-Edge (1B, 3B) + benchmark vs Microsoft BitNetCodeNatively-trained 1-bit models from TII. Benchmark: 53.17% vs 51.54% avg.
14Update bitnet.cpp to January 2026 parallel kernelsCodeIntegrate ELUT parallel kernels for +1.15–2.1x additional CPU speedup.
15Settings persistence + consent managementCodeSave user preferences, consent settings, and node configuration across sessions.
16Public testnet with 50+ community nodes targetComLaunch campaign: Show HN, r/LocalLLaMA. Target 50+ active nodes.
v0.7.0 Smart Layer Planned

Consensus Inference, Conversation Memory, reputation system, SAPO prototype — 10 tasks

View 10 tasks
#TaskTypeDetail
17Consensus Inference: confidence-based routing (SLM-MUX style)CodeRoute queries based on model confidence scores. Avoids groupthink via orchestrated protocols.
18Consensus Inference: cross-verification protocolCodeMultiple models verify each other’s outputs. Majority voting + synthesis.
19Consensus Inference: user-facing protocol selectionCodeSpeed/Quality/Verify mode selection in UI. Energy budget controls.
20Consensus Inference: provenance loggingCodeRecord which nodes, which protocol, energy used. Full audit trail on ledger.
21Conversation Memory: 3-tier architecture (Hot/Warm/Cold)CodeHot = current session, Warm = recent context, Cold = long-term storage.
22Conversation Memory: Fact Store with heuristic + LLM extractionCodeExtract and store facts from conversations. Heuristic + LLM-based extraction pipeline.
23Conversation Memory: semantic recall via embeddingsCodeRetrieve relevant memories using semantic similarity. Local embedding model.
24Node reputation system (reliability scoring)CodeScore nodes on uptime, quality, energy efficiency. Published on ledger.
25Anti-Sybil protections (IP + hardware fingerprint)CodePrevent fake nodes. IP diversity + hardware attestation.
26SAPO integration prototype (reasoning rollout sharing)R&DDecentralized experience sharing for reasoning improvement. Based on Gensyn research.
v0.7.5 R&D + Documentation Planned

Whitepaper v2, PT-BitNet R&D, KV-Cache proof of concept, benchmarks — 6 tasks

View 6 tasks
#TaskTypeDetail
27Whitepaper v2DocComplete WP v2: Consensus Inference, KV-Cache, ARIA-LM, reputation system, competitor analysis.
28PT-BitNet R&D: ternarize Qwen3-14BR&DPost-training ternarization of Qwen3-14B (28 GB → ~3 GB). Quality benchmarks.
29KV-Cache NVMe: proof of concept (save/restore between sessions)R&DPrototype SSD-based KV-cache persistence. Save and restore context across sessions.
30Benchmark Falcon-Edge vs Microsoft BitNet (publication content)R&DRigorous comparison. Publishable results for community credibility.
31Reproduce SLM-MATRIX results on ARIA infrastructureR&DValidate multi-agent debate accuracy claims on ARIA’s consensus system.
32Update threat model for distributed featuresDocAddress new attack vectors from DHT, consensus, memory, and reputation systems.
v0.8.0 Extended Context & Knowledge Planned

KV-Cache NVMe paging, Knowledge Network, desktop UI — 10 tasks

View 10 tasks
#TaskTypeDetail
33KV-Cache NVMe: attention-aware eviction + prefetch (SpeCache-style)CodeSparse attention identifies critical keys. Evict cold keys to SSD, prefetch hot keys.
34KV-Cache NVMe: 2-bit quantization (KIVI)CodeCompress KV-cache to 2-bit with <2% accuracy loss. Enables massive context on limited RAM.
35KV-Cache NVMe: async I/O overlapped with computeCodePipeline SSD reads with GPU/CPU compute to hide I/O latency.
36Target: 500K+ tokens on 8GB RAM laptopCodeIntegration of eviction + quantization + async I/O to reach 500K+ token context.
37Knowledge Network: P2P embedding sharing (Kademlia-based)CodeDistribute document embeddings across the P2P network via DHT.
38Knowledge Network: privacy-preserving embeddings (Eguard projection)CodeProjection-based defense reduces inversion attacks from >95% to <5%.
39Knowledge Network: standardized embedding model (community-governed)CodeCommunity-selected embedding model. Community governance for model updates.
40Knowledge Network: semantic search across distributed documentsCodeQuery the network for relevant documents. Federated semantic search.
41Infinite Context Mode UI in desktop appCodeDesktop UI for managing extended context. Visual indicators for hot/cold cache.
42Conversation Memory Manager UI in desktop appCodeBrowse, search, and manage stored memories. Fact store visualization.
v0.9.0 ARIA-LM: Community Model Planned

Community-evolving language model, SAPO, LoRA merging, distillation — 9 tasks

View 9 tasks
#TaskTypeDetail
43ARIA-LM base model selection and distillation pipelineCodeSelect base 1-bit model. Build distillation pipeline from frontier models.
44SAPO: decentralized reasoning rollout collectionCodeCollect reasoning traces from network nodes. Quality filtering via reputation.
45SAPO: quality validation via reputation + cross-verificationCodeOnly high-reputation nodes contribute training data. Cross-verify quality.
46LoRA merging: community-contributed adapters (TIES-Merging/MergeKit)CodeCommunity submits LoRA adapters. Merge using TIES-Merging or MergeKit.
47LoRA anti-poisoning: benchmark testing before mergeCodeAutomated benchmark suite validates adapter quality before community merge.
48Meta-LoRA: CPU-only adapter generation from datasetsCodeGenerate LoRA adapters on CPU from user-provided datasets. No GPU required.
49RAG communautaire: document contribution and retrievalCodeCommunity document sharing via Knowledge Network. Contribute and retrieve.
50Consensus Inference → distillation pipelineCodeBest debate outputs become training data for ARIA-LM. Self-improving loop.
51ARIA-LM v0.1 releaseComFirst community model release. Benchmarks, documentation, download.
v1.0.0 Production Network Planned

Stable production network, reputation system, hardened infrastructure — 6 tasks

View 6 tasks
#TaskTypeDetail
52Distributed KV-Cache across P2P networkCodeShare KV-cache across nodes. Collaborative context extension.
53Context Relay: map-reduce for long documentsCodeSplit long documents across nodes. Map-reduce style processing.
54Smart Inference Routing (local vs P2P vs cloud fallback)CodeIntelligent routing: try local first, then P2P network, then cloud API fallback.
55Universal API Bridge (OpenAI-compatible from any app)CodeDrop-in replacement for OpenAI API. Works with any OpenAI-compatible application.
56Production security hardeningInfraP2P network audit. Bug bounty program. Third-party security review.
57Production launchComStable production network for public use. Launch campaign: HN, Reddit, Twitter, Product Hunt.
v1.1.0+ Beyond — Long-term Vision Vision

Mobile, mesh networking, automation, gamification, MoE+1-bit research

  • Mobile companion app — iOS/Android with on-device inference (1B–3B models)
  • ARIA Mesh Mode — Offline local inference via Bluetooth/WiFi Direct
  • Live Network Globe — 3D visualization of the global ARIA network
  • Automation Studio — No-code AI workflows and task automation
  • Team/Family sharing — LAN-based node sharing for groups
  • Gamification & achievements — Reward engagement and contribution milestones
  • MoE + 1-bit research — 100B+ params, ~1 GB memory, frontier-class on laptops

At a Glance

9
Planned Versions
54
Identified Tasks
v0.5.5
Current Version