The World's First Blockchain That Thinks
Whitepaper v2.0 — March 2026
The current AI landscape is dominated by centralized services where a handful of corporations control the world's most powerful language models. Users have no ownership over the intelligence they help create, no way to verify how their data is used, and no provenance for AI-generated outputs.
MACLLM Chain addresses these fundamental issues by creating a decentralized AI network where:
When users interact with centralized AI services, their queries, conversations, and data reside on corporate servers. Users cannot verify whether their data is used for training, shared with third parties, or truly deleted upon request.
AI outputs lack traceability. There is no way to verify what information sources were used, what confidence level the system had, or which computational resources contributed to the answer.
Users who teach AI through feedback and corrections generate value that accrues entirely to the service provider. Knowledge contributors receive no ongoing compensation for the intelligence they help create.
MACLLM Chain consists of two complementary layers:
User Node Cognitive Node Knowledge Node
| | |
v v v
+----------------------------------------------------------+
| ORIN Chain (Body) |
| Ledger | Consensus | mCoin | Provenance | Smart Rules |
+----------------------------------------------------------+
| | |
v v v
+----------------------------------------------------------+
| MACLLM Brain (Mind) |
| 6-Stage Pipeline | 64-Axis Composite Embedding (65,536d) | Matrix Sharding |
| Triple Semantic Encoder (Koopman + PPMI + Neural) | GenerativeEngine |
+----------------------------------------------------------+
| Node Type | Role | Requirements | Reward |
|---|---|---|---|
| User Node | Submit queries, receive answers | Wallet + mCoin | — |
| Cognitive Node | Provide compute (blind matrix operations) | 20-40W device + Cognitive Test pass | 30% of query fee |
| Knowledge Node | Expert knowledge contribution | Domain expertise verification | 20% royalty per use |
Unlike transformer-based LLMs that require billions of parameters and massive GPU clusters, MACLLM uses a probability matrix architecture that runs on commodity hardware (Apple Mac Mini M1, 16GB RAM).
Every query is represented as a 64-axis composite embedding (64 × 1,024d = 65,536 dimensions) powered by a Triple Semantic Encoder (Koopman Eigenfunctions + PPMI+SVD + Mini Neural). The axes are organized across six core groups:
| Group | Axes | Purpose |
|---|---|---|
| Semantic | topic_clarity, intent_confidence, complexity, specificity | Input understanding |
| Quality | confidence, evidence_strength, consistency, freshness_need | Post-retrieval quality |
| Routing | source_type, knowledge_distance, creativity_need, depth_required | Decision-making |
| Safety | sensitivity, risk_score, trust_level, anomaly_score | Guardrails |
| Cognitive | reasoning_depth, abstraction_level, multi_step, creativity_need | Reasoning control |
| Meta | self_consistency, prediction_accuracy, learning_value, novelty | Self-awareness |
The full 64-axis space fits in a single CPU cache line as a 64×64 matrix = 32KB. The Triple Semantic Encoder produces a 1,024-dimensional vector per axis: Koopman Eigenfunctions (342d) capture temporal dynamics, PPMI+SVD (342d) encodes co-occurrence statistics from 100K documents, and a Mini Neural encoder (340d) provides discriminative triplet-trained representations.
Inspired by how the human brain compresses experience into principles rather than memorizing raw facts:
| Level | Content | Count | Size |
|---|---|---|---|
| L1 | Raw entries (procedures) | 243,000 | ~2GB |
| L2 | Concepts (ConceptGraph) | 50,000 | ~50MB |
| L3 | Principles | 10,000 | 7.7MB |
| L4 | Meta-rules | 1,000 | 330KB |
| L5 | Axioms | 100 | 41KB |
The brain comprises 84+ modules including: ThinkingPipeline (6-stage reasoning), CompositionEngine (16-module text generation), GenerativeEngine (creates text without LLM using ConceptGraph + PatternBank + FlowModel), Triple Semantic Encoder (Koopman + PPMI + Neural), Koopman Dynamics (predictive modeling), GeniusNetwork (multi-engine assembly), and SelfThinking (autonomous learning).
ORIN uses BFT (Byzantine Fault Tolerant) consensus with BLS signature aggregation and VRF-based leader rotation. Finality is immediate within a shard (no forks). A minimum of 2/3+1 validators must sign each block.
The ledger enforces a Fixed-Supply Invariant: the total supply of mCoin (1 billion) is verified at every checkpoint. No minting is allowed after genesis. State-first validation means transactions are checked against current state without replaying history.
Every query/answer pair is recorded as a provenance entry containing: query hash, answer hash, contributing cognitive nodes, knowledge sources, confidence score, latency, and axis snapshot. Raw data is never stored on-chain — only cryptographic hashes.
Traditional consensus mechanisms waste energy (PoW) or favor the wealthy (PoS). Proof of Cognition validates that nodes can actually perform cognitive work.
Nodes must maintain a cognitive score of ≥60% to remain active. Reputation is updated using exponential moving average: rep = 0.9 × old + 0.1 × score. Failing nodes are automatically deactivated.
The core intellectual property of MACLLM Chain is protected through four layers:
| Layer | Protection | What Attacker Sees |
|---|---|---|
| 1. Sharded Knowledge | Each node holds only a partial matrix | Incomplete data |
| 2. Encrypted Computation | AES + Post-Quantum (Kyber) | Ciphertext |
| 3. Compiled Engine | Compiled binary, not source code | Obfuscated binary |
| 4. Matrix Language | Probability values without schema | Meaningless numbers |
Four countermeasures prevent statistical pattern recognition:
Total overhead: +0.06ms. Crucially, more usage makes the system harder to analyze, not easier.
The probability matrix is split into N partitions (default: 3) with configurable overlap (default: 10%) for redundancy. Each Cognitive Node receives only its assigned shard and performs blind matrix multiplication without knowing what the values represent.
The Brain reassembles partial results by averaging overlapping regions. Rotation changes shard boundaries periodically, and noise is added before transmission to nodes.
| Recipient | Share | Rationale |
|---|---|---|
| Cognitive Nodes | 30% | Compute providers |
| Knowledge Nodes | 20% | Expert knowledge royalty |
| Treasury | 20% | Development + operations |
| Architect | 20% | Core IP creator |
| MACLLM Holders | 10% | Investor returns |
MACLLM Chain separates the unstoppable (ORIN chain for money/tokens) from the controllable (AI layer):
Following the principle "Compress Down, Not Scale Up" — like the human brain, which doesn't grow larger when it becomes smarter:
| Level | Method | Concurrent Users | Hardware Change |
|---|---|---|---|
| 0 (Today) | Single Brain | ~10 | 1 Mac Mini |
| 1 | 64-axis + compression + batch | ~300 | None |
| 2 | + Cognitive Nodes network | ~3,000 | Community nodes |
| 3 | + Brain replication | ~30,000+ | Additional Mac Minis |
ORIN Chain provides comprehensive developer tooling to lower the barrier to building on the network:
A zero-dependency Python client (orin/sdk.py) wraps the full REST API with typed dataclasses. Developers can query, transfer mCoin, check provenance, and monitor chain health in 3 lines of code. The SDK auto-detects local vs. public endpoints and supports API key authentication.
The ORIN API (FastAPI) exposes 15+ endpoints for query flow, chain inspection, account management, and node monitoring. Read endpoints are public; write endpoints enforce per-IP rate limiting (30/min write, 120/min read). API keys use SHA-256 hashing with admin-only generation via localhost. Interactive documentation is available via Swagger UI at /docs.
A client-side block explorer provides real-time visibility into blocks, transactions, provenance records, and account balances with search and auto-refresh capabilities.
Chain state (blocks, accounts, provenance) persists across restarts using SQLite in WAL mode. The system auto-restores full in-memory state from disk on startup, ensuring zero data loss during upgrades or crashes.
MACLLM Brain maintains per-user conversation memory across sessions. Topics, entities, and summaries are persisted to disk and injected as context when users return, enabling personalized long-term interactions without external databases.
As of March 2026, the working prototype demonstrates:
| Metric | Result |
|---|---|
| Brain Benchmark Score | 922/1000 (92.2%) (1,000 questions, 10 categories incl. edge cases) |
| Self-Sufficiency | 100% (0 external LLM calls) |
| Query Flow Latency | 0.21ms average (6-step on-chain flow) |
| Chain Tests | 42/42 passing |
| Cognitive Test | 5/5 nodes pass (3 mechanisms) |
| Supply Invariant | Valid (1B mCoin verified at every block) |
| Knowledge Base | 243,000+ entries with 428 core facts, 5-level compression |
| Brain Modules | 84+ modules in production |
| Persistence | SQLite WAL mode, auto-restore on startup |
| API Security | SHA-256 API keys + per-IP rate limiting |
| Developer Tools | Python SDK, CLI, Block Explorer, Swagger UI |
| Phase | Status | Deliverable |
|---|---|---|
| Phase 1: BrainV3 | ✅ Complete | 64-axis composite embedding (65,536d), Triple Semantic Encoder, 5-level compression, 922/1000 benchmark (92.2%), generative engine |
| Phase 2: ORIN MVP | ✅ Complete | BFT consensus, 6-step flow, mCoin, provenance, matrix sharding, Proof of Cognition |
| Phase 3: Launch | ✅ Complete | Pitch deck, live demo, whitepaper, investor outreach |
| Phase 4: Public Network | ✅ Complete | orin.macllm.ai, SDK, CLI, explorer, wallet, dashboard, API keys, persistence |
| Phase 5: Enhancement | ✅ Complete | 200-question benchmark, cross-session memory, long-form answers, RAG improvements |
| Phase 6: Production | In Progress | Rust/Go core rewrite, multi-machine deployment, smart contracts, P2P networking |
MACLLM Chain represents a fundamentally new approach to AI: one where intelligence is decentralized but protected, where every unit of energy performs actual cognitive work, and where contributors are automatically compensated through on-chain provenance. The working MVP demonstrates that this vision is technically feasible on commodity hardware, with a brain that scores 922/1000 (92.2%) on benchmarks and a chain that processes queries in sub-millisecond time with full provenance.
The system's unique combination of probability matrix architecture, Proof of Cognition consensus, multi-layer knowledge protection, and dual token economics creates a platform that becomes both more powerful and more secure as it grows — the opposite of traditional centralized AI services.
MACLLM Chain Whitepaper v2.0 — March 2026
Contact: [email protected] | macllm.ai
© 2026 MACLLM. All rights reserved.