Overview
Edge Hive brings Cognitive Hive AI to edge devices—from smartwatches to desktops. Instead of sending every request to the cloud, Edge Hive runs AI inference locally, providing privacy by default, offline capability, and reduced latency.
When local specialists can't handle a request with sufficient confidence, Edge Hive seamlessly falls back to cloud specialists—giving you the best of both worlds.
Local-First Architecture
Edge Hive follows a local-first philosophy:
- Process locally by default – Your data never leaves the device unless necessary
- Encrypted at rest – All beliefs, cache, and configuration use AES-256-GCM
- Confidence-based routing – Only escalate to cloud when local confidence is low
- Offline capable – Full functionality without internet connectivity
- Battery aware – Graceful degradation based on power state
Device Support
Edge Hive automatically adapts to device capabilities:
| Device Class | Max Model | Beliefs | Cache | Strategy |
|---|---|---|---|---|
| Watch | None (cloud only) | 100 | 1 MB | Aggressive sync |
| Phone | 500M params | 10K | 50 MB | Balanced |
| Tablet | 1B params | 50K | 100 MB | Balanced |
| Laptop | 7B params | 100K | 500 MB | Local-first |
| Desktop | 70B params | 1M | 1 GB | Local-first |
Supported Local Models
Edge Hive supports popular small language models in GGUF format with quantization:
SmolLM
135M, 360M, 1.7B variants. Excellent for constrained devices.
Qwen2
0.5B, 1.5B variants. Strong multilingual support.
Phi-3-mini
3.8B params. Microsoft's efficient reasoning model.
Llama-3.2
1B, 3B variants. Meta's latest small models.
Gemma-2
2B variant. Google's efficient architecture.
Custom
Any GGUF model via llama.cpp bindings.
Edge Specialists
Edge Hive includes specialized agents optimized for on-device processing:
Cognitive Loop
Every request flows through the Edge Hive cognitive loop:
-
Request Arrives
User input or system event triggers the loop
-
Check Cache
Look for cached responses to similar queries
-
Enrich Context
Add user preferences, time, location from belief store
-
Find Best Specialist
Route to the specialist with highest domain match
-
Confidence Check
If confidence ≥ threshold: process locally. Otherwise: delegate to cloud
-
Cache & Learn
Store successful responses, update beliefs from interaction
Security Implementation
Edge Hive implements comprehensive security at every layer:
Encryption at Rest
- AES-256-GCM for all persistent data
- PBKDF2 key derivation (100K iterations)
- Per-user encryption keys
- User passwords never stored
Network Security
- TLS 1.3 mandatory (ws:// rejected)
- HMAC-SHA256 message signing
- Replay attack protection
- Certificate validation required
// Each user gets isolated, encrypted storage
// macOS: ~/Library/Application Support/EdgeHive/users/{hash}/
// Linux: ~/.local/share/edge-hive/users/{hash}/
// Windows: %APPDATA%\EdgeHive\users\{hash}\
// Directory structure (all encrypted)
users/{user_hash}/
├── beliefs.enc // Encrypted belief store
├── cache.enc // Encrypted response cache
├── config.enc // Encrypted configuration
└── session.enc // Encrypted session tokens
Using Edge Hive
use edge_hive::{EdgeHive, DeviceClass, LocalModel};
// Initialize Edge Hive with automatic device detection
let hive = EdgeHive::new()
.with_model(LocalModel::auto()) // Auto-select based on device
.with_fallback("wss://hive.example.com")
.with_confidence_threshold(0.7)
.build().await?;
// Authenticate user (derives encryption key)
hive.authenticate("user@example.com", password).await?;
// Process request - automatically routes local vs cloud
let response = hive.process("What's on my calendar today?").await?;
// Check if processed locally
if response.processed_locally {
println("Handled by local specialist: {}", response.specialist);
println("Inference: {} tokens/sec", response.tokens_per_second);
}
Privacy by Default
Edge Hive keeps your data on your device. Calendar queries, health data, personal preferences—all processed locally. Only when the local model can't handle a request does it escalate to the cloud, and even then, only the minimum necessary context is shared.