What it is right now
Shard is a functioning early-stage system that lets: • Browsers act as Scout nodes to contribute WebGPU compute • A libp2p mesh for P2P networking • Verifier nodes run stronger local models to validate and finalize inference • A demo web app you can try live today • Clients fall back gracefully if WebGPU isn’t available • Rust daemon + Python API + web UI all wired together
It’s essentially a shared inference fabric — think distributed GPU from volunteers’ browsers + stronger hosts that stitch results into reliable responses. The repo includes tooling and builds for desktop, web, and daemon components.
Why it matters
There’s a growing gap between massive models and accessible compute. Shard aims to: • Harness idle WebGPU in browsers (scouts) • Validate and “finish” results on robust verifier nodes • Enable decentralized inference without centralized cloud costs • Explore community-driven compute networks for AI tasks
This isn’t just a demo — it’s a full stack P2P inference system with transport, networking, and workflow management.
Current limitations • Early stage, not production hardened • Needs more tests, documentation, and examples • Security and incentive layers are future work • UX around joining scheduler/mesh could improve
Come build with me
If you’re into decentralized compute, AI infrastructure, web GPU, or mesh networks — I’d love feedback, contributions, and ideas. Let’s talk about where shared inference networks could go next.