0G Storage Achieves 2 GB/s Throughput on Mainnet for AI Data
0G Storage Achieves 2 GB/s Throughput on Mainnet for AI Data
🚀 2 GB/s decentralized storage

0G Storage has reached 2 GB/s throughput on its live mainnet, verified through StorageScan. The platform is designed specifically for AI workloads requiring high-speed data movement.
Key specifications:
- 2 GB/s production throughput (not benchmark)
- 30+ MB/s sustained mainnet performance
- Dual-layer architecture: Log Layer for sequential data, KV Layer for mutable queries
- $11/TB/month pricing
- 57,893 files and 52.96 GB currently stored
Comparison with alternatives:
- Filecoin: ~45s retrieval time, $0.19/TB/month, sealed sectors
- Arweave: ~33 GiB/day upload, $5-8/GB one-time, permanent storage
0G Storage supports both immutable and mutable data models, with native compute integration for AI training data, model weights, and agent memory storage.
Most decentralized storage supports one data model: immutable files. 0G supports two. Log Layer for large sequential data. KV Layer for real-time mutable queries. Both on the same network, integrated with Compute and Chain.
0G Storage: 30+ MB/s mainnet throughput. Dual-layer architecture. Native compute integration. Full comparison with verified data: 0g.ai/blog/0g-storag…
The numbers side by side: 0G Storage: 30+ MB/s, $11/TB/month, mutable + immutable Filecoin: ~45s retrieval, $0.19/TB/month, sealed sectors Arweave: ~33 GiB/day upload, ~$5-8/GB one-time, permanent Which fits your workload?
AI models need to move data fast. Not "fast for crypto" fast. Actually fast. 0G Storage: 2 GB/s throughput. Live on mainnet. Verified on StorageScan. That's not a benchmark. That's production.
Decentralized storage that doesn't compromise on speed. 0G Storage: up to 2 GB/s throughput. Training data, model weights, agent memory — stored on infrastructure you don't have to trust. You verify it. Visit storagescan.0g.ai for live stats.
57,893 files. 52.96 GB stored on-chain. 0G Storage is live and being used — decentralized, verifiable, built for AI data at scale. This is what the infrastructure layer looks like when it actually works.
Verification Framework Released for 107B Parameter Distributed Training
A distributed AI training network has published a verification framework that proves honest training across all nodes for their 107 billion parameter model - completed nine months ago, 48% larger than recent 72B benchmarks. **Key developments:** - Model trained with 107B parameters across distributed nodes - New verification framework addresses trust in decentralized training - Solves the core challenge: proving honest participation without centralized oversight **Why it matters:** Distributed AI training faces a fundamental problem - how to verify that remote nodes actually performed legitimate training work rather than submitting fraudulent results. This framework provides cryptographic proof of honest training. The two-step approach separates model training from verification, allowing networks to scale AI development while maintaining integrity across independent participants. [Read the technical details]()
Major Payment Giants Enable AI Agent Transactions as 18,000+ Agents Go On-Chain
**Payment infrastructure is rapidly adapting to AI agents:** - Visa launched a CLI tool for AI agent payments - Stripe integrated machine payments on Solana - Over 18,000 agents are already transacting on-chain - CFTC Chair stated "AI needs blockchain" **The shift is fundamental:** Traditional payment systems were built for human users, but the emerging agent economy requires different infrastructure. As agents begin executing autonomous transactions at scale, the need for blockchain-based identity, verified compute, and decentralized storage becomes critical. This isn't speculative—major financial institutions are building the rails now.
Hardware-Level AI Verification Talk at EthCC Cannes
Jake Salerno will present on AI verification at EthCC Cannes on April 1, addressing a critical infrastructure gap in decentralized AI systems. **Key Focus:** - Hardware-level proof of honest AI inference - Moving beyond trust-based or log-based verification - Building verification as a core component of AI systems The talk explores how to cryptographically verify what an AI model actually computed - described as the missing layer for the agent economy. This addresses the fundamental challenge of ensuring AI outputs are authentic and untampered. A detailed analysis of verification approaches is available at [0g.ai/blog](https://0g.ai/blog/why-verification-matters-decentralized-ai-training).
AI Infrastructure Attracts $620M as Market Shifts from Generic Chains
$620 million flowed into a single AI subnet over 12 months, signaling a fundamental market shift. **Major institutional moves:** - Grayscale filed its first AI crypto investment vehicle - Kraken launched 0G margin trading The market is moving beyond generic Layer 1 blockchains toward **purpose-built AI infrastructure** that integrates chain, compute, storage, and data availability in unified stacks. This follows broader trends where Bitcoin miners, facing profitability pressures around $500 per BTC, are pivoting their GPU infrastructure toward AI compute as a yielding asset. Wall Street is actively financing this transition. The era of general-purpose chains appears to be giving way to specialized infrastructure designed specifically for AI workloads.
0G Labs Combines DiLoCoX and TEE for Production-Ready Decentralized AI Training
**0G Labs has deployed two critical technologies for decentralized AI training:** - **DiLoCoX**: Delivers 357x speedup in communication efficiency - **TEE (Trusted Execution Environment)**: Ensures computation integrity through hardware verification These dual capabilities address the core bottlenecks in distributed AI systems - communication overhead and trust in computation results. The combination enables production-grade decentralized AI training at scale. This development positions 0G Labs as a key infrastructure provider in the emerging decentralized AI landscape, where both speed and verifiability are essential for practical deployment.