The Graph Introduces Knowledge Graph Standard GRC-20

🤔 Knowledge Graphs Explained Finally

By The Graph
Dec 9, 2024, 4:47 PM
twitter

The Graph is launching GRC-20, web3's first knowledge graph standard.​ Knowledge graphs function as digital maps that show relationships between information pieces and adapt as data evolves.​

Key aspects:

  • Creates a common language for web3 data organization
  • Enables building customized applications on shared data
  • Supports true interoperability across platforms

Learn more at the Builders Office Hours with @yanivgraph on Thursday, December 5 at 5PM UTC via @GraphDevs livestream.​

Sources

Knowledge graphs are changing how AI understands data! The GRC-20 spec from @geoprotocol introduces web3-native standards for organizing information, making it more accessible & reliable for AI applications ☑️ ✨ Shaping the future of decentralized knowledge

The Graph
The Graph
@graphprotocol

✨ The Graph is pioneering a new data standard for web3 with GRC-20, a proposed common language for data across web3. Just as ERC-20 standardized value on Ethereum, GRC-20 will standardize data, information & knowledge and bring web3 to life 🌐 thegraph.com/blog/grc20-kno…

343
Reply

📢 ICYMI: @geoprotocol has introduced GRC-20 - the new data standard for web3! ✔️ Flexible knowledge graphs 🔄 True interoperability 🌐 Unlimited composability Learn how GRC-20 is shaping the future of decentralized data 🔽

The Graph
The Graph
@graphprotocol

✨ The Graph is pioneering a new data standard for web3 with GRC-20, a proposed common language for data across web3. Just as ERC-20 standardized value on Ethereum, GRC-20 will standardize data, information & knowledge and bring web3 to life 🌐 thegraph.com/blog/grc20-kno…

240
Reply

🌐 GRC-20 isn’t just for web3—it's for the entire knowledge economy. Imagine the worlds of healthcare, news, media, and politics benefiting from interconnected data. It’s time to standardize knowledge with GRC-20.

The Graph
The Graph
@graphprotocol

✨ The Graph is pioneering a new data standard for web3 with GRC-20, a proposed common language for data across web3. Just as ERC-20 standardized value on Ethereum, GRC-20 will standardize data, information & knowledge and bring web3 to life 🌐 thegraph.com/blog/grc20-kno…

431
Reply
Read more about The Graph

The Graph Publishes Technical Roadmap and 2025 Grants Report

The Graph has released its technical roadmap and 2025 grants report, providing transparency into the protocol's development direction and funding activities. **Key Updates:** - Technical roadmap now available for community review - 2025 grants report published, detailing funding allocations - Public quarterly call scheduled for March 31 **What's Next:** The Graph Foundation will host a community call on March 31 to discuss: - Strategic direction - Product updates - Protocol developments - Economic considerations The call is open to all community members interested in the protocol's future.

Web3 Infrastructure Reaches Revenue Reality Check

**The Graph Foundation's Team Lead Nick signals a fundamental shift in web3's evolution.** For the first time, blockchain projects are prioritizing **revenue and sustainability** over growth at all costs. This marks a departure from previous hype cycles toward genuine business fundamentals. **Key developments:** - Industry conversations now center on financial viability - Projects moving from expansion mode to institutional-grade infrastructure - Focus shifting from quantity to quality in blockchain data solutions This maturation follows The Graph's Q3 pivot toward faster sync times, real-time cross-chain data, and privacy-first frameworks—building the foundation for traditional finance integration with blockchain technology.

Aave Data Now Queryable in Plain English via The Graph

**Aave data access has been simplified for researchers, builders, and analysts.** Users can now query real-time lending markets, liquidations, flash loans, health factors, and governance data through plain English commands via The Graph Protocol subgraphs. **Key improvements:** - No GraphQL knowledge required - No manual unit conversions needed - Real-time data instead of stale screenshots - Natural language queries This builds on Aave's 2024 upgrade of 20 subgraphs across 10 networks to The Graph's decentralized infrastructure, making blockchain data more accessible to the Aave ecosystem.

🏦 Traditional Finance Embraces Blockchain Infrastructure

Traditional financial institutions are increasingly building on blockchain technology, creating demand for enterprise-grade data infrastructure. **Key Requirements:** - Compliance with regulatory standards - High-throughput data processing - Reliable operation at scale Substreams has emerged as a solution designed specifically to meet these institutional needs. The technology enables: - Robust data pipelines for blockchain applications - Multi-chain support for large-scale DeFi protocols - Infrastructure capable of handling billions in transaction volume This shift represents a significant step in blockchain adoption, as traditional finance moves beyond experimentation to production-grade implementations requiring institutional-level reliability and compliance.

Tempo Blockchain Now Supports The Graph's Subgraph Indexing

Tempo Blockchain Now Supports The Graph's Subgraph Indexing

**Tempo's Moderato and Mainnet networks are now compatible with The Graph's Subgraphs**, enabling developers to convert raw blockchain data into queryable GraphQL APIs without building custom infrastructure. **Key benefits for Tempo developers:** - Transform onchain data into clean, accessible APIs - No need for custom indexing infrastructure - Direct integration with The Graph's data querying tools This integration follows The Graph's recent release of four major Subgraph features, including: - **Dev Mode** for faster local testing - **Subgraph Composition** for modular, reusable data structures - **Aggregations** for instant analytics queries - **Declarative eth_calls** for parallel contract reads Developers building on Tempo can now leverage these tools to create more efficient data layers for their applications.