Skip to main content

Indexing With Flair

Real-time and historical custom data indexing for any EVM chain.

Flair offers reusable indexing primitives (such as fault-tolerant RPC ingestors, custom processors, re-org aware database integrations) to make it easy to receive, transform, store and access your on-chain data.

flair architecture

Why Flair?

Compared to other alternatives the main reasons are:

  • 🚀 Adopting parallel and distributed processing paradigm means high scalability and resiliency for your indexing stack. Instead of constrained sequential processing (e.g Subgraph).
  • 🧩 Focused on primitives, which means on the left you plug-in an RPC and on the right you output the data to any destination database.
  • 🚄 Native real-time stream processing for certain data workload (such as aggregations, rollups) for things like total volume per pool, or total portfolio per user wallet.
  • ☁️ Managed cloud services avoid DevOps and irrelevant engineering costs for dApp developers.
  • 🧑‍💻 Avoid decentralization overhead (consensus, network hops, etc) since we believe to enable best UX for dApps reading data must be as close to the developers as possible.

Features

  • ✅ Listen to any EVM chain with just an RPC URL.
    • Free managed RPC URLs for +8 popular chains already included.
    • Works with both websocket and https-only RPCs.
  • ✅ Track and ingest any contract for any event topic.
    • Auto-track new contracts deployed from factory contracts.
  • Custom processor scripts with Javascript runtime (with Typescript support)
    • Make external API or Webhook calls to third-party or your backend.
    • Get current or historical USD value of any ERC20 token amount of any contract address on any chain.
    • Use any external NPM library.
  • Stream any stored data to your destination database (Postgres, MongoDB, MySQL, Kafka, Elasticsearch, Timescale, etc).

Getting Started

1️⃣ Clone the starter boilerplate template and follow the instructions

git clone https://github.com/flair-sdk/starter-boilerplate.git
# ... follow instructions in README.mdx

Boilerplate instructions will create a new cluster, generate an API Key, and set up a manifest.yml to index your first contract with sample custom processor scripts.

Learn more about the structure of manifest.yml.

2️⃣ Configure Telos RPC nodes

Set a unique namespace, Telos chainId and RPC endpoint in your config. Remember that you can add up to 10 RPC endpoints for resiliency.

{
  "cluster": "dev",
  "namespace": "my-awesome-telos-indexing-dev",
  "indexers": [
    {
      "chainId": 40,
      "enabled": true,
      "ingestionFilterGroup": "default",
      "processingFilterGroup": "default",
      "sources": [
        # Highly-recommended to have at least 1 websocket endpoint
        "wss://mainnet.telos.net/evm",
        # You can put multiple endpoints for failover
        "https://mainnet.telos.net/evm"
      ]
    }
  ]
}

3️⃣ Sync some historical data using backfill command. Remember that enabled: true flag in your config enabled your indexer to capture data in real-time already.

# backfill certain contracts or block ranges
pnpm flair backfill --chain 40 --address 0x4550AaB24fFc973dFf67882a6758cac43C5Cc605 -d backward --max-blocks 10000
# backfill for a specific block number, if you have certain events you wanna test with
pnpm flair backfill --chain 40 -b 299322153
# backfill for the recent data in the last X minute
pnpm flair backfill --chain 40 --min-timestamp="30 mins ago" -d backward

4️⃣ Query your custom indexed data.

5️⃣ Stream the data to your own database.

Examples

Explore real-world usage of Flair indexing primitives for various use-cases.

DeFi

NFT

Need help?

Our engineers are available to help you at any stage.