By Hemera Protocol
Note
The Hemera Indexer is a work-in-progress project. If you need to use it in production, please consult the Hemera Team first.
Hemera Protocol is a decentralized, account-centric programmable indexing network created to function as a public goods data infrastructure, enhancing the capabilities of data networks in web3. This platform supports many web3 applications, from straightforward to intricate, such as blockchain explorers, on-chain asset portfolios, social graphs, zero-knowledge (ZK) coprocessors, community quality auditing, and web3 identities. All these can benefit from or be built on top of Hemera.
As the foundation of the Hemera Protocol, the blockchain indexer plays a crucial role. It is the primary component that enables efficient and organized access to blockchain data. Initially inspired by open-source projects like Ethereum ETL, we expanded its capabilities as the Ethereum ecosystem evolved, with the emergence of more Layer 2 chains and new ERC standards. Recognizing the need for a robust solution, we decided to develop our own indexer as the first step in building the Hemera Protocol Network. As of July 5, 2024, the initial open-source version of the Hemera Indexer offers comprehensive functionality, allowing for the indexing of any EVM-compatible chains and providing all necessary data for a basic blockchain explorer. In the coming weeks, we plan to incorporate additional features from our in-house version into the open-source version.
The system can export the following entities:
- Blocks
- Transactions
- Logs
- ERC20 / ERC721 / ERC1155 tokens
- ERC20 / ERC721 / ERC1155 Token transfers
- ERC20 / ERC721 / ERC1155 Token balance
- Contracts
- Traces / Internal transactions
- L1 -> L2 Transactions
- L2 -> L1 Transactions
- Rollup Batches
- DA Transactions
- User Operations
The data can be exported into the following formats:
- Postgresql SQL
- JSONL
- CSV
Entity Types are high-level categories that group related data models. They are defined in the EntityType
enum and can be combined using bitwise operations.
- Specified using the
-E
or--entity-types
option - Examples: EXPLORER_BASE, EXPLORER_TOKEN, EXPLORER_TRACE, etc.
- Multiple types can be combined using commas
Output Types correspond to more detailed data models and are typically associated with specific Entity Types.
- Specified using the
-O
or--output-types
option - Examples: Block, Transaction, Log, Token, AddressTokenBalance, etc.
- Takes precedence over Entity Types if specified
- Directly corresponds to data class names in the code (Domain)
-
Entity Types are used to generate a set of Output Types:
- The
generate_output_types
function maps Entity Types to their corresponding Output Types. - Each Entity Type yields a set of related data classes (Output Types).
- The
-
When specifying Output Types directly:
- It overrides the Entity Type selection.
- Allows for more granular control over the exported data.
It's important to note that when using the --output-types
option, you should specify the names that directly correspond to the data class names in the code. For example:
--output-types Block,Transaction,Log,Token,ERC20TokenTransfer
These names should match exactly with the data class definitions in your codebase. The Output Types are essentially the same as the data class names, allowing for precise selection of the data models you wish to export.
-
Using Entity Types:
--entity-types EXPLORER_BASE,EXPLORER_TOKEN
This will generate Output Types including Block, Transaction, Log, Token, ERC20TokenTransfer, etc.
-
Using Output Types:
--output-types Block,Transaction,Token
This will only generate the specified Output Types, regardless of Entity Types.
When developing or using this system, consider the following:
- Entity Types provide a broader, category-based selection of data.
- Output Types offer more precise control over the exact data models to be exported.
- The choice between using Entity Types or Output Types depends on the specific requirements of the data export task.
These names should match exactly with the data class definitions in your codebase. The Output Types are essentially the same as the data class names, allowing for precise selection of the data models you wish to export.
- Ability to select arbitrary block ranges for more flexible data indexing
- Option to choose any entities for targeted data extraction
- Automated reorg detection process to ensure data consistency and integrity
- VM Instance (or your local machine)
- RPC Node of your EVM-compatible blockchain
We recommend you have this configuration to run Hemera Indexer:
- 4-Core CPU
- at least 8 GB RAM
- an SSD drive with enough space left
Based on the 2024 Ethereum, every 10k blocks, which is approximately 1.5 million transactions, consumes disk size as below
- 9GB PostgreSQL database
- 9.3GB CSV file, if you opt for the CSVV export.
- 15GB JSON file, if you opt for the JSON export That is about 35GB for every 25k blocks, for every 100k blocks, we recommend at least 150 GB for every 100k eth blocks.
If you don't have a VM in place, you can create VMs from cloud providers. Create an AWS EC2 Instance
The Indexer will consume a large number of RPC requests. Make sure you have a robust and fast RPC endpoint. Most of the time, the RPC endpoint will be the bottleneck for the indexer.
git clone https://github.com/HemeraProtocol/hemera-indexer.git
or
git clone [email protected]:HemeraProtocol/hemera-indexer.git
We recommend running from docker containers using the provided docker-compose.yaml
.
If you prefer running from source code, please check out Run From Source Code.
If you have trouble running the following commands, consider referring to the official docker installation guide for the latest instructions.
# Add Docker's official GPG key:
sudo apt-get update
sudo apt-get install ca-certificates curl
sudo install -m 0755 -d /etc/apt/keyrings
sudo curl -fsSL https://download.docker.com/linux/ubuntu/gpg -o /etc/apt/keyrings/docker.asc
sudo chmod a+r /etc/apt/keyrings/docker.asc
# Add the repository to Apt sources:
echo \
"deb [arch=$(dpkg --print-architecture) signed-by=/etc/apt/keyrings/docker.asc] https://download.docker.com/linux/ubuntu \
$(. /etc/os-release && echo "$VERSION_CODENAME") stable" | \
sudo tee /etc/apt/sources.list.d/docker.list > /dev/null
sudo apt-get update
# Install docker and docker compose
sudo apt-get install docker-ce docker-ce-cli containerd.io docker-buildx-plugin docker-compose-plugin
docker compose version
sudo yum update -y
sudo yum install docker -y
sudo service docker start
sudo systemctl enable docker
sudo usermod -a -G docker ec2-user
newgrp docker
docker --version
docker run hello-world
DOCKER_CONFIG=${DOCKER_CONFIG:-$HOME/.docker}
mkdir -p $DOCKER_CONFIG/cli-plugins
curl -SL https://github.com/docker/compose/releases/download/v2.29.6/docker-compose-linux-x86_64 -o $DOCKER_CONFIG/cli-plugins/docker-compose
chmod +x $DOCKER_CONFIG/cli-plugins/docker-compose
docker compose version
cd hemera-indexer
cd docker-compose
Alternatively, you might want to edit environment variables in docker-compose.yaml
. Please check
out configuration manual on how to configure the environment variables.
vim docker-compose.yaml
Now, run the following command to spin up the containers.
sudo docker compose up
You should be able to see similar logs from your console that indicate Hemera Indexer is running properly.
[+] Running 5/0
âś” Container redis Created 0.0s
âś” Container postgresql Created 0.0s
âś” Container indexer Created 0.0s
âś” Container indexer-trace Created 0.0s
âś” Container hemera-api Created 0.0s
Attaching to hemera-api, indexer, indexer-trace, postgresql, redis
Skip this step if you already have both installed.
sudo apt update
sudo apt install make
To deploy your project, simply run:
make development
This command will:
- Create a Python virtual environment
- Activate the virtual environment
- Install necessary system packages
- Install Python dependencies
After running this command, your environment will be set up and ready to use.
Remember to activate the virtual environment (source ./venv/bin/activate
) when you want to work on your project in the future.
source ./venv/bin/activate
Hemera Indexer requires a PostgreSQL database to store all indexed data. You may skip this step if you already have a PostgreSQL set up.
Follow the instructions about how to set up a PostgreSQL database here: Setup PostgreSQL on Ubuntu.
Configure the OUTPUT
or --output
parameter according to your PostgreSQL role information. Check out Configure Hemera Indexer for details.
E.g. postgresql://${YOUR_USER}:${YOUR_PASSWORD}@${YOUR_HOST}:5432/${YOUR_DATABASE}
.
Please check out Configure Hemera Indexer on how to configure the indexer.
python hemera.py stream \
--provider-uri https://ethereum.publicnode.com \
--postgres-url postgresql://devuser:devpassword@localhost:5432/hemera_indexer \
--output jsonfile://output/eth_blocks_20000001_20010000/json,csvfile://output/hemera_indexer/csv,postgresql://devuser:devpassword@localhost:5432/eth_blocks_20000001_20010000 \
--start-block 20000001 \
--end-block 20010000 \
# alternatively you can spin up a separate process for traces, as it takes more time
# --entity-types trace,contract,coin_balance
--entity-types EXPLORER_BASE \
--block-batch-size 200 \
--batch-size 200 \
--max-workers 32
Once you have successfully bootstrapped Hemera Indexer, you should be able to view similar logs as below.
2024-06-25 16:37:38,456 - root [INFO] - Using provider https://eth.llamarpc.com
2024-06-25 16:37:38,456 - root [INFO] - Using debug provider https://eth.llamarpc.com
2024-06-25 16:37:38,485 - alembic.runtime.migration [INFO] - Context impl PostgresqlImpl.
2024-06-25 16:37:38,485 - alembic.runtime.migration [INFO] - Will assume transactional DDL.
2024-06-25 16:37:38,502 - alembic.runtime.migration [INFO] - Context impl PostgresqlImpl.
2024-06-25 16:37:38,502 - alembic.runtime.migration [INFO] - Will assume transactional DDL.
2024-06-25 16:37:39,485 - root [INFO] - Current block 20167548, target block 20137200, last synced block 20137199, blocks to sync 1
2024-06-25 16:37:39,486 - ProgressLogger [INFO] - Started work. Items to process: 1.
2024-06-25 16:37:40,267 - ProgressLogger [INFO] - 1 items processed. Progress is 100%.
2024-06-25 16:37:40,268 - ProgressLogger [INFO] - Finished work. Total items processed: 1. Took 0:00:00.782177.
2024-06-25 16:37:40,283 - exporters.postgres_item_exporter [INFO] - Exporting items to table block_ts_mapper, blocks end, Item count: 2, Took 0:00:00.014799
2024-06-25 16:37:40,283 - ProgressLogger [INFO] - Started work.
Hemera Indexer allows you to export the blockchain data to a database, or to JSON/CSV files.
Use any PostgreSQL client to connect to your PostgreSQL instance, please make sure the user
, password
, and port
is the same as your configuration.
By default, the PostgreSQL port is open on and mapped to port 5432 of your ec2 instance, you can verify or change it in the PostgreSQL section of the docker-compose.yaml
.
If you are using any cloud services, make sure the PostgreSQL port is accessible by updating the network rules.
If you are using AWS and EC2, you can check out this post on how to configure the security group.
By default, the docker-compose.yaml
mounts the output
folder to docker-compose/output
, assuming that you are running from docker-compose
folder.
You can find exported results in docker-compose/output
.
The database and exported file locations are the same as what you configured in OUTPUT
or --output
parameter.
E.g., If you specify the OUTPUT
or --output
parameter as below
# Command line parameter
python hemera.py stream \
--provider-uri https://ethereum.publicnode.com \
--postgres-url postgresql://devuser:devpassword@localhost:5432/hemera_indexer \
--output jsonfile://output/eth_blocks_20000001_20010000/json,csvfile://output/hemera_indexer/csv,postgresql://devuser:devpassword@localhost:5432/eth_blocks_20000001_20010000 \
--start-block 20000001 \
--end-block 20010000 \
--entity-types EXPLORER_BASE \
--block-batch-size 200 \
--batch-size 200 \
--max-workers 32
# Or using environment variable
export OUTPUT = postgresql://user:password@localhost:5432/hemera_indexer,jsonfile://output/json, csvfile://output/csv
You will be able to find those results in the output
folder of your current location.
Here are some important concepts to understand from Hemera Indexer:
Dataclasses are extensively utilized during the indexing process. All outputs from Hemera Indexer are defined as dataclasses. Each indexing job depends on certain dataclasses to run and generates one or more dataclasses as the output.
A job is an indexing task that has input (dependency) dataclasses and output dataclasses. You can define one or more jobs to run during the indexing process. The Hemera Indexer will automatically determine the job dependencies and decide the order in which the jobs will run.
An Entity Type is a higher-level aggregation of desired output. For example, EXPLORER_BASE refers to basic explorer data, including blocks, transactions, and logs. By specifying an entity type, you can easily set up for your use case.
Hemera indexer can read configuration from cmd line arguments or environment variables.
- If you run Hemera Indexer in Docker, then the environment variable is easier to configure.
- If you prefer running from Source Code, command line arguments are more intuitive.
Run with python hemera.py stream --help
to get the latest instructions for arguments.
- If the name of the parameter is in
UPPER_CASE
then it's an environment variable. - If the name of the parameter starts with
--
then it's a command line argument.
Avoid specifying the same parameter from both the environment variable and the command line argument.
[Default: https://ethereum-rpc.publicnode.com
]
The URI of the web3 rpc provider, e.g. file://$HOME/Library/Ethereum/geth.ipc
or https://ethereum-rpc.publicnode.com
.
[Default: https://ethereum-rpc.publicnode.com
]
The URI of the web3 debug rpc provider, e.g. file://$HOME/Library/Ethereum/geth.ipc
or https://ethereum-rpc.publicnode.com
.
[Required]
The PostgreSQL connection URL that the Hemera Indexer used to maintain its state. e.g. postgresql://user:[email protected]:5432/postgres
.
[Required] You may specify the output parameter so Hemera Indexer will export the data to CSV or JSON files. If not specified the data will be printed to the console.
If you have multiple outputs, use "," to concat the files.
The file location will be relative to your current location if you run from source code, or the output
folder as configured in docker-compose.yaml
.
e.g.
postgresql://user:password@localhost:5432/hemera_indexer
: Output will be exported to your postgres.jsonfile://output/json
: Json files will be exported to folderoutput/json
csvfile://output/csv
: Csv files will be exported to folderoutput/csv
console,jsonfile://output/json,csvfile://output/csv
: Multiple destinations are supported.
[Default: EXPLORER_BASE
]
The list of entity types to export. e.g. EXPLORER_BASE
, EXPLORER_TOKEN
, EXPLORER_TRACE
.
The list of output types to export, corresponding to more detailed data models. Specifying this option will prioritize these settings over the entity types specified in -E. Available options include: Block, Transaction, Log, Token, AddressTokenBalance, etc.
You may spawn up multiple Hemera Indexer processes, each of them specifying different output types to accelerate the indexing process. For example, indexing trace
data may take much longer than other entities, you may want to run a separate process to index trace
data. Checkout docker-compose/docker-compose.yaml
for examples.
[Default: head
]
The database version to initialize the database. Using the Alembic script's revision ID to specify a version.
e.g. head
, indicates the latest version.
Or base
, indicates the empty database without any table.
Default value: head
The block number to start from, e.g. 0
, 1000
, etc.
If you don't specify this, Hemera Indexer will read the last synced block from the PostgreSQL database and resume from it.
The block number that ends at, e.g. 10000
, 20000
, etc.
[Default: 1000
]
The number of block records to write to each file.
[Default: 10
]
Seconds to sleep between each sync with the latest blockchain state.
[Default: 10
]
The number of non-debug rpc requests to batch in a single request.
[Default: 1
]
The number of debug rpc to batch in a single request.
[Default: 1
]
The number of blocks to batch in a single sync round.
[Default: 5
]
The number of workers, e.g. 4
, 5
, etc.
The log file to use. e.g. path/to/logfile.log
.
[Default: file_sync_record
]
How to store the sync record data. e.g. pg_base
. This means sync record data will store in pg as base
be key. or you can use file_base
which means sync record data will store in file as base
be file name',
[Default: memory
]
Use an alternative cache solution to store middleware data generated by the Indexer. e.g redis
, cache data will store in redis, redis://localhost:6379, or memory which means cache data will store in memory.