A set of microservices that crawler data from a namada node, store them in a postgres database and serve them via a REST api.
🔧 This is currently being worked on. Don't expect things to work! 🔧
When using this project as a backend for Namadillo, always checkout the latest tag, as the main
branch could have an incompatible set of APIs.
If you feel like contributing to the this project, feel free to choose an issued labeled as bug
or good-first-issue
. If you want to add a new feature, we ask you to first open an issue so we discuss about it.
Our git process requires you to target main
if developing a new feature. Instead, if fixing a bug, you should checkout the latest tag maintaining branch (e.g 1.0.0-maint
).
The indexer is composed of a set microservices and a webserver, each one of these lives in his own crate. Each microservice is responsible of indexing some data from the chain and store them in the postgres database. Right now, there are 4 microservices:
chain
: goes block by block and fetches information from transactions (e.g balances)pos
: fetches the validator set each new epoch.rewards
: fetches PoS rewards each new epoch.governance
: fetches new proposal and the corresponding votes.parameters
: fetches the chain parameters.transactions
: fetches transaction starting from block height 0 (or the latest processed block height).
The webserver
is responsible to serve the data via a REST API, which are described in the swagger.yml
file in the project root.
We host a HTML version of the swagger at https://anoma.github.io/namada-indexer/.
- Create the
.env
file in the root of the project. You can use the.env_sample
file as a reference:
cp .env_sample .env
- Set the
TENDERMINT_URL
with the Namada RPC url:- Either create a local chain
- Or use a Public RPC
- Install just
- Run
just docker-up
- Install rust/cargo
- Update the
.env
values to match your setup, for example:DATABASE_URL=postgres://postgres:[email protected]:5433/namada-indexer TENDERMINT_URL=http://127.0.0.1:27657 CACHE_URL=redis://[email protected]:6379 PORT=5001
- Use the
run.sh
script inside each package. Keep in mind that PoS package have to be run always while other service might not
Instead of fetching data from a running network, for testing porpuses it's also possible to populate the databse with some random data.
cargo build
cd seeder && cargo run -- --database-url postgres://postgres:[email protected]:5433/namada-indexer
It's possible to only run the webserver and have access to the data via API.