A production-grade ETL pipeline for processing financial market data using Apache Airflow, dbt, and PostgreSQL.
This project demonstrates a modern data engineering pipeline that processes financial market data from Alpha Vantage API. It showcases best practices in data engineering including data validation, testing, documentation, and monitoring.
- Apache Airflow: Workflow orchestration
- PostgreSQL: Data warehouse
- dbt: Data transformation
- Docker: Containerization
- Python: Programming language
- Alpha Vantage API: Data source
- Docker and Docker Compose
- Python 3.9+
- Make (optional, for using Makefile commands)
- Clone the repository:
git clone https://github.com/javid912/datapipe-analytics.git
cd datapipe-analytics
- Create and activate a virtual environment:
python -m venv venv
source venv/bin/activate # On Windows: .\venv\Scripts\activate
- Copy the example environment file:
cp .env.example .env
- Start the services:
docker-compose up -d
airflow/
: Apache Airflow DAGs and custom operatorsdbt/
: Data transformation models and testssrc/
: Source code for data extraction and loadingtests/
: Unit and integration testsdocker/
: Dockerfile and related configurations
Please read CONTRIBUTING.md for details on our code of conduct and the process for submitting pull requests.
This project is licensed under the MIT License - see the LICENSE file for details.