Modern Data Platform
Batch + streaming data pipelines using Airflow, Kafka, and dbt.
Problem Statement
Multiple disconnected data sources created inconsistent reporting and delayed analytics.
Architecture Diagram
Data Sources→Airflow→Data Lake→dbt→Warehouse→BI
Tech Stack
Apache AirflowApache KafkadbtPostgreSQLSnowflake
Data Flow Explanation
Data Flow Explanation
This architecture unifies event and batch ingestion into a single governed platform. Airflow orchestrates ingestion and quality checks, Kafka handles high-throughput events, and dbt transforms curated data models into an analytics-ready warehouse.
Delivery Outcome
- Reduced dashboard latency with standardized transformations.
- Improved trust in executive metrics with reproducible pipelines.
- Enabled scalable onboarding of new product data domains.
Screenshots

