
I design and operate production data platforms and pipelines, focusing on reliable ETL, SQL performance tuning, and observability. I automate infra, enforce data quality and testing, and enable analytics at scale with Spark, query engines and orchestration.
Designed and deployed the Docker Compose infrastructure for a real-time data pipeline using Kafka, Flink, Iceberg, Trino, MinIO, and Superset.
Docker containerized and configurable Airflow data pipeline for collecting and storing stock and cryptocurrency market data.
ETL pipeline using Pulumi for infrastructure as code, integrating AWS services and Snowflake for automated data flow.
Streamlit Python-based web application to analyze historical stock data.
Analyzing and monitoring Net Promoter Score (NPS) performance for healthcare companies using SQL and Power BI.
Improved the library’s data manipulation and reporting functionalities, supporting the development of efficient data pipelines and enabling scalable data solutions for analytics and modeling.