Seasoned Data Engineer with a software‑developer’s mindset and a track record of designing, implementing, and operating high‑volume, mission‑critical data systems. I build scalable, maintainable, cloud‑native pipelines that deliver trusted data for analytics and machine learning—while keeping costs in check and teams moving fast.

GitHub Contributions

My work philosophy is to treat data infrastructure as code: automated, versioned, observable, and tested. I partner with product and analytics stakeholders to translate business requirements into robust ETL/ELT architectures, and I mentor engineers to adopt best practices around CI/CD, modular design, and data quality.

Certifications

Technology Stack

Amazon Web Services Data Pipelines Scalability Data Processing Data Architecture Amazon Redshift Security Controls Data Warehousing Cloud Computing Database Design AWS Kinesis Performance Tuning Apache Airflow Extract, Transform, Load (ETL) Infrastructure as Code (IaC) Terraform Amazon CloudWatch CI/CD Dataflow IT Automation Query Languages Data Lakes SQL Vector Databases Apache Kafka Amazon S3 Apache Spark Star Schema Apache Hadoop Machine Learning