Seasoned Data Engineer with a software‑developer’s mindset and a track record of designing, implementing, and operating high‑volume, mission‑critical data systems. I build scalable, maintainable, cloud‑native pipelines that deliver trusted data for analytics and machine learning—while keeping costs in check and teams moving fast.
My work philosophy is to treat data infrastructure as code: automated, versioned, observable, and tested. I partner with product and analytics stakeholders to translate business requirements into robust ETL/ELT architectures, and I mentor engineers to adopt best practices around CI/CD, modular design, and data quality.
Certifications
Technology Stack
Amazon Web Services
Data Pipelines
Scalability
Data Processing
Data Architecture
Amazon Redshift
Security Controls
Data Warehousing
Cloud Computing
Database Design
AWS Kinesis
Performance Tuning
Apache Airflow
Extract, Transform, Load (ETL)
Infrastructure as Code (IaC)
Terraform
Amazon CloudWatch
CI/CD
Dataflow
IT Automation
Query Languages
Data Lakes
SQL
Vector Databases
Apache Kafka
Amazon S3
Apache Spark
Star Schema
Apache Hadoop
Machine Learning



