A data engineer who loves turning complex data challenges into streamlined, scalable solutions. With expertise in building end-to-end data pipelines, architecting cloud infrastructures, and mastering cutting-edge tools like Microsoft Fabric and Delta Lake, I craft reliable data foundations that empower smarter decisions. Always curious, driven by a passion to innovate and make data work smarter — not harder.
Developed and optimized scalable data pipelines on Microsoft Fabric using Apache Spark and Delta Tables for efficient, reliable, and high-performance data processing.
Designed and implemented scalable data pipelines using Azure Data Factory and Azure Databricks to ingest, transform, and store data in ADLS Gen2 for advanced analytics.
Engineered scalable news ingestion and transformation pipelines in Microsoft Fabric using Dataflows, PySpark Notebooks, and Delta Lake on ADLS Gen2 to support sentiment analysis and real-time reporting.
Curated a well-organized digital garden in Obsidian using HTML to showcase comprehensive solutions and insights for a wide array of SQL problems.
Built a rich digital garden in Obsidian with HTML to break down and explain advanced PySpark concepts through practical examples and structured notes.
Automated machine learning pipeline for predictive analytics and model deployment.
Real-time data processing pipeline using Kafka and Spark Streaming.
Implemented data governance policies and metadata management system.
Designed and implemented a scalable enterprise data lake architecture.
Whether you're looking to build a new data infrastructure, optimize existing pipelines, or implement advanced analytics, I'd love to discuss how I can help.