Orchestrated end-to-end data pipelines by designing and implementing robust ETL/ELT solutions using Azure Data Factory (ADF), Databricks (PySpark, DLT), and DBT, processing high-volume data from diverse sources (SQL Server, ADLS, volumes) into Delta Lake and Unity Catalog, ensuring data quality and accessibility.
Optimized data performance and architecture by applying Spark optimization techniques (auto compact, auto optimize) and leveraging Delta Lake partitioning strategies, significantly enhancing query performance and reducing storage costs; automated incremental ingestion via DLT pipelines and Spark Structured Streaming.
Pioneered Microsoft Fabric adoption by gaining hands-on expertise and developing a real-time medallion architecture project, integrating ADF, PySpark, and Dataflow Gen 2 to load, transform, and cleanse data into Lakehouse tables, and created SQL Semantic Models for Power BI integration.