1. Role Objective
- Build, operate, and govern production-grade data and analytics solutions that span Databricks (Pipelines, Delta Live Tables, Genie, Agent Bricks) and Microsoft Fabric (Data Engineering, Lakehouse, Data Warehouse, Power BI).
- Deliver fast, reliable, and cost-optimized data flows while maintaining enterprise-grade security and observability.
2. Core Responsibilities
Architecture & Design
- Design end-to-end ingestion, transformation, and serving layers across Databricks and Fabric.
- Define datamodel standards (star schema, CDC, semistructured handling).
Pipeline Development
- Implement CICD-ready pipelines using Databricks Pipelines/Jobs API and Fabric pipelines (SparkSQL, notebooks).
- Enable realtime streaming (Event Hub/Kafka → Structured Streaming → Fabric Lakehouse).
Data Quality & Governance
- Register assets in Unity Catalog & Fabric Lakehouse catalog; enforce rowlevel security, data masking, and Purview lineage