Description

  • 5+ years of experience in Databricks and ETL development to design, build, and optimize robust data pipelines and ETL automation.
  • Collaborate effectively with data architects, DBAs, product owners, and other stakeholders to ensure alignment on project goals, requirements, and solutions.
  • Design, develop, and maintain scalable data pipelines in Azure using Python and Databricks, focusing on efficient ETL processes for large datasets.
  • Build and optimize data lakes and backend architectures, ensuring smooth data extraction, transformation, and loading (ETL) across systems.
  • Work with healthcare data (claims, pharma, PHI/PII), ensuring compliance with data privacy regulations and implementing security measures like tokenization.
  • Leverage Azure cloud technologies, particularly Databricks, to implement large-scale data processing and analytics solutions.
  • Integrate middleware solutions like MuleSoft for data processing between SAP and Databricks.
  • Drive data transformation and automation tasks, focusing on scalable and efficient solutions within the engineering team.
  • Familiarity with healthcare or pharma-related data, with expertise in handling sensitive data (PHI/PII) and implementing security measures such as tokenization

Education

Any Gradute