Description

Job Description: 
-DataHub, Committer, Spark-based frameworks, and machine learning for anomaly detection
-Strong expertise in metadata management, data cataloging, data lineage, data governance
-Proficiency in Java, Python, and REST API development
-Strong experience with Apache Spark for ETL pipeline design and custom framework development.
-Expertise in metadata ingestion from systems like data lakes, databases, and ETL tools.
-Hands-on experience with AWS services and cost optimization strategies.

Preferred Qualifications
-Knowledge of data governance regulations like GDPR, CCPA, or HIPAA.
-Experience with infrastructure-as-code tools such as Terraform or Helm.
-Familiarity with other metadata management tools like Amundsen, Collibra, or Alation.
-Understanding of version control, CI/CD pipelines, and open-source development practices.

Education

Any Graduate