Job Description:
-DataHub, Committer, Spark-based frameworks, and machine learning for anomaly detection
-Strong expertise in metadata management, data cataloging, data lineage, data governance
-Proficiency in Java, Python, and REST API development
-Strong experience with Apache Spark for ETL pipeline design and custom framework development.
-Expertise in metadata ingestion from systems like data lakes, databases, and ETL tools.
-Hands-on experience with AWS services and cost optimization strategies.
Preferred Qualifications
-Knowledge of data governance regulations like GDPR, CCPA, or HIPAA.
-Experience with infrastructure-as-code tools such as Terraform or Helm.
-Familiarity with other metadata management tools like Amundsen, Collibra, or Alation.
-Understanding of version control, CI/CD pipelines, and open-source development practices.
Any Graduate