Roles & Responsibility
Data Integration Engineer
Role Overview
This role will be responsible for conducting ingestion and systems integration activities on the Databricks platform This role requires a deep understanding of
Data Lake ingestion and integration processes and best practices
ETLELT design
CICD
System integration tools
Data pipeline management
About the role
Integrate data from a variety of source systems and knowledge to tailor integration approaches on a persystem basis
Liaise with Technical Data Steward and understanding business requirements for ingestion and integration
Document ingestion and integration flows and pipelines
Use Airflow to schedule and automate ingestion jobs
Manage metadata and master data in technical data catalog
Manage maintain and oversee ETLELT pipelines on the Databricks platform
Ensure ELTETL design meets required security and compliance guidelines and ensure PII management flagging and risk assessment during ingestion and integration
Conduct pipeline monitoring
Ensure adherence to SDLC best practices
About you
Bring 2 years of experience in data engineering ingestion pipelining and ETLELT
Hold a bachelors degree in computer science engineering statistics or related field
Have handson experience with and understanding of the following
SparkScala
SQL
PythonPySpark or similar programming language
Databricks
ETLELT development monitoring and pipelining using tools such as Apache Airflow
mibo
Data quality guidelines
CICD pipelines
Agile
Git and version control
Any Graduate