Description

Key Responsibilities ? Design and develop scalable ETL workflows to manage and transform complex datasets. ? Collaborate with data architects, data scientists, and business analysts to understand data requirements and optimize data flows. ? Implement data validation, cleansing, and transformation processes to ensure data integrity. ? Monitor and troubleshoot ETL jobs, identifying and resolving data inconsistencies or performance issues. ? Maintain comprehensive documentation for ETL processes and data flows to support data governance and audit requirements. Qualifications ? Bachelor’s or Master’s degree in Computer Science, Information Systems, or a related field. ? 6+ years of experience with end-to-end ETL processes, data modeling, and data integration. ? Proficiency with ETL tools (e.g., Informatica, Talend, Apache NiFi) and scripting languages like Python for advanced data manipulation. ? Advanced SQL skills for querying, performance tuning, and troubleshooting within large datasets. ? Experience with data warehousing concepts and tools (e.g., Amazon Redshift, Snowflake, Google BigQuery). ? Hands-on experience with Big Data technologies (e.g., Hadoop, Spark) for handling large volumes of data. ? Strong understanding of data governance, data quality frameworks, and data security protocols. ? Experience with job scheduling and monitoring tools (e.g., Apache Airflow, Control-M) to manage complex ETL workflows. ? Proven ability to optimize ETL processes for performance and scalability. ? Familiarity with cloud-based ETL environments (AWS Glue, Azure Data Factory) is a plus. ? Excellent problem-solving skills and a detail-oriented approach to data validation. Technical Skills: ETL, SQL, Data Warehousing, Informatica, Apache NiFi, Talend, Data Modeling, Python, Big Data, Hadoop, Spark, Amazon Redshift, Snowflake, Google BigQuery, Apache Airflow

Education

Bachelor's or Master's degrees