Job Description:
- We are seeking a detail-oriented Data Test Engineer to join our data migration and cloud modernization team.
- The ideal candidate will have hands-on experience testing complex ETL pipelines, data migration workflows, and cloud data platforms like Snowflake, with exposure to legacy ETL tools such as Ab Initio or Informatica.
- Experience in automating data validation, performance testing, and supporting real-time ingestion using Kafka or similar technologies is essential.
Key Responsibilities:
- Design, develop, and execute test plans for data migration projects moving data from legacy systems to Snowflake.
- Validate data pipelines developed using ETL tools like Ab Initio and Informatica, ensuring data quality, accuracy, and integrity.
- Develop automated test scripts and frameworks using Python for data validation, reconciliation, and regression testing.
- Perform end-to-end data validation including schema validation, volume checks, transformation logic verification, and performance benchmarking.
- Test real-time data ingestion workflows integrating Kafka, Snowpipe, and Snowflake COPY commands.
- Collaborate closely with development, data engineering, and DevOps teams to identify defects, track issues, and ensure timely resolution.
- Participate in designing reusable test automation frameworks tailored for cloud data platforms.
- Ensure compliance with data governance, security, and regulatory requirements during testing.
- Document test cases, results, and provide clear reporting to stakeholders.
- Support CI/CD pipelines by integrating automated testing into the deployment workflow.
Required Skills and Experience:
- 5+ years in data testing or quality assurance with strong experience in data validation and ETL testing.
- Hands-on experience testing data migrations to Snowflake or other cloud data warehouses.
- Familiarity with legacy ETL tools like Ab Initio or Informatica and their testing methodologies.
- Proficient in scripting languages such as Python for test automation and data validation.
- Knowledge of real-time data streaming platforms such as Kafka, Kinesis, or equivalents.
- Strong SQL skills for writing complex queries to validate data integrity and transformations.
- Experience with automated testing tools and frameworks for data quality checks.
- Understanding of cloud environments, particularly AWS services (S3, Lambda, Glue).
- Familiarity with CI/CD tools and practices to integrate automated testing.
Preferred Qualifications:
- Experience with performance and load testing of data pipelines.
- Knowledge of data governance and compliance frameworks.
- Exposure to BI tools such as Tableau, Power BI for validating data consumption layers.
- Certifications in data quality or cloud platforms (Snowflake, AWS) are a plus.