Key Responsibilities:
· Design and build Talend pipelines to implement Type 2 Slowly Changing Dimensions (SCD) models in Snowflake.
· Analyze and assess existing on-premises data sources (DB2, Oracle, Hadoop) for migration and integration.
· Develop and optimize data replication strategies using IBM Data Replicator and Qlik Replicate.
· Implement one-time data migration processes for history and archives and configure pipelines for CDC-based updates.
· Collaborate with data architects and business teams to define and enforce data modeling standards in Snowflake.
· Perform data profiling, validation, and reconciliation to ensure data integrity and consistency during migrations.
· Monitor and troubleshoot data pipelines, ensuring scalability, reliability, and performance.
· Document pipeline designs, workflows, and data mappings for compliance and audit requirements.
Qualifications:
· Proficiency in Talend ETL development and integration with Snowflake.
· Hands-on experience with IBM Data Replicator and Qlik Replicate.
· Strong knowledge of Snowflake database architecture and Type 2 SCD modeling.
· Expertise in containerized DB2, DB2, Oracle, and Hadoop data sources.
· Understanding of Change Data Capture (CDC) processes and real-time data replication patterns.
· Experience with SQL, Python, or Shell scripting for data transformations and automation.
· Familiarity with cloud platforms (AWS, Azure) and DevOps practices for pipeline automation.
Preferred Skills:
· Experience in data governance frameworks and metadata management.
· Working knowledge of version control tools (e.g., Git) and CI/CD pipelines.
· Exposure to Kafka or other streaming platforms for data ingestion.
· Strong troubleshooting and performance optimization capabilities
Any Graduate