Responsibilities: Designing, developing, and managing data integration processes, integrating data from diverse sources, transforming it to meet business requirements, and loading it into target systems.
- Daily Tasks: Developing and maintaining data integration solutions, ensuring data quality and integrity, optimizing data integration processes, supporting business intelligence and analytics, maintaining documentation and compliance.
- Skills: Proficiency in Pyspark, Apache Spark, and Python for data processing large datasets, experience with AWS tools and technologies (S3, EMR, Glue, Athena, RedShift, Postgres, RDS, Lambda, PySpark), experience with databases and data warehouses, ETL development, system integration, and CI/CD implementation.
- Desired Qualifications: Bachelor's degree in computer science, information technology, or a related field, 7-10+ years of experience in data engineering, database design, ETL processes, 5+ years of experience with programming languages such as PySpark, Python, and AWS tools and technologies, 3+ years of experience working with databases and data warehouses.
- Personal Attributes: Excellent problem-solving and communication skills, attention to detail, commitment to data quality, continuous learning mindset.
- Other Requirements: Providing technical guidance to junior developers, participating in agile development processes, understanding of data security, privacy, and compliance, and ability to collaborate with multi-functional teams
Bachelor's degree