Ability to create clean, easy to maintain, modify and support code.
Understanding the pre-processing analytical method.
Strong expertise in Unix shell scripting.
Expertise with ETL tools, specifically DataStage.
Expertise in creating and automated unit test cases.
Proficient in Python best practices, logging, and strong debugging skills.
Data querying and storage including SQL, Parquet, XML, JSON.
Understanding of Python environments, package management.
Familiar with git source control, CI/CD, agile development practices.
Able to work independently and ask for guidance when needed.
Strong communication skills and delivery focused.
Must be a self-starter and self-motivated.
Strong SQL knowledge, able to read and understand SQL.
Required Experience
8+ years of overall experience in data warehousing and data lakes.
Must have minimum 2 years of experience on GCP.
Experience with designs and implementation of ETL (Extract, transform, load) components to load data from disparate source systems using SQL Server and Microsoft BI tools.
Experience and strong knowledge in writing complex SQL statements, VBA scripts, DAX.
Experience in distributed computing frameworks such as Spark.
Education Requirements
Bachelor’s Degree in Computer Science, Computer Engineering or a closely related field.