Specialize in data integration from multiple external and internal sources in batch and real-time.
Coordinate and enhance data quality and availability across the Coca-Cola data estate.
Automate and optimize existing analytic workloads by recognizing patterns of data and technology usage.
Manage own learning and contribute to technical skill building of the team.
Embrace the engineering mindset and systems thinking.
Gain deep technical expertise in the data movement patterns, practices and tools.
Design and implement the platform and frameworks required for optimal extraction, transformation and loading of data from a wide variety of data sources.
Required Skills
Solid Python development skills.
Expertise with distributed computing and big data architectures.
In-depth knowledge of AWS – IAM, Neptune, EC2.
Highly proficient in SQL.
Required Experience
2+ years of designing and deploying data analytics pipelines in cloud using Azure Data Services.
3+ years using relational databases running ETL process and managing large data transformations.
3+ years experience in building and performance tuning Spark code (Scala preferred).
Machine Learning or Data Science experience or exposure.
Experience with Big Data migrations/transformations programs in the Data Warehousing and/or Business Intelligence areas.
Hands-on experience with Hadoop and Kafka.
Experience in building self-contained applications using Docker, Vagrant, Chef.