Description

Responsibilities

  • Specialize in data integration from multiple external and internal sources in batch and real-time.
  • Coordinate and enhance data quality and availability across the Coca-Cola data estate.
  • Automate and optimize existing analytic workloads by recognizing patterns of data and technology usage.
  • Manage own learning and contribute to technical skill building of the team.
  • Embrace the engineering mindset and systems thinking.
  • Gain deep technical expertise in the data movement patterns, practices and tools.
  • Design and implement the platform and frameworks required for optimal extraction, transformation and loading of data from a wide variety of data sources.

Required Skills

  • Solid Python development skills.
  • Expertise with distributed computing and big data architectures.
  • In-depth knowledge of AWS – IAM, Neptune, EC2.
  • Highly proficient in SQL.

Required Experience

  • 2+ years of designing and deploying data analytics pipelines in cloud using Azure Data Services.
  • 3+ years using relational databases running ETL process and managing large data transformations.
  • 3+ years experience in building and performance tuning Spark code (Scala preferred).
  • Machine Learning or Data Science experience or exposure.
  • Experience with Big Data migrations/transformations programs in the Data Warehousing and/or Business Intelligence areas.
  • Hands-on experience with Hadoop and Kafka.
  • Experience in building self-contained applications using Docker, Vagrant, Chef.
  • Additional AWS tech exp: Container orchestration (ECS/EKS), Compute (Lambda), Event Chain Services (SQS, Kinesis), Graph ML (DGL, Graph Notebook, etc).

Education Requirements

  • Bachelor’s Degree in Computer Science, Computer Engineering or a closely related field.


 

Education

Any Graduate