Description

Job Description:

  • 8-10 Years of total experience.
  • Ability to do code reviews and build the best practices.
  • Strong data engineering experience in AWS cloud.
  • Ingesting and processing data following best practices (error handling, re-processing, performance optimization, etc.) in both ELT and ETL processes.
  • Experience in managing data pipelines and providing production and operations support.
  • Strong experience in the AWS data ecosystem.
  • Strong experience in AWS native components and services such as S3, Glue, Lambda, EMR, Athena, Dynamo CloudFormation, and Step Functions...etc.
  • Experience in writing stored procedures and UDF functions in Python and SQL.
  • Experience in Databricks
  • Good experience in Big Data echo systems, Apache Spark, PySpark with Python
  • Familiarity with at least one cloud data integration tool such as Matillion, Nifi, Fivetran, HVR, SnapLogic, Talend.

Skills:

  • INFOSPHERE INFORMATION SERVER, DATASTAGE, ETL, AWS Cloud, S3, GLUE, LAMBDA, EMR, DYNAMO, Cloud Formation, Data Bricks, Big Data, Apache Spark, PySpark with Python

Nice to have.

  • Strong SQL programming skills.
  • Experience in code versioning using GIT/SVN.
  • Knowledge and experience in Redshift is highly preferred.
  • Collaborate with BI Engineers and other groups to ensure easy-to-use data access.
  • Strong problem-solving and communication skills.


 

Education

Any Graduate