Description

Mandatory Skills:

  • Programming Languages: Scala, Spark, PySpark, Python, SQL
  • Big Data Technologies: Hadoop, Hive, Pig, MapReduce
  • ETL & Data Engineering: Data Warehouse Design, ETL, Data Analytics, Data Mining, Data Cleansing
  • Cloud Platforms: GCP, Azure
  • Tools & Frameworks: Apache Hadoop, Airflow, Kubernetes, Containers
  • Other Skills: Data pipeline creation, optimization, troubleshooting, and data validation
  • Work Experience:
    • 6+ years in Data Warehouse and Big Data technologies
    • 4+ years of hands-on experience with Scala, Spark, PySpark, Python, and SQL
    • 3+ years in strategic data planning, governance, and standard procedures
    • 3+ years in Agile environments

Good to Have Skills:

  •  
    • Experience in Data Analytics, Machine Learning, and optimization
    • Understanding of Java, ReactJS, Node.js
    • Knowledge of managing big data workloads and containerized environments
    • Experience in analyzing large datasets and optimizing data workflows

Skills:

Scala, Spark, Airflow, Big Data, GCP

Education

Any Graduate