Mandatory Skills:
- Programming Languages: Scala, Spark, PySpark, Python, SQL
- Big Data Technologies: Hadoop, Hive, Pig, MapReduce
- ETL & Data Engineering: Data Warehouse Design, ETL, Data Analytics, Data Mining, Data Cleansing
- Cloud Platforms: GCP, Azure
- Tools & Frameworks: Apache Hadoop, Airflow, Kubernetes, Containers
- Other Skills: Data pipeline creation, optimization, troubleshooting, and data validation
- Work Experience:
- 6+ years in Data Warehouse and Big Data technologies
- 4+ years of hands-on experience with Scala, Spark, PySpark, Python, and SQL
- 3+ years in strategic data planning, governance, and standard procedures
- 3+ years in Agile environments
Good to Have Skills:
-
- Experience in Data Analytics, Machine Learning, and optimization
- Understanding of Java, ReactJS, Node.js
- Knowledge of managing big data workloads and containerized environments
- Experience in analyzing large datasets and optimizing data workflows
Skills:
Scala, Spark, Airflow, Big Data, GCP