Job Description
Mandatory Skills: Microsoft Azure, Hadoop, Spark, Databricks, Airflow, Kafka, Py spark
Requirments:
Experience working with distributed technology tools for developing Batch and Streaming pipelines using.
SQL, Spark, Python
Airflow
Scala
Kafka
Experience in Cloud Computing, e.g., AWS, GCP, Azure, etc.
Able to quickly pick up new programming languages, technologies, and frameworks.
Strong skills building positive relationships across Product and Engineering.
Able to influence and communicate effectively, both verbally and written, with team members and business stakeholders
Experience with creating/ configuring Jenkins pipeline for smooth CI/CD process for Managed Spark jobs, build Docker images, etc.
Working knowledge of Data warehousing, Data modelling, Governance and Data Architecture
Experience working with Data platforms, including EMR, Airflow, Data bricks (Data Engineering & Delta Lake components)
Experience working in Agile and Scrum development process.
Experience in EMR/ EC2, Data bricks etc.
Experience working with Data warehousing tools, including SQL database, Presto, and Snowflake
Experience architecting data product in Streaming, Server less and Microservices Architecture and platform.
Any Graduate