Description

Job Description

Handson with Spark (dataframe), Spark SQL, Databricks, AWS Glue, Scala/Spark or PySpark, Kafka or another streaming technology. Good learning & cross skilling ability. Design & Architecture of big data systems. Experience in ETL, Data Governance, Data Quality, Good understanding of data operations. AWS or Azure cloud. Should have experience with solutioning & consulting (MUST HAVE).

 

Roles & Responsibilities

Design/develop big data pipelines using Spark (batch/streaming) and other related technology stacks. MUST HAVE experience in solutioning big data systems and consultancy as SME on big data technologies.

Education

Any Graduate