Description

Skills Needed:
·        10+ years of experience in designing and building data pipelines in large-scale distributed systems.
·        Lead the design, development, and maintenance of scalable batch and real-time data processing pipelines.
·        Proficiency with Google Cloud Platform (GCP) and tools such as BigQuery, Dataflow, Pub/Sub, GCS.
·        Strong experience with the Hadoop Big Data Ecosystem: HDFS, Hive, Pig, HBase, YARN, and others.
·        Proficiency in Python and Scala for data engineering and scripting.

Education

Any Graduate