Skills Needed:
· 10+ years of experience in designing and building data pipelines in large-scale distributed systems.
· Lead the design, development, and maintenance of scalable batch and real-time data processing pipelines.
· Proficiency with Google Cloud Platform (GCP) and tools such as BigQuery, Dataflow, Pub/Sub, GCS.
· Strong experience with the Hadoop Big Data Ecosystem: HDFS, Hive, Pig, HBase, YARN, and others.
· Proficiency in Python and Scala for data engineering and scripting.
Any Graduate