Description

Minimum of 12 years of experience in designing and building production data pipelines from data ingestion to consumption within a hybrid big data architecture using GCP, Hadoop, Hive, HDFS, Hbase, Spark etc.
Expertise in one of the programming languages- Scala/Java/Python.
Experience with Data lake, data warehouse ETL build and design, and Data migration from legacy systems including Hadoop, Exadata, Oracle Teradata, or Netezza etc.
Deep understanding of GCP like
DataProc, Composer, AIRFlow and WireSafe
Data Transfer tools from GCP
Strong execution knowledge on BigQuery
GCP CiCD & SDK knowledge
Strong in PySpark , patterns of Structured streaming and Batch
Basic in JAVA

Education

Any Graduate