Description

Job Description

  • Strong hands-on experience with developing ETL solutions with PySpark and Python.
  • Create ELT data pipeline, optimize Spark clusters on containerized environments.
  • Good knowledge of S3 bucket, Kubernetes, OpenShift, NAS storage systems.
  • Development experience with one or more ETL tools such as IBM DataStage, Talend, Apache Airflow, SSIS, etc.,
  • Experience with working with security (CyberArk), file transfer (SFTP), reporting (Tableau), file system (NAS/S3), data (MSaaS, MongoDb, Hive) in conjuntion with ETL pipelines

Education

Any Graduate