Description

Responsibilities:

  • Build and develop tools to operate and support a massive data and compute infrastructure (10s of PB) through aggressive process automation.
  • Provide tenants with self-serve tools to provision Hadoop resources and environments.
  • Understand and improve the usability, reliability, and scalability of open-source Apache Hadoop services to optimize for the needs of Bloomberg application teams.

Requirements:

  • 3+ years experience Infrastructure as Code (IaC) practices and technologies like Ansible, Chef, or Terraform.
  • Systems programming experience in Python, Go or Java.
  • A degree in Computer Science, Engineering or similar field of study or equivalent work experience.
  • Experience with Chaos Engineering or testing strategies for infrastructure and platform operations.
  • Solid understanding of Linux Operating system, Shell scripting, OS troubleshooting
  • Strong problem-solving and communication skills.
  • Experience with workflow automation tools like Airflow, Argo, etc.
  • CI/CD experience like Jenkins, TeamCity, etc. to manage SDLC of infrastructure and platform automation tooling.
  • Experience in automating deployments of technologies in the Hadoop ecosystem (HDFS, HBase, Hive, Spark, Oozie, etc.)
  • Experience with Cloud Native computing technologies like Kubernetes, and Containers.
  • Operates well in Agile team cadences

Education

Any Graduate