Build and develop tools to operate and support a massive data and compute infrastructure (10s of PB) through aggressive process automation.
Provide tenants with self-serve tools to provision Hadoop resources and environments.
Understand and improve the usability, reliability, and scalability of open-source Apache Hadoop services to optimize for the needs of Bloomberg application teams.
Requirements:
3+ years experience Infrastructure as Code (IaC) practices and technologies like Ansible, Chef, or Terraform.
Systems programming experience in Python, Go or Java.
A degree in Computer Science, Engineering or similar field of study or equivalent work experience.
Experience with Chaos Engineering or testing strategies for infrastructure and platform operations.
Solid understanding of Linux Operating system, Shell scripting, OS troubleshooting
Strong problem-solving and communication skills.
Experience with workflow automation tools like Airflow, Argo, etc.
CI/CD experience like Jenkins, TeamCity, etc. to manage SDLC of infrastructure and platform automation tooling.
Experience in automating deployments of technologies in the Hadoop ecosystem (HDFS, HBase, Hive, Spark, Oozie, etc.)
Experience with Cloud Native computing technologies like Kubernetes, and Containers.