Description

  • Engineer advanced AI capabilities to power Research Digitization, Banking and Global Market solutions
  • Evaluation and subsequent implementation of Data/Model Parallelism libraries and techniques, AI Observability & Monitoring solutions, Vector databases and inference engines
  • Deploy, manage, performance tune and scale containerized applications using Kubernetes. Architect clusters leveraging fit for purpose hardware for AI workloads
  • Provide subject matter expertise in distributed and parallel computing. Analyze ML and data processing workloads to identity latency contributors, inefficient compute utilization and provide remediation recommendations
  • Automate AI infrastructure provisioning
  • Create reports for AI infrastructure usage and cost reports
  • Collaborate with multi-functional geographically distributed teams to drive cutting edge AI infrastructure and solutions

What do you need to succeed?

Must Have

  • Minimum of 2 years of hands-on AI Engineering experience
  • Degree in Computer Science or Engineering
  • Experience building, scaling, managing and monitoring ML pipelines
  • Experience with Python, Kubernetes, distributed computing
  • Solid understanding of machine learning, generative AI, agents, multi-agent collaboration, MCP servers, with focus on real-world implementation
  • Proficient in CI/CD principles, version control and best practices for deploying AI workloads to production.

Nice to Have

  • Azure Databricks, Spark, Snowflake, NVIDIA NIM, MLFlow
  • Experience building and deploying MCP Servers, AI Agents
  • Experience with any vector database.
  • Experience with Terraform

Education

Any Gradute