Description

About The Role We're building a next-generation data analytics platform on Google Cloud Platform to power in-app workflows and analytics for our users. Our stack includes Python microservices, Airflow for pipeline orchestration, and a React/Next.js frontend. You'll join a small, cross-functional team responsible for end-to-end service development, deployment, and operational excellence.

What You'll Do

  • Design, implement, and maintain backend services and APIs in Python
  • Build and optimize data pipelines using Apache Airflow
  • Collaborate with product and frontend teams to define clear service contracts
  • Develop infrastructure-as-code for GCP resources (Pub/Sub, Cloud Functions, BigQuery, Cloud Storage)
  • Ensure reliability: write tests, set up monitoring/alerting, troubleshoot production issues
  • Participate in code reviews, mentor junior engineers, and help evolve our best practices


What We're Looking For

  • 4+ years of professional Python development experience
  • Hands-on experience with Apache Airflow (authoring DAGs, operators, scheduling)
  • Strong working knowledge of Google Cloud Platform services (Compute Engine, Cloud Functions, BigQuery, Pub/Sub, IAM)
  • Experience containerizing applications (Docker) and deploying with CI/CD (GitHub Actions, Cloud Build, etc.)
  • Solid understanding of SQL and relational databases; bonus for NoSQL (Firestore/Datastore)
  • Familiarity with RESTful API design
  • Commitment to code quality: automated tests, linting, type checking


Nice-to-Haves

  • Experience with Terraform or other IaC tools
  • Knowledge of Kubernetes and serverless architectures
  • Background in event-driven or streaming data systems (Dataflow, Kafka)
  • Exposure to security best practices in cloud environments
  • Experience performing statistical analysis and data modeling (e.g., using NumPy, pandas, SciPy)
  • Familiarity with machine learning frameworks and workflows (e.g., scikit-learn, TensorFlow, Pytorch)

Education

Any Graduate