Description

• Develop and maintain backend systems using Python, PySpark ensuring high performance, scalability, and reliability
• Participate in the design and implementation of data engineering and ETL pipelines using PySpark
• Collaborate with cross-functional teams to identify and prioritize project requirements
• Mentor and guide engineers, providing technical guidance and code reviews
• Stay up-to-date with the latest technologies and frameworks, and apply this knowledge to improve existing systems and processes
• Lead the technical direction of the team, including architecture, design, and implementation of software systems
• Collaborate with product owners to define and prioritize requirements
• Develop and maintain technical documentation, including architecture diagrams, and technical specifications
• Participate in code reviews, ensuring high-quality code and adherence to coding standards
• Collaborate with DevOps and Operations teams to ensure smooth deployment and operation of applications

• Expert in DevOps practices and tools for CI/CD pipelines.
• Willingness to learn new technologies and adapt to new challenges
• Proven experience in leading technical teams, including technical guidance, mentoring, and coaching
• Strong technical leadership skills, including the ability to make technical decisions, prioritize tasks, and manage technical resources
• Excellent communication and collaboration skills, with the ability to work effectively with cross-functional teams
• Ability to drive technical innovation, including researching new technologies, evaluating technical options, and recommending technical solutions
• Strong problem-solving skills, with the ability to debug complex issues, optimize system performance, and ensure high-quality software delivery
• Experience with Agile development methodologies, including Scrum or Kanban, and the ability to apply these principles to lead the team

• Experience with cloud platforms, including Amazon Web Services (AWS), Microsoft Azure, or Google Cloud Platform (GCP)
• Experience with PySpark for Data Engineering/ETL pipelines
• Familiarity with containerization using Docker and orchestration using Kubernetes

Education

Any Gradute