Description

Key Responsibilities: 

 

  • Design, develop, and maintain ETL pipelines and data workflows using Python and PySpark .
  • Build and manage RESTful APIs for data access and integrations .
  • Work closely with data scientists and analysts to deliver clean, reliable, and well-structured data.
  • Optimize SQL queries, stored procedures, and performance for Oracle DB and MySQL.
  • Implement and automate CI/CD pipelines for efficient deployment.
  • Write unit and integration tests using PyTest .
  • Parse and process XML files, handle file operations efficiently using Python.
  • Ensure data quality through validation, cleansing, and monitoring pipelines in real-time.
  • Follow Agile methodologies for iterative development and delivery.
  • Apply Object-Oriented and Functional Programming principles in service design and architecture. 


     

Requirements:

 

  • Strong experience in Python (Backend only) with focus on data structures, OOPs, and algorithms.
  • Hands-on experience with PySpark for large-scale data processing.
  • Solid knowledge of Django or Flask web frameworks.
  • Expertise in SQL with deep understanding of query tuning and stored procedures.
  • Proficiency in file handling, XML parsing, and data validation using Python.
  • Familiarity with REST API design and integration.
  • Good knowledge of CI/CD tools like Git, Jenkins.
  • Exposure to cloud platforms like GCP (preferred), AWS, or Azure.
  • Excellent communication and interpersonal skills.
  • Strong attention to detail and problem-solving mindset. 


     

Good to have Skills:

 

  • Certification in cloud (e.g., GCP, AWS) or Python-related technologies.
  • Experience in working with financial domain applications or data-heavy systems

Education

Any Gradute