Description

Key Responsibilities

  • Design and build scalable data pipelines using technologies such as Hadoop, Hive, PySpark, and Python
  • Integrate and manage data services with Amazon S3, ensuring secure and efficient object storage
  • Develop and maintain data models and database designs to support business needs
  • Implement and manage job scheduling using Autosys
  • Create data visualizations and dashboards using Power BI and Dremio
  • Write and maintain Unix/Shell scripts and contribute to CI/CD pipelines
  • Troubleshoot and optimize data workflows and transformations
  • Collaborate with stakeholders to understand data requirements and deliver solutions
  • Participate in technical assessments and problem-solving discussions

Preferred Qualifications

  • 4+ years of hands-on experience in software or data engineering
  • Strong proficiency in big data technologies and cloud platforms (AWS required; GCP is a plus)
  • Solid understanding of database fundamentals (e.g., MySQL or similar)
  • Experience with automation using Spark, Python, and Hadoop/Hive
  • Familiarity with financial services or domain-specific data is a plus
  • Excellent communication skills with the ability to articulate technical work clearly

Required Qualifications

  • 4+ years of experience in Software Engineering or equivalent (including consulting, military, or education)


 

Education

Any Graduate