Description

In person interview is required

  • Design, develop, and maintain scalable data pipelines using Apache Spark and Java.
  • Implement data processing workflows and ETL processes to ingest, transform, and store large volumes of data.
  • Collaborate with data scientists, analysts, and other stakeholders to understand data requirements and deliver high-quality data solutions.
  • Optimize and tune data processing jobs for performance and cost-efficiency.
  • Ensure data quality, integrity, and security across all data pipelines and storage solutions.
  • Develop and maintain data models, schemas, and documentation.
  • Monitor and troubleshoot data pipeline issues, ensuring high availability and reliability.
  • Hands-on experience with AWS services, including S3, EMR, Lambda, and Glue. Snowflake
  • Experience with SQL and NoSQL databases.
  • CICd/ Jules /Spinnaker

Education

Any Graduate