Description

Key Responsibilities:

Design, build, and optimize data processing pipelines using Apache Spark (Core, SQL, Streaming).

Develop high-performance backend services using Java.

Integrate Spark jobs with various data sources (e.g., HDFS, Kafka, S3, Hive).

Write efficient, testable, and reusable code following best practices.

Work with large-scale datasets to perform data transformations and aggregations.

Collaborate with data engineers, analysts, and DevOps to ensure scalable solutions.

Participate in performance tuning and debugging of Spark jobs and Java services.

Support deployment and monitoring of batch and real-time jobs

Education

Any Graduate