Description

Design, develop, and optimize large-scale data processing workflows using Hadoop and Hive.

Build and maintain ETL pipelines for structured and semi-structured data across distributed systems.

Write efficient HiveQL queries and perform query tuning to ensure high performance.

Manage data stored in HDFS and work with other Hadoop ecosystem tools like Sqoop, Oozie, and Spark.

Collaborate with data analysts, engineers, and stakeholders to deliver scalable data solutions.

Ensure data quality, integrity, and security across all stages of the data lifecycle.

Key Technical Skills:

Hadoop, Hive, HDFS, HiveQL, MapReduce, Spark, Sqoop, Oozie, Python, Shell scripting, ETL, Data Warehousing

Education

Any Gradute