Description

Job Description:-

Key Responsibilities

Data Pipeline Development

Design and develop robust ETL/ELT pipelines for ingesting, transforming, and loading data from various sources.

Ensure data accuracy, consistency, and availability for downstream analytics and machine learning models.

Data Infrastructure & Architecture

Build and maintain scalable data architectures (e.g., data lakes, data warehouses).

Optimize data storage, query performance, and reliability using modern data tools and cloud services.

Collaboration & Integration

Work with stakeholders to understand data requirements and translate them into technical solutions.

Collaborate with software engineers and DevOps to integrate data systems with applications and services.

Data Quality & Governance

Implement and maintain data validation, monitoring, and quality checks.

Ensure compliance with data governance, security, and privacy policies.

Tooling & Automation

Automate manual processes and data workflows where possible.

Contribute to tooling that improves productivity and transparency across data operations.

Qualifications

Required:

Bachelor's degree in Computer Science, Engineering, or a related field (or equivalent experience)

3+ years of experience as a data engineer or similar role

Proficiency in SQL and at least one programming language (e.g., Python, Java, Scala)

Experience with modern data tools and frameworks (e.g., Apache Airflow, dbt, Spark, Kafka)

Hands-on experience with cloud platforms (AWS, GCP, or Azure) and services like S3, Redshift, BigQuery, or Snowflake

Strong understanding of data modeling, schema design, and performance optimization

Education

Any Graduate