Job Description:-
Data Pipeline Development
Design and develop robust ETL/ELT pipelines for ingesting, transforming, and loading data from various sources.
Ensure data accuracy, consistency, and availability for downstream analytics and machine learning models.
Data Infrastructure & Architecture
Build and maintain scalable data architectures (e.g., data lakes, data warehouses).
Optimize data storage, query performance, and reliability using modern data tools and cloud services.
Collaboration & Integration
Work with stakeholders to understand data requirements and translate them into technical solutions.
Collaborate with software engineers and DevOps to integrate data systems with applications and services.
Data Quality & Governance
Implement and maintain data validation, monitoring, and quality checks.
Ensure compliance with data governance, security, and privacy policies.
Tooling & Automation
Automate manual processes and data workflows where possible.
Contribute to tooling that improves productivity and transparency across data operations.
Required:
Bachelor's degree in Computer Science, Engineering, or a related field (or equivalent experience)
3+ years of experience as a data engineer or similar role
Proficiency in SQL and at least one programming language (e.g., Python, Java, Scala)
Experience with modern data tools and frameworks (e.g., Apache Airflow, dbt, Spark, Kafka)
Hands-on experience with cloud platforms (AWS, GCP, or Azure) and services like S3, Redshift, BigQuery, or Snowflake
Strong understanding of data modeling, schema design, and performance optimization
Any Graduate