Job Description:-
Design, build, and maintain ETL/ELT pipelines to process structured and unstructured data from various sources.
Develop and optimize data models (both batch and real-time) for analytics and business intelligence use cases.
Collaborate with data scientists, analysts, and software engineers to ensure data is available, consistent, and accessible.
Implement and maintain data quality, governance, and security controls across the data lifecycle.
Work with large-scale data warehouses, data lakes, and cloud-native data platforms (e.g., AWS, Azure, GCP).
Write clean, efficient, and maintainable code in languages such as Python, SQL, or Scala.
Monitor data pipelines and troubleshoot issues related to data latency, accuracy, and completeness.
Maintain metadata and data documentation to ensure data discoverability and reproducibility.
Evaluate new technologies and tools to continuously improve data engineering practices.
Bachelor’s or Master’s degree in Computer Science, Engineering, Information Systems, or related field.
3+ years of experience in data engineering or similar role.
Proficient in SQL and at least one programming language such as Python, Scala, or Java.
Hands-on experience with ETL tools (e.g., Apache Airflow, dbt, Talend, Informatica).
Solid understanding of data warehousing concepts and platforms such as Snowflake, Redshift, BigQuery, or Synapse.
Experience with cloud platforms like AWS (S3, Glue, Lambda), Azure, or GCP.
Familiarity with distributed data processing frameworks (e.g., Spark, Kafka, Flink).
Experience with version control (Git) and CI/CD workflows
Any Graduate