Job Description:-
Design and implement scalable, efficient, and secure data pipelines using GCP services such as BigQuery, Dataflow, Pub/Sub, Cloud Storage, and Cloud Composer.
Build data ingestion processes from various structured and unstructured data sources (e.g., APIs, files, databases).
Develop ETL/ELT frameworks and ensure timely and accurate delivery of data to analytics teams and business stakeholders.
Optimize data storage and querying performance within BigQuery and other GCP storage services.
Collaborate with data analysts, data scientists, and software engineers to understand data needs and deliver high-quality datasets.
Automate data quality checks, monitoring, and alerting for production data workflows.
Ensure compliance with data governance, security policies, and best practices in cloud data architecture.
Maintain documentation for data flows, architecture, and key decisions.
Bachelor’s degree in Computer Science, Engineering, Information Systems, or a related field.
3+ years of experience in data engineering, with at least 1–2 years focused on GCP.
Hands-on experience with BigQuery, Dataflow (Apache Beam), Pub/Sub, Cloud Storage, and Cloud Composer.
Proficiency in SQL, Python, and/or Java for data processing and pipeline development.
Experience with orchestration tools (e.g., Apache Airflow) and CI/CD pipelines.
Strong understanding of data modeling, data warehousing, and distributed computing.
Familiarity with modern data lake/data mesh architectures.
Excellent problem-solving and communication skills.
Any Graduate