Description

Key Skills: GCP Data Services & Storage, Big Data & Processing Frameworks, Monitoring & Optimization, Project Management & Agile, ETL & Data Pipelines.

Roles & Responsibilities:

  • Build and manage ETL pipelines using Google Cloud Dataflow.
  • Design and implement data storage solutions using BigQuery and Google Cloud Storage.
  • Develop and optimize data pipelines to handle large-scale data ingestion, transformation, and storage.
  • Work closely with data scientists and analysts to ensure data quality, accuracy, and availability.
  • Integrate Cloud Pub/Sub for real-time data streaming.
  • Manage and monitor data flow processes and troubleshoot issues as they arise.
  • Ensure the security of data through proper encryption, access control, and compliance with data privacy regulations.
  • Collaborate with cross-functional teams to meet business requirements and optimize data workflows.

Skills Required:

  • Google Cloud Storage (GCS) - Experience in handling structured and unstructured data, lifecycle management, and access control.
  • Google Cloud Dataflow (Apache Beam) - Experience in building scalable ETL/ELT pipelines for batch and streaming data processing.
  • Google Dataprep - Understanding of data preparation, cleansing, and transformation for analytics.
  • IAM & Role-Based Access Control (RBAC) - Knowledge of access management and security best practices for GCP services.
  • Google Cloud Operations Suite (Stackdriver) - Experience in monitoring, logging, and alerting for data workflows.

Educations: Graduate with a relevant field

Education

Any Graduate