Description

What are we looking for? 

 

To be successful in this role a strong interest and knowledge within data engineering is needed, both when it comes to developing data pipelines as well as designing data models.

 

Technical skills and core competencies

  1. Strong understanding of Data Architecture and models and experience leading data driven projects.
  2. Solid expertise with strong opinions on Data Modelling paradigms such as Kimball, Inmon, Data Marts, Data Vault, Medallion etc.
  3. Strong experience with Cloud Based data strategies and big data technologies – AWS Preferred. Ability to create backend services in Python that enables the data pipelines is required.
  4. Demonstrated experience on designing data platforms on AWS for batch and stream processing pipelines.
  5. Hands-on experience using AWS Managed and other big data services such as EMR, Glue, S3, Kinesis, DynamoDB, ECS is a must.
  6. Strong understanding of working of Apache Spark is a must.
  7. Strong understanding of various Data Lake/Lakehouse storage formats such as Delta, Iceberg, Hudi
  8. Experience designing data lakehouse with Medallion architecture is desirable.
  9. Solid experience in designing data pipelines for ETL with expert knowledge on ingestion, transformation, and data quality is a must
  10. Hands-on experience in SQL is a must.
  11. Expertise designing ETL pipelines combining Python + SQL is required.
  12. Understanding of data manipulation libraries in python like Pandas, Polars, DuckDB is desired
  13. Experience in designing the Data visualization with different tools such as Tableau and PowerBI is desirable.
  14. Working knowledge of other Data Platforms on Azure, Databricks, Snowflake is desirable but not must.

 

Responsibilities
  • Participate in the design and developing features in the existing Data Warehouse.
  • Provide leadership in establishing connection between Engineering, product and analytics/data scientists team.
  • Design, implement, update existing/new batch ETL pipelines
  • Define and implement data architecture.
  • Partner with both engineers and data analysts to build reliable datasets that can be trusted, understood, and used by the rest of the company.
  • Work with various data orchestration tools (Apache Airflow, Dagster, Prefect and others)
  • Embrace a fast-paced start-up environment.
  • You should be passionate about your job and enjoy a fast-paced international working environment.
  • Background or experience in the telecom industry is a plus but not a requirement.
  • Love automating and enjoy monitoring
Experience8 – 12 years

Education

Any Graduate