Description

Key Responsibilities

  • Data Analysis and Reporting : Work with business analysts and stakeholders to understand their data needs and deliver data solutions. Requires a strong understanding of data processing and hands-on experience with tools such as SQL, Pandas, and Jupiter.
  • Data Service Platform Development : Identify recurring analytical and reporting needs and implement reusable solutions within the internal data service platform. Proficiency in Python, Django, and JavaScript is required, along with strong debugging skills.
  • Data Processing Pipelines : Design, build, and maintain reliable and scalable daily data pipelines. Experience with Apache Airflow or other ETL orchestration tools is essential.
  • External Data Interfaces : Develop and maintain data services for external systems based on internal datasets. Requires knowledge of RESTful APIs and message queue technologies such as IBM MQ or Apache Kafka.
  • Data Platform Infrastructure : Work with cloud-based infrastructure components, including EC2, Docker, RDS, Redshift, and S3 to support the data platform.

Qualifications
Must:

  • Bachelor's degree in Data Engineering, Computer Science, Information Technology, or a related field
  • Minimum of 3 years' experience in a data engineering role
  • Proficiency in Python and Pandas
  • Strong SQL skills

Preferred:

  • Experience with JavaScript and web debugging tools
  • Familiarity with Apache Airflow
  • Experience with Django
  • Experience with AWS and related cloud technologies
  • Fluent in Mandarin


 

Education

Bachelor's degree