Data Analysis and Reporting : Work with business analysts and stakeholders to understand their data needs and deliver data solutions. Requires a strong understanding of data processing and hands-on experience with tools such as SQL, Pandas, and Jupiter.
Data Service Platform Development : Identify recurring analytical and reporting needs and implement reusable solutions within the internal data service platform. Proficiency in Python, Django, and JavaScript is required, along with strong debugging skills.
Data Processing Pipelines : Design, build, and maintain reliable and scalable daily data pipelines. Experience with Apache Airflow or other ETL orchestration tools is essential.
External Data Interfaces : Develop and maintain data services for external systems based on internal datasets. Requires knowledge of RESTful APIs and message queue technologies such as IBM MQ or Apache Kafka.
Data Platform Infrastructure : Work with cloud-based infrastructure components, including EC2, Docker, RDS, Redshift, and S3 to support the data platform.
Qualifications Must:
Bachelor's degree in Data Engineering, Computer Science, Information Technology, or a related field
Minimum of 3 years' experience in a data engineering role
Proficiency in Python and Pandas
Strong SQL skills
Preferred:
Experience with JavaScript and web debugging tools
Familiarity with Apache Airflow
Experience with Django
Experience with AWS and related cloud technologies