Description

Job Description:
• Bachelor's or master's degree in computer science, Engineering, or a related field.
• Strong programming skills in languages such as Python, PySpark, SQL etc.
• Experience in Build and optimize ETL workflows using tools/technologies such as Spark, Snowflake, Airflow and/or Azure Data factory, Glue, Redshift etc.
• Craft and optimize complex SQL queries and stored procedures for data transformation, aggregation, and analysis.
• Develop and maintain data models ensuring scalability and optimal performance.
• Utilize Snowpark for data processing within the Snowflake platform.
• Integrate Snowflake for efficient data storage and retrieval.
• Exposure to API integrations to facilitate data workflows.
• Experience in implementing CI-CD pipelines through DevOps platforms.
• Good experience in cloud infrastructure such as Azure, AWS or GCP
Skills:

Good to have experience in Docker, Kubernetes etc
Exposure in HTML, CSS, Javascript/JQuery, Node.js, Angular/React
Experience in API development, Flask/Django is a bonus


Responsibilities:
• Collaborate with software engineers, business stake holders and/or domain experts to translate business requirements into product features, tools, projects.
• Develop, implement, and deploy ETL solutions.
• Preprocess and analyze large datasets to identify patterns, trends, and insights.
• Evaluate, validate, and optimize data models to ensure efficiency, and generalizability.
• Monitor and maintain the performance of data pipeline, data models in production environments, identifying opportunities for improvement and update as needed.
• Document development processes, results, and lessons learned to facilitate knowledge sharing and continuous improvement

Education

Bachelor's or Master's degrees