Description

Required Skills:

8+ years of experience in data engineering, with a proven track record of designing, building, and maintaining large-scale data platforms.

3+ years of experience in a lead or architect role, demonstrating strong leadership and mentorship skills.

2+ year experience in SAS coding

2+ year of experience working on Google or AWS cloud developing data solutions. Certifications preferred

Expert-level understanding of data warehousing concepts, distributed systems, and cloud-based data platforms (AWS, GCP).

Hands-on experience with Python/Pyspark/Scala and basic libraries for machine learning is required

Experience with containerization and related technologies (e.g. Docker, Kubernetes) is a must

Knowledge of agile (scrum) development methodology is a plus

Strong development/automation skills

Can-do attitude on solving complex business problems, good interpersonal and teamwork skills

Possess great communication skills and must be able to directly work with business team to collect, analyze and put together appropriate requirements and solutions

Responsibilities:

Ability to discuss requirements with key stakeholders, document dataflow and capture requirements. Critical thinking target state on adaptable solution and recommend teams to drive best solution into implementation

Implement scalable, high-performance data pipelines to support complex data processing needs and enforce data governance policies, ensuring data quality, Integration to audit tools

Drive innovation by researching and recommending cutting-edge technologies and solutions to enhance data processing, storage, and analysis capabilities

Able to develop high-level and detailed technical design and solutions considering scalability, security, and performance of systems

Identify opportunities for process improvements and contribute to optimization efforts individually

Identify and resolve technical issues that arise during the solution life cycle

Experience in business process modeling and documentation tools

Responsible for moving all legacy SAS Workloads to cloud/Container platform solutions

Ability to research and assess open-source technologies and components to recommend and integrate into the design and implementation

Analyze Current Model SAS code and convert SAS based pipelines to PySpark to execute on distributed ecosystems

Education

Any Gradute