Description

We are actively seeking a highly skilled Data Engineer to join our team for a long-term onsite opportunity in Wilmington, DE. This is a hands-on role requiring strong experience with Python, Spark, AWS, and Terraform. Candidates must be local or within commuting distance, as the role is 5 days onsite with Face-to-Face interviews required.

Key Responsibilities:

Develop and manage scalable data pipelines for batch processing of large datasets (TBs).

Implement and manage cloud infrastructure using Terraform and AWS services including S3, EMR, SNS, SQS, Redshift.

Build and maintain CI/CD and DataOps pipelines using Python.

Utilize Spark or PySpark for big data processing.

Collaborate with DevOps and engineering teams to deliver robust solutions.

Support infrastructure with Kubernetes and contribute to Java-based systems where needed.

Must-Have Skills:

Strong experience with Python and Spark / PySpark

Hands-on with AWS (S3, EMR mandatory)

Proficient in Terraform

DevOps/DataOps experience, including CI/CD pipeline development

Experience in Java and working with large-scale batch data processing

Nice-to-Have Skills:

Experience with Snowflake, SQL, Redshift

Familiarity with SNS, SQS, and Kubernetes

Education

Any Gradute