We are actively seeking a highly skilled Data Engineer to join our team for a long-term onsite opportunity in Wilmington, DE. This is a hands-on role requiring strong experience with Python, Spark, AWS, and Terraform. Candidates must be local or within commuting distance, as the role is 5 days onsite with Face-to-Face interviews required.
Key Responsibilities:
Develop and manage scalable data pipelines for batch processing of large datasets (TBs).
Implement and manage cloud infrastructure using Terraform and AWS services including S3, EMR, SNS, SQS, Redshift.
Build and maintain CI/CD and DataOps pipelines using Python.
Utilize Spark or PySpark for big data processing.
Collaborate with DevOps and engineering teams to deliver robust solutions.
Support infrastructure with Kubernetes and contribute to Java-based systems where needed.
Must-Have Skills:
Strong experience with Python and Spark / PySpark
Hands-on with AWS (S3, EMR mandatory)
Proficient in Terraform
DevOps/DataOps experience, including CI/CD pipeline development
Experience in Java and working with large-scale batch data processing
Nice-to-Have Skills:
Experience with Snowflake, SQL, Redshift
Familiarity with SNS, SQS, and Kubernetes
Any Gradute