• Develop data pipelines using Cloud-Native tools like AWS Glue based on Apache Spark and Step Functions.
• Assemble large, complex data sets that meet functional / non-functional business requirements.
• Leverage serverless cloud services to prepare (extract and transform) and load large numbers of datasets for data processing.
• Extend standard ETL tool capabilities using Glue, Python/PySpark, Step Functions, SQS and Athena.
• Implement overall solution comprising of ETL jobs, Lambda and Python.
• Support the implementation of data analytics products.
• Develop and integrate custom developed software solutions to leverage automated deployment technologies.
• Develop, prototype, and deploy solutions in AWS Cloud.
• Coordinate closely with functional team to ensure requirements are clearly understood.
• Analyze (through proof of concept, performance, and end-to-end testing) and effectively coordinate the infrastructure/service needs working with the architecture and Data Center teams.
• Closely work with the architecture team to review the design and ETL code.
• Use industry leading DevOps tools such as AWS CloudFormation.
• Communicate key project data to team members and build team cohesion and effectiveness.
• Leverage Atlassian tool suite like JIRA and Confluence to track activities.
• Identify and apply best practices and standard operating procedures.
• Create innovative solutions to meet the technical needs of customers.
Basic Qualifications:
• Experience working in the software development lifecycle, with strong experience on ETL based development.
• Experience working with databases like DynamoDB. Experience with querying tools like SQL.
• Experience with containerization tools like Kubernetes.
• Experience using Delta Lake.
• Experience with data catalog tools like AWS Glue Catalog and DataHub.
• Experience working with programming languages (Python required)
• Experience working in a fast-paced development environment with drive to completion.
• Experience with development using Amazon Web Services (AWS) big data technologies.
• Well versed with using version control systems (CodeCommit preferred)
• Well versed with using issue/problem tracking systems (Jira preferred)
• Candidate must have bachelor’s with 8-12 years of prior relevant experience or master’s with 6-10 years of prior relevant experience.
Preferred Qualifications:
• Working experience on AWS Glue and Glue studio.
• Experience building processes supporting data transformation, data structures, metadata, dependency and workload management.
• Experience with Amazon QuickSight.
• Prior experience working with federal government contract
Any Gradute