Description

• Knowledge in core data-related AWS services, ability to ingest and transform data, orchestrate data pipelines while applying programming concepts, design data models, manage data life cycles, and ensure data quality.

• designs, builds, and maintains data systems using Amazon Web Services (AWS)

• Create data models, data pipelines, and data storage solutions 

• Build ETL (Extract, Transform, Load) pipelines and workflows 

• Create documentation for data pipelines, processes, and best practices 

• SQL, Programming (Python), PySpark, and AWS Services

• Proficiency in AWS services, SQL, and data warehousing

• Experience with data modeling, warehousing, and building ETL pipelines

• Proficiency with AWS Tools: experience using AWS Glue, AWS Lambda, Amazon Kinesis, Amazon EMR , Amazon Athena, Amazon DynamoDB, Amazon Cloudwatch, Amazon SNS and AWS Step Functions.

• Programming Skills: Strong experience with modern programming languages such as Python, Spark, and Scala.

• Expertise in Data Storage Technologies: In-depth knowledge of Data Warehouse, Database technologies, and Big Data Eco-system technologies such as AWS Redshift, AWS RDS, and Hadoop.

• Experience with AWS Data Lakes: Proven experience working with AWS data lakes on AWS S3 to store and process both structured and unstructured data sets

Education

Any Gradute