Work with overall Core team and build generic loading framework using Java, Angular, Apache Spark, Hive/Impala, Parquet and related Big Data Technologies.
Ensure high quality software development, with complete documentation and traceability.
Develop comprehensive knowledge of how areas of business, such as architecture and infrastructure, integrate to accomplish business goals.
Resolve variety of high impact problems/projects through in-depth evaluation of complex business processes, system processes and industry standards.
Translate functional requirement or specifications into technical solutions and implement the solutions with plans and executions.
Manage and implement data processes (Data Quality reports).
Develop data profiling, deduping logic, matching logic for analysis.
Required Skills
Good to have Data warehouse exposure Technical/Functional Proficiency.
Expertise with Hadoop, Spark and Hive implementations with strong programming experience in Hive, Java, Scala, Python, SQL.
Excellent communication skills (verbal and written).
Must be able to manage multiple priorities and meet deadlines.
Required Experience
Strong hands-on working experience in multiple Big data and Cloud technologies.
Experience in Big data technologies like HDFS, Hive. Hbase, Apache Spark Kafka.
Strong experience in validating testing and have quality product at early stage of lifecycle.
Agile/Scrum methodology experience is required.
Experience in SCMs like GIT; and tools like JIRA.
Experience in working on Linux shell scripting.
Experience in RDMS and No SQL databases.
Experience in Migration of SQL server code, VBA code to Other.
Education Requirements
Bachelor’s Degree in Computer Science, Computer Engineering or a closely related field.