Description

Responsibilities

  • Work with overall Core team and build generic loading framework using Java, Angular, Apache Spark, Hive/Impala, Parquet and related Big Data Technologies.
  • Ensure high quality software development, with complete documentation and traceability.
  • Develop comprehensive knowledge of how areas of business, such as architecture and infrastructure, integrate to accomplish business goals.
  • Resolve variety of high impact problems/projects through in-depth evaluation of complex business processes, system processes and industry standards.
  • Translate functional requirement or specifications into technical solutions and implement the solutions with plans and executions.
  • Manage and implement data processes (Data Quality reports).
  • Develop data profiling, deduping logic, matching logic for analysis.

Required Skills

  • Good to have Data warehouse exposure Technical/Functional Proficiency.
  • Expertise with Hadoop, Spark and Hive implementations with strong programming experience in Hive, Java, Scala, Python, SQL.
  • Excellent communication skills (verbal and written).
  • Must be able to manage multiple priorities and meet deadlines.

Required Experience

  • Strong hands-on working experience in multiple Big data and Cloud technologies.
  • Experience in Big data technologies like HDFS, Hive. Hbase, Apache Spark Kafka.
  • Strong experience in validating testing and have quality product at early stage of lifecycle.
  • Agile/Scrum methodology experience is required.
  • Experience in SCMs like GIT; and tools like JIRA.
  • Experience in working on Linux shell scripting.
  • Experience in RDMS and No SQL databases.
  • Experience in Migration of SQL server code, VBA code to Other.

Education Requirements

  • Bachelor’s Degree in Computer Science, Computer Engineering or a closely related field.


 

Education

Any Graduate