Description

We are looking for a Data Architect with creativity and results-oriented critical thinking to meet complex challenges and develop new strategies for acquiring, analyzing, modeling and storing data.

In this role you will guide the company into the future and utilize the latest technology and information management methodologies to meet our requirements for effective logical data modeling, metadata management and database warehouse domains. You will be working with experts in a variety of industries, including computer science and software development, as well as department heads and senior executives to integrate new technologies and refine system performance.

We reward dedicated performance with exceptional pay and benefits, as well as tuition reimbursement and career growth opportunities.

What You?ll Do

  • Define data retention policies
  • Monitor performance and advise any necessary infrastructure changes
  • Mentor junior engineers and work with other architects to deliver best in class solutions
  • Implement ETL / ELT process and orchestration of data flows
  • Recommend and drive adoption of newer tools and techniques from the big data ecosystem

Expertise You?ll Bring

  • 10+ years in industry, building and managing big data systems
  • Building, monitoring, and optimizing reliable and cost-efficient pipelines for SaaS is a must
  • Building stream-processing systems, using solutions such as Storm or Spark-Streaming
  • Dealing and integrating with data storage systems like SQL and NoSQL databases, file systems and object storage like s3
  • Reporting solutions like Pentaho, PowerBI, Looker including customizations
  • Developing high concurrency, high performance applications that are database-intensive and have interactive, browser-based clients
  • Working with SaaS based data management products will be an added advantage
  • Proficiency and expertise in Cloudera / Hortonworks
  • Spark
  • HDF and NiFi
  • RDBMS, NoSQL like Vertica, Redshift, Data Modelling with physical design and SQL performance optimization
  • Messaging systems, JMS, Active MQ, Rabbit MQ, Kafka
  • Big Data technology like Hadoop, Spark, NoSQL based data-warehousing solutions
  • Data warehousing, reporting including customization, Hadoop, Spark, Kafka, Core java, Spring/IOC, Design patterns
  • Big Data querying tools, such as Pig, Hive, and Impala
  • Open-source technologies and databases (SQL & NoSQL)
  • Proficient understanding of distributed computing principles
  • Ability to solve any ongoing issues with operating the cluster
  • Scale data pipelines using open-source components and AWS services
  • Cloud (AWS), provisioning, capacity planning and performance analysis at various levels
  • Web-based SOA architecture implementation with design pattern experience will be an added advantage

Education

Bachelor's Degree