Description

We are looking for a Data Architect with creativity and results-oriented critical thinking to meet complex challenges and develop new strategies for acquiring, analyzing, modeling and storing data.

In this role you will guide the company into the future and utilize the latest technology and information management methodologies to meet our requirements for effective logical data modeling, metadata management and database warehouse domains. You will be working with experts in a variety of industries, including computer science and software development, as well as department heads and senior executives to integrate new technologies and refine system performance.

We reward dedicated performance with exceptional pay and benefits, as well as tuition reimbursement and career growth opportunities.

What You?ll Do

Define data retention policies 
Monitor performance and advise any necessary infrastructure changes 
Mentor junior engineers and work with other architects to deliver best in class solutions 
Implement ETL / ELT process and orchestration of data flows 
Recommend and drive adoption of newer tools and techniques from the big data ecosystem

Expertise You?ll Bring

10+ years in industry, building and managing big data systems 
Building, monitoring, and optimizing reliable and cost-efficient pipelines for SaaS is a must 
Building stream-processing systems, using solutions such as Storm or Spark-Streaming 
Dealing and integrating with data storage systems like SQL and NoSQL databases, file systems and object storage like s3 
Reporting solutions like Pentaho, PowerBI, Looker including customizations 
Developing high concurrency, high performance applications that are database-intensive and have interactive, browser-based clients 
Working with SaaS based data management products will be an added advantage 
Proficiency and expertise in Cloudera / Hortonworks 
Spark 
HDF and NiFi 
RDBMS, NoSQL like Vertica, Redshift, Data Modelling with physical design and SQL performance optimization 
Messaging systems, JMS, Active MQ, Rabbit MQ, Kafka 
Big Data technology like Hadoop, Spark, NoSQL based data-warehousing solutions 
Data warehousing, reporting including customization, Hadoop, Spark, Kafka, Core java, Spring/IOC, Design patterns 
Big Data querying tools, such as Pig, Hive, and Impala 
Open-source technologies and databases (SQL & NoSQL) 
Proficient understanding of distributed computing principles 
Ability to solve any ongoing issues with operating the cluster 
Scale data pipelines using open-source components and AWS services 
Cloud (AWS), provisioning, capacity planning and performance analysis at various levels 
Web-based SOA architecture implementation with design pattern experience will be an added advantage

Education

Any Graduate