Description

As an Data Engineer, you will be a part of an Agile team to build healthcare applications and implement new features while adhering to the best coding development standards.

 

Responsibilities: -

  • Design and develop end-to-end business intelligence solutions using AWS and Databricks.
  • Build and maintain data models, ETL pipelines, and data warehousing solutions to support reporting and analytics.
  • Develop interactive dashboards, reports, and data visualizations that provide meaningful insights to stakeholders.
  • Develop and optimize ETL processes using Databricks and Apache Spark to ingest, clean, and transform large datasets.
  • Integrate data from various sources, including relational databases, APIs, and cloud storage, into a centralized data warehouse.
  • Ensure data accuracy, consistency, and availability across all BI platforms.
  • Work closely with business analysts, data engineer, and other stakeholders to gather requirements and translate them into technical solutions.
  • Collaborate with cross-functional teams to ensure BI solutions align with business objectives and data governance policies.
  • Provide technical guidance and support to stakeholders on best practices for data visualization and reporting.
  • Implement and configure AWS services, such as S3 and AWS DMS, to support CDC and analytics workloads.
  • Utilize Databricks to build scalable data processing workflows and enhance the performance of BI solutions.
  • Leverage Databricks' Delta Lake for real-time data processing and analytics.
  • Monitor and optimize the performance of ETL processes, data models, and BI reports to ensure efficiency and scalability
  • Troubleshoot and resolve performance issues, ensuring minimal downtime and optimal user experience.
  • Implement caching, indexing, and other performance-enhancing techniques in BI tools.
  • Develop and maintain visually compelling dashboards and reports using Power BI.
  • Implement best practices for data visualization, ensuring that reports are intuitive, interactive, and actionable.
  • Automate the delivery of reports and alerts to stakeholders, ensuring timely access to critical insights.
  • Ensure that BI solutions comply with data governance policies, including data quality, security, and privacy requirements.
  • Implement role-based access controls (RBAC) and data encryption to protect sensitive information.
  • Maintain accurate documentation of data sources, ETL processes, and reporting logic.
  • Strong proficiency in SQL, Python, and Apache Spark. Experience with data modeling, ETL development, and data warehousing. Hands-on experience with Power BI.
  • Knowledge of AWS services such as S3, Lambda, AWS DMS, API Gateway.
  • Strong analytical and problem-solving skills, with a keen eye for detail.
  • Experience with Databricks’ Delta Lake and real-time data processing.
  • Familiarity with data governance and data quality frameworks. Understanding of machine learning concepts and their application in BI.
  • Experience with AzureDevops for CI/CD.
  • Experience with Terraform.
  • Experience with on-prem to AWS data connectivity

Education

Any Gradute