Description

Key Skills: ETL/ELT, RDF, OWL, SPARQL, Neo4j, AWS Neptune, ArangoDB, Python, SQL, Cypher, Semantic Modeling, Cloud Data Pipelines, Data Quality, Knowledge Graph, Graph Query Optimization, Semantic Search.

Roles and Responsibilities:

  • Design and build advanced data pipelines for integrating structured and unstructured data into graph models.
  • Develop and maintain semantic models using RDF, OWL, and SPARQL.
  • Implement and optimize data pipelines on cloud platforms such as AWS, Azure, or GCP.
  • Model real-world relationships through ontologies and hierarchical graph data structures.
  • Work with graph databases such as Neo4j, AWS Neptune, ArangoDB for knowledge graph development.
  • Collaborate with cross-functional teams including AI/ML and business analysts to support semantic search and analytics.
  • Ensure data quality, security, and compliance throughout the pipeline lifecycle.
  • Monitor, debug, and enhance performance of graph queries and data transformation workflows.
  • Create clear documentation and communicate technical concepts to non-technical stakeholders.
  • Participate in global team meetings and knowledge-sharing sessions to align on data standards and architectural practices.

Experience Requirement:

  • 3-8 years of hands-on experience in ETL/ELT engineering and data integration.
  • Experience working with graph databases such as Neo4j, AWS Neptune, or ArangoDB.
  • Proven experience implementing knowledge graphs, including semantic modeling using RDF, OWL, and SPARQL.
  • Strong Python and SQL programming skills, with proficiency in Cypher or other graph query languages.
  • Experience designing and deploying pipelines on cloud platforms (AWS preferred).
  • Track record of resolving complex data quality issues and optimizing pipeline performance.
  • Previous collaboration with data scientists and product teams to implement graph-based analytics or semantic search features.

Education:  Any Graduation

Education

Any Graduate