What you’ll do
Apply the knowledge of data characteristics and data supply pattern, develop rules and tracking process to support data quality model.
Prepare data for analytical use by building data pipelines to gather data from multiple sources and systems.
Integrate, consolidate, cleanse and structure data for use by our clients in our solutions.
Perform design, creation, and interpretation of large and highly complex datasets.
Stay up-to-date with the latest trends and advancements in GCP and related technologies, actively proposing and evaluating new solutions.
Understand best practices for data management, maintenance, reporting and security and use that knowledge to implement improvements in our solutions.
Implement security best practices in pipelines and infrastructure.
Communicate with the global teams in regions with Equifax presence to share best practices and guide adoption strategies.
Develop and implement data quality checks and troubleshoot data anomalies.
Provide guidance and mentorship to junior data engineers.
Review dataset implementations performed by junior data engineers.
What you’ll need
BS degree in a STEM major or equivalent discipline; Master’s Degree strongly preferred.
5+ years of experience as a data engineer or related role.
Experience using programming languages such as Python or SQL and intermediate level experience with scripting languages.
Intermediate level understanding and experience with Google Cloud Platforms and overall cloud computing concepts, as well as basic knowledge of other cloud environments.
Experience building and maintaining moderately-complex data pipelines, troubleshooting issues, transforming and entering data into a data pipeline in order for the content to be digested and usable for future projects.
Experience designing and implementing moderately complex data models and experience enabling optimization to improve performance.
Demonstrates advanced Git usage and CI/CD integration skills.
What could set you apart
Experience with commercial credit data products.
Cloud certification.
Experience in performing analysis using Google BigQuery, BigTable, and other Google Cloud Platform technologies, Google DataFlow, Scala+ Spark or PySpark.
Experience working with data schema / formats such as JSON, AVRO, PARQUET.
Experience or advanced degree with focus on Entity Resolution concepts, prototype or hand-on project.
Experience leading international engagements and collaborating with international teams.
#LI-Hybrid
#LI-KD1
We offer comprehensive compensation and healthcare packages, 401k matching, paid time off, and organizational growth potential through our online learning platform with guided career tracks.
Any Graduate