Job Description:
We are seeking an expert-level Data Scientist with a strong focus on Big Data Engineering to join our team in Austin, Texas. This is a critical onsite role for a specialist who possesses comprehensive and authoritative knowledge in building and managing robust data infrastructure. As a Data Scientist (Big Data Engineer) Specialist, you will be instrumental in extracting value from large datasets, proactively sourcing and analyzing information to enhance business understanding, and developing AI-powered tools to automate key processes.
This position will play a vital role in the implementation of our Fraud Analytics roadmap. You will be responsible for establishing and maintaining new data engineering practices, as well as providing crucial support to our team of data analysts and data scientists. Your expertise in designing, building, and managing efficient, scalable, and reliable data pipelines will be essential to our success.
Responsibilities:
- Design, build, and manage highly efficient, scalable, and reliable data pipelines to support comprehensive data integration efforts.
- Proactively identify and fetch information from diverse data sources, ensuring data quality and integrity.
- Perform in-depth analysis of large datasets to provide actionable insights into business performance and trends.
- Develop and implement AI and machine learning models to automate critical business processes and enhance efficiency.
- Collaborate closely with data analysts and data scientists to understand their data needs and provide the necessary data infrastructure and support.
- Establish and enforce best practices for data engineering, including data governance, data quality, and data security.
- Optimize data pipelines for performance and scalability, ensuring they can handle growing data volumes and processing demands.
- Troubleshoot and resolve complex data-related issues, ensuring data availability and reliability.
- Stay current with the latest advancements in big data technologies and data engineering practices.
- Document data pipelines, data models, and data engineering processes clearly and comprehensively.
- Contribute to the overall architecture and strategy of our data analytics platform.
Candidate Skills and Qualifications:
Minimum Requirements:
- 10+ years of experience in designing, building, and managing efficient, scalable, and reliable data pipelines to support data integration efforts.
- Strong proficiency in Python for data manipulation, analysis, and pipeline development.
- Expert-level skills in SQL for data querying, transformation, and database management.
- Significant experience with Databricks on Azure for big data processing and analytics.
- Demonstrated ability to work effectively in an onsite environment in Austin, Texas (NO REMOTE WORK).
Preferred Skills:
- Experience with other big data technologies and frameworks (e.g., Spark, Hadoop, Kafka).
- Knowledge of data warehousing concepts and technologies.
- Experience with cloud-based data services on Azure.
- Familiarity with machine learning libraries and frameworks in Python (e.g., scikit-learn, TensorFlow, PyTorch).
- Experience in fraud analytics or a related domain.
- Strong understanding of data governance and data quality principles.
- Excellent problem-solving and analytical skills.
- Strong communication and collaboration skills.
- Bachelor's or Master's degree in Computer Science, Data Science, Engineering, or a related field.