Employment Type: Contract / Remote
About the Role
We are seeking a Senior Data Quality Engineer to support data platform initiatives for a leading global streaming and media technology company. This role is focused on ensuring the accuracy, reliability, and integrity of large-scale data systems and pipelines that power analytics, marketing, and business insights across the organization.
As a Senior Data Quality Engineer, you will work closely with data engineers, analysts, and product teams to validate complex data pipelines and ensure that every release meets strict quality standards before reaching production. Your work will play a critical role in maintaining trusted data that supports key business decisions and customer insights.
What You Will Do
- Validate ETL processes, transformation logic, and business rules across modern data platforms such as Snowflake and Databricks before code changes are released to production.
- Partner closely with data engineering teams to proactively identify potential risks or data issues early in the development lifecycle.
- Ensure the quality of every release through comprehensive testing strategies, including unit, integration, regression, and end-to-end data validation.
- Design and implement automated and reusable tests to improve coverage, reduce regression risk, and accelerate development cycles.
- Translate business and technical requirements into data validation scenarios that verify KPIs, metrics, and reporting accuracy.
- Define and manage data quality metrics, SLAs, and validation frameworks for assigned data domains.
- Improve and expand the existing data testing frameworks with a strong focus on automation and scalability.
- Collaborate with data analysts, product managers, and engineering teams to ensure data accuracy, completeness, and usability.
- Serve as a subject matter expert (SME) for data quality practices, mentoring junior engineers and supporting best practices across teams.
Basic Qualifications
- 5+ years of experience validating data pipelines, ETL processes, and data warehouse systems.
- Expert-level SQL skills with experience analyzing and validating large datasets (terabytes or more).
- Hands-on experience with modern data platforms such as Snowflake, Hive, Databricks, or similar technologies.
- Strong Python programming skills with experience building automated tests for data pipelines.
- Familiarity with data pipeline orchestration and processing tools such as Airflow and Spark.
- Strong analytical thinking and problem-solving abilities with high attention to detail.
- Excellent collaboration and communication skills with the ability to work effectively across technical and business teams.
Preferred Qualifications
- Experience with BDD testing frameworks such as Behave.
- Experience working in cloud environments such as AWS.
- Familiarity with data observability and quality monitoring tools such as Deequ, Great Expectations, Monte Carlo, or Datafold.
- Experience with AI-assisted development tools such as GitHub Copilot or similar developer productivity tools.
- Experience working in streaming, digital media, or large-scale consumer technology environments.
Location Preference
While this role is open to candidates across the U.S., candidates based in California or Utah are strongly preferred.
Job Type: Full-time
Pay: $100,000.00 - $150,000.00 per year
Work Location: Hybrid remote in Irvine, CA 92618