Qureos

FIND_THE_RIGHTJOB.

ETL Engineer

JOB_REQUIREMENTS

Hires in

Not specified

Employment Type

Not specified

Company Location

Not specified

Salary

Not specified

Freelancing Opportunity

Job Title: ETL Engineer/ML

Experience: 7+ Years

Job Overview

We are seeking a highly skilled Machine Learning / ETL Engineer to join our Business Intelligence IT team. The ideal candidate will design, develop, and optimize large-scale ETL/ELT pipelines, integrate diverse data sources, and apply Machine Learning frameworks for data quality and anomaly detection. You will collaborate closely with analysts, data scientists, and business stakeholders to deliver scalable, reliable, and actionable data-driven solutions.

This is a critical role in ensuring data integrity, scalability, and performance of our enterprise data platforms across cloud environments such as Snowflake, AWS, and GCP.

Key Responsibilities

  • Design, architect, and implement robust, high-performance ETL/ELT pipelines for large-scale data processing.
  • Collect and integrate data from multiple sources (databases, APIs, web services) into data lakes or warehouses.
  • Develop and maintain scalable data processing solutions using Snowflake, AWS, and GCP platforms.
  • Implement data quality, anomaly detection, and monitoring systems using Machine Learning frameworks.
  • Optimize query performance and manage large-scale data models, schemas, and metadata.
  • Support data governance through documentation, access control, lineage tracking, and retention policies.
  • Collaborate with cross-functional teams (data scientists, analysts, engineers, and business units) to translate requirements into actionable technical solutions.
  • Participate in CI/CD and automation initiatives to streamline data workflows and improve efficiency.
  • Ensure data accuracy, reliability, and availability for analytical and reporting purposes.

Qualifications

  • Bachelor’s degree in Computer Science, Information Technology, or a related field.
  • 7+ years of experience in data warehouse architecture, ETL/ELT pipelines, and data engineering.
  • 5+ years of experience in data modeling and writing advanced SQL queries across Snowflake, Oracle, and Columnar Databases.
  • Strong knowledge of data modeling, schema design, and data lakes.
  • Experience with AWS services such as S3, Lambda, and Data Pipeline.
  • Hands-on experience with Machine Learning algorithms for data quality and anomaly detection.

Technical Skills

  • Programming: Python, SQL
  • ETL Tools: Databricks, IBM DataStage (preferred)
  • Cloud Platforms: Snowflake, AWS, GCP, BigQuery
  • Data Processing Frameworks: Apache Spark (or equivalent big data tools)
  • Workflow Orchestration: Apache Airflow, AWS MWAA
  • DevOps Tools: CI/CD pipelines, automation frameworks
  • Data Architecture: Data lakes, warehouses, multi-cluster architecture, shareable data features
  • Bonus: Familiarity with Databricks, DataStage, and advanced Snowflake capabilities

Job Types: Part-time, Freelance
Contract length: 12 months

Pay: ₹30,000.00 - ₹40,000.00 per month

Benefits:

  • Work from home

Experience:

  • total work: 6 years (Required)

Shift availability:

  • Night Shift (Required)
  • Overnight Shift (Required)

Work Location: Remote

© 2025 Qureos. All rights reserved.