Qureos

FIND_THE_RIGHTJOB.

Data Engineer

JOB_REQUIREMENTS

Hires in

Not specified

Employment Type

Not specified

Company Location

Not specified

Salary

Not specified

Position Overview

We are seeking a skilled Data Streaming Engineer to join our engineering team ((US Client Project). The successful

candidate will play a critical role in developing and maintaining real-time data pipelines, modernizing

analytical workloads, and supporting operational excellence in cloud-based environments.

This role requires a self-starter with a proven track record of owning issues, finding creative solutions to

complex problems, and driving continual improvement initiatives. Strong organizational skills and a

thirst for knowledge are essential.

Responsibilities

Streaming ETL Pipeline Development

  • Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream

API)

  • Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and

PostgreSQL CDC

  • Implement transformation logic including joins, enrichment, normalization, schema mapping,

and event-time semantics

  • Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal

services

Modernization Away from dbt (Data Build Tool) & Snowflake

  • Assist in migration of analytical and transformation workloads from MongoDB, dbt (Data Build

Tool), and Snowflake to Flink/PostgreSQL

  • Translate existing dbt models, DAGs, and downstream dependencies into equivalent Flink

transformations or SQL-based operations

  • Support design of new schema models, data quality checks, and lineage patterns to replace

dbt/Snowflake functionality

  • Participate in validation and regression testing to ensure fidelity during migration

Operational Support

  • Deploy and update Flink jobs on containerized cloud-based platforms (ECS)
  • Handle routine operational tasks including job configuration updates, savepoint management,

and monitoring

  • Troubleshoot failing jobs and ensure pipeline reliability

Collaboration & Documentation

  • Work closely with platform engineering teams to define requirements for migration and pipeline

development

  • Produce clear documentation for data flows, schemas, and newly created Flink jobs

Requirements

  • 4+ years of experience in data engineering or streaming pipeline development
  • Practical experience with Apache Flink DataStream API
  • Hands-on experience with Apache Kafka as a CDC or event streaming backbone
  • Familiarity with Debezium-based connectors or similar CDC frameworks
  • Solid understanding of SQL, CTEs, relational modeling, and PostgreSQL
  • Experience working with dbt and Snowflake, with ability to analyze models and transformation

logic

  • Demonstrated ability to develop software in Java for pipeline development
  • Familiarity with cloud environments (AWS preferred)
  • Experience refactoring or replacing dbt models with streaming transformations
  • Exposure to schema evolution strategies and metadata management
  • Some experience with CI/CD workflows for data pipelines
  • Familiarity with containerization concepts (Docker, ECS/EKS, or Kubernetes)

Core Competencies

  • Detail-oriented with strong problem-solving and troubleshooting skills
  • Strong desire for operational excellence and continuous improvement
  • Ability to translate analytical requirements into scalable data flows
  • Comfortable working within established infrastructure and collaborating with senior engineers

on larger changes

  • Effective written and verbal communication skills in English, especially for documentation and

collaboration

Our Values

  • Be Committed: Work hard, own the problem, keep learning
  • Create Trust: Deliver on promises, build trust with clients and colleagues
  • Be Bold: Experiment, speak up
  • Deliver High Performance: Prepare thoroughly, commit fully, and deliver results

Job Types: Full-time, Permanent

Pay: ₹800,000.00 - ₹1,200,000.00 per year

Benefits:

  • Health insurance
  • Paid sick time
  • Provident Fund

Ability to commute/relocate:

  • Bengaluru, Karnataka: Reliably commute or planning to relocate before starting work (Required)

Application Question(s):

  • Skills Required: Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream

API)

  • Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and

PostgreSQL CDC

  • Implement transformation logic including joins, enrichment, normalization, schema mapping,

and event-time semantics

  • Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal

services

Experience:

  • Data Streaming Engineer : 4 years (Required)

Work Location: In person

Similar jobs

No similar jobs found

© 2026 Qureos. All rights reserved.