Position Overview
We are seeking a skilled Data Streaming Engineer to join our engineering team ((US Client Project). The successful
candidate will play a critical role in developing and maintaining real-time data pipelines, modernizing
analytical workloads, and supporting operational excellence in cloud-based environments.
This role requires a self-starter with a proven track record of owning issues, finding creative solutions to
complex problems, and driving continual improvement initiatives. Strong organizational skills and a
thirst for knowledge are essential.
Responsibilities
Streaming ETL Pipeline Development
- Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream
API)
- Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and
PostgreSQL CDC
- Implement transformation logic including joins, enrichment, normalization, schema mapping,
and event-time semantics
- Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal
services
Modernization Away from dbt (Data Build Tool) & Snowflake
- Assist in migration of analytical and transformation workloads from MongoDB, dbt (Data Build
Tool), and Snowflake to Flink/PostgreSQL
- Translate existing dbt models, DAGs, and downstream dependencies into equivalent Flink
transformations or SQL-based operations
- Support design of new schema models, data quality checks, and lineage patterns to replace
dbt/Snowflake functionality
- Participate in validation and regression testing to ensure fidelity during migration
Operational Support
- Deploy and update Flink jobs on containerized cloud-based platforms (ECS)
- Handle routine operational tasks including job configuration updates, savepoint management,
and monitoring
- Troubleshoot failing jobs and ensure pipeline reliability
Collaboration & Documentation
- Work closely with platform engineering teams to define requirements for migration and pipeline
development
- Produce clear documentation for data flows, schemas, and newly created Flink jobs
Requirements
- 4+ years of experience in data engineering or streaming pipeline development
- Practical experience with Apache Flink DataStream API
- Hands-on experience with Apache Kafka as a CDC or event streaming backbone
- Familiarity with Debezium-based connectors or similar CDC frameworks
- Solid understanding of SQL, CTEs, relational modeling, and PostgreSQL
- Experience working with dbt and Snowflake, with ability to analyze models and transformation
logic
- Demonstrated ability to develop software in Java for pipeline development
- Familiarity with cloud environments (AWS preferred)
- Experience refactoring or replacing dbt models with streaming transformations
- Exposure to schema evolution strategies and metadata management
- Some experience with CI/CD workflows for data pipelines
- Familiarity with containerization concepts (Docker, ECS/EKS, or Kubernetes)
Core Competencies
- Detail-oriented with strong problem-solving and troubleshooting skills
- Strong desire for operational excellence and continuous improvement
- Ability to translate analytical requirements into scalable data flows
- Comfortable working within established infrastructure and collaborating with senior engineers
on larger changes
- Effective written and verbal communication skills in English, especially for documentation and
collaboration
Our Values
- Be Committed: Work hard, own the problem, keep learning
- Create Trust: Deliver on promises, build trust with clients and colleagues
- Be Bold: Experiment, speak up
- Deliver High Performance: Prepare thoroughly, commit fully, and deliver results
Job Types: Full-time, Permanent
Pay: ₹800,000.00 - ₹1,200,000.00 per year
Benefits:
- Health insurance
- Paid sick time
- Provident Fund
Ability to commute/relocate:
- Bengaluru, Karnataka: Reliably commute or planning to relocate before starting work (Required)
Application Question(s):
- Skills Required: Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream
API)
- Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and
PostgreSQL CDC
- Implement transformation logic including joins, enrichment, normalization, schema mapping,
and event-time semantics
- Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal
services
Experience:
- Data Streaming Engineer : 4 years (Required)
Work Location: In person