We are seeking a skilled Data Streaming Engineer to join our engineering team ((US Client Project). The successful
candidate will play a critical role in developing and maintaining real-time data pipelines, modernizing
analytical workloads, and supporting operational excellence in cloud-based environments.
This role requires a self-starter with a proven track record of owning issues, finding creative solutions to
complex problems, and driving continual improvement initiatives. Strong organizational skills and a
thirst for knowledge are essential.
Responsibilities
Streaming ETL Pipeline Development
Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream
API)
Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and
PostgreSQL CDC
Implement transformation logic including joins, enrichment, normalization, schema mapping,
and event-time semantics
Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal
services
Modernization Away from dbt (Data Build Tool) & Snowflake
Assist in migration of analytical and transformation workloads from MongoDB, dbt (Data Build
Tool), and Snowflake to Flink/PostgreSQL
Translate existing dbt models, DAGs, and downstream dependencies into equivalent Flink
transformations or SQL-based operations
Support design of new schema models, data quality checks, and lineage patterns to replace
dbt/Snowflake functionality
Participate in validation and regression testing to ensure fidelity during migration
Operational Support
Deploy and update Flink jobs on containerized cloud-based platforms (ECS)
Handle routine operational tasks including job configuration updates, savepoint management,
and monitoring
Troubleshoot failing jobs and ensure pipeline reliability
Collaboration & Documentation
Work closely with platform engineering teams to define requirements for migration and pipeline
development
Produce clear documentation for data flows, schemas, and newly created Flink jobs
Requirements
4+ years of experience in data engineering or streaming pipeline development
Practical experience with Apache Flink DataStream API
Hands-on experience with Apache Kafka as a CDC or event streaming backbone
Familiarity with Debezium-based connectors or similar CDC frameworks
Solid understanding of SQL, CTEs, relational modeling, and PostgreSQL
Experience working with dbt and Snowflake, with ability to analyze models and transformation
logic
Demonstrated ability to develop software in Java for pipeline development
Familiarity with cloud environments (AWS preferred)
Experience refactoring or replacing dbt models with streaming transformations
Exposure to schema evolution strategies and metadata management
Some experience with CI/CD workflows for data pipelines
Familiarity with containerization concepts (Docker, ECS/EKS, or Kubernetes)
Core Competencies
Detail-oriented with strong problem-solving and troubleshooting skills
Strong desire for operational excellence and continuous improvement
Ability to translate analytical requirements into scalable data flows
Comfortable working within established infrastructure and collaborating with senior engineers
on larger changes
Effective written and verbal communication skills in English, especially for documentation and
collaboration
Our Values
Be Committed: Work hard, own the problem, keep learning
Create Trust: Deliver on promises, build trust with clients and colleagues
Be Bold: Experiment, speak up
Deliver High Performance: Prepare thoroughly, commit fully, and deliver results
Job Types: Full-time, Permanent
Pay: ?800,000.00 - ?1,200,000.00 per year
Benefits:
Health insurance
Paid sick time
Provident Fund
Ability to commute/relocate:
Bengaluru, Karnataka: Reliably commute or planning to relocate before starting work (Required)
Application Question(s):
Skills Required: Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream
API)
Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and
PostgreSQL CDC
Implement transformation logic including joins, enrichment, normalization, schema mapping,
and event-time semantics
Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal
services
Experience:
Data Streaming Engineer : 4 years (Required)
Work Location: In person
Beware of fraud agents! do not pay money to get a job
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.