Data Engineer

Year    KA, IN, India

Job Description

Position Overview



We are seeking a skilled Data Streaming Engineer to join our engineering team ((US Client Project). The successful

candidate will play a critical role in developing and maintaining real-time data pipelines, modernizing

analytical workloads, and supporting operational excellence in cloud-based environments.

This role requires a self-starter with a proven track record of owning issues, finding creative solutions to

complex problems, and driving continual improvement initiatives. Strong organizational skills and a

thirst for knowledge are essential.

Responsibilities

Streaming ETL Pipeline Development

Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream
API)

Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and
PostgreSQL CDC

Implement transformation logic including joins, enrichment, normalization, schema mapping,
and event-time semantics

Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal
services

Modernization Away from dbt (Data Build Tool) & Snowflake

Assist in migration of analytical and transformation workloads from MongoDB, dbt (Data Build
Tool), and Snowflake to Flink/PostgreSQL

Translate existing dbt models, DAGs, and downstream dependencies into equivalent Flink
transformations or SQL-based operations

Support design of new schema models, data quality checks, and lineage patterns to replace
dbt/Snowflake functionality

Participate in validation and regression testing to ensure fidelity during migration
Operational Support

Deploy and update Flink jobs on containerized cloud-based platforms (ECS) Handle routine operational tasks including job configuration updates, savepoint management,
and monitoring

Troubleshoot failing jobs and ensure pipeline reliability
Collaboration & Documentation

Work closely with platform engineering teams to define requirements for migration and pipeline
development

Produce clear documentation for data flows, schemas, and newly created Flink jobs

Requirements



4+ years of experience in data engineering or streaming pipeline development Practical experience with Apache Flink DataStream API Hands-on experience with Apache Kafka as a CDC or event streaming backbone Familiarity with Debezium-based connectors or similar CDC frameworks Solid understanding of SQL, CTEs, relational modeling, and PostgreSQL Experience working with dbt and Snowflake, with ability to analyze models and transformation
logic

Demonstrated ability to develop software in Java for pipeline development Familiarity with cloud environments (AWS preferred) Experience refactoring or replacing dbt models with streaming transformations Exposure to schema evolution strategies and metadata management Some experience with CI/CD workflows for data pipelines Familiarity with containerization concepts (Docker, ECS/EKS, or Kubernetes)

Core Competencies



Detail-oriented with strong problem-solving and troubleshooting skills Strong desire for operational excellence and continuous improvement Ability to translate analytical requirements into scalable data flows Comfortable working within established infrastructure and collaborating with senior engineers
on larger changes

Effective written and verbal communication skills in English, especially for documentation and
collaboration

Our Values



Be Committed: Work hard, own the problem, keep learning Create Trust: Deliver on promises, build trust with clients and colleagues Be Bold: Experiment, speak up Deliver High Performance: Prepare thoroughly, commit fully, and deliver results
Job Types: Full-time, Permanent

Pay: ?800,000.00 - ?1,200,000.00 per year

Benefits:

Health insurance Paid sick time Provident Fund
Ability to commute/relocate:

Bengaluru, Karnataka: Reliably commute or planning to relocate before starting work (Required)
Application Question(s):

Skills Required: Develop and maintain ETL and real-time processing pipelines using Apache Flink (DataStream
API)

Build and extend CDC ingestion pipelines using Debezium, Apache Kafka, MongoDB CDC, and
PostgreSQL CDC

Implement transformation logic including joins, enrichment, normalization, schema mapping,
and event-time semantics

Integrate Flink jobs with downstream systems such as PostgreSQL, object storage, and internal
services

Experience:

Data Streaming Engineer : 4 years (Required)
Work Location: In person

Beware of fraud agents! do not pay money to get a job

MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.


Job Detail

  • Job Id
    JD5067129
  • Industry
    Not mentioned
  • Total Positions
    1
  • Job Type:
    Full Time
  • Salary:
    Not mentioned
  • Employment Status
    Permanent
  • Job Location
    KA, IN, India
  • Education
    Not mentioned
  • Experience
    Year