Etl Engineer

Year    Remote, IN, India

Job Description

Freelancing Opportunity



Job Title:

ETL Engineer/ML

Experience:

7+ Years

Job Overview



We are seeking a

highly skilled Machine Learning / ETL Engineer

to join our

Business Intelligence IT team

. The ideal candidate will design, develop, and optimize large-scale

ETL/ELT pipelines

, integrate diverse data sources, and apply

Machine Learning frameworks

for data quality and anomaly detection. You will collaborate closely with analysts, data scientists, and business stakeholders to deliver scalable, reliable, and actionable data-driven solutions.

This is a critical role in ensuring

data integrity, scalability, and performance

of our enterprise data platforms across cloud environments such as

Snowflake, AWS, and GCP

.

Key Responsibilities



Design, architect, and implement

robust, high-performance ETL/ELT pipelines

for large-scale data processing. Collect and integrate data from multiple sources (databases, APIs, web services) into data lakes or warehouses. Develop and maintain scalable

data processing solutions

using

Snowflake, AWS, and GCP

platforms. Implement

data quality, anomaly detection, and monitoring systems

using Machine Learning frameworks. Optimize query performance and manage large-scale data models, schemas, and metadata. Support

data governance

through documentation, access control, lineage tracking, and retention policies. Collaborate with cross-functional teams (data scientists, analysts, engineers, and business units) to translate requirements into actionable technical solutions. Participate in

CI/CD and automation initiatives

to streamline data workflows and improve efficiency. Ensure

data accuracy, reliability, and availability

for analytical and reporting purposes.

Qualifications



Bachelor's degree

in Computer Science, Information Technology, or a related field.

7+ years

of experience in

data warehouse architecture, ETL/ELT pipelines, and data engineering

.

5+ years

of experience in

data modeling

and writing

advanced SQL

queries across

Snowflake, Oracle

, and

Columnar Databases

. Strong knowledge of

data modeling, schema design, and data lakes

. Experience with

AWS services

such as

S3, Lambda, and Data Pipeline

. Hands-on experience with

Machine Learning algorithms

for

data quality and anomaly detection

.

Technical Skills



Programming:

Python, SQL

ETL Tools:

Databricks, IBM DataStage (preferred)

Cloud Platforms:

Snowflake, AWS, GCP, BigQuery

Data Processing Frameworks:

Apache Spark (or equivalent big data tools)

Workflow Orchestration:

Apache Airflow, AWS MWAA

DevOps Tools:

CI/CD pipelines, automation frameworks

Data Architecture:

Data lakes, warehouses, multi-cluster architecture, shareable data features

Bonus:

Familiarity with Databricks, DataStage, and advanced Snowflake capabilities
Job Types: Part-time, Freelance
Contract length: 12 months

Pay: ₹30,000.00 - ₹40,000.00 per month

Benefits:

Work from home
Experience:

total work: 6 years (Required)
Shift availability:

Night Shift (Required) Overnight Shift (Required)
Work Location: Remote

Beware of fraud agents! do not pay money to get a job

MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.


Job Detail

  • Job Id
    JD4528546
  • Industry
    Not mentioned
  • Total Positions
    1
  • Job Type:
    Part Time
  • Salary:
    Not mentioned
  • Employment Status
    Permanent
  • Job Location
    Remote, IN, India
  • Education
    Not mentioned
  • Experience
    Year