. The ideal candidate will design, develop, and optimize large-scale
ETL/ELT pipelines
, integrate diverse data sources, and apply
Machine Learning frameworks
for data quality and anomaly detection. You will collaborate closely with analysts, data scientists, and business stakeholders to deliver scalable, reliable, and actionable data-driven solutions.
This is a critical role in ensuring
data integrity, scalability, and performance
of our enterprise data platforms across cloud environments such as
Snowflake, AWS, and GCP
.
Key Responsibilities
Design, architect, and implement
robust, high-performance ETL/ELT pipelines
for large-scale data processing.
Collect and integrate data from multiple sources (databases, APIs, web services) into data lakes or warehouses.
Develop and maintain scalable
data processing solutions
using
Snowflake, AWS, and GCP
platforms.
Implement
data quality, anomaly detection, and monitoring systems
using Machine Learning frameworks.
Optimize query performance and manage large-scale data models, schemas, and metadata.
Support
data governance
through documentation, access control, lineage tracking, and retention policies.
Collaborate with cross-functional teams (data scientists, analysts, engineers, and business units) to translate requirements into actionable technical solutions.
Participate in
CI/CD and automation initiatives
to streamline data workflows and improve efficiency.
Ensure
data accuracy, reliability, and availability
for analytical and reporting purposes.
Qualifications
Bachelor's degree
in Computer Science, Information Technology, or a related field.
7+ years
of experience in
data warehouse architecture, ETL/ELT pipelines, and data engineering
.
5+ years
of experience in
data modeling
and writing
advanced SQL
queries across
Snowflake, Oracle
, and
Columnar Databases
.
Strong knowledge of
data modeling, schema design, and data lakes
.
Experience with
AWS services
such as
S3, Lambda, and Data Pipeline
.
Hands-on experience with
Machine Learning algorithms
for
data quality and anomaly detection
.
Technical Skills
Programming:
Python, SQL
ETL Tools:
Databricks, IBM DataStage (preferred)
Cloud Platforms:
Snowflake, AWS, GCP, BigQuery
Data Processing Frameworks:
Apache Spark (or equivalent big data tools)
Workflow Orchestration:
Apache Airflow, AWS MWAA
DevOps Tools:
CI/CD pipelines, automation frameworks
Data Architecture:
Data lakes, warehouses, multi-cluster architecture, shareable data features
Bonus:
Familiarity with Databricks, DataStage, and advanced Snowflake capabilities
Job Types: Part-time, Freelance
Contract length: 12 months
Pay: ₹30,000.00 - ₹40,000.00 per month
Benefits:
Work from home
Experience:
total work: 6 years (Required)
Shift availability:
Night Shift (Required)
Overnight Shift (Required)
Work Location: Remote
Beware of fraud agents! do not pay money to get a job
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.