Are you passionate about unlocking the power of data to drive innovation and transform business outcomes? Join our cutting-edge Data Engineering team and be a key player in delivering scalable, secure, and high-performing data solutions across the enterprise. As a
Data Engineer
, you will play a central role in designing and developing modern data pipelines and platforms that support data-driven decision-making and AI-powered products. With a focus on
Python
,
SQL
,
AWS
,
PySpark
, and
Databricks
, you'll enable the transformation of raw data into valuable insights by applying engineering best practices in a cloud-first environment.
We are looking for a highly motivated professional who can work across teams to build and manage robust, efficient, and secure data ecosystems that support both analytical and operational workloads.
Accountabilities:
Design, build, and optimize scalable data pipelines using
PySpark
,
Databricks
, and
SQL
on
AWS cloud platforms
.
Collaborate with data analysts, data scientists, and business users to understand data requirements and ensure reliable, high-quality data delivery.
Implement batch and streaming data ingestion frameworks from a variety of sources (structured, semi-structured, and unstructured data).
Develop reusable, parameterized ETL/ELT components and data ingestion frameworks.
Perform data transformation, cleansing, validation, and enrichment using
Python
and
PySpark
.
Build and maintain data models, data marts, and logical/physical data structures that support BI, analytics, and AI initiatives.
Apply best practices in software engineering, version control (Git), code reviews, and agile development processes.
Ensure data pipelines are well-tested, monitored, and robust with proper logging and alerting mechanisms.
Optimize performance of distributed data processing workflows and large datasets.
Leverage AWS services (such as S3, Glue, Lambda, EMR, Redshift, Athena) for data orchestration and lakehouse architecture design.
Participate in data governance practices and ensure compliance with data privacy, security, and quality standards.
Contribute to documentation of processes, workflows, metadata, and lineage using tools such as
Data Catalogs
or
Collibra
(if applicable).
Drive continuous improvement in engineering practices, tools, and automation to increase productivity and delivery quality.
Essential Skills / Experience:
4 to 6 years
of professional experience in
Data Engineering
or a related field.
Strong programming experience with
Python
and experience using Python for data wrangling, pipeline automation, and scripting.
Deep expertise in writing complex and optimized
SQL
queries on large-scale datasets.
Solid hands-on experience with
PySpark
and distributed data processing frameworks.
Expertise working with
Databricks
for developing and orchestrating data pipelines.
Experience with
AWS cloud
services such as
S3
,
Glue
,
EMR
,
Athena
,
Redshift
, and
Lambda
.
Practical understanding of ETL/ELT development patterns and data modeling principles (Star/Snowflake schemas).
Experience with job orchestration tools like
Airflow
,
Databricks Jobs
, or
AWS Step Functions
.
Understanding of data lake, lakehouse, and data warehouse architectures.
Familiarity with DevOps and CI/CD tools for code deployment (e.g., Git, Jenkins, GitHub Actions).
Strong troubleshooting and performance optimization skills in large-scale data processing environments.
Excellent communication and collaboration skills, with the ability to work in cross-functional agile teams.
Desirable Skills / Experience:
AWS or Databricks certifications (e.g., AWS Certified Data Analytics, Databricks Data Engineer Associate/Professional).
Exposure to
data observability
,
monitoring
, and
alerting
frameworks (e.g., Monte Carlo, Datadog, CloudWatch).
Experience working in healthcare, life sciences, finance, or another regulated industry.
Familiarity with data governance and compliance standards (GDPR, HIPAA, etc.).
Knowledge of modern data architectures (Data Mesh, Data Fabric).
Exposure to streaming data tools like Kafka, Kinesis, or Spark Structured Streaming.
Experience with data visualization tools such as Power BI, Tableau, or QuickSight.
Work Environment & Collaboration:
We value a hybrid, collaborative environment that encourages shared learning and innovation. You will work closely with product owners, architects, analysts, and data scientists across geographies to solve real-world business problems using cutting-edge technologies and methodologies. We encourage flexibility while maintaining a strong in-office presence for better team synergy and innovation.
About Agilisium -
Agilisium, is an AWS technology Advanced Consulting Partner that enables companies to accelerate their "Data-to-Insights-Leap.
With $25+ million in annual revenue and over 40% year-over-year growth, Agilisium is one of the fastest-growing IT solution providers in Southern California.
Our most important asset? People.
Talent management plays a vital role in our business strategy.
We're looking for "drivers"; big thinkers with growth and strategic mindset -- people who are committed to customer obsession, aren't afraid to experiment with new ideas.
And we are all about finding and nurturing individuals who are ready to do great work.
* At Agilisium, you'll collaborate with great minds while being challenged to meet and exceed your potential
Beware of fraud agents! do not pay money to get a job
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.