Job Description

We are seeking an experienced Senior Data Engineer with strong hands-on expertise in Databricks and AWS to design, develop, and optimize scalable data pipelines and modern data platforms. The ideal candidate will have deep experience with big data ecosystems, cloud-based data solutions (AWS), and ETL/ELT frameworks using Python, PySpark, and SQL.


Key Responsibilities:


Design, build, and maintain data pipelines and ETL workflows using Databricks (PySpark, Spark SQL) and AWS services (S3, EMR, Glue, Lambda, RDS, DynamoDB).


Implement data lakehouse architectures integrating Delta Lake, Snowflake, and AWS storage layers.


Develop and deploy automated and scalable data ingestion frameworks to process structured and unstructured data.


Build and optimize data models (Star/Snowflake schemas) for analytical and reporting use cases.


Orchestrate workflows using Airflow or similar scheduling tools; manage job dependencies and monitoring.


Collaborate with analytics, data science, and business teams to deliver high-quality, reliable, and well-documented data assets.


Ensure data quality, governance, and security best practices are maintained throughout pipelines.


Troubleshoot and tune performance for Spark clusters and SQL queries to optimize cost and efficiency.


Integrate with modern tools such as DBT, Kubernetes, and CI/CD pipelines for continuous delivery of data solutions.


Required Skills & Experience:


10+ years of experience in data engineering, with proven work on large-scale distributed data systems.


Strong hands-on experience with Databricks (PySpark, Spark SQL, Delta Lake).


Expertise in AWS Cloud Services: EMR, S3, Glue, Lambda, RDS, EC2, CloudWatch, IAM, etc.


Proficiency in Python and SQL for data transformation and automation.


Experience with Snowflake (SnowSQL, Snowpipe, schema design, and optimization).


Solid understanding of data warehousing concepts (ETL/ELT, dimensional modeling, data partitioning).


Experience with Airflow for scheduling and monitoring data pipelines.


Familiarity with CI/CD tools (Jenkins, Git, Docker, Kubernetes) for automated deployments.


Working knowledge of NoSQL databases (MongoDB, DynamoDB, Cassandra).


Excellent debugging, problem-solving, and performance-tuning skills.


.

About Virtusa




Teamwork, quality of life, professional and personal development: values that Virtusa is proud to embody. When you join us, you join a team of 27,000 people globally that cares about your growth -- one that seeks to provide you with exciting projects, opportunities and work with state of the art technologies throughout your career with us.


Great minds, great potential: it all comes together at Virtusa. We value collaboration and the team environment of our company, and seek to provide great minds with a dynamic place to nurture new ideas and foster excellence.


Virtusa was founded on principles of equal opportunity for all, and so does not discriminate on the basis of race, religion, color, sex, gender identity, sexual orientation, age, non-disqualifying physical or mental disability, national origin, veteran status or any other basis covered by appropriate law. All employment is decided on the basis of qualifications, merit, and business need.

Beware of fraud agents! do not pay money to get a job

MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.


Job Detail

  • Job Id
    JD5094644
  • Industry
    Not mentioned
  • Total Positions
    1
  • Job Type:
    Full Time
  • Salary:
    Not mentioned
  • Employment Status
    Permanent
  • Job Location
    AP, IN, India
  • Education
    Not mentioned
  • Experience
    Year