We are looking for a skilled PySpark Developer with hands on experience in Reltio MDM to join our data engineering team The ideal candidate will be responsible for designing and implementing scalable data processing solutions using PySpark and integrating with Reltiocloud native MDM platform
Key Responsibilities
Develop and maintain data pipelines using PySpark in distributed computing environments eg AWS EMR Databricks
Integrate and synchronize data between enterprise systems and the Reltio MDM platform
Design and implement data transformation cleansing and enrichment processes
Collaborate with data architects business analysts and Reltio solution architects to ensure high quality data modeling
Work on API based integration between Reltio and upstream downstream applications
Optimize PySpark jobs for performance and cost efficiency
Ensure data quality integrity and governance throughout the pipeline
Troubleshoot and resolve data and performance issues in existing workflows
Required Skills Qualifications
5 to 7 years of experience in PySpark development and distributed data processing
Strong understanding of Apache Spark DataFrames and Spark SQL
Experience with Reltio MDM including entity modeling survivorship rules match merge configuration
Proficiency in working with REST APIs and JSON data formats
Experience with cloud platforms like AWS and data services eg S3 Lambda step function
Good knowledge of data warehousing concepts ETL workflows and data modeling
Familiarity with CI CD practices and version control tools like Git
Strong problem solving and communication skills
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.