Our organization is developing a scalable and resilient data platform to power customer-facing web, mobile, reporting, and data exchange (inbound/outbound) capabilities. Currently, we leverage an enterprise data warehouse--ODP (One Data Platform)--managed by the central data team (CDO organization), with data ingested on a daily and hourly basis.
To enable near-real-time visibility and support operational decision-making, we aim to integrate with external systems such as Siebel CRM, Nuvolo CRM, and ServiceMax (Salesforce). We are seeking an experienced Data Architect to lead the design, validation, and implementation of robust streaming and batch data pipelines, business logic layers, and data stores--ensuring high standards of data quality, compliance, and governance.
Key Responsibilities:
Lead architecture and solution design for batch, near real-time data pipelines integrating heterogeneous data sources.
Define data ingestion, transformation, modeling, and serving strategies using native AWS services (e.g., Lambda, Kinesis, DMS, Glue, Redshift etc..) or any better tools.
Evaluate and integrate with external systems:
Siebel CRM: direct DB connection with delta update via timestamp
Nuvolo CRM: webhook event ingestion
Service Max: Salesforce streaming API
Customer SFTP: file-based ingestion
Design scalable frameworks for:
Delta detection and state change tracking
Business logic execution (state derivation, rule engine integration if needed)
Schema/version management
Regional data partitioning (US, EMEA, APAC)
Collaborate with internal application and data engineering teams to ensure alignment with downstream consumers (Solr, Power BI, SFTP exports)
Establish monitoring, alerting, retries, and error handling frameworks for data pipelines.
Support model changes and schema evolution processes.
Required Skills & Experience
Skill Area Expertise
Cloud Platform :
Advanced proficiency in AWS ecosystem (Glue, Lambda, Kinesis/MSK, Step Functions, Redshift, Aurora, S3, Lake Formation)
Data Integration:
Experience with multi-source data integration (DB, webhook, API, SFTP), near real-time and batch
ETL/ELT Design
:Hands-on experience designing data lakes, data lakehouses, streaming ETL, CDC pipelines
Data Modeling
:Expertise in dimensional modeling (Star/Snowflake)
Architecture
:Designing modular, scalable, and region-aware architecture with multi-tenant or regional separation considerations
Security & Governance
:Familiarity with data access controls, encryption, IAM policies
Monitoring :
CloudWatch, X-Ray, or similar tools for pipeline observability
Tools:
Experience with Airflow, Airbyte, dbt, Great Expectations, or similar tools is a plus
Preferred Qualifications
10+ years of experience in Data Engineering / Architecture roles
Prior experience building near-real-time streaming architectures
Healthcare or enterprise data compliance experience (HIPAA, GDPR) is a plus
Strong communication skills to collaborate with internal engineering and BI teams
Engagement Deliverables
End-to-end data pipeline architecture document with source integration
Reference implementation of key pipelines (one per source system)
Schema design and data catalog documentation
Data lineage and change management strategy
Knowledge transfer and handover to internal team
Interested candidates , Kindly reach out to hr@icoresoftwaresystems.com or 863 742 5983 Jesilina HR
Job Type: Full-time
Pay: ?1,000,000.00 - ?1,500,000.00 per year
Work Location: In person
Beware of fraud agents! do not pay money to get a job
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.