Senior Data Architect

Year    KA, IN, India

Job Description

Project Background:

Our organization is developing a scalable and resilient data platform to power customer-facing web, mobile, reporting, and data exchange (inbound/outbound) capabilities. Currently, we leverage an enterprise data warehouse--ODP (One Data Platform)--managed by the central data team (CDO organization), with data ingested on a daily and hourly basis.

To enable near-real-time visibility and support operational decision-making, we aim to integrate with external systems such as Siebel CRM, Nuvolo CRM, and ServiceMax (Salesforce). We are seeking an experienced Data Architect to lead the design, validation, and implementation of robust streaming and batch data pipelines, business logic layers, and data stores--ensuring high standards of data quality, compliance, and governance.

Key Responsibilities:



Lead architecture and solution design for batch, near real-time data pipelines integrating heterogeneous data sources. Define data ingestion, transformation, modeling, and serving strategies using native AWS services (e.g., Lambda, Kinesis, DMS, Glue, Redshift etc..) or any better tools. Evaluate and integrate with external systems: Siebel CRM: direct DB connection with delta update via timestamp Nuvolo CRM: webhook event ingestion Service Max: Salesforce streaming API Customer SFTP: file-based ingestion Design scalable frameworks for: Delta detection and state change tracking Business logic execution (state derivation, rule engine integration if needed) Schema/version management Regional data partitioning (US, EMEA, APAC) Collaborate with internal application and data engineering teams to ensure alignment with downstream consumers (Solr, Power BI, SFTP exports) Establish monitoring, alerting, retries, and error handling frameworks for data pipelines. Support model changes and schema evolution processes.

Required Skills & Experience



Skill Area Expertise


Cloud Platform :

Advanced proficiency in AWS ecosystem (Glue, Lambda, Kinesis/MSK, Step Functions, Redshift, Aurora, S3, Lake Formation)

Data Integration:

Experience with multi-source data integration (DB, webhook, API, SFTP), near real-time and batch

ETL/ELT Design

:Hands-on experience designing data lakes, data lakehouses, streaming ETL, CDC pipelines

Data Modeling

:Expertise in dimensional modeling (Star/Snowflake)

Architecture

:Designing modular, scalable, and region-aware architecture with multi-tenant or regional separation considerations

Security & Governance

:Familiarity with data access controls, encryption, IAM policies

Monitoring :

CloudWatch, X-Ray, or similar tools for pipeline observability

Tools:

Experience with Airflow, Airbyte, dbt, Great Expectations, or similar tools is a plus

Preferred Qualifications



10+ years of experience in Data Engineering / Architecture roles Prior experience building near-real-time streaming architectures Healthcare or enterprise data compliance experience (HIPAA, GDPR) is a plus Strong communication skills to collaborate with internal engineering and BI teams

Engagement Deliverables



End-to-end data pipeline architecture document with source integration Reference implementation of key pipelines (one per source system) Schema design and data catalog documentation Data lineage and change management strategy Knowledge transfer and handover to internal team

Interested candidates , Kindly reach out to hr@icoresoftwaresystems.com or 863 742 5983 Jesilina HR



Job Type: Full-time

Pay: ?1,000,000.00 - ?1,500,000.00 per year

Work Location: In person

Beware of fraud agents! do not pay money to get a job

MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.


Job Detail

  • Job Id
    JD4018520
  • Industry
    Not mentioned
  • Total Positions
    1
  • Job Type:
    Full Time
  • Salary:
    Not mentioned
  • Employment Status
    Permanent
  • Job Location
    KA, IN, India
  • Education
    Not mentioned
  • Experience
    Year