Overview:
Join the Prodapt team in building a unified, cloud-native environment for scalable machine learning inferencing. You will help design, develop, and optimize robust workflows that empower data scientists and engineers to efficiently deploy, serve, and monitor ML models at scale, supporting both real-time and batch inference use cases.
Responsibilities:
Develop, maintain, and enhance model deployment workflows using Seldon, Docker, and Kubernetes for scalable inferencing.
Build and optimize REST and gRPC endpoints for serving ML models, ensuring secure and reliable access from internal tools and pipelines.
Integrate with AI Hub for unified model registration, endpoint management, and monitoring.
Support both online (real-time) and offline (batch) inference, leveraging batch inference capabilities.
Manage container images and model artifacts using Docker Hub, Artifact Registry, and Google Cloud Storage.
Implement and maintain CI/CD pipelines for automated model deployment and endpoint promotion.
Ensure robust security, compliance, and governance, including role-based access control and audit logging.
Collaborate with data scientists, ML engineers, and platform teams to deliver production-grade inferencing solutions.
Participate in code reviews, architecture discussions, and continuous improvement of the inferencing platform.
Requirements:
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.