Job Decsription:
Design, build, and maintain end-to-end ML pipelines for experiment tracking, feature management, model versioning, and retraining.
Develop and deploy scalable APIs for model inference and AI services.
Implement and optimize LLM training, fine-tuning, and inference workflows.
Work extensively on distributed training and serving of large-scale models using GPU architectures.
Apply model and data parallelism using frameworks such as DeepSpeed and vLLM.
Optimize models to improve latency, accuracy, and cost efficiency.
Implement LLMOps and DevOps best practices for production-grade AI systems.
Containerize and orchestrate applications using Docker and Kubernetes.
Integrate and manage LLM orchestration frameworks like LangChain, LangGraph, LangFlow, Flowise, and LlamaIndex.
Monitor and maintain AI systems using logging and observability tools.
MNCJobsIndia.com will not be responsible for any payment made to a third-party. All Terms of Use are applicable.