Flag job

Report

LLM Ops Engineer

Min Experience

3 years

Location

Gurgaon, Haryana, India

JobType

full-time

About the job

Info This job is sourced from a job board

About the role

Company Description:

WNS, part of Capgemini, is an Agentic AI-powered leader in intelligent operations and transformation, serving more than 700 clients across 10 industries, including Banking and Financial Services, Healthcare, Insurance, Shipping and Logistics, and Travel and Hospitality. We bring together deep domain excellence – WNS’ core differentiator – with AI-powered platforms and analytics to help businesses innovate, scale, adapt and build resilience in a world defined by disruption.

Our purpose is clear: to enable lasting business value by designing intelligent, human-led solutions that deliver sustainable outcomes and a differentiated impact. With three global headquarters across four continents, operations in 13 countries, 65 delivery centers and more than 66,000 employees, WNS combines scale, expertise and execution to create meaningful, measurable impact.

Login into: https://www.wns.com/ to know more!

Job Description:

We are hiring an LLM Ops Engineer to join our AI Research team, a highly technical group working on cutting-edge advancements in the AI industry. The team focuses on building scalable, production-grade LLM systems, fine-tuning strategies, evaluation frameworks, and next-generation deployment architectures.

This role requires hands-on experience operating LLMs beyond simple API integration. The ideal candidate understands the architectural, operational, and evaluation complexities that differentiate LLMOps from traditional MLOps.

Responsibilities:- 

  • Manage the end-to-end lifecycle of LLMs: registry, packaging, versioning, deployment, monitoring, and rollback.
  • Deploy and operate self-hosted / open-source LLMs (not limited to OpenAI API usage).
  • Design and manage scalable inference infrastructure, including GPU-aware deployments.
  • Implement CI/CD pipelines for LLM deployment and continuous evaluation.
  • Monitor system performance including latency, throughput, token usage, cost, drift (model and data), and hallucinations.
  • Ensure secure, compliant, and resilient cloud-based model deployments.
  • Collaborate with research and engineering for deployments.

Skills:-

  • Strong hands-on experience with LLM handling, hosting, and operationalization.
  • Clear understanding of how LLMOps differs from traditional MLOps (prompt management, non-deterministic outputs, semantic evaluation, token economics, guardrails etc.).
  • Experience with Kubernetes, Docker, and containerized deployments.
  • Cloud expertise (AWS / Azure / GCP) including compute, storage, IAM, networking, and monitoring.
  • Experience building scalable inference and model-serving architectures.
  • Familiarity with tools such as MLflow, Kubeflow etc. (good to have).
  • Understanding of vector databases, RAG systems, and evaluation frameworks (preferred).
  • Knowledge of GenAI security considerations (prompt injection, data leakage prevention).
Qualifications:
  • Bachelor’s degree in Computer Science, Engineering, or related field.
  • DevOps certification (e.g., AWS DevOps Engineer, Azure DevOps, or equivalent).
  • 3–5 years of experience in MLOps, LLMOps, ML Engineering, or related roles.
  • Bachelor’s or master’s degree in computer science, Artificial Intelligence, Data Science, or a related technical field.
  • Demonstrated experience deploying ML/LLM systems in production environments.

About the company

Provides global business process management and digital transformation services.

Skills

Kubernetes
Docker
AWS
Azure
GCP
MLflow
Kubeflow
GPU