Lead AI Engineer

November 18, 2025

Job Description

YOUR IMPACT

We are seeking a highly skilled AI Systems Engineer to lead the design, development, and optimization of Retrieval-Augmented Generation (RAG) pipelines and multi-agent AI workflows within enterprise-scale environments.The role requires deep technical expertise across LLM orchestrationcontext engineering, and production-grade deployment practices. You will work cross-functionally with data, platform, and product teams to build scalable, reliable, and context-aware AI systems that power next-generation enterprise intelligence solutions.

What The Role Offers

  • Be part of an enterprise AI transformation team shaping the future of LLM-driven applications.
  • Work with cutting-edge technologies in AI orchestration, RAG, and multi-agent systems.
  • Opportunity to architect scalable, secure, and context-aware AI systems deployed across global enterprise environments.
  • Collaborative environment fostering continuous learning and innovation in Generative AI systems engineering.
  • Architect, implement, and optimize enterprise-grade RAG pipelines covering data ingestion, embedding creation, and vector-based retrieval.
  • Design, build, and orchestratemulti-agent workflows using frameworks such as LangGraphCrew AI, or AI Development Kit (ADK) for collaborative task automation.
  • Engineer prompts and contextual templates to enhance LLM performance, accuracy, and domain adaptability.
  • Integrate and manage vector databases (pgvector, Milvus, Weaviate, Pinecone) for semantic search and hybrid retrieval.
  • Develop and maintain data pipelines for structured and unstructured data using SQL and NoSQL systems.
  • Expose RAG workflows through APIs using FastAPI or Flask, ensuring high reliability and performance.
  • Containerize, deploy, and scale AI microservices using DockerKubernetes, and Helm within enterprise-grade environments.
  • Implement CI/CD automation pipelines via GitLab or similar tools to streamline builds, testing, and deployments.
  • Collaborate with cross-functional teams (Data, ML, DevOps, Product) to integrate retrieval, reasoning, and generation into end-to-end enterprise systems.
  • Monitor and enhance AI system observability using PrometheusGrafana, and OpenTelemetry for real-time performance and reliability tracking.
  • Integrate LLMs with enterprise data sources and knowledge graphs to deliver contextually rich, domain-specific outputs.

What You Need To Succeed

  • Education: Bachelors or Masters degree in Computer ScienceArtificial Intelligence, or related technical discipline.
  • Experience: 5 – 10 years in AI/ML system developmentdeployment, and optimization within enterprise or large-scale environments.
  • Deep understanding of Retrieval-Augmented Generation (RAG) architecture and hybrid retrieval mechanisms.
  • Proficiency in Python with hands-on expertise in FastAPIFlask, and REST API design.
  • Strong experience with vector databases (pgvector, Milvus, Weaviate, Pinecone).
  • Proficiency in prompt engineering and context engineering for LLMs.
  • Hands-on experience with containerization (Docker) and orchestration (Kubernetes, Helm) in production-grade deployments.
  • Experience with CI/CD automation using GitLabJenkins, or equivalent tools.
  • Familiarity with LangChainLangGraphGoogle ADK, or similar frameworks for LLM-based orchestration.
  • Knowledge of AI observabilitylogging, and reliability engineering principles.
  • Understanding of enterprise data governancesecurity, and scalability in AI systems.
  • Proven track record of building and maintaining production-grade AI applications with measurable business impact.
  • Experience in fine-tuning or parameter-efficient tuning (PEFT/LoRA) of open-source LLMs.
  • Familiarity with open-source model hostingLLM governance frameworks, and model evaluation practices.
  • Knowledge of multi-agent system design and Agent-to-Agent (A2A) communication frameworks.
  • Exposure to LLMOps platforms such as LangSmithWeights & Biases, or Kubeflow.
  • Experience with cloud-based AI infrastructure (AWS Sagemaker, Azure OpenAI, GCP Vertex AI).
  • Working understanding of distributed systemsAPI gateway management, and service mesh architectures.
  • Strong analytical and problem-solving mindset with attention to detail.
  • Effective communicator with the ability to collaborate across technical and business teams.
  • Self-motivated, proactive, and capable of driving end-to-end ownership of AI system delivery.
  • Passion for innovation in LLM orchestrationretrieval systems, and enterprise AI solutions.

Role: 

Search EngineerIndustry Type: 

Software ProductDepartment: 

Engineering – Software & QAEmployment Type: 

Full Time, PermanentRole Category: 

Software Development

EducationUG: 

Any GraduatePG: 

LLM in Law, Any Postgraduate

Key Skills

Skills highlighted with ‘‘ are preferred keyskills

nosql

AiPrometheusLlmDevopsGrafanaMicroservicesModel EvaluationRestDockerAws SagemakerAwsData GovernancePythonFlaskApi GatewayAzureArchitectureArtificial IntelligenceHrSqlJenkinsGcpGitlabKubernetes