Job Description
YOUR IMPACT
We are seeking a highly skilled AI Systems Engineer to lead the design, development, and optimization of Retrieval-Augmented Generation (RAG) pipelines and multi-agent AI workflows within enterprise-scale environments.The role requires deep technical expertise across LLM orchestration, context engineering, and production-grade deployment practices. You will work cross-functionally with data, platform, and product teams to build scalable, reliable, and context-aware AI systems that power next-generation enterprise intelligence solutions.
What The Role Offers
- Be part of an enterprise AI transformation team shaping the future of LLM-driven applications.
- Work with cutting-edge technologies in AI orchestration, RAG, and multi-agent systems.
- Opportunity to architect scalable, secure, and context-aware AI systems deployed across global enterprise environments.
- Collaborative environment fostering continuous learning and innovation in Generative AI systems engineering.
- Architect, implement, and optimize enterprise-grade RAG pipelines covering data ingestion, embedding creation, and vector-based retrieval.
- Design, build, and orchestratemulti-agent workflows using frameworks such as LangGraph, Crew AI, or AI Development Kit (ADK) for collaborative task automation.
- Engineer prompts and contextual templates to enhance LLM performance, accuracy, and domain adaptability.
- Integrate and manage vector databases (pgvector, Milvus, Weaviate, Pinecone) for semantic search and hybrid retrieval.
- Develop and maintain data pipelines for structured and unstructured data using SQL and NoSQL systems.
- Expose RAG workflows through APIs using FastAPI or Flask, ensuring high reliability and performance.
- Containerize, deploy, and scale AI microservices using Docker, Kubernetes, and Helm within enterprise-grade environments.
- Implement CI/CD automation pipelines via GitLab or similar tools to streamline builds, testing, and deployments.
- Collaborate with cross-functional teams (Data, ML, DevOps, Product) to integrate retrieval, reasoning, and generation into end-to-end enterprise systems.
- Monitor and enhance AI system observability using Prometheus, Grafana, and OpenTelemetry for real-time performance and reliability tracking.
- Integrate LLMs with enterprise data sources and knowledge graphs to deliver contextually rich, domain-specific outputs.
What You Need To Succeed
- Education: Bachelors or Masters degree in Computer Science, Artificial Intelligence, or related technical discipline.
- Experience: 5 – 10 years in AI/ML system development, deployment, and optimization within enterprise or large-scale environments.
- Deep understanding of Retrieval-Augmented Generation (RAG) architecture and hybrid retrieval mechanisms.
- Proficiency in Python with hands-on expertise in FastAPI, Flask, and REST API design.
- Strong experience with vector databases (pgvector, Milvus, Weaviate, Pinecone).
- Proficiency in prompt engineering and context engineering for LLMs.
- Hands-on experience with containerization (Docker) and orchestration (Kubernetes, Helm) in production-grade deployments.
- Experience with CI/CD automation using GitLab, Jenkins, or equivalent tools.
- Familiarity with LangChain, LangGraph, Google ADK, or similar frameworks for LLM-based orchestration.
- Knowledge of AI observability, logging, and reliability engineering principles.
- Understanding of enterprise data governance, security, and scalability in AI systems.
- Proven track record of building and maintaining production-grade AI applications with measurable business impact.
- Experience in fine-tuning or parameter-efficient tuning (PEFT/LoRA) of open-source LLMs.
- Familiarity with open-source model hosting, LLM governance frameworks, and model evaluation practices.
- Knowledge of multi-agent system design and Agent-to-Agent (A2A) communication frameworks.
- Exposure to LLMOps platforms such as LangSmith, Weights & Biases, or Kubeflow.
- Experience with cloud-based AI infrastructure (AWS Sagemaker, Azure OpenAI, GCP Vertex AI).
- Working understanding of distributed systems, API gateway management, and service mesh architectures.
- Strong analytical and problem-solving mindset with attention to detail.
- Effective communicator with the ability to collaborate across technical and business teams.
- Self-motivated, proactive, and capable of driving end-to-end ownership of AI system delivery.
- Passion for innovation in LLM orchestration, retrieval systems, and enterprise AI solutions.
Role:
Search EngineerIndustry Type:
Software ProductDepartment:
Engineering – Software & QAEmployment Type:
Full Time, PermanentRole Category:
Software Development
EducationUG:
Any GraduatePG:
LLM in Law, Any Postgraduate
Key Skills
Skills highlighted with ‘‘ are preferred keyskills
AiPrometheusLlmDevopsGrafanaMicroservicesModel EvaluationRestDockerAws SagemakerAwsData GovernancePythonFlaskApi GatewayAzureArchitectureArtificial IntelligenceHrSqlJenkinsGcpGitlabKubernetes