What Is the role?
We are looking for an AI Engineer who can build and ship AI-powered applications — not just prototype them. You’ll design RAG pipelines, build AI agents, integrate foundation models into real products, and keep them running reliably in production on AWS and Azure. This is a hands-on engineering role, not a research role.
Key Responsibilities
AI Application Development:
- Build production RAG systems: chunking strategies, embedding pipelines, retrieval, reranking, and response generation
- Design and implement AI agents using LangGraph or similar frameworks for multi-step reasoning and tool use
- Integrate foundation models (Claude, GPT, Llama, Mistral) via AWS Bedrock and Azure AI Foundry into existing applications
- Implement structured output, function calling, and tool use patterns for reliable model interactions
- Build and maintain prompt templates, manage prompt versioning, and iterate based on evaluation results
AI Infrastructure & Operations:
- Deploy and manage AI workloads on AWS Bedrock and Azure AI Foundry
- Set up observability for AI systems — tracing, latency monitoring, token usage, and quality metrics using tools like LangSmith or Langfuse
- Implement evaluation pipelines: automated evals, LLM-as-judge, and regression testing for model outputs
- Manage vector databases for semantic search (Pinecone, Qdrant, or pgvector)
- Optimize costs: model selection, caching, batching, and routing between models based on task complexity
Safety & Quality:
- Implement guardrails for content filtering, PII detection, and hallucination mitigation
- Design fallback strategies for when models fail or produce low-confidence outputs
- Build human-in-the-loop review workflows where needed
General:
- Write clean, testable Python (and/or TypeScript) code
- Containerize AI services with Docker and deploy to cloud infrastructure
- Collaborate with product and engineering teams to identify where AI adds real value
Required Skills
AI/ML Fundamentals:
- 2+ years building AI-driven applications that run in production (not just notebooks)
- Strong understanding of how LLMs work: tokenization, context windows, temperature, and their practical implications
- Hands-on experience building RAG systems: document ingestion, chunking, embeddings, retrieval strategies, and reranking
- Prompt engineering that goes beyond basics: structured outputs, chain-of-thought, few-shot examples, and system prompts
- Experience with at least one agent framework (LangGraph, LangChain, or LlamaIndex) for building multi-step workflows
Cloud AI Platforms:
- Production experience with AWS Bedrock or Azure AI Foundry (model access, guardrails, knowledge bases)
- Familiarity with managed AI services: AWS Bedrock Agents, Azure AI Search, Azure Document Intelligence
- Understanding of model selection trade-offs: cost, latency, quality, and context window size
Observability & Evaluation:
- Experience setting up tracing and monitoring for AI systems (LangSmith, Langfuse, or similar)
- Ability to design and run evaluation pipelines: automated scoring, LLM-as-judge, and regression tests
- Understanding of key metrics: latency, token usage, retrieval relevance, answer quality
Vector Databases & Search:
- Hands-on experience with at least one vector database (Pinecone, Qdrant, pgvector, or Weaviate)
- Understanding of embedding models, similarity search, and hybrid search (vector + keyword)
Engineering:
- Strong Python skills — this is your primary language
- Experience with Docker and containerized deployments
- Comfortable with Git, CI/CD, and working in a team codebase
- Basic cloud infrastructure knowledge (AWS or Azure)
Preferred Skills
Advanced AI:
- Experience with fine-tuning (LoRA, QLoRA) for domain-specific use cases
- Knowledge of MCP (Model Context Protocol) for connecting AI agents to external tools and data
- Multi-agent system design and orchestration patterns
- Streaming architectures for real-time AI responses
- Synthetic data generation for training and evaluation
Cloud & Infrastructure:
- AWS SageMaker for custom model hosting
- Azure AI Studio for model experimentation and deployment
- Infrastructure as Code (CDK, Terraform) for AI infrastructure
- Kubernetes for scaling AI workloads
Additional:
- Experience with multimodal AI (vision + text)
- Knowledge of AI security: prompt injection prevention, data leakage mitigation
- Cost optimization strategies for LLM-heavy workloads (caching, model routing, batching)
- TypeScript/Node.js for building AI-powered APIs alongside Python services
Personal Qualities
- You debug systematically — AI systems fail in non-obvious ways and you’re patient enough to trace through them
- Clear communicator — can explain model behavior and trade-offs to non-AI teammates
- Pragmatic about tool choices — you pick what works, not what’s trending
- Comfortable with ambiguity — AI projects often require experimentation before a clear path emerges
- Keeps up with the field without chasing every new paper or framework
We offer you
- Competitive Compensation
- Professional Growth
- Cutting-Edge Technologies
- Highly motivated & collaborative Team
- Challenging Projects
- Work-Life Balance
- Opportunities for Advancement
- Employee Well-being