We’re seeking a Senior AI Engineer to design and ship production-grade agentic AI systems that automate complex workflows end-to-end. This is a hands-on, senior role with significant technical ownership. You’ll work closely with the Chief Architect, product, engineering, and domain experts to translate ambiguous, high-impact problems into reliable AI-driven user experiences.
What success looks like:
Ship AI capabilities that measurably improve user outcomes (quality, time saved, throughput)
Build systems that are reliable by design: evals, observability, safety, and cost/latency controls from day one
Iterate quickly using a tight loop of instrument → evaluate → improve → deploy
We’re seeking a Senior AI Engineer to design and ship production-grade agentic AI systems that automate complex workflows end-to-end. This is a hands-on, senior role with significant technical ownership. You’ll work closely with the Chief Architect, product, engineering, and domain experts to translate ambiguous, high-impact problems into reliable AI-driven user experiences.
What success looks like:
Ship AI capabilities that measurably improve user outcomes (quality, time saved, throughput)
Build systems that are reliable by design: evals, observability, safety, and cost/latency controls from day one
Iterate quickly using a tight loop of instrument → evaluate → improve → deploy
What You’ll Do
Agentic AI Feature & Workflow Development
Build and integrate AI-driven features using LLM APIs (OpenAI / Azure OpenAI, Anthropic, Gemini on Vertex AI)
Design and implement tool-using agents (structured function calling, schema validation, retries, fallbacks)
Build multi-agent workflows when appropriate (e.g., planner/worker, reviewer/critic, specialist routing) and know when a simpler architecture is better
Create agentic workflows such as document understanding, extraction, reasoning over evidence, task automation, and multi-step decision support
Own context engineering end-to-end:
dynamic context assembly (retrieval + state + tool outputs)
context budgeting and compression/summarization
grounding strategies to reduce hallucinations and improve consistency
Implement retrieval-augmented generation (RAG) and search workflows using off-the-shelf vector stores and embedding services
Evaluation, Quality & Iteration (Core)
Establish evaluation frameworks for accuracy, reliability, and output quality
Build task-specific eval suites: golden datasets, adversarial cases, regression tests, and rubric-based scoring
Set up automated evaluation pipelines and release gates (CI/CD-friendly) tied to prompt/model/version changes
Define and monitor online metrics (e.g., task success rate, human override rate, safety flags, latency, cost) and run experiments/A-B tests where appropriate
Use LLM-as-judge responsibly: calibrate, validate, and pair with human labels when needed
Engineering, Integration & Observability
Develop scalable backend services and APIs that incorporate AI functionality
Integrate AI pipelines into existing cloud, microservices, and event-driven architectures
Implement observability and analytics for all AI features (tracing, evaluations, prompt versioning, cost tracking) Example tooling: Langfuse (and/or OpenTelemetry-compatible stacks)
Ensure reliability, uptime, performance, and security of AI services
Build internal tooling for evaluation, testing, prompt/version management, and safe deployment
Product & Collaboration
Partner with product managers, designers, the Chief Architect, and domain SMEs to shape AI-first solutions
Rapidly prototype concepts and iterate based on user feedback and measurable eval results
Translate business problems into well-structured AI workflows without requiring ML model training
Document system behavior, known failure modes, and operational playbooks
Governance & Safety
Implement guardrails, checks, and fallback logic for safe and predictable AI behavior
Help define and follow compliance, privacy, and responsible AI guidelines
Design for safe tool execution (bounded actions, permissions, escalation paths, human-in the-loop review
What You Bring
Core Strengths (Required)
Strong software engineering background (Python preferred) and experience shipping backend services
Deep hands-on experience building agentic LLM systems from first principles: agent loops, tool interfaces, planning/replanning, memory/state, and failure handling
Strong context engineering ability: retrieval strategies, routing, grounding, context budgeting, and long-context tradeoffs
Strong evaluation discipline: golden datasets, regression gating, automated eval pipelines, and online monitoring
Practical experience with LLM APIs (OpenAI/Azure OpenAI/Anthropic/Gemini) and AI orchestration frameworks
Excellent debugging, systems thinking, and problem decomposition skills
Comfortable operating in fast-paced, ambiguous environments with high ownership
Signals We Value
You’ve shipped an LLM/agent system in production and can clearly explain:
the failure modes you discovered
the evals you built to catch regressions
how you improved cost/latency while increasing quality
how you monitored and iterated safely over time
You keep up with industry developments (model releases, frameworks, best practices) and can translate them into pragmatic improvement
Nice to Have
Experience with cloud platforms (AWS and/or GCP), microservices, and event-driven systems
Experience with observability stacks (OpenTelemetry, Datadog, Honeycomb) and AI-specific tooling (e.g., Langfuse, Braintrust, HumanLoop, W&B Weave)
Experience with workflow orchestration for long-running jobs (Temporal, Celery, Airflow)
Experience building enterprise AI features (permissions, auditability, compliance constraints)
Experience with safety/policy layers (PII handling, prompt injection defenses, sandboxed tool execution)
Why Join Us
Build core AI capabilities that directly impact users and product strategy
Work on cutting-edge, real-world agentic systems—focused on applied engineering (no model training required)
High ownership, fast iteration cycles, and strong cross-functional collaboration
Competitive compensation and opportunities for rapid advancement
What Your First 90 Days Could Look Like
Ship one production agent workflow end-to-end with:
tracing + observability
an offline eval suite with regression gates
cost/latency targets and monitoring
documented failure modes and fallback path