Build & optimize AI Agents that survive production

Build & optimize AI Agents that survive production

Hands-on workshop - The Fifth Elephant Pune Edition

Tickets

Loading…

About the workshop

It’s easy to ship a magical agent demo. It’s much harder to ship an agent that works for real users: noisy inputs, partial context, flaky tools, ambiguous goals, and “tiny prompt changes” that break everything.

In this hands-on workshop, we’ll build a small but realistic agent in Python + DSPy, then turn it into something you can actually run in production: structured I/O, tool contracts, tracing, evals, and automatic optimization.

You’ll leave with a concrete engineering workflow (an “agent improvement loop”) that you can take back to your team:

instrument → collect failures → convert to evals → optimise → ship via CI.

The techniques are framework-agnostic; we’ll use DSPy because it makes optimisation and modularity explicit in code.

Level: Intermediate

Workshop duration: 3 hours

Prerequisites:

  • Comfortable with Python (APIs, functions, virtualenv/uv)
  • Basic familiarity with LLMs (prompts, tool calling concepts)
  • Laptop + internet access

Target audience

  • Software engineers / platform engineers building LLM features
  • SDETs / QA engineers working on evals and reliability
  • Engineering managers and tech leads who need a production-ready approach

Workshop outline

  1. Anatomy of a production agent (15 min)
    Agent loop, tool contracts, ground-truth checks, stop conditions, failure modes.

  2. Build the agent in DSPy (60 min)
    Signatures + modules, tool wiring, structured outputs, error handling.

  3. Observability & evals (45 min)
    Tracing, failure buckets, creating an eval set from real-ish cases, measuring baseline.

  4. Optimization (45 min)
    Few-shot baselines → DSPy optimiser run → compare metrics + inspect deltas.

  5. Shipping the improvement loop (15 min)
    Minimal CI pattern: run evals on PR, regressions gate merges, version prompts/programs.

Set-up

The workshop skeleton and requirements can be found in this repo: https://github.com/unravel-team/real-agents-workshop

Key takeaways

  • A practical mental model of agents: goal → plan/act loop → tool calls → ground-truth checks → stop conditions.
  • How to build agents as maintainable software (signatures/modules) instead of brittle prompt blobs.
  • How to add observability + evals so you can debug “why it failed” and measure progress.
  • How to use DSPy optimisers (few-shot + program/prompt optimisation) to improve quality systematically.
  • A repeatable CI workflow to keep agents improving safely as users and requirements change.

About the instructors

Kiran Kulkarni is the founder of Unravel.tech, where he helps teams build production-grade AI systems—agentic workflows, evaluation pipelines, and reliability/observability practices. He’s been a founding engineer and engineering leader across data + AI systems and loves turning “cool demos” into software that survives real users.

Utkarsh Dighe is a senior engineer at Unravel.tech, where he designs and builds pragmatic solutions using Agentic AI to tackle complex problems across domains. He takes an engineering-first approach, focusing on reliability, robustness, and scalability—ensuring systems don’t just work in theory, but hold up under real-world usage.

How to attend this workshop

This workshop is part of The Fifth Elephant Pune edition and is open for The Fifth Elephant annual members. If you wish to attend The Fifth Elephant Pune edition, pick up an annual membership.

This workshop is open to 40 participants only. Seats will be available on first-come-first-serve basis. 🎟️

Contact information ☎️

For inquiries about the workshop, contact +91-7676332020 or write to info@hasgeek.com

Hosted by

Jumpstart better data engineering and AI futures