2025 Conference Tracks
This track covers the key architectural choices and infra strategies behind scaling AI and LLM systems in production—from bare metal to Kubernetes, GPU scheduling to inference optimization. Learn what it really takes to build and operate reliable GenAI and agent platforms at scale.
This track explores the design patterns shaping modern agents, from prompt engineering and tool integration to memory and planning strategies, focusing on real-world systems, not just frameworks. It also covers the infrastructure, safety checks, and governance required to deploy agents reliably and securely in production environments, with expert presenters sharing their insights on the challenges of running agents at scale.
This track covers the key architectural choices and infrastructure strategies behind scaling AI and LLM systems in production, from bare metal to Kubernetes, GPU scheduling to inference optimization. It also addresses the complexities of managing model, data, and pipeline versions in a reproducible, team-friendly way, alongside the unique challenges of deploying ML in regulated, resource-constrained, or air-gapped environments. Expert speakers will share insights on building and operating reliable GenAI and agent platforms at scale while navigating the tradeoffs when cloud-based solutions aren’t an option.
This track focuses on scoping and delivering complex AI projects, exploring how teams are adapting their scoping processes to account for LLMs, agents, and evolving project boundaries in fast-moving environments. It also dives into the strategies behind AI product development, from aligning business goals to driving successful delivery and scaling. Expert presenters will share practical insights on navigating the complexities of AI product strategy and execution.