Bringing your first LLM into Production or Beware of your Boss’s Nephew by Christian Hidber
Blog posts and notebooks on «build you chatbot in 5 minutes» or «chat with your data now» are abundant. Deploying a Large Language Model (LLM) application in a production environment—and keeping it running smoothly—is a significantly more complex endeavor.
Through LLM application demos from an industrial planning software, we'll explore how to select promising applications and understand the differences between ad-hoc and production prompts. We dive hands-on into llm-as-a-judge evaluations and online-monitoring of retrieval-based solutions. The talk concludes with some of the issues you might encounter, as well as some – opinionated – recommendations.
About Christian:
Christian lives in Zurich, Switzerland and works as a consultant focusing on real-world machine learning applications. . He earned his PhD in mathematics from ETH Zurich and completed a postdoctoral fellowship at the International Computer Science Institute in Berkeley. Christian has been developing and architecting IT solutions for the last 2 decades. Currently he’s applying artificial intelligence to Geberit’s planning software ProPlanner.