The talk will cover why and how to monitor LLMs deployed to production, aiming for their enhanced performance and efficiency.
We will focus on the state-of-the-art solutions for detecting hallucinations, split into two types:
1. Uncertainty Quantification
2. LLM self-evaluation
In the Uncertainty Quantification part, we will discuss algorithms to leverage token probabilities to estimate the quality of model responses. This includes simple accuracy estimation and more advanced methods for estimating Semantic Uncertainty or any classification metric.
In the LLM self-evaluation part, we will cover using (potentially the same) LLM to quantify the quality of the answer. We will also cover state-of-the-art algorithms such as SelfCheckGPT and LLM-eval.
You will build an intuitive understanding of the LLM monitoring methods, their strengths and weaknesses, and learn how to easily set up an LLM monitoring system, creating hallucination-free LLMs.
CTO at NannyML
We are looking for passionate people willing to cultivate and inspire the next generation of leaders in tech, business, and data science. If you are one of them get in touch with us!