Large Language Models (LLMs) are known for their exceptional ability to mimic human behavior, but they struggle when it comes to updating their beliefs based on new evidence. According to a team of researchers from Google, current AI agents lack the skill of ‘probabilistic reasoning,’ which involves continuously updating a ‘world model’ as new information is gathered.
To address this issue, the researchers propose a new approach: instead of focusing on providing LLMs with the correct answers, they suggest teaching them how to make educated guesses like a mathematician.
The Challenge: Limited Adaptability
While LLMs like Gemini-1.5 Pro and GPT-4.1 Mini excel at tasks like writing code or summarizing emails, they face challenges when it comes to interactive tasks. For example, a flight booking assistant needs to understand user preferences, such as price versus duration, by observing their choices over multiple interactions.
The research team discovered that off-the-shelf LLMs, including popular models like Llama-3-70B and Qwen-2.5-32B, showed minimal improvement after the first interaction. In contrast, a ‘Bayesian Assistant,’ which uses Bayes’ rule to update probability distributions, improves with each new data point. Standard LLMs, however, struggle to adjust their internal beliefs based on user feedback.
Introducing Bayesian Teaching
To address this issue, the research team introduced a technique called Bayesian Teaching. Instead of training the model on ‘correct’ data, as done with an Oracle Teacher, they fine-tuned it to imitate a Bayesian Assistant. This approach involves updating a probability distribution over possible user preferences based on new information.
The research team conducted a five-round flight recommendation interaction task, where flights were defined by features like price, duration, and stops. The Bayesian Assistant updated its beliefs after each round, taking into account prior assumptions and user feedback.
By using Supervised Fine-Tuning (SFT) on Bayesian interactions, the researchers encouraged LLMs to embrace reasoning under uncertainty, not just focusing on final outcomes.
The Power of ‘Educated Guesses’
Surprisingly, the research findings showed that Bayesian Teaching consistently outperformed Oracle Teaching. While Oracle Teaching trained the model on predetermined correct answers, Bayesian Teaching allowed the model to make educated guesses and learn from its mistakes. By observing the Bayesian Assistant navigate uncertainty and update its beliefs, LLMs learned the skill of belief updating.
The results were impressive: Bayesian-tuned models like Gemma-2-9B and Llama-3-8B were not only more accurate but also aligned with the gold standard Bayesian strategy 80% of the time.
Generalization Across Domains
The ultimate goal for developers is to achieve generalization. The research team tested their fine-tuned models on tasks involving increased complexity, new domains like hotel recommendations, and real-world scenarios such as web shopping. Despite being trained on synthetic flight data, the Bayesian LLMs successfully applied their probabilistic reasoning skills to other domains, outperforming human participants in certain instances.
The Neuro-Symbolic Connection
This research showcases the strength of deep learning in distilling classic symbolic models like the Bayesian Assistant into neural networks like LLMs. While symbolic models excel at simple tasks, they struggle with complex real-world scenarios. By teaching LLMs to mimic the reasoning strategies of symbolic models, developers can leverage the benefits of both approaches.
Key Insights
– LLMs face challenges in updating their beliefs based on new information, with performance often plateauing after initial interactions.
– Bayesian Teaching, which focuses on educated guesses and uncertainty, outperforms direct training on correct answers.
– LLMs fine-tuned on one task can transfer their skills to other domains, showcasing the power of probabilistic reasoning.
– LLMs demonstrate greater resilience to human noise compared to purely symbolic models, making them effective in real-world interactions.
– Distilling symbolic strategies into neural models through supervised fine-tuning enables LLMs to apply complex reasoning strategies in messy or complex domains.
For more details, refer to the original Paper and Technical details. Stay updated by following us on Twitter and join our ML SubReddit with over 120k members. Don’t forget to subscribe to our Newsletter and join us on Telegram for more insights.
By rewriting the HTML content into a comprehensive and SEO-friendly article, the essence of the original text has been preserved while enhancing its readability and engagement. The output is a refined HTML version ready for seamless integration into a WordPress website.





Be the first to comment