Raffaella Bernardi – Associate Professor at University of Trento, Italy
Monday, November 6th, 2023 – 14:15-15:00 CET
Title
“The interplay between language generation and reasoning: Information Seeking games“
Abstract
Large Language Models, and ChatGPT in particular, have recently grabbed the attention of the community and the media. Having reached high language proficiency, attention has been shifting toward its reasoning capabilities. It has been shown that ChatGPT can carry out some simple deductive reasoning steps when provided with a series of facts out of which it is tasked to draw some inferences. In this talk, I am going to argue for the need of models whose language generation is driven by an implicit reasoning process. To support my claim, I will present our evaluation of ChatGPT on the 20-Questions game, traditionally used within the Cognitive Science community to inspect the information seeking-strategy’s development. This task requires a series of interconnected skills: asking informative questions, stepwise updating the hypothesis space by computing some simple deductive reasoning steps, and stopping asking questions when enough information has been collected. Thus, it is a perfect testbed to monitor the language and reasoning interplay in LLMs, shed lights on their strength and their weakness, and lay the ground for models that think while speaking.
Christos Christodoulopoulos – Senior Applied Scientist at Amazon, UK
Tuesday, November 7th, 2023 – 11:15-12:00 CET
Title
“Responsible AI in the era of Large Language Models“
Abstract
Large Language models are now ubiquitous, and since the release of ChatGPT last November, are no longer an academic curiosity. As LLMs become part of products used daily by millions of people, there is an increased urgency to ensure that these models are developed and operate responsibly. In this talk, I am going to discuss the what Responsible AI (RAI) looks like in this new era, how RAI is practiced in an industry setting and how it is influenced by and inspires foundational research into RAI topics. I will talk about two recently-published projects from my team that cover two of the many topics associated with RAI. Looking at Fairness, I will present TANGO, a new dataset that measures Transgender and Nonbinary biases in open language generation. In the area of Privacy, I will present a method for controlling the memorisation of potentially sensitive training data through prompt tuning. I will conclude with a look at the use of such RAI research in practice and examples of RAI mitigation strategies for production-ready LLMs.