FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER.

AGI-Driven Loss of Democratic Control. Meet AI Researcher, Professor Yoshua Bengio

The Royal Swedish Academy of Engineering Sciences (IVA)

02Sept24.

566 views. Streamed live on 2 Sept 2024. Professor Yoshua Bengio is recognized worldwide as one of the leading experts in artificial intelligence. He is best known for his groundbreaking work in deep learning, for which he received the 2018 A.M. Turing Award (the “Nobel Prize of Computing”) along with Geoffrey Hinton and Yann LeCun. Yoshua Bengio is a professor at the Université de Montréal, and the founder and scientific director of Mila – Quebec AI Institute. He is also co-director of the CIFAR Learning in Machines & Brains program as a Senior Fellow, and serves as the scientific director of IVADO – an interdisciplinary and cross-sectoral research and training consortium with a mission to develop and promote robust, reasoning, and responsible AI. Professor Yoshua Bengio will participate in the seminar online, live from Montreal, Canada. The seminar is organized by IVA’s Information Technology Division. Find more from the Royal Swedish Academy of Engineering Sciences (IVA) Facebook:   / iva1919   Linkedin:   / iva   Webb: https://www.iva.se/

Learn more:

  • Can a Bayesian Oracle Prevent Harm from an Agent? Bengio et al
  • ABSTRACT. Is there a way to design powerful AI systems based on machine learning methods that would satisfy probabilistic safety guarantees? With the long-term goal of obtaining a probabilistic guarantee that would apply in every context, we consider estimating a context-dependent bound on the probability of violating a given safety specification. Such a risk evaluation would need to be performed at run-time to provide a guardrail against dangerous actions of an AI. Noting that different plausible hypotheses about the world could produce very different outcomes, and because we do not know which one is right, we derive bounds on the safety violation probability predicted under the true but unknown hypothesis. Such bounds could be used to reject potentially dangerous actions. Our main results involve searching for cautious but plausible hypotheses, obtained by a maximization that involves Bayesian posteriors over hypotheses. We consider two forms of this result, in the iid case and in the non-iid case, and conclude with open problems towards turning such theoretical results into practical AI guardrails.
  • Bounding the probability of harm from an AI to create a guardrail. Published 29 August 2024 by yoshuabengio

FOR EDUCATIONAL AND KNOWLEDGE SHARING PURPOSES ONLY. NOT-FOR-PROFIT. SEE COPYRIGHT DISCLAIMER.