Discover this podcast and so much more

Podcasts are free to enjoy without a subscription. We also offer ebooks, audiobooks, and so much more for just $11.99/month.

#151 Asa Cooper: How Will We Know If AI Is Fooling Us?

#151 Asa Cooper: How Will We Know If AI Is Fooling Us?

FromEye On A.I.


#151 Asa Cooper: How Will We Know If AI Is Fooling Us?

FromEye On A.I.

ratings:
Length:
60 minutes
Released:
Nov 5, 2023
Format:
Podcast episode

Description

This episode is sponsored by Celonis ,the global leader in process mining. AI has landed and enterprises are adapting. To give customers slick experiences and teams the technology to deliver. The road is long, but you’re closer than you think. Your business processes run through systems. Creating data at every step. Celonis reconstructs this data to generate Process Intelligence. A common business language. So AI knows how your business flows. Across every department, every system and every process. With AI solutions powered by Celonis enterprises get faster, more accurate insights. A new level of automation potential. And a step change in productivity, performance and customer satisfaction Process Intelligence is the missing piece in the AI Enabled tech stack.   Go to https://celonis.com/eyeonai to find out more.   Welcome to episode 151 of the ‘Eye on AI’ podcast. In this episode, host Craig Smith sits down with Asa Strickland Cooper, a postdoctoral researcher at NYU, who is at the forefront of language model safety.  This episode takes us on a journey through the complexities of AI situational awareness, the potential for consciousness in language models, and the future of AI safety research. Craig and Asa delve into the nuances of AI situational awareness and its distinction from sentience. Asa, with his rich background in NLP and AI safety from Edinburgh University, shares insights from his post-doc work at NYU, discussing collaborative efforts on a paper that has garnered attention for its take on situational awareness in large language models (LLMs).  We explore the economic drivers behind creating AI with such capabilities and the role of scaling versus algorithmic innovation in achieving this milestone. We also delve into the concept of agency in LLMs, the challenges of post-deployment monitoring, and the effectiveness of current measures in detecting situational awareness. To wrap things off, we break down the importance of source trustworthiness and the model's ability to discern reliable information, a critical aspect of AI safety and functionality, so make sure to watch till the end.   Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI (00:00) Preview and Introduction (02:30) Asa's NLP Expertise and the Safety of Language Models (06:05) Breaking Down AI's Situational Awareness  (13:44) Evolution of AI: Predictive Models to AI Coworkers (20:29) New Frontier in AI Development? (27:14) Measuring AI's Awareness (33:49) Innovative Experiments with LLMs (40:51) The Consequences of Detecting Situational Awareness in AI (44:07) How To Train AI On Trusted Sources (49:52) What Is The Future of AI Training? (56:35) AI Safety: Public Concerns and the Path Forward**  
Released:
Nov 5, 2023
Format:
Podcast episode

Titles in the series (100)

Eye on A.I. is a biweekly podcast, hosted by longtime New York Times correspondent Craig S. Smith. In each episode, Craig will talk to people making a difference in artificial intelligence. The podcast aims to put incremental advances into a broader context and consider the global implications of the developing technology. AI is about to change your world, so pay attention.