62 minutes | May 24, 2023
#123 Aidan Gomez: How AI Language Models Will Shape The Future
Welcome to Eye on AI, the podcast that keeps you informed about the latest trends, obstacles, and possibilities in the realm of artificial intelligence. In this episode, we have the privilege of engaging in a thought-provoking discussion with Aidan Gomez, an exceptional AI developer and co-founder of Cohere. Aidan’s passion lies in enhancing the efficiency of massive neural networks and effectively deploying them in the real world. Drawing from his vast experience, which includes leading a team of researchers at For.ai and conducting groundbreaking research at Google Brain, Aidan provides us with unique insights and anecdotes that shed light on the AI landscape. During our conversation, Aidan explains his collaboration with the legendary Geoffrey Hinton and their remarkable project at Google Brain. We delve into the intricate architecture of AI systems, demystifying the construction of the transformative transformer algorithm. Aidan generously shares his knowledge on the creation of attention within these models and the complexities of scaling such systems. As we explore the fascinating domain of language models, Aidan discusses their learning process, bridging the gap between code and data. We uncover the immense potential of these models to suggest other large-scale counterparts. We gain invaluable insights into Aidan’s journey as a co-founder of Cohere, an innovative platform revolutionizing the utilization of language technology. Tune in to Eye on AI now to immerse yourself in a captivating conversation that will expand your understanding of this ever-develop field. (00:00) Preview (00:33) Introduction & sponsorship (02:00) Aidan's background with machine learning & AI (05:10) Geoffrey Hinton & Aidan Gomez working together (07:55) Aidan Gomez & Google Brain's project (12:53) Aidan's role in building AI architecture (15:25) How the transformer algorithm is built (18:25) How do you create attention? (20:40) How do you scale the model? (25:10) How language models learn from code and data (29:55) Did you know the potential of the project? (34:15) Can LLMs suggest other large models? (36:45) How Aidan Gomez started Cohere (41:10) How do people use Cohere? (46:50) Examples of language technology models (48:40) How Cohere handles hallucinations (52:53) The dangers of AI Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI
56 minutes | May 10, 2023
#122 Connor Leahy: Unveiling the Darker Side of AI
Welcome to Eye on AI, the podcast that explores the latest developments, challenges, and opportunities in the world of artificial intelligence. In this episode, we sit down with Connor Leahy, an AI researcher and co-founder of EleutherAI, to discuss the darker side of AI. Connor shares his insights on the current negative trajectory of AI, the challenges of keeping superintelligence in a sandbox, and the potential negative implications of large language models such as GPT4. He also discusses the problem of releasing AI to the public and the need for regulatory intervention to ensure alignment with human values. Throughout the podcast, Connor highlights the work of Conjecture, a project focused on advancing alignment in AI, and shares his perspectives on the stages of research and development of this critical issue. If you’re interested in understanding the ethical and social implications of AI and the efforts to ensure alignment with human values, this podcast is for you. So join us as we delve into the darker side of AI with Connor Leahy on Eye on AI. (00:00) Preview (00:48) Connor Leahy’s background with EleutherAI & Conjecture (03:05) Large language models applications with EleutherAI (06:51) The current negative trajectory of AI (08:46) How difficult is keeping super intelligence in a sandbox? (12:35) How AutoGPT uses ChatGPT to run autonomously (15:15) How GPT4 can be used out of context & negatively (19:30) How OpenAI gives access to nefarious activities (26:39) The problem with the race for AGI (28:51) The goal of Conjecture and advancing alignment (31:04) The problem with releasing AI to the public (33:35) FTC complaint & government intervention in AI (38:13) Technical implementation to fix the alignment issue (44:34) How CoEm is fixing the alignment issue (53:30) Stages of research and development of Conjecture Craig Smith Twitter: https://twitter.com/craigss Eye on A.I. Twitter: https://twitter.com/EyeOn_AI
56 minutes | Apr 28, 2023
Danny Tobey: At the Intersection of Law and Artificial Intelligence
In this podcast, we sit down with Danny Tobey, an attorney with the global law firm DLA Piper, to discuss the changing legal dynamics surrounding artificial intelligence. As one of the leading experts in the field, Danny provides valuable insights into the current state of legislation and regulation, the efforts of regulatory bodies like the Federal Trade Commission in tackling issues related to AI, and how the law firm of the future will look as AI continues to transform the economy. With the growing impact of AI on all aspects of our lives, the legal profession is facing unique challenges and opportunities. Danny brings a wealth of knowledge and experience to the conversation, having worked with clients in industries ranging from healthcare to financial services to consumer products. Throughout the podcast, Danny explores the ethical and legal implications of AI, as well as the ways in which AI is already reshaping the legal industry. He provides thoughtful perspectives on how the legal profession can adapt and evolve to meet the demands of an AI-driven economy, and the role that lawyers and regulatory bodies will play in shaping the future of this transformative technology. Whether you're a legal professional looking to stay on top of the latest developments in AI, or simply interested in the ways that AI is changing the legal landscape, this podcast is sure to offer valuable insights and food for thought. So join us as we dive deep into the intersection of law and artificial intelligence with Danny Tobey. Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
41 minutes | Apr 13, 2023
Yoshua Bengio: Pausing More Powerful AI Models and His Work on World Models
In this episode of the Eye on A.I. podcast, host Craig Smith interviews Yoshua Bengio, one of the founding fathers of deep learning and a Turing Award winner. Bengio shares his insights on the famous pause letter, which he signed along with other prominent A.I. researchers, calling for a more responsible approach to the development of A.I. technologies. He discusses the potential risks associated with increasingly powerful A.I. models and the importance of ensuring that models are developed in a way that aligns with our ethical values. Bengio also talks about his latest research on world models and inference machines, which aim to provide A.I. systems with the ability to reason for reality and make more informed decisions. He explains how these models are built and how they could be used in a variety of applications, such as autonomous vehicles and robotics. Throughout the podcast, Bengio emphasises the need for interdisciplinary collaboration and the importance of addressing the ethical implications of A.I. technologies. Don’t miss this insightful conversation with one of the most influential figures in A.I. on Eye on A.I. podcast! Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
35 minutes | Mar 30, 2023
Edo Liberty: Solving ChatGPT Hallucinations With Vector Embeddings
Welcome to the latest episode of our podcast featuring Edo Liberty, an AI expert and former creator of SageMaker at Amazon’s AI labs. In this episode, Edo discusses how his team at Pinecone.io is tackling the problem of hallucinations in large language models like ChatGPT. Edo’s approach involves using vector embeddings to create a long-term memory database for large language models. By converting authoritative and trusted information into vectors, and loading them into the database, the system provides a reliable source of information for large language models to draw from, reducing the likelihood of inaccurate responses. Throughout the episode, Edo explains the technical details of his approach and shares some of the potential applications for this technology, including AI systems that rely on language processing. Edo also discusses the future of AI and how this technology could revolutionise the way we interact with computers and machines. With his insights and expertise in the field, this episode is a must-listen for anyone interested in the latest developments in AI and language processing. We have a new sponsor this week: NetSuite by Oracle, a cloud-based enterprise resource planning software to help businesses of any size manage their financials, operations, and customer relationships in a single platform. They've just rolled out a terrific offer: you can defer payments for a full NetSuite implementation for six months. That's no payment and no interest for six months, and you can take advantage of this special financing offer today at netsuite.com/EYEONAI Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
43 minutes | Mar 15, 2023
Ilya Sutskever: The Mastermind Behind GPT-4 and the Future of AI
In this podcast episode, Ilya Sutskever, the co-founder and chief scientist at OpenAI, discusses his vision for the future of artificial intelligence (AI), including large language models like GPT-4. Sutskever starts by explaining the importance of AI research and how OpenAI is working to advance the field. He shares his views on the ethical considerations of AI development and the potential impact of AI on society. The conversation then moves on to large language models and their capabilities. Sutskever talks about the challenges of developing GPT-4 and the limitations of current models. He discusses the potential for large language models to generate a text that is indistinguishable from human writing and how this technology could be used in the future. Sutskever also shares his views on AI-aided democracy and how AI could help solve global problems such as climate change and poverty. He emphasises the importance of building AI systems that are transparent, ethical, and aligned with human values. Throughout the conversation, Sutskever provides insights into the current state of AI research, the challenges facing the field, and his vision for the future of AI. This podcast episode is a must-listen for anyone interested in the intersection of AI, language, and society. Timestamps: (00:04) Introduction of Craig Smith and Ilya Sutskever. (01:00) Sutskever's AI and consciousness interests. (02:30) Sutskever's start in machine learning with Hinton. (03:45) Realization about training large neural networks. (06:33) Convolutional neural network breakthroughs and imagenet. (08:36) Predicting the next thing for unsupervised learning. (10:24) Development of GPT-3 and scaling in deep learning. (11:42) Specific scaling in deep learning and potential discovery. (13:01) Small changes can have big impact. (13:46) Limits of large language models and lack of understanding. (14:32) Difficulty in discussing limits of language models. (15:13) Statistical regularities lead to better understanding of world. (16:33) Limitations of language models and hope for reinforcement learning. (17:52) Teaching neural nets through interaction with humans. (21:44) Multimodal understanding not necessary for language models. (25:28) Autoregressive transformers and high-dimensional distributions. (26:02) Autoregressive transformers work well on images. (27:09) Pixels represented like a string of text. (29:40) Large generative models learn compressed representations of real-world processes. (31:31) Human teachers needed to guide reinforcement learning process. (35:10) Opportunity to teach AI models more skills with less data. (39:57) Desirable to have democratic process for providing information. (41:15) Impossible to understand everything in complicated situations. Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
34 minutes | Mar 2, 2023
Ben Sorscher: Data Pruning for Efficient Machine Learning
In this episode, Ben Sorscher, a PhD student at Stanford, sheds light on the challenges posed by the ever-increasing size of data sets used to train machine learning models, specifically large language models. The sheer size of these data sets has been pushing the limits of scaling, as the cost of training and the environmental impact of the electricity they consume becomes increasingly enormous. As a solution, Ben discusses the concept of “data pruning” - a method of reducing the size of data sets without sacrificing model performance. Data pruning involves selecting the most important or representative data points and removing the rest, resulting in a smaller, more efficient data set that still produces accurate results. Throughout the podcast, Ben delves into the intricacies of data pruning, including the benefits and drawbacks of the technique, the practical considerations for implementing it in machine learning models, and the potential impact it could have on the field of artificial intelligence. Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
55 minutes | Feb 16, 2023
Yann LeCun: Filling the Gap in Large Language Models
In this episode, Yann LeCun, a renowned computer scientist and AI researcher, shares his insights on the limitations of large language models and how his new joint embedding predictive architecture could help bridge the gap. While large language models have made remarkable strides in natural language processing and understanding, they are still far from perfect. Yann LeCun points out that these models often cannot capture the nuances and complexities of language, leading to inaccuracies and errors. To address this gap, Yann LeCun introduces his new joint embedding predictive architecture - a novel approach to language modelling that combines techniques from computer vision and natural language processing. This approach involves jointly embedding text and images, allowing for more accurate predictions and a better understanding of the relationships between original concepts and objects. Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
37 minutes | Feb 1, 2023
Terry Sejnowski: NeurIPS and the Future of AI
In this episode, Terry Sejnowski, an AI pioneer, chairman of the NeurIPS Foundation, and co-creator of Boltzmann Machines, delves into the latest developments in deep learning and their potential impact on our understanding of the human brain. Terry Sejnowski begins by discussing the NeurIPS conference - one of the most significant events in the field of artificial intelligence - and its role in advancing research and innovation in deep learning. He shares insights into the latest breakthroughs in the field, including the repurposing of the sleep-wake cycle of Boltzmann Machines in Geoff Hinton's new Forward-Forward algorithm. Throughout the episode, Terry Sejnowski shares his expertise on the intersection of artificial intelligence and neuroscience, exploring how advances in deep learning may help us better understand the complexities of the human brain. He discusses how researchers are using AI techniques to study brain activity and the potential implications for fields such as medicine and psychology. Overall, this episode will be of particular interest to those interested in the latest developments in artificial intelligence and their potential applications in neuroscience and related fields. Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
59 minutes | Jan 19, 2023
Geoffrey Hinton: Unpacking The Forward-Forward Algorithm
In this episode, Geoffrey Hinton, a renowned computer scientist and a leading expert in deep learning, provides an in-depth exploration of his groundbreaking new learning algorithm - the forward-forward algorithm. Hinton argues this algorithm provides a more plausible model for how the cerebral cortex might learn, and could be the key to unlocking new possibilities in artificial intelligence. Throughout the episode, Hinton discusses the mechanics of the forward-forward algorithm, including how it differs from traditional deep learning models and what makes it more effective. He also provides insights into the potential applications of this new algorithm, such as enabling machines to perform tasks that were previously thought to be exclusive to human cognition. Hinton shares his thoughts on the current state of deep learning and its future prospects, particularly in neuroscience. He explores how advances in deep learning may help us gain a better understanding of our own brains and how we can use this knowledge to create more intelligent machines. Overall, this podcast provides a fascinating glimpse into the latest developments in artificial intelligence and the cutting-edge research being conducted by one of its leading pioneers. Craig Smith Twitter: https://twitter.com/craigssEye on A.I. Twitter: https://twitter.com/EyeOn_AI
60 minutes | Jan 2, 2023
Setting the stage for 2023
To set the stage for some terrific conversations I have coming to you in the new year, in this episode we go back to some earlier conversations that talk about how we got to where we are in deep learning and how those early threads continue to lead innovation.
40 minutes | Nov 9, 2022
AI Supply Chain Optimization
This week I talk to Bob Rogers, a Harvard trained astrophysicist who once built digital twins of black holes to better understand them, and now builds digital twins of supply chains to help make them more efficient and resilient.
45 minutes | Oct 20, 2022
NO-CODE WITH AKKIO
Jonathon Reilly, co-founder of Akkio, a no-code AI platform, talks about how users with a web browser and an idea have the power to bring AI to life themselves without having to write code.
43 minutes | Oct 5, 2022
MLOps with ClearML
Moses Guttmann, founded ClearML, talks about the evolution of the MLOps industry over the past few years and ClearML's contribution to it.
20 minutes | Sep 21, 2022
Bratin Saha, head of Amazon's machine learning services, talks about Amazon's growing dominance in model building and deploying AI, about the company's SageMaker platform, and whether anyone can compete with the behemoth.
29 minutes | Sep 8, 2022
AUTOMATED CODE GENERATION
Peter Schrammel, one of the founders of Diffblue, an automated unit-test writing software company, speaks about the increasing automatic generation of code and how he sees such automation increasing the productivity of developers.
30 minutes | Aug 25, 2022
Michael Kearns on Privacy
Michael Kearns, a computer scientist professor at the University of Pennsylvania and an Amazon scholar talks about differential privacy, how Amazon's research approach differs from its peers, and how AI will eventually permeate all aspects of our lives.
37 minutes | Aug 10, 2022
Jacki Morie, a senior XPRIZE advisor, talks about the ANA Avatar XPRIZE, a competition focused on creating a physical avatar system that will seamlessly transport human skills and experience to distant locations. The four-year competition is in its final stretch.
41 minutes | Jul 28, 2022
VITAL & MINT
Aaron Patzer, founder of the personal finance app MINT and more recently founder of the AI-based healthcare company Vital, talks about keeping customer data private and the promise, giving emergency room patients information with AI and finding friendly solutions to anxiety-producing problems.
32 minutes | Jul 15, 2022
Amazon's Rohit Prasad
Rohit Prasad, Amazon's Senior Vice President and Head Scientist for Alexa, speaks about the development of conversational AI and virtual assistants and the merging of IoT sensor data into ambient intelligence - AI that is always present and immediately accessible.