The conversation features André Carpati, discussing the future of AI agents, his work in AI education, and his broader views on intelligence and technological progress. Carpati argues that predictions of rapid advancement in AI agents are overblown, suggesting a "decade of agents" is more realistic than a single year. He highlights current AI agents' limitations in intelligence, multimodality, and continuous learning, stating they are not yet suitable for tasks like employee-level work.
Carpati draws on his 15 years of experience in AI to contextualize present optimism and pessimism, referencing the early excitement around deep learning and reinforcement learning in gaming. He critiques the over-reliance on gaming as a training ground for general AI, advocating for real-world interaction and knowledge work as more relevant targets.
He uses an analogy of building "ghosts" rather than "animals," emphasizing that current AI development is rooted in imitation of human-generated internet data rather than replicating evolution or animalistic learning processes. Carpati contrasts this with Richard Sutton's approach of building AI to mimic animals, expressing skepticism about whether this is fully achievable.
The conversation explores the nuances of in-context learning, suggesting it might be a form of gradient descent within a token window, completing patterns within the model's weights. The discussion touches on the differences between pre-training and in-context learning, highlighting how much more information the models can store. He also brings up an analogy of models resembling humans and the models recalling information like a "hazy recollection"
Carpati talks about how to make the best version of models by stripping them from knowledge and maintaining a "cognitive core" He outlines how people are pushing this field forward by using reasoning and the prefrontal cortex, and not doing much for like the basal ganglia.
Carpati offers insights on software engineering with AI, noting that autocomplete is now helpful but isn't ready for advanced coding and code that has never been written before. He emphasizes that agents tend to make too many cognitive errors, and he doesn't think it's an actual net gain.
The discussion touches on the challenges of reinforcement learning and the difficulty of assigning partial credit in process supervision. He brings up the idea of using LLMs as judges.
Looking at AI, he describes his concerns with AI becoming more powerful, but also more out of control. He describes how the field can become a pot that is completely autonomous.
Discussing Eureka, Carpati envisions building something like "Starfleet Academy," an elite educational institution focused on technical knowledge. He stresses that the best option is a one-on-one tutor and wants to build an education model around that. As well, he hopes to build "ramps to knowledge" to help advance the field and teach students.
Lastly, Carpati discusses his previous role at Tesla in building a self driving car. He states that it will still take a while, and thinks the demo stage is often useless. He is building an educational tool for self-driving cars as well.
He also offers insights for any educator from any field, stating that they should start by laying out the first order and then attack the other terms.