Thursday, January 22, 2026
Technology
129 min read

Princeton's Tom Griffiths on Connecting Language and Artificial Intelligence

MIT Sloan Management Review
January 20, 20262 days ago
Connecting Language and (Artificial) Intelligence: Princeton’s Tom Griffiths

AI-Generated Summary
Auto-generated

Princeton professor Tom Griffiths discusses his book "The Laws of Thought," exploring how mathematics underpins both human and artificial intelligence. The conversation covers three core frameworks: rules and symbols, neural networks, and probability. Griffiths highlights how these models connect cognitive science and large language models, and examines the distinctions between human and machine intelligence, particularly in areas like judgment and curation.

In this bonus episode of the Me, Myself, and AI podcast, Princeton University professor and artificial intelligence researcher Tom Griffiths joins host Sam Ransbotham to unpack The Laws of Thought, his new book exploring how math has been used for centuries to understand how minds — human and machine — actually work. Tom walks through three main frameworks shaping intelligence today — rules and symbols, neural networks, and probability — and he explains why modern AI only makes sense when you see how those pieces fit together. The conversation connects cognitive science, large language models, and the limits of human versus machine intelligence. Along the way, Tom and Sam dig into language, learning, and what humans still do better — like judgment, curation, and metacognition. Subscribe to Me, Myself, and AI on Apple Podcasts or Spotify. Transcript Allison Ryder: Hi, everyone. While we’re on winter break, we’re dropping a couple of bonus episodes featuring cutting-edge academic researchers. On today’s episode, Sam is joined by professor and director of Princeton’s Computational Cognitive Science Lab, Tom Griffiths. Tom is the author of the forthcoming book The Laws of Thought and joins Sam today to speak about AI’s mathematical and linguistic backgrounds. It was a fascinating conversation, and I hope you enjoy it. Tom Griffiths: I’m Tom Griffiths from Princeton University, and you’re listening to Me, Myself, and AI. Sam Ransbotham: Welcome to Me, Myself, and AI, a podcast from MIT Sloan Management Review exploring the future of artificial intelligence. I’m Sam Ransbotham, professor of analytics at Boston College. I’ve been researching data, analytics, and AI at MIT SMR since 2014, with research articles, annual industry reports, case studies, and now 12 seasons of podcast episodes. In each episode, corporate leaders, cutting-edge researchers, and AI policy makers join us to break down what separates AI hype from AI success. Hi, listeners. Thanks, everyone, for joining us again. Our guest today is Tom Griffiths, professor of psychology and computer science, and the director of the Princeton Laboratory for Artificial Intelligence. Tom has a new book, The Laws of Thought, which I suspect our listeners will enjoy learning about. Tom, [it’s] great to have you on the podcast. Tom Griffiths: Thanks, Sam. [It’s] great to be here. Sam Ransbotham: Why don’t we start with … I think people know it’s kind of fun to be your professor because people know what professors are, but maybe let’s start with a little bit of a bio. Can you give us some background on what your roles are with the lab at Princeton? Tom Griffiths: Princeton, like a lot of other educational institutions, has been trying to figure out how to respond to all of the things that are happening with AI in the world at the moment. The AI lab is the starting point for doing that in terms of thinking about being able to make some targeted investments in research areas where we see potential for [a] transformative impact for AI in a way that’s maybe more nimble than a traditional academic institution might. Sam Ransbotham: There’s a lot going on within universities trying to figure out what exactly all this means, and I guess all of society. But let’s start with The Laws of Thought. Can you explain in some simple terms what these laws are and how they relate to human cognition and artificial intelligence? Tom Griffiths: The idea behind the book is … I think all of us in school learn about the laws of nature, right? These [are] sort of principles of physics or something like that, that tell us about how … the world around us works. One interesting thing is that the same scientists who hundreds of years ago were trying to figure out what those laws of nature were, using math to describe the physical world, were just as interested in using math to try and understand the mental world, the world inside us. The book is really the story of that effort. It turns out understanding our inside world is a bit harder than understanding our outside world. It took us a little bit longer to figure out what the fundamental principles are. It charts the story, from people first introducing this idea of using mathematics to understand the mind, through some of the first discoveries about what kinds of mathematical principles could be used for explaining how minds work — things like mathematical logic — to the discovery that that was not going to get us all the way to understanding things like how people learn complex concepts that have fuzzy boundaries, things like languages, and then ideas like artificial neural networks, which are very popular at the moment in artificial intelligence, and then probability and statistics as another approach that really helps us understand why it is that some of those AI methods actually work. Sam Ransbotham: I think you approach this from three different frameworks: rules and symbols is one framework, neural networks is another, and then Bayesian probability is a third. Maybe I’m grossly oversimplifying these three big prongs in the book. Maybe take a minute and explain what each of those pieces are and then, more importantly, how they all weave together. Tom Griffiths: You got it. Those are the three big pieces. Basically the story is that I think the origins of people trying to think about mathematical principles for understanding the mind are really tied up in that rules and symbols approach. That was because that seemed like the first tool that we had that really described something like how thought worked. So if you go back to the origins of logic, the title of the book, The Laws of Thought, is a phrase that’s used by George Boole, who was working in the 19th century and sort of figured out some of the first principles of mathematical logic. Those principles turned into the principles that underlie our computers today, through the work of Alan Turing and John von Neumann and others. When psychologists were trying to work out how to rigorously study something that you can’t see or touch — something inside our heads, our minds — they discovered that those mathematical principles of logic were actually really useful for expressing rigorous, precise hypotheses about how minds work. So that was the starting point for what we now call cognitive science, which is trying to use these mathematical principles to figure out how minds work. For a while it seemed like that was going pretty well. It turned out that those systems of rules and symbols worked well for describing things like deductive reasoning, things like problem-solving or planning, things like even the structure of languages through the work of people like Noam Chomsky. But after a while they started to realize that maybe that wasn’t going to be all that we’d need in order to understand how minds work. One of the big problems for that rules and symbols approach was explaining learning. It helps us to explain how we reason, helps us to explain what the structures or languages are like, but it doesn’t help us to explain how those things get into our heads in the first place. How do we learn these kinds of strategies for thinking? How do we learn what the structure of those languages are? And it also didn’t work for capturing some of the … fuzzy boundaries that we see in real concepts. … If you ask people, “Is an olive a fruit?” people are quite uncertain about [the answer] in a way that’s maybe hard to capture if you’re really just thinking in terms of something like logic. In the 1960s-1970s, people started to explore different ways of thinking about the mind in terms of different kinds of mathematics, and thinking about things like maybe our concepts are related to … we can think about something in the world as being a point in space, where it’s an abstract space that picks out the features that thing has. Maybe a concept is a region in that space. Now a new kind of mathematics is needed for describing these kinds of continuous representations. It turns out that when you start thinking in those terms, you end up getting to new ways of thinking about how to solve that learning problem. And that’s where artificial neural networks come in. They’re essentially a way of thinking about how to represent things as points in space and then learn the relationships between those points in space so you can map from one space to another. So that solved a bunch of problems that we had for logic. But then we have a bunch of other questions. For example, if we look at things like our large language models [LLMs] today that are very successful in doing all sorts of things like learning language. Really understanding why it is that they’re able to do that requires us to take one more step and think about a different kind of mathematical idea. [Those] ideas come from probability and statistics. Statistics is really the science of inductive inference. It tells us what we can infer from data, and probability theory gives us a tool for understanding how we can work with uncertainty and how we can make inferences from the data that we see. Sam Ransbotham: Is there a fourth one? We’ve got three nice things, and each time you pointed out some aspect of them that [was] strong, and some aspect that led to a limitation. Is there [a] number four out there that we need that we haven’t figured out yet? Or is it just a matter of getting these three mixed in the right proportions and emphasized in the right ways? Tom Griffiths: I think these three are actually pretty good. … A funny thing that happened to me when I was writing this book is that I’ve been teaching these kinds of ideas to undergraduates for 20 years. When I give my class on computational approaches to understanding cognition, I would normally start that class by saying, “Unlike taking a class in physics or something like that, where you can expect to hear the answers, we’re still figuring these things out. We have good ways of asking the questions [even though] we haven’t quite … got to those answers.” But I actually think in the last 10 years, in the period that I was working on the book, I think there’s been a change in how much we understand about these things and how well they fit together. We can kind of start to see some glimmers of really figuring out what those laws of thought might look like. Sam Ransbotham: [Are these] laws of thought going to be understanding the borders between these better, or is it going to be some sort of complementarity between them, or some sort of combination in a unique way? Tom Griffiths: Complementarity and combination are, I think, the two ways to think about this. One thing that we’ve started to realize is that there [are] different ways that you can provide an explanation for something like the human mind. So, again, if you’re a physical scientist and you want to explain a phenomenon, say, the behavior of an animal, you could think about explaining that phenomenon at lots of different levels. You could explain it in terms of the environment that the animal is in. You could explain it in terms of the muscles and bones of the animal that are doing certain kinds of things and the nerves and so on. You could explain it in terms of the chemical reactions that are happening to produce those things, or you could explain it in terms of the atoms and molecules that are interacting. There are all of these different levels of analysis that we’re used to thinking about when we think about physical systems. One of the insights of cognitive science, something that goes back to a theoretical neuroscientist called David Marr, is the idea that there are similar kinds of levels of analysis that we can think about when we’re trying to understand something like human behavior or the behavior of an intelligent system more generally. Marr suggested you could think about this in terms of three levels. The most abstract is what he called the computational level, which is: What’s the abstract problem the system is solving, and what does the solution to that problem look like? And then, more concrete than that, there’s what’s called the algorithmic level, which is: What are the actual processes that are going on inside that system? What [are] the algorithms that are being executed to produce that solution? Then the third is what’s called the implementation level, which is: How is that algorithm implemented inside the brain? One important insight here is that these three different systems of mathematics that we’ve been talking about don’t need to be fighting with one another. They can be cooperating by giving us explanations that operate at different levels of analysis. In particular, logic and probability theory are at that most abstract level. They kind of describe how an ideal agent should solve problems that [it] faces, problems like, how do I figure out what’s true based on the things that I already know? Or what inferences can I draw from the things that I’ve seen already? The neural networks give us a story about how you can actually create systems that implement different kinds of algorithms that are strategies for approximating solutions to those more abstract, more idealized kinds of mathematical systems. So part of the reason why I think maybe three is enough is that logic tells us how to solve what we call deductive problems, problems that require figuring out what’s true when we have all the information. Probability theory tells us how to solve what are called inductive problems, problems where we don’t have all the information and we have to kind of do the best we can to figure out what to do based on what we know. And then our brains somehow solve both of those kinds of problems using things that look very much like the kinds of structures of artificial neural networks. So that combination of three things actually gives us ways of describing the abstract problems we solve as well as the kinds of physical systems that might actually implement solutions to those problems. Sam Ransbotham: Language is a pervasive idea through the book. We can give a nod to my frequent coauthor David Kiron, who’s super into [Ludwig] Wittgenstein and “If you can’t express it in language, it doesn’t exist” type of ideas. It’s a major function of the book. What’s the relationship between maybe the fact that we’re coming up with all this in a moment of English language dominance and would all these things be the same if we’d had them 200 years ago with a French language dominance? We, at the same time, have another layer of mathematical language, which is pervasive. Is coding language another one of these languages that is going to be pervasive? How do all these connections between language and intelligence link up in your mind? Tom Griffiths: Language is a recurring example because it has characteristics that line up with all of these three kinds of mathematical ways of thinking about minds. So the rules and symbols part is what you can think about in terms of a traditional … way of thinking about grammar in language: A sentence has a noun and a verb, and you know they’re combined in particular ways, and you can move them around in certain kinds of ways. That was this sort of important insight that Noam Chomsky had that really organized most of 20th century linguistics and continues to influence the way people think about these things. But that is not enough to explain everything that happens involving language. One of the challenges that Chomsky had was explaining how it is that human children come to speak language, because there wasn’t really a good way to formalize learning in that rules and symbols approach. So he ended up concluding you just had to build it all in. And then, based on the limited information you get, you’ve got enough constraints in the system that the right thing comes out. Neural networks offered a way to think about how you could learn things like language from data by showing that even if you had something that was described by a system of rules and symbols, it didn’t need to be implemented as a system of rules and symbols. Some of the key insights that came from [the] early work [of] using neural networks were that you could take grammar and you could have a neural network learn that grammar. It could do so without ever having explicitly represented a noun or a verb or any of those kinds of things. So that gives us a different way of thinking about what language is and a way of understanding how it is that languages might be learned. And then, probability theory helps us to understand how, in general, we could imagine those processes of learning working, and sort of understand what it is that they do. So if we think about what language is, it also has this inductive component, where when I say something, you’re trying to figure out what I’m trying to communicate to you, and you’re making an inference from the things that are coming out of my mouth in order to know what’s going on. And your brain is also trying to solve a prediction problem, where it’s trying to cue up, “What are the concepts that are likely to come up in this conversation next?” in a way that makes it easy for you to understand the things that we’re saying. So if you think about language as a probabilistic object, then we can understand some of the things about how it works and how it’s learned in a way that goes beyond just how we might actually be able to create something like a neural network that instantiates that language. When we look at things like large language models, I think they’re actually a great illustration of how these three things come together. First of all, in order to create these systems [that] are able to demonstrate a remarkable amount of intelligence, we need to train them on something [that] has this kind of rules and symbols structure. Large language models aren’t just trained on natural language, and they’re not just trained on English. So you get trained on English. You get trained on French. They also get trained on a large amount of computer code. They’re getting a lot of symbolic structure built into them through that training process. And that’s part of what underlies the intelligence that they manifest. They’re based on these big artificial neural networks. So that’s that second piece. That’s the reason why it’s possible to learn those things. And then the third piece is the way that they’re trained is by predicting the next token [that] is going to appear in a sequence, the next word or part of a word that they’re going to see based on all of the words, the parts of words that they’ve seen before. So that training is explicitly setting this up as a probabilistic model. But what it’s trying to do is to learn a probability distribution over sequences of tokens. And what it’s trying to do when you’re interacting with it is make inferences about what sequences of tokens you might want it to generate based on the sequences of tokens that you’ve typed into it. Sam Ransbotham: Let me push back though a little bit. … You mentioned the rules and symbols. It feels like that’s a place where these current implementations of large language models might be a bit weak. I think one of the examples you gave in your book is that we can pick up a word from the first use, or we can even make up words that should be real words, even if they aren’t really words. We do that based off of an understanding of how those symbols work and how those tokens work. … People can learn a word on their first hearing. They don’t require the 14 million images of ImageNet to learn. You kind of implied that given this large corpus of knowledge, it would extract those rules and symbols. Why not push them a little bit more and give them some rules and symbols to work with? Rather than learn about gravity, you say, “Hey, here’s gravity. This is how it works — always, not just in the four examples that you’ve seen. It works that way all the time.” Where’s that? Maybe I’ll push back a little bit to say [that] when I’ve seen some examples of some of the large language models struggling with math, they’ve struggled not with “What is 2 plus 2?” Because they’ve seen millions of examples of that. But if you take a superlong number and add it to another superlong number, you tend to get a random, superlong, other number rather than the symbolic representation. You mentioned, I think, in one of the chapters, [that] we understand numbers without ever having seen that specific number before. Is there room for more symbolic processing here, [a] more rules-based approach? Tom Griffiths: Yeah. I think you highlighted two of the important ways in which current AI systems differ from human cognition, and two of the kinds of things where we can imagine learning things from how human minds work that might make those AI systems better. So those two things [involve] generalization — being able to generalize in a systematic way beyond the data that they’re provided and learning from small amounts of data. I gave the example of kids learning language. A kid learns language on the order of 10 years, whereas the kinds of large language models that are deployed today require more on the order of 10,000 years of continuous speech or something like that in order to reach the level of competence that they reach. So those are places where we have opportunities to learn from people. The first of those, this point about generalization, is really about have you formed the right kinds of representations of a domain such that when you start to see things that go beyond the training data you’ve seen, you’re able to then respond to those in ways that are consistent with what you should have learned, in order to represent the domain that you’re operating in? That remains an outstanding problem for language models. Part of the way that they’re able to do this is they’ve been exposed to so much linguistic data that they’re able to do very well without necessarily needing to do a lot of extrapolation beyond the kinds of data that they’ve seen before. I have colleagues who have developed paradigms for measuring the extent to which they can extrapolate. They can actually do fairly well. You can do things like have them compose ideas together that they’ve not encountered before and come up with new kinds of things when you put those pieces together. But I think that’s something where there are still limitations in the systematicity of generalization. One of the things that surprises us about large language models is they sometimes behave in ways that make very little sense to us. That’s because we’re expecting them to generalize in ways that are like the ways that we’re used to human beings generalizing. The other part of that, the learning part, I think, is perhaps one of the keys to thinking about how we can make systems that generalize better. Because being able to learn from less data very much requires being able to engage in good systematic generalizations. The way we talk about that in machine learning and in cognitive science is in terms of what we call inductive bias. Inductive bias is what the learner brings to a problem. That means that they favor some solutions over others. So if you see only a limited amount of data, there are many possible ways that you could explain [the] data that you saw. How do you choose between those many possible explanations? If you’re learning a language, how do you choose which structure of the language you’re going to infer? Inductive bias is the thing that breaks ties there. It tells us, “You should think about it this way rather than this way.” Humans undoubtedly have a systematic set of inductive biases that are not instantiated in our neural networks. One of the important challenges for both AI and cognitive science is figuring out what those human inductive biases are and figuring out how to put them into things like these kinds of neural network systems. That’s something that I work on in my lab. It’s something that a lot of cognitive scientists are actively thinking about at the moment. Sam Ransbotham: One of the things I enjoyed was with each chunk in your book, you take a little bit of a historical path through how we got to this point. You mentioned before Boole and the development of Boolean logic and Bayes, and how that added to the equation. Now I’m going to be mean here and say that “OK, these are some summaries of historical information leading to a path.” My mean part is, “What did you, Tom the human, add to this book?” Were I to summarize the development of language or the development of Bayesian thought or probabilistic reasoning, and I put that in an LLM, and I ask, “Give me four paragraphs about that.” What did you the human add to this equation, to this book that would not have been done by that summarization process? Tom Griffiths: There are a couple of things. One thing is the book actually involves a lot of primary source research. The story of where the book came from is partly that I realized that our field of cognitive science is one where, at the time when I started this project, many of the people who were there at the sort of birth of modern cognitive science in the 1950s were still alive, and I was able to go around and interview them and collect their stories. So a lot of the book is really telling those stories and using those to explain where those ideas come from. But in terms of the writing, I think the thing that I am able to do as a human author is engage in what we call theory of mind in cognitive science. This is me thinking about what it is that is going to make sense to a reader and that’s going to be appealing, not just as a story but also clear in terms of conveying those ideas and putting those together in a structure that makes sense for the reader. I’m not going to claim that’s something only humans are going to be able to do forever. I think, at the moment, it’s something that still humans are better at doing than the current models that we have. But we actually have been doing experiments in my lab and showing that large language models are not bad at putting together a curriculum for people to help them learn a concept, sort of figuring out, “You need to introduce this simpler idea first and then this other idea, and then it can put them together and so on.” So they’re definitely able to extract some of the structures that we use for solving these problems through our own intuitions about pedagogy and theory of mind and so on. Sam Ransbotham: Maybe a different thing is that there [are] a lot of stories that you didn’t include in the book. There’s a lot of curation that’s happening. That’s one thing that I guess I’m thinking about a lot. We have the machines capable of [having] effectively infinite memory of all possible stories. There’s definitely value in you stitching out and saying, “These were important for this reason, and this is important for another reason,” because these were some big steps. In doing so you’ve inherently had to leave out some things. There’s a curation going on there. I feel like that’s something that you knew what to focus on that was important for the story, and maybe, like you say, the machines aren’t quite there for that. Tom Griffiths: This is actually, I think, a good connection to a question that I get asked a lot: What’s going to happen to the kinds of jobs that humans do? Because we’ve previously seen technology replace certain kinds of labor — physical labor [of] various kinds, manufacturing, things like that. And now we’re seeing machines start to replace cognitive labor, which is different. As a psychologist, one of the things that I think about is there’s another kind of labor that maybe is going to become even more important, which is metacognitive labor. This is what you’re doing as a manager when you’re thinking about, “Who is going to be the best employee to do this job, and how should I describe it to them in order to ask them to do it so that it ends up being done in a way that’s effective?” Or thinking about for yourself: “What strategy should I use to solve a problem?” and “What’s the right way to approach this problem?” What we’re starting to do with these machines is outsource the cognitive component of the work. They’re able to do some of that for us, but we’re still having to do a lot of the metacognitive part. So that curation process you’re describing is a good example of something [that] is not telling the story; it’s figuring out the structure around that. That’s going to be the analogs of the prompt that you’d be providing to the AI system to maybe tell that story for you. Sam Ransbotham: Are there other things besides metacognition? What else is in that list? Tom Griffiths: The first thing, I would say, is metacognition is quite a big item. I’m giving it the same status as physical labor and cognition. There are lots of people today whose jobs are metacognitive jobs, in management roles. And I think that’s going to be a skill set that becomes more and more important. One of the things that happens in graduate school — well at least when I work with my graduate students — is not just learning how to do research but also learning how to think about what a good research project is. What I say to my students is there’s a difference between the projects that we could do and the projects that we should do. Figuring out how to prioritize and work out what the best ideas are [is] one of the hardest things for people to learn, and that’s part of why it takes quite a long time to do a Ph.D. I think that kind of skill set is going to be one that becomes increasingly valuable as it becomes easier to execute on these kinds of things. Sam Ransbotham: One of the things that you mentioned and others have mentioned, too, is some of these constraints that are constraints on human intelligence are not the same as the constraints on machines. … What are some of those constraints [where] you see differences between machines and humans? And then what are the implications of there not being those constraints in the future? Tom Griffiths: The three that I would normally highlight are: We have limited lifespans, limited time in this world. That means limited data we can learn from, limited compute. So [we have] limited cognitive resources, because we just carry around 2 or 3 pounds of neural tissue, and we have to do everything with that. And then, [we have] limited bandwidth for communication. If I want to share some of the data or compute that I have with you, I have to do it through this very inefficient mechanism that we’re using right now of making honking noises. That set of constraints, I would say, [is] what makes human intelligence what it is, right? … We’ve evolved minds in response to those constraints. If you look at what’s going on for AI systems, really none of those things are true. We are able to turn up the knob of compute as high as it can go. It’s somewhat limited at the moment because we’re running out of money and energy resources to be able to keep building data centers. But that’s something where the expectation that we should have is that we should, over time, continue to be able to have greater compute capacity for training the systems. That translates into being able to train systems on much more data, and all of [our] breakthrough AI systems have been trained on more data than human beings will ever experience. AlphaGo [has] many human lifetimes of playing [the] game of Go, and our language models today have many human lifetimes of linguistic data. And then bandwidth: You can take one AI system that’s been trained on one set of things and then train it on some more things, or transfer the weights that it has between machines or split them up in all these ways. This idea of foundation models is that you can have one model and then copy it many times and then fine-tune those to solve different problems, and that’s just fundamentally different from humans. For those reasons, I think we’re going to see a meaningful divergence between the kinds of minds that humans are and the kinds of minds that the AI systems are. We shouldn’t expect them to be the same because they’re operating under different constraints, but we can still learn meaningful things about one another by comparing these different species when we take into account the fact that we’ve evolved in these different environments. Sam Ransbotham: I liked your “We communicate by making honking noises to each other.” Is that holding us back? Is there a need for a language 3.0 type of a thing to move us to better bandwidth? Tom Griffiths: René Descartes wrote about this idea all the way back when he was starting to think about math for the physical world. He talked about this idea that maybe there’s a similar sort of structure to language. You could imagine creating a language where just hearing somebody say something in that language, you know what that thing is. In the same way that if I say “10,000, 500, and 42,” even though that’s not a string you’ve ever heard before, it tells you exactly the thing that it’s referring to, you can sort of figure it out from the expression. [Gottfried] Leibniz was also obsessed with this idea. He had this idea that he called the universal character, which was a language in which you would be able to express things and then perform some mathematical operations on those things, and then figure out what the consequences of those things were. Just by expressing things in that language, you would know whether those things were true or false, and whether they were compatible with other things, and so on. That’s what motivated him to think about mathematical logic. There was this spirit that traced through the book in terms of thinking about what the consequences of having these kinds of mathematical formalisms for understanding thought can be. So it’s an interesting question: Can you turn all of that back around and come up with better languages for humans? There’s a little bit of work along these lines. Bean Kim and colleagues at Google have a paper [that] looks at neologisms for language models, where it’s looking at what are places where introducing a new term can help a language model capture a concept that’s relevant to the operations that model is doing but also help us understand what it is that the language model is doing? I think you can imagine building better interfaces between us and AI systems by allowing language to evolve in ways that capture concepts that are relevant to those systems. That sounds like an interesting cognitive science problem. Sam Ransbotham: We may be headed that way anyway by communicating with emojis. It doesn’t convey well on this audio format, but the chapter you’ve got about the representation of those universal symbols and how those worked, I didn’t know about that beforehand. So I was able to pick up on that. One of the things on your website says — and I’ll quote you — “It’s natural to ask, ‘What makes human intelligence special?’” So if it’s natural, let me ask it: What makes human intelligence special? Tom Griffiths: I would say those things that I mentioned, the constraints, are the things that really shape the nature of human intelligence. But I think it’s maybe a mistake to think about that as being special. Rather, maybe we should think about that as being different. I think there’s a tendency that people have when they talk about AI to think about intelligence being a one-dimensional scale. So people ask questions like, “Have we made superintelligence?” “Have we made whatever the next iteration of this is?” I think that’s a very limited way of thinking about what it is minds are and what intelligence is. I think it’s maybe better to think about intelligence as being shaped by the kinds of problems that minds have to solve and the kinds of constraints that they have to solve those problems under. That’s maybe more like a sort of evolutionary way of looking at things, where we can imagine minds being shaped by those problems and constraints. But it’s something where if we apply that to thinking about AI, we’re going to have expectations that humans and AI systems are going to be meaningfully different. There are ways that you can imagine making AI systems better by incorporating things that come from people. That’s part of what makes it exciting to think about these things from the perspective of cognitive science. But I don’t think it’s obligatory that we do that, because we could come up with completely different ways of solving those problems that make it possible for us to make AI systems that can do the things that we want AI systems to do, without them necessarily having to be exactly like us. I think when people talk about [artificial general intelligence] and wanting to make systems that are like people but better in whatever way, my reaction to that is to say, “Well, maybe we should just think about them as being different from us but having a set of capabilities that that are perhaps harmonious with and complementary to the abilities that we have as humans.” I think another thing that we should think about when we’re trying to think about these differences between humans and AI systems is exactly what it is that we want to use our AI systems for. In general, we have a higher bar for the behavior of our AI systems than we have for a human being. That’s appropriate, right? If we’re going to be deploying a system intentionally and we could make it better, we should try and make it better. I think some of these things, like reasoning and being able to solve math problems, and so on, that are within the capacity of AI systems but are in some ways a challenge for our current large language model-based systems, are places where there’s opportunities to use what people call a neuro-symbolic approach, where you build in aspects of logic or other kinds of mathematical tools that these systems can use to be able to do things that make them better than the baseline language models are. That’s an important thing to do if you want to make better AI systems. It’s not necessarily something that’s going to help us understand human minds better. Human minds are built out of the same sort of messy stuff that we’re trying to build our large language models out of and mess up in the same kinds of ways as those models when you have them solve math problems or do other kinds of things. Part of that is suggesting that, again, there’s a kind of divergence that could happen where we need to do a bunch more work to figure out how to make our AI systems reliable in ways … in order to be deployable, but we have enough of the basic principles figured out that we still have good insight into human cognition already. Sam Ransbotham: What I like about that is it speaks to the idea that LLMs, though wonderful, [are] not necessarily the end of the chapter of the book. There are whole different ways of approaching these problems that may or may not have strengths and weaknesses. Just like you mentioned with the growth of the different approaches, whether it’s symbols and rules or Bayesian logic or neural networks, we can realize some limitations and then build on them and try to combine those in different ways. So maybe there’s hope for some entirely new approaches. This is all interesting. Given these laws of thought, what should our listeners change about anything that they do tomorrow? Tom Griffiths: I think one thing is it might change the way that you think about what AI systems are doing. I think we all have a model of how minds work that’s based on interacting with other humans. And when we think about our AI systems, we think about them using the same set of tools that we use for thinking about other humans. And that can be … misleading in a few ways. One way is that we can have incorrect assumptions about how they’re going to generalize. We say, “This AI system solved this olympiad math problem that’s extremely hard.” And then if we think about a human being who is able to do that, you’d say, “Oh, they must be incredibly smart. They must be able to do all sorts of things that I can’t do.” But in fact, that’s a relatively narrow piece of the profile of these systems. In some ways, getting better at solving these kinds of problems might make them less good at solving other kinds of problems, and there’s a kind of balancing act that goes on there. So when you try and make generalizations [such as], “If this kind of machine can solve this problem, it’s going to be able to solve this other problem,” I think if you approach it as not something that’s like us but rather something that’s shaped by the way in which it’s been trained and the constraints that it operates under, and all these other kinds of things, our expectations about how the systems are going to generalize would be different. So that might make you a little more pessimistic about timelines for building [artificial general intelligence], right? We shouldn’t make the same generalizations from the peaks, covering the entire surface of the abilities of these systems. I think the other thing it might do is help us to imagine futures that are perhaps less scary in terms of the way that we imagine these systems affecting human societies. If we start thinking about AI systems as being different from us, then that suggests this view of complementarity, where there are going to be things that you’re going to be better at, and there are going to be things the AI is going to be better at. Just like you try and figure out how to divide jobs up across people who have different kinds of abilities, thinking about how it is that we’re going to divide the kinds of things that we want to be able to accomplish between humans who can do certain kinds of things and AI systems that can do other kinds of things is maybe a healthier way of thinking about this than sort of imagining that we’re going to be completely replaced. Sam Ransbotham: It’s fascinating talking to you. We’ve been talking about the Laws of Thought, which is coming out [on] Feb. 10. Tom, thanks for taking the time to talk with us today. Tom Griffiths: Thank you. Sam Ransbotham: Thanks for listening today. On our next episode, I’ll speak with Nobel Prize-winning economist Daron Acemoglu about his research at MIT. Please join us. Allison Ryder: Thanks for listening to Me, Myself, and AI. Our show is able to continue, in large part, due to listener support. Your streams and downloads make a big difference. If you have a moment, please consider leaving us an Apple Podcasts review or a rating on Spotify. And share our show with others you think might find it interesting and helpful.

Rate this article

Login to rate this article

Comments

Please login to comment

No comments yet. Be the first to comment!
    Tom Griffiths on AI & Language: Princeton Insights