REVIEW of ChatGPT and the Future of AI: The Deep Language Revolution by T. Sejnowski – AI-Human convergences and divergences
Nigel P. Daly, PhD 戴 禮
Coaching Communication-and-Language Fitness (CLF) + Performance enhancing AI ??+??=? | Writer | Applied Linguistics Researcher
ChatGPT and the Future of AI: The Deep Language Revolution? emerges from the computational neuroscientist Terrence Sejnowski’s interest in Large Language Models (LLMs) like ChatGPT and their potential to reshape human productivity, creativity, learning, and cognition. The subtitle?"The Deep Language Revolution"?highlights how advances in deep learning, particularly transformer-based architectures, have enabled LLMs like ChatGPT to revolutionize human-machine interaction through their mastery of language. Sejnowksi uses the term Deep language revolution to underscore the profound societal and cultural impacts of these technologies, from transforming industries and education to raising ethical questions about their role in reshaping communication and cognition.?He makes the compelling analogy between the conceptual leaps enabled by LLMs and the discovery of the DNA double helix: AI as a technology could bridge the kind of gaps in understanding intelligence that DNA has done in biology.
Why this book is important??
This book is extremely useful and insightful for educators, technologists, and policymakers navigating AI’s integration into daily life. Because the author is an expert in computation and neuroscience, he is able to demystify the capabilities of both LLMs and brains, and he has the rare skillset to explain both fields and their intersections with rare clarity and rigor.
Terrence Sejnowski is a renowned computational neuroscientist and AI researcher who has pioneered work in computational neuroscience and neural networks. I rediscovered him 2 weeks when I listened to him as a guest on the Huberman Podcast (Nov. 19, 2024) as part of his promotion efforts for this book that was released at the end of October 2024. Sejnowski, a professor at the Salk Institute and co-founder of the Neural Information Processing Systems (NeurIPS) conference, has consistently bridged the gap between biological brains and artificial intelligence. The first time I discovered him was when I read his collaboration with Barbara Oakley to write the 2021 book entitled Uncommon sense teaching—Practical insights in brain science to help students learn and develop a hugely popular course on Coursera called “Learning how to learn”.?Sejnowksi is therefore uniquely positioned to offer insights in to computation, neuroscience, learning, and education.
Some unique features of the book??
Extensive Use of ChatGPT: Sejnowski frequently integrates ChatGPT sections as a simplifier of complex technical concepts and also a chapter summarizer. This not only helps reinforce key points but also makes the book an interactive demonstration of LLM utility.?In fact, he playfully acknowledges ChatGPT as a co-author.
Interdisciplinary Analysis: The book bridges neuroscience, cognitive science, and AI computer science and in so doing, explores the technical and philosophical overlaps between human cognition and artificial intelligence.?
Focus on Prompt Engineering: The book uniquely highlights the importance of crafting detailed prompts. This is a critical skill to maximize LLM efficiency and parallels the evolution of programming in earlier technological revolutions.?
Summary of the Three Parts?
The book contains 14 chapters over 3 parts: Living with large language models,? Transformers, and Back to the future. The following lists some key insights from each part.?
Part I: Living with Large Language Models?
Part II: Transformers?
Sejnowski explains transformers as the deep learning architecture central to LLMs. They enable models to process and generate language by attending to relationships between words in a sequence, regardless of their position. This architecture, which uses self-attention mechanisms, allows LLMs to scale effectively, generalize across tasks, and exhibit emergent capabilities like reasoning and creative problem-solving (Vaswani et al., 2017).?
Part III: Back to the Future?
LLMs and Human Cognition: Similarities and Differences?
Similarities
Both human brains and generative AI systems like Large Language Models (LLMs) share the ability to generalize across complex inputs and create novel outputs. For example, LLMs learn patterns from vast datasets in a manner akin to humans recognizing linguistic structures through exposure. A striking parallel lies in their reliance on high-dimensional processing: LLMs operate in vast parameter spaces, with GPT-4 utilizing over a trillion parameters, comparable to the 86 billion neurons and trillions of synaptic connections in the human brain. Furthermore, LLMs act as mirrors to reflect the knowledge, biases, and intent of their users—an analogy that aligns with how humans adapt their social cognition based on environmental cues. This mirror effect is illustrated through the "reverse Turing test," where the intelligence of LLM outputs often reflects the sophistication of the user’s prompts.?
领英推荐
There are compelling parallels between human language production and how LLMs generate text, which suggests shared underlying mechanisms despite vastly different substrates. Both systems rely on probabilistic models: humans subconsciously predict and select words based on linguistic context, while LLMs calculate probabilities using vast training data to determine the most likely next word. For example, LLMs like GPT-4 excel at producing coherent, contextually appropriate language by leveraging transformer architectures, which mimic human brain processes like working memory and attention to prioritize relevant information in a sequence (I went into fine detail on this here).?
Sejnowski also notes that both humans and LLMs exhibit creativity through recombination of learned elements. In humans, this ability comes in the form of metaphors, analogies, and novel expressions, while LLMs achieve similar results by blending patterns from diverse training data. A good example is ChatGPT’s ability to produce poetry and complex narratives, which reflects a capacity for structured, meaningful output that closely mirrors human linguistic creativity. However, while humans produce language grounded in embodied experience and social context, LLMs operate without true understanding, relying solely on statistical correlations?
Differences?
Despite these similarities, there are fundamental differences that underscore the limitations of LLMs compared to human cognition. Unlike humans, who learn through embodied experiences involving sensory input and emotional/social context, LLMs are "brains in a vat," trained on static text without real-world interaction. This lack of embodiment restricts their understanding of concepts tied to physical experiences. Additionally, human learning is multimodal and develops over years through trial, error, and adaptation, while LLMs rely on massive computational training upfront. For instance, training GPT-3 required approximately $12 million in computational resources, highlighting a resource-intensive process that contrasts with the efficiency of human development. Lastly, humans possess metacognition—the ability to think about their own thinking—whereas LLMs simulate understanding but lack genuine self-awareness or introspection, limiting their autonomy and adaptability.?
Takeaways for Education
1.?Personalized Learning Companions?
2.?Immersive Storytelling for Language and History?
?
3.?AI-Powered Peer Review and Feedback?
Takeaways for the Workplace
?AI as a Real-Time Communication Coach?
2.?Dynamic Idea Generation and Problem Solving?
3.?Enhanced Employee Onboarding and Training?
?
PhD Candidate in Rhetoric, Theory and Culture at North Dakota State University. A bilingual writer, columnist, rhetorician and poet. AI-driven Pedagogy. Gender, Body, Sexuality, Disability.
3 个月Excellent review! Thanks for introducing the book!
Ancora Imparo
3 个月Excellent introduction. Thank you!