Jay McClelland
Language in an Integrated Understanding System
Speaker
-
Jay McClelland
Jay McClelland
Jay McClelland uses neural network models to capture human cognitive abilities including language understanding, memory, and mathematical cognition. With David Rumelhart he co-led the project resulting in Parallel Distributed Processing, a two-volume work presenting these models in the mid-1980’s. Today he is a professor in the Psychology and (by courtesy) linguistics Departments at Stanford, where he is the director of the Center for Mind, Brain, computation and Technology.
Abstract →
Jay McClelland
Language in an Integrated Understanding System
I consider language from the perspective that it is a part of a system for understanding and communicating about situations. In humans, this ability emerges gradually from experience and depends on domain-general principles of biological neural networks: connection-based learning, distributed representation, and context-sensitive, mutual constraint satisfaction-based processing. Current artificial language processing systems rely on the same domain general principles, embodied in artificial neural networks. Indeed, recent progress in this field depends on query-based attention, which extends the ability of these systems to exploit context and has contributed to remarkable breakthroughs. Nevertheless, most current models focus exclusively on language-internal tasks, limiting their ability to perform tasks that depend on understanding situations. These systems also lack memory for the contents of prior situations outside of a fixed contextual span. I will describe the organization of the brain’s distributed understanding system, which includes a fast learning system that addresses the memory problem. I will then sketch a framework for future models of understanding drawing equally on cognitive neuroscience and artificial intelligence and exploiting query-based attention. Finally, I will highlight relevant current directions and consider further developments needed to fully capture human-level language understanding in a computational system.