Learning World Dynamics With Structured Representations in a Computational Model of Event Cognition
AbstractThe segmentation of the sensory stream into discrete events is a central process in cognition, with implications for memory, learning, and attention. Event cognition can be modeled computationally as the process of inferring underlying events from an unlabeled sequence of scenes, learning the dynamics of different classes of events using the event schema abstraction, and using trained event schemata to predict future scenes and recall past events. In this paper we explore the problem of representing scenes and event schemata so that an event schema can be trained on sequences of generated scenes to learn the dynamics of a single class of events, providing a benchmark for the segmentation process. In particular, scenes are encoded with structured representations to capture their semantic content, allowing for the decodable vectorization of complex semantic relationships. Event schemata, which are transition functions that learn event dynamics, are represented as recurrent neural networks that operate on sequences of observed scene vectors to predict the next observed scene. This paper addresses problems in scene representation, event schema architecture, and training methods, arriving upon a framework for learning world dynamics that can be used as a foundation for event segmentation models.
Citable link to this pagehttp://nrs.harvard.edu/urn-3:HUL.InstRepos:38811530
- FAS Theses and Dissertations