Publications by authors named "Georg Ostrovski"

Article Synopsis
  • The paper introduces a new approach to analyzing asymmetric games involving two populations, showing that they can be simplified into two single-population symmetric games.
  • It discusses how the asymmetric bimatrix game can be broken down into symmetric counterparts by studying the separate payoff tables, which leads to easier analysis and understanding.
  • A key finding is that Nash equilibria in the original asymmetric game correspond to specific Nash equilibria in the symmetric games, providing insights into their evolutionary dynamics through simpler examples.
View Article and Find Full Text PDF

Artificial neural networks are remarkably adept at sensory processing, sequence learning and reinforcement learning, but are limited in their ability to represent variables and data structures and to store data over long timescales, owing to the lack of an external memory. Here we introduce a machine learning model called a differentiable neural computer (DNC), which consists of a neural network that can read from and write to an external memory matrix, analogous to the random-access memory in a conventional computer. Like a conventional computer, it can use its memory to represent and manipulate complex data structures, but, like a neural network, it can learn to do so from data.

View Article and Find Full Text PDF

The theory of reinforcement learning provides a normative account, deeply rooted in psychological and neuroscientific perspectives on animal behaviour, of how agents may optimize their control of an environment. To use reinforcement learning successfully in situations approaching real-world complexity, however, agents are confronted with a difficult task: they must derive efficient representations of the environment from high-dimensional sensory inputs, and use these to generalize past experience to new situations. Remarkably, humans and other animals seem to solve this problem through a harmonious combination of reinforcement learning and hierarchical sensory processing systems, the former evidenced by a wealth of neural data revealing notable parallels between the phasic signals emitted by dopaminergic neurons and temporal difference reinforcement learning algorithms.

View Article and Find Full Text PDF