Publications by authors named "Susanne Still"

Information engines can convert thermal fluctuations of a bath at temperature T into work at rates of order k_{B}T per relaxation time of the system. We show experimentally that such engines, when in contact with a bath that is out of equilibrium, can extract much more work. We place a heavy, micron-scale bead in a harmonic potential that ratchets up to capture favorable fluctuations.

View Article and Find Full Text PDF

A 1929 Gedankenexperiment proposed by Szilárd, often referred to as "Szilárd's engine", has served as a foundation for computing fundamental thermodynamic bounds to information processing. While Szilárd's original box could be partitioned into two halves and contains one gas molecule, we calculate here the maximal average work that can be extracted in a system with N particles and q partitions, given an observer which counts the molecules in each partition, and given a work extraction mechanism that is limited to pressure equalization. We find that the average extracted work is proportional to the mutual information between the one-particle position and the vector containing the counts of how many particles are in each partition.

View Article and Find Full Text PDF

This Letter exposes a tight connection between the thermodynamic efficiency of information processing and predictive inference. A generalized lower bound on dissipation is derived for partially observable information engines which are allowed to use temperature differences. It is shown that the retention of irrelevant information limits efficiency.

View Article and Find Full Text PDF

The difficulty of obtaining appreciable quantities of biologically important molecules in thermodynamic equilibrium has long been identified as an obstacle to life's emergence, and determining the specific nonequilibrium conditions that might have given rise to life is challenging. To address these issues, we investigate how the concentrations of life's building blocks change as a function of the distance from equilibrium on average, in two example settings: (i) the synthesis of heavy amino acids and (ii) their polymerization into peptides. We find that relative concentrations of the heaviest amino acids can be boosted by four orders of magnitude, and concentrations of the longest peptide chains can be increased by hundreds of orders of magnitude.

View Article and Find Full Text PDF

Recently proposed information-exploiting systems extract work from a single heat bath by using temporal correlations on an input tape. We study how enforcing time-continuous dynamics, which is necessary to ensure that the device is physically realizable, constrains possible designs and drastically diminishes efficiency. We show that these problems can be circumvented by means of applying an external, time-varying protocol, which turns the device from a "passive," free-running machine into an "actively" driven one.

View Article and Find Full Text PDF

A system responding to a stochastic driving signal can be interpreted as computing, by means of its dynamics, an implicit model of the environmental variables. The system's state retains information about past environmental fluctuations, and a fraction of this information is predictive of future ones. The remaining nonpredictive information reflects model complexity that does not improve predictive power, and thus represents the ineffectiveness of the model.

View Article and Find Full Text PDF

We provide a fresh look at the problem of exploration in reinforcement learning, drawing on ideas from information theory. First, we show that Boltzmann-style exploration, one of the main exploration methods used in reinforcement learning, is optimal from an information-theoretic point of view, in that it optimally trades expected return for the coding cost of the policy. Second, we address the problem of curiosity-driven learning.

View Article and Find Full Text PDF

We introduce an approach to inferring the causal architecture of stochastic dynamical systems that extends rate-distortion theory to use causal shielding--a natural principle of learning. We study two distinct cases of causal inference: optimal causal filtering and optimal causal estimation. Filtering corresponds to the ideal case in which the probability distribution of measurement sequences is known, giving a principled method to approximate a system's causal structure at a desired level of representation.

View Article and Find Full Text PDF

We present a neuromorphic pattern generator for controlling the walking gaits of four-legged robots which is inspired by central pattern generators found in the nervous system and which is implemented as a very large scale integrated (VLSI) chip. The chip contains oscillator circuits that mimic the output of motor neurons in a strongly simplified way. We show that four coupled oscillators can produce rhythmic patterns with phase relationships that are appropriate to generate all four-legged animal walking gaits.

View Article and Find Full Text PDF

Clustering provides a common means of identifying structure in complex data, and there is renewed interest in clustering as a tool for the analysis of large data sets in many fields. A natural question is how many clusters are appropriate for the description of a given system. Traditional approaches to this problem are based on either a framework in which clusters of a particular shape are assumed as a model of the system or on a two-step procedure in which a clustering criterion determines the optimal assignments for a given number of clusters and a separate criterion measures the goodness of the classification to determine the number of clusters.

View Article and Find Full Text PDF

Entropy and information provide natural measures of correlation among elements in a network. We construct here the information theoretic analog of connected correlation functions: irreducible N-point correlation is measured by a decrease in entropy for the joint distribution of N variables relative to the maximum entropy allowed by all the observed N-1 variable distributions. We calculate the "connected information" terms for several examples and show that it also enables the decomposition of the information that is carried by a population of elements about an outside source.

View Article and Find Full Text PDF