Annurev Neuro 062012 170325 2
Annurev Neuro 062012 170325 2
ANNUAL
REVIEWS Further Decoding Neural
Click here for quick links to
Annual Reviews content online,
including:
Representational Spaces Using
• Other articles in this volume
• Top cited articles
• Top downloaded articles
Multivariate Pattern Analysis
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
1
Department of Psychological and Brain Sciences, Center for Cognitive Neuroscience,
Dartmouth College, Hanover, New Hampshire 03755;
email: james.v.haxby@dartmouth.edu, andrew.c.connolly@dartmouth.edu,
swaroopgj@gmail.com
2
Center for Mind/Brain Sciences (CIMeC), University of Trento, Rovereto,
Trentino 38068, Italy
435
NE37CH22-Haxby ARI 30 June 2014 9:14
Contents
INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 436
CORE CONCEPT: REPRESENTATIONAL SPACES . . . . . . . . . . . . . . . . . . . . . . . . . . . 437
MULTIVARIATE PATTERN CLASSIFICATION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 439
REPRESENTATIONAL SIMILARITY ANALYSIS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 442
BUILDING A COMMON MODEL OF A NEURAL
REPRESENTATIONAL SPACE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 447
STIMULUS-MODEL-BASED ENCODING AND DECODING . . . . . . . . . . . . . . . . . 449
MULTIPLEXED TOPOGRAPHIES FOR POPULATION RESPONSES . . . . . . . . . 450
FUTURE DIRECTIONS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 452
Individual and Group Differences . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 452
Between-Area Transformations in a Processing Pathway . . . . . . . . . . . . . . . . . . . . . . . . . . 453
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
INTRODUCTION
Information is encoded in patterns of neural activity. This information can come from our ex-
perience of the world or can be generated by thinking. One of the great challenges for sys-
tems neuroscience is to break this code. Developing algorithms for decoding neural activity in-
volves many modalities of measurement—including single-unit recording, electrocorticography
(ECoG), electro- and magnetoencephalography (EEG and MEG), and functional magnetic res-
onance imaging (fMRI)—in various species. All decoding methods are multivariate analyses of
brain activity patterns that are distributed across neurons or cortical regions. These methods are
referred to generally as multivariate pattern analysis (MVPA). This review focuses on the progress
made in the past decade and a half in the development of methods for decoding human neural ac-
tivity as measured with fMRI. We make occasional references to decoding analyses of single-unit
recording data in monkeys and of ECoG and MEG data in humans to illustrate the general utility
of decoding methods and to indicate the potential for multimodal decoding.
Prior to the discovery that within-area patterns of response in fMRI carried information that
afforded decoding of stimulus distinctions (Haxby et al. 2001, Cox & Savoy 2003, Haxby 2012), it
MEG: magnetoen- was generally believed that the spatial resolution of fMRI allowed investigators to ask only which
cephalography task or stimulus activated a region globally. Thus, fMRI studies focused on associating brain
fMRI: functional regions with functions. A region’s function was identified by determining which task activated it
magnetic resonance most strongly. The introduction of decoding using MVPA has revolutionized fMRI research by
imaging changing the questions that are asked. Instead of asking what a region’s function is, in terms of a
Multivariate pattern single brain state associated with global activity, fMRI investigators can now ask what information
analysis (MVPA): is represented in a region, in terms of brain states associated with distinct patterns of activity, and
analysis of brain how that information is encoded and organized.
activity patterns with
Multivariate pattern (MVP) classification distinguishes patterns of neural activity associated
methods such as
pattern classification, with different stimuli or cognitive states. The first demonstrations of MVP classification showed
RSA, hyperalignment, that different high-level visual stimulus categories (faces, animals, and objects) were associated
or stimulus-model- with distinct patterns of brain activity in the ventral object vision pathway (Haxby et al. 2001, Cox
based encoding and & Savoy 2003). Subsequent work has shown that MVP classification can also distinguish many
decoding
other brain states, for example low-level visual features in the early visual cortex (Haynes & Rees
2005, Kamitani & Tong 2005) and auditory stimuli in the auditory cortex (Formisano et al. 2008,
REPRESENTATIONAL SPACE
Representational space is a high-dimensional space in which each neural response or stimulus is expressed as a
vector with different values for each dimension. In a neural representational space, each pattern feature is a measure
of local activity, such as a voxel or a single neuron. In a stimulus representational space, each feature is a stimulus
attribute, such as a physical attribute or semantic label.
Staeren et al. 2009), as well as more abstract brain states such as intentions (Haynes et al. 2007,
Soon et al. 2008) and the contents of working memory (Harrison & Tong 2009).
Whereas MVP classification simply demonstrates reliable distinctions among brain states, more
recently introduced methods characterize how these brain states are organized. Representational
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
similarity analysis (RSA) (Kriegeskorte et al. 2008a) analyzes the geometry of representations
in terms of the similarities among brain states. RSA can show that the representations of the
same set of stimuli in two brain regions have a different structure (Kriegeskorte et al. 2008a,b;
by Kansas State University on 07/19/14. For personal use only.
Connolly et al. 2012a,b), whereas MVP classification may find that the classification accuracy is
equivalent in those regions. Stimulus-model-based encoding and decoding algorithms show that
brain activity patterns can be related to the constituent features of stimuli or cognitive states. This
innovation affords predictions of patterns of brain response to novel stimuli based on their features
(Kay et al. 2008, Mitchell et al. 2008). It also affords reconstruction of stimuli from brain activity
patterns based on predictions of the stimulus features (Miyawaki et al. 2008, Naselaris et al. 2009,
Nishimoto et al. 2011, Horikawa et al. 2013).
Several excellent reviews have focused on MVP classification (Norman et al. 2006, Haynes &
Rees 2006, O’Toole et al. 2007, Pereira et al. 2009, Tong & Pratte 2012), RSA (Kriegeskorte
& Kievet 2013), or stimulus-model-based encoding and decoding (Naselaris et al. 2011). Here
we integrate neural decoding methods into a common framework organized around the concept Representational
of high-dimensional representational spaces (see sidebar). In all these methods, brain activity similarity analysis
(RSA): analysis of the
patterns are analyzed as vectors in high-dimensional representational spaces. Neural decoding pattern of similarities
then analyzes these spaces in terms of (a) reliably distinctive locations of pattern response vectors among response
(MVP classification), (b) the proximity of these vectors to each other (RSA), or (c) mapping of vectors
vectors from one representational space to another—from one subject’s neural representational Response vector:
space to a model space that is common across subjects (hyperalignment) or from stimulus feature a brain activity pattern
spaces to neural spaces (stimulus-model-based encoding). expressed as the
response strengths for
features of that
pattern, e.g., voxels,
CORE CONCEPT: REPRESENTATIONAL SPACES
single neurons, or
The core concept that underlies neural decoding and encoding analyses is that of high-dimensional model dimensions
representational vector spaces. Neural responses—brain activity patterns—are analyzed as vectors Hyperalignment:
in a neural representational space. Brain activity patterns are distributed in space and time. The transformation of
elements, or features, of these patterns are local measures of activity, and each of these local individual
representational spaces
measures is a dimension in the representational space. Thus, if neural responses measured with
into a model
fMRI have 1,000 voxels, the representational space is 1,000-dimensional. If a population response representational space
has spike rates for 600 cells, the representational space is 600-dimensional. If fMRI responses with in which each
1,000 voxels include six time points, the response vectors are analyzed in a 6,000-dimensional space. dimension has a
For fMRI, measures of local activity are usually voxels (volume elements in brain images), but common tuning
function
there are numerous alternatives, such as nodes on the cortical surface, the average signal for an
area, a principal or independent component, or a measure of functional connectivity between a pair
of locations. For single-unit recording, local measures can be single-neuron spike rates, multiple-
unit spike rates, or local field potentials, among other possibilities. Similarly, EEG and MEG
responses can be analyzed as time-varying patterns of activity distributed over sensors or sources,
Machine learning:
a branch of artificial with numerous possibilities for converting activity into frequencies, principal or independent
intelligence that builds components, or measures of synchrony between sources.
and evaluates The computational advantages of representational vector spaces extend beyond neural repre-
induction algorithms sentational spaces to representational spaces for stimuli or cognitive states. For example, a visual
that learn data patterns
stimulus can be modeled as a set of features based on response properties of neurons in V1, as
and associate them
with labels higher-order visual features, or as a set of semantic labels. Sounds—voices and music—can be mod-
eled as sets of acoustic features, words can be modeled as sets of semantic features, actions as sets
Pattern feature:
a single element in a of movement and goal features, etc. Once the description of the stimulus is in a representational
distributed pattern, space, various computational manipulations can be applied for relating stimulus representational
such as a voxel, a single spaces to neural representational spaces.
neuron, or a model
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
All the major varieties of neural decoding and encoding analyses follow from this conversion
space dimension
of patterns of brain activity or stimuli to single points in high-dimensional representational vector
Tuning function: the spaces. MVP classification uses machine learning methods to define decision boundaries in a neu-
by Kansas State University on 07/19/14. For personal use only.
profile of differential
ral representational space that best distinguish a set of response vectors for one brain state from
responses to stimuli or
brain states for a single others. RSA analyzes the similarity between response vectors as distances in the representational
pattern feature space. Stimulus-model-based encoding predicts the location of the neural response vector for a
new stimulus on the basis of the coordinates of that stimulus in a stimulus feature space. Stimulus-
model-based decoding tests whether the encoding-based prediction allows correct classification
of neural response vectors to new stimuli. Building a model of a neural representational space
that is common across brains requires hyperalignment to rotate the coordinate axes of individual
representational spaces to minimize the difference in the locations of response vectors for the
same stimuli. Thus, stimuli and other cognitive events are represented as vectors in neural rep-
resentational spaces as well as in stimulus representational spaces, and the computational task for
understanding representation becomes one of characterizing the geometries within spaces and
relating the geometries of these spaces to each other.
Numerically, a set of response vectors in a representational space is a matrix in which each col-
umn is a local pattern feature (e.g., voxel) and each row is a response vector (Figure 1). The values
in each column reflect the differential responses of that pattern feature to conditions or stimuli.
This profile of differential responses is called the tuning function. All the neural decoding and
encoding methods can be understood in terms of analyzing or manipulating the geometry of the
response vectors in a high-dimensional space. Computationally, these analyses and manipulations
are performed using linear algebra. Here we illustrate the concepts related to high-dimensional
representational spaces in two-dimensional figures by showing only two dimensions at a time—
the equivalent of a two-voxel brain or a two-neuron population. The linear algebra for these
two-dimensional toy examples is the same as the linear algebra for representational spaces with
many more dimensions and larger matrices. Most of the algorithms that we discuss here can
be implemented using PyMVPA (http://www.pymvpa.org; Hanke et al. 2009), a Python-based
software platform that includes tutorials and sample data sets.
The geometries in a neural representational space, as defined here, are distinctly different
from the geometries of cortical anatomy and of cortical topographies. A cortical topography
can be thought of as a two-dimensional manifold. Neural encoding can be thought of as the
problem of projecting high-dimensional representations into this low-dimensional topography.
Decoding is the problem of projecting a neural response in a two-dimensional topography into
a high-dimensional representational space. The methods that we discuss here are computational
algorithms that attempt to model these transformations.
Voxel y
Pattern features
-1.6
16
1.6
Conditions
Voxel x
Flatten
-1.6
-1.7
1.7
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
x y
by Kansas State University on 07/19/14. For personal use only.
Figure 1
Multivariate pattern analysis (MVPA) is a family of methods that treats the measured fMRI signal as a set of pattern vectors stored in an
N × M matrix with N observations (e.g., stimulus conditions, time points) and M features (e.g., voxels, cortical surface nodes) define an
M-dimensional vector space. The goal of MVPA analyses is to analyze the structure of these high-dimensional representational spaces.
In a neural representational space, brain responses are vectorized, thus discarding the spatial re-
lationships among cortical locations and the temporal relationships among time points. Thus, the
approaches that we present here do not attempt to model the spatial structure of cortical topogra-
phies or how high-dimensional functional representations are packed into these topographies. An
approach to modeling cortical topographies based on the principle of spatial continuity of func-
tion can be found in work by Aflalo & Graziano (2006, 2011) and Graziano & Aflalo (2007a,b).
Anatomy and, in particular, cortical topography are important aspects of neural representation, of
course. Although decoding methods discard anatomical and topographic information when brain
responses are analyzed in high-dimensional representational spaces, the anatomical location of
a representational space can be investigated using searchlight analyses (Kriegeskorte et al. 2006,
Chen et al. 2011, Oosterhof et al. 2011), and the topographic organization of that representation
can be recovered by projecting response vectors and linear discriminants from a common model
representational space into individual subjects’ topographies (Haxby et al. 2011).
Voxel y Predictions
(e.g., stimuli or time points)
-1.
-1
1.
1.7
7 ? ?
Conditions
Voxel x
Targets
0
?
-1
-1.
-1.7
1.7
x y 0
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Testing data
by Kansas State University on 07/19/14. For personal use only.
?
?
? -1.
17
x y
Figure 2
MVP classification analyses involve partitioning data matrices into different sets for training and testing a pattern classifier. A classifier
is trained to designate sectors of the vector space to the labels provided for the samples in the training set. Test samples are then
classified as belonging to the labeled class associated with the sector in which they reside. Classification accuracy is measured as the
proportion of predicted labels that match the actual label (target) for each test item. A confusion matrix provides information about the
patterns of correct classifications (on the diagonal) and misclassifications (off diagonal).
An MVP classification analysis begins with dividing the data into independent training and test
data sets (Figure 2). The decision rules that determine the confines of each class of neural response
vectors are developed on training data. The border between sectors for different conditions is called
a decision surface. The validity of the classifier is then tested on the independent test data. For
Test data: the data valid generalization testing, the test data must play no role in the development of the classifier,
set used to test the including data preprocessing (Kriegeskorte et al. 2009). Each test data response vector is then
validity of a decision classified as another exemplar of the condition associated with the sector in which it is located.
rule that was derived
Classifier accuracy is the percentage of test vectors that are correctly classified. A more revealing
on training data
assessment of classifier performance is afforded by examining the confusion matrix. A confusion
Training data: the
matrix presents the frequencies for all classifications of each experimental condition, including the
portion of a data set
that is used to derive details about misclassifications. Examination of misclassifications adds information about which
the decision rule for conditions are most distinct and which are more similar. This information is analyzed using
pattern classification additional methods in RSA (see next section). Examination of classification accuracy for each
Decision surface: a condition separately can alert the investigator to whether average accuracy is really dependent on
surface that defines the a small number of conditions, rather than an accurate reflection of performance across all or most
boundary between conditions. Thus, average classification accuracy is a useful metric but discards information that
sectors in a
can be discovered by examining the classification confusion matrix. Confusion matrices are shown
representational space
and is used to classify in Figure 3 for two category perception experiments (Haxby et al. 2011; Connolly et al. 2012a,b).
vectors From the first experiment, on the perception of faces and objects, the confusion matrix reveals
that if misclassified, faces are classified as other faces and objects as other objects. Moreover,
the classifier cannot distinguish female from male faces. From the second experiment, on the
100
90
80
70
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Stimulus category
Percent classifications
60
by Kansas State University on 07/19/14. For personal use only.
50
40
30
20
10
Figure 3
Confusion matrices for two experiments that measured responses to visual objects in human ventral temporal (VT) cortex. The patterns
of misclassifications show that when items are misclassified they are more likely to be confused with items from the same superordinate
category: faces and small objects (top); and primates, birds, and bugs (bottom). Classification performed on data matrices from the same
subject (i.e., the same set of features) produces higher overall accuracies (a) than does between-subject classification (BSC) (b) where the
features (voxels) have been aligned on the basis of a standard anatomical template.
perception of animals, the confusion matrix reveals that misclassifications are usually within animal
class (primates, birds, and insects) and rarely across classes.
MVP classification uses machine learning classifiers to develop the decision rules. In general,
different classifiers produce similar results, and some classifiers tend to perform better than others.
A seminal MVP classification study (Haxby et al. 2001) used a one-nearest-neighbor classifier that
classified a test response vector as the category for the training data vector that was closest in the
neural representational space. Distance between vectors was measured using correlation, which is
the cosine of the angle between mean-centered vectors. Because a single vector was used for each
class in the training data—the mean pattern for each class in half the data—the decision surfaces
were linear hyperplanes separating each pair of unique classes. Nearest-neighbor methods are
fast and conceptually clear, but most have found that other classifiers provide slightly higher
Between-subject
classification (BSC): accuracies. Cox & Savoy (2003) were the first to use support vector machine (SVM) (Cortez &
classification of a Vapnik 1995) classifiers for fMRI. SVM classifiers fine-tune the position of the decision surface
subject’s response on the basis of the vectors that are closest to the surface, i.e., the support vectors, by maximizing
vectors based on a the distances from the surface to these borderline cases. Other regression-based methods, such as
classifier built on other
linear discriminant analysis (LDA) (e.g., Carlson et al. 2003, O’Toole et al. 2005), are also effective
subjects’ data
and can include regularization methods for selecting features, such as sparse multinomial logistic
Within-subject
regression (SMLR) (Yamashita et al. 2008). Most MVP classification analyses have used linear
classification (WSC):
classification of a classifiers—meaning that the decision surface is planar—some for theoretically driven reasons
subject’s response (Kamitani & Tong 2005) but mostly for simplicity and to avoid overfitting the noise in the training
vectors based on a data, which leads to larger performance decrements in generalization testing.
classifier built on that
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Until recently, almost all MVP classification analyses had built a new classifier for each indi-
subject’s own data
vidual brain. Cox & Savoy (2003) showed that classifier performance dropped drastically if based
Dissimilarity matrix on other subjects’ data. The performance decrement for between-subject classification (BSC) rel-
by Kansas State University on 07/19/14. For personal use only.
θ3
1-cosθ2
Voxel x
1-cosθ3
θ1 θ2
1-cosθ1
(e.g., stimuli or time points)
Pattern features
(e.g., voxels)
Conditions
Euclidean distance
Voxel y
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
d3
x y
by Kansas State University on 07/19/14. For personal use only.
Voxel x
d2 d1 d3
d1
d2
Feature-based Feature-independent
representational space representational space
Figure 4
Representational similarity analysis examines the patterns of distances between vectors in the high-dimensional vector space. Measures
of angular similarity such as cosine and Pearson product-moment correlation are standard measures that are most sensitive to the
relative contributions of feature dimensions. These similarity measures are transformed into dissimilarities by subtracting them from 1.
Another standard measure of the distance between vectors is Euclidean distance, which is more sensitive to overall differences in vector
length or magnitude.
Investigation of the similarity of neural representations from fMRI data dates back to an early
paper by Edelman et al. (1998), which was the first to use multidimensional scaling to visualize the
representational space for visual objects. Two groups reanalyzed data from an early MVP classifi-
cation study on the distributed representation of faces and objects (Haxby et al. 2001; reanalyzed
by Hanson et al. 2004; O’Toole et al. 2005, 2007) and found similar similarity structures using
different distance measures, one based on intermediate-layer weights in a neural network classifier
and the other based on misclassifications. Both found that the strongest dissimilarity was between
the animate (human faces and cats) and inanimate (houses, chairs, shoes, bottles, and scissors)
categories and, within the inanimate domain, a strong dissimilarity between houses and all the
smaller objects. These basic elements of the structure of face, animal, and object representations
were corroborated and greatly amplified by subsequent studies in monkeys (Kiani et al. 2007) and
humans (Kriegeskorte et al. 2008b). Kiani et al. (2007) measured the responses of single neurons
in the monkey inferior temporal (IT) cortex to a large variety of faces, animals, and objects and cal-
culated correlations among response vectors as indices of the similarity of the population response
vectors. The results revealed the major distinctions between animate and inanimate stimuli, with
a clear distinction between faces and bodies, but went deeper to show a similarity structure for the
Figure 5
a
Three examples of Animate Inanimate
representational Body
similarity analysis
Face
(RSA). (a)
Dendrogram derived
from multiple
single-unit recordings
ce
nd
dy
bo b
dy
ts
sh
ile
r
Ca
cross-species RSA
rfl
ac
al -lim
ec
fa
fa
Fi
pt
bo
bo
Ha
dy
tte
yf
ns
by Kansas State University on 07/19/14. For personal use only.
Re
al
an
im ur
rd
an
ke
Bu
ri
im
an Fo
Bi
on
he
An
Hu
Hu
Ot
M
monkeys, however,
with less definition of
subordinate categories
Animate Inanimate
Inanimate Animate
Natural Artificial
Artificial Natural
c
High
Low
representations of animals that appears to reflect knowledge about similarities among species (see
Figure 5a). Kriegeskorte et al. (2008a,b) used a subset of the stimuli from the Kiani et al. study
in an fMRI study in humans and showed that the basic similarity structure of representations
in human ventral temporal cortex is quite similar to that for representations in the monkey IT
cortex (Figure 5b). Although Kriegeskorte et al. did not show the detailed structure among
animal species that was evident in the monkey data, subsequent targeted studies by Connolly
et al. (2012b) show that this structure is also evident in the human ventral temporal (VT) cortex
(Figure 5c).
RSA can be applied in many different ways to discover the structure of neural representational
geometries. These approaches include data-driven analyses and model-driven analyses. Data-
driven RSA discovers and describes the similarity structures that exist in different cortical fields.
Model-driven RSA searches for cortical fields whose similarity structures are predicted using
stimulus or cognitive models, including behavioral ratings of perceived similarity.
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Cortical fields that have representational spaces with similarity structures of interest can be
identified in numerous ways. Cortical fields that show significant MVP classification can be further
analyzed with RSA to describe the similarity structure of the conditions that can be distinguished
by Kansas State University on 07/19/14. For personal use only.
(Figure 6a). Cortical fields can also be identified by virtue of having similarity structures that
Figure 6
MVPA searchlight analyses (Kriegeskorte et al. 2006) for identifying cortical fields of interest (from Connolly et al. 2012a). MVP
classification accuracies (a) and consistency in local similarity structures across subjects (b) identify similarly large swaths of the visually
responsive cortex. Clustering of voxels based on similarities between locally defined searchlight dissimilarity matrix (DSMs) provides a
means to identify cortical fields with unique shared structure such as the lateral occipital complex (LOC) (c) and the early visual
cortex (d ).
are consistent across subjects (Figure 6b). The similarity structures in different cortical fields
identified in these ways, however, may differ. Connolly et al. (2012b) developed a clustering
method for finding different similarity structures in different cortical locations that were shared
across subjects (Figure 6c,d).
Understanding the similarity structure in a cortical field requires examining that structure. Of-
ten, the DSM itself is too complicated and high-dimensional to see the structure clearly. The full
DSM, therefore, is often distilled into a lower-dimensional illustration to facilitate examination.
The most common methods for reducing the dimensionality of a DSM are hierarchical clustering
to produce a dendrogram (e.g., Figure 6a,c), multidimensional scaling (MDS), and related meth-
ods such as DISTATIS (Abdi et al. 2009, 2012a). The dendrogram in Figure 5c (Connolly et al.
2012b), for example, reveals that animal species from the same class are most similar to each other
and that vertebrate classes (primates and birds) are more similar to each other than they are to the
invertebrate class (insects). MDS often reveals that a low-dimensional subspace can account for
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
a large portion of a similarity structure. For example, in Connolly et al. (2012b), the similarities
among animal classes were captured by a single dimension—the animacy continuum—that ranged
from primates to birds to insects and was associated with a distinctive coarse scale topography in
by Kansas State University on 07/19/14. For personal use only.
human VT cortex that had previously been attributed to the distinction between animate and
inanimate stimuli. Finer within-class distinctions (e.g., moths versus ladybugs), however, were
based on other dimensions and finer-scale topographies.
The meaning of the similarity structure in a representational space can also be investigated by
comparing it to a DSM generated by a model of the experimental conditions based on stimulus
features or behavioral ratings (Kriegeskorte et al. 2008a). For example, Connolly et al. (2012b)
showed that the DSM in the first cluster (Figure 6b) correlated highly with a DSM based on
behavioral ratings of similarities among animals, whereas the second cluster correlated highly with
a DSM based on visual features from a model of V1 neuron responses. Carlin et al. (2011) used
RSA to investigate the representation of the direction of another’s eye gaze that is independent of
head angle. They constructed a model similarity structure of gaze direction with invariance across
head angle. In addition, they constructed models of similarity structure due to confounding factors,
such as image similarity, and searched for areas with a similarity structure that correlated with their
DSM of interest after partialling out any shared variance due to confounding factors—illustrating
how RSA may be used to test a well-controlled model.
One of the great advantages of RSA is that it strips a cluster of response vectors out of a feature-
based representational space into a representational space based on relative distances among vec-
tors. This format allows comparison of representational geometries across subjects, across brain
regions, across measurement modalities, and even across species. The second-order isomorphism
across these spaces is afforded by the feature-independent format of DSMs. For example, between-
subject similarity of DSMs has been exploited to afford between-subject MVP classification (Abdi
et al. 2012b, Raizada & Connolly 2012).
The feature-independent second-order isomorphism, however, does have some cost. Stripping
representational spaces of features makes it impossible to compare population codes in terms of the
constituent tuning functions of those features. Thus, one cannot investigate whether the spaces in
different subjects share the same feature tuning functions or how these tuning function codes differ
for different brain regions. One cannot predict the response to a new stimulus in a subject on the
basis of the responses to that stimulus in other subjects. One cannot predict the tuning function for
individual neural features in terms of stimulus features, precluding investigators from predicting
the response pattern vector for a new stimulus on the basis of its features. The next two sections
review methods that do afford these predictions using hyperalignment and stimulus-model-based
encoding and decoding.
Haxby et al. 2011, Conroy et al. 2013). Algorithms for function-based, rubber-sheet alignment
of the cortical manifold, based on either the tuning functions or the functional connectivity of
cortical nodes, improve BSC but still do not afford BSC accuracies that are equivalent to WSC
by Kansas State University on 07/19/14. For personal use only.
Individual subject
448
features (e.g., voxels) Common model dimensions
-2.2
2.2
2 2 -1.6
16
Haxby
Voxel x
-1
1.6
6
30 June 2014
-0.8
08
·
Dimension a
x y a b
9:14
Connolly
·
Individual subject features
Voxel y -0.6
.6
Dimension b
Guntupalli
Individual subject
features (e.g., voxels) Common model dimensions
-2.5
2.5
Voxel x
-1.7
Dimension a
x y a b
= -1.6
16
-1.7
Voxel x
-2.2
22
2.2
Dimension a
x y a b
Individual subject features
Figure 7
Intersubject hyperalignment of neural representational spaces. Hyperalignment aligns individual subjects’ representational spaces into a common model representational
space using high-dimensional rotation characterized by an orthogonal transformation matrix for each subject. A dimension in this common model space is a weighted
sum of voxels in each individual subject (a column of the transformation matrix), which is functionally equivalent across subjects as reflected by the alignment of pattern
vectors in the common model space.
NE37CH22-Haxby ARI 30 June 2014 9:14
a
= y1* + y2* + y3* + ... + y35*pc35
b
= y1* + y2* + y3* + ... + y35*pc35
Figure 8
Modeling response patterns as weighted sums of common model dimensions. (a) Any response pattern in a
subject can be modeled as a weighted sum of patterns representing common model dimensions.
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
(b) Category-selective regions defined by contrasts, such as faces-versus-objects for the fusiform face area,
can be modeled in the same way.
by Kansas State University on 07/19/14. For personal use only.
The transformation matrix provides a set of basis functions that can model any pattern of brain
response. Each column in the matrix, corresponding to one common model dimension, is a set
of weights distributed across voxels. A pattern of brain response in those voxels is a weighted
sum of these patterns of weights (Figure 8a). Models of VT cortex based on single dimensions,
such as contrasts that define category-selective regions, are modeled well in the 35-dimensional
model space (Figure 8b), but these single dimensions account for only a small portion of the
variance in responses to a dynamic and varied natural stimulus such as the movie. For example,
the contrast between responses to faces and responses to objects, which defines the fusiform face
area (FFA) (Kanwisher et al. 1997), accounts for only 12% of the variance that is accounted for by
the 35-dimensional model. This result indicates that models based on simple, univariate contrasts
are insufficient as models of neural representational spaces.
The use of a complex, dynamic stimulus is essential for deriving transformation matrix param-
eters that afford general validity across a wide range of stimuli. Transformation matrices can also
be calculated on the basis of responses to more controlled experiments, such as the category per-
ception experiments. These transformation matrices are valid for modeling the response vectors
for stimuli in that experiment but, when applied to data from other experiments, do not afford
BSC of new stimuli (Haxby et al. 2011). This result indicates that data from a limited sampling of
brain states, such as those sampled in a standard category perception experiment, do not provide
a sufficient basis for building a common model of a neural representational space.
et al. 2009), and acoustic features of music (Casey et al. 2012). These features can be continuous or
binary. Thus, each stimulus is characterized as a vector in the high-dimensional stimulus feature
space (Figure 9). The response for each feature in a neural representational space (e.g., voxel) is
ECoG:
electrocorticography then modeled as a weighted sum of the stimulus features. For example, a V1 voxel may have the
strongest weights for Gabor filters of a certain orientation and spatial frequency in a particular
location, reflecting the orientation selectivity and retinotopic receptive field for that voxel. The
prediction equation for each voxel is calculated on the basis of regularized regression analysis of
responses to the stimuli in the training data. The result is a linear transformation matrix in which
each column is a neural pattern feature and each row is a stimulus feature (Figure 9). The values
in each column are the regression weights for predicting the response of that neural feature given
a set of stimulus feature values.
Applying the encoding parameter transformation matrix to the stimulus feature values for a
new stimulus thus predicts the response in each voxel (Figure 9). This new stimulus can be any
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
new stimulus in the same domain that can be described with the same stimulus features. For ex-
ample, using Gabor filters, investigators can predict the response to any natural still image (Kay
et al. 2008, Naselaris et al. 2009). Using motion energy filters, the response to any video can be
by Kansas State University on 07/19/14. For personal use only.
predicted (Nishimoto et al. 2011). Using acoustic features, the response to any clip of music can
be predicted (Casey et al. 2012). The validity of the transformation can then be tested using either
MVP classification or Bayesian reconstruction of the stimulus. Each type of validation testing
involves analysis of response vectors for new stimuli. For MVP classification, neural response vec-
tors are predicted for stimuli in the validation testing set using the linear transformation matrix
estimated from an independent set of training stimuli. The classifier then tests whether the mea-
sured response vector is more similar to the predicted response vector for that stimulus than to
predicted response vectors for other stimuli in the testing set. For Bayesian reconstruction, the al-
gorithm identifies predicted response vectors generated from a large set of stimuli (priors) that are
most similar to the measured response vector. The reconstructed stimulus is then produced from
those matching stimuli. For example, Nishimoto et al. (2011) found the 30 videos that generated
predicted response vectors that most closely matched a measured response vector. They then pro-
duced a video by averaging those 30 video priors. The reconstructed videos bear an unmistakable
resemblance to the viewed video, albeit lacking detail, providing a convincing proof of concept.
Related methods have been used to reconstruct 10 × 10 contrast images (Miyawaki et al.
2008) and decode the contents of dream imagery (Horikawa et al. 2013). A study of brain activity
measured over the auditory language cortex using electrocorticography (ECoG) in surgery patients
used related methods to reconstruct the spectrogram for spoken words, producing auditory stimuli
that closely resembled the original words (Pasley et al. 2012).
In general, stimulus-model-based decoding is limited to Bayesian reconstruction based on sim-
ilarities to a set of priors. Simply generating a stimulus based on predicted features is infeasible
because the dimensionality of the neural representational space, given current brain measure-
ment techniques, is much lower than that of stimulus feature spaces that are complete enough to
construct a stimulus. Some researchers have speculated, however, that perception also involves a
process of Bayesian reconstruction based on similarity to prior perceptual experiences, a process
that could operate effectively with incomplete specification of stimulus features (Friston & Kiebel
2009).
× =
Voxel x
-1.0
10
1.0 -1.5
1.5
.5
x y ? Measured
Stimulus features
-1 2
-1.2 -1 1
-1.1
Figure 9
Stimulus-model-based encoding and decoding involve deriving a transformation matrix that affords prediction of responses of neural features from stimulus-model
features. Each column in the encoding transformation matrix corresponds to a neural feature and provides the weights for stimulus features to estimate that neural
feature’s response. Decoding responses to novel stimuli involves comparing measured response patterns to predicted patterns for stimulus priors.
The fact that neural activity resampled into this coarse spatial grid can be effectively decoded
suggests that the topographies for distinct neural representations of the decoded information
include lower spatial frequency patterns. In fact, decoding of different distinctions appears to be
based on topographies of different spatial scales. Coarse distinctions, such as the difference between
animate and inanimate stimuli (Martin 2007, Mahon et al. 2009) or animal species at different levels
on the animacy continuum (Connolly et al. 2012b), are found in a coarse topography from lateral
to medial VT cortex. Finer distinctions, such as those between old and young human faces (Op
de Beeck et al. 2010, Brants et al. 2011) or between two types of birds (Connolly et al. 2012b), are
carried by finer-scale topographies. The finer-scale topographies for subordinate distinctions, such
as old versus young faces, are not restricted to the areas of maximal activity for the superordinate
category (Op de Beeck et al. 2010, Brants et al. 2011, Haxby et al. 2011). Thus, the topographies
for different distinctions appear to be overlapping and exist at multiple scales.
The topographic organization of the cortex reflects the problem of projecting a high-
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
dimensional representational space, composed of features with complex, interrelated tuning func-
tions, into a two-dimensional manifold. Kohonen (1982, 2001) proposed that cortical maps self-
organize to locate neurons with related tuning functions close to each other. Aflalo & Graziano
by Kansas State University on 07/19/14. For personal use only.
(2006, 2011; Graziano & Aflalo 2007a,b) have used this principle of spatial continuity of func-
tion to account for the topographic organization of the motor cortex (Aflalo & Graziano 2006)
and the coarse-scale topographic organization of the extrastriate visual cortex (Aflalo & Graziano
2011). Others have used this principle to account for multiplexed functional topographies in the
primary visual (Durbin & Mitchison 1990) and auditory (Schreiner 1995) cortices. Accounting
for seemingly disordered multiplexed functional topographies requires an adequate model of the
high-dimensional representational space. Haxby et al. (2011) showed that the topographies in
VT cortex that support a wide range of stimulus distinctions, including distinctions among re-
sponses to complex video segments, can be modeled with 35 basis functions. These pattern bases
are of different spatial scales and define gradients in different locations. Low-dimensional mod-
els of neural representation, such as those exemplified by category-selective regions (Kanwisher
2010, Weiner & Grill-Spector 2011), however, are not sufficient to model complex, multiplexed
functional topographies that support these distinctions.
FUTURE DIRECTIONS
Recent advances in computational methods for neural decoding have revealed that the information
provided from measurement of human brain activity is far more detailed and specific than was
previously thought possible. This review of the current state of the art shows that these methods
can be integrated in a framework organized around the concept of high-dimensional represen-
tational spaces. The development of algorithms for neural encoding and decoding, however, has
only begun. Although the power of neural decoding is limited by brain activity measurement
methods—and further technological breakthroughs will bring greater power and sensitivity to
neural decoding projects—new computational methods can direct investigators to additional im-
portant topics. Three areas for future investigation are addressed below.
individual and group differences would facilitate studies of how factors such as development,
education, genetics, and clinical disorders influence neural representation.
manifolds, affording viewpoint-invariant identity recognition (DiCarlo & Cox 2007, Freiwald &
Tsao 2010). Determining the structure of these transformations and the role of input from multiple
regions is a major challenge for future work.
by Kansas State University on 07/19/14. For personal use only.
Multimodality Decoding
fMRI has very coarse temporal resolution. Neural decoding studies with other measurement
modalities, such as single-unit recording (e.g., Hung et al. 2005, Freiwald & Tsao 2010), ECoG
(Pasley et al. 2012), and MEG (Carlson et al. 2011, Sudre et al. 2012), have shown how population
codes for different types of information emerge over time as measured in tens of milliseconds.
Similar tracking of population codes has been demonstrated with fMRI but is severely limited by
the temporal characteristics of the hemodynamic response (Kohler et al. 2013). Using multiple
modalities, representational spaces could be modeled in which some dimensions reflect different
time points for the same spatial feature and other dimensions reflect spatiotemporal gradients or
wavelets. The potential of multimodal neural decoding is largely unexplored.
DISCLOSURE STATEMENT
The authors are not aware of any affiliations, memberships, funding, or financial holdings that
might be perceived as affecting the objectivity of this review.
LITERATURE CITED
Abdi H, Dunlop JP, Williams LJ. 2009. How to compute reliability estimates and display confidence and toler-
ance intervals for pattern classifiers using the Bootstrap and 3-way multidimensional scaling (DISTATIS).
NeuroImage 45:89–95
Abdi H, Williams LJ, Conolly AC, Gobbini MI, Dunlop JP, Haxby JV. 2012a. Multiple Subject Barycentric
Discriminant Analysis (MUSUBADA): how to assign scans to categories without using spatial normal-
ization. Comp. Math. Methods Med. 2012:634165
Abdi H, Williams LJ, Valentin D, Bennani-Dosse M. 2012b. STATIS and DISTATIS: optimum multi-
table principal component analysis and three way metric multidimensional scaling. Wiley Interdiscip. Rev.
Comput. Stat. 4:124–67
Aflalo TN, Graziano MSA. 2006. Possible origins of the complex topographic organization of motor cortex:
reduction of a multidimensional space onto a two-dimensional array. J. Neurosci. 26:6288–97
Aflalo TN, Graziano MSA. 2011. Organization of the macaque extrastriate cortex re-examined using the
principle of spatial continuity of function. J. Neurophysiol. 105:305–20
Brants M, Baeck A, Wagemans J, Op de Beeck H. 2011. Multiple scales of organization for object selectivity
in ventral visual cortex. NeuroImage 56:1372–81
Carlin JD, Calder AJ, Kriegeskorte N, Nili H, Rowe JB. 2011. A head view-invariant representation of gaze
direction in anterior superior temporal cortex. Curr. Biol. 21:1817–21
Carlson TA, Hogendoorn H, Kanai R, Mesik J, Turret J. 2011. High temporal resolution decoding of object
position and category. J. Vis. 11:1–17
Carlson TA, Schrater P, He S. 2003. Patterns of activity in the categorical representations of objects. J. Cogn.
Neurosci. 15:704–17
Casey M, Thompson J, Kang O, Raizada R, Wheatley T. 2012. Population codes representing musical timbre
for high-level fMRI categorization of music genres. In Machine Learning and Interpretation in Neuroimag-
ing, Ser. Vol. 7263, ed. G Langs, I Rish, M Grosse-Wentrup, B Murphy, pp. 36–41. Berlin/Heidelberg:
Springer-Verlag
Chen Y, Namburi P, Elliott LT, Heinzle J, Soon CS, et al. 2011. Cortical surface-based searchlight decoding.
NeuroImage 56:582–92
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Connolly AC, Gobbini MI, Haxby JV. 2012a. Three virtues of similarity-based multi-voxel pattern analysis:
an example from the human object vision pathway. In Understanding Visual Population Codes (UVPC):
Toward A Common Multivariate Framework for Cell Recording and Functional Imaging, ed. N Kriegeskorte,
by Kansas State University on 07/19/14. For personal use only.
Haynes JD, Rees G. 2006. Decoding mental states from brain activity in humans. Nat. Rev. Neurosci. 7:523–34
Haynes JD, Sakai K, Rees G, Gilbert S, Frith C, Passingham RE. 2007. Reading hidden intentions in the
human brain. Curr. Biol. 17:323–28
Horikawa T, Tamaki M, Miyawaki Y, Kamitani Y. 2013. Neural decoding of visual imagery during sleep.
Science 340:639–42
Hung CP, Kreiman G, Poggio T, DiCarlo JJ. 2005. Fast readout of object identity from macaque inferior
temporal cortex. Science 310:863–66
Kamitani Y, Tong F. 2005. Decoding the visual and subjective contents of the human brain. Nat. First paper to show that
Neurosci. 8:679–85 MVPA can decode an
Kanwisher N. 2010. Functional specificity in the human brain: a window into the functional architecture of early visual feature,
the mind. Proc. Natl. Acad. Sci. USA 107:11163–70 namely edge
orientation.
Kanwisher N, McDermott J, Chun MM. 1997. The fusiform face area: a module in human extrastriate cortex
specialized for face perception. J. Neurosci. 17:4302–11
Kay KN, Naselaris T, Prenger RJ, Gallant JL. 2008. Identifying natural images from human brain Introduced
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Kohler PJ, Fogelson SV, Reavis EA, Meng M, Guntupalli JS, et al. 2013. Pattern classification precedes
region-average hemodynamic response in early visual cortex. NeuroImage 78:249–60
Kohonen T. 1982. Self-organizing formation of topologically correct feature maps. Biol. Cybern. 43:59–69
Kohonen T. 2001. Self-Organizing Maps. Berlin: Springer
Kriegeskorte N, Goebel R, Bandettini P. 2006. Information-based functional brain mapping. Proc. Natl. Acad.
Sci. USA 103:3863–68 Introduces RSA as a
Kriegeskorte N, Kievet RA. 2013. Representational geometry: integrating cognition, computation, and the common format for the
brain. Trends Cogn. Sci. 17:401–12 geometry of
Kriegeskorte N, Mur M, Bandettini P. 2008a. Representational similarity analysis—connecting the representational spaces.
branches of systems neuroscience. Front. Syst. Neurosci. 2:4
Kriegeskorte N, Mur M, Ruff DA, Kiani R, Bodurka J, et al. 2008b. Matching categorical object First demonstration that
representations in inferior temporal cortex of man and monkey. Neuron 60:1126–41 RSA affords comparison
Kriegeskorte N, Simmons WK, Bellgowan PSF, Baker CI. 2009. Circular analysis is systems neuroscience: of representational
the dangers of double dipping. Nat. Neurosci. 12:535–40 spaces across species.
Mahon BZ, Anzellotti S, Schwarzbach J, Zampini M, Caramazza A. 2009. Category-specific organization in
the human brain does not require visual experience. Neuron 63:397–405
Martin A. 2007. The representation of object concepts in the brain. Annu. Rev. Psychol 58:25–45 This paper introduced
decoding of words and
Mitchell TM, Shinkareva SV, Carlson A, Chang K-M, Malave VL, et al. 2008. Predicting human brain
concepts based on
activity associated with the meanings of nouns. Science 320:1191–95
semantic feature
Miyawaki Y, Uchida H, Yamashita O, Sato M, Morito Y, et al. 2008. Visual image reconstruction from human models.
brain activity using a combination of multiscale local image decoders. Neuron 60:915–29
Naselaris T, Kay KN, Nishimoto S, Gallant JL. 2011. Encoding and decoding in fMRI. NeuroImage
56:400–10 Reviews methods for
Naselaris T, Prenger RJ, Kay KN, Oliver M, Gallant JL. 2009. Bayesian reconstruction of natural images stimulus-model-based
from human brain activity. Neuron 63:902–15 encoding and decoding.
Nishimoto S, Vu AT, Naselaris T, Bejamini Y, Yu B, Gallant JL. 2011. Reconstructing visual experience from
brain activity evoked by natural movies. Curr. Biol. 21:1641–46
Norman KA, Polyn SM, Detre GJ, Haxby JV. 2006. Beyond mind-reading: multi-voxel pattern analysis of
fMRI data. Trends Cogn. Sci. 10:424–30
Oosterhof NN, Wiestler T, Downing PE, Diedrichsen J. 2011. A comparison of volume-based and surface-
based multi-voxel pattern analysis. NeuroImage 56:593–600
Op de Beeck H, Brants M, Baeck A, Wagemans J. 2010. Distributed subordinate specificity for bodies, faces,
and buildings in human ventral visual cortex. NeuroImage 49:3414–25
O’Toole AJ, Jiang F, Abdi H, Haxby JV. 2005. Partially distributed representations of objects and faces in
ventral temporal cortex. J. Cogn. Neurosci. 17:580–90
O’Toole AJ, Jiang F, Abdi H, Pénard N, Dunlop JP, Parent MA. 2007. Theoretical, statistical, and practical
perspectives on pattern-based classification approaches to the analysis of functional neuroimaging data.
J. Cogn. Neurosci. 191:735–52
Pasley BN, David SV, Mesgarani N, Flinker A, Shamma SA, et al. 2012. Reconstructing speech from human
auditory cortex. PLoS Biol. 10:e1001251
Pereira F, Mitchell T, Botvinick M. 2009. Machine learning classifiers and fMRI: a tutorial overview.
NeuroImage 45(Suppl. 1):S199–209
Raizada RDS, Connolly AC. 2012. What makes different people’s representations alike: neural similarity space
solves the problem of across-subject fMRI decoding. J. Cogn. Neurosci. 24:868–77
Sabuncu M, Singer BD, Conroy B, Bryan RE, Ramadge PJ, Haxby JV. 2010. Function-based intersubject
alignment of human cortical anatomy. Cereb. Cortex 20:130–40
Schönemann PH. 1966. A generalized solution of the orthogonal procrustes problem. Psychometrika 31:1–10
Schreiner CE. 1995. Order and disorder in auditory cortical maps. Curr. Opin. Neurobiol. 5:489–96
Shinkareva SV, Malave VL, Mason RA, Mitchell TM, Just MA. 2011. Commonality of neural representations
Annu. Rev. Neurosci. 2014.37:435-456. Downloaded from www.annualreviews.org
Annual Review of
Contents Neuroscience
v
NE37-FrontMatter ARI 23 June 2014 16:8
Indexes
Errata
vi Contents
Annual Reviews
It’s about time. Your time. It’s time well spent.
well as all scientists and users of statistics about major methodological advances and the computational tools that
allow for their implementation. It will include developments in the field of statistics, including theoretical statistical
by Kansas State University on 07/19/14. For personal use only.
underpinnings of new methodology, as well as developments in specific application domains such as biostatistics
and bioinformatics, economics, machine learning, psychology, sociology, and aspects of the physical sciences.
Complimentary online access to the first volume will be available until January 2015.
table of contents:
• What Is Statistics? Stephen E. Fienberg • High-Dimensional Statistics with a View Toward Applications
• A Systematic Statistical Approach to Evaluating Evidence in Biology, Peter Bühlmann, Markus Kalisch, Lukas Meier
from Observational Studies, David Madigan, Paul E. Stang, • Next-Generation Statistical Genetics: Modeling, Penalization,
Jesse A. Berlin, Martijn Schuemie, J. Marc Overhage, and Optimization in High-Dimensional Data, Kenneth Lange,
Marc A. Suchard, Bill Dumouchel, Abraham G. Hartzema, Jeanette C. Papp, Janet S. Sinsheimer, Eric M. Sobel
Patrick B. Ryan • Breaking Bad: Two Decades of Life-Course Data Analysis
• The Role of Statistics in the Discovery of a Higgs Boson, in Criminology, Developmental Psychology, and Beyond,
David A. van Dyk Elena A. Erosheva, Ross L. Matsueda, Donatello Telesca
• Brain Imaging Analysis, F. DuBois Bowman • Event History Analysis, Niels Keiding
• Statistics and Climate, Peter Guttorp • Statistical Evaluation of Forensic DNA Profile Evidence,
• Climate Simulators and Climate Projections, Christopher D. Steele, David J. Balding
Jonathan Rougier, Michael Goldstein • Using League Table Rankings in Public Policy Formation:
• Probabilistic Forecasting, Tilmann Gneiting, Statistical Issues, Harvey Goldstein
Matthias Katzfuss • Statistical Ecology, Ruth King
• Bayesian Computational Tools, Christian P. Robert • Estimating the Number of Species in Microbial Diversity
• Bayesian Computation Via Markov Chain Monte Carlo, Studies, John Bunge, Amy Willis, Fiona Walsh
Radu V. Craiu, Jeffrey S. Rosenthal • Dynamic Treatment Regimes, Bibhas Chakraborty,
• Build, Compute, Critique, Repeat: Data Analysis with Latent Susan A. Murphy
Variable Models, David M. Blei • Statistics and Related Topics in Single-Molecule Biophysics,
• Structured Regularizers for High-Dimensional Problems: Hong Qian, S.C. Kou
Statistical and Computational Issues, Martin J. Wainwright • Statistics and Quantitative Risk Management for Banking
and Insurance, Paul Embrechts, Marius Hofert
Access this and all other Annual Reviews journals via your institution at www.annualreviews.org.