Skip to main content

ORIGINAL RESEARCH article

Front. Integr. Neurosci., 30 June 2009
Volume 3 - 2009 | https://doi.org/10.3389/neuro.07.016.2009

Prefrontal activity predicts monkeys’ decisions during an auditory category task

Department of Otorhinolaryngology: Head and Neck Surgery, University of Pennsylvania School of Medicine, Philadelphia, PA, USA
The neural correlates that relate auditory categorization to aspects of goal-directed behavior, such as decision-making, are not well understood. Since the prefrontal cortex (PFC) plays an important role in executive function and the categorization of auditory objects, we hypothesized that neural activity in the PFC should predict an animal’s behavioral reports (decisions) during a category task. To test this hypothesis, we tested PFC activity that was recorded while monkeys categorized human spoken words (Russ et al., 2008b ). We found that activity in the ventrolateral PFC, on average, correlated best with the monkeys’ choices than with the auditory stimuli. This finding demonstrates a direct link between PFC activity and behavioral choices during a non-spatial auditory task.

Introduction

Neural correlates of auditory categories are found throughout the cortex. For example, neurons in the auditory cortex of humans and non-human animals code speech sounds and other perceptual categories (Guenther et al., 2004 ; Ohl et al., 2001 ; Poeppel et al., 2004 ; Selezneva et al., 2006 ; Steinschneider et al., 1995 ). The auditory cortex also represents multi-modal categories, such as visual and auditory communication signals and bimodal-looming representations (Hoffman et al., 2008 ; Maier et al., 2008 ). More abstract categorical representations are found in the ventrolateral prefrontal cortex (vPFC), a cortical region involved in non-spatial auditory cognition (Rauschecker and Tian, 2000 ; Russ et al., 2008a ,b ): in the vPFC, neurons code the functional meaning of vocalizations (food quality and food versus non-food) (Cohen et al., 2006 ; Gifford III et al., 2005 ).
Categories are useful because they provide an efficient means to represent information (Freedman et al., 2001 , 2002 , Miller et al., 2002 ; Shepard, 1987 ; Spence, 1937 ). They are useful since by associating a new exemplar with an established category, information that has been learned previously is available to a signal receiver, such as a listener or a viewer. Consequently, categories allow these receivers a flexible way to process and represent novel stimuli, a fundamental property of goal-directed behavior (Miller et al., 2002 ; Shepard, 1987 ; Spence, 1937 ).
Recently, we tested the neural correlates that relate auditory categories to a component of goal-directed behavior, namely decision-making (Russ et al., 2008b ). While we recorded from vPFC neurons, monkeys listened to a “reference” stimulus and “test” stimulus and reported whether these stimuli were the same or different. Using methods from signal-detection theory (Britten et al., 1992 ; Green and Swets, 1966 ; Gu et al., 2007 ), we found that vPFC activity correlated better with the monkeys’ actual choices (i.e., their decisions) than with what the monkeys should have chosen (i.e., the acoustic/perceptual relationship between the reference and test stimuli).
However, the results from this previous study were limited since the signal-detection methods that we employed used different metrics to quantify the different components of the task (e.g., what the monkeys should have chosen and what they actually chose). Thus, it is not clear the degree to which individual vPFC neurons code the monkeys’ decisions (actually chose) relative to other components of a category task.
To address this question, we constructed and trained a simple neural network – a multi-layer perceptron (MLP) (Bishop, 1995 ; Hertz et al., 1991 ; Rosenblatt, 1962 ; Widrow and Lehr, 1990 ) – to quantify how well the spike trains of a vPFC neuron coded three components of our task: (1) the perceptual category of the test stimulus, (2) the relationship between the reference stimulus and the test stimulus (i.e., what the monkeys should choose), and (3) the monkey’s decisions (i.e., their actual choice). The latter components were analyzed in the previous study with signal-detection methods but the first component is a new analysis for this study. The advantage of the MLP, in contrast to our previous analyses (Russ et al., 2008b ), is that, on a neuron-by-neuron basis, we can quantify each of these three task-related components with a comparable metric. We found that vPFC neurons, on average, coded the monkeys’ decisions better than they coded the perceptual category of the test stimulus and better than they code the relationship between the reference and test stimuli. These results confirm and extend our earlier study (Russ et al., 2008b ) by demonstrating that, on average, vPFC activity reflects the decision-making processes that monkeys make during a non-spatial auditory task. These data also provide a direct link between single neurons and behavioral choices in the vPFC on a non-spatial auditory task.

Materials and Methods

The neural data set analyzed here has been the subject of a previous study (Russ et al., 2008b ). We recorded from neurons in the vPFC from one male and one female rhesus monkey (Macaca mulatta). Under isofluorane anesthesia, the monkeys were implanted with a scleral search coil, head-positioning cylinder, and a recording chamber. vPFC recordings were obtained from the male rhesus’ left hemisphere and from the female’s right hemisphere. All recordings were guided by pre- and post-operative magnetic resonance images of each monkey’s brain. The recording cylinder was centered on the region of the PFC that overlaps with areas 12/45 as defined by Romanski and colleagues (Romanski and Goldman-Rakic, 2002 ; Romanski et al., 1999a ). Stereotaxically, this region is centered ∼26 mm anterior (relative to the interaural axis) and ∼20 mm lateral. The Dartmouth Institutional Animal Care and Use Committee approved the experimental protocols.

Auditory Stimuli

The prototype stimuli were the spoken words bad and dad. In humans, these stimuli differ in their place of articulation. The prototypes were digitized recordings of an American adult female and were provided by Dr. Michael Kilgard. Morphed versions of the prototypes were created using the STRAIGHT (Kawahara et al., 1999 ) software package, which is run in the Matlab (The Mathworks Inc.) programming environment. Morphing was accomplished by calculating the shortest trajectory between the fundamental and formant frequencies of the two prototypes. Morphed versions of the two prototypes were created at 20, 40, 50, 60, and 80% of the distance along this trajectory. Spectrograms of the two prototypes and some of the morphed stimuli are shown in Figure 1 B.
Figure 1. Same-different task and behavioral performance. (A) Following two to four presentations of the reference stimulus, a test stimulus was presented. The reference stimulus was always one of the two prototype spoken words (bad or dad). The test stimulus was a morphed version of the prototypes. If the monkeys perceived that the reference and test stimuli were the same, they made a saccade to a leftward target. If the monkeys perceived that the reference and test stimuli were different, they made a saccade to a rightward target. (B) Spectrographic representations of the prototype spoken words and two of the morphs. In this example, the reference stimulus is bad. Consequently, it is the 100% morph, whereas dad is the 0% morph; see Section “Materials and Methods” for more details. When the reference stimulus is the spoken word dad, the morph percentages are reversed (e.g., the 0% morph is bad and the 100% morph is dad). The axes for all of the spectrograms are seen in the leftmost spectrogram. (C) The average performance of the monkeys from those recording sessions reported in this manuscript. The monkeys’ performance is shown as a function of the reference stimulus: the prototype spoken word bad (left column) or dad (right column). A 0% morph means that the test stimulus was a different prototype than the reference stimulus (e.g., the reference stimulus was the prototype bad and the test stimulus was the prototype dad). A 100% morph means that the test and reference stimuli were the same (e.g., both were the prototype dad). Other values represent morphed stimuli between these two extremes. Error bars are standard error of the means.

Same-Different Task

As schematized in Figure 1 A, the task began with two to four presentations of a “reference” stimulus that was followed by the presentation of a “test” stimulus. The reference and test stimuli were 500 ms in duration. The inter-stimulus interval averaged 1600 ms. The stimuli were presented from a speaker (Pyle, PLX32) that was in front of the monkey at a level of 70 dB SPL. The reference stimulus was always one of the two prototype words. The test stimulus was either (1) one of the two prototypes or (2) a morph of one of the two prototypes. The 100% morph was operationally defined to be the same prototype as the reference stimulus; therefore, the 0% morph was the other prototype. 700-ms after test-stimulus offset, two LEDs were illuminated. If the test stimulus was a 0–40% morph, the monkeys were rewarded when they successfully reported that the reference and test stimuli were different by making a saccade to the LED that was 20° to the right of the speaker. If the test stimulus was a 60–100% morph, the monkeys were rewarded when they successfully reported that the reference and test stimuli were the same by making a saccade to the LED that was 20° to the left of the speaker. When the test stimulus was a 50% morph, which has been shown to be a (categorical) perceptual boundary in rhesus (see Figure 1 C) (Kuhl and Padden, 1982 , 1983 ), the monkeys were rewarded based on their overall performance (Grunewald et al., 2002 ). The reward was given 300 ms after the monkey fixated one of the target LEDs; the reward was a drop of juice or water (∼0.25–0.5 ml/correct trial).

Recording Procedure

Single-unit extracellular recordings were obtained with tungsten microelectrodes (Frederick Haer & Co., Bowdoin, MA, USA) seated inside a stainless-steel guide tube. The electrode and guide tube were advanced into the brain with a hydraulic microdrive (Narishige MO-95). The electrode signal was amplified (Bak MDA-4I) and band-pass filtered (Krohn-Hite 3700) between 0.6 and 6.0 kHz. Single-unit activity was isolated using a two-window, time-voltage discriminator (Bak DDIS-1). The time of occurrence of each action potential was stored for on- and off-line analyses.
The vPFC was identified by its anatomical location and its neurophysiological properties (Cohen et al., 2004 ; Romanski and Goldman-Rakic, 2002 ). The vPFC is located anterior to the arcuate sulcus and Area 8a and lies below the principal sulcus. vPFC neurons were further characterized by their strong responses to auditory stimuli.
Once a neuron was isolated, the monkeys participated in blocks of trials of the same-different task. Since vPFC neurons respond broadly to a wide range of auditory stimuli (Cohen et al., 2007 ; Romanski et al., 2005 ; Russ et al., 2008a ), we did not tailor the reference and test stimuli to the neuron’s response characteristics. In each block of trials, there were six trials in which the test stimulus was a 0% morph, six trials in which the test stimulus was a 100% morph, and two trials of each of the remaining morphs (i.e., the 20, 40, 50, 60, and 80% morphs). The test stimulus was chosen in a balanced pseudorandom order. We recorded at least five blocks of trials with each reference stimulus. Consequently, for each neuron, we typically had ≥200 spike trains that were available for subsequent off-line data analysis.

Data Analysis

A fully connected MLP (Bishop, 1995 ; Hertz et al., 1991 ; Rosenblatt, 1962 ; Widrow and Lehr, 1990 ) was constructed that contained an input layer, a single hidden layer with 100 nodes, and an output layer that contained 2 nodes (see Figure 2 ); as is commonly done, the number of output nodes equaled the number of items that were learned by the network (i.e., 2) (Hertz et al., 1991 ; LeCun et al., 1998 ). The number of nodes of the input layer depended on the statistical properties of the spike trains (see below). The activation function of hidden units was tanh(x).
Figure 2. Schematic of the multi-layer perceptron (MLP). For each neuron, the interspike intervals of M spike trains underwent principal-component analysis (PCA) to extract relevant features. For each spike train, the N components derived from the PCA were fed into the input layer of the MLP sequentially. The hidden layer of the network contained 100 nodes and the output layer contained 2 nodes. The network was a fully connected feedforward network. A winner-take-all approach applied to the output node: thus, the node with the highest activation determined the MLP’s answer.
The MLP was not designed to mimic brain function. Instead, it was a platform that we used to quantify how well vPFC neurons coded different task-related components of the same-different task (see below). The advantage of the MLP, in contrast to our previous study (Russ et al., 2008b ), is that, on a neuron-by-neuron basis, we can quantify these different task-related components with a comparable metric. We chose a MLP versus a relatively simpler linear classifier for two reasons: (1) MLPs generally perform better than linear classifiers (LeCun et al., 1998 ) and (2) since our data set is not linearly separable (results not shown), a linear classifier is not appropriate (Hertz et al., 1991 ).
The input to the MLP was a vector of interspike-interval times. For each neuron, the interspike-interval times during the 1000-ms period that started with the test-stimulus onset was calculated on a trial-by-trial basis of the same-different task; this time period ended before the LEDs were introduced into the environment (see Figure 1 A).
A principal-component analysis (PCA) pre-processed these interspike-interval times (see Figure 2 ). PCA extracts the relevant features of a dataset and also reduces its dimensionality (Battiti, 1994 ; Jain and Zongker, 1997 ; LeCun et al., 1998 ; Liu and Wang, 1998 ; Trier et al., 1996 ; Turk and Pentland, 1991 ). Indeed, the PCA decreased the dimensionality of our dataset by ∼25%. Such pre-processing is commonly used in neural networks, like ours, that are trained to recognize patterns and classify data. Dimensionality reduction, besides decreasing computation time, also improves the capacity of classifiers, such as a MLP, to generalize to novel inputs (Cristianini and Shawe-Tayler, 2000 ; Jain and Zongker, 1997 ).
For our PCA, a matrix was constructed for each neuron in which each row was a trial and the columns contained the interspike-interval times. Using this matrix, we calculated the principal components and the projections of each vector of interspike-interval times (i.e., the row of the matrix) in the principal-component space; each row vector was zero padded, relative to the longest trial, so that each row vector in the matrix was the same length. These projections formed the actual inputs to the MLP.
We trained different MLPs to decode three different components of the same-different task.
The first component was the perceptual category of the test stimulus or the “test-stimulus category.” For this training, one output node of the MLP network mapped neural activity elicited by a 60–100% test morph to one prototype stimulus (e.g., bad). The second output node mapped neural activity elicited by the 0–40% test morph to the other prototype stimulus (e.g., dad). These cut-offs were based on our monkeys’ behavioral performance which indicated that the monkeys perceived the bad–dad transition in a categorical manner (see Figure 1 C); this behavior is consistent with the extant literature (Eimas et al., 1971 ; Kuhl and Miller, 1975 , 1978 ; Kuhl and Padden, 1982 , 1983 ). The MLP training was conducted independent of both (1) the reference stimulus and (2) the monkey’s behavioral reports (actual choices or decisions).
The second component was the relationship between the reference stimulus and the test stimulus – that is, what the monkeys should choose based on the acoustic/perceptual relationship between the reference and test stimuli (see Figure 1 C). For this training, one output node of the MLP network was trained to associate activity elicited by a 60–100% test morph with the reference stimulus. The other output node was trained such that activity elicited by a 0–40% test morph was not associated with the reference stimulus. Once again, these cut-offs were based on our monkeys’ behavioral performance which indicated that they perceived the bad–dad transition in a categorical manner (see Figure 1 C). This MLP training was done independently of the monkeys’ behavioral reports (actual choices or decisions).
The third and final component was the monkeys’ behavioral reports or decisions – that is, what the monkeys actually chose. For this training, one output node of the MLP network mapped neural activity with the monkey’s decision that the reference and test stimuli were the same. The second output node mapped neural activity with the monkey’s decision that the reference and test stimuli were different. This MLP training was accomplished independent of both the reference and test stimuli.
For all three of these task-related components (i.e., test-stimulus category, the relationship between reference and test stimuli, and monkeys’ decisions), we included both successful and error trials. We applied a winner-take-all rule to the output layer. Hence, the output node with the highest activation determined the MLP’s answer.
The MLP was trained using classic backpropogation (Almeida, 1988 ; Chauvin and Rumelhart, 1995 ; Hertz et al., 1991 ; Pineda, 1987 ; Rohwer, 1987 ); the weights of the network were updated after propagating each spike train. Since the training of a MLP is stochastic, we trained a population of independent MLPs to generate bounds on the amount of information contained in vPFC activity. That is, for each neuron, an independent set of 50 MLPs was trained to decode the test stimulus from vPFC activity; a second independent set of 50 MLPs was trained to decode the relationship between the reference and test stimuli; and a third independent set was trained to decode the monkeys’ decisions. Each MLP was initialized with a different set of initial random weights and was trained and tested independently.
For each MLP, one half of a neuron’s spike trains (data) was used as the training set. The other half of the neuron’s spike trains was used to evaluate the MLP’s performance (decoding capacity). This approach is called the “split-sample” method (LeCun et al., 1998 ; Prechelt, 1998 ).
When a MLP is over-trained, its capacity to generalize to novel inputs becomes poorer while its performance on the training set continues to improve. To test for over-training, we instantiated an algorithm called “early-stopping” (Caruana et al., 2000 ; Nelson and Illingworth, 1991 ; Prechelt, 1998 ). In early-stopping, a subset of data, the “validation set,” is removed from the training set. After each training epoch (i.e., a complete cycle of training that used all of the available data from the training set), the MLP is tested with the validation set. Validation error typically decreases as training commences and as the MLP learns the data. However, with more training, the validation error increases as the MLP overfits the data. Consequently, the evaluation set is tested on the weights that produced the lowest validation error.
In our instantiation of this algorithm, we randomly removed 20 spike trains from the training set to be used as a validation set. After each training epoch, the performance of the MLP on the validation set was evaluated. This process was repeated iteratively for 1000 training epochs to find the set of weights that produced the lowest validation error. After training was complete (1000 epochs), the “best” MLP was built using this set of weights. Finally, the evaluation set of spike trains was fed into this best MLP to test its decoding capacity on this novel data set.

Results

Neurophysiological Recordings

We recorded from 91 vPFC “auditory” neurons while the monkeys participated in the same-different task (Figure 1 A); “auditory” neurons had reliably different firing rates during the 500-ms period that began with test-stimulus onset than during the 500-ms period that occurred prior to test-stimulus onset (t-test, p < 0.05). Approximately equal numbers of neurons were collected from both monkeys; we could not identify any differences between the data collected from the two monkeys so the data were treated as a unitary dataset. For 51 of these 91 neurons, we collected blocks of data in which both bad and dad were the reference stimulus. In the other 38 neurons, we only collected blocks of data in which either bad (22 neurons) or dad was the reference stimulus (16 neurons). All 51 of these neurons were classified as “auditory”; these neurons had reliably different firing rates during the 500-ms period that began with the test-stimulus onset than during the 500-ms period that occurred prior to the test-stimulus onset (t-test, p < 0.05).
The response profile from a vPFC neuron is shown in Figure 3 . The data in Figure 3 A were generated when the reference stimulus was bad. The neuron generally had a high firing rate when the monkey reported that the reference and test stimuli were different (see the blue colors). In contrast, when the monkey reported that the reference and test stimuli were the same (see red-purple colors), the neuron had a relatively lower firing rate. There are several interpretations of this neuron’s response. First, it could reflect the monkey’s decisions: high activity when the monkey reported that the reference and test stimuli were different and lower activity when the monkey reported that these stimuli were the same. Similarly, this activity could reflect the relationship between the reference and test stimuli. These two cases are different in that the former reflects what the monkey actually chose, whereas the latter reflects what the monkey should choose. Alternatively, this activity might reflect the monkey’s perception of the test-stimulus (see Figure 1 C) or the test stimulus’ acoustic features. For example, high responses could indicate when the monkey perceived the test stimulus as dad, and lower responses could indicate when he perceived it as bad.
Figure 3. An example of neural activity from a single vPFC neuron during the same-different task. In (A), the reference stimulus was the prototype spoken word bad. In (B), the reference stimulus was the prototype spoken word dad. For both panels, the rasters and spike-density histograms are aligned relative to the onset of the test stimulus. The morph value of the test stimulus is indicated by color as shown by the color bar: 0% morphs are the lightest blue color and 100% morphs are the purple color. When the test stimulus was a 100% morph, it was identical to the reference stimulus. The arrows in (A) indicate the approximate times of each of the reference stimuli. In these two panels, only successful trials are shown.
These factors can be disambiguated, in part, by looking at the neuron’s response to the other reference stimulus, dad. These data are displayed in Figure 3 B. If the neuron was coding the monkey’s perception of the test stimulus, for example, the neuron should continue to elicit high responses when the monkey perceived the test stimulus as dad (see blue colors) and lower responses when he perceived it as bad (see red-purple colors). However, this pattern is not observed. Instead, we continue to find high levels of activity when the monkey reported that the reference and test stimuli were different and lower levels of activity when the monkey reported that the stimuli were the same. Consequently, vPFC activity does not appear to reflect the monkeys’ percept of the test stimulus but better reflects more abstract components of the same-different task relating to what he should chose or actually chose. To further test whether vPFC activity reflects (1) the monkeys’ decisions (actually chose), (2) the relationship between the reference and test stimuli (should choose), or (3) perceptual category of the test stimulus, we trained independent sets of MLPs to decode these three task-related components from vPFC activity.

MLP Decoding

We tested how well a MLP decoded three different components of the same-different task from vPFC activity: (1) perceptual category of test-stimulus (i.e., the test-stimulus category), (2) the relationship between the reference and test stimuli (i.e., what the monkeys should choose), and (3) the monkeys’ actual choice (i.e., their behavioral reports or decisions); see Section “Materials and Methods” for more details. To quantify their capacity to code each these three task-related components, 50 MLPs were independently trained for each of these three components on a neuron-by-neuron basis.
The MLPs were able to learn the relationship between vPFC activity and the desired task-related component. Figure 4 shows two examples of this relationship when two MLPs were trained to decode decision-related neural activity. The data in Figure 4 A were generated from the neuron whose activity most reliably reflected the monkey’s decisions (in terms of the MLP’s performance on the validation dataset); whereas the data in Figure 4 B were generated from the neuron whose activity least reliably reflected the monkey’s decisions.
Figure 4. Learning examples from two neurons. Panel (A) was generated from the neuron whose activity most reliably reflected the monkey’s decisions (in terms of the MLP’s performance on the validation dataset); whereas the data in panel (B) were generated from the neuron whose activity least reliably reflected the monkey’s decisions. The graphs in both panels show the average training (red data) and validation (blue data) history; these average values were calculated from the 50 individual training histories that were generated from each of the 50 MLPs constructed for each neuron. The error function (dotted black line) is also shown as a function of training epoch; the error function is yes where T is the theoretical value [0 or 1] and O is the actual value, which is a continuous function, of the output node i for each spike train j. The error function is calculated as a function of each training epoch.
As can be seen, with more training, the MLP’s performance on the training set improved monotonically. However, whereas the MLP’s performance on the validation set in Figure 4 A improved greatly with training before reaching an asymptote, the MLP’s performance on the validation set in Figure 4 B did not show substantial improvement. Nevertheless, for both cases, the MLP’s performance on the validation set reached its maximum point relatively early in the training period. Thus, we often found that the 1000 training epochs were redundant. It is important to note, though, that this did not impact any interpretation of the data from the evaluation set since, in accordance with the early-stopping algorithm (Caruana et al., 2000 ; Nelson and Illingworth, 1991 ; Prechelt, 1998 ), we used the weights from the MLP that gave the best performance, relative to the validation set, and not the weights generated on the last training epoch (see Section “Materials and Methods”).
To quantify the MLP’s performance for each vPFC neuron, we calculated the proportion of times that each of the 50 MLPs correctly decoded the test-stimulus category from the evaluation set. From these 50 decodings, we formed a distribution of “proportion correct.” Next, an analogous distribution was created from 50 MLPs that were (1) trained to decode the relationship between the reference and test stimuli and (2) trained to decode the monkeys’ decisions. Finally using a one-way ANOVA with post hoc comparisons, we tested, on a neuron-by-neuron basis, whether the proportion-correct distributions from the “test-stimulus-trained” MLPs, the “relationship-trained” MLPs, and the “decision-trained MLPs” were reliably different. We found that a significant population (n = 17/51; binomial probability; p < 0.05) of vPFC neurons decoded the monkeys’ decisions reliably better (p < 0.05) than they decoded the test stimulus and the relationship between reference and test stimuli. A non-significant population (n = 4/51) of vPFC neurons decoded the relationship between the reference and test stimuli better, and a small but significant population (n = 6/51; p < 0.05) of vPFC neurons decoded the test stimulus better. Finally, for 24 vPFC neurons, there was no reliable difference between the proportion-correct distributions. A chi-squared test indicated that this distribution of significant neurons was significantly different than that expected by chance (p < 0.05).
Next, for each neuron, we calculated the mean value for each of these three proportion-correct distributions. We defined this mean value as the “average decoding capacity” (ADC). Next, these three values were pair-wise correlated on a neuron-by-neuron basis. The results of these correlations are shown in Figure 5 . In each panel of this figure, a data point’s position along the horizontal and vertical axis represents two of the three possible ADC values for each neuron. For example in Figure 5 A, a data point’s position along the horizontal axis represents a neuron’s relationship-trained-ADC value, whereas a data point’s position along the vertical axis represents the neuron’s decision-trained-ADC value. As can be seen, on average, vPFC neurons had both significantly larger decision-trained-ADC values than relationship-trained-ADC values (Figure 5 A) and test-stimulus-trained-ADC values (Figure 5 B) (Wilcoxon test, p < 0.05). However, the test-stimulus-ADC values and the relationship-trained-ADC values were not reliably different (Figure 5 C).
Figure 5. Population analysis of average decoding capacity (ADC). In each panel, on a neuron-by-neuron basis, two of the three ADC values are compared. In panel (A), the decision-trained-ADC values and relationship-trained-ADC values are compared. In panel (B), the decision-trained-ADC values and test-stimulus-trained-ADC values are compared. In panel (C), the test-stimulus-trained-ADC values and relationship-trained-ADC values are compared. In each panel, the solid line is the line of equal ADC value.

Sensitivity to the Type of Stimuli in the Training Set

To assess how robust the decoding capacity of the MLPs was to the particulars of the training set, we performed two additional analyses. First, for each neuron, we trained three new sets of MLPs with data only from those trials in which the test stimulus was a prototype stimulus (i.e., a 0 or 100% morph) and then tested the MLPs with data generated from trials in which the test stimulus was a prototype or a morphed stimulus. The purpose of this analysis was to test whether the MLPs could generalize from vPFC test-stimulus activity elicited by prototype test stimuli to vPFC activity elicited by the morphed test stimuli. This MLP-training paradigm also mimicked the actual training that the monkeys received: prior to recording, the test stimulus was always one of the two prototypes, but when recording began, morphed stimuli were also used as test stimuli.
The results of this analysis can be found in Figure 6 . Decision-trained-ADC values were, on average, significantly greater (Wilcoxon; p < 0.05) than both relationship-trained-ADC values (Figure 6 A) and test-stimulus-trained-ADC values (Figure 6 B). Whereas the magnitude of the ADC values is less than that seen in Figure 5 , the relative difference between decision-trained-ADC values and the other ADC values was substantially larger than that seen in Figure 5 . Also, unlike Figure 5 , on average, the relationship-trained-ADC values were significantly greater (Wilcoxon; p < 0.05) than the test-stimulus-trained-ADC values (Figure 6 C).
Figure 6. Population analysis of ADC values when the networks were trained using vPFC activity from trials in which the test stimulus was a prototype stimulus only. In each panel, on a neuron-by-neuron basis, two of the three ADC values are compared. In panel (A), the decision-trained-ADC values and relationship-trained-ADC values are compared. In panel (B), the decision-trained-ADC values and test-stimulus-trained-ADC values are compared. In panel (C), the test-stimulus-trained-ADC values and relationship-trained-ADC values are compared. In each panel, the solid line is the line of equal ADC value.
Next, to further assess how robust the decoding capacity of the MLPs was to the particulars of the training set, we trained three new sets of MLPs using data from one reference-stimulus prototype and evaluated the MLPs using data from the other reference stimulus. For example, if a MLP was trained using data in which bad was the reference stimulus, it was evaluated using only data in which dad was the reference stimulus. If the activity of vPFC neurons was dependent on the reference stimulus, we would predict that if the MLP was trained only on data in which bad was the reference stimulus, the MLP should perform poorly when tested with data in which dad was the reference stimulus. However, if a vPFC neuron’s response is not dependent on the reference stimulus, the MLP should perform relatively better when tested with data using a different reference stimulus than it was trained. For each of the 50 MLPs and for each neuron, we randomly picked which reference stimulus was used during training.
The results of this analysis can be found in Figure 7 . As in Figure 6 , the magnitude of the ADC values are reduced but the general pattern remains the same as that seen in Figure 5 . That is, on average, (1) decision-trained-ADC values were significantly greater (Wilcoxon; p < 0.05) than both test-stimulus-trained-ADC and relationship-trained-ADC values (Figures 7 A,B) and (2) the test-stimulus-trained-ADC and relationship-trained-ADC values were not reliably different (Figure 7 C). Together, the results shown in Figures 6 and 7 suggest that, independent of the type of training, vPFC neurons code the monkeys’ decisions better than other components of the same-different task.
Figure 7. Population analysis of ADC values when the networks were trained using data generated from only one reference-stimulus prototype and evaluated with data from the other reference stimulus. In panel (A), the decision-trained-ADC values and relationship-trained-ADC values are compared. In panel (B), the decision-trained-ADC values and test-stimulus-trained-ADC values are compared. In panel (C), the test-stimulus-trained-ADC values and relationship-trained-ADC values are compared. In each panel, the solid line is the line of equal ADC value.

Discussion

MLPs that were trained with the spike trains of vPFC neurons were able, on average, to decode the monkeys’ decisions more accurately than they were able to decode both the perceptual category of the test stimulus (test-stimulus category) and the relationship between the reference and test stimuli. These results confirm our previous study by Russ et al. (2008b) and extend it by quantitatively demonstrating that vPFC activity, on average, correlates better with decision-related activity on a neuron-by-neuron basis (Figure 5 ).
Importantly, higher decision-trained-ADC values were still found when the MLPs were trained only with data in which test stimuli were prototypes and when the MLPs were trained only with data from one reference-stimulus prototype (Figures 6 and 7 ). Together, these data indicate further that, on average, vPFC activity correlates best with the monkeys actual choices (decisions) and not with the perceptual category of the test stimulus or what the monkeys should choose (i.e., the relationship between the reference and test stimuli). These results also reinforce the hypothesis that the vPFC and the auditory regions leading to the vPFC play an active role in aspects of non-spatial auditory cognition (Rauschecker and Tian, 2000 ; Russ et al., 2007 , 2008a ,b ); this hypothesis is discussed in further detail in the next section.

Relationship between vPFC and Auditory-Cortex Processing

Since spoken words, like bad, dad, and their morphs, can be considered to be auditory objects (Miller and Cohen, in press ), it might be fruitful to frame our interpretation of this study within a context of auditory-object analysis (Blank et al., 2002 , 2003 ; Darwin, 1997 ; De Santis et al., 2007 ; Griffiths and Warren, 2004 ; Micheyl et al., 2005 ; Murray et al., 2006 ; Nelken et al., 2003 ; Poremba et al., 2004 ; Rauschecker, 1998 ; Scott, 2005 ; Scott and Wise, 2004 ; Scott et al., 2000 , 2004 ; Sussman, 2004 ; Ulanovsky et al., 2003 ; Wise et al., 2001 ; Zatorre et al., 2004 ). The first step in auditory-object analysis is for the perceptual system to extract and code the spectrotemporal properties, localization cues, and other low-level features in the signal. These features are then “bound” together to form a representation of the object. The next components of auditory-object analysis involve computations that lead to the formation of increasingly abstract representations and to other perceptual/cognitive states that guide actions and decisions. We present these steps to be serial in nature only as a useful conceptual heuristic, which may not reflect true neural processing. Indeed, the cortex is likely to process an auditory object in a dynamic parallel system in which detection and discrimination are not separable processes but different read-out schemes (Geisler and Albrecht, 1996 ; Gold and Shadlen, 2007 ; Sternberg, 2001 ).
Where in the cortical hierarchy are auditory objects processed? The most likely pathway for vocalization processing is the so-called “ventral” processing stream, a pathway that processes the non-spatial attributes of an auditory stimulus (Rauschecker and Tian, 2000 ; Ungerleider and Mishkin, 1982 ). This pathway originates in the auditory cortex (Kaas and Hackett, 2000 ). The ventral stream is further defined by a series of projections that includes the anterior belt of the auditory cortex and regions of the prefrontal cortex (PFC), specifically the vPFC (Rauschecker and Tian, 2000 ; Romanski et al., 1999a ,b ).
Our analyses indicate that, on average, vPFC activity correlates better with neural functions that follow auditory-object formation. Specifically, vPFC activity appears to better reflect the abstract neural states involved in decision-making. As noted above, on average, the ADC of vPFC neurons was highest when the MLPs were trained to decode the monkeys’ actual decisions and not the “lower-level” components of the task such as perceptual category of the test stimulus (Figures 57 ). Consistent with these neurophysiological data, a transcranial-magnetic-stimulation study from our laboratory has provided direct evidence that the vPFC is causally involved in decision-making during the same-different task (Orr et al., 2008 ).
Where in this ventral processing stream do neurons code the perceptual features – specifically the categorical percept of bad or dad (see Figure 1 ) (Eimas et al., 1971 ; Kuhl and Miller, 1975 , 1978 ; Kuhl and Padden, 1982 , 1983 ) – and other components of the task such as comparison between the reference and test stimuli? We hypothesize that regions of the auditory cortex that are part of the ventral processing stream carry this type of information (Rauschecker and Tian, 2000 ; Russ et al., 2008a ,b ). Several pieces of data support this hypothesis. First, category-related information about human-phoneme differences is seen in the auditory cortex of the untrained rhesus monkeys and rats (Engineer et al., 2008 ; Steinschneider et al., 1995 ). Indeed, it is thought that the capacity to discriminate between human speech sounds, such as phonemes like ba and da, relies mainly on general bottom-up auditory mechanisms that are common to all vertebrates (Aslin et al., 2002 ). Second, preliminary data from our laboratory indicate that neurons in the superior temporal gyrus, a region of the auditory cortex that receives input from the primary auditory cortex and that projects to the vPFC (Rauschecker and Tian, 2000 ; Romanski et al., 1999a ,b ) codes the perceptual category of the test stimulus reliably better than decision-related activity (Lee and Cohen, unpublished observations). Finally, a number of studies suggest that at the level of the primary auditory cortex, if not earlier (Nelken et al., 2003 ), neurons are integrating the dynamic spectrotemporal properties of a stimulus, a fundamental requirement for object perception (Barbour and Wang, 2003 ; Bendor and Wang, 2007 ; Fishman et al., 2000 , 2001 ; Wang and Kadia, 2001 ; Wang et al., 2005 ).
Our hypothesis that decision-related circuitry is a product of computations occurring in the PFC and not the “sensory” cortex is supported by analogous studies in the visual system. Miller and colleagues argue that PFC neurons tend to reflect a stimulus’ membership in a category more than its physical properties, whereas neurons in the infratemporal cortex tend to be better correlated with its physical properties than PFC neurons (Freedman et al., 2003 ). More specifically, the responses of PFC neurons tend to vary with the rules mediating a task or the behavioral significance of stimuli, whereas responses in the infratemporal cortex tend to be invariant to these variables (Ashby and Spiering, 2004 ; Freedman et al., 2003 ).

Comparing the MLP with other Classifiers

In general, there is no a priori way to determine what type of classifier (e.g., MLP, linear classifier, etc.) works best with a particular data set. For example, LeCun et al. (1998) compared the performance of a variety of classifiers on a standardized database of handwriting samples and found that a perceptron with two hidden layers performed better than a perceptron with a single hidden-layer. However, other studies indicate that classification does not always improve with more hidden layers (Bishop, 1995 ).
In our study, we used a single hidden-layer perceptron for two reasons. First, the learning capacity of a single hidden-layer network is well described both theoretically and empirically (Bishop, 1995 ; Cybenko, 1989 ; LeCun et al., 1998 ; Siegelmann and Sontag, 1991 ). Second, this network can perform both linear and non-linear mappings between inputs and outputs; when the weights are small, a MLP implements a linear function. Overall, one has to be conservative when interpreting the data from the MLP since other classifiers may prove to be better for any particular data set.

Comments on Performance of the MLP

On average, the performance of the MLPs was relatively poor. This observation relates to at least three non-exclusive issues. First, we reported all of the results from a database that used a very minimal criterion for inclusion: responses to sounds. Consequently, the activity of some vPFC neurons was probably highly related to the task, whereas the activity of other neurons was not related to the task. Indeed, the categorization studies from Miller’s laboratory (Freedman et al., 2001 , 2002 ) indicate that only ∼20–25% of their PFC neurons are engaged in categorization. So, it is conceivable that only a small percentage of PFC neurons are engaged in a given task.
Second, the poor performance might be inherent to components of the MLP’s training. Each MLP was trained by randomly selecting the training set and the initial weights. Consequently, it is reasonable to speculate that some MLPs failed to extract the learning (generalization) rule that was taught during training. Indeed, some MLPs learned the rule, whereas others did not. Since there is no a priori way to determine whether a MLP is trained appropriately (Bishop, 1995 ; Hertz et al., 1991 ) (i.e., one MLP can be well trained or poorly trained by a random selection of the training set or the initial weights), we choose to look at the ADC (i.e., the mean value for each of proportion-correct distributions for each neuron) to avoid the bias inherent in looking at the results of a single MLP.
Finally, it is possible that better behavioral performance and/or better ADC values might have been obtained if we had used a different stimulus set, such as species-specific vocalizations. However, we chose not to use vocalizations since vocalizations do not differ along a single dimension (e.g., some are noisy and some are harmonic stacks) (Hauser, 1998 ). In contrast, the phonemes ba and da differ along a single dimension: their 2nd formant (Diehl et al., 2004 ). Furthermore, if we had morphed species-specific vocalizations, we would have induced potential changes in both the referent and semantic meaning of the vocalizations and had to control for these two confounds.

Alternative Interpretations

One possible alternative interpretation is that since the stimulus-presentation dynamics in our same-different task are similar to that used in oddball tasks and stimulus-specific adaptation (Näätänen, 1992 ; Ulanovsky et al., 2003 ), stronger “pop-out” vPFC responses might reflect the automatic detection (Näätänen, 1992 ) of uncommon test stimuli. However, several pieces of data argue against this possibility. First, if vPFC responses reflect detection of test stimuli that are acoustically distinct from the reference stimulus, then they should have responded strongly to any test stimulus that was acoustically distinct from the reference stimulus (i.e., the 0–80% morphs). However, since vPFC neurons respond weakly to several of these “novel” test stimuli (e.g., Figure 3 ), vPFC activity cannot reflect the presence of acoustically distinct test stimuli. Moreover, some vPFC neurons have a low firing rate when the reference and test stimuli were acoustically distinct but a relatively higher firing rate when the reference and test stimuli were the same (Russ et al., 2008b ); once again, a pattern of responsivity incompatible with the idea that vPFC neurons automatically signal the detection of acoustically distinct test stimuli with strong pop-out responses (Näätänen, 1992 ). Finally, if vPFC activity reflects the automatic detection of acoustically uncommon stimuli, we would expect that vPFC activity would habituate with repeated presentations of the reference stimuli as seen in stimulus-specific adaptation studies (Reches and Gutfreund, 2008 ; Ulanovsky et al., 2003 ). However, contrary to this hypothesis, separate studies from our lab have failed to note this pattern of activity (Gifford III et al., 2005 ; Russ et al., 2008b ). Thus, several lines of evidence indicate that vPFC activity does not reflect acoustically novel stimuli.
It is possible that the higher responses reflect the detection of stimuli that are novel semantically or perceptually (Strange et al., 2000 ). Indeed, we have argued previously that vPFC neurons reflect changes in a semantic difference between vocalizations (Gifford III et al., 2005 ). Under this hypothesis, we would only see enhanced responses that were perceptually distinct from the reference stimulus (i.e., morphs >50%), a pattern consistent with our data. However, even this version of the novelty-detection hypothesis would predict decreases in neural response to repeated presentations of the reference stimuli, an observation that we have failed to note in previous studies (Gifford III et al., 2005 ; Russ et al., 2008b ).
A second alternative interpretation is that vPFC neurons may not be coding decisions but may be correlated with saccadic-eye movement plans (Snyder et al., 2000 ). Our analyses indicate that such eye-movement plans cannot wholly explain our data. First, previous experiments from our group demonstrated that eye movements do not appear to be correlated with changes in neural activity in the vPFC (Gifford III et al., 2005 ). Second, using data from this study, a signal-detection metric could not identify a relationship between vPFC firing rates and the monkeys’ eye movements during the time period when they were saccading to one of the two LEDs (Russ and Cohen, unpublished observation). However, at this juncture, we cannot rule out the possibility that the vPFC is also involved in other components of the same-different task such as action selection (Miller and Cohen, 2001 ).

Conclusions

This study provides further support for the involvement of the PFC in decision-making: vPFC neurons, on average, report the monkeys’ decisions during a same-different task. Since the vPFC has been hypothesized to be at the apex of a network of auditory regions that specialize in the processing of non-spatial auditory information (Rauschecker and Tian, 2000 ; Russ et al., 2008a ,b ), future research should focus on how those regions that provide afferent input to the vPFC respond during the same-different task in order to better categorize the interactions between cortical regions.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgements

We would like to thank Heather Hersh for helpful comments on the preparation of this manuscript, Ashlee Ackelson and Selina Davis for excellent technical assistance, and Farshad Chowdhury and Lauren Wool for assistance with preliminary aspects of data collection. Yale E. Cohen was supported by grants from the NIDCD-NIH and the NIMH-NIH. Brian E. Russ was supported by an NRSA grant from the NIMH-NIH.

References

Almeida, L. B. (1988). Backpropagation in perceptrons with feedback. In Neural Computers (Neuss 1987), R. Eckmiller and Ch. von der Malsburg, eds (Berlin, Springer-Verlag), pp. 199–208.
Ashby, F. G., and Spiering, B. J. (2004). The neurobiology of category learning. Behav. Cogn. Neurosci. Rev. 3, 101–113.
Aslin, R. N., Werker, J. F., and Morgan, J. I. (2002). Innate phonetic boundaries revised. J. Acoust. Soc. Am. 112, 1257–1260.
Barbour, D. L., and Wang, X. (2003). Contrast tuning in auditory cortex. Science 299, 1073–1075.
Battiti, R. (1994). Using mutual information for selecting features in supervised neural net learning. IEEE Trans. Neural Netw. 5, 537–550.
Bendor, D., and Wang, X. (2007). Differential neural coding of acoustic flutter within primate auditory cortex. Nat. Neurosci. 10, 763–771.
Bishop, C. M. (1995). Neural Networks for Pattern Recognition. Oxford, Oxford University Press.
Blank, S. C., Bird, H., Turkheimer, F., and Wise, R. J. (2003). Speech production after stroke: the role of the right pars opercularis. Ann. Neurol. 54, 310–320.
Blank, S. C., Scott, S. K., Murphy, K., Warburton, E., and Wise, R. J. (2002). Speech production: Wernicke, Broca and beyond. Brain 125, 1829–1838.
Britten, K. H., Shadlen, M. N., Newsome, W. T., and Movshon, J. A. (1992). The analysis of visual motion: a comparison of neuronal and psychophysical performance. J. Neurosci. 12, 4745–4765.
Caruana, R., Lawrence, S., and Giles, L. (2000). Overfitting in neural nets: backpropagation, conjugate gradient, and early stopping. Adv. Neural Inf. Process. Syst. 13, 402–408.
Chauvin, Y., and Rumelhart, D. E. (eds.). (1995). Backpropagation: Theory, Architecture, and Applications. Hillsdale, NJ, Lawrence Erlbaum Associates.
Cohen, Y. E., Hauser, M. D., and Russ, B. E. (2006). Spontaneous processing of abstract categorical information in the ventrolateral prefrontal cortex. Biol. Lett. 2, 261–265.
Cohen, Y. E., Russ, B. E., Gifford, G. W. III, Kiringoda, R., and MacLean, K. A. (2004). Selectivity for the spatial and nonspatial attributes of auditory stimuli in the ventrolateral prefrontal cortex. J. Neurosci. 24, 11307–11316.
Cohen, Y. E., Theunissen, F. E., Russ, B. E., and Gill, P. (2007). The acoustic features of rhesus vocalizations and their representation in the ventrolateral prefrontal cortex. J. Neurophysiol. 97, 1470–1184.
Cristianini, N., and Shawe-Tayler, J. (2000). An Introduction to Support Vector Machines and Other Kernel-based Learning Methods. New York, NY, Cambridge University Press.
Cybenko, G. V. (1989). Approximation by superpositions of a sigmoidal function. Math. Control Signals Syst. 2, 303–314.
Darwin, C. J. (1997). Auditory grouping. Trends Cogn. Sci. 1, 327–333.
De Santis, L., Clarke, S., and Murray, M. M. (2007). Automatic and intrinsic auditory “what” and “where” processing in humans revealed by electrical neuroimaging. Cereb. Cortex 17, 9–17.
Diehl, R. L., Lotto, A. J., and Holt, L. L. (2004). Speech perception. Annu. Rev. Psychol. 55, 149–179.
Eimas, P. D., Siqueland, E. R., Jusczyk, P., and Vigorito, J. (1971). Speech perception in infants. Science 171, 303–306.
Engineer, C. T., Perez, C. A., Chen, Y. H., Carraway, R. S., Reed, A. C., Shetake, J. A., Jakkamsetti, V., Chang, K. Q., and Kilgard, M. P. (2008). Cortical activity patterns predict speech discrimination ability. Nat. Neurosci. 11, 603–608.
Fishman, Y. I., Reser, D. H., Arezzo, J. C., and Steinschneider, M. (2000). Complex tone processing in primary auditory cortex of the awake monkey. I. Neural ensemble correlates of roughness. J. Acoust. Soc. Am. 108, 235–246.
Fishman, Y. I., Volkov, I. O., Noh, M. D., Garell, P. C., Bakken, H., Arezzo, J. C., Howard, M. A., and Steinschneider, M. (2001). Consonance and dissonance of musical chords: neural correlates in auditory cortex of monkeys and humans. J. Neurophysiol. 86, 2761–2788.
Freedman, D. J., Riesenhuber, M., Poggio, T., and Miller, E. K. (2001). Categorical representation of visual stimuli in the primate prefrontal cortex. Science 291, 312–316.
Freedman, D. J., Riesenhuber, M., Poggio, T., and Miller, E. K. (2002). Visual categorization and the primate prefrontal cortex: neurophysiology and behavior. J. Neurophysiol. 88, 929–941.
Freedman, D. J., Riesenhuber, M., Poggio, T., and Miller, E. K. (2003). A comparison of primate prefrontal and inferior temporal cortices during visual categorization. J. Neurosci. 23, 5235–5246.
Geisler, W. S., and Albrecht, D. G. (1996). Visual cortex neurons in monkeys and cats: detection, discrimination, and identification. Vis. Neurosci. 14, 897–919.
Gifford, G. W. III, MacLean, K. A., Hauser, M. D., and Cohen, Y. E. (2005). The neurophysiology of functionally meaningful categories: macaque ventrolateral prefrontal cortex plays a critical role in spontaneous categorization of species-specific vocalizations. J. Cogn. Neurosci. 17, 1471–1482.
Gold, J. I., and Shadlen, M. N. (2007). The neural basis of decision making. Annu. Rev. Neurosci. 30, 535–574.
Green, D. M., and Swets, J. A. (1966). Signal Detection Theory and Psychophysics. New York, John Wiley and Sons, Inc.
Griffiths, T. D., and Warren, J. D. (2004). What is an auditory object? Nat. Rev. Neurosci. 5, 887–892.
Grunewald, A., Bradley, D. C., and Andersen, R. A. (2002). Neural correlates of structure-from-motion perception in macaque V1 and MT. J. Neurosci. 22, 6195–6207.
Gu, Y., DeAngelis, G. C., and Angelaki, D. E. (2007). A functional link between area MSTd and heading perception based on vestibular signals. Nat. Neurosci. 10, 1038–1047.
Guenther, F. H., Nieto-Castanon, A., Ghosh, S. S., and Tourville, J. A. (2004). Representation of sound categories in auditory cortical maps. J. Speech Lang. Hear. Res. 47, 46–57.
Hauser, M. D. (1998). Functional referents and acoustic similarity: field playback experiments with rhesus monkeys. Anim. Behav. 55, 1647–1658.
Hertz, J., Krogh, A., and Palmer, R. G. (1991). Introduction to the Theory of Neural Computation. New York, Perseus.
Hoffman, K. L., Ghazanfar, A. A., Gauthier, I., and Logothetis, N. K. (2008). Category-specific responses to faces and objects in primate auditory cortex. Front. Syst. Neurosci. 1, 1–8. doi: 10.3389/neuro.3306.3002.2007.
Jain, A., and Zongker, D. (1997). Feature selection: evaluation, application, and small sample performance. IEEE Trans. Pattern Anal. Mach. Intell. 19, 153–158.
Kaas, J. H., and Hackett, T. A. (2000). Subdivisions of auditory cortex and processing streams in primates. Proc. Natl. Acad. Sci. USA 97, 11793–11799.
Kawahara, H., Masuda-Katsuse, I., and de Cheveigne, A. (1999). Restructuring speech representations using a pitch-adaptive time-frequency smoothing and an instantaneous-frequency-based F0 extraction. Speech Commun. 27, 187–199.
Kuhl, P. K., and Miller, J. D. (1975). Speech perception by the chinchilla: voiced-voiceless distinction in alveolar plosive consonants. Science 190, 69–72.
Kuhl, P. K., and Miller, J. D. (1978). Speech perception by the chinchilla: identification function for synthetic VOT stimuli. J. Acoust. Soc. Am. 63, 905–917.
Kuhl, P. K., and Padden, D. M. (1982). Enhanced discriminability at the phonetic boundaries for the voicing feature in macaques. Percept. Psychophys. 32, 542–550.
Kuhl, P. K., and Padden, D. M. (1983). Enhanced discriminability at the phonetic boundaries for the place feature in macaques. J. Acoust. Soc. Am. 73, 1003–1010.
LeCun, Y., Bottou, L., Bengio, Y., and Haffner, P. (1998). Gradient-based learning applied to document recognition. Proc. IEEE 86, 2278–2324.
Liu, Z., and Wang, Y. (1998). Audio feature extraction and analysis for scene segmentation and classification. J. VLSI Signal Process. 20, 61–79.
Maier, J. X., Chandrasekaran, C., and Ghazanfar, A. A. (2008). Integration of bimodal looming signals through neuronal coherence in the temporal lobe. Curr. Biol. 18, 963–968.
Micheyl, C., Tian, B., Carlyon, R. P., and Rauschecker, J. P. (2005). Perceptual organization of tone sequences in the auditory cortex of awake macaques. Neuron 48, 139–148.
Miller, C. T., and Cohen, Y. E. (in press). Vocalization Processing. In Primate Neuroethology, A. Ghazanfar and M. L. Platt, eds (Oxford, Oxford University Press).
Miller, E., and Cohen, J. D. (2001). An integrative theory of prefrontal cortex function. Annu. Rev. Neurosci. 24, 167–202.
Miller, E. K., Freedman, D. J., and Wallis, J. D. (2002). The prefrontal cortex: categories, concepts, and cognition. Philos. Trans. R. Soc. Lond., B, Biol. Sci. 29, 1123–1136.
Murray, M. M., Camen, C., Gonzalez Andino, S. L., Bovet, P., and Clarke, S. (2006). Rapid brain discrimination of sounds of objects. J. Neurosci. 26, 1293–1302.
Näätänen, R. (1992). Attention and Brain Function. Hillsdale, NJ, Lawrence Erlbaum Associates.
Nelken, I., Fishbach, A., Las, L., Ulanovsky, N., and Farkas, D. (2003). Primary auditory cortex of cats: feature detection or something else? Biol. Cybern. 89, 397–406.
Nelson, M. C., and Illingworth, W. T. (1991). A Practical Guide to Neural Nets. Reading, MA, Addison-Wesley.
Ohl, F. W., Scheich, H., and Freeman, W. J. (2001). Change in pattern of ongoing cortical activity with auditory category learning. Nature 412, 733–736.
Orr, L. E., Russ, B. E., and Cohen, Y. E. (2008). Disruption of decision-making capacities in the rhesus macaque by prefrontal cortex TMS. In Proceedings of the Program No. 875.24. 2008 Neuroscience Meeting Planner. Washington, DC, Society for Neuroscience.
Pineda, F. J. (1987). Generalization of back-propagation to recurrent neural networks. Phys. Rev. Lett. 59, 2229–2232.
Poeppel, D., Guillemin, A., Thompson, J., Fritz, J., Bavelier, D., and Braun, A. R. (2004). Auditory lexical decision, categorical perception, and FM direction discrimination differentially engage left and right auditory cortex. Neuropsychologia 42, 183–200.
Poremba, A., Malloy, M., Saunders, R. C., Carson, R. E., Herscovitch, P., and Mishkin, M. (2004). Species-specific calls evoke asymmetric activity in the monkey’s temporal poles. Nature 427, 448–451.
Prechelt, L. (1998). Early stopping – but when? Lect. Notes Comput. Sci. 1524, 55–69.
Rauschecker, J. P. (1998). Cortical processing of complex sounds. Curr. Opin. Neurobiol. 8, 516–521.
Rauschecker, J. P., and Tian, B. (2000). Mechanisms and streams for processing of “what” and “where” in auditory cortex. Proc. Natl. Acad. Sci. USA 97, 11800–11806.
Reches, A., and Gutfreund, Y. (2008). Stimulus-specific adaptations in the gaze control system of the barn owl. J. Neurosci. 28, 1523–1533.
Rohwer, R., and Forrest, B. (1987). Training time-dependence in neural networks. In IEEE First International Conference on Neural Networks (San Diego 1987), Vol. II, M. Caudill and C. Butler, eds (New York, IEEE), pp. 701–708.
Romanski, L. M., Averbeck, B. B., and Diltz, M. (2005). Neural representation of vocalizations in the primate ventrolateral prefrontal cortex. J. Neurophysiol. 93, 734–747.
Romanski, L. M., Bates, J. F., and Goldman-Rakic, P. S. (1999a). Auditory belt and parabelt projections to the prefrontal cortex in the rhesus monkey. J. Comp. Neurol. 403, 141–157.
Romanski, L. M., Tian, B., Fritz, J., Mishkin, M., Goldman-Rakic, P. S., and Rauschecker, J. P. (1999b). Dual streams of auditory afferents target multiple domains in the primate prefrontal cortex. Nat. Neurosci. 2, 1131–1136.
Romanski, L. M., and Goldman-Rakic, P. S. (2002). An auditory domain in primate prefrontal cortex. Nat. Neurosci. 5, 15–16.
Rosenblatt, F. (1962). Principles of Neurodynamics: Perceptrons and Theory of Brain Mechanism. Washington, DC, Spartan Books.
Russ, B. E., Ackelson, A. L., Baker, A. E., and Cohen, Y. E. (2008a). Coding of auditory-stimulus identity in the auditory non-spatial processing stream. J. Neurophysiol. 99, 87–95.
Russ, B. E., Orr, L. E., and Cohen, Y. E. (2008b). Prefrontal neurons predict choices during an auditory same-different task. Curr. Biol. 18, 1483–1488.
Russ, B. E., Lee, Y.-S., and Cohen, Y. E. (2007). Neural and behavioral correlates of auditory categorization. Hear. Res. 229, 204–212.
Scott, S. K. (2005). Auditory processing – speech, space and auditory objects. Curr. Opin. Neurobiol. 15, 197–201.
Scott, S. K., Blank, C. C., Rosen, S., and Wise, R. J. (2000). Identification of a pathway for intelligible speech in the left temporal lobe. Brain 123(Pt 12), 2400–2406.
Scott, S. K., Rosen, S., Wickham, L., and Wise, R. J. (2004). A positron emission tomography study of the neural basis of informational and energetic masking effects in speech perception. J. Acoust. Soc. Am. 115, 813–821.
Scott, S. K., and Wise, R. J. (2004). The functional neuroanatomy of prelexical processing in speech perception. Cognition 92, 13–45.
Selezneva, E., Scheich, H., and Brosch, M. (2006). Dual time scales for categorical decision making in auditory cortex. Curr. Biol. 16, 2428–2433.
Shepard, R. N. (1987). Toward a universal law of generalization for psychological science. Science 237, 1317–1323.
Siegelmann, H. T., and Sontag, E. D. (1991). Turing computability with neural networks. Appl. Math. Lett. 4, 77–80.
Snyder, L. H., Batista, A. P., and Andersen, R. A. (2000). Intention-related activity in the posterior parietal cortex: a review. Vis. Res. 40, 1433–1441.
Spence, K. W. (1937). The differential response in animals to stimuli varying within a single dimension. Psychol. Rev. 44, 430–444.
Steinschneider, M., Schroeder, C. E., Arezzo, J. C., and Vaughan, H. G. Jr. (1995). Physiologic correlates of the voice onset time boundary in primary auditory cortex (A1) of the awake monkey: temporal response patterns. Brain Lang. 48, 326–340.
Sternberg, S. (2001). Separate modifiability, mental modules, and the use of pure and composite measures to reveal them. Acta Psychol (Amst.) 106, 147–246.
Strange, B. A., Henson, R. N., Friston, K. J., and Dolan, R. J. (2000). Brain mechanisms for detecting perceptual, semantic, and emotional deviance. Neuroimage 12, 425–433.
Sussman, E. S. (2004). Integration and segregation in auditory scene analysis. J. Acoust. Soc. Am. 117, 1285–1298.
Trier, D., Jain, A. K., and Taxt, T. (1996). Feature extraction methods for character recognition – a survey. Pattern Recognit. 29, 641–662.
Turk, M., and Pentland, M. (1991). Eigenfaces for recognition. J. Cogn. Neurosci. 3, 71–86.
Ulanovsky, N., Las, L., and Nelken, I. (2003). Processing of low-probability sounds by cortical neurons. Nat. Neurosci. 6, 391–398.
Ungerleider, L. G., and Mishkin, M. (1982). Two cortical visual systems. In Analysis of Visual Behavior, D. J. Ingle et al., eds (Cambridge, MA, MIT Press).
Wang, X., and Kadia, S. C. (2001). Differential representation of species-specific primate vocalizations in the auditory cortices of marmoset and cat. J. Neurophysiol. 86, 2616–2620.
Wang, X., Lu, T., Snider, R. K., and Liang, L. (2005). Sustained firing in auditory cortex evoked by preferred stimuli. Nature 435, 341–346.
Widrow, B., and Lehr, M. A. (1990). 30 years of adaptive neural networks. Proc. IEEE 78, 1415–1442.
Wise, R. J., Scott, S. K., Blank, S. C., Mummery, C. J., Murphy, K., and Warburton, E. A. (2001). Separate neural subsystems within ‘Wernicke’s area’. Brain 124, 83–95.
Zatorre, R. J., Bouffard, M., and Belin, P. (2004). Sensitivity to auditory object features in human temporal neocortex. J. Neurosci. 24, 3637–3642.
Keywords:
auditory, decision-making, prefrontal cortex, phoneme, rhesus, category
Citation:
Lee JH, Russ BE, Orr LE and Cohen YE (2009). Prefrontal activity predicts monkeys’ decisions during an auditory category task. Front. Integr. Neurosci.3:16. doi: 10.3389/neuro.07.016.2009
Received:
16 April 2009;
 Paper pending published:
26 May 2009;
Accepted:
14 June 2009;
 Published online:
30 June 2009.

Edited by:

Jennifer M. Groh, Duke University, USA

Reviewed by:

Amy Poremba, University of Iowa, USA
Kari L. Hoffman, York University, Canada
Copyright:
© 2009 Lee, Russ, Orr and Cohen. This is an open-access article subject to an exclusive license agreement between the authors and the Frontiers Research Foundation, which permits unrestricted use, distribution, and reproduction in any medium, provided the original authors and source are credited.
*Correspondence:
Yale E. Cohen, Department of Otorhinolaryngology: Head and Neck Surgery, University of Pennsylvania School of Medicine, 3400 Spruce-5 Ravdin, Philadelphia, PA 19104, USA. e-mail: ycohen@mail.med.upenn.edu
Download