Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 04 March 2014
Sec. Decision Neuroscience
This article is part of the Research Topic Toward a Unified View of the Speed-Accuracy Trade-Off: Behaviour, Neurophysiology and Modelling View all 12 articles

Using reaction times and binary responses to estimate psychophysical performance: an information theoretic analysis

  • Psychology Department, Sheffield University, Sheffield, UK

As the strength of a stimulus increases, the proportions of correct binary responses increases, which define the psychometric function. Simultaneously, mean reaction times (RT) decrease, which collectively define the chronometric function. However, RTs are traditionally ignored when estimating psychophysical parameters, even though they may provide additional Shannon information. Here, we extend Palmer et al's (2005) proportional-rate diffusion model (PRD) by: (a) fitting individual RTs to an inverse Gaussian distribution, (b) including lapse rate, (c) point-of-subjective-equality (PSE) parameters, and, (d) using a two-alternative forced choice (2AFC) design based on the proportion of times a variable comparison stimulus is chosen. Maximum likelihood estimates of mean RT values (from fitted inverse Gaussians) and binary responses were fitted both separately and in combination to this extended PRD (EPRD) model, to obtain psychophysical parameter values. Values estimated from binary responses alone (i.e., the psychometric function) were found to be similar to those estimated from RTs alone (i.e., the chronometric function), which provides support for the underlying diffusion model. The EPRD model was then used to estimate the mutual information between binary responses and stimulus strength, and between RTs and stimulus strength. These provide conservative bounds for the average amount of Shannon information the observer gains about stimulus strength on each trial. For the human experiment reported here, the observer gains between 2.68 and 3.55 bits/trial. These bounds are monotonically related to a new measure, the Shannon increment, which is the expected value of the smallest change in stimulus strength detectable by an observer.

1. Introduction

For over a 100 years, it has been known that the ability to discriminate between two stimuli increases as a sigmoidal function of the difference between those stimuli, where this is traditionally measured using binary observer responses. However, when an observer makes a response, there is a trade-off between speed, or reaction time (RT), and accuracy of responses. This speed-accuracy trade-off has been the subject of numerous papers, notably (Ratcliff, 1978; Harvey, 1986; Swanson and Birch, 1992; Wichmann and Hill, 2001; Palmer et al., 2005), and more recently in Bonnet et al. (2008).

Here, we propose four extensions to the proportional-rate diffusion model (PRD) proposed in Palmer et al. (2005). First, we introduce a new parameter, the point-of-subjective-equality (PSE), which takes account of systematic shifts or bias in observer perception. This parameter is incorporated into the chronometric and psychometric functions. Second, we use a maximum likelihood estimate (MLE) of the RT mean based on a physically motivated diffusion model of RTs which involves fitting individual RTs to an inverse Gaussian distribution. Third, we take account of lapses in observer concentration by introducing a lapse rate parameter, which is estimated simultaneously with other psychophysical parameters. Fourth, we use a two-alternative forced choice (2AFC) design where the psychometric function is defined, not by the proportion of correct responses (range 50–100%), but by the proportion of times a variable comparison stimulus is chosen in preference to a fixed reference stimulus (range 0–100%). Note that the 2AFC experimental procedure is the same whether one chooses to measure the proportion of correct responses or the proportion of times a variable comparison stimulus is chosen.

Once the model has been fitted to these data, it can be used to estimate the mutual information (Shannon and Weaver, 1949; MacKay, 2003; Stone, 2014) between binary responses and stimulus strength, and between RT and stimulus strength. Finally, the mutual information provides a value for the Shannon increment, which is the expected value of the smallest change in stimulus strength detectable by an observer.

2. The Proportional-Rate Diffusion Model

We provide a brief summary of Palmer et al's PRD model (Palmer et al., 2005) here, and describe extensions below. In the experiment described in Palmer et al. (2005), an observer is presented with an array of moving dots. Stimulus strength x is defined by coherence (i.e., the percentage of dots moving in the same direction), and the observer is required to indicate which one of two directions the dots are moving. Note that coherence, and therefore stimulus strength x, varies between zero and some upper bound.

The PRD model is based on a diffusion model of RT, where the mean RT τPRD varies as a sigmoidal function of x

τ¯PRD=AKxtanh(KAx)+τ¯res,    (1)

where K is a measure of observer sensitivity, and A represents a decision boundary associated with RT. The first term on the right hand side represents the time to make a decision, and τres is a fixed residual RT (e.g., time to respond after a decision is made). Notice that this model requires that the mean RT τPRD decreases monotonically as the motion signal increases above zero, a requirement which will be relaxed in the model proposed below.

Within the PRD model, the probability PPRD of making a correct response is defined by the logistic psychometric function

PPRD=11+e2AK|x|,    (2)

where |x| indicates the absolute value of x. In Equation (2), the product AK acts as a single parameter which modulates the steepness of the sigmoidal function, and therefore acts as a measure of sensitivity to changes in stimulus strength. Note that the stimulus strength cannot fall below zero in Palmer et al's moving dot experiment, and that, when the stimulus motion strength is x = 0%, the observer has to guess, so that PPRD = 0.5, whereas if x = 100% then PPRD = 1.0.

3. The Extended Proportional-Rate Diffusion (EPRD) Model

The model proposed here is based on the assumption that responses arise from a two-alternative forced choice (2AFC) procedure. On each trial, the observer is presented with two stimuli, and the task is to choose the stronger stimulus, where strength can be defined in terms of differences in any physical quantity, such as speed, luminance, or contrast. The two stimuli are a reference stimulus with a stimulus value sR that remains constant within a specific subset of trials, and a comparison stimulus with a value sC that varies between trials. A comparison response is obtained if the observer chooses the comparison stimulus. The stimulus strength x within one trial is defined as the difference between the reference value sR and the comparison value sC, specifically x = sCsR.

We measure performance in terms of the proportion P of times that a variable comparison stimulus is chosen in preference to the fixed reference stimulus, which we define as a comparison stimulus response, so P varies between zero and one. A direct translation from PPRD to P would guarantee that a stimulus strength of zero corresponds to P = 0.5. However, if observer perception is biased, such that a stimulus difference of x = 0 is not perceived as zero, then a stimulus strength of zero would not coincide with P = 0.5. This perceptual bias can be accommodated with a second modification, a new parameter sPSE, which is the point-of-subjective-equality (PSE) between the comparison and reference stimuli. Specifically, sPSE is the value sC of the comparison stimulus which is perceived to be the same as the value sR of reference stimulus.

Given that the stimulus strength is x = sCsR, the perceived stimulus strength x′ is

x=sCsPSE    (3)
    =xΔx,    (4)

where Δx is the error in the perceived value of sC. The probability of choosing the comparison stimulus is defined as

P=11+e2AKx.    (5)

Note that the product AK effectively acts as a single parameter, and will be treated as such for binary response data (but not for RT data, see below).

In order to take account of observer lapses in concentration, which result in a pure guess, we introduce a lapse rate parameter γ. Evidence presented in Wichmann and Hill (2001) suggests that failure to take account of the lapse rate can lead to substantial errors in estimated psychophysical parameter values. If the lapse rate were zero then we would expect that P = 0 for highly negative stimulus strengths, and that P = 1 for highly positive stimulus strengths, so that observed deviations from P = 0 and P = 1 at extreme stimulus strengths can be used to provide an estimate of the lapse rate. Thus, the lapse rate parameter limits the lower and upper bounds of the psychometric function to Pmin = γ/2 and Pmax = 1 − γ/2, respectively, such that1

P=[11+e2AKx0.5] (1γ)+0.5.    (6)

Thus, the three parameters to be estimated for Equation (6) define the vector variable

θP=(sPSE,AK,γ).    (7)

Similarly, we model the observer's mean RT for a perceived stimulus strength x′ as

τ¯=Akxtanh (KAx)+τ¯res.    (8)

Here, the effects of A and K are separable, and so the four parameters to be estimated for Equation (8) define the vector variable

θτ=(sPSE,A,K,τ¯res).    (9)

The lapse rate parameter is not included here because lapses have no predictable effect on RT.

Finally, we can adapt results from Luce (1986) and Palmer et al. (2005) to relate RT to response probability. The mean decision time is defined as τdec = τiτres, so that Equations (5, 8) can be combined to provide a mapping between mean decision time τdec and the probability P of choosing the comparison stimulus

τ¯dec=(A/K)(2P1)x.(10)

Thus, if the perceived stimulus strength x′ has a large positive or negative value then P = 0 or P = 1 (respectively), and so τdec = A/(K|x′|) in both cases. This predicts that, for a given perceived stimulus strength, the probability of choosing the comparison stimulus is proportional to the mean decision time.

4. Using Observer Responses

For each trial, we obtain a RT and a binary response from the observer, which indicates whether the observer has chosen the comparison stimulus or the reference stimulus. At each stimulus strength xi, the comparison and reference stimuli are presented to the observer on Ni trials, and the number of times the observer chooses the comparison and reference stimulus is recorded as ni and Nini, respectively. For a given putative value of Pi, a standard binomial model gives the probability of the observed binary responses as

p(ni|Ni,Pi)=CNini×Pini×(1Pi)Nini,    (11)

where Pi is a function of the parameters Ak, γ and PSE as defined in Equation (6). The maximum likelihood estimate of Pi is the proportion of comparison stimulus responses Pi = ni/Ni.

When considered over all Nx values of x, the probability of observing the set of all binary responses is defined by the log likelihood function

LP=logi=1NxCNiniPini(1Pi)Nini    (12)
     =i=1NxnilogPi+i=1Nx(Nini)log(1Pi)+i=1NxlogCNini,     (13)

where the final term does not depend on parameter values, and can be discarded unless the exact value of the likelihood is required. Recall that each Pi is determined by Equation (6), which is a function of the EPRD parameter values θP = (A, K, γ, PSE). The maximum likelihood estimate (MLE) of θP is obtained by finding EPRD parameter values θP that maximize LP.

If the number of trials at each stimulus strength is large then Equation (13) can be approximated by a Gaussian function. At a given stimulus strength xi, the observed proportion of binary responses is Pi, which is assumed to be the probability Pi plus a noise term ηP, so that Pi = Pi + ηP. If the noise ηP has a Gaussian distribution with variance vP,i then

p(Pi|A,k,xi)=12πvP,iexp(PiPi)22vP,i,    (14)

where Pi is defined as a function of A, k, x′ in Equation (6), and the variances vP,i can be estimated from the data as vP,i = NiPi(1 − Pi). Results for the Gaussian approximation in Equation (14) were found to be very similar to those for Equation (13). Results reported here are based on Equation (13).

5. Using Reaction Times

RTs tend to be short if the comparison stimulus value is very different from the reference stimulus, but as the comparison and reference stimuli become more similar, so the RT increases, as shown in Figure 4B. Here, we use RTs in a two stage process. First, a mean RT value is estimated at each stimulus strength. These mean RT values are then used as data for the RTτ model, which is used to estimate EPRD model parameters.

5.1. Inverse Gaussian Model of Individual RTs

It is commonly assumed that the RT is the time required for the cumulative amount of perceptual evidence to reach some criterion value (Ratcliff, 1978; Smith, 1990). Specifically, this evidence accumulation is assumed to consist of a Brownian diffusion process with positive drift, which can be likened to a the total distance traveled in a one-dimensional biased random walk. If a Brownian process is allowed to run for a fixed time then it is well known that the final distribution of values (e.g., evidence) has a Gaussian distribution. However, it is less well known that if a Brownian diffusion process is allowed to run until it reaches a fixed criterion value then the time taken to reach that value has an inverse Gaussian or Wald distribution (see Figure 3). Therefore, if the amount of evidence required to make a response is stable for a given observer then RTs are appropriately modeled using an inverse Gaussian distribution2.

FIGURE 1
www.frontiersin.org

Figure 1. How the entropy H(x) in stimulus strength x is accounted for by the entropy H(τ) in RT (τ) and entropy H(P) in the probability P of a particular binary response r. The entropies of x, P, and τ are represented by the discs X, Y, and Z, respectively. The mutual information between x and P is I(x, P) = (a + b), and the mutual information between x and τ is I(x, τ) = (a + c).

FIGURE 2
www.frontiersin.org

Figure 2. Schematic illustration of typical stimulus shown to observer on a single trial. The observer has to choose the face that looks wider. The stimulus in the experiment used was a picture of the actor James Corden's face, with all background details removed (see http://illusionoftheyear.com/2010/the-fat-face-thin-fft-illusion).

FIGURE 3
www.frontiersin.org

Figure 3. Reaction times fitted with an inverse Gaussian (Equation 15). Each dot represents 1 of 20 RTs for a stimulus value (width scaling) of s = 1.05.

If RTs have an inverse Gaussian distribution with mean τi then the probability of a single observed RT τij associated with the jth presentation of the stimulus value xi is

p(τij|τ¯i,λi)=(λi2πτij3)1/2×exp [λi(τijτ¯i)22τ¯i2τij],    (15)

where the variance of this distribution is

vτi=τ¯i33/λi.    (16)

Each of the Nx stimulus strengths is presented Ni times. For one model RT mean, the probability of the observed Ni RTs (one RT per trial) defines the log likelihood function

Lτ,i=logj=1Nip(τij|τ¯i,λi).    (17)

Maximizing Equation (17) with respect to the parameters τi and λi yields a maximum likelihood estimate (MLE) of both parameters at one stimulus strength xi. Even though the algebraic mean and the MLE mean are identical (Tweedie, 1957) for the inverse Gaussian, the fitting process provides the parameter estimate λi, which is vital for subsequent calculations.

5.2. Model RTτ: Using Mean Reaction Times

For a given stimulus strength xi, the predicted mean RT τi varies as a tanh function of xi, as defined in Equation (8). The central limit theorem allows us to assume that the distribution of mean RTs of the inverse Gaussian pdf at a given stimulus strength xi is Gaussian with mean τi and variance vτ,i. Therefore, the likelihood of the EPRD mean τi from Equation (8) is

p(τ¯i|τ¯i(θτ))=12πvτ¯,ie(τ¯iτ¯i)2/(2vτ¯,i).    (18)

The variance of an inverse Gaussian distribution of RT values with mean τi is vτi (Equation 16), so the variance vτi of a distribution of means (where each mean is based on Ni samples) is

vτ¯i=τ¯i3λiNi.    (19)

Thus, we can assess the fit of the inverse Gaussian mean RTs τi to the EPRD mean RTs τi of Equation (8) as follows. The probability of the Nx mean RTs τi (one mean RT per stimulus strength) defines the log likelihood function

Lτ¯=logi=1Nxp(τ¯i|τ¯i)    (20)
    =1/2i=1Nx(τ¯iτ¯i)2vτ¯,i1/2i=1Nxlog2πvτ¯,i,    (21)

where τi is defined in Equation (8), so that the parameters to be estimated for model RTτ are θτ = (A, k, γ, PSE, τres) to fit the overall variation in mean RT with stimulus strength x.

In summary, we have three estimates of the mean RT at each stimulus strength: the algebraic mean τobsi, the MLE mean of the inverse Gaussian or Wald pdf τi (from Equation 17), which collectively are used as data to estimate the means τi (one per stimulus strength) obtained from the fitted EPRD model (from Equation 21). The MLE means τi are shown as crosses in Figure 4B, and the means τi are corresponding points on the fitted curve, respectively.

FIGURE 4
www.frontiersin.org

Figure 4. The psychometric function (A) and chronometric function (B), from the face inversion experiment for one observer. The width scaling factor s applied to the comparison image is indicated on the abscissa. The vertical dashed line marks the point-of-subjective-equality (PSE) at s = 1.032. (A) Each dot represents the observed proportion of trials for which the observer chose the comparison stimulus, and the fitted psychometric function is defined in Equation 6. (B) Each dot represents the RT of a single trial for the same responses as in Figure 4A (RTs greater than 2 s are not shown). The fitted chronometric function is defined in Equation 8. The dashed curve joins the fitted (inverse Gaussian) mean RTs, each of which was obtained by maximizing Equation 17. The solid curves in (A, B) (Equations 6, 8, respectively) were fitted using combined binary and mean RT data by maximizing Equation 22. A graph similar to (A) was obtained for model LP (i.e., using only binary responses), and a graph similar to (B) was obtained for model Lτ (i.e., using only mean RTs).

We also have two estimates of the probability of a comparison stimulus response at each stimulus strength: the observed proportion of comparison stimulus responses (which is the MLE Pi = ni/Ni), and the mean Pi (one per stimulus strength) obtained from fitting the EPRD model (Equation 13) to the MLE means Pi. These are shown as dots in Figure 4A, and as corresponding points on the fitted curve, respectively.

6. Using Binary Responses and RTs

In the absence of knowledge regarding the covariance between the noise in mean RT and binary response probability, we are forced to assume this covariance is zero. In other words, we assume that LP and Lτ provide independent estimates of the EPRD model parameters. In this case, estimates based on combined RT and binary response probability are obtained by maximizing the sum of likelihoods

LC=LP+Lτ¯.    (22)

However, the implausibility of this independence assumption means that we will not take seriously any results based on Equation (22).

7. Information Theory

The amount of Shannon information (Shannon and Weaver, 1949; MacKay, 2003; Stone, 2014) that the observer gains about the stimulus is reflected in both the binary responses and RTs. Specifically, the average Shannon information that each mean RT provides about the stimulus strength x is the mutual information I(x, τ) between x and the mean RT. Similarly, the average Shannon information that binary responses provide about the stimulus strength x is the mutual information I(x,P) between x and the probability of a comparison stimulus binary response.

More importantly, the total amount of Shannon information that the observer has about the stimulus cannot be less than the amount of Shannon information implicit in the observer's combined binary and RT responses. In other words, the total mutual information, as measured by an experimenter, between observer responses and stimulus strength provides a lower bound for the amount of Shannon information that the observer has about the stimulus strength. Thus, each the mutual information value provided in this paper constitutes a conservative estimate of the amount of information that the observer gains about the stimulus.

7.1. Evaluating I(x, P)

The mutual information I(x, P) between stimulus strength s and the probability P that the observer chooses the comparison stimulus (i.e., r = 1) is

I(x,P)=xPp(x,P)logp(x,P)p(x)p(P)dPdx    (23)
          =H(x)+H(P)H(x,P) bits,    (24)

where H(x) and H(P) are the differential entropies of p(x) and p(P), respectively, and H(x, P) is the differential entropy of the joint distribution p(x, P). All logarithms in this paper use base 2, so information is measured in bits. Substituting p(x, P) = p(P|x)p(x), yields

I(x,P)=xp(x)Pp(P|x)logp(P|x)p(P)dPdx    (25)
          =H(P)H(P|x) bits,    (26)

where H(P|x) is the differential entropy of the noise in the measurements P. Given Bayes' rule, p(P|x) = p(x|P)p(P)/p(x), we can recognize the mutual information as the differential entropy H(P) of the prior distribution minus the differential entropy H(P|x) of the posterior distribution.

We can evaluate Equation (25) by summing over discrete versions of the variables x and P. Recall that the observed proportion of responses r = 1 at a given stimulus strength xi is Pi = ni/Ni, so that

I(x,P)=k=1Nxp(xk) [i=1Nxp(Pi|xk)logp(Pi|xk)p(Pi)] bits.    (27)

We assume that the probability of stimulus values is locally uniform, so that p(xk) = 1/Nk. In order to evaluate Equation (27), we require expressions for p(Pi|xk) and p(Pi).

7.1.1. Evaluating the posterior p(Pi|xk)

Using Equation (5) across a range of x values, the fitted value of P at xk is Pk. Assuming a binomial distribution, the probability of the observed proportion Pi given a fitted value Pk at xk is

p(Pi|xk)=CNiniPkni(1Pk)Nini,    (28)

where p(Pi|xk) = p(Pi|Pk), and p(Pi|xk) values are normalized to ensure that ∑i p(Pi|xk) = 1.

7.1.2. Evaluating the prior p(Pi)

The distribution of binary responses is binomial with a mean equal to the grand mean PG of all NG binary responses of an observer

PG=1NGi=1NGri,    (29)

where ri = 1 if and only if a response corresponds to the observer choosing the comparison stimulus. The observer's prior probability of the binary responses for the ith stimulus strength is therefore

p(Pi)=CNiniPGni(1PG)Nini,    (30)

where p(Pi) values are normalized to ensure that ∑i p(Pi) = 1.

7.2. Evaluating I(x, τ)

Following the same line of reasoning as above, the mutual information I(x, τ) between stimulus strength and mean RT is

I(x,τ¯)=xp(x)τ¯p(τ¯|x)logp(τ¯|x)p(τ¯)dτ¯dx    (31)
         =H(τ¯)H(τ¯|x) bits,    (32)

where H(τ|x) is the differential entropy of the noise in the measurements τ.

We can evaluate Equation (31) by summing over discrete versions of the variables x and τ

I(x,τ¯)=k=1Nxp(xk) [i=1Nip(τ¯i|xk)logp(τ¯i|xk)p(τ¯i)]  bits,    (33)

where p(τi|xk) is defined by the EPRD model (Equation 8) with a fitted value τk, so that

p(τ¯i|xk)=p(τ¯i|τ¯k(θτ)),    (34)

as in Equation (18). As before, we assume that the probability of stimulus values is uniform, so that p(xk) = 1/Ni.

7.2.1. Evaluating the posterior p(τi|xk)

The posterior is defined in Equation (18), but is repeated here with changed subscripts for clarity

p(τ¯i|xk)=12πvτ¯kexp [(τ¯iτ¯k)22vτ¯k],    (35)

where vτk is defined in Equation (19), and p(τi|xk) values are normalized to ensure that ∑i p(τi|xk) = 1.

7.2.2. Evaluating the prior p(τi)

A parametric form for the observer's prior probability distribution p(τ) of individual RTs was estimated from the entire set of that observer's grand total of NG RTs. These were fitted to an inverse Gaussian distribution to obtain a grand mean τG and a parameter λG. This pdf has a variance

vG=τ¯G3/λG.    (36)

At each stimulus strength xi, the RT mean is based on a sample of Ni RTs, and the central limit theorem suggests that the distribution of means is approximately Gaussian with a variance

vg=vG/Ni.    (37)

Therefore, the prior probability density of each inverse Gaussian mean τi is

p(τ¯i)=12πvgexp [(τ¯iτ¯G)22vg],    (38)

where p(τi) values are normalized to ensure that ∑i p(τi) = 1.

7.3. The Shannon Information of a Single Response

So far we have derived expressions for the Shannon information implicit in the average RT τi and also in the average binary response, which is summarized as the proportion Pi of comparison responses, for a stimulus strength xi. Here, we derive an expression for the Shannon information associated with a single trial; first for RTs, and then for binary responses.

As the number of trials at each stimulus strength is increased, so the variance in each mean RT decreases, and the central limit theorem ensures that the distribution of means becomes increasingly Gaussian. The mutual information between two variables (e.g., mean RT and stimulus strength) depends on the signal to noise ratio SNR

I1/2log2(1+SNR),    (39)

where SNR is the signal variance expressed as a fraction of the noise variance in the measurement (Shannon and Weaver, 1949). If the distribution of mean RTs is Gaussian then the distribution of differences Δτ between mean RT τ and the grand mean RT (at one stimulus strength) must also be Gaussian. Because the mutual information is defined in Equation (32) to be the differential entropy of τ minus the differential entropy of the noise Δτ in τ, we can assume equality in Equation (39) (Rieke et al., 1997). In fact, we do not need to rely on the central limit theorem here, because even if the perturbing noise Δτ is not Gaussian, Shannon's Theorem 18 (Shannon and Weaver, 1949) implies equality in Equation (39), so that

I=1/2log2(1+SNR) bits.    (40)

We already have a value for the mutual information I(x, τ) from Equation (27), so we can re-arrange Equation (40) to find the SNR associated with τ

SNRτ¯=22I(x,τ¯)1 bits.    (41)

However, the mutual information I(x, τ) obtained from Equation (27) tells us how much average Shannon information each mean RT provides about stimulus strength, whereas we want to know how much average information each individual RT provides about stimulus strength. Because the value of SNR in Equation (41) is based on mean RTs, each of which involves Ni trials, the variance of the measurement noise has been reduced by a factor of Ni relative to the noise in the RT of a single trial (provided this noise is iid). This implies that the value of SNR for a single trial is

SNRτ=SNRτ¯/Ni    (42)
         =(22I(x,τ¯)1)/Ni bits.    (43)

If we substitute SNRτ into Equation (40) then we obtain an estimate of the average Shannon information I(x, τ) implicit in the observer's RT in a single trial

I(x,τ)=12log2 [1+(22I(x,τ¯)1)Ni] bits.    (44)

A similar line of reasoning implies that the average Shannon information I(x, r) implicit in the observer's binary response r in a single trial is

I(x,r)=12log2 [1+(22I(x,P)1)Ni] bits.    (45)

In order to compare mutual information estimates for the different variables τ and r, the calculations for I(x, τ) and I(x, r) should be based on the same range of stimulus strengths x.

7.4. Defining the Shannon Increment

The mutual information between stimulus strength and (binary or RT) responses can be used to define the smallest average detectable difference in stimulus strength, which we call the Shannon increment (SI). We first define the effective stimulus range xrange as the range of stimulus strengths x associated with response probabilities between P = ϵ and P = 1 − ϵ, for some small value ϵ. Then the SI is related to the mutual information I by

SI=xrange2I,    (46)

where the value 2 is based on the assumption that information is measured in bits (i.e., using log to the base 2), and SI has the same units as stimulus strength. Because SI decreases monotonically with mutual information, it should become asymptotically closer to the true value of SI as the number of trials or stimulus strengths is increased.

A brief explanation for this definition is as follows. Consider a range of stimulus strengths xrange which give rise to “noisy” observer responses y = f(x), where these responses are samples from a probability density function p(y(x)), and where the mutual information between x and y is I bits. One way to interpret SI involves assuming that p(y(x)) is uniform. In this case, on average, knowing the value of y reduces the possible range of x values to an interval Δx = xrange/2I, which we can recognize as being equal to the SI.

8. Fat-Face Thin: A Demonstration Experiment

We used the EPRD models described above to estimate the PSE and other key parameters for a simple demonstration experiment using a human observer. On each trial, the observer was presented with a colored picture of an upright face and an inverted face (see Figure 2) on a computer screen, and was required to indicate which one appeared to be wider by pressing a left/right computer key. For half of the trials, the reference stimulus was an upright face, and the comparison stimulus was an inverted version of the same face, and these were swapped for the other half of the trials. The width of the comparison image was determined by 1 of 21 stretch factors s = 0.90, 0.91, …, 1.10, but the height of both stimuli was kept constant. The stimulus strength was defined to be x = s − 1, so that x varied between −0.1 and 0.1. For a given value of si, the observer was presented with the same stimulus pair for a total of Ni = 20 trials. Stimuli were shown in random order, and the left/right position of reference/comparison stimuli was counterbalanced across trials.

8.1. Results

Each of three models defined by LP, Lτ, and LC was used to fit a psychometric and/or a chronometric function to the data from one subject, as shown in Figure 4. Maximum likelihood parameter estimation was implemented in MatLab using the Nelder–Mead simplex method. The parameter estimates for each model are summarized in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Results for three models.

8.2. Using Binary Responses: Model LP

Based on 420 binary responses, maximizing LP (Equation 12) yields a psychometric function similar to that in Figure 4A, and a PSE of sPSE = 1.031. This maximum likelihood estimate implies that an inverted face must be 3.1% wider than an upright face in order for the two faces to be perceived as the same width. Numerical estimation of the Hessian matrix of second derivatives of Equation (12) at sPSE yields a standard error (se) of 0.003, which implies that sPSE is significantly different from s = 1 (p < 0.001). The values of three parameters were estimated for this model, the PSE, Ak, and γ, and the product Ak is quoted in Table 1 for comparison with other works.

8.3. Using Mean Reaction Times: Model Lτ

Each of 21 mean RTs (one per stimulus strength) was first estimated by maximizing Equation (17), based on 20 RTs per stimulus strength. Using these 21 mean RTs, Lτ (Equation 21), was maximized with respect to four parameters (PSE, A, k, and τres) to yield a chronometric function similar to that in Figure 4B. The estimated PSE is sPSE = 1.034 (se = 0.004, p < 0.001).

8.4. Using Mean RTs and Observer Responses: Model LC

Based on 42 data points (the 21 estimated mean RTs used for Lτ plus 21 corresponding binary response probabilities used for LP), maximizing LC (Equation 22) yields the psychometric function and the chronometric function in Figures 4A,B, respectively, and a PSE of 1.032 (se = 0.003, p < 0.001). There are five parameters to be estimated for this model, the PSE, A, k, τres, and γ.

8.5. Shannon Information

The mutual information I(x, τ) between x and τ is the entropy in p(τ) and p(x) shared by the joint distribution p(x, τ). Using Equation (33), this evaluates to I(x, τ) = 2.79 bits. Using Equation (44) with Ni = 20, this implies that the mutual information I(x, τ) for a single RT is I(x, τ) = 0.87 bits, and is represented by the intersection of regions X and Z.

Similarly, Equation (27) can be used to estimate the mutual information between x and P, which comes to I(x, P) = 4.82 bits. Using Equation (45) with Ni = 20, this implies that the mutual information I(x, r) for a single binary response r is I(x, r) = 2.68 bits, and is represented by the intersection of regions X and Y.

We can use I(x, τ) and I(x, r) to provide lower and upper bounds on the total amount of mutual information Itot between x and the combined variables (r, τ), which can be considered to be a vector variable. If τ and r provide independent information about x (i.e., if a = 0 in Figure 1) then the maximum value of Itot is

max(Itot)=I(x,τ)+I(x,r)    (47)
                 =0.87+2.68    (48)
                 = 3.55bits.    (49)

However, if all of the information I(x, τ) provided by τ about x is the same as part of the information provided by r about x (i.e., if c = 0 in Figure 1) then Itot cannot be less than I(x, r). To take account of the possibility that all of the information I(x, r) provided by r about x is the same as part of the information provided by τ about x, we can write

min(Itot)=max(I(x,τ),I(x,r))    (50)
                =max(0.87,2.68)    (51)
                =2.68 bits.    (52)

Thus, on average, each trial provides the observer with between 2.68 and 3.55 bits.

8.6. Shannon Increment

Using a conservative estimate of mutual information of I = 2.68 bits suggests that the observer can discriminate differences between the reference and comparison stimulus with an average resolution of about one part in 6.39 (= 22.68) of the effective range xrange of stimulus strengths. Note that the range of scaling values used srange = 0.2 (i.e., 0.9 … 1.1) equals the range of stimulus strengths xrange = 0.2 (i.e., −0.1 … 0.1). Therefore, the SI for the width scaling factor is

SI=xrange/2I    (53)
    = 0.2/6.39    (54)
    = 0.031,    (55)

where we have assumed ϵ = 0 here. Thus, on average, the smallest change in scaling factor (between reference and comparison stimulus) detectable by the observer is SI = 0.031.

9. Discussion

We have shown how the PRD model from Palmer et al. (2005) can be extended to make use of individual RTs, which can be combined with binary observer responses to estimate key psychophysical parameters in a 2AFC design.

A key feature of diffusion-based models is that they treat each RT as the end-point of an accumulation of evidence. If we take this type of evidence-accumulation process seriously then it makes sense to model the distribution of RT values as an inverse Gaussian distribution (for reasons described in section 5).

A striking result is the difference between the log likelihoods associated with the binary response model and the RT model, despite the fact that the binary response model has fewer free parameters than the RT model, and that both models provide similar PSE estimates which (based on their sems, not shown) are not significantly different. These log likelihood values suggest that the EPRD model provides a better fit to the RT data than it does to the binary response data. This difference in likelihoods suggests that the parameter estimates obtained using the combined RT and response data is dominated by the binary data likelihood term.

Self-evidently, both the RT and binary responses of an observer depend on the stimulus strength x. However, in general, it is not known if RT or binary response data provide more Shannon information about the value of x. More importantly, and more subtley, it is not known if they provide the same information about x, or if they merely provide the same amount of information about x (see Figure 1).

We can gain some insight into the nature of this problem by considering the proportion of the differential entropy in stimulus values accounted for by the corresponding differential entropy in observer responses. At one extreme, if an observer is told to respond as quickly as possible then the RTs should provide relatively large amounts of mutual information regarding stimulus strength, whereas the binary responses carry relatively little mutual information (because speeded responses tend to be inaccurate Hanks et al., 2011). In this case, the RT entropy at a given stimulus strength will be relatively small, because RTs will be tightly coupled to the stimulus strength, whereas the binary response entropy at a given stimulus strength will be relatively large (because these responses are inaccurate, and therefore not tightly coupled to the stimulus strength). However, when considered across different stimulus strengths, the tight coupling between RT and stimulus strength will give rise to a relatively large RT entropy, and most of this entropy will be shared with stimulus strength entropy (which defines a large mutual information between RT and stimulus strength). In contrast, these fast, inaccurate responses across stimulus strengths will be associated with a relatively small range of response probability values (e.g., P ≈ 0.5), which will therefore have a relatively small entropy, most of which is not shared with the stimulus strength entropy (which defines a small mutual information between binary responses and stimulus strength). In summary, fast responses should yield high entropy RT values, which share a large proportion of their entropy with the stimulus strength, combined with low entropy P values which share a small proportion of their entropy with the stimulus strength. At the other extreme, if an observer is told to be as accurate as possible then this should yield high entropy P values which share a large proportion of their entropy with the stimulus strength, combined with low entropy RT values which share a small proportion of their entropy with the stimulus strength. In summary, the entropy in stimulus strength can be shared with entropy in both accuracy (P) and speed (RT). However, as there is probably only a finite amount of such shared entropy (mutual information) available, we predict that it can be realized experimentally as maximum speed or maximum accuracy, but not both.

The scenario considered above can be represented geometrically, as in Figure 1. If we compare the mutual information between τ and x with the mutual information between r and x then it is possible that they have the same magnitude [e.g., (a + c) = (a + b), as in Figure 1]. However, the fact that both τ and x have the same amount of mutual information (i.e., they account for the same amount of entropy in x) does not imply that they account for the same entropy in x. Formally, the fact that (a + c) = (a + b) does not imply that (a + c) ≡ (a + b). This matters because, even if I(x, τ) = I(x, r), we could not conclude that I(x, τ) ≡ I(x, r), and so we could not conclude that τ and r provide mutually redundant information. Thus, we cannot dismiss τ simply because r accounts for more entropy in x than τ does (or vice versa). Indeed, this is precisely the situation that we have in the results reported here, and provides reasonable grounds for making use of both RT and binary response data in general.

Unfortunately, we have been unable to derive an expression for the total mutual information between the joint variables (RT and binary responses) and stimulus strength I(τ, P; x′) (i.e., the area [a + b + c] in Figure 1), although it may be possible to do so using Equation (10) [where the entropy of the difference between P and τ is H(τ, P|x′)]. The precise effect of the instructions given to observers on mutual information, and the proposed invariance of the total mutual information with respect to instructions, clearly require further research (Soukoreff and MacKenzie, 2009).

The Shannon increment (SI) is similar in spirit to the more conventional just noticeable difference (JND). However, the JND has an arbitrary value, and (despite its name) there is no reason to suppose that a JND is indeed just noticeable. The SI is monotonically related to the average amount of Shannon information an observer gains regarding a single presentation of a stimulus, and is a measure of the perceptual resolution with which a parameter is represented by the observer.

10. Conclusion

We have presented an extended proportional-rate diffusion model, which takes account of both individual RTs and binary responses for maximum likelihood estimation of key psychophysical parameters (e.g., PSE, slope) of the psychometric and chronometric functions. The fact that these psychophysical parameters have similar estimated values when computed independently for two models based on RTs alone or on binary responses alone provides support for the underlying physical basis of this class of diffusion models.

An information-theoretic analysis was used to estimate the average amount of Shannon information that each RT provided about the stimulus value, and also the average amount of Shannon information that each binary response provided about the stimulus value. This analysis provides bounds for the average amount of Shannon information that the observer gains about the stimulus value from one presentation, which was found to be between 2.68 and 3.55 bits/trial for the experiment used here.

Conflict of Interest Statement

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

Thanks to Steve Snow, Nathan Lepora, and Tom Stafford for reading an early draft of this paper, and to two referees for their detailed comments.

Footnotes

1. ^Notice that, if the lapse rate is γ = 0.01 then the upper and lower bounds are 0.995 and 0.005, respectively, because half of the observer's guesses will be correct, on average.

2. ^For reference, the Wald distribution is the distribution of first passage times of a biased Brownian process, and is qualitatively similar to the log-normal distribution, which is often used to model RT.

References

Bonnet, C., Ars, J., and Ferrer, S. (2008). Reaction times as a measure of uncertainty. Psicothema 20, 43–48.

Pubmed Abstract | Pubmed Full Text

Hanks, T., Mazurek, M., Kiani, R., Hopp, E., and Shadlen, M. (2011). Elapsed decision time affects the weighting of prior probability in a perceptual decision task. J. Neurosci. 31, 6339–6352. doi: 10.1523/JNEUROSCI.5613-10.2011

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Harvey, L. (1986). Efficient estimation of sensory thresholds. Behav. Res. Methods Instrum. Comput. 18, 623–632. doi: 10.3758/BF03201438

CrossRef Full Text

Luce, R. D. (1986). Response Times: Their Role in Inferring Elementary Mental Organization. New York, NY: Oxford University Press.

MacKay, D. (2003). Information Theory, Inference, and Learning Algorithms. Cambridge: Cambridge University Press.

Palmer, J., Huk, A., and Shadlen, M. (2005). The effect of stimulus strength on the speed and accuracy of a perceptual decision. J. Vis. 5, 376–404. doi: 10.1167/5.5.1

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Ratcliff, R. (1978). A theory of memory retrieval. Psychol. Rev. 85, 59–108. doi: 10.1037/0033-295X.85.2.59

CrossRef Full Text

Rieke, F., Warland, D., van Steveninck, R., and Bialek, W. (1997). Spikes: Exploring the Neural Code. Cambridge, MA: MIT Press.

Shannon, C., and Weaver, W. (1949). The Mathematical Theory of Communication. Urbana, IL: University of Illinois Press.

Smith, P. (1990). A note on the distribution of response times for a random walk with Gaussian increments. J. Math. Psychol. 34, 445–459. doi: 10.1016/0022-2496(90)90023-3

CrossRef Full Text

Soukoreff, R. W., and MacKenzie, I. S. (2009). “An informatic rationale for the speed-accuracy tradeoff,” in Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, 2009 (San Antonio, TX), 2890–2896. doi: 10.1109/ICSMC.2009.5346580

CrossRef Full Text

Stone, J. (2014). Information Theory: A Tutorial Introduction. Sheffield: Sebtel Press.

Swanson, W., and Birch, E. (1992). Extracting thresholds from noisy psychophysical data. Percept. Psychophys. 51, 409–422. doi: 10.3758/BF03211637

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Tweedie, I. M. C. K. (1957). Statistical properties of inverse Gaussian distributions. Ann. Math. Stat. 28, 362–377. doi: 10.1214/aoms/1177706881

CrossRef Full Text

Wichmann, F. A., and Hill, N. (2001). The psychometric function: I. Fitting, sampling, and goodness of fit. Percept. Psychophys. 63, 1293–1313. doi: 10.3758/BF03194544

Pubmed Abstract | Pubmed Full Text | CrossRef Full Text

Appendix

Mathematical Symbols and Abbreviations

A an EPRD model parameter which is the amount of evidence required to trigger a response.

comparison stimulus response: a response indicating the comparison stimulus was chosen.

EPRD: extended proportional-rate diffusion model.

SI: Shannon increment, the smallest detectable change in a stimulus.

γ EPRD lapse rate parameter.

i index over stimulus strength x, with range i = 1, …, Nx.

j index over trials at one stimulus strength xi, with range j = 1, …, Ni.

k index over stimulus strength, with range k = 1, …, Nx.

K is a measure of sensitivity to changes in x in the EPRD model.

Ni number of trials at stimulus strength xi.

Nx number of different stimulus strengths.

PSE: point of subjective equality.

Pi proportion of comparison stimulus responses at stimulus strength xi, predicted by EPRD model.

Pi MLE mean, equal to observed proportion of comparison responses at stimulus strength xi.

r binary observer response (e.g., observer chooses comparison or reference stimulus).

sC variable stimulus value of the comparison stimulus.

sR fixed stimulus value of the reference stimulus.

sPSE value of the comparison stimulus which the observer perceives as being the same as the reference stimulus.

τi MLE mean of inverse Gaussian RT at stimulus strength xi.

τi mean RT at stimulus strength xi, as predicted by EPRD model.

τdec,i mean decision RT at stimulus strength xi, as predicted by EPRD model.

τres mean residual RT (assumed the same at all stimulus strengths), as predicted by EPRD model, where τres = τdec,iτi.

θτ = (sPSE, A, K, γ, τres), five parameters for the RT component of the EPRD model.

θP = (sPSE, A K, γ), three parameters for the binary response component of the EPRD model.

vτ, i variance in mean RT.

xi stimulus strength.

xi perceived strength of stimulus with strength xi.

Keywords: psychometric function, chronometric function, point of subjective equality, diffusion model, reaction time, threshold, Shannon information, mutual information

Citation: Stone JV (2014) Using reaction times and binary responses to estimate psychophysical performance: an information theoretic analysis. Front. Neurosci. 8:35. doi: 10.3389/fnins.2014.00035

Received: 29 November 2013; Accepted: 10 February 2014;
Published online: 04 March 2014.

Edited by:

Dominic Standage, Queen's University, Canada

Reviewed by:

NaYoung So, Columbia University, USA
Wael Asaad, Brown University, USA

Copyright © 2014 Stone. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: James V. Stone, Psychology Department, Sheffield University, Western Bank, Sheffield S10 2TP, UK e-mail: j.v.stone@sheffield.ac.uk

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.