Abstract
Complex perceptual decisions, in which information must be integrated across multiple sources of evidence, are ubiquitous but are not well understood. Such decisions rely on sensory processing of each individual source of evidence, and are therefore vulnerable to bias if sensory processing resources are disproportionately allocated among visual inputs. To investigate this, we developed an implicit neurofeedback protocol embedded within a complex decision-making task to bias sensory processing in favor of one source of evidence over another. Human participants of both sexes (N = 30) were asked to report the average motion direction across two fields of oriented moving bars. Bars of different orientations flickered at different frequencies, thus inducing steady-state visual evoked potentials. Unbeknownst to participants, neurofeedback was implemented to implicitly reward attention to a specific “trained” orientation (rather than any particular motion direction). As attentional selectivity for this orientation increased, the motion coherence of both fields of bars increased, making the task easier without altering the relative reliability of the two sources of evidence. Critically, these neurofeedback trials were alternated with “test” trials in which motion coherence was not contingent on attentional selectivity, allowing us to assess the training efficacy. The protocol successfully biased sensory processing, resulting in earlier and stronger encoding of the trained evidence source. In turn, this evidence was weighted more heavily in behavioral and neural representations of the integrated average, although the two sources of evidence were always matched in reliability. These results demonstrate how biases in sensory processing can impact integrative decision-making processes.
SIGNIFICANCE STATEMENT Many everyday decisions require active integration of different sources of sensory information, such as deciding when it is safe to cross a road, yet little is known about how the brain prioritizes sensory sources in the service of adaptive behavior, or whether such decisions can be altered through learning. Here we addressed these questions using a novel behavioral protocol that provided observers with real-time feedback of their own brain activity patterns in which sensory processing was implicitly biased toward a subset of the available information. We show that, while such biases are a normal and adaptive mechanism for humans to process complex visual information, they can also contribute to suboptimal decision-making.
- feature-based attention
- forward encoding
- implicit neurofeedback
- integrative decision-making
- sensory processing
- SSVEPs
Introduction
To interact with a complex and dynamic visual world, humans accumulate samples of noisy evidence to guide perceptual decisions (Ratcliff and Rouder, 1998; Usher and McClelland, 2001). This process is well characterized for decisions based on isolated sources of evidence (Summerfield and Tsetsos, 2015) but is less understood for decisions that require observers to integrate information across multiple sources (Spitzer et al., 2017). Such “integrative” decisions are particularly vulnerable to bias because different sources of evidence can be subject to different degrees of prioritization of sensory processing (Mostert et al., 2015). Indeed, internal biases in the allocation of sensory resources are ubiquitous in visual processing (Appelle, 1972; Yantis, 1993; Theeuwes et al., 2010; Anderson, 2016; Failing and Theeuwes, 2018). Typical behavioral investigations probe such biases by introducing “bottom-up” differences in evidence presentation, manipulating factors, such as stimulus reliability (Fetsch et al., 2013; Zylberberg et al., 2016), probability, or the reward associated with different visual features (Gao et al., 2011; Summerfield and De Lange, 2014). To eliminate these bottom-up sources of bias, we embedded implicit neurofeedback into an integrative decision-making task, unconsciously rewarding observers for preferential attentional allocation to one source of evidence over another. An ideal observer should give equal weight to two overlapping stimuli of identical reliability (Rahnev and Denison, 2018). Here we asked whether neurofeedback can bias (or “adapt”) neural representations in favor of one source of evidence over another, even though this might be statistically suboptimal. This novel approach allowed us to directly assess how differences in the neural processing of concurrent stimuli relate to how they are weighted in integrative decisions.
During perceptual decisions, observers repeatedly sample the available information and stochastically accumulate evidence toward a decision threshold (Usher and McClelland, 2001; Gold and Shadlen, 2007; Ratcliff and McKoon, 2008). Studies of decision-making based on single stimuli have shown that factors, such as the rate of evidence accumulation, or the decision threshold, vary for different decision alternatives (Salzman et al., 1990; Gao et al., 2011; John-Saaltink et al., 2016). There is also evidence to suggest that integrative decisions are biased toward more reliable stimuli (Raposo et al., 2012; Brunton et al., 2013), even when this leads to suboptimal decisions (Rangelov et al., 2020). We set out to investigate the temporal dynamics of evidence accumulation during integrative decision-making by using an implicit neurofeedback protocol to unconsciously reward prioritized processing of one stimulus over another.
Implicit neurofeedback protocols, in which participants are blind to the specific pattern of neural activity manipulated, have been established as a valuable tool for determining causal links between neural activity and behavior (Purcell et al., 2010; Moxon and Foffani, 2015; Sorger et al., 2019). We embedded neurofeedback within our behavioral task as dynamic changes in task difficulty, thus entirely concealing that participants were receiving neurofeedback (deBettencourt et al., 2015). Participants were asked to report the average direction of motion across two overlapping sets of moving bars, which were tilted at −45° and 45° (Fig. 1d). The two sets of bars moved in different directions, and each set flickered at a different frequency. This flicker induced steady-state visual evoked potentials (SSVEPs), an established index of feature-based visual attention (i.e., prioritized processing of relevant visual features) (Müller et al., 2006; Andersen et al., 2011; Painter et al., 2014; Renton et al., 2019b). Using normalized subtractions of SSVEP amplitudes, we altered task difficulty in real time. Specifically, motion coherence across all bars increased when attention was allocated to bars of the trained orientation, and decreased when attention was allocated to bars of the untrained orientation. To assess carryover effects of neurofeedback, we included test trials in which motion coherence was held constant (i.e., no real-time neurofeedback); thus, the display was no longer contingent on participants' attention. We used a combination of time-frequency analyses and computational modeling to assess how neurofeedback impacted integrative decision-making processes.
Materials and Methods
Participants
Thirty participants (13 males, age mean = 21.93 years, SD = 4.40 years) volunteered to participate in the experiment after providing informed consent. All participants had normal or corrected-to-normal vision. The study was approved by the University of Queensland Human Research Ethics Committee, and the experiment was conducted in accordance with the relevant guidelines and regulations. Participants were paid $40 for attendance. Two participants were excluded from analysis because of technical problems with the EEG recordings.
Task overview
Participants were tasked with reporting the average direction of motion across a field of moving bars (Fig. 1d; Movie 1). The field was composed of two sets of bars oriented at −45° and 45°, which each moved in different average directions. Thus, the true average motion direction across all bars was the average of the motion directions of each set of bars. We recorded EEG data while participants performed the task, using SSVEPs as a measure of attentional selectivity to deliver neurofeedback in real time. The motion coherence of both sets of moving bars varied continuously, contingent on attentional selectivity for one orientation (the trained orientation) over the other (the untrained orientation). From the participants' perspective, both sets of moving bars were equally relevant for the task. However, unbeknownst to them, higher attentional selectivity for the trained orientation resulted in greater motion coherence for both sets of bars, making the averaging task easier (Fig. 1a). Conversely, lower attentional selectivity resulted in lower motion coherence for both sets of bars, making the averaging task more difficult (Fig. 1b). Crucially, at any point in time, bars at the trained and untrained orientations moved with equal coherence; thus, there was never any physical difference in signal quality between the sets. The trained orientation was switched halfway through the experiment, such that selectivity for one set of bars was encouraged in the first half of the experiment, and selectivity for the other set of bars was encouraged for the second half of the experiment.
Experiment protocol
Individual trial design
On each trial, the moving bars were displayed for 3 s. SSVEPs were calculated over a 1 s sliding window, so that bar movement was completely random for the first second of each trial (Fig. 1c). Neurofeedback was implemented over the next 2 s of the trial (Fig. 1d). Movement coherence was calculated frame by frame (120 Hz refresh rate), based on the SSVEP selectivity for the trained orientation over the untrained orientation. At the end of each trial, participants used a computer mouse to rotate a dial toward the perceived average motion direction (Fig. 1d; Movie 1). To avoid biasing participants toward any particular motion direction based on the initial dial position, the dial appeared in the direction of the first mouse displacement following response cue onset. A feedback line indicating the true average motion direction appeared for 1 s immediately after participants clicked the mouse button to register a response.
Block design and counterbalancing
The experiment consisted of 12 blocks of 60 trials. On 72 of the 720 trials (10%), no neurofeedback was presented, and the motion coherence remained constant at 50% of the participant's calculated threshold. These test trials were presented pseudo-randomly, such that one test trial was presented in every group of 10 trials. During neurofeedback, motion coherence constantly shifted, contingent on the current attentional state. Any changes in SSVEP amplitude during this feedback could be attributed to these changes in the display. However, if a shift in feature-based attention was entrained by the feedback, it should emerge on trials in which there was no change in motion coherence and the displays were effectively identical for the trained and untrained orientations. The inclusion of test trials therefore allowed us to assess whether neurofeedback was effective in inducing a shift in feature-based attention. To overcome any additional noise added by the participant's preexisting biases toward one orientation over another, we used a within-subjects design, sequentially training each participant on both orientations. During the first half of the experiment, neurofeedback encouraged attention to bars of one orientation. The other orientation was trained during the second half of the experiment. Other than the changed feedback contingency, the experiment was identical in the first and second halves, and no indication was given to participants that anything had changed. The order of training of the two orientations was counterbalanced between participants, such that −45° was trained first for half of the participants and 45° was trained first for the other half. Both experimenter and participant were blind as to which orientation was trained in the first and second halves of the experiment. Each of the 16 possible directions of motion was presented on 45 trials in total. These trials were randomly interleaved throughout the experiment. Orientation and motion direction were fully counterbalanced, such that neurofeedback encouraged attention to the trained orientation rather than to any particular motion direction.
Frequency tagging design
All moving bars oscillated between black (RGB: 0, 0, 0) and white (RGB: 255, 255, 255) in a sinusoidal flicker with a 50% duty cycle to induce frequency tags. All bars of each orientation (−45°, 45°) flickered at the same frequency (13 Hz, 15 Hz). These frequencies were chosen to be beyond the range of endogenous oscillations in the theta and α ranges, which are more likely to show large variations in amplitude with visual stimulation (Demiralp et al., 2007; Rajagovindan and Ding, 2011; Renton et al., 2019a). The flicker frequencies were counterbalanced between the trained and untrained orientations, separately for neurofeedback and test trials. For each block, the trained orientation flickered at 13 Hz and the untrained orientation flickered at 15 Hz, for 50% of neurofeedback trials and 50% of test trials; on the remainder of the trials, the trained orientation flickered at 15 Hz, and the untrained orientation flickered at 13 Hz. Thus, neurofeedback encouraged attention toward the trained orientation rather than any particular flicker frequency.
Stimulus design
The moving bars subtended 0.06° × 0.19° of visual angle, and moved at a rate of 1.19° of visual angle/s across the display, within a circular aperture with a radius of 4.46° of visual angle. Bars that moved beyond the boundary of this circular area immediately reappeared at 180.00° of polar angle from the point of disappearance, maintaining the same angular trajectory. There were 400 moving bars of each orientation (−45.00°, 45.00°), and thus 800 moving bars in total. The starting position of each bar was randomly selected on each trial, with the constraint that it was within the circular display area. The drawing order of all 800 bars was also randomized trial-by-trial, such that any flickering bar was as likely to be above the others as below. This ensured that the two flicker frequencies were equally salient. Over the course of the experiment, the bars moved in 16 possible directions of motion (0.00°, 22.50°, 45.00°, 67.50°, 90.00°, 112.50°, 135.00°, 157.50°, 180.00°, 202.50°, 225.00°, 247.50°, 270.00°, 292.50°, 315.00°, 337.50°). On any given trial, the two directions of motion presented were always 67.50° apart. Thus, there were 16 possible average directions of motion (11.25°, 33.75°, 56.25°, 78.75°, 101.25°, 123.75°, 146.25°, 168.75°, 191.25°, 213.75°, 236.25°, 258.75°, 281.25°, 303.75°, 326.25°, 348.75°). A red (RGB: 255, 0, 0) fixation cross, which subtended 0.12° × 0.12° of visual angle, was presented at the center of the display, on top of all the moving bars. At the beginning of each block of 60 trials, a black (RGB: 0, 0, 0) screen announced which block would follow in white (RGB: 255, 255, 255), 40 point, Arial font (e.g., “BLOCK 1”). After an enforced 10 s break, the word “[ENTER]” appeared in 20 point font below the block number to announce that participants could proceed to the next block by a button press when ready. After each trial, participants indicated the average direction of motion across all moving bars. The response screen was black (RGB: 0, 0, 0) and contained a white (RGB: 255, 255, 255) circle with a radius of 4.46° of visual angle. This circle outlined the area within which the bars had moved during the trial.
Display computer specifications
The display was presented at a viewing distance of 57 cm on a 24-inch ASUS VG248QE monitor with a refresh rate of 120 Hz and resolution of 1920 × 1080. Stimuli were presented using the Psychophysics Toolbox (Brainard and vision, 1997) running in MATLAB R2017a (64 bit) under Windows 10 (64 bit). The experiment was run on a Dell Precision Tower 5810 desktop computer containing an Intel Xeon E7-4809 v2 CPU and NVIDIA QUADRO M4000 GPU.
EEG recording and processing
EEG recording
EEG data were sampled at 1200 Hz using a g.USBamp amplifier (g.tec Medical Engineering) from 20 active Ag/AgCl scalp electrodes arranged according to the international standard 10-20 system for electrode placement in a nylon head cap (Oostenveld and Praamstra, 2001). The electrode positions were as follows: Iz, O1, O10, O2, O9, Oz, P10, P3, P4, P7, P8, P9, PO10, PO3, PO4, PO7, PO8, PO9, POz, and Pz. This group of electrodes was chosen because SSVEPs are most strongly represented at occipitoparietal electrode sites (e.g., Renton et al., 2019b). The ground electrode was positioned at FCz, and an active Ag/AgCl ear clip electrode was used as the reference. EEG data were filtered in real time with a notch filter at 48-52 Hz and a 1-100 Hz bandpass filter. The experiment was run across three instances of MATLAB. EEG data were originally read into the first instance, then sent to a second instance for real-time processing and classification. The resulting data stream representing selectivity for the trained stimulus was then sent to the third instance of MATLAB from which the experimental task was displayed. EEG data were streamed between MATLAB instances/threads using the FieldTrip memory buffer (Oostenveld et al., 2011).
Neurofeedback
Motion coherence
Neurofeedback was implemented by manipulating the motion coherence of the moving bars. To create an average direction of motion across many moving bars, the direction of motion of each moving line was drawn from a normally distributed probability density function, with its center at the average direction of motion. The SD of this distribution of motion directions was manipulated to change the motion coherence (Waskom et al., 2018). Thus, motion was highly coherent when this distribution was narrow (Fig. 1a), and low in coherence when this distribution was broad (Fig. 1b). Neurofeedback was implemented to implicitly “reward” participants for attending to the trained orientation. Thus, as attentional selectivity for the trained orientation increased, the motion coherence of both fields of bars increased, making the task easier (Fig. 2a). The SD of the distribution of motion directions was updated every 54 ms.
Real-time SSVEP selectivity
Attentional selectivity was calculated as the difference in the Z scores of the SSVEPs (13 Hz, 15 Hz) to the trained and untrained orientations. We used this difference score, rather than the absolute amplitude of the SSVEP to the trained orientation, to specifically target a relative shift in attention between the two orientations rather than the participant's arousal or sustained attention to the task. Each iteration of the real-time data analysis loop (computation speed: 54 ms; ∼19 Hz), read the last 1 s of EEG data from a memory buffer, performed an FFT (Fig. 2b,c), normalized the SSVEPs, and wrote the resulting attentional selectivity scores to a second memory buffer accessed by the task presentation thread. Before the experimental task began, the EEG data recorded during the staircase procedure (described below) were analyzed to find the four electrodes for which SSVEP amplitudes were highest at each flicker frequency. The real-time SSVEP amplitudes at the two flicker frequencies were taken as the average FFT amplitude across these four predefined best electrodes at the first harmonic (f0) of the frequency of interest. SSVEP amplitudes were continuously added to a distribution of SSVEP amplitudes for each frequency over the course of the experiment (Fig. 2d). As flicker frequency was fully counterbalanced with bar orientation, these distributions contained a roughly even number of samples from when the SSVEP represented the trained and untrained orientation. After each sample, SSVEP amplitudes were added to frequency-specific SSVEP amplitude population distributions, and each sample was transformed to a Z score using the distribution mean (µ) and SD (σ, Fig. 2e). The population of SSVEP amplitudes was initially built up during the staircase procedure, and thus already contained 80 trials worth of samples on the first trial of the experiment. This ensured that the distribution provided a reliable estimate of the population distributions from the beginning of the neurofeedback. Once the Z scores for the SSVEPs to both flicker frequencies were calculated, the Z score for the untrained orientation was subtracted from that for the trained orientation (Fig. 2f). This difference in Z scores was linearly translated to motion coherence (Fig. 2g). When the Z score difference was zero (i.e., no selectivity for the trained or untrained motion direction), motion coherence was set to 80% of the threshold at which participants could no longer perform the task. Motion coherence was scaled around this threshold such that a Z score difference of >2.5 resulted in motion coherence with an SD of 0 (i.e., 100% motion coherence).
Staircase procedure
A staircase procedure was implemented to determine the threshold of motion coherence around which participants were unable to perform the averaging task. The staircase procedure established a baseline performance threshold, at which participants were within 67.5° of the average direction of movement on 50% of trials. Neurofeedback was implemented as a normal distribution centered at 80% of this threshold. The procedure involved four concurrent staircases, starting with the motion coherence SDs of 1.0°, 40.0°, 80.0°, and 120.0°. Trials lasted 2 s and were coded as incorrect if the participant's answer was > 67.5° from the true average. Each individual staircase was run for 20 trials, resulting in 80 staircase trials altogether. The threshold was taken as the average result across the four staircases. The mean threshold SD of dot motion directions was 41.36° (95% CI: 35.65°-48.07°, range 10°-76°).
EEG preprocessing for offline analyses
For offline analysis, noisy electrodes, identified via visual inspection, were replaced with a linear interpolation based on the nearest channels. For each trial, EEG data across all 20 channels were linearly detrended and baseline-corrected. All frequency-based analyses were performed using the average of the four EEG channels that showed the highest SSVEP amplitude for each participant and flicker frequency (13 Hz, 15 Hz). To assess which EEG channels to use for the SSVEP analyses, we computed grand-average ERPs for each participant by averaging the EEG data across every 3 s trial at each EEG channel. These ERPs were submitted to FFTs, and the four electrodes with the highest amplitudes for each frequency (13 Hz, 15 Hz) were used for all SSVEP analyses. The electrodes included in this group for the 13 Hz SSVEP, listed from most common to least common, were as follows: Oz (19 participants, -19p), O2 (17p), POz (16p), O1 (14p), PO8 (9p), PO4 (9p), Pz (7p), PO3 (6p), P3 (5p), P4 (4p), Iz (2p), O9 (1p), PO7 (1p), O10 (1p), P8 (1p). The electrodes included in this group for the 15 Hz SSVEP, listed from most common to least common, were as follows: Oz (18p), POz (17p), O2 (14p), O1 (11p), PO4 (11p), PO3 (9p), Pz (6p), PO8 (5p), O9 (3p), PO7 (3p), P4 (3p), P7 (2p), P3 (2p), O10 (2p), Iz (2p), P9 (1p), PO10 (1p), P10 (1p), P8 (1p). Trials on which the absolute amplitude at any of these channels exceeded 200 µV were excluded from further analysis. As frequency and orientation were counterbalanced and orthogonalized, this approach optimized SSVEP amplitudes equally for the trained and untrained orientations.
To assess the time course of SSVEP selectivity for the trained over the untrained orientation, we performed a time-frequency analysis. For this analysis, we generated two ERPs for each participant: one for trials in which the trained and untrained orientations flickered at 13 and 15 Hz, respectively, and one for the reverse case. These ERPs included an additional 2 s of data on either side of each trial to preserve precision at the trial beginning and end. We applied Morlet wavelet time-frequency transforms to each of these ERPs, extracting SSVEPs at both 13 and 15 Hz (Brainstorm) (Tadel et al., 2011). These transformations used mother wavelets at each target frequency with FWHM set to 0.50 s, fixing temporal resolution at 0.20 s and spectral resolution at 0.80 Hz for each frequency. The resulting SSVEPs were averaged across the two flicker frequencies to output the average time course of the neural response to the trained and untrained orientations. This approach was applied separately for neurofeedback and test trials.
Statistical tests and models
All statistical tests were performed using MATLAB (2016b). Circular statistics were computed using “Circstat: a Matlab toolbox for circular statistics” (Berens, 2009) (https://github.com/circstat/circstat-matlab). Bayes factors (BFs) were computed using the “bayesFactor” package for MATLAB (Rouder et al., 2009; Morey and Rouder, 2011; Krekelberg, 2021) (https://github.com/klabhub/bayesFactor).
Permutation tests
Time-series effects were assessed statistically using Monte Carlo permutation simulations (Chung and Romano, 2013). For the time-frequency analyses, we sought to identify periods during which SSVEPs to the trained and untrained motion directions were significantly different. For this analysis, we generated two ERPs for each participant: one for trials in which the trained and untrained orientation flickered at 13 and 15 Hz, respectively, and one for the reverse case. We applied Morlet wavelet transforms to these ERPs to assess the average time course of the neural response to the trained and untrained orientations (see EEG preprocessing for offline analyses). The difference between these two time-series was taken to reflect SSVEP selectivity for the trained orientation (SSVEPtrained – SSVEPuntrained). We implemented a Monte Carlo simulation to generate a permutation distribution of the possible values for this SSVEP selectivity score that would be expected by chance. Over 1000 iterations, we shuffled the labels for the flicker frequency of the trained and untrained orientations on individual trials, before generating the ERPs for the time-frequency analysis. This resulted in 1000 permuted time-series' of SSVEP selectivity per participant.
For the forward encoding analysis, we sought to identify periods of significant tuning to the trained, untrained, and average motion directions. For these analyses, we trained a forward encoding model based on idealized tuning curves centered at each of the possible motion directions (0°-360°). For each trial, we rotated the labels for the predicted channel responses relative to the true motion direction presented on that trial (−180° to 180°). For example, if the presented motion direction on a given trial was 180°, we would adjust the labels for this trial such that channel responses for 0° were relabeled −180° and channel responses at 360° were relabeled 180°. Tuning metrics were derived as the vector sum of channel responses aligned to the cardinal axis of the encoded motion direction (see Forward encoding model). In other words, the tuning metric represented the motion energy of channel responses in the direction of the encoded motion direction. We implemented a Monte Carlo simulation to generate a permutation distribution of possible values for this tuning metric that would be expected by chance. Over 1000 permutations, we shuffled the labels for the true motion direction presented on each trial, such that the tuning metric was calculated around different combinations of random angles. This resulted in 1000 permuted time courses of tuning to the encoded motion direction per participant.
For each of these metrics, we averaged the true and permuted time-series data across participants to identify statistically significant periods across the grand-average time-series. This averaging across participants was performed in a second permutation step, in which we generated 1,000,000 grand-average permuted time courses by averaging different combinations of the individual participants, permuted time courses on each iteration. Each permuted group average time course represented different combinations of each of the 28 participants. 1000 permuted time courses. Every participant was still included in every permuted average; we simply sampled a different one of the participant's permuted time courses on each permutation iteration. To control for multiple comparisons, the time-series data from these 1,000,000 grand averages were then all grouped within a single distribution to calculate the thresholds for significance. Using a criterion of p < 0.05 (two-tailed) on this full distribution, we designated points on the true grand-average time course as significantly different from chance if they fell below the 2.5th percentile or above the 97.5th percentile on the distribution of the permuted data.
Forward encoding model
Forward encoding models are based on the premise that EEG data should reflect a weighted combination of tuned neuronal responses. This principle can be used to estimate the strength of tuning to certain motion directions using weight matrices, which have in turn been estimated using idealized tuning curves (Brouwer and Heeger, 2009; Garcia et al., 2013; Smout et al., 2019; Tang et al., 2020). This approach allowed us to estimate the strength of tuning to the motion directions of the trained and untrained orientations as well as the grand-average motion direction over the course of each 3 s trial. To prepare the data for the forward encoding analysis, we submitted single-trial EEG data from each 3 s trial to a Morlet wavelet transform, with the FWHM set to 0.2, to maximize temporal resolution. The time-frequency transforms at the average of the second harmonic frequencies (26 Hz, 30 Hz) were used to assess how well the motion directions were encoded over time during the trial (Garcia et al., 2013). The model was evaluated using k-fold cross-validation; thus, the data were iteratively separated into train (B1) and test (B2) sets, such that each of the 12 experimental trial blocks served as a test set in turn (Fig. 3a).
To generate an idealized set of channel outputs tuned to each motion direction, we populated a vector with the 16 possible motion directions for the trained or untrained motion direction (0.0°-337.5°, in 22.5° increments), or for the average motion directions (33.75°-371.25°, in 22.5° increments), depending on the motion direction currently encoded. For each trial, we subtracted the presented motion direction from this vector, such that the position corresponding to the presented motion direction was set to 0 (i.e., −180.0°, −157.5°, −135.0°, −112.5°, −90.0°, −67.5°, −45.0°, −22.0°, 0.0°, 22.5°, 45.0°, 67.5°, 90.00°, 112.5°, 135.0°, 157.5°). The cosine of this vector, raised to the seventh power, formed a biologically plausible set of channel outputs (C1) representing the idealized neuronal response to each motion direction (Garcia et al., 2013). Negative values in this design matrix were set to zero (Fig. 3b). Training and testing were performed independently across time using a 20 ms average sliding window, sampled at 100 Hz. This meant that each iteration of training occurred using a single data matrix (B1, m electrodes × n trials), which should be related to the hypothetical channel outputs (C1, k channels × n trials) by a weight matrix (W, m electrodes × k channels) as follows:
This weight matrix was estimated, using linear regression, as follows (Fig. 3c):
And the channel responses (C2) for the test data (B2) were in turn based on the estimated weight matrix (Fig. 3d) as follows:
Channel responses for each trial in the experiment were estimated as each experimental block was tested in turn. This process was performed 3 times for each participant: once selecting the trained motion direction on each trial, once selecting the untrained motion direction, and once for the grand-average motion direction. These responses for each trial were rearranged such that the encoded motion directions were centered in the vector of channel responses. Thus, 0°, set as the central position in the matrix of estimated channel responses, always represented the actual motion direction presented on each trial. Further, when we encoded the grand mean motion direction, we flipped the matrix of estimated channel responses such that the trained motion direction was always to the right of the average, and the untrained motion direction was always to the left. Once each trial was recentered, the average channel responses to the trained, untrained, and grand-average motion directions were calculated for each participant. These average responses were smoothed using a sliding average (80 sample window; MATLAB movmean function).
To assess how strongly the trained, untrained, and average motion directions were encoded throughout the trial, we calculated a single metric representing the degree of tuning at each time point. Channel responses were represented as polar vectors, where each vector's angle was the corrected motion direction (−180° to 180°), and each vector's amplitude was the channel response at that motion direction. We calculated the x and y components of each of these vectors. The x components are oriented in the direction of the encoded motion direction (0°). Thus, adding the x components of vectors corresponding to each motion direction resulted in a metric that represents the net response to the tuned motion direction.
Mixture distribution modeling of behavioral responses
To assess how attentional selectivity affected response bias during integrative decision-making, we first sought to identify guesses and target reversals. To this end, we fit a mixture distribution model for each participant's distribution of response errors (uncorrected, i.e., trained feature equally likely to fall on either side of the average) to describe these as a probabilistic mixture of noisy target responses (centered around true average), nontarget responses (centered 180° from true average), and guesses (uniform distribution at ∼360°; www.paulbays.com/toolbox/) (Bays et al., 2009; Schneegans and Bays, 2016). One participant was excluded from behavioral data analyses for poor performance, as their responses were best modeled by a uniform (i.e., flat) distribution from −180° to 180° relative to the correct response.
Participants' estimates of average motion direction on each trial (10) should reflect the weighted average of the two presented motion directions. To quantify these decision weights for the trained and untrained motion directions, a multiple-regression model with a term for each of the presented motion directions was fitted to each participant's target responses using ordinary least squares as follows:
Source code and data availability
Source code for the experimental task is as follows: https://github.com/air2310/Neurofeedback-Investigations-of-Integrative-Decision-Making. Data are available on request from corresponding author.
Results
Implicit training: debriefing questionnaire
The neurofeedback protocol embedded within the integrative decision-making task aimed to implicitly shift attention toward each of two orientations in turn. Participants were informed that the display might change over the course of the experiment depending on the EEG recording, but were not provided with any further detail. To probe whether or not participants had guessed the calculation underlying neurofeedback, we used a debriefing questionnaire after the experiment. Questions included: “Please describe any strategies you used to do the task, and whether they changed over the course of the experiment.” and “Was there anything you did that made the task easier or more difficult?” Participants typically described their strategies with phrases, such as “blurring my eyes,” “focusing on the fixation cross,” and “following the overall flow.” No participant described focusing more on bars of one orientation than another. This suggests that participants were not consciously aware that the task difficulty was dependent on their attention to one set of bars over another and that neurofeedback was indeed delivered implicitly. As the trained orientation was counterbalanced within subjects and the experimenter was blind to the training order, this suggests that any differences in the processing of the trained and untrained stimuli cannot be attributed to subject or experimenter biases and must result from neurofeedback training.
Sensory processing biases: neurofeedback training of attention
Frequency tagging
Before assessing for differences between SSVEPs to the trained and untrained orientations, we first confirmed that frequency tagging elicited reliable SSVEPs by calculating neural responses to the individual flicker frequencies. To this end, we calculated the grand mean ERP across all neurofeedback and test trials. Taking the FFT of these grand mean ERPs revealed that, overall, SSVEPs were evident at the flicker frequencies, and were similar in amplitude for neurofeedback and test trials (f1: 13 Hz; f2 15 Hz; Fig. 4a). SSVEPs on both neurofeedback and test trials were centered over occipitoparietal electrode sites (Fig. 4b,c), consistent with previous frequency tagging studies involving dynamic visual displays (e.g., Renton et al., 2019b). Thus, frequency tagging was successful, and neurofeedback and test trials evoked a similar neural response overall.
SSVEP selectivity during test trials
On test trials, motion coherence was not contingent on real-time SSVEP feedback. Instead, coherence remained fixed at 50% of the threshold at which participants could no longer perform the task. Any difference in attentional selectivity between the trained and untrained orientations during these trials thus provided a pure index of any trained carryover effects from the neurofeedback trials. To assess whether any such attentional selectivity for the trained feature arose on either trial type, we calculated 3 s trial ERPs for each participant, separately for the neurofeedback and test trials in which the trained orientation flickered at either 13 or 15 Hz (counterbalanced across the experiment). These ERPs were submitted to FFTs, and the amplitudes at 13 and 15 Hz were averaged to output SSVEPs to the trained and untrained orientations. To assess attentional selectivity, we calculated the difference in SSVEPs to the trained and untrained orientations, separately for neurofeedback and test trials. The topographical distribution of SSVEP selectivity (ΔµV) for the trained over the untrained orientation on test trials shows attentional selectivity for the trained orientation across a number of occipitoparietal electrode sites (Fig. 4d). To assess whether this selectivity for the trained orientation was statistically significant, mean SSVEP amplitudes (µV) evoked by the trained and untrained orientation were calculated using the mean of the four electrodes with the highest SSVEP amplitudes for each flicker frequency and participant. These mean SSVEPs were submitted to a t test, which showed that SSVEPs evoked by the trained orientation (mean = 0.69, SD = 0.42) were significantly larger than those evoked by the untrained orientation (mean = 0.65, SD = 0.38, t(27) = 2.28, p = 0.03, BF10 = 1.84; Fig. 4f). Thus, neurofeedback was successful in shifting feature-based attention toward one orientation over the other. This attentional selectivity is based on an average across the entire trial.
We next asked at which point in the 3 s trials attentional selectivity emerged (Fig. 5a). To assess the time course of attentional selectivity for the trained over the untrained orientation, we calculated the difference between the Morlet wavelet-transformed ERPs to these two signals (Fig. 5c,e). We used a permutation test based on a Monte Carlo simulation to assess whether attentional selectivity for either the trained or untrained orientation was significant at any point during the trial. The results of this analysis revealed significant attentional selectivity for the trained over the untrained orientation on test trials, which emerged in the first second after the onset of coherent motion (Fig. 5g). This attentional selectivity for the trained orientation was significant with a criterion of p < 0.05 (two-tailed) from 163 to 889 ms after the onset of coherent motion. Thus, overall, the neurofeedback-induced enhancement of the trained orientation was most prominent in the first second of coherent motion within a trial.
SSVEP selectivity during neurofeedback training
On neurofeedback trials, motion coherence varied continuously, contingent on real-time attentional selectivity for one set of oriented bars over the other. Thus, motion coherence was higher when the trained orientation was attended and lower when the untrained orientation was attended. On these neurofeedback trials, the topographical distribution of the difference between SSVEPs to the trained and untrained orientation (ΔµV) revealed no effect at any electrode site (Fig. 4e). To assess any possible difference statistically, mean SSVEP amplitudes evoked by the trained and untrained orientations were calculated using the mean of the four electrodes with the highest SSVEP amplitudes for each flicker frequency and participant. These mean SSVEP amplitudes (µV) were submitted to a paired-samples t test, which revealed no significant difference between SSVEPs evoked by the trained (mean = 0.65, SD = 0.40) and untrained orientations during neurofeedback trials (mean = 0.66, SD = 0.41, t(27) = −0.99, p = 0.33, BF10 = 0.31; Fig. 4g).
We next asked whether attentional selectivity emerged at any point in the 3 s trial. To do this, we performed a time-frequency analysis of the neurofeedback trial ERPs (Fig. 5b). Using the same procedure as for the test trials, we used Morlet wavelet transforms to calculate the average time course of the neural response to the trained and untrained orientations (Fig. 5d,f), and determined the difference between these two SSVEPs (Fig. 5h). To assess whether attentional selectivity for either the trained or untrained orientation was significant at any point during the trial, we used a permutation test based on a Monte Carlo simulation. The results of this analysis revealed no significant attentional selectivity for either orientation at any point during the trial (Fig. 5h). It should be noted that this does not indicate that participants never switched their attention between stimuli, but rather suggests that there was no time during which participants reliably attended more to one feature than another.
Thus, there was no overall difference in attentional selectivity for the trained and untrained orientations during neurofeedback trials. Recall, however, that SSVEPs during these neurofeedback trials were influenced by dynamic changes in motion coherence, and thus cannot unambiguously determine any stable attentional biases in favor of one orientation over the other. The test trials, in which neurofeedback was removed and thus motion coherence was stable, revealed an attentional bias. We next asked how this bias in sensory processing affected the perceptual representation of motion direction for the two stimuli, as well as the integration of the two motion signals into a perceptual average.
Integrative decision-making processes: neural representations of the trained, untrained, and average motion directions
Forward encoding model
Recall that the participants' task was to integrate the motion directions of the two overlapping stimulus fields to produce their average motion direction. The two sets of bars always moved with identical motion coherence, and each trial was followed by feedback indicating the true average motion direction. Thus, an ideal observer should weigh the two sets of bars equally in the grand average (Rahnev and Denison, 2018). However, the neurofeedback embedded within this integrated decision-making task was successful in training participants to prioritize the processing of the trained set of bars over the untrained set of bars, as evidenced by the SSVEP bias observed on the test trials. This gives us the opportunity to observe how any perceptual biases might arise, independent of any bottom-up differences in the integrated stimuli. To this end, we set out to visualize and describe the temporal dynamics of evidence accumulation for the trained, untrained, and grand-average motion directions over the course of the trial. As continuous EEG data were recorded throughout the experiment, we used a forward encoding approach in which we trained separate models for the motion directions of the two presented stimulus fields, and a further model for the average motion direction.
Forward encoding results: trained and untrained motion directions
Before assessing how the two motion directions were integrated into an average, we first assessed how the representation of the trained and untrained motion directions built up separately during the trial. Figure 6a and b show the average strength of encoding at each channel, centered around motion directions for the trained (Fig. 6a) and untrained (Fig. 6b) orientations, throughout the 2 s of coherent motion. These surface plots reveal that the motion direction of the trained orientation was strongly encoded early in the trial, and built up from the moment of coherent motion onset (Fig. 6a). This strong representation of the trained motion direction early in the trial tapered off after ∼1 s of coherent motion. By contrast, encoding to the untrained motion direction was suppressed early in the trial, but with a peak emerging toward the end of the trial (Fig. 6b).
To statistically determine how encoding of the trained and untrained motion directions differed, we computed tuning metrics to summarize the degree of tuning along the trained and untrained motion directions. We first asked when tuning to the trained and untrained motion directions differed using a permutation test based on a Monte Carlo simulation. By iteratively recalculating the mean difference in tuning to the trained and untrained motion directions with the labels for trained and untrained shuffled, we generated a permutation distribution to represent the magnitude of difference that would be expected by chance. This showed that tuning to the trained motion direction was significantly stronger than tuning to the untrained motion direction from 153 to 906 ms after the onset of coherent motion (Fig. 6d; p < 0.05). Notably, this period of significantly stronger encoding of the trained orientation overlaps the period of significant attentional selectivity for the trained orientation on test trials (Fig. 5g).
We next asked when the tuning metrics for the trained and untrained motion directions, individually, differed from what would be expected by chance. To this end, we compared the tuning metrics for the trained and untrained motion directions to permuted distributions of tuning curves created by shuffling the labels for the true motion directions presented on each trial. The permutation distribution therefore represented the range of likely tuning amplitudes that might be expected if the modeled motion direction was arbitrary. Setting α = 0.05 (two-tailed), we found that tuning to the trained motion direction was significantly stronger than chance within the first second of coherent motion, briefly from 404 to 464 ms and again from 565 to 836 ms after the onset coherent motion (Fig. 6e). By contrast, tuning to the untrained motion direction was actually weaker than might be expected by chance over the first second of the trial, from 103 to 695 ms after the onset of coherent motion. Stronger than chance tuning for this motion direction only emerged toward the end of the trial, from 1708 to 1969 ms after the onset of coherent motion (Fig. 6f). Notably, these results were generated using both the neurofeedback and test trials to maximize accuracy and reliability of the forward encoding model. However, the pattern of results described here held true for both neurofeedback trials and test trials when separated. This pattern of results suggests that, while the trained and untrained orientations were presented concurrently, they were encoded sequentially in the brain, with a strong representation of the trained motion direction prioritized because of neurofeedback.
Forward encoding results: average motion direction
We next asked how the trained and untrained motion directions were integrated into the average motion direction during the trial. Notably, this motion direction was never actually displayed during the trial but needed to be built up internally through integration of the motion directions of the two overlapping stimulus fields. Applying a forward encoding model to this motion direction allowed us to assess how neurofeedback-induced differences in the representations of the trained and untrained motion directions affected the buildup of the neural representation of the integrated average. Figure 6c shows the average strength of encoding of each motion direction channel, centered around the average motion direction, for the 2 s of coherent motion. These were flipped after model training such that the trained motion direction was always to the right of the average, and the untrained motion direction was always to the left of the average. This surface plot shows that the representation of the average motion direction built up gradually during the trial, and grew stronger until the offset of coherent motion (Fig. 6c). Notably, the peak response appears to be shifted away from the true average response toward the trained motion direction (22.5°).
In order to quantify the strength of tuning to the average motion direction over the course of the trial, we generated a summary tuning metric by calculating the vector sum of all channel responses in the direction of the encoded motion direction (i.e., channel response ×cos(θ); see Fig. 6b,c). This metric indicated the strength of tuning to the average motion direction at each point in the trial (Fig. 6g). To assess when this tuning metric differed from what might be expected by chance, we compared it with a permuted distribution of average tuning metrics generated by shuffling the labels for the true average motion directions presented on each trial. The permutation distribution therefore represented the range of likely tuning amplitudes that might be expected if the modeled motion direction was arbitrary. Setting α at 0.05 (two-tailed), we found that tuning to the true average motion direction was significantly stronger than chance from 1087 ms after the onset of coherent motion until the end of the trial. Notably, this tuning to the average motion direction occurred after the offset of significant tuning to the trained motion direction but overlapped with significant tuning to the untrained motion direction.
To investigate the shift in peak tuning away from the true average motion direction and toward the trained motion direction (Fig. 6c), we generated a summary bias metric by calculating the vector sum of all channel responses perpendicular to the direction of the average motion direction (i.e., channel response ×sin(θ)). Positive values of this metric indicate a shift in tuning toward the trained motion direction, and negative values indicate a shift in tuning toward the untrained motion direction. We used the same permutation test as that used to assess periods of significant tuning to the average motion direction to assess periods of significant shifts in tuning direction away from this average. Setting α at 0.05 (two-tailed), we found that tuning direction was shifted significantly away from the true average motion direction and toward the trained motion direction from 484-866 ms after the onset of coherent motion (corresponding with the period of significant tuning to the trained motion direction). To visualize how the amplitude and direction of tuning to the average motion direction evolved over time, we plotted the mean tuning vector for each time point after the onset of coherent motion (Movie 2). y components of these vectors were used to calculate the tuning metric, whereas the x components were used to generate the bias metric. The video shows that, during the period of significant bias, tuning was centered around the trained motion direction. Tuning then shifted toward the true average, growing in amplitude over time.
To maximize accuracy and reliability of the forward encoding model, forward encoding results were generated using both the neurofeedback and test trials. However, it should be noted that the pattern of results described here holds true for both neurofeedback and test trials when modeled separately. Thus, overall, we found that the trained orientation was encoded first, followed by the untrained orientation, before integration into a representation of the average motion direction. Further, the integrated representation of the average motion direction was shifted toward the trained motion direction. We next asked how this encoding process affected behavioral responses on each trial.
Integrative decision-making outcomes: behavioral reporting of the average motion
Having shown that neurofeedback biased the allocation of attention toward the trained motion direction, and that in turn, this motion direction was encoded earlier and more strongly than the untrained motion direction, we next asked how neurofeedback affected the behavioral outcome of the integrated decision at the end of each trial.
Response bias distribution
To visualize the distribution of response errors relative to the true average, we calculated the difference between responses and the true average motion direction on each trial, such that zero represented perfect accuracy. Mixture distribution modeling of these responses was used to separate target responses from guesses and nontarget responses (reversals; see Materials and Methods). Target responses were then transformed such that errors with positive values were closer to the trained motion direction, and errors with negative values were closer to the untrained motion direction (Fig. 7a,b). We used the circular statistics toolbox to calculate descriptive statistics for each participant's transformed distribution of response errors. On test trials, these distributions had an average circular mean of 1.28° (95% CI: −0.82° to 3.38°), circular median of 1.39° (95% CI: −1.78° to 4.56°), and mode of 0.74° (95% CI: −7.91° to 9.39°; Fig. 7a). On neurofeedback trials, these distributions had an average circular mean of −0.98° (95% CI: −2.52° to 0.56°), circular median of −0.43° (95% CI: −1.48° to 0.62), and mode of −1.67° (95% CI: −7.95° to 4.61°; Fig. 7a). Thus, participants' responses appeared to be shifted slightly toward the trained motion direction on test trials, and if anything appeared shifted toward the untrained motion direction on neurofeedback trials. We next set out to formally test how the trained and untrained motion directions were weighted in participants' responses.
Feature decision weightings
Participants' estimates of average motion direction on each trial should reflect the weighted average of the two presented motion directions. To assess whether the trained or untrained motion direction was weighted more heavily in participants' responses, we used a regression approach to compute the least-squares estimates of the weightings of the trained and untrained orientations for each participant (see Materials and Methods). This analysis revealed that for test trials, in which participants attended significantly more strongly to the trained orientation, the trained orientation (mean = 0.69, SD = 0.31) was indeed weighted more heavily in participants' responses than the untrained orientation (mean = 0.64, SD = 0.33, t(26) = 2.59, p = 0.016, BF10 = 3.24; Fig. 7c). By contrast, on neurofeedback trials, in which there was no overall difference in attentional selectivity for the trained and untrained orientations, weights for the trained orientation (mean = 0.56, SD = 0.33) did not differ significantly from weights for the untrained orientation (mean = 0.57, SD = 0.33; t(26) = −1.53, p = 0.138, BF10 = 0.57; Fig. 7c).
Response bias by attentional selectivity
On test trials, participants attended more strongly to the trained feature, and in turn weighted this feature more heavily in their integrated average responses. To more directly assess this relationship, we compared response bias (°) for the trials on which SSVEP selectivity for the trained orientation was strongest, with response bias for the trials on which SSVEP selectivity was weakest. To this end, we calculated an SSVEP selectivity score (SSVEPtrained – SSVEPuntrained) for each trial by first Z scoring the single-trial SSVEPs relative to the population of the participant's single-trial SSVEPs to each flicker frequency. For each participant, we performed a median split based on these SSVEP selectivity scores (separately for neurofeedback and test trials), to obtain two equally sized samples of trials in which SSVEP selectivity for the trained orientation was at its strongest or weakest. We then calculated the circular mean of each participant's response bias scores (Δ°) for each of these groups of trials (where bias toward the trained orientation is positive). For test trials, responses were significantly more biased (°) toward the trained set of bars on trials with stronger attentional selectivity for these bars (mean = 2.95, SD = 10.65) than on trials with weaker attentional selectivity for these bars (mean = −3.21, SD = 13.25; t(26) = 2.21, p = 0.036, BF10 = 1.62; Fig. 7e).
On neurofeedback trials, participants did not attend more to one orientation than another overall, and in turn, they weighted both sets of bars equally in their responses on these trials. Partitioning responses on these trials according to single-trial SSVEP selectivity can be used to reveal more about this relationship. At first glance, it is evident that the response bias was much more variable when SSVEP selectivity for the trained feature was weak (i.e., stronger SSVEP selectivity for untrained, mean = −1.14, SD = 8.43), than when SSVEP selectivity for the trained feature was strong (mean = −1.35, SD = 3.49). Indeed, the response bias scores failed Bartlett's test for homogeneity of variance (p < 0.001), and therefore were not submitted to a t test. To more closely evaluate this relationship, we calculated the circular SD of each participant's response bias scores for each group of trials (SSVEP selectivity for the trained orientation, relatively high vs relatively low). This showed that responses were significantly more variable when SSVEP selectivity for the trained orientation was relatively low (mean = 46.05, SD = 15.92) versus relatively high (mean = 43.15, SD = 15.98, t(26) = 6.15, p < 0.001, BF10 = 1105; Fig. 7f). This sizable difference in response accuracy demonstrates that the neurofeedback protocol successfully made the task more difficult when participants attended to the untrained orientation, and easier when participants attended to the trained orientation.
Neurofeedback dynamics
Our neurofeedback protocol was designed to bias sensory processing toward one visual feature over another. We have shown that this protocol was successful in inducing an attentional bias which transferred to test trials, on which no neurofeedback was delivered and for which there was nothing in the display to encourage prioritized processing of one stimulus over another. To facilitate the design of future neurofeedback studies using similar protocols, we investigated the dynamics of feedback to assess how this was represented in the brain, and whether any particular features of the design might yield even stronger effects. These analyses were limited to neurofeedback trials (on which feedback was presented).
Feedback variability
Neurofeedback was designed to represent the real-time difference in attentional selectivity for the trained and untrained features. To this end, at each moment in the trial, we first calculated the SSVEPs at each of the two flicker frequencies (13 Hz, 15 Hz), which were counterbalanced trial-by-trial with orientation (trained, untrained). These SSVEPs were Z-scored relative to the population of SSVEPs at that flicker frequency to assess whether they were relatively high or low compared with what might be expected for that frequency (for an example trial, see Fig. 8a). For each time point, we calculated SSVEP selectivity for the trained feature as the difference between these two Z scores (for an example trial, see Fig. 8b). This SSVEP selectivity score was used to control the motion coherence of the moving bars, making the task easier when attentional selectivity for the trained orientation was higher (for an illustration of how motion coherence corresponded to different SSVEP selectivity scores on an example trial, see Fig. 8c).
It is worth asking whether this feedback signal represented rapid, within-trial fluctuations in the allocation of attention between the two features, or a more stable attentional state that varied across trials. To address this question, we first calculated the SD of feedback (SSVEP selectivity) values within each trial, taking the average of these values, and then calculated the average feedback values for each trial and took the SD across trials. If feedback largely reflected momentary fluctuations in attentional state, we would expect larger within-trials SDs, whereas if feedback was tracking a more slowly changing state, we would expect the SD to be larger across trials. The analysis revealed that feedback varied significantly more within trials (mean = 0.90, SD = 0.22) than across trials (mean = 0.59, SD = 0.14, t(27) = −13.97, p < 0.001, BF10 = 9.74 × 1010; Fig. 8d), which suggests that feedback was targeting moment-by-moment differences in information processing more than gradual changes in attentional state across the experiment.
Time-frequency analysis of the test (non-neurofeedback) trials showed that attentional selectivity for the trained feature was strongest in the first half of the coherent motion period, tapering off toward the end of the trial. We next investigated whether neurofeedback was more variable earlier in the trial as well, as this would have reinforced the training contingencies more strongly during this period. To this end, we calculated the SD of feedback scores for the first half of each trial (0-1 s after the onset of coherent motion) and for the second half of each trial (1-2 s after the onset of coherent motion), and calculated the average of these SDs for each participant. Feedback was indeed more variable earlier (mean = 0.82, SD = 0.19) than later in the trial (mean = 0.78, SD = 0.18, t(27) = 7.2, p < 0.001, BF10 = 1.56 × 105; Fig. 8e). To examine feedback on an even finer timescale, we calculated the SD of feedback values across trials for each time point in the trial. For each participant, we used a standard linear regression to regress SD scores against time within the trial, and found that the slope (mean = −0.05, 95% CI: −0.08, −0.02) was significantly more negative than chance (t(27) = −3.02, p = 0.006, BF10 = 7.63). Thus, neurofeedback was most variable soon after the onset of coherent motion, and became progressively less variable toward the end of each trial. The correspondence in timing between high variance feedback periods on neurofeedback trials and significant attentional selectivity on test trials suggests that participants may have been more strongly conditioned to attend to the trained orientation during this time.
Learning curves and switch costs
In this experiment, we implemented “bidirectional” neurofeedback; that is, participants were first trained on one orientation, then on the other, thus controlling for the trained orientation within-subjects. This design choice meant we could be certain that any neural or behavioral biases arose from neurofeedback training rather than potential preexisting orientation preferences among participants. However, this advantage was balanced against potentially attenuated training effects resulting from switching the neurofeedback contingency mid-experiment. To determine whether there were such switch costs, we first compared the training effects for the first and second halves of the experiment. We began by regressing ordinal trial number within the relevant half of the experiment (i.e., how many trials the feature had been trained for), against SSVEP selectivity on individual test (i.e., non-neurofeedback) trials, using a simple linear regression. The slope of this linear regression indicated the rate at which SSVEP selectivity for the trained orientation changed over the course of the training block. This slope value did not differ significantly from 0 in either the first half (mean = −0.02, SD = 0.19, t(27) = −0.52, p = 0.610, BF10 = 0.23) or second half (mean = 0.03, SD = 0.20, t(27) = 0.73, p = 0.471, BF10 = 0.26) of the experiment. Indeed, there was no significant difference in slope between experiment halves (t(27) = −0.79, p = 0.435, BF10 = 0.27). Thus, it appears that rather than steadily increasing over the course of training, effects emerged relatively rapidly and then remained at a constant level.
To assess how switching the neurofeedback contingency affected the magnitude of these consistent training effects, we next compared SSVEP selectivity on individual test (i.e., without neurofeedback) trials across the two halves of the experiment. We found that SSVEP selectivity for the trained orientation (ΔµV) was significantly larger in the first half of the experiment (mean = 0.11, SD = 0.18) than in the second half (mean = −0.02, SD = 0.20, t(27) = 2.39, p = 0.024, BF10 = 2.22). Thus, switching the neurofeedback contingency did indeed reduce the magnitude of the training effect. To more closely examine how SSVEP selectivity changed in response to the switched contingency, we focused on the test trials immediately preceding and following the switch. For this analysis, we split each training block into thirds, and compared the end of the first training block with the beginning and middle of the second training block. SSVEP selectivity immediately after the switch (mean = −0.13, SD = 0.29) was significantly lower than that immediately before the switch (mean = 0.13, SD = 0.29, t(27) = 3.78, p < 0.001, BF10 = 2.22; Fig. 9). This reversal in amplitude shows that the contingency learned in the first half of the experiment carried over to the second half, even though it was no longer appropriate. However, by the middle of the second block, SSVEP selectivity had increased again (mean = 0.03, SD = 0.37, t(27) = 1.80, p = 0.083, BF10 = 0.82), recovering to a point where it was no longer significantly different from preswitch (end of first training block) selectivity (t(27) = 1.15, p = 0.258, BF10 = 0.37; Fig. 9). Thus, switching the feedback contingency mid-experiment did indeed dampen training effects. However, it appears that neurofeedback training was still effective in both halves of the experiment.
Discussion
By using an implicit neurofeedback protocol embedded within an integrative, perceptual decision-making task, we were able to bias sensory processing in favor of one source of evidence over another, despite the sources remaining matched for reliability (Mostert et al., 2015). This unique approach revealed how integrative decision-making processes are sensitive to biases in early sensory processing stages (Andersen et al., 2011). The induced attentional bias affected the order in which different sources of evidence were encoded, as well as the strength of encoding of the two sources of evidence. As a result, sources of evidence that were prioritized in early sensory processing were weighted more heavily in behavioral reports of the integrated average.
Our neurofeedback protocol biased neural responses to visual inputs, such that the processing of one featural attribute (in this case, a specific orientation) was prioritized over a concurrent and equally salient feature (Treue and Trujillo, 1999; Saenz et al., 2002; Maunsell and Treue, 2006). This prioritization was evident in the difference between SSVEPs evoked by the different stimulus sets on test trials, even though motion coherence was matched between the stimuli and there was no benefit to prioritizing one stimulus set over the other on these trials (Müller et al., 2006; Sorger et al., 2019). Our neurofeedback protocol was unique in that it directly targeted SSVEPs: visually evoked patterns of neural activity, which scale in amplitude with attentional allocation to the evoking object (Andersen et al., 2011; Norcia et al., 2015). Neurofeedback protocols have been used to improve performance on SSVEP-based BCIs, but these protocols have targeted SSVEPs indirectly, basing feedback on endogenous α oscillations (Wan et al., 2016), or algorithm classification accuracy (Zhang et al., 2010). In the past decade, several fMRI neurofeedback studies have endeavored to train complex spatiotemporal patterns of stimulus-evoked activity, allowing researchers to directly target neural representations of visual stimuli (Shibata et al., 2011, 2016; Scharnowski et al., 2012; Nan et al., 2013; Robineau et al., 2014; deBettencourt et al., 2015; Amano et al., 2016; Cortese et al., 2016; Habes et al., 2016). Applying this approach to SSVEPs allowed us to access a near real-time readout of visual processing, and provide feedback at a much higher temporal resolution than fMRI typically allows. This high temporal resolution, together with the flexible and dynamic nature of the signal we targeted, likely accounts for the very rapid training we achieved (Wolfe and Horowitz, 2004). Many similar neurofeedback studies targeting visual processing used multiple sessions spanning several days to effect change (Zhang et al., 2010; Strehl, 2014; Wan et al., 2016; Watanabe et al., 2017). By contrast, our neurofeedback protocol was “bidirectional,” such that the same signal was both upregulated and downregulated within a single experimental session (Sorger et al., 2019). While this approach allowed us to attribute the observed training effects definitively to the neurofeedback protocol, we found that it diminished the overall magnitude of the training effects across the experiment. In future studies, it will be important to balance the advantages of such controls over experimental design against the potential for larger, and perhaps more enduring, training effects. In the present study, while the training effect was somewhat diminished by our implementation of “bidirectional” control, it was nevertheless sufficient to bias integrative decision-making.
We investigated how the neurofeedback-induced bias in sensory processing of two sources of evidence affected integrative decision-making processes. Forward encoding analyses showed that neurofeedback training resulted in changes in both the sequential order and strength of encoding of the concurrently presented motion stimuli (Eldar et al., 2018; Tang et al., 2018, 2020). Interestingly, such “serial” encoding of display elements in integrative decision-making is characteristic of more complex decisions involving the integration of many distinct sources of evidence (Eldar et al., 2018). Serial processing allows more resources to be dedicated to the processing of each stimulus, allowing the trained stimulus to be encoded not only earlier in the trial, but also more strongly than would be possible under parallel processing (Treisman, 1969; Bergen and Julesz, 1983). It is notable that these differences in the encoding of trained and untrained stimuli were evident in both neurofeedback and test trials. However, the difference in SSVEP amplitude between these stimuli was only apparent on test trials. By contrast, the forward encoding results revealed that there were critical differences in the way the two sets of stimuli were encoded even during neurofeedback (Ho et al., 2012; Scolari et al., 2012). Moreover, the fact that these differences in the representations of the two stimuli were evident in the absence of a clear SSVEP difference suggests that the encoding results are not purely the result of a shift in sensory processing first to one stimulus, then to the other (Müller et al., 2006; Ales and Norcia, 2009). Rather, the forward encoding results more likely represent differences in sensory decision processes, with the representation of the trained motion direction prioritized over that of the untrained motion direction (Mostert et al., 2015). It should be noted that neurofeedback targeted the relative neural response to bars of the two different orientations. These orientations were entirely decoupled from motion direction. Thus, the paradigm never directly rewarded the representation of any particular motion direction, but rather encouraged sensory processing of a particular source of evidence. In time, evidence from this source was encoded earlier and more strongly than evidence from the untrained source.
These differences in the order and strength of encoding of the trained and untrained motion direction were reflected in both the neural representation and behavioral reports of the integrated average. Forward encoding results showed that this neural representation built up gradually during the trial, consistent with evidence accumulation through sequential sampling (Usher and McClelland, 2001; Ratcliff and McKoon, 2008; Kelly and O'Connell, 2015; van Vugt et al., 2019). This encoded representation was shifted away from the true average and toward the trained motion direction, and this shift was also reflected in behavior. Indeed, evidence from the stimulus prioritized in early sensory processing was weighted more strongly in participants' behavioral reports of the integrated average. Further, this response bias was strongest when the attentional bias was strongest. This pattern of results is similar to that found for sources of evidence that differ in reliability. In such studies, shifts in mean responses toward more reliable sources of evidence have been attributed to more precise neural representations, leading to stronger decision-weightings for this information (Landy et al., 1995; Zylberberg et al., 2016; Rahnev and Denison, 2018). However, while it may be optimal for integrative decision-making processes to discount genuinely unreliable sources of evidence, our findings suggest that this discounting still occurs when differences in the precision of representation of two sources of evidence originate from internal biases in early sensory processing, rather than from true differences in reliability (Trommershäuser, 2009; Ma, 2010).
Integrative perceptual decision-making is an essential cognitive function, allowing us to navigate complex and dynamic visual environments. The mechanics of these decision processes and their interaction with low-level visual processing are therefore worthy of further investigation. For example, in studies of single-stimulus decision-making, perceptual biases toward rewarded response alternatives have been shown to be strongest when responses are required within the first 500 ms after stimulus onset, tapering off as stimulus presentation times approach 2 s (Gao et al., 2011). This reduction in bias with additional sampling has been explained in terms of dynamic integration of stimulus information over time, and can be well modeled by a leaky accumulator. In leaky accumulator models, evidence accumulation is subject to leakage over time, allowing for the correction of any initial biases as more evidence is accumulated (Usher and McClelland, 2001; Tsetsos et al., 2012). It would be valuable to know whether a similar model might explain evidence integration across different sources of evidence, especially given our forward encoding results, which showed that the encoded representation of the average motion direction was initially strongly biased toward the trained motion direction. Further, here we operationalized integrative decision-making as the perceptual averaging of a single stimulus feature (motion direction) across different sources of evidence. It remains unclear whether these effects would persist when integration occurs across different stimulus features, which could be represented by separate populations of neurons (Emmanouil and Treisman, 2008). Consider crossing a busy road, for example. This is a complex decision-making task that relies on the integration of multiple distinct visual features, such as the color of the traffic lights as well as the speed and trajectory of nearby vehicles. Errors during such “complex” integrative perceptual decisions could prove fatal, and it is thus important we seek a better understanding of these processes.
There are many inherent biases in how sensory processing resources are allocated across visual inputs, with differences occurring across levels of statistical regularity (Appelle, 1972), previous reward associations (Anderson, 2016; Failing and Theeuwes, 2018), and current top-down goals (Yantis, 1993; Theeuwes et al., 2010). Our findings show that such biases in sensory processing can lead to differences in the precision and timing of the representation of evidence from different sources, leading to biased integrative perceptual decision outcomes. The ability to rapidly and implicitly induce such biased sensory processing within one experimental session has potentially wide-ranging applications in both basic and translational neuroscience. Neurofeedback protocols of this nature could be applied to address unanswered questions about how visual selective attention interacts with other cognitive functions (e.g., expectation) (Bang and Rahnev, 2017), and could be harnessed for cognitive training to enhance performance in neurologic patients (Robineau et al., 2014; Sitaram et al., 2017).
Footnotes
The authors declare no competing financial interests.
J.B.M. and A.I.R were supported by the ARC Center of Excellence for Integrative Brain Function ARC Center Grant CE140100007.
- Correspondence should be addressed to Angela I. Renton at angie.renton23{at}gmail.com