13.07.2015 Views

Sweating the Small Stuff: Does data cleaning and testing ... - Frontiers

Sweating the Small Stuff: Does data cleaning and testing ... - Frontiers

Sweating the Small Stuff: Does data cleaning and testing ... - Frontiers

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Lages <strong>and</strong> JaworskaHow predictable are voluntary decisions?STUDY 1: SPONTANEOUS MOTOR DECISIONSIn this replication study we investigated response bias <strong>and</strong> responsedependency of binary decisions in a spontaneous motor task. Weclosely replicated <strong>the</strong> study by Soon et al. (2008) in terms of stimuli,task, <strong>and</strong> instructions (Soon et al., 2008) but without monitoringfMRI brain activity since we are mainly interested in behavioralcharacteristics.METHODSSubjects were instructed to relax while fixating on <strong>the</strong> center of ascreen where a stream of r<strong>and</strong>om letters was presented in 500 msintervals. At some point, when participants felt <strong>the</strong> urge to do so,<strong>the</strong>y immediately pressed one of two buttons with <strong>the</strong>ir left or rightindex finger. Simultaneously, <strong>the</strong>y were asked to remember <strong>the</strong> letterthat appeared on <strong>the</strong> screen at <strong>the</strong> time when <strong>the</strong>y believed <strong>the</strong>irdecision to press <strong>the</strong> button was made. Shortly afterward, <strong>the</strong> lettersfrom three preceding trials <strong>and</strong> an asterisk were presented onscreen r<strong>and</strong>omly arranged in a two-by-two matrix. The participantswere asked to select <strong>the</strong> remembered letter in order to report<strong>the</strong> approximate time point when <strong>the</strong>ir decision was formed. If<strong>the</strong> participant chose <strong>the</strong> asterisk it indicated that <strong>the</strong> rememberedletter was not among <strong>the</strong> three preceding intervals <strong>and</strong> <strong>the</strong>voluntary decision occurred more than 1.5 s ago. Subjects wereasked to avoid any form of preplanning for choice of movementor time of execution.PARTICIPANTSAll participants (N = 20, age 17–25, 14 female) were students atGlasgow University. They were naïve as to <strong>the</strong> aim of <strong>the</strong> study,right-h<strong>and</strong>ed,<strong>and</strong> with normal or corrected-to-normal visual acuity.The study was conducted according to <strong>the</strong> Declaration ofHelsinki ethics guidelines. Informed written consent was obtainedfrom each participant before <strong>the</strong> study.RESULTSFollowing Soon et al. (2008) we computed for each participant <strong>the</strong>frequency of a left or right response. If we assume that <strong>the</strong> spontaneousdecision task produces independent responses <strong>the</strong>n <strong>the</strong>process can be modeled by a binomial distribution where probabilityfor a left <strong>and</strong> right response may vary from participant toparticipant.P[t(x) = s|θ] =( ns)θ s (1 − θ) n−sThe observed <strong>data</strong> t(x) is simply <strong>the</strong> sum of s left (right) responses,n is <strong>the</strong> total number of responses, <strong>and</strong> θ is a parameter that reflects<strong>the</strong> unknown probability of responding Left (Right) with θ ∈ [0,1].The hypo<strong>the</strong>sis of a balanced response corresponds to a responserate of θ = 0.5. Ra<strong>the</strong>r than trying to affirm this null hypo<strong>the</strong>siswe can test whe<strong>the</strong>r <strong>the</strong> observed number of left (right) responsesdeviates significantly from <strong>the</strong> null hypo<strong>the</strong>sis by computing <strong>the</strong>corresponding p-value (two-sided).p (two - sided) =n−s∑x i =0t (x i ) +n∑t (x i )x i =sWe found that response frequencies of 4 out of 20 participants (2out of 20 if adjusted for multiple tests according to Sidak–Dunn)significantly deviated from a binomial distribution with equalprobabilities (p < 0.05, two-sided). Soon et al. (2008) excluded24 out of 36 participants who exceeded a response criterion that isequivalent to a binomial test with p < 0.11 (two-sided). Bode et al.(2011) applied a similar response criterion but did not documentselection of participants. They reported exclusion of a single participantfrom <strong>the</strong>ir sample of N = 12 due to relatively unbalanceddecisions <strong>and</strong> long trial durations; responses from <strong>the</strong> remaining11 subjects were included in <strong>the</strong>ir analyses. In <strong>the</strong> present study 8out of 20 participants did not meet Soon et al.’s response criterion(for details see Table A1 in Appendix).Selection of participants is a thorny issue. While <strong>the</strong> intentionmay have been to select participants who made truly spontaneous<strong>and</strong> <strong>the</strong>refore independent decisions <strong>the</strong>y selected participantswho generated approximately balanced responses. This assumptionis fallible since subjects’ response probabilities are unlikely tobe perfectly balanced <strong>and</strong> <strong>the</strong> null hypo<strong>the</strong>sis of θ = 0.5 can bedifficult to affirm.Excluding 2/3 of <strong>the</strong> subjects reduces generalizability of results<strong>and</strong> imposing <strong>the</strong> assumption of no response bias on <strong>the</strong> remainingsubjects seems inappropriate because <strong>the</strong>se participants can stillhave true response probabilities θ that are systematically differentfrom 0.5.To give an example of how a moderate response bias mayaffect prediction accuracy of a trained classifier, consider a participantwho generates 12 left <strong>and</strong> 20 right responses in 32 trials.Although this satisfies <strong>the</strong> response criterion mentioned above, aclassifier trained on this <strong>data</strong> is susceptible to response bias. If <strong>the</strong>classifier learns to match <strong>the</strong> individual response bias predictionaccuracy may exceed <strong>the</strong> chance level of 50%. (If, for example, <strong>the</strong>classifier trivially predicts <strong>the</strong> more frequent response <strong>the</strong>n thisstrategy leads to 62.5% ra<strong>the</strong>r than 50% correct predictions in ourexample.)To alleviate <strong>the</strong> problem of response bias Soon et al. (2008)<strong>and</strong> Bode et al. (2011) not only selected among participants butalso designated equal numbers of left (L) <strong>and</strong> right (R) responsesfrom <strong>the</strong> experimental trials before entering <strong>the</strong> <strong>data</strong> into <strong>the</strong>irclassification analysis. It is unclear how <strong>the</strong>y sampled trials buteven if <strong>the</strong>y selected trials r<strong>and</strong>omly <strong>the</strong> voxel activities beforeeach decision are drawn from an experiment with unbalanced L<strong>and</strong> R responses. As a consequence <strong>the</strong> problem does not dissipatewith trial selection. After selecting an equal number of L<strong>and</strong> R responses from <strong>the</strong> original <strong>data</strong> set this subsample stillhas an unbalanced number of L <strong>and</strong> R responses in <strong>the</strong> precedingtrials so that <strong>the</strong> distribution of all possible pairs of successiveresponses in trial t − 1 <strong>and</strong> trial t (LL, LR, RR, RL) is not uniform.Since <strong>the</strong>re are more Right responses in <strong>the</strong> original <strong>data</strong>set we are more likely to sample more RR “stay” trials <strong>and</strong> lessLR “switch” trials as well as more RL “switch” trials compared toLL “stay” trials. The exact transition probabilities for <strong>the</strong>se eventsdepend on <strong>the</strong> individual response pattern. Switching <strong>and</strong> stayingbetween successive responses creates a confounding variable thatmay introduce spurious correlations between voxel activities fromprevious responses <strong>and</strong> <strong>the</strong> predicted responses. This confoundmay be picked up when training a linear support vector machine<strong>Frontiers</strong> in Psychology | Quantitative Psychology <strong>and</strong> Measurement March 2012 | Volume 3 | Article 56 | 143

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!