___________________________________________________________________________________________________________________ Internal–External Attention Task ___________________________________________________________________________________________________________________ Script Author: David Nitz (dave@millisecond.com) for Millisecond Software, LLC last updated: 02-23-2022 by K. Borchert (katjab@millisecond.com) for Millisecond Software, LLC Script Copyright © 02-23-2022 Millisecond Software ___________________________________________________________________________________________________________________ BACKGROUND INFO ___________________________________________________________________________________________________________________ This script implements the Internal-External Attention Task Paradigm that uses a probe-detection paradigm to measure internal (e.g. focusing on bodily sensations) vs external attention (e.g. visual stimuli on the screen). For details on the procedure implemented by this script refer to Mansell, W., Clark, D. M., & Ehlers, A. (2003). Internal versus external attention in social anxiety: an investigation using a novel paradigm. Behaviour Research and Therapy, 41, 555–572. ___________________________________________________________________________________________________________________ TASK DESCRIPTION ___________________________________________________________________________________________________________________ Participants are asked to react to a series of probes as quickly as possible while looking at a series of pictures on a computer screen that are each displayed for app. 25 s. Pictures include faces of happy, angry, and neutral facial expressions of males and females as well as objects. Two types of probes are used: Participants are asked to press the Spacebar whenever they feel a slight vibration, claimed to be due to changes in their physiology (internal) and measured by some sensor, and whenever they see an "E" flashed onto the screen (external). ___________________________________________________________________________________________________________________ DURATION ___________________________________________________________________________________________________________________ the default set-up of the script takes appr. 8 minutes to complete ___________________________________________________________________________________________________________________ DATA FILE INFORMATION ___________________________________________________________________________________________________________________ The default data stored in the data files are: (1) Raw data file: 'ieat_raw*.iqdat' (a separate file for each participant) build: The specific Inquisit version used (the 'build') that was run computer.platform: the platform the script was run on (win/mac/ios/android) date, time: date and time script was run subject, group, with the current subject/groupnumber session: with the current session id blockcode, blocknum: the name and number of the current block (built-in Inquisit variable) trialcode, trialnum: the name and number of the currently recorded trial (built-in Inquisit variable) Note: trialnum is a built-in Inquisit variable; it counts all trials run; even those that do not store data to the data file such as feedback trials testround: counts the testrounds (1-4) blockCount: counts the test blocks (1-4) per test round trialcount: counts the trials per test block (1-8) imageCategory: the image category tested in the current block: angry, happy, neutral, object probetype: 1 = internal probe; 2 = external probe probeOnset: probe stimulus onset asynchrony in ms (onset of probe after image onset) picture.currentpic.currentitem: stores the currently presented picture pictype: the type of the current picture (female-angry, male-angry, female-happy, male-happy etc.) response: the participant's response (scancode of response buttons) 57 = spacebar press 0 = no response correct: accuracy of response: 1 = correct response (spacebar press within 3000ms of presentation of probe); 0 = otherwise latency: the response latency (in ms); measured from onset of probe nprobes: counts the number of probes per pictype in a block nProbesTestround: counts the number of total probes run in a test round probeseq: stores the current fixed probe sequence (one of 4 possible) probepattern: stores the currently Probe Timing Patterns: 1 designates 'x1' (see list.x1) 2 designates 'x2' (see list.x2) 3 designates 'y1' (see list.y1) 4 designates 'y2' (see list.y2) (2) Summary data file: 'ieat_summary*.iqdat' (a separate file for each participant) inquisit.version: Inquisit version run computer.platform: the platform the script was run on (win/mac/ios/android) startdate: date script was run starttime: time script was started subjectid: assigned subject id number groupid: assigned group id number sessionid: assigned session id number elapsedtime: time it took to run script (in ms); measured from onset to offset of script completed: 0 = script was not completed (prematurely aborted); 1 = script was completed (all conditions run) propCorrect_Int: Proportion Spacebar press with latencies <= responseTimeout (here: 3000ms) for internal probes propCorrect_Ext: Proportion Spacebar press with latencies <= responseTimeout (here: 3000ms) for external probes meanRT_Int: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for internal probes meanRT_Ext: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for external probes AB: attention bias (difference: int-ext) => positive: participant was faster to attend to external than internal probes => negative: participant was faster to attend to internal than external probes meanRT_angry_int: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for internal probes on angry faces meanRT_angry_ext: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for external probes on angry faces AB_angry: attention bias (difference: int-ext) for angry faces => positive: participant was faster to attend to external than internal probes for angry faces => negative: participant was faster to attend to internal than external probes for angry faces meanRT_happy_int: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for internal probes on happy faces meanRT_happy_ext: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for external probes on happy faces AB_happy: attention bias (difference: int-ext) for happy faces => positive: participant was faster to attend to external than internal probes for happy faces => negative: participant was faster to attend to internal than external probes for happy faces meanRT_neutral_int: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for internal probes on neutral faces meanRT_neutral_ext: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for external probes on neutral faces AB_neutral: attention bias (difference: int-ext) for neutral faces => positive: participant was faster to attend to external than internal probes for neutral faces => negative: participant was faster to attend to internal than external probes for neutral faces meanRT_object_int: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for internal probes on objects meanRT_object_ext: mean latency (in ms) of pressing spacebar with latencies <= responseTimeout for external probes on objects AB_object: attention bias (difference: int-ext) for objects => positive: participant was faster to attend to external than internal probes for objects => negative: participant was faster to attend to internal than external probes for objects ___________________________________________________________________________________________________________________ EXPERIMENTAL SET-UP ___________________________________________________________________________________________________________________ Practice: 8 practice trials (1 practice image with 4 external and 4 internal probe trials; order is fixed) Test: * four test rounds are run (32 trials per test round => 128 trials total) * each test round runs four image category blocks (angry, happy, neutral faces, objects) - the image categories run in a blocked design: * each image category block runs 8 trials (4 external, 4 internal) => each image category runs 32 trials across the four test rounds (16 internal, 16 external) * there are 4 different probe sequences: each image category block selects one of the sequences randomly with the constraint that across the four test rounds, each image category runs each probe sequence once. * there are 4 different probe SOA patterns: each image category block selects one of four Probe SOAs pattern (see list.x1, list.x2, list.y1, list.y2 for more information) randomly with the constraint that each image category runs each of the four patterns once across the four test rounds * the individual pictures of each image category block are selected randomly (e.g. the four angry pictures are randomly assigned to testround1-testround4) * the four test rounds differ in the order of the four image category blocks (order is determined by a Latin Square) * four groupnumbers run 4 different sequences of the four testrounds: Groupnumber 1: runs "angry, happy, neutral, object" (round1), "happy, neutral, object, angry" (round2), "neutral, object, angry, happy" (round3), "object, angry, happy, neutral" (round4) Groupnumber 2: runs "happy, neutral, object, angry" (round1), "neutral, object, angry, happy" (round2), "object, angry, happy, neutral" (round3), "angry, happy, neutral, object" (round4) Groupnumber 3: runs "neutral, object, angry, happy" (round1), "object, angry, happy, neutral" (round2), "angry, happy, neutral, object" (round3), "happy, neutral, object, angry" (round4) Groupnumber 3: runs "object, angry, happy, neutral" (round1), "angry, happy, neutral, object" (round2), "happy, neutral, object, angry" (round3), "neutral, object, angry, happy" (round4) Trial Sequence: Stimulus (e.g. object) for assigned SOA -> probe (for 100ms) -> response (measured from onset of probe; max response time: 3000ms) ___________________________________________________________________________________________________________________ STIMULI ___________________________________________________________________________________________________________________ NOTE: The stimuli used in this script are not the same as in Mansell et al. (2003). Original face and object images are not in the public domain and/or may not be redistributed. Face images courtesy of the Face-Place Face Database Project (http://www.face-place.org/). Copyright 2008, Michael J. Tarr, Center for the Neural Basis of Cognition, Carnegie Mellon University (http://www.tarrlab.org/). Funding provided by NSF award 0339122. Face stimuli released under the Creative Commons Attribution Share Alike license (https://creativecommons.org/licenses/by-sa/3.0/). Object images courtesy of the Object Data Bank. Copyright 1996, Brown University, Providence, RI. All Rights Reserved. Permission to use, copy, modify, and distribute this software and its documentation for any purpose other than its incorporation into a commercial product is hereby granted without fee, provided that the above copyright notice appear in all copies and that both that copyright notice and this permission notice appear in supporting documentation, and that the name of Brown University not be used in advertising or publicity pertaining to distribution of the software without specific, written prior permission. Images produced by this software are also copyright Brown University and may not be used for any commercial purpose. BROWN UNIVERSITY DISCLAIMS ALL WARRANTIES WITH REGARD TO THIS SOFTWARE, INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR ANY PURPOSE. IN NO EVENT SHALL BROWN UNIVERSITY BE LIABLE FOR ANY SPECIAL, INDIRECT OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. ___________________________________________________________________________________________________________________ INSTRUCTIONS ___________________________________________________________________________________________________________________ see section Editable Instructions ___________________________________________________________________________________________________________________ EDITABLE CODE ___________________________________________________________________________________________________________________ check below for (relatively) easily editable parameters, stimuli, instructions etc. Keep in mind that you can use this script as a template and therefore always "mess" with the entire code to further customize your experiment. The parameters you can change are: /responseTimeOut: the response TimeOut in ms (default: 3000ms) after 3000ms the response will be coded as an error response