Return to the Social Evaluation Learning Task page
						
								SOCIAL EVALUATION LEARNING TASK
SCRIPT INFO

Script Author: Katja Borchert, Ph.D. (katjab@millisecond.com) for Millisecond Software, LLC
Date: 03-14-2018
last updated: 04-25-2018 by K.Borchert (katjab@millisecond.com) for Millisecond Software LLC

Script Copyright © 04-25-2018 Millisecond Software


BACKGROUND INFO 

											*Purpose*
This script implements a Social Evaluation Learning Task as described in:

Button, K.S., Kounali, D., Stapinski, L., Rapee, R.M, Lewis, G., & Munafò, M.R.M (2015).
Fear of Negative Evaluation Biases Social Evaluation Inference: Evidence from a
Probabilistic Learning Task. PLOS ONE | DOI:10.1371/journal.pone.0119456

Stimuli published in:
Button KS, Browning M, Munafo MR, Lewis G (2012) Social inference and social anxiety: evidence of a
fear-congruent self-referential learning bias. J Behav Ther Exp Psychiatry 43: 1082–1087. doi: 10.
1016/j.jbtep.2012.05.004 PMID: 22699043


											  *Task*
Participants encounter 6 computer personas across 2 learning tasks during which they have to 
learn whether the computer personas like them (self referential task, SR) or like 'George' (other referential task, OR).
At the end of the SR condition, participants have to guess if the computer persona (e.g. "Alex") likes them based on 
learning what 'Alex' thinks of them. To learn what 'Alex' thinks of them, participants are given word pairs
(e.g. "witty" vs. "dull") and are asked to choose the word that corresponds to what 'Alex' thinks of them.
Feedback contingencies corresponded to 3 rules, "like", "neutral" and ""dislike", with P[positive word correct] = 0.8, 0.5 and
0.2, respectively. Each feedback contingency block is coupled with a different persona.
In the OR condition, participants have to guess if the computer persona (e.g. "Charlie") likes a person named
'George' using the same set-up.

Note: due to the nature of using 4 spatially oriented response keys this task requires
a machine with an attached keyboard. Script "socialevaluationlearningtask_MI.iqx" runs the same task
with mouse/touchscreen input.


DATA FILE INFORMATION: 
The default data stored in the data files are:

(1) Raw data file: 'socialevaluationlearningtask_raw*.iqdat' (a separate file for each participant)*

build:							Inquisit build
computer.platform:				the platform the script was run on
date, time, subject, group:		date and time script was run with the current subject/groupnumber 
/conditionOrder:				SR->OR vs. OR -> SR (Note: SR = self referential; OR = other ('George') referential)

blockcode, blocknum:			the name and number of the current block
trialcode, trialnum: 			the name and number of the currently recorded trial
									(Note: not all trials that are run might record data; by default data is collected unless /recorddata = false is set for a particular trial/block) 

/condition:						1 = SR (self rating) vs. 2 = OR (other rating) => current rating block																	
/LearningTrialCounter:			counts learning trials in each 'rule block' (block.LIKE, block.DISLIKE, block.NEUTRAL) 
									
/persona:						the 'name' of the computer persona currently 'judging'
/rule:							rule to be learned: 
									"positive" (persona likes me/George); 
									"negative" (persona does not like me/George)
									"neutral"  (persona is indifferent towards me/George)									
									
/posWord_location:				stores the current position of the positive word (1-4)
/negWord_location:				stores the current position of the negative word (1-4)
									Note: 1: upper left; 2: upper right; 3 = lower right; 4 = lower left (clockwise)	
									
/posWord:						stores the currently presented positive word
/negWord:						stores the currently presented negative word
	
/incorrResp:					stores the response key (in scancode) associated with the incorrect word
/corrResp:						stores the response key (in scancode) associated with the correct word
									Note: 16 = q; 25 = p, 44 = z; 50 = m

response:						the participant's response (scancode of selected response key)
correct:						accuracy of response: 1 = correct response; 0 = otherwise

									Learning Trials:
									LIKE/NEUTRAL condition: 1 = participant chose positive word; 0 = otherwise
									DISLIKE condition: 1 = participant chose negative word; 0 = otherwise
									
									Rating Trials:
									correct = 1 has no meaning
																		
latency: 						the response latency (in ms); measured from: onset of word pairs

/feedback:						1 = positive feedback for positive words/negative feedback for negative words
								0 = negative feedback for positive words/positive feedback for negative words
									Note: across every 10 trials, the selection of 1/0 should reflect the feedback-contingency of the block
									Example: condition 'LIKE': across every 10 trials, 8 trials have values.feedback = 1; 2 trials have values.feedback = 0
/selectFeedbackStim:			1 = correct; 2 = incorrect

/rating:						converts the rating into a discrete scale from:
								0 = completely dislike to 100 = completely like


(2) Summary data file: 'socialevaluationlearningtask_summary*.iqdat' (a separate file for each participant)*

script.startdate:				date script was run
script.starttime:				time script was started
script.subjectid:				subject id number
script.groupid:					group id number
/conditionOrder:				SR->OR vs. OR -> SR (Note: SR = self referential; OR = other ('George') referential)
script.elapsedtime:				time it took to run script (in ms)
computer.platform:				the platform the script was run on
/completed:						0 = script was not completed (prematurely aborted); 1 = script was completed (all conditions run)

/persona_SR:					the names used in the SR condition
/persona_OR:					the names used in the OR condition

/rating_LIKE_SR:				stores the rating for the LIKE SR condition on a scale from 0 = totally dislike to 100 = totally like
/rating_DISLIKE_SR:				stores the rating for the DISLIKE SR condition on a scale from 0 = totally dislike to 100 = totally like
/rating_NEUTRAL_SR:				stores the rating for the NEUTRAL SR condition on a scale from 0 = totally dislike to 100 = totally like

/rating_LIKE_OR:				stores the rating for the LIKE OR condition on a scale from 0 = totally dislike to 100 = totally like
/rating_DISLIKE_OR:				stores the rating for the DISLIKE OR condition on a scale from 0 = totally dislike to 100 = totally like
/rating_NEUTRAL_OR:				stores the rating for the NEUTRAL OR condition on a scale from 0 = totally dislike to 100 = totally like

/propLikeResp_LIKE_SR:			proportion of selecting the positive word in SR "LIKE" condition
								(=> proportion correct responses in SR 'LIKE' condition)
/meanRT_likeResp_LIKE_SR:		mean response time in ms of selecting positive word in SR "LIKE" condition

/propLikeResp_DISLIKE_SR:		proportion of selecting the positive word in SR "DISLIKE" condition
								(=> proportion error responses in SR 'DISLIKE' condition)
/meanRT_likeResp_DISLIKE_SR:	mean response time in ms of selecting positive word in SR "DISLIKE" condition

/propLikeResp_NEUTRAL_SR:		proportion of selecting the positive word in SR "NEUTRAL" condition
								(=> proportion correct responses in SR 'NEUTRAL' condition)
/meanRT_likeResp_NEUTRAL_SR:	mean response time in ms of selecting positive word in SR "NEUTRAL" condition

/propLikeResp_LIKE_OR:			proportion of selecting the positive word in OR "LIKE" condition
								(=> proportion correct responses in OR 'LIKE' condition)
/meanRT_likeResp_LIKE_OR:		mean response time in ms of selecting positive word in OR "LIKE" condition

/propLikeResp_DISLIKE_OR:		proportion of selecting the positive word in OR "DISLIKE" condition
								(=> proportion errir responses in OR 'DISLIKE' condition)
/meanRT_likeResp_DISLIKE_OR:	mean response time in ms of selecting positive word in OR "DISLIKE" condition

/propLikeResp_NEUTRAL_OR:		proportion of selecting the positive word in OR "NEUTRAL" condition
								(=> proportion correct responses in OR 'NEUTRAL' condition)
/meanRT_likeResp_NEUTRAL_OR:	mean response time in ms of selecting positive word in OR "NEUTRAL" condition


* separate data files: to change to one data file for all participants (on Inquisit Lab only), go to section
"DATA" and follow further instructions


EXPERIMENTAL SET-UP

* order of SR (self-referential) vs. OR (other-referential) conditions is counterbalanced by groupnumber
odd groupnumbers : SR->OR
even groupnumbers: OR->SR

* personas are selected randomly for each condition from a pool
of three gender-neutral (English speaking countries) names (see list.personas_SR and list.personas_OR)

SR condition: (reference: self)
- 3 blocks: like, neutral, dislike (order is randomly determined)
- each block runs 32 trials, randomly selecting word pairs from a pool of 64 word pairs (no repeats within a block)
- one word of each word pair is 'positive', the other 'negative'
- the two words are randomly assigned to one of 4 screen locations (upper left, upper right, lower right, lower left)

- LIKE-block: across every 10 trials, participants receive positive feedback if they select the positive description 8 times
(2 randomly selected times, participants get negative feedback for selecting the positive word;
the contingeny is reversed for selecting the negative description)

- NEUTRAL-block: across every 10 trials, participants receive positive feedback if they select the positive description 5 times
(5 randomly selected times, participants get negative feedback for selecting the positive word;
the contingeny is reversed for selecting the negative description)

- DISLIKE-block: across every 10 trials, participants receive positive feedback if they select the positive description 2 times
(8 randomly selected times, participants get negative feedback for selecting the positive word;
the contingeny is reversed for selecting the negative description)

- at the end of each block, participants are asked to decide whether the computer liked them or disliked them and with what
probability (e.g. '70% liked me')

OR condition (reference: a person named George)
- same set up as SR blocks


STIMULI
see Button et al (2012)
stimuli under section "Editable Stimuli"

Note: all elements that present text can be found under section 'Editable Stimuli', 'Editable Instructions'
or 'Editable Lists' for easy editing/translating of the task

INSTRUCTIONS
generated based on Button et al (2015) and an eprime script running the task
see section "Editable Instructions"

EDITABLE CODE:
check below for (relatively) easily editable parameters, stimuli, instructions etc. 
Keep in mind that you can use this script as a template and therefore always "mess" with the entire code to further customize your experiment.

The parameters you can change are:

/topleftkey:				scancode of the top left key (default: 16 -> q)
/toprightkey:				scancode of the top right key (default: 25 -> p)
/bottomleftkey:				scancode of the bottom left key (default: 44 -> z)
/bottomrightkey:			scancode of the top left key (default: 16 -> q)
							Note: if the response keys are changed, picture.keyboard needs to be updated

/feedbackDuration:			the duration (in ms) of the feedback (default: 2000ms)

Copyright © Millisecond Software. All rights reserved.
Contact | Terms of Service | Privacy Statement | Security Statement | GDPR