-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathsummarize.py
285 lines (223 loc) · 10.8 KB
/
summarize.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
"""
A set of functions that take the output of the read_file functions
and return a dictionary of summary statistics.
"""
def mean(x):
"""
A roll your own function to get the mean of a list of numbers
so that we don't have to use Numpy. This ignores missing values, so
mean([1,2,3,None]) = 6/3 = 2
:param x: a list of numbers
:return average: a float, the mean of the numbers in x
"""
# Ignore any "None" values
new_array = [i for i in x if i is not None]
n_entries = len(new_array)
# Calculate the mean of the new array the old-fashioned way
if n_entries > 0:
average = float(sum([i for i in new_array]))/n_entries
else: # If the array has no legitimate values, set the average to none
average = None
return average
def get1(task_dict_list):
"""
:param task_dict_list: list of dictionaries output
by read_file function
:return a dictionary of summary summary statistics including
T1_BadTouchesAllTrials
T1_BadTouchesFirst
"""
# Calculate the average number of bad touches per trial
bad_touches = [x['NumBadTouches'] for x in task_dict_list]
scores = [x['Score-incorrect only'] for x in task_dict_list]
# We should always have 12 trials, and be looking at the first
# six trials versus the last six trials, but I've added this
# midpoint variable in case for some reason there's a different
# number of trials.
mid_point = len(bad_touches) / 2
# Again, since there should be equal number of trials contributing
# to avg_first and average_last, avg_total could be computed by
# just doing mean([avg_first, avg_last]), but it's safer this way
avg_total = mean(bad_touches)
avg_first = mean(bad_touches[:mid_point])
avg_last = mean(bad_touches[mid_point:])
score_first = mean([score for score in scores[:mid_point] if score])
score_last = mean([score for score in scores[mid_point:] if score])
score_total = mean([score for score in scores if score])
return {'T1_BadTouchesAllTrials': avg_total, 'T1_BadTouchesFirst': avg_first, 'T1_BadTouchesLast': avg_last,
'T1_ScoreFirst': score_first, 'T1_ScoreLast': score_last, 'T1_ScoreAllTrials': score_total}
def get2(task_dict_list):
"""
:param task_dict_list: A list of dictionaries from task 2 log file.
Each entry of the list corresponds to a trial.
:return: a dict object with 12 keys corresponding to summary data for task 2.
This is grossly hard-coded, but it's correct
and produces a flat dictionary, which will
be easier to write to the summary file.
"""
###############################################################
# Organize Trials into separate lists according to trial type #
###############################################################
non_switch_rule_trials = []
switch_rule_trials = []
non_switch_side_trials = []
switch_side_trials = []
same_trials = []
opposite_trials = []
for trial in task_dict_list:
if trial['SwitchRule'] is True:
switch_rule_trials.append(trial)
elif trial['SwitchRule'] is False:
non_switch_rule_trials.append(trial)
if trial['SwitchSide'] is True:
switch_side_trials.append(trial)
elif trial['SwitchSide'] is False:
non_switch_side_trials.append(trial)
if trial['TargetSide'] == "same":
same_trials.append(trial)
elif trial['TargetSide'] == "opposite":
opposite_trials.append(trial)
###############################################################
# Calculate summary statistics for the different trial types. #
# Note that trials for which there is missing or incomplete #
# data are not included in these calculations. #
###############################################################
switch_rule_accuracy = mean([x['Correct'] for x in switch_rule_trials])
non_switch_rule_accuracy = mean([x['Correct'] for x in non_switch_rule_trials])
switch_side_accuracy = mean([x['Correct'] for x in switch_side_trials])
non_switch_side_accuracy = mean([x['Correct'] for x in non_switch_side_trials])
same_accuracy = mean([x['Correct'] for x in same_trials])
opposite_accuracy = mean([x['Correct'] for x in opposite_trials])
switch_rule_rt = mean([x['ReactionTime'] for x in switch_rule_trials])
non_switch_rule_rt = mean([x['ReactionTime'] for x in non_switch_rule_trials])
switch_side_rt = mean([x['ReactionTime'] for x in switch_side_trials])
non_switch_side_rt = mean([x['ReactionTime'] for x in non_switch_side_trials])
same_rt = mean([x['ReactionTime'] for x in same_trials])
opposite_rt = mean([x['ReactionTime'] for x in opposite_trials])
return {'T2_SwitchRuleAvgAccuracy': switch_rule_accuracy, 'T2_NonSwitchRuleAvgAccuracy': non_switch_rule_accuracy,
'T2_SwitchSideAvgAccuracy': switch_side_accuracy, 'T2_NonSwitchSideAvgAccuracy': non_switch_side_accuracy,
'T2_SameAccuracy': same_accuracy, 'T2_OppositeAccuracy': opposite_accuracy, 'T2_SwitchRuleRT':
switch_rule_rt, 'T2_NonSwitchRuleRT': non_switch_rule_rt, 'T2_SwitchSideRT': switch_side_rt,
'T2_NonSwitchSideRT': non_switch_side_rt, 'T2_SameRT': same_rt, 'T2_OppositeRT': opposite_rt}
def get3(task_dict_list):
"""
Reads Task 3 list of dictionaries. Importantly, each list member for
task three corresponds to an individual dot press, not an individual trial
(there may be 1,2, or 3 dot presses per trial).
EXAMPLE INPUT: would consist of a list of dictionaries like these:
{'Delay': 0.1,
'DistanceFromCenter': 38.38547,
'DotPressed': 4,
'EarlyResponse': False,
'NumDots': 1,
'Rank': 1,
'ReactionTime': 1.040039,
'ShownDots': 4,
'TimeOut': False,
'TouchPosition': (263.0, 369.0),
'TrialNum': 56},
"""
trials = []
trial_num = None
# Builds a list within which each trial is represented by a list,
# within which each dot has a (distance, delay tuple)
# TODO: This could be implemented more efficiently either as a dictionary or a different list structure
for trial in task_dict_list:
if trial['TrialNum'] == trial_num:
trials[-1].append((trial['DistanceFromCenter'], trial['Delay']))
else:
trials.append([])
trial_num = trial['TrialNum']
trials[-1].append((trial['DistanceFromCenter'], trial['Delay']))
# Now we have a list of lists of (distance, delay) tuples.
# Calculate average accuracy by load, determined by the length of the list
acc_by_load = {}
for i in range(1, 4):
load_means = [mean([dot[0] for dot in trial if dot[0]]) for trial in trials if len(trial) == i]
acc_by_load[i] = mean([x for x in load_means if x is not None])
acc_by_delay = {}
for delay in [0.1, 3]:
delay_means = [mean([dot[0] for dot in trial if dot[0] is not None])
for trial in trials if trial[0][1] == delay]
acc_by_delay[delay] = mean([x for x in delay_means if x is not None])
return {'T3_Load1Distance': acc_by_load[1], 'T3_Load2Distance': acc_by_load[2], 'T3_Load3Distance': acc_by_load[3],
'T3_Delay0.1Distance': acc_by_delay[0.1], 'T3_Delay3Distance': acc_by_delay[3]}
def get4(task_dict_list):
"""
Reads Task 4 dictionary (output from read_file)
and outputs summary data. Input looks like:
[{'AvgDistanceFromCenter': 15.30155,
'AvgResponseTime': 0.645187,
'Block': 1,
'PercentCorrect': 100.0},
{'AvgDistanceFromCenter': 22.55198,
'AvgResponseTime': 0.5908622,
'Block': 2,
'PercentCorrect': 100.0},
{'AvgDistanceFromCenter': 17.30606,
'AvgResponseTime': 0.6249163,
'Block': 3,
'PercentCorrect': 100.0},
{'AvgDistanceFromCenter': 20.74315,
'AvgResponseTime': 0.6071847,
'Block': 4,
'PercentCorrect': 100.0},
{'AvgDistanceFromCenter': 21.26595,
'AvgResponseTime': 0.5916294,
'Block': 5,
'PercentCorrect': 100.0}]
"""
rand_blocks = (1, 4)
rule_blocks = (2, 3, 5)
random_mean = mean([block['AvgResponseTime'] for block in task_dict_list if block['Block'] in rand_blocks])
rule_mean = mean([block['AvgResponseTime'] for block in task_dict_list if block['Block'] in rule_blocks])
block4_mean = task_dict_list[3]['AvgResponseTime']
block5_mean = task_dict_list[4]['AvgResponseTime']
return {'T4_RandomRT': random_mean, 'T4_RuleRT': rule_mean, 'T4_Block4RT': block4_mean,
'T4_Block5RT': block5_mean}
def get5(task_dict_list):
"""
Reads Task 5 dictionary list (output from read_file)
and outputs summary data. Input looks like:
{'AvgDistancePerTarget': 170.9032,
'AvgFirstTen': 0.5186402,
'AvgLastTen': 0.5875,
'AvgLocation': 692.6765,
'AvgTargetsPerArea': (5.0, 3.0, 6.0, 3.0, 4.0, 5.0, 5.0, 3.0),
'AvgTimePerAction': 0.5415921,
'AvgTimePerTarget': 0.5734504,
'Duration': 19.51416,
'EndCondition': 'completed',
'NumBadTouches': 2,
'NumGoodTouches': 34,
'NumRepeats': 0,
'StandardDeviation': 10.20111,
'Task': 1}
"""
values = map(mean, ([x['NumBadTouches'] for x in task_dict_list], [x['NumRepeats'] for x in task_dict_list],
[x['AvgDistancePerTarget'] for x in task_dict_list]))
return zip(['T5_NumBadTouches', 'T5_NumRepeats', 'T5_AvgDistancePerTarget'], values)
def get6(task_dict_list):
"""
:param task_dict_list: list of dictionaries output
by readTask1LogFile function
"""
# Calculate the average number of bad touches per trial
bad_touches = [x['NumBadTouches'] for x in task_dict_list]
scores = [x['Score-incorrect only'] for x in task_dict_list]
# We should always have 12 trials, and be looking at the first
# six trials versus the last six trials, but I've added this
# midpoint variable in case for some reason there's a different
# number of trials.
mid_point = len(bad_touches) / 2
# Again, since there should be equal number of trials contributing
# to avg_first and average_last, avg_total could be computed by
# just doing mean([avg_first, avg_last]), but it's safer this way
avg_total = mean(bad_touches)
avg_first = mean(bad_touches[:mid_point])
avg_last = mean(bad_touches[mid_point:])
score_first = mean([score for score in scores[:mid_point] if score])
score_last = mean([score for score in scores[mid_point:] if score])
score_total = mean([score for score in scores if score])
return {'T6_BadTouchesAllTrials': avg_total, 'T6_BadTouchesFirst': avg_first, 'T6_BadTouchesLast': avg_last,
"T6_ScoreFirst": score_first, "T6_ScoreLast": score_last, "T6_ScoreAllTrials": score_total}