evijit HF staff commited on
Commit
a226d69
β€’
1 Parent(s): fcd3e1e

Update model_data/model_a_data.json

Browse files
Files changed (1) hide show
  1. model_data/model_a_data.json +304 -174
model_data/model_a_data.json CHANGED
@@ -1,219 +1,349 @@
1
  {
2
  "metadata": {
3
- "Name": "Model A",
4
- "Provider": "TechCorp",
5
- "Version": "2.1",
6
- "Release Date": "2023-09-15",
7
  "Type": "Large Language Model",
8
  "Modalities": ["Text-to-Text"]
9
  },
10
  "scores": {
11
  "Bias, Stereotypes, and Representational Harms Evaluation": {
12
- "Comprehensive Evaluation Methodology": {
13
  "status": "Yes",
14
- "source": "Both",
15
- "applicable_evaluations": [
16
- "Evaluations at various stages (data collection, preprocessing, model architecture, training, deployment)",
17
- "Both intrinsic (e.g., embedding analysis) and extrinsic (e.g., downstream task performance) evaluation methods"
18
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  },
20
- "Inclusive Protected Class Consideration": {
21
  "status": "No",
22
- "source": null,
23
- "applicable_evaluations": [
24
- "Evaluation of non-standard protected classes (e.g., socioeconomic status, education level, regional differences)",
25
- "Consideration of intersectionality and how identity aspects interact",
26
- "Assessment of potential harms to non-typical groups (e.g., by profession or hobbies)"
27
- ]
28
- },
29
- "Cultural and Linguistic Diversity": {
30
- "status": "Yes",
31
- "source": "3P",
32
- "applicable_evaluations": [
33
- "Tests of model performance and biases across languages and cultures",
34
- "Consideration of how protected categories may shift in meaning across regions"
35
- ]
36
  },
37
- "Stereotype and Harmful Association Detection": {
38
  "status": "Yes",
39
- "source": "1P",
40
- "applicable_evaluations": [
41
- "Detection of stereotypical word associations in text models",
42
- "Sentiment analysis and toxicity measurements, especially regarding specific groups"
43
- ]
 
 
 
 
 
 
 
 
 
 
 
 
44
  },
45
- "Performance Disparities Assessment": {
46
- "status": "No",
47
- "source": null,
48
- "applicable_evaluations": [
49
- "Detailed breakdowns of performance metrics (accuracy, precision, recall) for various subgroups",
50
- "Performance analysis for disadvantaged subgroups",
51
- "Intersectionality considerations in performance analysis"
52
- ]
 
 
 
 
 
 
 
53
  }
54
  },
55
  "Cultural Values and Sensitive Content Evaluation": {
56
- "Hate Speech and Toxicity Evaluation": {
57
- "status": "Yes",
58
- "source": "Both",
59
- "applicable_evaluations": [
60
- "Assessments of harmful text generation",
61
- "Evaluations of toxicity, hurtfulness, or offensiveness"
62
- ]
63
  },
64
- "Cultural Value Representation": {
65
- "status": "No",
66
- "source": null,
67
- "applicable_evaluations": [
68
- "Use of pre-existing scholarship (e.g., World Values Survey, Geert Hofstede's work)",
69
- "Inductive and participatory evaluations grounded in specific cultural contexts",
70
- "Assessments of ethical scenarios and political value representation"
71
- ]
72
- },
73
- "Diverse Cultural Context": {
74
  "status": "Yes",
75
- "source": "3P",
76
- "applicable_evaluations": [
77
- "Assessments that don't equate nationality with cultural context",
78
- "Representation of differing cultural values within countries"
79
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80
  }
81
  },
82
  "Disparate Performance": {
83
- "Subpopulation Performance Analysis": {
84
- "status": "Yes",
85
- "source": "1P",
86
- "applicable_evaluations": [
87
- "Non-aggregated (disaggregated) evaluation results with in-depth breakdowns across subpopulations",
88
- "Metrics such as subgroup accuracy, calibration, AUC, recall, precision, min-max ratios"
89
- ]
90
  },
91
- "Cross-lingual and Dialect Evaluation": {
92
- "status": "No",
93
- "source": null,
94
- "applicable_evaluations": [
95
- "Cross-lingual prompting on standard benchmarks",
96
- "Examination of performance across dialects",
97
- "Analysis of hallucination disparity across languages"
98
- ]
99
- },
100
- "Image Generation Quality Assessment": {
101
  "status": "N/A",
102
- "source": null,
103
- "applicable_evaluations": []
 
 
 
 
 
104
  }
105
  },
106
  "Environmental Costs and Carbon Emissions Evaluation": {
107
- "Energy Consumption Measurement": {
108
  "status": "Yes",
109
- "source": "1P",
110
- "applicable_evaluations": [
111
- "Measurement of energy used in training, testing, and deploying the system",
112
- "Evaluation of compute power consumption"
113
- ]
 
 
 
 
 
 
 
 
 
114
  },
115
- "Carbon Footprint Quantification": {
116
- "status": "No",
117
- "source": null,
118
- "applicable_evaluations": [
119
- "Use of tools like CodeCarbon or Carbontracker",
120
- "Measurement of carbon emissions for training and inference",
121
- "Conversion of energy consumption to carbon emissions"
122
- ]
123
- },
124
- "Hardware Resource Evaluation": {
125
  "status": "Yes",
126
- "source": "1P",
127
- "applicable_evaluations": [
128
- "Assessment of CPU, GPU, and TPU usage",
129
- "Measurement of FLOPS (Floating Point Operations)"
130
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
131
  }
132
  },
133
  "Privacy and Data Protection Evaluation": {
134
- "Data Minimization and Consent Practices": {
135
  "status": "Yes",
136
- "source": "Both",
137
- "applicable_evaluations": [
138
- "Implementation of data minimization practices",
139
- "Use of opt-in data collection methods",
140
- "Assessment of active consent for collecting, processing, and sharing data"
141
- ]
142
- },
143
- "Memorization and Data Leakage Evaluation": {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
144
  "status": "Yes",
145
- "source": "1P",
146
- "applicable_evaluations": [
147
- "Examination of the maximum amount of discoverable information given training data",
148
- "Evaluation of extractable information without training data access"
149
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
150
  },
151
- "Personal Information Revelation Assessment": {
152
- "status": "No",
153
- "source": null,
154
- "applicable_evaluations": [
155
- "Direct prompting tests to reveal Personally Identifiable Information (PII)",
156
- "Use of tools like ProPILE to audit PII revelation likelihood",
157
- "Evaluation of the system's ability to infer personal attributes"
158
- ]
 
 
 
 
 
 
 
159
  }
160
  },
161
  "Financial Costs Evaluation": {
162
- "Comprehensive Cost Evaluation": {
163
- "status": "Yes",
164
- "source": "1P",
165
- "applicable_evaluations": [
166
- "Estimation of infrastructure and hardware costs",
167
- "Calculation of labor hours from researchers, developers, and crowd workers",
168
- "Tracking of compute costs using low-cost or standard pricing per instance-hour"
169
- ]
170
- },
171
- "Storage and Training Cost Analysis": {
172
- "status": "Yes",
173
- "source": "1P",
174
- "applicable_evaluations": [
175
- "Assessment of storage costs for both datasets and resulting models",
176
- "Consideration of in-house vs. cloud storage options",
177
- "Evaluation of training costs based on in-house GPUs or per-hour-priced instances"
178
- ]
179
- },
180
- "Hosting and Inference Cost Evaluation": {
181
- "status": "No",
182
- "source": null,
183
- "applicable_evaluations": [
184
- "Evaluation of low-latency serving costs",
185
- "Assessment of inference costs based on token usage",
186
- "Consideration of factors such as initial prompt length and requested token response length"
187
- ]
188
  }
189
  },
190
  "Data and Content Moderation Labor Evaluation": {
191
- "Crowdwork Standards Compliance": {
192
- "status": "No",
193
- "source": null,
194
- "applicable_evaluations": [
195
- "Assessment of compliance with Criteria for Fairer Microwork",
196
- "Evaluation against Partnership on AI's Responsible Sourcing of Data Enrichment Services guidelines",
197
- "Comparison with Oxford Internet Institute's Fairwork Principles"
198
- ]
199
- },
200
- "Crowdworker Demographics and Compensation": {
201
  "status": "Yes",
202
- "source": "3P",
203
- "applicable_evaluations": [
204
- "Documentation of crowd workers' demographics",
205
- "Transparency in reporting instructions given to crowdworkers",
206
- "Assessment of how crowdworkers were evaluated and compensated"
207
- ]
208
- },
209
- "Psychological Support and Content Exposure": {
210
- "status": "No",
211
- "source": null,
212
- "applicable_evaluations": [
213
- "Documentation of immediate trauma support availability",
214
- "Assessment of long-term professional psychological support provision",
215
- "Evaluation of practices for controlling exposure to traumatic material"
216
- ]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
217
  }
218
  }
219
  }
 
1
  {
2
  "metadata": {
3
+ "Name": "StarCoder2",
4
+ "Provider": "BigCode",
5
+ "URL": "https://huggingface.co/bigcode/starcoder2-15b",
 
6
  "Type": "Large Language Model",
7
  "Modalities": ["Text-to-Text"]
8
  },
9
  "scores": {
10
  "Bias, Stereotypes, and Representational Harms Evaluation": {
11
+ "1.1 Bias Detection Overview": {
12
  "status": "Yes",
13
+ "sources": [
14
+ {
15
+ "type": "🌐",
16
+ "detail": "https://arxiv.org/abs/2402.19173",
17
+ "name": "BOLD - Bias in Open-ended Language Generation Dataset"
18
+ },
19
+ {
20
+ "type": "🌐",
21
+ "detail": "https://arxiv.org/abs/2402.19173",
22
+ "name": "WinoBias"
23
+ }
24
+ ],
25
+ "questions": {
26
+ "Evaluations at various stages (data collection, preprocessing, AI system architecture, training, deployment)": false,
27
+ "Have intrinsic properties of the AI system been evaluated for bias (e.g., embedding analysis)": false,
28
+ "Have extrinsic bias evaluations been run (e.g., downstream task performance)": true,
29
+ "Have evaluations been run across all applicable modalities": true,
30
+ "Have bias evaluations been run that take the form of automatic quantitative evaluation": true,
31
+ "Have bias evaluations been run with human participants?": false
32
+ }
33
  },
34
+ "1.2 Protected Classes and Intersectional Measures": {
35
  "status": "No",
36
+ "sources": [],
37
+ "questions": {
38
+ "Do evaluations cover all applicable legal protected categories for in-scope uses of the system?": false,
39
+ "Do evaluations cover additional subgroups that are likely to be harmed based on other personal characteristics": false,
40
+ "Evaluation of how different aspects of identity interact and compound in AI system behavior": false,
41
+ "Evaluation of AI system biases for legal protected categories and additional relevant subgroups": false
42
+ }
 
 
 
 
 
 
 
43
  },
44
+ "1.3 Measurement of Stereotypes and Harmful Associations": {
45
  "status": "Yes",
46
+ "sources": [
47
+ {
48
+ "type": "🌐",
49
+ "detail": "https://arxiv.org/abs/2402.19173",
50
+ "name": "HONEST - Hurtful Sentence Completion in English Language Models"
51
+ },
52
+ {
53
+ "type": "🌐",
54
+ "detail": "https://arxiv.org/abs/2402.19173",
55
+ "name": "RealToxicityPrompts"
56
+ }
57
+ ],
58
+ "questions": {
59
+ "Measurement of known stereotypes in AI system outputs": true,
60
+ "Measurement of other negative associations and assumptions regarding specific groups": true,
61
+ "Measurement of stereotypes and negative associations across in-scope contexts": false
62
+ }
63
  },
64
+ "1.4 Bias Evaluation Transparency and Documentation": {
65
+ "status": "Yes",
66
+ "sources": [
67
+ {
68
+ "type": "🌐",
69
+ "detail": "https://arxiv.org/abs/2402.19173",
70
+ "name": "Evaluation Documentation"
71
+ }
72
+ ],
73
+ "questions": {
74
+ "Sufficient documentation of evaluation methods (including code and datasets) to replicate findings": true,
75
+ "Sufficient documentation of evaluation results (including intermediary statistics) to support comparison to other AI systems": true,
76
+ "Documentation of bias mitigation measures, including their secondary impacts": false,
77
+ "Documentation of bias monitoring approaches post-release/deployment if applicable": false
78
+ }
79
  }
80
  },
81
  "Cultural Values and Sensitive Content Evaluation": {
82
+ "2.1 Cultural Variation Overview": {
83
+ "status": "N/A",
84
+ "sources": [],
85
+ "questions": {}
 
 
 
86
  },
87
+ "2.2 Cultural Diversity and Representation": {
88
+ "status": "N/A",
89
+ "sources": [],
90
+ "questions": {}
91
+ },
92
+ "2.3 Generated Sensitive Content across Cultural Contexts": {
 
 
 
 
93
  "status": "Yes",
94
+ "sources": [
95
+ {
96
+ "type": "🌐",
97
+ "detail": "https://arxiv.org/abs/2402.19173",
98
+ "name": "HONEST - Hurtful Sentence Completion in English Language Models"
99
+ },
100
+ {
101
+ "type": "🌐",
102
+ "detail": "https://arxiv.org/abs/2402.19173",
103
+ "name": "RealToxicityPrompts"
104
+ }
105
+ ],
106
+ "questions": {
107
+ "Has the AI system been evaluated for its likelihood of facilitating generation of threatening or violent content": true,
108
+ "Has the AI system been evaluated for its likelihood of facilitating generation of targeted harassment or discrimination": false,
109
+ "Has the AI system been evaluated for its likelihood of facilitating generation of hate speech": false,
110
+ "Has the AI system been evaluated for its likelihood of exposing its direct users to content embedding values and assumptions": false,
111
+ "Has the AI system been evaluated for its likelihood of exposing its direct users to categories of content that might be inappropriate": true,
112
+ "Has the AI system been evaluated for its likelihood of exposing its direct users to content that might have additional negative psychological impacts": false,
113
+ "Has the evaluation of the AI system's behaviors explicitly considered cultural variation": false
114
+ }
115
+ },
116
+ "2.4 Cultural Variation Transparency and Documentation": {
117
+ "status": "N/A",
118
+ "sources": [],
119
+ "questions": {}
120
  }
121
  },
122
  "Disparate Performance": {
123
+ "3.1 Disparate Performance Overview": {
124
+ "status": "N/A",
125
+ "sources": [],
126
+ "questions": {}
 
 
 
127
  },
128
+ "3.2 Identifying Target Groups for Disparate Performance Evaluation": {
129
+ "status": "N/A",
130
+ "sources": [],
131
+ "questions": {}
132
+ },
133
+ "3.3 Subgroup Performance Analysis": {
 
 
 
 
134
  "status": "N/A",
135
+ "sources": [],
136
+ "questions": {}
137
+ },
138
+ "3.4 Disparate Performance Evaluation Transparency and Documentation": {
139
+ "status": "N/A",
140
+ "sources": [],
141
+ "questions": {}
142
  }
143
  },
144
  "Environmental Costs and Carbon Emissions Evaluation": {
145
+ "4.1 Environmental Costs Overview": {
146
  "status": "Yes",
147
+ "sources": [
148
+ {
149
+ "type": "🌐",
150
+ "detail": "https://mlco2.github.io/impact/#compute",
151
+ "name": "Machine Learning Emissions Calculator"
152
+ }
153
+ ],
154
+ "questions": {
155
+ "Evaluations of different processes within development and deployment": false,
156
+ "Have evaluations been run across all applicable modalities?": true,
157
+ "Have evaluations been run on standardized benchmarks or metrics?": true,
158
+ "Have evaluations taken into account community feedback from regions affected by data center power consumption?": false,
159
+ "Do evaluations consider the full supply chain including environmental impact of hardware components and data centers used?": false
160
+ }
161
  },
162
+ "4.2 Energy Cost and Environmental Impact of Development": {
 
 
 
 
 
 
 
 
 
163
  "status": "Yes",
164
+ "sources": [
165
+ {
166
+ "type": "🌐",
167
+ "detail": "https://mlco2.github.io/impact/#compute",
168
+ "name": "Machine Learning Emissions Calculator"
169
+ }
170
+ ],
171
+ "questions": {
172
+ "Accounting of FLOPS across development stages": true,
173
+ "Evaluation of energy consumption using standardized tracking tools": true,
174
+ "Evaluation of carbon impact accounting for regional energy sources": true,
175
+ "Evaluation of hardware lifecycle environmental impact": false
176
+ }
177
+ },
178
+ "4.3 Energy Cost and Environmental Impact of Deployment": {
179
+ "status": "N/A",
180
+ "sources": [],
181
+ "questions": {}
182
+ },
183
+ "4.4 Environmental Costs Transparency and Documentation": {
184
+ "status": "Yes",
185
+ "sources": [
186
+ {
187
+ "type": "🌐",
188
+ "detail": "https://mlco2.github.io/impact/#compute",
189
+ "name": "Machine Learning Emissions Calculator"
190
+ }
191
+ ],
192
+ "questions": {
193
+ "Documentation about equipment and infrastructure specifications": true,
194
+ "Sufficient documentation of evaluation methods including components covered": false,
195
+ "Sufficient documentation of evaluation methods to replicate findings": true,
196
+ "Sufficient documentation of evaluation results for comparison": true
197
+ }
198
  }
199
  },
200
  "Privacy and Data Protection Evaluation": {
201
+ "5.1 Privacy and Data Protection Overview": {
202
  "status": "Yes",
203
+ "sources": [
204
+ {
205
+ "type": "🏒",
206
+ "detail": "PII detection and redaction using an NER model"
207
+ },
208
+ {
209
+ "type": "🌐",
210
+ "detail": "https://huggingface.co/spaces/bigcode/in-the-stack",
211
+ "name": "Opt-out tool for users"
212
+ },
213
+ {
214
+ "type": "🌐",
215
+ "detail": "https://arxiv.org/abs/2402.19173",
216
+ "name": "Asleep at the Keyboard Security Benchmark"
217
+ }
218
+ ],
219
+ "questions": {
220
+ "Evaluations at various stages (data collection, preprocessing, AI system architecture, training, deployment)": true,
221
+ "Have intrinsic properties of the AI system been evaluated for privacy vulnerabilities": false,
222
+ "Have extrinsic privacy evaluations been run": true,
223
+ "Have evaluations been run across all applicable modalities": true,
224
+ "Have privacy evaluations been run that take the form of automatic quantitative evaluation": true,
225
+ "Have privacy evaluations been run with human participants?": false
226
+ }
227
+ },
228
+ "5.2 Privacy, Likeness, and Publicity Harms": {
229
+ "status": "N/A",
230
+ "sources": [],
231
+ "questions": {}
232
+ },
233
+ "5.3 Intellectual Property and Information Security": {
234
  "status": "Yes",
235
+ "sources": [
236
+ {
237
+ "type": "🏒",
238
+ "detail": "Membership test to find if generated code was copied from the training corpus"
239
+ },
240
+ {
241
+ "type": "🏒",
242
+ "detail": "Code attribution tool to find the original author and license of the generated code"
243
+ },
244
+ {
245
+ "type": "🌐",
246
+ "detail": "https://arxiv.org/abs/2402.19173",
247
+ "name": "Asleep at the Keyboard Security Benchmark"
248
+ }
249
+ ],
250
+ "questions": {
251
+ "Has the AI system been evaluated for its likelihood of reproducing other categories of information from its training data": true,
252
+ "Has the system been evaluated for other information security risks for in-scope uses": false
253
+ }
254
  },
255
+ "5.4 Privacy Evaluation Transparency and Documentation": {
256
+ "status": "Yes",
257
+ "sources": [
258
+ {
259
+ "type": "🏒",
260
+ "detail": "Documentation of training data information risk categories and consent status"
261
+ }
262
+ ],
263
+ "questions": {
264
+ "Documentation of the categories of training data that present information risk": true,
265
+ "Documentation of evaluation methods to replicate findings": true,
266
+ "Documentation of evaluation results to support comparison": true,
267
+ "Documentation of evaluation limitations": false,
268
+ "Documentation of deployment considerations": false
269
+ }
270
  }
271
  },
272
  "Financial Costs Evaluation": {
273
+ "6.1 Financial Costs Overview": {
274
+ "status": "N/A",
275
+ "sources": [],
276
+ "questions": {}
277
+ },
278
+ "6.2 Development and Training Costs": {
279
+ "status": "N/A",
280
+ "sources": [],
281
+ "questions": {}
282
+ },
283
+ "6.3 Deployment and Operation Costs": {
284
+ "status": "N/A",
285
+ "sources": [],
286
+ "questions": {}
287
+ },
288
+ "6.4 Financial Cost Documentation and Transparency": {
289
+ "status": "N/A",
290
+ "sources": [],
291
+ "questions": {}
 
 
 
 
 
 
 
292
  }
293
  },
294
  "Data and Content Moderation Labor Evaluation": {
295
+ "7.1 Labor Evaluation Overview": {
 
 
 
 
 
 
 
 
 
296
  "status": "Yes",
297
+ "sources": [
298
+ {
299
+ "type": "🏒",
300
+ "detail": "PII annotations by human annotators with fair wage"
301
+ }
302
+ ],
303
+ "questions": {
304
+ "Evaluation of labor practices at various stages": true,
305
+ "Have labor conditions been evaluated for different worker categories": true,
306
+ "Have labor evaluations been run across all applicable task types": false,
307
+ "Have labor practices been evaluated against established industry standards": true,
308
+ "Have labor evaluations included both direct employees and contracted workers": false,
309
+ "Have evaluations considered different regional and jurisdictional contexts": true
310
+ }
311
+ },
312
+ "7.2 Working Conditions and Compensation": {
313
+ "status": "Yes",
314
+ "sources": [
315
+ {
316
+ "type": "🏒",
317
+ "detail": "PII annotations by human annotators with fair wage"
318
+ }
319
+ ],
320
+ "questions": {
321
+ "Assessment of compensation relative to local living wages and industry standards": true,
322
+ "Assessment of job security and employment classification": false,
323
+ "Evaluation of workplace safety, worker protections and rights": false,
324
+ "Assessment of worker autonomy and task assignment practices": false,
325
+ "Evaluation of power dynamics and worker feedback mechanisms": false
326
+ }
327
+ },
328
+ "7.3 Worker Wellbeing and Support": {
329
+ "status": "N/A",
330
+ "sources": [],
331
+ "questions": {}
332
+ },
333
+ "7.4 Labor Practice Documentation and Transparency": {
334
+ "status": "Yes",
335
+ "sources": [
336
+ {
337
+ "type": "🏒",
338
+ "detail": "PII annotations by human annotators with fair wage"
339
+ }
340
+ ],
341
+ "questions": {
342
+ "Documentation of labor evaluation methodology and frameworks used": true,
343
+ "Documentation of worker demographics and task distribution": false,
344
+ "Documentation of support systems, worker protections": false,
345
+ "Documentation of incident reporting and resolution procedures": false
346
+ }
347
  }
348
  }
349
  }