Spaces:
Runtime error
Runtime error
Update app.py
Browse filesadding headers
app.py
CHANGED
@@ -22,7 +22,9 @@ with gr.Blocks() as demo:
|
|
22 |
<p style="margin-bottom: 14px; font-size: 100%"> As AI-enabled Text-to-Image models are becoming increasingly used, characterizing the social biases they exhibit is a necessary first step to lowering their risk of discriminatory outcomes. <br> We compare three such models: <b> Stable Diffusion v.1.4, Stable Diffusion v.2. </b>, and <b> Dall-E 2 </b>, prompting them to produce images of different <i> professions </i> and <i> identity characteristics </i>. <br> Since artificial depictions of fictive humans have no inherent gender or ethnicity nor do they belong to socially-constructed groups, we pursued our analysis <i> without </i> ascribing gender and ethnicity categories to the images generated, still finding clear evidence of ethnicity and gender biases. You can explore these findings in the sections below: </p>
|
23 |
''')
|
24 |
|
25 |
-
|
|
|
|
|
26 |
|
27 |
with gr.Accordion("Identity group results (ethnicity and gender)", open=False):
|
28 |
gr.HTML('''
|
@@ -38,7 +40,10 @@ with gr.Blocks() as demo:
|
|
38 |
gr.HTML('''
|
39 |
<p style="margin-bottom: 14px; font-size: 100%"> You can see that the models reflect many societal biases -- for instance representing Native Americans wearing traditional headdresses, non-binary people with stereotypical haircuts and glasses, and East Asian men with features that amplify ethnic stereotypes. <br> This is problematic because it reinforces existing cultural stereotypes and fails to represent the diversity that is present in all identity groups.</p>
|
40 |
''')
|
41 |
-
|
|
|
|
|
|
|
42 |
|
43 |
with gr.Accordion("Bias Exploration", open=False):
|
44 |
gr.HTML('''
|
@@ -50,7 +55,7 @@ with gr.Blocks() as demo:
|
|
50 |
<p style="margin-bottom: 14px; font-size: 100%"> Using the <b><a href='https://huggingface.co/spaces/society-ethics/DiffusionClustering' style='text-decoration: underline;' target='_blank'> Diffusion Cluster Explorer</a></b>, we can see that the top cluster for the CEO and director professions is <b> Cluster 4</b>: </p> ''')
|
51 |
with gr.Column():
|
52 |
ceo_img = gr.Image(Image.open("images/bias/ceo_dir.png"), label = "CEO Image", show_label=False)
|
53 |
-
|
54 |
with gr.Row():
|
55 |
with gr.Column():
|
56 |
gr.HTML('''
|
@@ -70,6 +75,10 @@ with gr.Blocks() as demo:
|
|
70 |
with gr.Column(scale=2):
|
71 |
cluster4 = gr.Image(Image.open("images/bias/Cluster2.png"), label = "Cluster 2 Image", show_label=False)
|
72 |
|
|
|
|
|
|
|
|
|
73 |
with gr.Accordion("Comparing model generations", open=False):
|
74 |
gr.HTML('''
|
75 |
<p style="margin-bottom: 14px; font-size: 100%"> One of the goals of our study was allowing users to compare model generations across professions in an open-ended way, uncovering patterns and trends on their own. This is why we created the <a href='https://huggingface.co/spaces/society-ethics/DiffusionBiasExplorer' style='text-decoration: underline;' target='_blank'> Diffusion Bias Explorer </a> and the <a href='https://huggingface.co/spaces/society-ethics/Average_diffusion_faces' style='text-decoration: underline;' target='_blank'> Average Diffusion Faces </a> tools. We show some of their functionalities below: </p> ''')
|
@@ -90,6 +99,9 @@ with gr.Blocks() as demo:
|
|
90 |
gr.HTML('''
|
91 |
<p style="margin-bottom: 14px; font-size: 100%"> Looking at the average faces for a given profession across multiple models can help see the dominant characteristics of that profession, as well as how much variation there is (based on how fuzzy the image is). <br> In the images shown here, we can see that representations of these professions significantly differ across the three models, while sharing common characteristics, e.g. <i> postal workers </i> all wear caps. <br> Also, the average faces of <i> hairdressers </i> seem more fuzzy than the other professions, indicating a higher diversity among the generations compared to other professions. <br> Look at the <a href='https://huggingface.co/spaces/society-ethics/Average_diffusion_faces' style='text-decoration: underline;' target='_blank'> Average Diffusion Faces </a> tool for more examples! </p>''')
|
92 |
|
|
|
|
|
|
|
93 |
with gr.Accordion("Exploring the pixel space of generated images", open=False):
|
94 |
gr.HTML('''
|
95 |
<br>
|
@@ -104,7 +116,7 @@ with gr.Blocks() as demo:
|
|
104 |
### All of the tools created as part of this project:
|
105 |
""")
|
106 |
gr.HTML('''
|
107 |
-
<p style="margin-bottom: 10px; font-size:
|
108 |
<a href='https://huggingface.co/spaces/society-ethics/Average_diffusion_faces' style='text-decoration: underline;' target='_blank'> Average Diffusion Faces </a> <br>
|
109 |
<a href='https://huggingface.co/spaces/society-ethics/DiffusionBiasExplorer' style='text-decoration: underline;' target='_blank'> Diffusion Bias Explorer </a> <br>
|
110 |
<a href='https://huggingface.co/spaces/society-ethics/DiffusionClustering' style='text-decoration: underline;' target='_blank'> Diffusion Cluster Explorer </a> <br>
|
|
|
22 |
<p style="margin-bottom: 14px; font-size: 100%"> As AI-enabled Text-to-Image models are becoming increasingly used, characterizing the social biases they exhibit is a necessary first step to lowering their risk of discriminatory outcomes. <br> We compare three such models: <b> Stable Diffusion v.1.4, Stable Diffusion v.2. </b>, and <b> Dall-E 2 </b>, prompting them to produce images of different <i> professions </i> and <i> identity characteristics </i>. <br> Since artificial depictions of fictive humans have no inherent gender or ethnicity nor do they belong to socially-constructed groups, we pursued our analysis <i> without </i> ascribing gender and ethnicity categories to the images generated, still finding clear evidence of ethnicity and gender biases. You can explore these findings in the sections below: </p>
|
23 |
''')
|
24 |
|
25 |
+
gr.Markdown("""
|
26 |
+
## Identity group results (ethnicity and gender)
|
27 |
+
""")
|
28 |
|
29 |
with gr.Accordion("Identity group results (ethnicity and gender)", open=False):
|
30 |
gr.HTML('''
|
|
|
40 |
gr.HTML('''
|
41 |
<p style="margin-bottom: 14px; font-size: 100%"> You can see that the models reflect many societal biases -- for instance representing Native Americans wearing traditional headdresses, non-binary people with stereotypical haircuts and glasses, and East Asian men with features that amplify ethnic stereotypes. <br> This is problematic because it reinforces existing cultural stereotypes and fails to represent the diversity that is present in all identity groups.</p>
|
42 |
''')
|
43 |
+
gr.Markdown("""
|
44 |
+
## Bias Exploration
|
45 |
+
""")
|
46 |
+
|
47 |
|
48 |
with gr.Accordion("Bias Exploration", open=False):
|
49 |
gr.HTML('''
|
|
|
55 |
<p style="margin-bottom: 14px; font-size: 100%"> Using the <b><a href='https://huggingface.co/spaces/society-ethics/DiffusionClustering' style='text-decoration: underline;' target='_blank'> Diffusion Cluster Explorer</a></b>, we can see that the top cluster for the CEO and director professions is <b> Cluster 4</b>: </p> ''')
|
56 |
with gr.Column():
|
57 |
ceo_img = gr.Image(Image.open("images/bias/ceo_dir.png"), label = "CEO Image", show_label=False)
|
58 |
+
|
59 |
with gr.Row():
|
60 |
with gr.Column():
|
61 |
gr.HTML('''
|
|
|
75 |
with gr.Column(scale=2):
|
76 |
cluster4 = gr.Image(Image.open("images/bias/Cluster2.png"), label = "Cluster 2 Image", show_label=False)
|
77 |
|
78 |
+
gr.Markdown("""
|
79 |
+
## Comparing model generations
|
80 |
+
""")
|
81 |
+
|
82 |
with gr.Accordion("Comparing model generations", open=False):
|
83 |
gr.HTML('''
|
84 |
<p style="margin-bottom: 14px; font-size: 100%"> One of the goals of our study was allowing users to compare model generations across professions in an open-ended way, uncovering patterns and trends on their own. This is why we created the <a href='https://huggingface.co/spaces/society-ethics/DiffusionBiasExplorer' style='text-decoration: underline;' target='_blank'> Diffusion Bias Explorer </a> and the <a href='https://huggingface.co/spaces/society-ethics/Average_diffusion_faces' style='text-decoration: underline;' target='_blank'> Average Diffusion Faces </a> tools. We show some of their functionalities below: </p> ''')
|
|
|
99 |
gr.HTML('''
|
100 |
<p style="margin-bottom: 14px; font-size: 100%"> Looking at the average faces for a given profession across multiple models can help see the dominant characteristics of that profession, as well as how much variation there is (based on how fuzzy the image is). <br> In the images shown here, we can see that representations of these professions significantly differ across the three models, while sharing common characteristics, e.g. <i> postal workers </i> all wear caps. <br> Also, the average faces of <i> hairdressers </i> seem more fuzzy than the other professions, indicating a higher diversity among the generations compared to other professions. <br> Look at the <a href='https://huggingface.co/spaces/society-ethics/Average_diffusion_faces' style='text-decoration: underline;' target='_blank'> Average Diffusion Faces </a> tool for more examples! </p>''')
|
101 |
|
102 |
+
gr.Markdown("""
|
103 |
+
## Exploring the Pixel Space of Generated Images
|
104 |
+
""")
|
105 |
with gr.Accordion("Exploring the pixel space of generated images", open=False):
|
106 |
gr.HTML('''
|
107 |
<br>
|
|
|
116 |
### All of the tools created as part of this project:
|
117 |
""")
|
118 |
gr.HTML('''
|
119 |
+
<p style="margin-bottom: 10px; font-size: 100%">
|
120 |
<a href='https://huggingface.co/spaces/society-ethics/Average_diffusion_faces' style='text-decoration: underline;' target='_blank'> Average Diffusion Faces </a> <br>
|
121 |
<a href='https://huggingface.co/spaces/society-ethics/DiffusionBiasExplorer' style='text-decoration: underline;' target='_blank'> Diffusion Bias Explorer </a> <br>
|
122 |
<a href='https://huggingface.co/spaces/society-ethics/DiffusionClustering' style='text-decoration: underline;' target='_blank'> Diffusion Cluster Explorer </a> <br>
|