“Harm and Bias Evaluation and Solution for Adobe Firefly,” a Presentation from Adobe

embeddedvision 61 views 20 slides Sep 06, 2024
Slide 1
Slide 1 of 20
Slide 1
1
Slide 2
2
Slide 3
3
Slide 4
4
Slide 5
5
Slide 6
6
Slide 7
7
Slide 8
8
Slide 9
9
Slide 10
10
Slide 11
11
Slide 12
12
Slide 13
13
Slide 14
14
Slide 15
15
Slide 16
16
Slide 17
17
Slide 18
18
Slide 19
19
Slide 20
20

About This Presentation

For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/09/harm-and-bias-evaluation-and-solution-for-adobe-firefly-a-presentation-from-adobe/

Rebecca Li, Machine Learning Engineering Manager at Adobe, presents the “Harm and Bias Evaluation and Solution for Adobe...


Slide Content

Harm and Bias Evaluation
and Solution for Adobe
Firefly
Dr. Xiaoyang (Rebecca) LI
ML Engineer Manager, Firefly Eval Science
Adobe

Harm & Bias
2© 2024 Adobe
Unaddressed Harm & Bias issues can stop models & products
Microsoft Tay, 2016 Meta’s Galactica, 2022

Mitigate Harmful Bias & Unsafe Content
Results
from
Adobe
Firefly
?
What is Human Bias?

There are many types of representational harms
Footer Goes Here 4
•Missing representations (no result)
•Mislabeled identities / Inaccurate depictions
•Stereotyping
•Over and under representation
•Dehumanization
•Cultural/religious insensitivities

Adobe Solution: AI Ethics Program
5© 2024 Adobe

Design
6© 2024 Adobe
Before launch, AI ethics committee reviews the
scientific evaluation of the possible intentional &
unintentional generation of:
•Nudity (adult & child)
•Ethnicity & identity (mis)representation
•Violence & gore
•Hate content (signs, hand gestures, etc.)
Interdisciplinary team: go/no-go decision

Goal:
•Mitigate bias in AI training sets and commit to preventing the
perpetuation of stereotypes and harm in model output
Approach: Data filtering & enrichment
•Human and automatic moderation removes harmful content
•Smart algorithms enhance data quality with limited resources.
•Data acquisition to close data gaps
Development: Model Training
7© 2024 Adobe

Goal:
•Testing of Firefly features and product to mitigate against harmful
biases and stereotypes
Approach:
•Automated testing(ML metrics & models)
•Human Evaluation (semi-automatic pipeline)
Development: Model Testing and Evaluation
8© 2024 Adobe

Case Study: Quantifying Harm Types for Text to Image
9

Case Study: Quantifying Likelihood of Nudity for GenFill
10© 2024 Adobe

Case Study: Evaluating Bias for Generated Humans
11© 2024 Adobe

Case Study: Evaluating Bias for Generated Humans
12© 2024 Adobe

Result Analysis: Evaluating Bias for Skin Color Diversity
Image Model B
Image Model A
Skin Color

Result Analysis: Evaluating Bias for Gender Diversity
Image Model B
Image Model A
Gender

Result Analysis: Evaluating Bias for Age Diversity
Age
Image Model B
Image Model A

Harm:
•Can’t load harm prompt
•Strict control on unintentional harm
Bias
•Increasing variety during prompt inference
•Customized debiasing strategies for different regions
Deployment
16© 2024 Adobe

Feedback by Users
17© 2024 Adobe
harmful and biased content feedback on output quality

Address potential harm & bias issues online/in real-time
•Requires guardrails in pre- & post-processing
•Mixture of approaches, interdisciplinary team
Proactive & reactive mitigation
•Source content to improve human representation diversity
•Red teaming, social media watch
Harm & Bias Mitigation Strategy
18© 2024 Adobe

Adobe Firefly
https://firefly.adobe.com/
Adobe AI Ethic
https://www.adobe.com/ai/overview/ethics.html
Reducing biased and harmful outcomes in generative AI
https://adobe.design/stories/leading-design/reducing-biased-and-harmful-outcomes-
in-generative-ai
Resource
19© 2024 Adobe

Questions?
20© 2024 Adobe