File size: 4,857 Bytes
6009572 992c10c 6009572 992c10c 6009572 992c10c 4d89098 6009572 992c10c 6009572 992c10c 96bf3a1 4d89098 992c10c 6009572 992c10c 6009572 d15ea2f 6009572 d15ea2f 6009572 d15ea2f 6009572 3b56755 2ad8f6f 3b56755 5be28ba 1d5b781 cb12a9e 5be28ba 1d5b781 cb12a9e 1d5b781 3b56755 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 |
from dataclasses import dataclass
from enum import Enum
@dataclass
class Task:
benchmark: str
metric: str
col_name: str
# Select your tasks here
# ---------------------------------------------------
class Tasks(Enum):
# task_key in the json file, metric_key in the json file, name to display in the leaderboard
task0 = Task("anli_r1", "acc", "ANLI")
task1 = Task("logiqa", "acc_norm", "LogiQA")
NUM_FEWSHOT = 0 # Change with your few shot
# ---------------------------------------------------
# Your leaderboard name
TITLE = """<h1 align="center" id="space-title">UnlearnDiffAtk: Unlearned Diffusion Model Benchmark</h1>"""
# subtitle
SUB_TITLE = """<h2 align="center" id="space-title">Effective and efficient adversarial prompt generation approach for unlearned diffusion model evaluations.</h2>"""
# What does your leaderboard evaluate?
INTRODUCTION_TEXT = """
This benchmark evaluates the <strong>robustness and utility retaining</strong> of safety-driven unlearned diffusion models (DMs) across a variety of tasks. For more details, please visit the [project](https://www.optml-group.com/posts/mu_attack).
- The <strong>robustness</strong> of unlearned DM is evaluated through our proposed adversarial prompt attack, [UnlearnDiffAtk](https://github.com/OPTML-Group/Diffusion-MU-Attack), which has been accepted to ECCV 2024.
- The <strong>utility retaining</strong> of unlearned DM is evaluated through FID and CLIP score on the generated images using [10K randomly sampled COCO caption prompts](https://github.com/OPTML-Group/Diffusion-MU-Attack/blob/main/prompts/coco_10k.csv).
Demo of our offensive method: [UnlearnDiffAtk](https://huggingface.co/spaces/Intel/UnlearnDiffAtk)\\
Demo of our defensive method: [AdvUnlearn](https://huggingface.co/spaces/Intel/AdvUnlearn)
"""
EVALUATION_QUEUE_TEXT = """
<strong>\[Evaluation Metrics\]</strong>:
- Pre-Attack Success Rate (<strong>Pre-ASR</strong>): lower is better;
- Post-attack success rate (<strong>Post-ASR</strong>): lower is better;
- Fréchet inception distance(<strong>FID</strong>): evaluate distributional quality of image generations, lower is better;
- <strong>CLIP Score</strong>: measure contextual alignment with prompt descriptions, higher is better.
<strong>\[DM Unlearning Tasks\]</strong>:
- NSFW: Nudity
- Style: Van Gogh
- Objects: Church, Tench, Parachute, Garbage Truck
"""
# Which evaluations are you running? how can people reproduce what you have?
LLM_BENCHMARKS_TEXT = f"""
For more details of Unlearning Methods used in this benchmarks:
- [Adversarial Unlearning (AdvUnlearn)](https://github.com/OPTML-Group/AdvUnlearn);
- [Erased Stable Diffusion (ESD)](https://github.com/rohitgandikota/erasing);
- [Forget-Me-Not (FMN)](https://github.com/SHI-Labs/Forget-Me-Not);
- [Ablating Concepts (AC)](https://github.com/nupurkmr9/concept-ablation);
- [Unified Concept Editing (UCE)](https://github.com/rohitgandikota/unified-concept-editing);
- [concept-SemiPermeable Membrane (SPM)](https://github.com/Con6924/SPM);
- [Saliency Unlearning (SalUn)](https://github.com/OPTML-Group/Unlearn-Saliency);
- [EraseDiff (ED)](https://github.com/JingWu321/EraseDiff);
- [ScissorHands (SH)](https://github.com/JingWu321/Scissorhands);
- [Mass Concept Erasure (MACE)](https://github.com/Shilin-LU/MACE);
- [Reliable and Efficient Concept Erasure (RECE)](https://github.com/CharlesGong12/RECE);
- [Training-Free and Adaptive Guard (SAFREE)](https://github.com/jaehong31/SAFREE).
<strong>We will evaluate your model on UnlearnDiffAtk Benchmark!</strong> \\
Open a [github issue](https://github.com/OPTML-Group/Diffusion-MU-Attack/issues) or email us at zhan1853@msu.edu!
"""
CITATION_BUTTON_LABEL = "Copy the following snippet to cite these results"
CITATION_BUTTON_TEXT = r"""
@inproceedings{zhang2023generate,
title={To Generate or Not? Safety-Driven Unlearned Diffusion Models Are Still Easy To Generate Unsafe Images... For Now},
author={Zhang, Yimeng and Jia, Jinghan and Chen, Xin and Chen, Aochuan and Zhang, Yihua and Liu, Jiancheng and Ding, Ke and Liu, Sijia},
booktitle={European Conference on Computer Vision},
year={2024}
}
@article{zhang2024defensive,
title={Defensive Unlearning with Adversarial Training for Robust Concept Erasure in Diffusion Models},
author={Zhang, Yimeng and Chen, Xin and Jia, Jinghan and Zhang, Yihua and Fan, Chongyu and Liu, Jiancheng and Hong, Mingyi and Ding, Ke and Liu, Sijia},
journal={arXiv preprint arXiv:2405.15234},
year={2024}
}
"""
CONTRIBUTOR_BUTTON_LABEL = "Contributors are listed as follows:"
CONTRIBUTOR_BUTTON_TEXT = f"""
• OPTML @ Michigan State University:
Sijia Liu, Yimeng Zhang, JInghan Jia, Aochuan Chen, Yihua Zhang, Jiacheng Liu
• SCAI @ Arizona State University:
Maitreya Patel, Abhiram Kusumba
• Intel Corp:
Kyle Min, Ke Ding, Xin Chen
"""
|