Update README.md
Browse files
README.md
CHANGED
@@ -8,13 +8,14 @@ pipeline_tag: text-classification
|
|
8 |
|
9 |
## Model Details
|
10 |
|
11 |
-
RADAR-Vicuna-7B is an AI-text detector trained via
|
12 |
based on [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
|
13 |
|
14 |
- **Developed by:** [TrustSafeAI](https://huggingface.co/TrustSafeAI)
|
15 |
-
- **Model type:** An encoder-only language model based on the transformer architecture.
|
16 |
-
- **License:** Non-commercial license
|
17 |
-
- **Trained from model:** [RoBERTa](https://arxiv.org/abs/1907.11692)
|
|
|
18 |
|
19 |
### Model Sources
|
20 |
|
@@ -24,7 +25,7 @@ based on [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
|
|
24 |
|
25 |
## Uses
|
26 |
Users could use this detector to assist them in detecting text generated by large language models.
|
27 |
-
Please
|
28 |
|
29 |
## Get Started with the Model
|
30 |
Please refer to the following guidelines to see how to locally run the downloaded model or use our API service hosted on Huggingface Space.
|
@@ -34,15 +35,15 @@ Please refer to the following guidelines to see how to locally run the downloade
|
|
34 |
## Training Pipeline
|
35 |
|
36 |
We propose adversarial learning between a paraphraser and our detector. The paraphraser's goal is to make the AI-generated text more like human-writen and the detector's goal is to
|
37 |
-
promote it's ability to identify the
|
38 |
|
39 |
-
- **(Step 1) Training Data preparation**: Before training, we use Vicuna-7B to generate AI-text by performing text
|
40 |
|
41 |
-
- **(Step 2) Update the paraphraser** During training, the paraphraser will do paraphrasing on the AI-text generated in **Step 1**. And then collect the reward returned by the detector to update the
|
42 |
|
43 |
-
- **(Step 3) Update the detector** The detector is optimized using the logistic loss on the human-text,
|
44 |
|
45 |
-
See more details in
|
46 |
|
47 |
## Ethical Considerations
|
48 |
We suggest users use our tool to assist with identifying AI-written content at scale and with discretion. If the detection result is to be used as evidence, further validation steps
|
|
|
8 |
|
9 |
## Model Details
|
10 |
|
11 |
+
RADAR-Vicuna-7B is an AI-text detector trained via adversarial learning between the detector and a paraphraser on human-text corpus ([OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext)) and AI-text corpus generated
|
12 |
based on [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
|
13 |
|
14 |
- **Developed by:** [TrustSafeAI](https://huggingface.co/TrustSafeAI)
|
15 |
+
- **Model type:** An encoder-only language model based on the transformer architecture (RoBERTa).
|
16 |
+
- **License:** [Non-commercial license](https://huggingface.co/lmsys/vicuna-7b-v1.1#model-details) (inherited from Vicuna-7B-v1.1)
|
17 |
+
- **Trained from model:** [RoBERTa](https://arxiv.org/abs/1907.11692)
|
18 |
+
|
19 |
|
20 |
### Model Sources
|
21 |
|
|
|
25 |
|
26 |
## Uses
|
27 |
Users could use this detector to assist them in detecting text generated by large language models.
|
28 |
+
Please note that this detector is trained on AI-text generated by Vicuna-7B-v1.1. As the model only supports [non-commercial use](https://huggingface.co/lmsys/vicuna-7b-v1.1#model-details), the intended users are **not allowed to involve this detector into commercial activities**.
|
29 |
|
30 |
## Get Started with the Model
|
31 |
Please refer to the following guidelines to see how to locally run the downloaded model or use our API service hosted on Huggingface Space.
|
|
|
35 |
## Training Pipeline
|
36 |
|
37 |
We propose adversarial learning between a paraphraser and our detector. The paraphraser's goal is to make the AI-generated text more like human-writen and the detector's goal is to
|
38 |
+
promote it's ability to identify the AI-text.
|
39 |
|
40 |
+
- **(Step 1) Training Data preparation**: Before training, we use Vicuna-7B to generate AI-text by performing text completion based on the prefix span of human-text in [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
|
41 |
|
42 |
+
- **(Step 2) Update the paraphraser** During training, the paraphraser will do paraphrasing on the AI-text generated in **Step 1**. And then collect the reward returned by the detector to update the paraphraser using Proxy Proximal Optimization loss.
|
43 |
|
44 |
+
- **(Step 3) Update the detector** The detector is optimized using the logistic loss on the human-text, AI-text and paraphrased AI-text.
|
45 |
|
46 |
+
See more details in Sections 3 and 4 of this [paper](https://arxiv.org/pdf/2307.03838.pdf).
|
47 |
|
48 |
## Ethical Considerations
|
49 |
We suggest users use our tool to assist with identifying AI-written content at scale and with discretion. If the detection result is to be used as evidence, further validation steps
|