gregH commited on
Commit
7989976
·
1 Parent(s): dd5d10d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -1
README.md CHANGED
@@ -1,3 +1,50 @@
1
  ---
2
- license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ {}
3
  ---
4
+
5
+ <br>
6
+
7
+ # RADAR Model Card
8
+
9
+ ## Model Details
10
+
11
+ RADAR-Vicuna-7B is an AI-text detector trained via adversrial learning between the detector and a paraphraser on human-text corpus([OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext)) and ai-text corpus generated
12
+ from [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext)
13
+
14
+ - **Developed by:** [TrustSafeAI](https://huggingface.co/TrustSafeAI)
15
+ - **Model type:** An encoder-only language model based on the transformer architecture.
16
+ - **License:** Non-commercial license
17
+ - **Trained from model:** [RoBERTa](https://arxiv.org/abs/1907.11692).
18
+
19
+ ### Model Sources
20
+
21
+ - **Project Page:** https://radar.vizhub.ai/
22
+ - **Paper:** https://arxiv.org/abs/2307.03838
23
+ - **IBM Blog Post:** https://research.ibm.com/blog/AI-forensics-attribution
24
+
25
+ ## Uses
26
+ Users could use this detector to assist them in detecting text generated by large language models.
27
+ Please be noted that this detector is trained on ai-text generated by Vicuna-7B-v1.1. As the model only supports [non-commercial use](https://huggingface.co/lmsys/vicuna-7b-v1.1#model-details).
28
+ So the intended users are not allowed to involve this detector into commercial activities.
29
+
30
+ ## Get Started with the Model
31
+ Please refer the following google colab demo to see how to locally run the downloaded model or use our API service hosted on Huggingface Space.
32
+ - Google Colab Demo: https://colab.research.google.com/drive/1XmqDFsSpLJ67EGXUl9MpapeJ-XozRRpq#scrollTo=w8s2p5lwkShg
33
+ - Huggingface API Documentation: https://trustsafeai-radar-ai-text-detector.hf.space/?view=api
34
+
35
+ ## Training Pipeline
36
+
37
+ We propose adversarial learning between a paraphraser and our detector. The paraphraser's goal is to make the AI-generated text more like human-writen and the detector's goal is to
38
+ promote it's ability to distinguish the human-text and ai-text.
39
+
40
+ - **(Step 1) Training Data preparation**: Before training, we use Vicuna-7B to generate AI-text by performing text-completion based on the prefix span of human-text in [OpenWebText](https://huggingface.co/datasets/Skylion007/openwebtext).
41
+
42
+ - **(Step 2) Update the paraphraser** During training, the paraphraser will do paraphrasing on the AI-text generated in **Step 1**. And then colect the reward returned by the detector to update the parphraser using Proxy Proximal Optimization loss.
43
+
44
+ - **(Step 3) Update the detector** The detector is optimized using the logistic loss on the human-text,ai-text and paraphrased ai-text.
45
+
46
+ See more details in the Section 3 and 4 of this [paper](https://arxiv.org/pdf/2307.03838.pdf).
47
+
48
+ ## Ethical Considerations
49
+ we suggest users use our tool to assist with identifying AI-written content at scale and with discretion. If the detection result is to be used as evidence, further validation steps
50
+ are necessary as RADAR cannot always make correct predictions.