File size: 2,022 Bytes
d762240 0159a4b e8e7469 0159a4b b3c4a71 0704ca0 b3c4a71 e8e7469 0704ca0 b3c4a71 0704ca0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
---
license: apache-2.0
---
## Introduction
This is a fine-tuned LM in our papers below and the related GitHub repo is [here](https://github.com/metaphors/TibetanPLMsFineTuning).
***[Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model (Cao et al., WWW 2024 Workshop - SocialNLP)](https://dl.acm.org/doi/10.1145/3589335.3652503)***
***[Pay Attention to the Robustness of Chinese Minority Language Models! Syllable-level Textual Adversarial Attack on Tibetan Script (Cao et al., ACL 2023 Workshop - TrustNLP)](https://aclanthology.org/2023.trustnlp-1.4)***
## Citation
If you think our work useful, please kindly cite our paper.
```
@inproceedings{10.1145/3589335.3652503,
author = {Cao, Xi and Qun, Nuo and Gesang, Quzong and Zhu, Yulei and Nyima, Trashi},
title = {Multi-Granularity Tibetan Textual Adversarial Attack Method Based on Masked Language Model},
year = {2024},
isbn = {9798400701726},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3589335.3652503},
doi = {10.1145/3589335.3652503},
booktitle = {Companion Proceedings of the ACM on Web Conference 2024},
pages = {1672–1680},
numpages = {9},
keywords = {language model, robustness, textual adversarial attack, tibetan},
location = {Singapore, Singapore},
series = {WWW '24}
}
```
```
@inproceedings{cao-etal-2023-pay-attention,
title = "Pay Attention to the Robustness of {C}hinese Minority Language Models! Syllable-level Textual Adversarial Attack on {T}ibetan Script",
author = "Cao, Xi and
Dawa, Dolma and
Qun, Nuo and
Nyima, Trashi",
booktitle = "Proceedings of the 3rd Workshop on Trustworthy Natural Language Processing (TrustNLP 2023)",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.trustnlp-1.4",
pages = "35--46"
}
``` |