|
--- |
|
language: "zh" |
|
tags: |
|
- agriculture-domain |
|
- agriculture |
|
widget: |
|
- text: "[MASK]是許多亞洲國家的主要糧食作物。" |
|
--- |
|
# agriculture-bert-base-chinese |
|
|
|
This is a bert model for agriculture domain. The self-supervised learning approach of MLM was used to train the model. |
|
- Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. |
|
- This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT internally masks the future tokens. |
|
- It allows the model to learn a bidirectional representation of the sentence. |
|
|
|
```python |
|
from transformers import pipeline |
|
fill_mask = pipeline( |
|
"fill-mask", |
|
model="gigilin7/agriculture-bert-base-chinese", |
|
tokenizer="gigilin7/agriculture-bert-base-chinese" |
|
) |
|
res = fill_mask("[MASK]是許多亞洲國家的主要糧食作物。") |
|
``` |