julien-c HF staff commited on
Commit
780b585
1 Parent(s): 96f17bd

Migrate model card from transformers-repo

Browse files

Read announcement at https://discuss.huggingface.co/t/announcement-all-model-cards-will-be-migrated-to-hf-co-model-repos/2755
Original file history: https://github.com/huggingface/transformers/commits/master/model_cards/krevas/finance-koelectra-small-generator/README.md

Files changed (1) hide show
  1. README.md +41 -0
README.md ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: ko
3
+ ---
4
+
5
+ # 📈 Financial Korean ELECTRA model
6
+
7
+ Pretrained ELECTRA Language Model for Korean (`finance-koelectra-small-generator`)
8
+
9
+ > ELECTRA is a new method for self-supervised language representation learning. It can be used to
10
+ > pre-train transformer networks using relatively little compute. ELECTRA models are trained to
11
+ > distinguish "real" input tokens vs "fake" input tokens generated by another neural network, similar to
12
+ > the discriminator of a GAN.
13
+
14
+ More details about ELECTRA can be found in the [ICLR paper](https://openreview.net/forum?id=r1xMH1BtvB)
15
+ or in the [official ELECTRA repository](https://github.com/google-research/electra) on GitHub.
16
+
17
+ ## Stats
18
+
19
+ The current version of the model is trained on a financial news data of Naver news.
20
+
21
+ The final training corpus has a size of 25GB and 2.3B tokens.
22
+
23
+ This model was trained a cased model on a TITAN RTX for 500k steps.
24
+
25
+ ## Usage
26
+
27
+ ```python
28
+ from transformers import pipeline
29
+
30
+ fill_mask = pipeline(
31
+ "fill-mask",
32
+ model="krevas/finance-koelectra-small-generator",
33
+ tokenizer="krevas/finance-koelectra-small-generator"
34
+ )
35
+
36
+ print(fill_mask(f"내일 해당 종목이 대폭 {fill_mask.tokenizer.mask_token}할 것이다."))
37
+ ```
38
+
39
+ # Huggingface model hub
40
+
41
+ All models are available on the [Huggingface model hub](https://huggingface.co/krevas).