Lynxpda commited on
Commit
44dd7a4
1 Parent(s): 53e9f36

Upload folder using huggingface_hub

Browse files
2_Dense/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e8a24e740662ddf8ec666af2558717744596d1fe23fdb76beca28210e9048038
3
  size 2364028
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:343f53d79815d1296ed1b563314badea919b8e9eaac97f1cda7df6d563da61fa
3
  size 2364028
README.md CHANGED
@@ -5,12 +5,10 @@ tags:
5
  - sentence-transformers
6
  - feature-extraction
7
  - sentence-similarity
8
- license: apache-2.0
9
- datasets:
10
- - Lynxpda/wiki-bt-veps-russian
11
  ---
12
 
13
- # LaBSE-veps
14
 
15
  This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
16
 
@@ -49,7 +47,7 @@ The model was trained with the parameters:
49
 
50
  **DataLoader**:
51
 
52
- `torch.utils.data.dataloader.DataLoader` of length 8158 with parameters:
53
  ```
54
  {'batch_size': 8, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
55
  ```
@@ -64,7 +62,7 @@ The model was trained with the parameters:
64
  Parameters of the fit()-Method:
65
  ```
66
  {
67
- "epochs": 1,
68
  "evaluation_steps": 100,
69
  "evaluator": "__main__.ChainScoreEvaluator",
70
  "max_grad_norm": 1,
@@ -74,7 +72,7 @@ Parameters of the fit()-Method:
74
  },
75
  "scheduler": "warmupcosine",
76
  "steps_per_epoch": null,
77
- "warmup_steps": 1000,
78
  "weight_decay": 0.01
79
  }
80
  ```
 
5
  - sentence-transformers
6
  - feature-extraction
7
  - sentence-similarity
8
+
 
 
9
  ---
10
 
11
+ # {MODEL_NAME}
12
 
13
  This is a [sentence-transformers](https://www.SBERT.net) model: It maps sentences & paragraphs to a 768 dimensional dense vector space and can be used for tasks like clustering or semantic search.
14
 
 
47
 
48
  **DataLoader**:
49
 
50
+ `torch.utils.data.dataloader.DataLoader` of length 334 with parameters:
51
  ```
52
  {'batch_size': 8, 'sampler': 'torch.utils.data.sampler.RandomSampler', 'batch_sampler': 'torch.utils.data.sampler.BatchSampler'}
53
  ```
 
62
  Parameters of the fit()-Method:
63
  ```
64
  {
65
+ "epochs": 5,
66
  "evaluation_steps": 100,
67
  "evaluator": "__main__.ChainScoreEvaluator",
68
  "max_grad_norm": 1,
 
72
  },
73
  "scheduler": "warmupcosine",
74
  "steps_per_epoch": null,
75
+ "warmup_steps": 500,
76
  "weight_decay": 0.01
77
  }
78
  ```
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "sentence-transformers/LaBSE",
3
  "architectures": [
4
  "BertModel"
5
  ],
 
1
  {
2
+ "_name_or_path": "Lynxpda/LaBSE-veps",
3
  "architectures": [
4
  "BertModel"
5
  ],
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f4519c9e5844fb9ebe636bc96ba56d83bb8f178a93d38fe0cdb2b47380b49a6
3
  size 1883730160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abba0b3a86565dc1f2ac9763e3fb4da6762ae12dbc675a81dd810ea0df47866c
3
  size 1883730160
tokenizer_config.json CHANGED
@@ -47,12 +47,19 @@
47
  "do_lower_case": false,
48
  "full_tokenizer_file": null,
49
  "mask_token": "[MASK]",
 
50
  "model_max_length": 512,
51
  "never_split": null,
 
52
  "pad_token": "[PAD]",
 
 
53
  "sep_token": "[SEP]",
 
54
  "strip_accents": null,
55
  "tokenize_chinese_chars": true,
56
  "tokenizer_class": "BertTokenizer",
 
 
57
  "unk_token": "[UNK]"
58
  }
 
47
  "do_lower_case": false,
48
  "full_tokenizer_file": null,
49
  "mask_token": "[MASK]",
50
+ "max_length": 256,
51
  "model_max_length": 512,
52
  "never_split": null,
53
+ "pad_to_multiple_of": null,
54
  "pad_token": "[PAD]",
55
+ "pad_token_type_id": 0,
56
+ "padding_side": "right",
57
  "sep_token": "[SEP]",
58
+ "stride": 0,
59
  "strip_accents": null,
60
  "tokenize_chinese_chars": true,
61
  "tokenizer_class": "BertTokenizer",
62
+ "truncation_side": "right",
63
+ "truncation_strategy": "longest_first",
64
  "unk_token": "[UNK]"
65
  }