elucidator8918 commited on
Commit
79a4e71
·
verified ·
1 Parent(s): 34aec67

Push model using huggingface_hub.

Browse files
Files changed (3) hide show
  1. README.md +3 -203
  2. config.json +3 -44
  3. model.safetensors +1 -1
README.md CHANGED
@@ -1,209 +1,9 @@
1
  ---
2
- base_model: sentence-transformers-testing/stsb-bert-tiny-safetensors
3
- datasets:
4
- - QuotaClimat/frugalaichallenge-text-train
5
- language:
6
- - en
7
- license: apache-2.0
8
- model_name: frugal-ai-text-tiny-final
9
- pipeline_tag: text-classification
10
  tags:
11
  - model_hub_mixin
12
  - pytorch_model_hub_mixin
13
- - climate
14
  ---
15
 
16
- # Model Card for Model ID
17
-
18
- <!-- Provide a quick summary of what the model is/does. -->
19
-
20
- Classify text into 8 categories of climate misinformation with General Text Embedding Architecture.
21
-
22
- ## Model Details
23
-
24
- ### Model Description
25
-
26
- <!-- Provide a longer summary of what this model is. -->
27
-
28
- Fine trained BERT for classifying climate information as part of the Frugal AI Challenge, for submission to https://huggingface.co/frugal-ai-challenge and scoring on accuracy and efficiency. Trainied on only the non-evaluation 80% of the data, so it's (non-cheating) score will be lower.
29
-
30
- - **Developed by:** Andre Bach
31
- - **Funded by [optional]:** N/A
32
- - **Shared by [optional]:** Andre Bach
33
- - **Model type:** Text classification
34
- - **Language(s) (NLP):** ['en']
35
- - **License:** apache-2.0
36
- - **Finetuned from model [optional]:** sentence-transformers-testing/stsb-bert-tiny-safetensors
37
-
38
- ### Model Sources [optional]
39
-
40
- <!-- Provide the basic links for the model. -->
41
-
42
- - **Repository:** frugal-ai-text-tiny-final
43
- - **Paper [optional]:** [More Information Needed]
44
- - **Demo [optional]:** [More Information Needed]
45
-
46
- ## Uses
47
-
48
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
49
-
50
- ### Direct Use
51
-
52
- <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
53
-
54
- [More Information Needed]
55
-
56
- ### Downstream Use [optional]
57
-
58
- <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
59
-
60
- [More Information Needed]
61
-
62
- ### Out-of-Scope Use
63
-
64
- <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
65
-
66
- [More Information Needed]
67
-
68
- ## Bias, Risks, and Limitations
69
-
70
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
71
-
72
- [More Information Needed]
73
-
74
- ### Recommendations
75
-
76
- <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
77
-
78
- Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
79
-
80
- ## How to Get Started with the Model
81
-
82
- Use the code below to get started with the model.
83
-
84
- [More Information Needed]
85
-
86
- ## Training Details
87
-
88
- ### Training Data
89
-
90
- <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
91
-
92
- [More Information Needed]
93
-
94
- ### Training Procedure
95
-
96
- <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
97
-
98
- #### Preprocessing [optional]
99
-
100
- [More Information Needed]
101
-
102
-
103
- #### Training Hyperparameters
104
-
105
- - **Training regime:** {'max_dataset_size': 'full', 'model_path': 'sentence-transformers-testing/stsb-bert-tiny-safetensors', 'max_length': 256, 'num_epochs': 7, 'batch_size': 16} <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
106
-
107
- #### Speeds, Sizes, Times [optional]
108
-
109
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
110
-
111
- [More Information Needed]
112
-
113
- ## Evaluation
114
-
115
- <!-- This section describes the evaluation protocols and provides the results. -->
116
-
117
- ### Testing Data, Factors & Metrics
118
-
119
- #### Testing Data
120
-
121
- <!-- This should link to a Dataset Card if possible. -->
122
-
123
- [More Information Needed]
124
-
125
- #### Factors
126
-
127
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
128
-
129
- [More Information Needed]
130
-
131
- #### Metrics
132
-
133
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
134
-
135
- {'train_loss': 0.9096412720250302, 'train_acc': 0.6826765188834154, 'test_loss': 1.1881297435853388, 'test_acc': 0.5783429040196882}
136
-
137
- ### Results
138
-
139
- [More Information Needed]
140
-
141
- #### Summary
142
-
143
-
144
-
145
- ## Model Examination [optional]
146
-
147
- <!-- Relevant interpretability work for the model goes here -->
148
-
149
- [More Information Needed]
150
-
151
- ## Environmental Impact
152
-
153
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
154
-
155
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
156
-
157
- - **Hardware Type:** [More Information Needed]
158
- - **Hours used:** [More Information Needed]
159
- - **Cloud Provider:** [More Information Needed]
160
- - **Compute Region:** [More Information Needed]
161
- - **Carbon Emitted:** [More Information Needed]
162
-
163
- ## Technical Specifications [optional]
164
-
165
- ### Model Architecture and Objective
166
-
167
- [More Information Needed]
168
-
169
- ### Compute Infrastructure
170
-
171
- [More Information Needed]
172
-
173
- #### Hardware
174
-
175
- [More Information Needed]
176
-
177
- #### Software
178
-
179
- [More Information Needed]
180
-
181
- ## Citation [optional]
182
-
183
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
184
-
185
- **BibTeX:**
186
-
187
- [More Information Needed]
188
-
189
- **APA:**
190
-
191
- [More Information Needed]
192
-
193
- ## Glossary [optional]
194
-
195
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
196
-
197
- [More Information Needed]
198
-
199
- ## More Information [optional]
200
-
201
- [More Information Needed]
202
-
203
- ## Model Card Authors [optional]
204
-
205
- [More Information Needed]
206
-
207
- ## Model Card Contact
208
-
209
- [More Information Needed]
 
1
  ---
 
 
 
 
 
 
 
 
2
  tags:
3
  - model_hub_mixin
4
  - pytorch_model_hub_mixin
 
5
  ---
6
 
7
+ This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
8
+ - Library: [More Information Needed]
9
+ - Docs: [More Information Needed]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
config.json CHANGED
@@ -1,45 +1,4 @@
1
  {
2
- "_name_or_path": "sentence-transformers-testing/stsb-bert-tiny-safetensors",
3
- "architectures": [
4
- "BertModel"
5
- ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
- "hidden_act": "gelu",
9
- "hidden_dropout_prob": 0.1,
10
- "hidden_size": 128,
11
- "id2label": {
12
- "0": "LABEL_0",
13
- "1": "LABEL_1",
14
- "2": "LABEL_2",
15
- "3": "LABEL_3",
16
- "4": "LABEL_4",
17
- "5": "LABEL_5",
18
- "6": "LABEL_6",
19
- "7": "LABEL_7"
20
- },
21
- "initializer_range": 0.02,
22
- "intermediate_size": 512,
23
- "label2id": {
24
- "LABEL_0": 0,
25
- "LABEL_1": 1,
26
- "LABEL_2": 2,
27
- "LABEL_3": 3,
28
- "LABEL_4": 4,
29
- "LABEL_5": 5,
30
- "LABEL_6": 6,
31
- "LABEL_7": 7
32
- },
33
- "layer_norm_eps": 1e-12,
34
- "max_position_embeddings": 512,
35
- "model_type": "bert",
36
- "num_attention_heads": 2,
37
- "num_hidden_layers": 2,
38
- "pad_token_id": 0,
39
- "position_embedding_type": "absolute",
40
- "torch_dtype": "float32",
41
- "transformers_version": "4.48.1",
42
- "type_vocab_size": 2,
43
- "use_cache": true,
44
- "vocab_size": 30522
45
- }
 
1
  {
2
+ "model_path": "prajjwal1/bert-tiny",
3
+ "num_labels": 8
4
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e5a6cee834b9f750f427a240974241f2f5edbe86181f6f274fc56251050779a
3
  size 17552376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66bfd74409fa519f38ef3d5132e6978382aa949e2013822c57888341d877dbfb
3
  size 17552376