Training in progress, epoch 0
Browse files- README.md +56 -5
- config.json +14 -77
- model.safetensors +2 -2
- preprocessor_config.json +16 -0
- runs/Aug12_15-39-39_surveillance/events.out.tfevents.1723457391.surveillance.95468.0 +3 -0
- training_args.bin +2 -2
README.md
CHANGED
@@ -3,6 +3,8 @@ license: cc-by-nc-4.0
|
|
3 |
base_model: MCG-NJU/videomae-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
|
|
|
|
6 |
model-index:
|
7 |
- name: videomae-base-finetuned-subset
|
8 |
results: []
|
@@ -14,6 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
|
|
14 |
# videomae-base-finetuned-subset
|
15 |
|
16 |
This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
|
|
|
|
|
|
|
17 |
|
18 |
## Model description
|
19 |
|
@@ -33,17 +38,63 @@ More information needed
|
|
33 |
|
34 |
The following hyperparameters were used during training:
|
35 |
- learning_rate: 5e-05
|
36 |
-
- train_batch_size:
|
37 |
-
- eval_batch_size:
|
38 |
- seed: 42
|
39 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
40 |
- lr_scheduler_type: linear
|
41 |
- lr_scheduler_warmup_ratio: 0.1
|
42 |
-
- training_steps:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
|
44 |
### Framework versions
|
45 |
|
46 |
-
- Transformers 4.
|
47 |
-
- Pytorch 2.
|
48 |
- Datasets 2.20.0
|
49 |
- Tokenizers 0.19.1
|
|
|
3 |
base_model: MCG-NJU/videomae-base
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
+
metrics:
|
7 |
+
- accuracy
|
8 |
model-index:
|
9 |
- name: videomae-base-finetuned-subset
|
10 |
results: []
|
|
|
16 |
# videomae-base-finetuned-subset
|
17 |
|
18 |
This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
|
19 |
+
It achieves the following results on the evaluation set:
|
20 |
+
- Loss: 1.9422
|
21 |
+
- Accuracy: 0.7011
|
22 |
|
23 |
## Model description
|
24 |
|
|
|
38 |
|
39 |
The following hyperparameters were used during training:
|
40 |
- learning_rate: 5e-05
|
41 |
+
- train_batch_size: 2
|
42 |
+
- eval_batch_size: 2
|
43 |
- seed: 42
|
44 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
45 |
- lr_scheduler_type: linear
|
46 |
- lr_scheduler_warmup_ratio: 0.1
|
47 |
+
- training_steps: 2640
|
48 |
+
|
49 |
+
### Training results
|
50 |
+
|
51 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
52 |
+
|:-------------:|:-------:|:----:|:---------------:|:--------:|
|
53 |
+
| 1.9136 | 0.0254 | 67 | 2.1529 | 0.1096 |
|
54 |
+
| 1.9987 | 1.0254 | 134 | 1.9985 | 0.2593 |
|
55 |
+
| 1.0754 | 2.0254 | 201 | 2.1637 | 0.1956 |
|
56 |
+
| 1.413 | 3.0254 | 268 | 1.9523 | 0.1473 |
|
57 |
+
| 1.3047 | 4.0254 | 335 | 2.5442 | 0.1956 |
|
58 |
+
| 1.2488 | 5.0254 | 402 | 2.0563 | 0.2330 |
|
59 |
+
| 1.0581 | 6.0254 | 469 | 1.9954 | 0.2264 |
|
60 |
+
| 0.9165 | 7.0254 | 536 | 1.7661 | 0.2769 |
|
61 |
+
| 1.5722 | 8.0254 | 603 | 2.2872 | 0.2264 |
|
62 |
+
| 0.9083 | 9.0254 | 670 | 2.1004 | 0.3363 |
|
63 |
+
| 0.8093 | 10.0254 | 737 | 1.2497 | 0.6549 |
|
64 |
+
| 0.4925 | 11.0254 | 804 | 3.5720 | 0.2813 |
|
65 |
+
| 0.4573 | 12.0254 | 871 | 1.5213 | 0.3604 |
|
66 |
+
| 1.1082 | 13.0254 | 938 | 1.5453 | 0.5934 |
|
67 |
+
| 0.8066 | 14.0254 | 1005 | 2.9169 | 0.2967 |
|
68 |
+
| 0.6615 | 15.0254 | 1072 | 2.1412 | 0.5780 |
|
69 |
+
| 0.146 | 16.0254 | 1139 | 2.5006 | 0.3978 |
|
70 |
+
| 0.3815 | 17.0254 | 1206 | 1.7907 | 0.5956 |
|
71 |
+
| 0.2124 | 18.0254 | 1273 | 1.6622 | 0.6527 |
|
72 |
+
| 0.5304 | 19.0254 | 1340 | 1.8988 | 0.5956 |
|
73 |
+
| 0.1519 | 20.0254 | 1407 | 2.7940 | 0.3934 |
|
74 |
+
| 0.486 | 21.0254 | 1474 | 2.6766 | 0.4198 |
|
75 |
+
| 0.5502 | 22.0254 | 1541 | 2.3451 | 0.5495 |
|
76 |
+
| 0.7527 | 23.0254 | 1608 | 1.7518 | 0.6462 |
|
77 |
+
| 0.3194 | 24.0254 | 1675 | 2.0738 | 0.5890 |
|
78 |
+
| 0.0189 | 25.0254 | 1742 | 2.9264 | 0.5407 |
|
79 |
+
| 0.2928 | 26.0254 | 1809 | 2.5495 | 0.5451 |
|
80 |
+
| 0.0036 | 27.0254 | 1876 | 1.8143 | 0.6989 |
|
81 |
+
| 0.3772 | 28.0254 | 1943 | 2.2384 | 0.6088 |
|
82 |
+
| 0.0044 | 29.0254 | 2010 | 1.7688 | 0.7033 |
|
83 |
+
| 0.7291 | 30.0254 | 2077 | 2.0591 | 0.6571 |
|
84 |
+
| 0.1553 | 31.0254 | 2144 | 2.0690 | 0.6505 |
|
85 |
+
| 0.5454 | 32.0254 | 2211 | 1.8762 | 0.7055 |
|
86 |
+
| 0.3096 | 33.0254 | 2278 | 2.2310 | 0.6440 |
|
87 |
+
| 0.0053 | 34.0254 | 2345 | 2.0907 | 0.6615 |
|
88 |
+
| 0.0024 | 35.0254 | 2412 | 2.4127 | 0.6022 |
|
89 |
+
| 0.0022 | 36.0254 | 2479 | 2.0037 | 0.6989 |
|
90 |
+
| 0.0026 | 37.0254 | 2546 | 2.0130 | 0.6725 |
|
91 |
+
| 0.0013 | 38.0254 | 2613 | 1.9391 | 0.6967 |
|
92 |
+
| 0.0017 | 39.0102 | 2640 | 1.9422 | 0.7011 |
|
93 |
+
|
94 |
|
95 |
### Framework versions
|
96 |
|
97 |
+
- Transformers 4.40.2
|
98 |
+
- Pytorch 2.0.1+cu117
|
99 |
- Datasets 2.20.0
|
100 |
- Tokenizers 0.19.1
|
config.json
CHANGED
@@ -12,87 +12,23 @@
|
|
12 |
"hidden_dropout_prob": 0.0,
|
13 |
"hidden_size": 768,
|
14 |
"id2label": {
|
15 |
-
"0": "
|
16 |
-
"1": "
|
17 |
-
"2": "
|
18 |
-
"3": "
|
19 |
-
"4": "
|
20 |
-
"5": "
|
21 |
-
"6": "08_EMPLOYEES_IN_GROUP",
|
22 |
-
"7": "09_EMPLOYEE_SITTING",
|
23 |
-
"8": "10_EMPLOYEE_SLEEPING",
|
24 |
-
"9": "11_EMPLOYEE_USING_LIFT",
|
25 |
-
"10": "12_EMPLOYEE_USING_MOBILE",
|
26 |
-
"11": "13_FIGHTING",
|
27 |
-
"12": "14_FILLING_FEEDBACK",
|
28 |
-
"13": "15_INCHARGE_IN_COUNTER",
|
29 |
-
"14": "16_LOADING",
|
30 |
-
"15": "17_MOPPING_FLOOR",
|
31 |
-
"16": "18_NO_ONE_IN_COUNTER",
|
32 |
-
"17": "19_OTHER_EMPLOYEES_DOING_BILLING",
|
33 |
-
"18": "20_RUNNING",
|
34 |
-
"19": "21_SECURITY_FRISKING",
|
35 |
-
"20": "22_SITTING_IN_COUNTER",
|
36 |
-
"21": "23_SITTING_ON_TABLE",
|
37 |
-
"22": "24_STANDING_ON_TABLE",
|
38 |
-
"23": "25_STORE_CLOSING",
|
39 |
-
"24": "26_STORE_OPENING",
|
40 |
-
"25": "27_TEAM_MEETING",
|
41 |
-
"26": "28_THROWING_BAGS",
|
42 |
-
"27": "29_UNLOADING",
|
43 |
-
"28": "30_NoFight",
|
44 |
-
"29": "31_FIRE_DETECTION",
|
45 |
-
"30": "33_SHOPLIFTING",
|
46 |
-
"31": "34_BILLINGPROCESS",
|
47 |
-
"32": "35_SHOWINGMERCHANDISE",
|
48 |
-
"33": "36_FOLDINGCLOTHES",
|
49 |
-
"34": "37_DELIVERYPRODUCT",
|
50 |
-
"35": "38_CARRYINGCLOTHESBAGS",
|
51 |
-
"36": "40_HIGHFOOTFALL",
|
52 |
-
"37": "41_STOCKINWARD"
|
53 |
},
|
54 |
"image_size": 224,
|
55 |
"initializer_range": 0.02,
|
56 |
"intermediate_size": 3072,
|
57 |
"label2id": {
|
58 |
-
"
|
59 |
-
"
|
60 |
-
"
|
61 |
-
"
|
62 |
-
"
|
63 |
-
"
|
64 |
-
"08_EMPLOYEES_IN_GROUP": 6,
|
65 |
-
"09_EMPLOYEE_SITTING": 7,
|
66 |
-
"10_EMPLOYEE_SLEEPING": 8,
|
67 |
-
"11_EMPLOYEE_USING_LIFT": 9,
|
68 |
-
"12_EMPLOYEE_USING_MOBILE": 10,
|
69 |
-
"13_FIGHTING": 11,
|
70 |
-
"14_FILLING_FEEDBACK": 12,
|
71 |
-
"15_INCHARGE_IN_COUNTER": 13,
|
72 |
-
"16_LOADING": 14,
|
73 |
-
"17_MOPPING_FLOOR": 15,
|
74 |
-
"18_NO_ONE_IN_COUNTER": 16,
|
75 |
-
"19_OTHER_EMPLOYEES_DOING_BILLING": 17,
|
76 |
-
"20_RUNNING": 18,
|
77 |
-
"21_SECURITY_FRISKING": 19,
|
78 |
-
"22_SITTING_IN_COUNTER": 20,
|
79 |
-
"23_SITTING_ON_TABLE": 21,
|
80 |
-
"24_STANDING_ON_TABLE": 22,
|
81 |
-
"25_STORE_CLOSING": 23,
|
82 |
-
"26_STORE_OPENING": 24,
|
83 |
-
"27_TEAM_MEETING": 25,
|
84 |
-
"28_THROWING_BAGS": 26,
|
85 |
-
"29_UNLOADING": 27,
|
86 |
-
"30_NoFight": 28,
|
87 |
-
"31_FIRE_DETECTION": 29,
|
88 |
-
"33_SHOPLIFTING": 30,
|
89 |
-
"34_BILLINGPROCESS": 31,
|
90 |
-
"35_SHOWINGMERCHANDISE": 32,
|
91 |
-
"36_FOLDINGCLOTHES": 33,
|
92 |
-
"37_DELIVERYPRODUCT": 34,
|
93 |
-
"38_CARRYINGCLOTHESBAGS": 35,
|
94 |
-
"40_HIGHFOOTFALL": 36,
|
95 |
-
"41_STOCKINWARD": 37
|
96 |
},
|
97 |
"layer_norm_eps": 1e-12,
|
98 |
"model_type": "videomae",
|
@@ -102,9 +38,10 @@
|
|
102 |
"num_frames": 16,
|
103 |
"num_hidden_layers": 12,
|
104 |
"patch_size": 16,
|
|
|
105 |
"qkv_bias": true,
|
106 |
"torch_dtype": "float32",
|
107 |
-
"transformers_version": "4.
|
108 |
"tubelet_size": 2,
|
109 |
"use_mean_pooling": false
|
110 |
}
|
|
|
12 |
"hidden_dropout_prob": 0.0,
|
13 |
"hidden_size": 768,
|
14 |
"id2label": {
|
15 |
+
"0": "08_EMPLOYEES_IN_GROUP",
|
16 |
+
"1": "12_EMPLOYEE_USING_MOBILE",
|
17 |
+
"2": "17_MOPPING_FLOOR",
|
18 |
+
"3": "21_SECURITY_FRISKING",
|
19 |
+
"4": "31_FIRE_DETECTION",
|
20 |
+
"5": "33_SHOPLIFTING"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
},
|
22 |
"image_size": 224,
|
23 |
"initializer_range": 0.02,
|
24 |
"intermediate_size": 3072,
|
25 |
"label2id": {
|
26 |
+
"08_EMPLOYEES_IN_GROUP": 0,
|
27 |
+
"12_EMPLOYEE_USING_MOBILE": 1,
|
28 |
+
"17_MOPPING_FLOOR": 2,
|
29 |
+
"21_SECURITY_FRISKING": 3,
|
30 |
+
"31_FIRE_DETECTION": 4,
|
31 |
+
"33_SHOPLIFTING": 5
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
},
|
33 |
"layer_norm_eps": 1e-12,
|
34 |
"model_type": "videomae",
|
|
|
38 |
"num_frames": 16,
|
39 |
"num_hidden_layers": 12,
|
40 |
"patch_size": 16,
|
41 |
+
"problem_type": "single_label_classification",
|
42 |
"qkv_bias": true,
|
43 |
"torch_dtype": "float32",
|
44 |
+
"transformers_version": "4.40.2",
|
45 |
"tubelet_size": 2,
|
46 |
"use_mean_pooling": false
|
47 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b5b6b15df4dfca3158b060a254876f8acab0d345103679bef2aa210411c37057
|
3 |
+
size 344949680
|
preprocessor_config.json
CHANGED
@@ -1,4 +1,20 @@
|
|
1 |
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
"crop_size": {
|
3 |
"height": 224,
|
4 |
"width": 224
|
|
|
1 |
{
|
2 |
+
"_valid_processor_keys": [
|
3 |
+
"videos",
|
4 |
+
"do_resize",
|
5 |
+
"size",
|
6 |
+
"resample",
|
7 |
+
"do_center_crop",
|
8 |
+
"crop_size",
|
9 |
+
"do_rescale",
|
10 |
+
"rescale_factor",
|
11 |
+
"do_normalize",
|
12 |
+
"image_mean",
|
13 |
+
"image_std",
|
14 |
+
"return_tensors",
|
15 |
+
"data_format",
|
16 |
+
"input_data_format"
|
17 |
+
],
|
18 |
"crop_size": {
|
19 |
"height": 224,
|
20 |
"width": 224
|
runs/Aug12_15-39-39_surveillance/events.out.tfevents.1723457391.surveillance.95468.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f665acf9940a591fabcb006ba0db11b26381e80f2325ff61d6f8573ba0207e91
|
3 |
+
size 20505
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d98c485376c1af8a9b392c443d4ff52d023a5d2038846af2832a02236b57028d
|
3 |
+
size 4603
|