gauti83 commited on
Commit
1e46112
1 Parent(s): 6f8f389

Training in progress, epoch 0

Browse files
README.md CHANGED
@@ -3,6 +3,8 @@ license: cc-by-nc-4.0
3
  base_model: MCG-NJU/videomae-base
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: videomae-base-finetuned-subset
8
  results: []
@@ -14,6 +16,9 @@ should probably proofread and complete it, then remove this comment. -->
14
  # videomae-base-finetuned-subset
15
 
16
  This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
 
 
 
17
 
18
  ## Model description
19
 
@@ -33,17 +38,63 @@ More information needed
33
 
34
  The following hyperparameters were used during training:
35
  - learning_rate: 5e-05
36
- - train_batch_size: 8
37
- - eval_batch_size: 8
38
  - seed: 42
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
  - lr_scheduler_warmup_ratio: 0.1
42
- - training_steps: 37920
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
 
44
  ### Framework versions
45
 
46
- - Transformers 4.44.0
47
- - Pytorch 2.2.1+cu121
48
  - Datasets 2.20.0
49
  - Tokenizers 0.19.1
 
3
  base_model: MCG-NJU/videomae-base
4
  tags:
5
  - generated_from_trainer
6
+ metrics:
7
+ - accuracy
8
  model-index:
9
  - name: videomae-base-finetuned-subset
10
  results: []
 
16
  # videomae-base-finetuned-subset
17
 
18
  This model is a fine-tuned version of [MCG-NJU/videomae-base](https://huggingface.co/MCG-NJU/videomae-base) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 1.9422
21
+ - Accuracy: 0.7011
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 5e-05
41
+ - train_batch_size: 2
42
+ - eval_batch_size: 2
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_ratio: 0.1
47
+ - training_steps: 2640
48
+
49
+ ### Training results
50
+
51
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
+ |:-------------:|:-------:|:----:|:---------------:|:--------:|
53
+ | 1.9136 | 0.0254 | 67 | 2.1529 | 0.1096 |
54
+ | 1.9987 | 1.0254 | 134 | 1.9985 | 0.2593 |
55
+ | 1.0754 | 2.0254 | 201 | 2.1637 | 0.1956 |
56
+ | 1.413 | 3.0254 | 268 | 1.9523 | 0.1473 |
57
+ | 1.3047 | 4.0254 | 335 | 2.5442 | 0.1956 |
58
+ | 1.2488 | 5.0254 | 402 | 2.0563 | 0.2330 |
59
+ | 1.0581 | 6.0254 | 469 | 1.9954 | 0.2264 |
60
+ | 0.9165 | 7.0254 | 536 | 1.7661 | 0.2769 |
61
+ | 1.5722 | 8.0254 | 603 | 2.2872 | 0.2264 |
62
+ | 0.9083 | 9.0254 | 670 | 2.1004 | 0.3363 |
63
+ | 0.8093 | 10.0254 | 737 | 1.2497 | 0.6549 |
64
+ | 0.4925 | 11.0254 | 804 | 3.5720 | 0.2813 |
65
+ | 0.4573 | 12.0254 | 871 | 1.5213 | 0.3604 |
66
+ | 1.1082 | 13.0254 | 938 | 1.5453 | 0.5934 |
67
+ | 0.8066 | 14.0254 | 1005 | 2.9169 | 0.2967 |
68
+ | 0.6615 | 15.0254 | 1072 | 2.1412 | 0.5780 |
69
+ | 0.146 | 16.0254 | 1139 | 2.5006 | 0.3978 |
70
+ | 0.3815 | 17.0254 | 1206 | 1.7907 | 0.5956 |
71
+ | 0.2124 | 18.0254 | 1273 | 1.6622 | 0.6527 |
72
+ | 0.5304 | 19.0254 | 1340 | 1.8988 | 0.5956 |
73
+ | 0.1519 | 20.0254 | 1407 | 2.7940 | 0.3934 |
74
+ | 0.486 | 21.0254 | 1474 | 2.6766 | 0.4198 |
75
+ | 0.5502 | 22.0254 | 1541 | 2.3451 | 0.5495 |
76
+ | 0.7527 | 23.0254 | 1608 | 1.7518 | 0.6462 |
77
+ | 0.3194 | 24.0254 | 1675 | 2.0738 | 0.5890 |
78
+ | 0.0189 | 25.0254 | 1742 | 2.9264 | 0.5407 |
79
+ | 0.2928 | 26.0254 | 1809 | 2.5495 | 0.5451 |
80
+ | 0.0036 | 27.0254 | 1876 | 1.8143 | 0.6989 |
81
+ | 0.3772 | 28.0254 | 1943 | 2.2384 | 0.6088 |
82
+ | 0.0044 | 29.0254 | 2010 | 1.7688 | 0.7033 |
83
+ | 0.7291 | 30.0254 | 2077 | 2.0591 | 0.6571 |
84
+ | 0.1553 | 31.0254 | 2144 | 2.0690 | 0.6505 |
85
+ | 0.5454 | 32.0254 | 2211 | 1.8762 | 0.7055 |
86
+ | 0.3096 | 33.0254 | 2278 | 2.2310 | 0.6440 |
87
+ | 0.0053 | 34.0254 | 2345 | 2.0907 | 0.6615 |
88
+ | 0.0024 | 35.0254 | 2412 | 2.4127 | 0.6022 |
89
+ | 0.0022 | 36.0254 | 2479 | 2.0037 | 0.6989 |
90
+ | 0.0026 | 37.0254 | 2546 | 2.0130 | 0.6725 |
91
+ | 0.0013 | 38.0254 | 2613 | 1.9391 | 0.6967 |
92
+ | 0.0017 | 39.0102 | 2640 | 1.9422 | 0.7011 |
93
+
94
 
95
  ### Framework versions
96
 
97
+ - Transformers 4.40.2
98
+ - Pytorch 2.0.1+cu117
99
  - Datasets 2.20.0
100
  - Tokenizers 0.19.1
config.json CHANGED
@@ -12,87 +12,23 @@
12
  "hidden_dropout_prob": 0.0,
13
  "hidden_size": 768,
14
  "id2label": {
15
- "0": "01_CASH_HANDED_OVER",
16
- "1": "02_CASH_LOCKER_CLOSING",
17
- "2": "03_CASH_LOCKER_OPENING",
18
- "3": "04_CUSTOMER_BAG_CHECK",
19
- "4": "05_EATING_FOOD_IN_STORE_",
20
- "5": "07_EMPLOYEE_HIDING_BEHIND_PARCELS",
21
- "6": "08_EMPLOYEES_IN_GROUP",
22
- "7": "09_EMPLOYEE_SITTING",
23
- "8": "10_EMPLOYEE_SLEEPING",
24
- "9": "11_EMPLOYEE_USING_LIFT",
25
- "10": "12_EMPLOYEE_USING_MOBILE",
26
- "11": "13_FIGHTING",
27
- "12": "14_FILLING_FEEDBACK",
28
- "13": "15_INCHARGE_IN_COUNTER",
29
- "14": "16_LOADING",
30
- "15": "17_MOPPING_FLOOR",
31
- "16": "18_NO_ONE_IN_COUNTER",
32
- "17": "19_OTHER_EMPLOYEES_DOING_BILLING",
33
- "18": "20_RUNNING",
34
- "19": "21_SECURITY_FRISKING",
35
- "20": "22_SITTING_IN_COUNTER",
36
- "21": "23_SITTING_ON_TABLE",
37
- "22": "24_STANDING_ON_TABLE",
38
- "23": "25_STORE_CLOSING",
39
- "24": "26_STORE_OPENING",
40
- "25": "27_TEAM_MEETING",
41
- "26": "28_THROWING_BAGS",
42
- "27": "29_UNLOADING",
43
- "28": "30_NoFight",
44
- "29": "31_FIRE_DETECTION",
45
- "30": "33_SHOPLIFTING",
46
- "31": "34_BILLINGPROCESS",
47
- "32": "35_SHOWINGMERCHANDISE",
48
- "33": "36_FOLDINGCLOTHES",
49
- "34": "37_DELIVERYPRODUCT",
50
- "35": "38_CARRYINGCLOTHESBAGS",
51
- "36": "40_HIGHFOOTFALL",
52
- "37": "41_STOCKINWARD"
53
  },
54
  "image_size": 224,
55
  "initializer_range": 0.02,
56
  "intermediate_size": 3072,
57
  "label2id": {
58
- "01_CASH_HANDED_OVER": 0,
59
- "02_CASH_LOCKER_CLOSING": 1,
60
- "03_CASH_LOCKER_OPENING": 2,
61
- "04_CUSTOMER_BAG_CHECK": 3,
62
- "05_EATING_FOOD_IN_STORE_": 4,
63
- "07_EMPLOYEE_HIDING_BEHIND_PARCELS": 5,
64
- "08_EMPLOYEES_IN_GROUP": 6,
65
- "09_EMPLOYEE_SITTING": 7,
66
- "10_EMPLOYEE_SLEEPING": 8,
67
- "11_EMPLOYEE_USING_LIFT": 9,
68
- "12_EMPLOYEE_USING_MOBILE": 10,
69
- "13_FIGHTING": 11,
70
- "14_FILLING_FEEDBACK": 12,
71
- "15_INCHARGE_IN_COUNTER": 13,
72
- "16_LOADING": 14,
73
- "17_MOPPING_FLOOR": 15,
74
- "18_NO_ONE_IN_COUNTER": 16,
75
- "19_OTHER_EMPLOYEES_DOING_BILLING": 17,
76
- "20_RUNNING": 18,
77
- "21_SECURITY_FRISKING": 19,
78
- "22_SITTING_IN_COUNTER": 20,
79
- "23_SITTING_ON_TABLE": 21,
80
- "24_STANDING_ON_TABLE": 22,
81
- "25_STORE_CLOSING": 23,
82
- "26_STORE_OPENING": 24,
83
- "27_TEAM_MEETING": 25,
84
- "28_THROWING_BAGS": 26,
85
- "29_UNLOADING": 27,
86
- "30_NoFight": 28,
87
- "31_FIRE_DETECTION": 29,
88
- "33_SHOPLIFTING": 30,
89
- "34_BILLINGPROCESS": 31,
90
- "35_SHOWINGMERCHANDISE": 32,
91
- "36_FOLDINGCLOTHES": 33,
92
- "37_DELIVERYPRODUCT": 34,
93
- "38_CARRYINGCLOTHESBAGS": 35,
94
- "40_HIGHFOOTFALL": 36,
95
- "41_STOCKINWARD": 37
96
  },
97
  "layer_norm_eps": 1e-12,
98
  "model_type": "videomae",
@@ -102,9 +38,10 @@
102
  "num_frames": 16,
103
  "num_hidden_layers": 12,
104
  "patch_size": 16,
 
105
  "qkv_bias": true,
106
  "torch_dtype": "float32",
107
- "transformers_version": "4.44.0",
108
  "tubelet_size": 2,
109
  "use_mean_pooling": false
110
  }
 
12
  "hidden_dropout_prob": 0.0,
13
  "hidden_size": 768,
14
  "id2label": {
15
+ "0": "08_EMPLOYEES_IN_GROUP",
16
+ "1": "12_EMPLOYEE_USING_MOBILE",
17
+ "2": "17_MOPPING_FLOOR",
18
+ "3": "21_SECURITY_FRISKING",
19
+ "4": "31_FIRE_DETECTION",
20
+ "5": "33_SHOPLIFTING"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
  },
22
  "image_size": 224,
23
  "initializer_range": 0.02,
24
  "intermediate_size": 3072,
25
  "label2id": {
26
+ "08_EMPLOYEES_IN_GROUP": 0,
27
+ "12_EMPLOYEE_USING_MOBILE": 1,
28
+ "17_MOPPING_FLOOR": 2,
29
+ "21_SECURITY_FRISKING": 3,
30
+ "31_FIRE_DETECTION": 4,
31
+ "33_SHOPLIFTING": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
  },
33
  "layer_norm_eps": 1e-12,
34
  "model_type": "videomae",
 
38
  "num_frames": 16,
39
  "num_hidden_layers": 12,
40
  "patch_size": 16,
41
+ "problem_type": "single_label_classification",
42
  "qkv_bias": true,
43
  "torch_dtype": "float32",
44
+ "transformers_version": "4.40.2",
45
  "tubelet_size": 2,
46
  "use_mean_pooling": false
47
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5f1318e883106ad5b5ed5348bb2fee4bc507e0a729b8e918967cd76b05ec666
3
- size 345048120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5b6b15df4dfca3158b060a254876f8acab0d345103679bef2aa210411c37057
3
+ size 344949680
preprocessor_config.json CHANGED
@@ -1,4 +1,20 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "crop_size": {
3
  "height": 224,
4
  "width": 224
 
1
  {
2
+ "_valid_processor_keys": [
3
+ "videos",
4
+ "do_resize",
5
+ "size",
6
+ "resample",
7
+ "do_center_crop",
8
+ "crop_size",
9
+ "do_rescale",
10
+ "rescale_factor",
11
+ "do_normalize",
12
+ "image_mean",
13
+ "image_std",
14
+ "return_tensors",
15
+ "data_format",
16
+ "input_data_format"
17
+ ],
18
  "crop_size": {
19
  "height": 224,
20
  "width": 224
runs/Aug12_15-39-39_surveillance/events.out.tfevents.1723457391.surveillance.95468.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f665acf9940a591fabcb006ba0db11b26381e80f2325ff61d6f8573ba0207e91
3
+ size 20505
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47f0da6ab5aa8218e99facc794e0186f6c4a36936bcb2dba6a54d641612fc738
3
- size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d98c485376c1af8a9b392c443d4ff52d023a5d2038846af2832a02236b57028d
3
+ size 4603