mwmathis commited on
Commit
d22ff20
1 Parent(s): ef415ae

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -19
README.md CHANGED
@@ -3,22 +3,18 @@ tags:
3
  - computer_vision
4
  - pose_estimation
5
  ---
 
6
 
7
- Copyright 2021-2023 by Mackenzie Mathis, Alexander Mathis, Shaokai Ye and contributors. All rights reserved.
8
 
 
 
9
 
10
- - Please cite **Ye et al 2023** if you use this model in your work https://arxiv.org/abs/2203.07436v1
11
- - If this license is not suitable for your business or project
12
- please contact EPFL-TTO (https://tto.epfl.ch/) or Mackenzie Mathis (mackenzie.mathis @ epfl.ch) for a full commercial license.
13
 
14
- This software may not be used to harm any animal deliberately!
15
-
16
-
17
- **MODEL CARD:**
18
-
19
- This model was trained a dataset called "TopViewMouse-5K." It was trained in Tensorflow 2 within the [DeepLabCut framework](www.deeplabcut.org).
20
- Full training details can be found in Ye et al. 2023.
21
- You can use this model simply with our light-weight loading package called [DLCLibrary](https://github.com/DeepLabCut/DLClibrary). Here is an example useage:
22
 
23
  ```python
24
  from pathlib import Path
@@ -30,7 +26,36 @@ model_dir.mkdir()
30
  download_huggingface_model("superanimal_topviewmouse", model_dir)
31
  ```
32
 
33
- **Training Data:**
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
34
 
35
  It consists of being trained together on the following datasets:
36
 
@@ -57,16 +82,58 @@ of Geneva (Switzerland; license GE01) (n=4 mice).
57
 
58
  Here is an image with examples from the datasets, the distribution of images per dataset, and the keypoint guide.
59
 
60
- Please note that each dataest was labeled by separate labs, seperate individuals, therefore while we map names
61
- to a unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary Note on annotator bias).
62
- You will also note the dataset is primarily using C56Blk6/J mice and only some CD1 examples.
63
- We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023),
64
- or fine-tune these weights with your own labeling.
65
-
66
  <p align="center">
67
  <img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690986892069-I1DP3EQU14DSP5WB6FSI/modelcard-TVM.png?format=1500w" width="95%">
68
  </p>
69
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
70
 
71
  1. Oliver Sturman, Lukas von Ziegler, Christa Schläppi, Furkan Akyol, Mattia Privitera, Daria Slominski, Christina Grimm, Laetitia Thieren, Valerio
72
  Zerbi, Benjamin Grewe, et al. Deep learning-based behavioral analysis reaches human accuracy and is capable of outperforming commercial
 
3
  - computer_vision
4
  - pose_estimation
5
  ---
6
+ # MODEL CARD:
7
 
8
+ ## Model Details
9
 
10
+ • SuperAnimal-TopViewMouse model developed by the [M.W.Mathis Lab](http://www.mackenziemathislab.org/) in 2023, trained to predict quadruped pose from images.
11
+ Please see [Shaokai Ye et al. 2023](https://arxiv.org/abs/2203.07436) for details.
12
 
13
+ The model is an DLRRNet and HRNet-w32 trained on our TopViewMouse-5K dataset.
 
 
14
 
15
+ • It was trained within the DeepLabCut framework. Full training details can be found in Ye et al. 2023.
16
+ You can use this model simply with our light-weight loading package called [DLCLibrary](https://github.com/DeepLabCut/DLClibrary).
17
+ Here is an example useage:
 
 
 
 
 
18
 
19
  ```python
20
  from pathlib import Path
 
26
  download_huggingface_model("superanimal_topviewmouse", model_dir)
27
  ```
28
 
29
+ ## Intended Use
30
+
31
+ • Intended to be used for pose tracking of lab mice videos filmed from an overhead view. The models can be used as a plug-and-
32
+ play solution if extremely high precision is not required (we benchmark the zero-shot performance in the paper). Otherwise, it is
33
+ recommended to also be used as the weights for transfer learning and fine-tuning.
34
+
35
+ • Intended for academic and research professionals working in fields related to animal behavior, neuroscience, biomechanics, and
36
+ ecology.
37
+
38
+ • Not suitable for other species and other camera views. Also not suitable for videos that look dramatically different from those we
39
+ show in the paper.
40
+
41
+ ## Factors
42
+
43
+ • Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and resolution of the
44
+ video frames. The present of objects might also cause false detections of the mice and keypoints. When two or more animals are
45
+ extremely close, it could cause the top-down detectors to only detect only one animal, if used without further fine-tuning.
46
+
47
+
48
+ ## Metrics
49
+ • Mean Average Precision (mAP)
50
+
51
+ • Root Mean Square Error (RMSE)
52
+
53
+ ## Evaluation Data
54
+
55
+ • The test split of TopViewMouse-5K and in the paper on two benchmarks, DLC Openfield and TriMouse
56
+
57
+
58
+ ## Training Data
59
 
60
  It consists of being trained together on the following datasets:
61
 
 
82
 
83
  Here is an image with examples from the datasets, the distribution of images per dataset, and the keypoint guide.
84
 
 
 
 
 
 
 
85
  <p align="center">
86
  <img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690986892069-I1DP3EQU14DSP5WB6FSI/modelcard-TVM.png?format=1500w" width="95%">
87
  </p>
88
 
89
+ ## Ethical Considerations
90
+
91
+ • Data was collected with IUCAC or other governmental approval. Each individual dataset used in training reports the ethics approval
92
+ they obtained.
93
+
94
+ ## Caveats and Recommendations
95
+
96
+ • The model may have reduced accuracy in scenarios with extremely varied lighting conditions or atypical mouse characteristics not
97
+ well-represented in the training data. For example, this dataset only has one set of white mice, therefore it may not generalize well
98
+ to diverse settings of white lab mice.
99
+
100
+ • Please note that each training dataset was labeled by separate labs and different individuals, therefore while we map names to a
101
+ unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary Note on
102
+ annotator bias).
103
+
104
+ • Note the dataset is primarily using C56Blk6/J mice and only some CD1 examples.
105
+
106
+ • We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these
107
+ weights with your own labeling.
108
+
109
+ ## License
110
+
111
+ Modified MIT.
112
+
113
+ Copyright 2023 by Mackenzie Mathis, Shaokai Ye, and contributors.
114
+
115
+ Permission is hereby granted to you (hereafter "LICENSEE") a fully-paid, non-exclusive,
116
+ and non-transferable license for academic, non-commercial purposes only (hereafter “LICENSE”)
117
+ to use the "MODEL" weights (hereafter "MODEL"), subject to the following conditions:
118
+
119
+ The above copyright notice and this permission notice shall be included in all copies or substantial
120
+ portions of the Software:
121
+
122
+ This software may not be used to harm any animal deliberately.
123
+
124
+ LICENSEE acknowledges that the MODEL is a research tool.
125
+ THE MODEL IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING
126
+ BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
127
+ IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
128
+ WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE MODEL
129
+ OR THE USE OR OTHER DEALINGS IN THE MODEL.
130
+
131
+ If this license is not appropriate for your application, please contact Prof. Mackenzie W. Mathis
132
+ (mackenzie@post.harvard.edu) and/or the TTO office at EPFL (tto@epfl.ch) for a commercial use license.
133
+
134
+ Please cite **Ye et al** if you use this model in your work https://arxiv.org/abs/2203.07436v2.
135
+
136
+ ## References
137
 
138
  1. Oliver Sturman, Lukas von Ziegler, Christa Schläppi, Furkan Akyol, Mattia Privitera, Daria Slominski, Christina Grimm, Laetitia Thieren, Valerio
139
  Zerbi, Benjamin Grewe, et al. Deep learning-based behavioral analysis reaches human accuracy and is capable of outperforming commercial