Update README.md
Browse files
README.md
CHANGED
@@ -3,22 +3,18 @@ tags:
|
|
3 |
- computer_vision
|
4 |
- pose_estimation
|
5 |
---
|
|
|
6 |
|
7 |
-
|
8 |
|
|
|
|
|
9 |
|
10 |
-
|
11 |
-
- If this license is not suitable for your business or project
|
12 |
-
please contact EPFL-TTO (https://tto.epfl.ch/) or Mackenzie Mathis (mackenzie.mathis @ epfl.ch) for a full commercial license.
|
13 |
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
**MODEL CARD:**
|
18 |
-
|
19 |
-
This model was trained a dataset called "TopViewMouse-5K." It was trained in Tensorflow 2 within the [DeepLabCut framework](www.deeplabcut.org).
|
20 |
-
Full training details can be found in Ye et al. 2023.
|
21 |
-
You can use this model simply with our light-weight loading package called [DLCLibrary](https://github.com/DeepLabCut/DLClibrary). Here is an example useage:
|
22 |
|
23 |
```python
|
24 |
from pathlib import Path
|
@@ -30,7 +26,36 @@ model_dir.mkdir()
|
|
30 |
download_huggingface_model("superanimal_topviewmouse", model_dir)
|
31 |
```
|
32 |
|
33 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
34 |
|
35 |
It consists of being trained together on the following datasets:
|
36 |
|
@@ -57,16 +82,58 @@ of Geneva (Switzerland; license GE01) (n=4 mice).
|
|
57 |
|
58 |
Here is an image with examples from the datasets, the distribution of images per dataset, and the keypoint guide.
|
59 |
|
60 |
-
Please note that each dataest was labeled by separate labs, seperate individuals, therefore while we map names
|
61 |
-
to a unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary Note on annotator bias).
|
62 |
-
You will also note the dataset is primarily using C56Blk6/J mice and only some CD1 examples.
|
63 |
-
We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023),
|
64 |
-
or fine-tune these weights with your own labeling.
|
65 |
-
|
66 |
<p align="center">
|
67 |
<img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690986892069-I1DP3EQU14DSP5WB6FSI/modelcard-TVM.png?format=1500w" width="95%">
|
68 |
</p>
|
69 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
70 |
|
71 |
1. Oliver Sturman, Lukas von Ziegler, Christa Schläppi, Furkan Akyol, Mattia Privitera, Daria Slominski, Christina Grimm, Laetitia Thieren, Valerio
|
72 |
Zerbi, Benjamin Grewe, et al. Deep learning-based behavioral analysis reaches human accuracy and is capable of outperforming commercial
|
|
|
3 |
- computer_vision
|
4 |
- pose_estimation
|
5 |
---
|
6 |
+
# MODEL CARD:
|
7 |
|
8 |
+
## Model Details
|
9 |
|
10 |
+
• SuperAnimal-TopViewMouse model developed by the [M.W.Mathis Lab](http://www.mackenziemathislab.org/) in 2023, trained to predict quadruped pose from images.
|
11 |
+
Please see [Shaokai Ye et al. 2023](https://arxiv.org/abs/2203.07436) for details.
|
12 |
|
13 |
+
• The model is an DLRRNet and HRNet-w32 trained on our TopViewMouse-5K dataset.
|
|
|
|
|
14 |
|
15 |
+
• It was trained within the DeepLabCut framework. Full training details can be found in Ye et al. 2023.
|
16 |
+
You can use this model simply with our light-weight loading package called [DLCLibrary](https://github.com/DeepLabCut/DLClibrary).
|
17 |
+
Here is an example useage:
|
|
|
|
|
|
|
|
|
|
|
18 |
|
19 |
```python
|
20 |
from pathlib import Path
|
|
|
26 |
download_huggingface_model("superanimal_topviewmouse", model_dir)
|
27 |
```
|
28 |
|
29 |
+
## Intended Use
|
30 |
+
|
31 |
+
• Intended to be used for pose tracking of lab mice videos filmed from an overhead view. The models can be used as a plug-and-
|
32 |
+
play solution if extremely high precision is not required (we benchmark the zero-shot performance in the paper). Otherwise, it is
|
33 |
+
recommended to also be used as the weights for transfer learning and fine-tuning.
|
34 |
+
|
35 |
+
• Intended for academic and research professionals working in fields related to animal behavior, neuroscience, biomechanics, and
|
36 |
+
ecology.
|
37 |
+
|
38 |
+
• Not suitable for other species and other camera views. Also not suitable for videos that look dramatically different from those we
|
39 |
+
show in the paper.
|
40 |
+
|
41 |
+
## Factors
|
42 |
+
|
43 |
+
• Based on the known robustness issues of neural networks, the relevant factors include the lighting, contrast and resolution of the
|
44 |
+
video frames. The present of objects might also cause false detections of the mice and keypoints. When two or more animals are
|
45 |
+
extremely close, it could cause the top-down detectors to only detect only one animal, if used without further fine-tuning.
|
46 |
+
|
47 |
+
|
48 |
+
## Metrics
|
49 |
+
• Mean Average Precision (mAP)
|
50 |
+
|
51 |
+
• Root Mean Square Error (RMSE)
|
52 |
+
|
53 |
+
## Evaluation Data
|
54 |
+
|
55 |
+
• The test split of TopViewMouse-5K and in the paper on two benchmarks, DLC Openfield and TriMouse
|
56 |
+
|
57 |
+
|
58 |
+
## Training Data
|
59 |
|
60 |
It consists of being trained together on the following datasets:
|
61 |
|
|
|
82 |
|
83 |
Here is an image with examples from the datasets, the distribution of images per dataset, and the keypoint guide.
|
84 |
|
|
|
|
|
|
|
|
|
|
|
|
|
85 |
<p align="center">
|
86 |
<img src="https://images.squarespace-cdn.com/content/v1/57f6d51c9f74566f55ecf271/1690986892069-I1DP3EQU14DSP5WB6FSI/modelcard-TVM.png?format=1500w" width="95%">
|
87 |
</p>
|
88 |
|
89 |
+
## Ethical Considerations
|
90 |
+
|
91 |
+
• Data was collected with IUCAC or other governmental approval. Each individual dataset used in training reports the ethics approval
|
92 |
+
they obtained.
|
93 |
+
|
94 |
+
## Caveats and Recommendations
|
95 |
+
|
96 |
+
• The model may have reduced accuracy in scenarios with extremely varied lighting conditions or atypical mouse characteristics not
|
97 |
+
well-represented in the training data. For example, this dataset only has one set of white mice, therefore it may not generalize well
|
98 |
+
to diverse settings of white lab mice.
|
99 |
+
|
100 |
+
• Please note that each training dataset was labeled by separate labs and different individuals, therefore while we map names to a
|
101 |
+
unified pose vocabulary, there will be annotator bias in keypoint placement (See Ye et al. 2023 for our Supplementary Note on
|
102 |
+
annotator bias).
|
103 |
+
|
104 |
+
• Note the dataset is primarily using C56Blk6/J mice and only some CD1 examples.
|
105 |
+
|
106 |
+
• We recommend if performance is not as good as you need it to be, first try video adaptation (see Ye et al. 2023), or fine-tune these
|
107 |
+
weights with your own labeling.
|
108 |
+
|
109 |
+
## License
|
110 |
+
|
111 |
+
Modified MIT.
|
112 |
+
|
113 |
+
Copyright 2023 by Mackenzie Mathis, Shaokai Ye, and contributors.
|
114 |
+
|
115 |
+
Permission is hereby granted to you (hereafter "LICENSEE") a fully-paid, non-exclusive,
|
116 |
+
and non-transferable license for academic, non-commercial purposes only (hereafter “LICENSE”)
|
117 |
+
to use the "MODEL" weights (hereafter "MODEL"), subject to the following conditions:
|
118 |
+
|
119 |
+
The above copyright notice and this permission notice shall be included in all copies or substantial
|
120 |
+
portions of the Software:
|
121 |
+
|
122 |
+
This software may not be used to harm any animal deliberately.
|
123 |
+
|
124 |
+
LICENSEE acknowledges that the MODEL is a research tool.
|
125 |
+
THE MODEL IS PROVIDED “AS IS”, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING
|
126 |
+
BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
|
127 |
+
IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY,
|
128 |
+
WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE MODEL
|
129 |
+
OR THE USE OR OTHER DEALINGS IN THE MODEL.
|
130 |
+
|
131 |
+
If this license is not appropriate for your application, please contact Prof. Mackenzie W. Mathis
|
132 |
+
(mackenzie@post.harvard.edu) and/or the TTO office at EPFL (tto@epfl.ch) for a commercial use license.
|
133 |
+
|
134 |
+
Please cite **Ye et al** if you use this model in your work https://arxiv.org/abs/2203.07436v2.
|
135 |
+
|
136 |
+
## References
|
137 |
|
138 |
1. Oliver Sturman, Lukas von Ziegler, Christa Schläppi, Furkan Akyol, Mattia Privitera, Daria Slominski, Christina Grimm, Laetitia Thieren, Valerio
|
139 |
Zerbi, Benjamin Grewe, et al. Deep learning-based behavioral analysis reaches human accuracy and is capable of outperforming commercial
|