|
--- |
|
license: cc-by-nc-4.0 |
|
tags: |
|
- sparsh |
|
- DIGIT |
|
- SSL pre-training |
|
pretty_name: touch-slide |
|
--- |
|
|
|
# Dataset Details |
|
|
|
Touch-Slide is a dataset inspired by [YCB-Slide](https://github.com/rpl-cmu/YCB-Slide). Its purpose is to increase the amount of data from multiple DIGIT sensors for self-supervised learning (SSL) pre-training of the Sparsh model. Touch-Slide consists of human-sliding interactions on toy kitchen objects using the DIGIT sensor. We used 9 objects, as shown below, and collected 5 trajectories for each, resulting in a total of 180k frames. |
|
|
|
<!-- insert figure from assets/touch_slide_objs.png --> |
|
![Touch-Slide](assets/touch_slide_objs.png) |
|
|
|
This is a visual example of how the dataset was collected, showcasing sliding interactions that capture trajectories rich in shear forces: |
|
|
|
![a](assets/bread-real.gif) ![b](assets/bread-digit.gif) |
|
|
|
|
|
|
|
## Uses |
|
|
|
This dataset does not include labels and is intended for self-supervised training only. It is specifically designed for training the Sparsh models listed in the Hugging Face [Sparsh collection](https://huggingface.co/collections/facebook/sparsh-67167ce57566196a4526c328). |
|
|
|
Please refer to the [Sparsh repo](https://github.com/facebookresearch/sparsh) for further information about usage. |
|
|
|
## Dataset Structure |
|
|
|
The dataset consists of 5 trajectories for each object. Each trajectory is stored as a pickle file, containing binarized tactile images. The structure is as follows: |
|
|
|
```bash |
|
Touch-Slide |
|
βββ object_0 # eg: bread |
|
β βββ dataset_0.pkl |
|
β βββ ... |
|
β βββ dataset_4.pkl |
|
βββ object_1 # eg: corn |
|
βββ ... |
|
``` |
|
|
|
This is a sample code for loading the dataset in pickle format and extracting the images: |
|
|
|
```python |
|
def load_pickle_dataset(file_dataset): |
|
with open(file_dataset, "rb") as f: |
|
all_frames = pickle.load(f) |
|
return all_frames |
|
|
|
def load_bin_image(io_buf): |
|
img = Image.open(io.BytesIO(io_buf)) |
|
img = np.array(img) |
|
return img |
|
|
|
frames = load_pickle_dataset('bread/dataset_0.pkl') |
|
img = load_bin_image(frames[0]) |
|
``` |
|
|
|
## BibTeX entry and citation info |
|
|
|
```bibtex |
|
@inproceedings{ |
|
higuera2024sparsh, |
|
title={Sparsh: Self-supervised touch representations for vision-based tactile sensing}, |
|
author={Carolina Higuera and Akash Sharma and Chaithanya Krishna Bodduluri and Taosha Fan and Patrick Lancaster and Mrinal Kalakrishnan and Michael Kaess and Byron Boots and Mike Lambeta and Tingfan Wu and Mustafa Mukadam}, |
|
booktitle={8th Annual Conference on Robot Learning}, |
|
year={2024}, |
|
url={https://openreview.net/forum?id=xYJn2e1uu8} |
|
} |
|
``` |
|
|