|
--- |
|
license: cc-by-4.0 |
|
pretty_name: SDSS 4d data cubes |
|
tags: |
|
- astronomy |
|
- compression |
|
- images |
|
dataset_info: |
|
config_name: tiny |
|
features: |
|
- name: image |
|
dtype: |
|
array4_d: |
|
shape: |
|
- 5 |
|
- 800 |
|
- 800 |
|
dtype: uint16 |
|
- name: ra |
|
dtype: float64 |
|
- name: dec |
|
dtype: float64 |
|
- name: pixscale |
|
dtype: float64 |
|
- name: ntimes |
|
dtype: int64 |
|
- name: nbands |
|
dtype: int64 |
|
splits: |
|
- name: train |
|
num_bytes: 558194176 |
|
num_examples: 2 |
|
- name: test |
|
num_bytes: 352881364 |
|
num_examples: 1 |
|
download_size: 908845172 |
|
dataset_size: 911075540 |
|
--- |
|
|
|
# GBI-16-4D Dataset |
|
|
|
GBI-16-4D is a dataset which is part of the AstroCompress project. It contains data assembled from the Sloan Digital SkySurvey (SDSS). Each FITS file contains a series of 800x800 pixel uint16 observations of the same portion of the Stripe82 field, taken in 5 bandpass filters (u, g, r, i, z) over time. The filenames give the |
|
starting run, field, camcol of the observations, the number of filtered images per timestep, and the number of timesteps. For example: |
|
|
|
```cube_center_run4203_camcol6_f44_35-5-800-800.fits``` |
|
|
|
contains 35 frames of 800x800 pixel images in 5 bandpasses starting with run 4203, field 44, and camcol 6. The images are stored in the FITS standard. |
|
|
|
# Usage |
|
|
|
You first need to install the `datasets` and `astropy` packages: |
|
|
|
```bash |
|
pip install datasets astropy |
|
``` |
|
|
|
There are two datasets: `tiny` and `full`, each with `train` and `test` splits. The `tiny` dataset has 2 4D images in the `train` and 1 in the `test`. The `full` dataset contains all the images in the `data/` directory. |
|
|
|
## Local Use (RECOMMENDED) |
|
|
|
You can clone this repo and use directly without connecting to hf: |
|
|
|
```bash |
|
git clone https://huggingface.co/datasets/AnonAstroData/GBI-16-4D |
|
``` |
|
|
|
```bash |
|
git lfs pull |
|
``` |
|
|
|
Then `cd GBI-16-4D` and start python like: |
|
|
|
```python |
|
from datasets import load_dataset |
|
dataset = load_dataset("./GBI-16-4D.py", "tiny", data_dir="./data/", writer_batch_size=1, trust_remote_code=True) |
|
ds = dataset.with_format("np") |
|
``` |
|
|
|
Now you should be able to use the `ds` variable like: |
|
|
|
```python |
|
ds["test"][0]["image"].shape # -> (55, 5, 800, 800) |
|
``` |
|
|
|
Note of course that it will take a long time to download and convert the images in the local cache for the `full` dataset. Afterward, the usage should be quick as the files are memory-mapped from disk. |
|
|
|
|
|
## Use from Huggingface Directly |
|
|
|
This method may only be an option when trying to access the "tiny" version of the dataset. |
|
|
|
To directly use from this data from Huggingface, you'll want to log in on the command line before starting python: |
|
|
|
```bash |
|
huggingface-cli login |
|
``` |
|
|
|
or |
|
|
|
``` |
|
import huggingface_hub |
|
huggingface_hub.login(token=token) |
|
``` |
|
|
|
Then in your python script: |
|
|
|
```python |
|
from datasets import load_dataset |
|
dataset = load_dataset("AstroCompress/GBI-16-4D", "tiny", writer_batch_size=1, trust_remote_code=True) |
|
ds = dataset.with_format("np") |
|
``` |
|
|
|
|
|
## Demo Colab Notebook |
|
We provide a demo collab notebook to get started on using the dataset [here](https://colab.research.google.com/drive/1wcz7qMqSAMST2kXFlL-TbwpYR26gYIYy?usp=sharing). |
|
|
|
|
|
## Utils scripts |
|
Note that utils scripts such as `eval_baselines.py` must be run from the parent directory of `utils`, i.e. `python utils/eval_baselines.py`. |
|
|