HuggingSpaces / models /download.sh
khulnasoft's picture
Create download.sh
3e8f8ea verified
raw
history blame
1.74 kB
#!/bin/bash
# Set models and datasets to download
models=(
"nlpconnect/vit-gpt2-image-captioning"
"lllyasviel/ControlNet"
"lllyasviel/sd-controlnet-canny"
"lllyasviel/sd-controlnet-depth"
"lllyasviel/sd-controlnet-hed"
"lllyasviel/sd-controlnet-mlsd"
"lllyasviel/sd-controlnet-openpose"
"lllyasviel/sd-controlnet-scribble"
"lllyasviel/sd-controlnet-seg"
"runwayml/stable-diffusion-v1-5"
"damo-vilab/text-to-video-ms-1.7b"
"microsoft/speecht5_asr"
"JorisCos/DCCRNet_Libri1Mix_enhsingle_16k"
"espnet/kan-bayashi_ljspeech_vits"
"facebook/detr-resnet-101"
"microsoft/speecht5_hifigan"
"microsoft/speecht5_vc"
"openai/whisper-base"
"Intel/dpt-large"
"facebook/detr-resnet-50-panoptic"
"facebook/detr-resnet-50"
"google/owlvit-base-patch32"
"impira/layoutlm-document-qa"
"ydshieh/vit-gpt2-coco-en"
"dandelin/vilt-b32-finetuned-vqa"
"lambdalabs/sd-image-variations-diffusers"
"facebook/maskformer-swin-base-coco"
"Intel/dpt-hybrid-midas"
)
datasets=("Matthijs/cmu-arctic-xvectors")
# Set the current directory
CURRENT_DIR=$(pwd)
# Download models
for model in "${models[@]}"; do
echo "----- Downloading from https://huggingface.co/${model} -----"
if [ -d "${model}" ]; then
(cd "${model}" && git pull && git lfs pull)
else
git clone --recurse-submodules "https://huggingface.co/${model}" "${model}"
fi
done
# Download datasets
for dataset in "${datasets[@]}"; do
echo "----- Downloading from https://huggingface.co/datasets/${dataset} -----"
if [ -d "${dataset}" ]; then
(cd "${dataset}" && git pull && git lfs pull)
else
git clone --recurse-submodules "https://huggingface.co/datasets/${dataset}" "${dataset}"
fi
done