Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
FROM python:3.10.13-slim-bullseye | |
ENV DEBIAN_FRONTEND=noninteractive | |
RUN apt-get update && \ | |
apt-get upgrade -y && \ | |
apt-get install -y --no-install-recommends \ | |
git \ | |
git-lfs \ | |
wget \ | |
curl \ | |
# python build dependencies \ | |
build-essential | |
RUN useradd -m -u 1000 user | |
USER user | |
ENV HOME=/home/user \ | |
PATH=/home/user/.local/bin:${PATH} | |
WORKDIR ${HOME}/app | |
RUN pip install --no-cache-dir -U pip setuptools wheel && \ | |
pip install "huggingface-hub" "hf-transfer" "gradio[oauth]>=4.28.0" "gradio_huggingfacehub_search==0.0.7" "APScheduler" | |
COPY --chown=1000 . ${HOME}/app | |
# TODO: revert once the PR is merged | |
# RUN git clone https://github.com/ggerganov/llama.cpp --depth 1 | |
RUN git clone https://github.com/ngxson/llama.cpp -b xsn/lora_convert_base_is_optional --depth 1 | |
RUN pip install -r llama.cpp/requirements.txt | |
ENV PYTHONPATH=${HOME}/app \ | |
PYTHONUNBUFFERED=1 \ | |
HF_HUB_ENABLE_HF_TRANSFER=1 \ | |
GRADIO_ALLOW_FLAGGING=never \ | |
GRADIO_NUM_PORTS=1 \ | |
GRADIO_SERVER_NAME=0.0.0.0 \ | |
GRADIO_THEME=huggingface \ | |
TQDM_POSITION=-1 \ | |
TQDM_MININTERVAL=1 \ | |
SYSTEM=spaces \ | |
LD_LIBRARY_PATH=/usr/local/cuda/lib64:${LD_LIBRARY_PATH} \ | |
PATH=/usr/local/nvidia/bin:${PATH} | |
ENTRYPOINT /bin/bash start.sh | |