|
#!/bin/bash |
|
set -o pipefail |
|
set -ex |
|
|
|
export DEBIAN_FRONTEND=noninteractive |
|
export PATH=/h2ogpt_conda/bin:$PATH |
|
export HOME=/workspace |
|
export CUDA_HOME=/usr/local/cuda-11.8 |
|
|
|
|
|
apt-get update && apt-get install -y \ |
|
git \ |
|
curl \ |
|
wget \ |
|
software-properties-common \ |
|
pandoc \ |
|
vim \ |
|
libmagic-dev \ |
|
poppler-utils \ |
|
tesseract-ocr \ |
|
libtesseract-dev \ |
|
libreoffice \ |
|
autoconf \ |
|
libtool |
|
|
|
|
|
apt-get upgrade -y |
|
|
|
|
|
wget https://repo.anaconda.com/miniconda/Miniconda3-py310_23.1.0-1-Linux-x86_64.sh && \ |
|
mkdir -p /h2ogpt_conda && \ |
|
bash ./Miniconda3-py310_23.1.0-1-Linux-x86_64.sh -b -u -p /h2ogpt_conda && \ |
|
conda install python=3.10 pygobject weasyprint -c conda-forge -y |
|
|
|
export PIP_EXTRA_INDEX_URL="https://download.pytorch.org/whl/cu118" |
|
|
|
bash docs/linux_install.sh |
|
|
|
chmod -R a+rwx /h2ogpt_conda |
|
|
|
|
|
export TIKTOKEN_CACHE_DIR=/workspace/tiktoken_cache |
|
python3.10 -c " |
|
import tiktoken |
|
from tiktoken_ext import openai_public |
|
# FakeTokenizer etc. needs tiktoken for general tasks |
|
for enc in openai_public.ENCODING_CONSTRUCTORS: |
|
encoding = tiktoken.get_encoding(enc) |
|
model_encodings = [ |
|
'gpt-4', |
|
'gpt-4-0314', |
|
'gpt-4-32k', |
|
'gpt-4-32k-0314', |
|
'gpt-3.5-turbo', |
|
'gpt-3.5-turbo-16k', |
|
'gpt-3.5-turbo-0301', |
|
'text-ada-001', |
|
'ada', |
|
'text-babbage-001', |
|
'babbage', |
|
'text-curie-001', |
|
'curie', |
|
'davinci', |
|
'text-davinci-003', |
|
'text-davinci-002', |
|
'code-davinci-002', |
|
'code-davinci-001', |
|
'code-cushman-002', |
|
'code-cushman-001' |
|
] |
|
for enc in model_encodings: |
|
encoding = tiktoken.encoding_for_model(enc) |
|
print('Done!') |
|
" |
|
|
|
|
|
|
|
export VLLM_CACHE=/workspace/.vllm_cache |
|
cd /h2ogpt_conda |
|
python -m venv vllm_env --system-site-packages |
|
|
|
|
|
|
|
/h2ogpt_conda/vllm_env/bin/python -m pip install https://github.com/vllm-project/vllm/releases/download/v0.2.7/vllm-0.2.7+cu118-cp310-cp310-manylinux1_x86_64.whl |
|
/h2ogpt_conda/vllm_env/bin/python -m pip install mosaicml-turbo |
|
|
|
/h2ogpt_conda/vllm_env/bin/python -m pip install https://h2o-release.s3.amazonaws.com/h2ogpt/triton-2.1.0-0-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.whl |
|
/h2ogpt_conda/vllm_env/bin/python -m pip install https://h2o-release.s3.amazonaws.com/h2ogpt/megablocks-0.5.0-cp310-cp310-linux_x86_64.whl |
|
|
|
/h2ogpt_conda/vllm_env/bin/python -m pip install ray pandas gputil==1.4.0 fschat==0.2.34 flash-attn==2.4.2 autoawq==0.1.8 ip install uvicorn[standard] |
|
mkdir $VLLM_CACHE |
|
chmod -R a+rwx /h2ogpt_conda |
|
|
|
|
|
mkdir -p /h2ogpt_conda/envs/vllm/bin |
|
ln -s /h2ogpt_conda/vllm_env/bin/python3.10 /h2ogpt_conda/envs/vllm/bin/python3.10 |
|
|
|
|
|
cd /workspace && make build_info.txt git_hash.txt |
|
cp /workspace/build_info.txt /build_info.txt |
|
cp /workspace/git_hash.txt /git_hash.txt |
|
|
|
mkdir -p /workspace/save |
|
chmod -R a+rwx /workspace/save |
|
|
|
|
|
rm -rf /workspace/Miniconda3-py310_23.1.0-1-Linux-x86_64.sh |
|
rm -rf /workspace/.cache/pip |
|
|