winglian commited on
Commit
039e2a0
1 Parent(s): 4fde300

bump versions of deps (#1621)

Browse files

* bump versions of deps

* bump transformers too

* fix xformers deps and include s3fs install

Files changed (3) hide show
  1. docker/Dockerfile +1 -1
  2. requirements.txt +8 -8
  3. setup.py +10 -7
docker/Dockerfile CHANGED
@@ -11,7 +11,7 @@ ARG PYTORCH_VERSION="2.1.2"
11
  ENV PYTORCH_VERSION=$PYTORCH_VERSION
12
 
13
  RUN apt-get update && \
14
- apt-get install -y --allow-change-held-packages vim curl nano libnccl2 libnccl-dev
15
 
16
  WORKDIR /workspace
17
 
 
11
  ENV PYTORCH_VERSION=$PYTORCH_VERSION
12
 
13
  RUN apt-get update && \
14
+ apt-get install -y --allow-change-held-packages vim curl nano libnccl2 libnccl-dev s3fs
15
 
16
  WORKDIR /workspace
17
 
requirements.txt CHANGED
@@ -1,22 +1,22 @@
1
  --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
2
  packaging==23.2
3
  peft==0.10.0
4
- transformers @ git+https://github.com/huggingface/transformers.git@43d17c18360ac9c3d3491389328e2fe55fe8f9ce
5
- tokenizers==0.15.0
6
- bitsandbytes==0.43.0
7
- accelerate==0.28.0
8
- deepspeed==0.13.1
9
  pydantic==2.6.3
10
  addict
11
  fire
12
  PyYAML>=6.0
13
  requests
14
- datasets==2.15.0
15
- flash-attn==2.5.5
16
  sentencepiece
17
  wandb
18
  einops
19
- xformers==0.0.22
20
  optimum==1.16.2
21
  hf_transfer
22
  colorama
 
1
  --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/
2
  packaging==23.2
3
  peft==0.10.0
4
+ transformers==4.40.2
5
+ tokenizers==0.19.1
6
+ bitsandbytes==0.43.1
7
+ accelerate==0.30.1
8
+ deepspeed==0.14.2
9
  pydantic==2.6.3
10
  addict
11
  fire
12
  PyYAML>=6.0
13
  requests
14
+ datasets==2.19.1
15
+ flash-attn==2.5.8
16
  sentencepiece
17
  wandb
18
  einops
19
+ xformers==0.0.23.post1
20
  optimum==1.16.2
21
  hf_transfer
22
  colorama
setup.py CHANGED
@@ -30,7 +30,7 @@ def parse_requirements():
30
 
31
  try:
32
  if "Darwin" in platform.system():
33
- _install_requires.pop(_install_requires.index("xformers==0.0.22"))
34
  else:
35
  torch_version = version("torch")
36
  _install_requires.append(f"torch=={torch_version}")
@@ -45,9 +45,12 @@ def parse_requirements():
45
  else:
46
  raise ValueError("Invalid version format")
47
 
48
- if (major, minor) >= (2, 1):
49
- _install_requires.pop(_install_requires.index("xformers==0.0.22"))
50
- _install_requires.append("xformers>=0.0.23")
 
 
 
51
  except PackageNotFoundError:
52
  pass
53
 
@@ -68,13 +71,13 @@ setup(
68
  dependency_links=dependency_links,
69
  extras_require={
70
  "flash-attn": [
71
- "flash-attn==2.5.5",
72
  ],
73
  "fused-dense-lib": [
74
- "fused-dense-lib @ git+https://github.com/Dao-AILab/flash-attention@v2.3.3#subdirectory=csrc/fused_dense_lib",
75
  ],
76
  "deepspeed": [
77
- "deepspeed==0.13.1",
78
  "deepspeed-kernels",
79
  ],
80
  "mamba-ssm": [
 
30
 
31
  try:
32
  if "Darwin" in platform.system():
33
+ _install_requires.pop(_install_requires.index("xformers==0.0.23.post1"))
34
  else:
35
  torch_version = version("torch")
36
  _install_requires.append(f"torch=={torch_version}")
 
45
  else:
46
  raise ValueError("Invalid version format")
47
 
48
+ if (major, minor) >= (2, 3):
49
+ _install_requires.pop(_install_requires.index("xformers==0.0.23.post1"))
50
+ _install_requires.append("xformers>=0.0.26.post1")
51
+ elif (major, minor) >= (2, 2):
52
+ _install_requires.pop(_install_requires.index("xformers==0.0.23.post1"))
53
+ _install_requires.append("xformers>=0.0.25.post1")
54
  except PackageNotFoundError:
55
  pass
56
 
 
71
  dependency_links=dependency_links,
72
  extras_require={
73
  "flash-attn": [
74
+ "flash-attn==2.5.8",
75
  ],
76
  "fused-dense-lib": [
77
+ "fused-dense-lib @ git+https://github.com/Dao-AILab/flash-attention@v2.5.8#subdirectory=csrc/fused_dense_lib",
78
  ],
79
  "deepspeed": [
80
+ "deepspeed==0.14.2",
81
  "deepspeed-kernels",
82
  ],
83
  "mamba-ssm": [