Xuehao commited on
Commit
73c4782
·
verified ·
1 Parent(s): e738fb2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -23,7 +23,7 @@ uv pip install vllm --torch-backend=auto
23
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
24
  import torch
25
 
26
- MODEL_PATH = "INC4AI/MiniMax-M2.5-int4-mixed-AutoRound"
27
 
28
  model = AutoModelForCausalLM.from_pretrained(
29
  MODEL_PATH,
@@ -50,7 +50,7 @@ print(response)
50
  ### VLLM Usage
51
 
52
  ```bash
53
- vllm serve INC4AI/MiniMax-M2.5-int4-mixed-AutoRound \
54
  --port 7777 \
55
  --host localhost \
56
  --trust-remote-code \
@@ -64,7 +64,7 @@ vllm serve INC4AI/MiniMax-M2.5-int4-mixed-AutoRound \
64
  ## Generate the Model
65
 
66
  ```bash
67
- auto-round --model_name MiniMaxAI/MiniMax-M2.5 --scheme w4a16 --ignore_layers gate --iters 0 --output_dir MiniMax-M2.5-int4-mixed-AutoRound
68
  ```
69
 
70
  ## Ethical Considerations and Limitations
 
23
  from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
24
  import torch
25
 
26
+ MODEL_PATH = "Intel/MiniMax-M2.5-int4-AutoRound"
27
 
28
  model = AutoModelForCausalLM.from_pretrained(
29
  MODEL_PATH,
 
50
  ### VLLM Usage
51
 
52
  ```bash
53
+ vllm serve Intel/MiniMax-M2.5-int4-AutoRound \
54
  --port 7777 \
55
  --host localhost \
56
  --trust-remote-code \
 
64
  ## Generate the Model
65
 
66
  ```bash
67
+ auto-round --model_name MiniMaxAI/MiniMax-M2.5 --scheme w4a16 --ignore_layers gate --iters 0 --output_dir MiniMax-M2.5-int4-AutoRound
68
  ```
69
 
70
  ## Ethical Considerations and Limitations