coeuslearning commited on
Commit
c1ded61
1 Parent(s): 1a8de41

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -7,6 +7,8 @@ import spaces
7
  import torch
8
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
9
 
 
 
10
  MAX_MAX_NEW_TOKENS = 2048
11
  DEFAULT_MAX_NEW_TOKENS = 1024
12
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
@@ -31,8 +33,7 @@ if not torch.cuda.is_available():
31
 
32
  if torch.cuda.is_available():
33
  model_id = "meta-llama/Llama-2-7b-chat-hf"
34
- auth_token = "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"
35
- model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16, device_map="auto", use_auth_token=auth_token)
36
  tokenizer = AutoTokenizer.from_pretrained(model_id)
37
  tokenizer.use_default_system_prompt = False
38
 
 
7
  import torch
8
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
9
 
10
+ !huggingface-cli login --token "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"
11
+
12
  MAX_MAX_NEW_TOKENS = 2048
13
  DEFAULT_MAX_NEW_TOKENS = 1024
14
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
 
33
 
34
  if torch.cuda.is_available():
35
  model_id = "meta-llama/Llama-2-7b-chat-hf"
36
+ model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16, device_map="auto")
 
37
  tokenizer = AutoTokenizer.from_pretrained(model_id)
38
  tokenizer.use_default_system_prompt = False
39