akhaliq HF Staff patrickhubermeta commited on
Commit
a221e28
·
verified ·
1 Parent(s): eb8ec5c

Update to account for MobileLLM-Pro repository split (#1)

Browse files

- Update to account for MobileLLM-Pro repository split (dab7631d3bafef88b7c5376ff04bac0cd38f9cfb)


Co-authored-by: Patrick Huber <[email protected]>

Files changed (1) hide show
  1. app.py +1 -3
app.py CHANGED
@@ -9,7 +9,6 @@ from huggingface_hub import login
9
  import spaces
10
 
11
  MODEL_ID = "facebook/MobileLLM-Pro"
12
- SUBFOLDER = "instruct" # use the chat template
13
  MAX_NEW_TOKENS = 256
14
  TEMPERATURE = 0.7
15
  TOP_P = 0.95
@@ -32,12 +31,11 @@ def _ensure_loaded():
32
  if _tokenizer is not None and _model is not None:
33
  return
34
  _tokenizer = AutoTokenizer.from_pretrained(
35
- MODEL_ID, trust_remote_code=True, subfolder=SUBFOLDER
36
  )
37
  _model = AutoModelForCausalLM.from_pretrained(
38
  MODEL_ID,
39
  trust_remote_code=True,
40
- subfolder=SUBFOLDER,
41
  torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
42
  low_cpu_mem_usage=True,
43
  device_map="auto" if torch.cuda.is_available() else None,
 
9
  import spaces
10
 
11
  MODEL_ID = "facebook/MobileLLM-Pro"
 
12
  MAX_NEW_TOKENS = 256
13
  TEMPERATURE = 0.7
14
  TOP_P = 0.95
 
31
  if _tokenizer is not None and _model is not None:
32
  return
33
  _tokenizer = AutoTokenizer.from_pretrained(
34
+ MODEL_ID, trust_remote_code=True
35
  )
36
  _model = AutoModelForCausalLM.from_pretrained(
37
  MODEL_ID,
38
  trust_remote_code=True,
 
39
  torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
40
  low_cpu_mem_usage=True,
41
  device_map="auto" if torch.cuda.is_available() else None,