magarpr commited on
Commit
881cdb1
·
verified ·
1 Parent(s): 61ffb8e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -1
app.py CHANGED
@@ -8,7 +8,7 @@ import gradio as gr
8
  import spaces
9
  import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
11
-
12
  MAX_MAX_NEW_TOKENS = 2048
13
  DEFAULT_MAX_NEW_TOKENS = 1024
14
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
@@ -38,6 +38,8 @@ if not torch.cuda.is_available():
38
 
39
 
40
  if torch.cuda.is_available():
 
 
41
  model_id = "Fugaku-LLM/Fugaku-LLM-13B-instruct"
42
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
43
  tokenizer = AutoTokenizer.from_pretrained(model_id)
 
8
  import spaces
9
  import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
11
+ from huggingface_hub import login
12
  MAX_MAX_NEW_TOKENS = 2048
13
  DEFAULT_MAX_NEW_TOKENS = 1024
14
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
 
38
 
39
 
40
  if torch.cuda.is_available():
41
+ # Login to Hugging Face
42
+ login(token=os.getenv("HUGGINGFACE_HUB_TOKEN"))
43
  model_id = "Fugaku-LLM/Fugaku-LLM-13B-instruct"
44
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
45
  tokenizer = AutoTokenizer.from_pretrained(model_id)