gabrielchua commited on
Commit
26ad4ba
1 Parent(s): ef30892

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -8
app.py CHANGED
@@ -1,3 +1,4 @@
 
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
  import torch
@@ -7,22 +8,20 @@ import requests
7
  from openai import OpenAI
8
 
9
  HF_API_KEY = os.getenv("HF_API_KEY")
10
- model_id = "meta-llama/Llama-Guard-3-1B"
11
-
12
 
13
  model = AutoModelForCausalLM.from_pretrained(
14
- model_id,
15
  torch_dtype=torch.bfloat16,
16
  device_map="auto",
17
  token=HF_API_KEY
18
-
19
- )
20
- tokenizer = AutoTokenizer.from_pretrained(model_id, token=HF_API_KEY)
21
 
22
 
23
  # Model 1: Llama Guard Model
 
24
  def llama_guard_moderation(input_text):
25
-
26
  conversation = [
27
  {
28
  "role": "user",
@@ -34,7 +33,6 @@ def llama_guard_moderation(input_text):
34
  ],
35
  }
36
  ]
37
-
38
  input_ids = tokenizer.apply_chat_template(
39
  conversation, return_tensors="pt"
40
  ).to(model.device)
 
1
+ import spaces
2
  import gradio as gr
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
  import torch
 
8
  from openai import OpenAI
9
 
10
  HF_API_KEY = os.getenv("HF_API_KEY")
11
+ MODEL_ID = "meta-llama/Llama-Guard-3-1B"
 
12
 
13
  model = AutoModelForCausalLM.from_pretrained(
14
+ MODEL_ID,
15
  torch_dtype=torch.bfloat16,
16
  device_map="auto",
17
  token=HF_API_KEY
18
+ ).to('cuda')
19
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, token=HF_API_KEY).to('cuda')
 
20
 
21
 
22
  # Model 1: Llama Guard Model
23
+ @spaces.GPU
24
  def llama_guard_moderation(input_text):
 
25
  conversation = [
26
  {
27
  "role": "user",
 
33
  ],
34
  }
35
  ]
 
36
  input_ids = tokenizer.apply_chat_template(
37
  conversation, return_tensors="pt"
38
  ).to(model.device)