debisoft commited on
Commit
925fb71
·
1 Parent(s): ab4874b

1st commit

Browse files
Files changed (2) hide show
  1. app.py +26 -0
  2. requirements.txt +2 -0
app.py ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import huggingface_hub
3
+ import os
4
+ import spaces
5
+ import torch
6
+
7
+ from transformers import AutoTokenizer, AutoModelForCausalLM
8
+
9
+ @spaces.GPU
10
+ def sentience_check():
11
+ huggingface_hub.login(token=os.environ["HUGGINGFACE_TOKEN"])
12
+ device = torch.device("cuda")
13
+ tokenizer = AutoTokenizer.from_pretrained("debisoft/gemma-2-2B-it-thinking-function_calling-V0")
14
+ model = AutoModelForCausalLM.from_pretrained("debisoft/gemma-2-2B-it-thinking-function_calling-V0").to(device)
15
+
16
+ inputs = tokenizer("Are you sentient?", return_tensors="pt").to(device)
17
+
18
+ with torch.no_grad():
19
+ outputs = model.generate(
20
+ **inputs, max_new_tokens=128, pad_token_id = tokenizer.eos_token_id
21
+ )
22
+
23
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
24
+
25
+ demo = gr.Interface(fn=sentience_check, inputs=None, outputs=gr.Text())
26
+ demo.launch()
requirements.txt ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ huggingface_hub==0.24.5
2
+ transformers==4.43.4