File size: 945 Bytes
de81baf
fde41fa
20fd975
de81baf
 
 
 
 
 
fde41fa
cabdd14
fde41fa
cabdd14
 
 
 
20fd975
fde41fa
20fd975
fde41fa
 
 
 
 
 
20fd975
 
fde41fa
20fd975
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
from transformers import AutoModelForCausalLM, AutoTokenizer
fine_tuned_model = "ashioyajotham/results"

# Load the tokenizer for the Falcon 7B model with remote code trust
tokenizer = AutoTokenizer.from_pretrained(fine_tuned_model, trust_remote_code=True)

# Set the padding token to be the same as the end-of-sequence token
tokenizer.pad_token = tokenizer.eos_token

def generate_code(input_text):
    inputs = tokenizer.encode(input_text, return_tensors='pt')
    # Use your fine-tuned model to generate code here
    generated_code = fine_tuned_model.generate(input_text, max_length=100, temperature = .7, do_sample=True, pad_token_id=tokenizer.eos_token_id)
    completion = tokenizer.decode(generated_code[0])

    return completion

import gradio as gr

iface = gr.Interface(
    fn=generate_code,
    inputs="text",
    outputs="text",
    title="Code Generation App",
    description="Generate code from text input."
)

iface.launch()