falcon_7b_coder / app.py
ashioyajotham's picture
Update app.py
de81baf
raw
history blame
945 Bytes
from transformers import AutoModelForCausalLM, AutoTokenizer
fine_tuned_model = "ashioyajotham/results"
# Load the tokenizer for the Falcon 7B model with remote code trust
tokenizer = AutoTokenizer.from_pretrained(fine_tuned_model, trust_remote_code=True)
# Set the padding token to be the same as the end-of-sequence token
tokenizer.pad_token = tokenizer.eos_token
def generate_code(input_text):
inputs = tokenizer.encode(input_text, return_tensors='pt')
# Use your fine-tuned model to generate code here
generated_code = fine_tuned_model.generate(input_text, max_length=100, temperature = .7, do_sample=True, pad_token_id=tokenizer.eos_token_id)
completion = tokenizer.decode(generated_code[0])
return completion
import gradio as gr
iface = gr.Interface(
fn=generate_code,
inputs="text",
outputs="text",
title="Code Generation App",
description="Generate code from text input."
)
iface.launch()