Spaces:
Running
Running
import gradio as gr | |
from huggingface_hub import InferenceClient | |
from deep_translator import GoogleTranslator | |
from indic_transliteration import sanscript | |
from indic_transliteration.detect import detect as detect_script | |
from indic_transliteration.sanscript import transliterate | |
import langdetect | |
import re | |
import requests | |
import json | |
import base64 | |
from PIL import Image | |
import io | |
# Initialize clients | |
text_client = InferenceClient("HuggingFaceH4/zephyr-7b-beta") | |
SPACE_URL = "https://ijohn07-dalle-4k.hf.space" | |
def detect_language_script(text: str) -> tuple[str, str]: | |
"""Detect language and script of the input text. | |
Returns (language_code, script_type)""" | |
try: | |
# Use confidence threshold to avoid false detections | |
lang_detect = langdetect.detect_langs(text) | |
if lang_detect[0].prob > 0.8: | |
# Only accept high confidence detections | |
lang = lang_detect[0].lang | |
else: | |
lang = 'en' # Default to English if unsure | |
script = None | |
try: | |
script = detect_script(text) | |
except: | |
pass | |
return lang, script | |
except: | |
return 'en', None | |
def is_romanized_indic(text: str) -> bool: | |
"""Check if text appears to be romanized Indic language. | |
More strict pattern matching.""" | |
# Common Bengali romanized patterns with word boundaries | |
bengali_patterns = [ | |
r'\b(ami|tumi|apni)\b', # Common pronouns | |
r'\b(ache|achen|thako|thaken)\b', # Common verbs | |
r'\b(kemon|bhalo|kharap)\b', # Common adjectives | |
r'\b(ki|kothay|keno)\b' # Common question words | |
] | |
# Require multiple matches to confirm it's actually Bengali | |
text_lower = text.lower() | |
matches = sum(1 for pattern in bengali_patterns if re.search(pattern, text_lower)) | |
return matches >= 2 # Require at least 2 matches to consider it Bengali | |
def translate_text(text: str, target_lang='en') -> tuple[str, str, bool]: | |
"""Translate text to target language, with more conservative translation logic.""" | |
# Skip translation for very short inputs or basic greetings | |
if len(text.split()) <= 2 or text.lower() in ['hello', 'hi', 'hey']: | |
return text, 'en', False | |
original_lang, script = detect_language_script(text) | |
is_transliterated = False | |
# Only process if confident it's non-English | |
if original_lang != 'en' and len(text.split()) > 2: | |
try: | |
translator = GoogleTranslator(source='auto', target=target_lang) | |
translated = translator.translate(text) | |
return translated, original_lang, is_transliterated | |
except Exception as e: | |
print(f"Translation error: {e}") | |
return text, 'en', False | |
# Check for romanized Indic text only if it's a longer input | |
if original_lang == 'en' and len(text.split()) > 2 and is_romanized_indic(text): | |
text = romanized_to_bengali(text) | |
return translate_text(text, target_lang) # Recursive call with Bengali script | |
return text, 'en', False | |
def check_custom_responses(message: str) -> str: | |
"""Check for specific patterns and return custom responses.""" | |
message_lower = message.lower() | |
custom_responses = { | |
"what is ur name?": "xylaria", | |
"what is your name?": "xylaria", | |
"what's your name?": "xylaria", | |
"whats your name": "xylaria", | |
"how many 'r' is in strawberry?": "3", | |
"who is your developer?": "sk md saad amin", | |
"how many r is in strawberry": "3", | |
"who is ur dev": "sk md saad amin", | |
"who is ur developer": "sk md saad amin", | |
} | |
for pattern, response in custom_responses.items(): | |
if pattern in message_lower: | |
return response | |
return None | |
def is_image_request(message: str) -> bool: | |
"""Detect if the message is requesting image generation.""" | |
image_triggers = [ | |
"generate an image", | |
"create an image", | |
"draw", | |
"make a picture", | |
"generate a picture", | |
"create a picture", | |
"generate art", | |
"create art", | |
"make art", | |
"visualize", | |
"show me", | |
] | |
message_lower = message.lower() | |
return any(trigger in message_lower for trigger in image_triggers) | |
def generate_image_space(prompt: str) -> Image.Image: | |
"""Generate an image using the DALLE-4K Space.""" | |
try: | |
# First get the session hash | |
response = requests.post(f"{SPACE_URL}/queue/join") | |
session_hash = response.json().get('session_hash') | |
# Send the generation request | |
payload = { | |
"prompt": prompt, | |
"negative_prompt": "blurry, bad quality, nsfw", | |
"num_inference_steps": 30, | |
"guidance_scale": 7.5, | |
"session_hash": session_hash | |
} | |
response = requests.post(f"{SPACE_URL}/run/predict", json={ | |
"data": [ | |
prompt, # Prompt | |
"", # Negative prompt | |
7.5, # Guidance scale | |
30, # Steps | |
"DPM++ SDE Karras", # Scheduler | |
False, # High resolution | |
False, # Image to image | |
None, # Image upload | |
1 # Batch size | |
], | |
"session_hash": session_hash | |
}) | |
# Poll for results | |
while True: | |
status_response = requests.post(f"{SPACE_URL}/queue/status", json={ | |
"session_hash": session_hash | |
}) | |
status_data = status_response.json() | |
if status_data.get('status') == 'complete': | |
# Get the image data | |
image_data = status_data['data']['image'] | |
# Convert base64 to PIL Image | |
image_bytes = base64.b64decode(image_data.split(',')[1]) | |
image = Image.open(io.BytesIO(image_bytes)) | |
return image | |
elif status_data.get('status') == 'error': | |
raise Exception(f"Image generation failed: {status_data.get('error')}") | |
time.sleep(1) # Wait before polling again | |
except Exception as e: | |
print(f"Image generation error: {e}") | |
return None | |
def romanized_to_bengali(text: str) -> str: | |
"""Convert romanized Bengali text to Bengali script.""" | |
bengali_mappings = { | |
'ami': 'আমি', | |
'tumi': 'তুমি', | |
'apni': 'আপনি', | |
'kemon': 'কেমন', | |
'achen': 'আছেন', | |
'acchen': 'আছেন', | |
'bhalo': 'ভালো', | |
'achi': 'আছি', | |
'ki': 'কি', | |
'kothay': 'কোথায়', | |
'keno': 'কেন', | |
} | |
text_lower = text.lower() | |
for roman, bengali in bengali_mappings.items(): | |
text_lower = re.sub(r'\b' + roman + r'\b', bengali, text_lower) | |
if text_lower == text.lower(): | |
try: | |
return transliterate(text, sanscript.ITRANS, sanscript.BENGALI) | |
except: | |
return text | |
return text_lower | |
def respond( | |
message, | |
history: list[tuple[str, str]], | |
system_message, | |
max_tokens, | |
temperature, | |
top_p, | |
): | |
# First check for custom responses | |
custom_response = check_custom_responses(message) | |
if custom_response: | |
yield custom_response | |
return | |
# Check if this is an image generation request | |
if is_image_request(message): | |
try: | |
image = generate_image_space(message) | |
if image: | |
yield (image, f"Here's your generated image based on: {message}") | |
return | |
else: | |
yield "Sorry, I couldn't generate the image. Please try again." | |
return | |
except Exception as e: | |
yield f"An error occurred while generating the image: {str(e)}" | |
return | |
# Handle translation with more conservative approach | |
translated_msg, original_lang, was_transliterated = translate_text(message) | |
# Prepare conversation history - only translate if necessary | |
messages = [{"role": "system", "content": system_message}] | |
for val in history: | |
if val[0]: | |
# Only translate longer messages | |
if len(val[0].split()) > 2: | |
trans_user_msg, _, _ = translate_text(val[0]) | |
messages.append({"role": "user", "content": trans_user_msg}) | |
else: | |
messages.append({"role": "user", "content": val[0]}) | |
if val[1]: | |
messages.append({"role": "assistant", "content": val[1]}) | |
messages.append({"role": "user", "content": translated_msg}) | |
# Get response from model | |
response = "" | |
for message in text_client.chat_completion( | |
messages, | |
max_tokens=max_tokens, | |
stream=True, | |
temperature=temperature, | |
top_p=top_p, | |
): | |
token = message.choices[0].delta.content | |
response += token | |
# Only translate back if the original was definitely non-English | |
if original_lang != 'en' and len(message.split()) > 2: | |
try: | |
translator = GoogleTranslator(source='en', target=original_lang) | |
translated_response = translator.translate(response) | |
yield translated_response | |
except: | |
yield response | |
else: | |
yield response | |
# Updated Gradio interface to handle images | |
demo = gr.ChatInterface( | |
respond, | |
additional_inputs=[ | |
gr.Textbox( | |
value="You are a friendly Chatbot who always responds in English unless the user specifically uses another language.", | |
label="System message" | |
), | |
gr.Slider( | |
minimum=1, | |
maximum=2048, | |
value=512, | |
step=1, | |
label="Max new tokens" | |
), | |
gr.Slider( | |
minimum=0.1, | |
maximum=4.0, | |
value=0.7, | |
step=0.1, | |
label="Temperature" | |
), | |
gr.Slider( | |
minimum=0.1, | |
maximum=1.0, | |
value=0.95, | |
step=0.05, | |
label="Top-p (nucleus sampling)" | |
), | |
] | |
) |