Spaces:
Running
on
Zero
Running
on
Zero
""" | |
Configuration file for Frame 0 Laboratory for MIA | |
BAGEL 7B integration with FLUX prompt optimization | |
""" | |
import os | |
import torch | |
from typing import Dict, Any | |
# Application Configuration | |
APP_CONFIG = { | |
"title": "Frame 0 Laboratory for MIA", | |
"description": "Advanced image analysis with BAGEL 7B and FLUX prompt optimization", | |
"version": "2.0.0", | |
"author": "Frame 0 Laboratory for MIA" | |
} | |
# BAGEL Model Configuration | |
BAGEL_CONFIG = { | |
"model_repo": "ByteDance-Seed/BAGEL-7B-MoT", | |
"local_model_path": "./model", | |
"cache_dir": "./model/cache", | |
"download_patterns": ["*.json", "*.safetensors", "*.bin", "*.py", "*.md", "*.txt"], | |
# Model parameters | |
"dtype": torch.bfloat16, | |
"device_map_strategy": "auto", | |
"max_memory_per_gpu": "80GiB", | |
"offload_buffers": True, | |
"force_hooks": True, | |
# Image processing | |
"vae_transform_size": (1024, 512, 16), | |
"vit_transform_size": (980, 224, 14), | |
# Inference parameters | |
"max_new_tokens": 512, | |
"temperature": 0.7, | |
"top_p": 0.9, | |
"do_sample": True | |
} | |
# Device Configuration for ZeroGPU | |
def get_device_config() -> Dict[str, Any]: | |
"""Determine optimal device configuration for BAGEL""" | |
device_config = { | |
"device": "cpu", | |
"use_gpu": False, | |
"gpu_count": 0, | |
"memory_efficient": True | |
} | |
if torch.cuda.is_available(): | |
gpu_count = torch.cuda.device_count() | |
device_config.update({ | |
"device": "cuda", | |
"use_gpu": True, | |
"gpu_count": gpu_count, | |
"gpu_memory_gb": torch.cuda.get_device_properties(0).total_memory / 1e9, | |
"multi_gpu": gpu_count > 1 | |
}) | |
elif hasattr(torch.backends, 'mps') and torch.backends.mps.is_available(): | |
device_config.update({ | |
"device": "mps", | |
"use_gpu": True, | |
"gpu_count": 1 | |
}) | |
return device_config | |
# BAGEL Device Mapping Configuration | |
def get_bagel_device_map(gpu_count: int) -> Dict[str, str]: | |
"""Configure device mapping for BAGEL model""" | |
# Same device modules that need to be on the same GPU | |
same_device_modules = [ | |
'language_model.model.embed_tokens', | |
'time_embedder', | |
'latent_pos_embed', | |
'vae2llm', | |
'llm2vae', | |
'connector', | |
'vit_pos_embed' | |
] | |
device_map = {} | |
if gpu_count == 1: | |
# Single GPU configuration | |
for module in same_device_modules: | |
device_map[module] = "cuda:0" | |
else: | |
# Multi-GPU configuration - keep critical modules on same device | |
first_device = "cuda:0" | |
for module in same_device_modules: | |
device_map[module] = first_device | |
return device_map | |
# Processing Configuration | |
PROCESSING_CONFIG = { | |
"max_image_size": 1024, | |
"image_quality": 95, | |
"supported_formats": [".jpg", ".jpeg", ".png", ".webp"], | |
"batch_size": 1, | |
"timeout_seconds": 120 # Increased for BAGEL processing | |
} | |
# FLUX Prompt Rules | |
FLUX_RULES = { | |
"remove_patterns": [ | |
r',\s*trending on artstation', | |
r',\s*trending on [^,]+', | |
r',\s*\d+k\s*', | |
r',\s*\d+k resolution', | |
r',\s*artstation', | |
r',\s*concept art', | |
r',\s*digital art', | |
r',\s*by greg rutkowski', | |
], | |
"camera_configs": { | |
"portrait": ", Shot on Hasselblad X2D 100C, 90mm f/2.5 lens at f/2.8, professional portrait photography", | |
"landscape": ", Shot on Phase One XT, 40mm f/4 lens at f/8, epic landscape photography", | |
"street": ", Shot on Leica M11, 35mm f/1.4 lens at f/2.8, documentary street photography", | |
"default": ", Shot on Phase One XF IQ4, 80mm f/2.8 lens at f/4, professional photography" | |
}, | |
"lighting_enhancements": { | |
"dramatic": ", dramatic cinematic lighting", | |
"portrait": ", professional studio lighting with subtle rim light", | |
"default": ", masterful natural lighting" | |
} | |
} | |
# Scoring Configuration | |
SCORING_CONFIG = { | |
"max_score": 100, | |
"score_weights": { | |
"prompt_quality": 0.3, | |
"technical_details": 0.25, | |
"artistic_value": 0.25, | |
"flux_optimization": 0.2 | |
}, | |
"grade_thresholds": { | |
95: {"grade": "LEGENDARY", "color": "#059669"}, | |
90: {"grade": "EXCELLENT", "color": "#10b981"}, | |
80: {"grade": "VERY GOOD", "color": "#22c55e"}, | |
70: {"grade": "GOOD", "color": "#f59e0b"}, | |
60: {"grade": "FAIR", "color": "#f97316"}, | |
0: {"grade": "NEEDS WORK", "color": "#ef4444"} | |
} | |
} | |
# Environment Configuration | |
ENVIRONMENT = { | |
"is_spaces": os.getenv("SPACE_ID") is not None, | |
"is_local": os.getenv("SPACE_ID") is None, | |
"log_level": os.getenv("LOG_LEVEL", "INFO"), | |
"debug_mode": os.getenv("DEBUG", "false").lower() == "true", | |
"space_id": os.getenv("SPACE_ID", ""), | |
"space_author": os.getenv("SPACE_AUTHOR_NAME", "") | |
} | |
# BAGEL Inference Prompts | |
BAGEL_PROMPTS = { | |
"image_analysis": "Describe this image in detail, including objects, people, setting, mood, and visual elements:", | |
"flux_prompt": "Generate a detailed FLUX prompt for this image, focusing on photographic and artistic elements:", | |
"detailed_description": "Provide a comprehensive analysis of this image including composition, lighting, colors, and artistic style:", | |
} | |
# Flash Attention Installation Command | |
FLASH_ATTN_INSTALL = { | |
"command": "pip install flash-attn --no-build-isolation", | |
"env": {"FLASH_ATTENTION_SKIP_CUDA_BUILD": "TRUE"}, | |
"shell": True | |
} | |
# Export main configurations | |
__all__ = [ | |
"APP_CONFIG", | |
"BAGEL_CONFIG", | |
"get_device_config", | |
"get_bagel_device_map", | |
"PROCESSING_CONFIG", | |
"FLUX_RULES", | |
"SCORING_CONFIG", | |
"ENVIRONMENT", | |
"BAGEL_PROMPTS", | |
"FLASH_ATTN_INSTALL" | |
] |