Create generation_engine.py
Browse files- generation_engine.py +850 -0
generation_engine.py
ADDED
|
@@ -0,0 +1,850 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
"""
|
| 2 |
+
Core code generation engine with LLM interactions and advanced generation logic.
|
| 3 |
+
"""
|
| 4 |
+
|
| 5 |
+
import uuid
|
| 6 |
+
import time
|
| 7 |
+
import re
|
| 8 |
+
from typing import Dict, List, Optional, Tuple, Generator, Any
|
| 9 |
+
import gradio as gr
|
| 10 |
+
|
| 11 |
+
from utils import get_inference_client, remove_code_block, extract_text_from_file,
|
| 12 |
+
create_multimodal_message, apply_search_replace_changes, cleanup_session_media, reap_old_media
|
| 13 |
+
from web_utils import extract_website_content, enhance_query_with_search
|
| 14 |
+
from code_processing import (
|
| 15 |
+
is_streamlit_code, is_gradio_code, extract_html_document,
|
| 16 |
+
parse_transformers_js_output, format_transformers_js_output, build_transformers_inline_html,
|
| 17 |
+
parse_svelte_output, format_svelte_output,
|
| 18 |
+
parse_multipage_html_output, format_multipage_output, validate_and_autofix_files,
|
| 19 |
+
inline_multipage_into_single_preview, apply_generated_media_to_html
|
| 20 |
+
)
|
| 21 |
+
from media_generation import MediaGenerator
|
| 22 |
+
from config import (
|
| 23 |
+
HTML_SYSTEM_PROMPT, TRANSFORMERS_JS_SYSTEM_PROMPT, SVELTE_SYSTEM_PROMPT, GENERIC_SYSTEM_PROMPT,
|
| 24 |
+
SEARCH_START, DIVIDER, REPLACE_END, TEMP_DIR_TTL_SECONDS
|
| 25 |
+
)
|
| 26 |
+
|
| 27 |
+
class GenerationEngine:
|
| 28 |
+
"""Advanced code generation engine with multi-model support and intelligent processing"""
|
| 29 |
+
|
| 30 |
+
def __init__(self):
|
| 31 |
+
self.media_generator = MediaGenerator()
|
| 32 |
+
self._active_generations = {}
|
| 33 |
+
self._generation_stats = {
|
| 34 |
+
'total_requests': 0,
|
| 35 |
+
'successful_generations': 0,
|
| 36 |
+
'errors': 0,
|
| 37 |
+
'avg_response_time': 0.0
|
| 38 |
+
}
|
| 39 |
+
|
| 40 |
+
def generate_code(self,
|
| 41 |
+
query: Optional[str] = None,
|
| 42 |
+
vlm_image: Optional[gr.Image] = None,
|
| 43 |
+
gen_image: Optional[gr.Image] = None,
|
| 44 |
+
file: Optional[str] = None,
|
| 45 |
+
website_url: Optional[str] = None,
|
| 46 |
+
settings: Dict[str, Any] = None,
|
| 47 |
+
history: Optional[List[Tuple[str, str]]] = None,
|
| 48 |
+
current_model: Dict = None,
|
| 49 |
+
enable_search: bool = False,
|
| 50 |
+
language: str = "html",
|
| 51 |
+
provider: str = "auto",
|
| 52 |
+
**media_options) -> Generator[Dict[str, Any], None, None]:
|
| 53 |
+
"""
|
| 54 |
+
Main code generation method with comprehensive options and streaming support
|
| 55 |
+
"""
|
| 56 |
+
start_time = time.time()
|
| 57 |
+
session_id = str(uuid.uuid4())
|
| 58 |
+
|
| 59 |
+
try:
|
| 60 |
+
self._active_generations[session_id] = {
|
| 61 |
+
'start_time': start_time,
|
| 62 |
+
'status': 'initializing',
|
| 63 |
+
'progress': 0
|
| 64 |
+
}
|
| 65 |
+
|
| 66 |
+
# Initialize and validate inputs
|
| 67 |
+
query = query or ''
|
| 68 |
+
history = history or []
|
| 69 |
+
settings = settings or {}
|
| 70 |
+
current_model = current_model or {'id': 'Qwen/Qwen3-Coder-480B-A35B-Instruct', 'name': 'Qwen3-Coder'}
|
| 71 |
+
|
| 72 |
+
# Update statistics
|
| 73 |
+
self._generation_stats['total_requests'] += 1
|
| 74 |
+
|
| 75 |
+
# Cleanup old resources
|
| 76 |
+
self._cleanup_resources(session_id)
|
| 77 |
+
|
| 78 |
+
# Determine if this is a modification request
|
| 79 |
+
has_existing_content = self._check_existing_content(history)
|
| 80 |
+
|
| 81 |
+
# Handle modification requests with search/replace
|
| 82 |
+
if has_existing_content and query.strip():
|
| 83 |
+
yield from self._handle_modification_request(query, history, current_model, provider, session_id)
|
| 84 |
+
return
|
| 85 |
+
|
| 86 |
+
# Process file inputs and website content
|
| 87 |
+
enhanced_query = self._process_inputs(query, file, website_url, enable_search)
|
| 88 |
+
|
| 89 |
+
# Select appropriate system prompt
|
| 90 |
+
system_prompt = self._select_system_prompt(language, enable_search, has_existing_content)
|
| 91 |
+
|
| 92 |
+
# Prepare messages for LLM
|
| 93 |
+
messages = self._prepare_messages(history, system_prompt, enhanced_query, vlm_image)
|
| 94 |
+
|
| 95 |
+
# Generate code with streaming
|
| 96 |
+
yield from self._stream_generation(
|
| 97 |
+
messages, current_model, provider, language,
|
| 98 |
+
enhanced_query, gen_image, session_id, media_options
|
| 99 |
+
)
|
| 100 |
+
|
| 101 |
+
# Update success statistics
|
| 102 |
+
self._generation_stats['successful_generations'] += 1
|
| 103 |
+
elapsed_time = time.time() - start_time
|
| 104 |
+
self._update_avg_response_time(elapsed_time)
|
| 105 |
+
|
| 106 |
+
except Exception as e:
|
| 107 |
+
self._generation_stats['errors'] += 1
|
| 108 |
+
error_message = f"Generation Error: {str(e)}"
|
| 109 |
+
print(f"[GenerationEngine] Error: {error_message}")
|
| 110 |
+
|
| 111 |
+
yield {
|
| 112 |
+
'code_output': error_message,
|
| 113 |
+
'history_output': self._convert_history_to_messages(history),
|
| 114 |
+
'sandbox': f"<div style='padding:2rem;text-align:center;color:#dc2626;background:#fef2f2;border:1px solid #fecaca;border-radius:8px;'><h3>Generation Failed</h3><p>{error_message}</p></div>",
|
| 115 |
+
'status': 'error'
|
| 116 |
+
}
|
| 117 |
+
finally:
|
| 118 |
+
# Cleanup generation tracking
|
| 119 |
+
self._active_generations.pop(session_id, None)
|
| 120 |
+
|
| 121 |
+
def _cleanup_resources(self, session_id: str):
|
| 122 |
+
"""Clean up temporary resources"""
|
| 123 |
+
try:
|
| 124 |
+
cleanup_session_media(session_id)
|
| 125 |
+
reap_old_media()
|
| 126 |
+
except Exception as e:
|
| 127 |
+
print(f"[GenerationEngine] Cleanup warning: {e}")
|
| 128 |
+
|
| 129 |
+
def _check_existing_content(self, history: List[Tuple[str, str]]) -> bool:
|
| 130 |
+
"""Check if there's existing content to modify"""
|
| 131 |
+
if not history:
|
| 132 |
+
return False
|
| 133 |
+
|
| 134 |
+
last_assistant_msg = history[-1][1] if history else ""
|
| 135 |
+
content_indicators = [
|
| 136 |
+
'<!DOCTYPE html>', '<html', 'import gradio', 'import streamlit',
|
| 137 |
+
'def ', 'IMPORTED PROJECT FROM HUGGING FACE SPACE',
|
| 138 |
+
'=== index.html ===', '=== index.js ===', '=== style.css ==='
|
| 139 |
+
]
|
| 140 |
+
|
| 141 |
+
return any(indicator in last_assistant_msg for indicator in content_indicators)
|
| 142 |
+
|
| 143 |
+
def _handle_modification_request(self, query: str, history: List[Tuple[str, str]],
|
| 144 |
+
current_model: Dict, provider: str, session_id: str) -> Generator:
|
| 145 |
+
"""Handle search/replace modification requests"""
|
| 146 |
+
try:
|
| 147 |
+
print("[GenerationEngine] Processing modification request")
|
| 148 |
+
|
| 149 |
+
client = get_inference_client(current_model['id'], provider)
|
| 150 |
+
last_assistant_msg = history[-1][1] if history else ""
|
| 151 |
+
|
| 152 |
+
# Create search/replace system prompt
|
| 153 |
+
system_prompt = f"""You are a code editor assistant. Generate EXACT search/replace blocks for the requested modifications.
|
| 154 |
+
|
| 155 |
+
CRITICAL REQUIREMENTS:
|
| 156 |
+
1. Use EXACTLY these markers: {SEARCH_START}, {DIVIDER}, {REPLACE_END}
|
| 157 |
+
2. The SEARCH block must match existing code EXACTLY (including whitespace)
|
| 158 |
+
3. Generate multiple blocks if needed for different changes
|
| 159 |
+
4. Only include specific lines that need to change with sufficient context
|
| 160 |
+
5. DO NOT include explanations outside the blocks
|
| 161 |
+
|
| 162 |
+
Example:
|
| 163 |
+
{SEARCH_START}
|
| 164 |
+
<h1>Old Title</h1>
|
| 165 |
+
{DIVIDER}
|
| 166 |
+
<h1>New Title</h1>
|
| 167 |
+
{REPLACE_END}"""
|
| 168 |
+
|
| 169 |
+
user_prompt = f"""Existing code:
|
| 170 |
+
{last_assistant_msg}
|
| 171 |
+
|
| 172 |
+
Modification request:
|
| 173 |
+
{query}
|
| 174 |
+
|
| 175 |
+
Generate the exact search/replace blocks needed."""
|
| 176 |
+
|
| 177 |
+
messages = [
|
| 178 |
+
{"role": "system", "content": system_prompt},
|
| 179 |
+
{"role": "user", "content": user_prompt}
|
| 180 |
+
]
|
| 181 |
+
|
| 182 |
+
# Generate modification instructions
|
| 183 |
+
response = self._call_llm(client, current_model, messages, max_tokens=4000, temperature=0.1)
|
| 184 |
+
|
| 185 |
+
if response:
|
| 186 |
+
# Apply changes
|
| 187 |
+
if '=== index.html ===' in last_assistant_msg:
|
| 188 |
+
modified_content = self._apply_transformers_js_changes(last_assistant_msg, response)
|
| 189 |
+
else:
|
| 190 |
+
modified_content = apply_search_replace_changes(last_assistant_msg, response)
|
| 191 |
+
|
| 192 |
+
if modified_content != last_assistant_msg:
|
| 193 |
+
updated_history = history + [(query, modified_content)]
|
| 194 |
+
|
| 195 |
+
yield {
|
| 196 |
+
'code_output': modified_content,
|
| 197 |
+
'history': updated_history,
|
| 198 |
+
'sandbox': self._generate_preview(modified_content, "html"),
|
| 199 |
+
'history_output': self._convert_history_to_messages(updated_history),
|
| 200 |
+
'status': 'completed'
|
| 201 |
+
}
|
| 202 |
+
return
|
| 203 |
+
|
| 204 |
+
# Fallback to normal generation if modification failed
|
| 205 |
+
print("[GenerationEngine] Search/replace failed, falling back to normal generation")
|
| 206 |
+
|
| 207 |
+
except Exception as e:
|
| 208 |
+
print(f"[GenerationEngine] Modification request failed: {e}")
|
| 209 |
+
|
| 210 |
+
def _process_inputs(self, query: str, file: Optional[str], website_url: Optional[str],
|
| 211 |
+
enable_search: bool) -> str:
|
| 212 |
+
"""Process file and website inputs, enhance with search if enabled"""
|
| 213 |
+
enhanced_query = query
|
| 214 |
+
|
| 215 |
+
# Process file input
|
| 216 |
+
if file:
|
| 217 |
+
file_text = extract_text_from_file(file)
|
| 218 |
+
if file_text:
|
| 219 |
+
file_text = file_text[:5000] # Limit size
|
| 220 |
+
enhanced_query = f"{enhanced_query}\n\n[Reference file content]\n{file_text}"
|
| 221 |
+
|
| 222 |
+
# Process website URL
|
| 223 |
+
if website_url and website_url.strip():
|
| 224 |
+
website_text = extract_website_content(website_url.strip())
|
| 225 |
+
if website_text and not website_text.startswith("Error"):
|
| 226 |
+
website_text = website_text[:8000] # Limit size
|
| 227 |
+
enhanced_query = f"{enhanced_query}\n\n[Website content to redesign]\n{website_text}"
|
| 228 |
+
elif website_text.startswith("Error"):
|
| 229 |
+
fallback_guidance = """
|
| 230 |
+
Since I couldn't extract the website content, please provide:
|
| 231 |
+
1. What type of website is this?
|
| 232 |
+
2. What are the main features you want?
|
| 233 |
+
3. What's the target audience?
|
| 234 |
+
4. Any specific design preferences?"""
|
| 235 |
+
enhanced_query = f"{enhanced_query}\n\n[Website extraction error: {website_text}]{fallback_guidance}"
|
| 236 |
+
|
| 237 |
+
# Enhance with web search
|
| 238 |
+
if enable_search:
|
| 239 |
+
enhanced_query = enhance_query_with_search(enhanced_query, True)
|
| 240 |
+
|
| 241 |
+
return enhanced_query
|
| 242 |
+
|
| 243 |
+
def _select_system_prompt(self, language: str, enable_search: bool, has_existing_content: bool) -> str:
|
| 244 |
+
"""Select appropriate system prompt based on context"""
|
| 245 |
+
if has_existing_content:
|
| 246 |
+
return self._get_followup_system_prompt(language)
|
| 247 |
+
|
| 248 |
+
# Add search enhancement to prompts if enabled
|
| 249 |
+
search_enhancement = """
|
| 250 |
+
|
| 251 |
+
Use web search results when available to incorporate the latest best practices, frameworks, and technologies.""" if enable_search else ""
|
| 252 |
+
|
| 253 |
+
if language == "html":
|
| 254 |
+
return HTML_SYSTEM_PROMPT + search_enhancement
|
| 255 |
+
elif language == "transformers.js":
|
| 256 |
+
return TRANSFORMERS_JS_SYSTEM_PROMPT + search_enhancement
|
| 257 |
+
elif language == "svelte":
|
| 258 |
+
return SVELTE_SYSTEM_PROMPT + search_enhancement
|
| 259 |
+
else:
|
| 260 |
+
return GENERIC_SYSTEM_PROMPT.format(language=language) + search_enhancement
|
| 261 |
+
|
| 262 |
+
def _get_followup_system_prompt(self, language: str) -> str:
|
| 263 |
+
"""Get follow-up system prompt for modifications"""
|
| 264 |
+
return f"""You are an expert developer modifying existing {language} code.
|
| 265 |
+
Apply the requested changes using SEARCH/REPLACE blocks with these markers:
|
| 266 |
+
{SEARCH_START}, {DIVIDER}, {REPLACE_END}
|
| 267 |
+
|
| 268 |
+
Requirements:
|
| 269 |
+
- SEARCH blocks must match existing code EXACTLY
|
| 270 |
+
- Provide multiple blocks for different changes
|
| 271 |
+
- Include sufficient context to make matches unique
|
| 272 |
+
- Do not include explanations outside the blocks"""
|
| 273 |
+
|
| 274 |
+
def _prepare_messages(self, history: List[Tuple[str, str]], system_prompt: str,
|
| 275 |
+
enhanced_query: str, vlm_image: Optional[gr.Image]) -> List[Dict]:
|
| 276 |
+
"""Prepare messages for LLM interaction"""
|
| 277 |
+
messages = [{'role': 'system', 'content': system_prompt}]
|
| 278 |
+
|
| 279 |
+
# Add history
|
| 280 |
+
for user_msg, assistant_msg in history:
|
| 281 |
+
# Handle multimodal content in history
|
| 282 |
+
if isinstance(user_msg, list):
|
| 283 |
+
text_content = ""
|
| 284 |
+
for item in user_msg:
|
| 285 |
+
if isinstance(item, dict) and item.get("type") == "text":
|
| 286 |
+
text_content += item.get("text", "")
|
| 287 |
+
user_msg = text_content if text_content else str(user_msg)
|
| 288 |
+
|
| 289 |
+
messages.append({'role': 'user', 'content': user_msg})
|
| 290 |
+
messages.append({'role': 'assistant', 'content': assistant_msg})
|
| 291 |
+
|
| 292 |
+
# Add current query
|
| 293 |
+
if vlm_image is not None:
|
| 294 |
+
messages.append(create_multimodal_message(enhanced_query, vlm_image))
|
| 295 |
+
else:
|
| 296 |
+
messages.append({'role': 'user', 'content': enhanced_query})
|
| 297 |
+
|
| 298 |
+
return messages
|
| 299 |
+
|
| 300 |
+
def _stream_generation(self, messages: List[Dict], current_model: Dict, provider: str,
|
| 301 |
+
language: str, query: str, gen_image: Optional[gr.Image],
|
| 302 |
+
session_id: str, media_options: Dict) -> Generator:
|
| 303 |
+
"""Stream code generation with real-time updates"""
|
| 304 |
+
|
| 305 |
+
try:
|
| 306 |
+
client = get_inference_client(current_model['id'], provider)
|
| 307 |
+
|
| 308 |
+
# Handle special model cases
|
| 309 |
+
if current_model["id"] == "zai-org/GLM-4.5":
|
| 310 |
+
yield from self._handle_glm_45_generation(client, messages, language, query, gen_image, session_id, media_options)
|
| 311 |
+
return
|
| 312 |
+
elif current_model["id"] == "zai-org/GLM-4.5V":
|
| 313 |
+
yield from self._handle_glm_45v_generation(client, messages, language, query, session_id, media_options)
|
| 314 |
+
return
|
| 315 |
+
|
| 316 |
+
# Standard streaming generation
|
| 317 |
+
completion = self._create_completion_stream(client, current_model, messages)
|
| 318 |
+
content = ""
|
| 319 |
+
|
| 320 |
+
# Process stream with intelligent updates
|
| 321 |
+
for chunk in completion:
|
| 322 |
+
chunk_content = self._extract_chunk_content(chunk, current_model)
|
| 323 |
+
|
| 324 |
+
if chunk_content:
|
| 325 |
+
content += chunk_content
|
| 326 |
+
|
| 327 |
+
# Yield periodic updates based on language type
|
| 328 |
+
if language == "transformers.js":
|
| 329 |
+
yield from self._handle_transformers_streaming(content)
|
| 330 |
+
elif language == "svelte":
|
| 331 |
+
yield from self._handle_svelte_streaming(content)
|
| 332 |
+
else:
|
| 333 |
+
yield from self._handle_standard_streaming(content, language)
|
| 334 |
+
|
| 335 |
+
# Final processing with media integration
|
| 336 |
+
final_content = self._finalize_content(content, language, query, gen_image, session_id, media_options)
|
| 337 |
+
|
| 338 |
+
yield {
|
| 339 |
+
'code_output': final_content,
|
| 340 |
+
'history': [(query, final_content)],
|
| 341 |
+
'sandbox': self._generate_preview(final_content, language),
|
| 342 |
+
'history_output': self._convert_history_to_messages([(query, final_content)]),
|
| 343 |
+
'status': 'completed'
|
| 344 |
+
}
|
| 345 |
+
|
| 346 |
+
except Exception as e:
|
| 347 |
+
raise Exception(f"Streaming generation failed: {str(e)}")
|
| 348 |
+
|
| 349 |
+
def _handle_glm_45_generation(self, client, messages, language, query, gen_image, session_id, media_options):
|
| 350 |
+
"""Handle GLM-4.5 specific generation"""
|
| 351 |
+
try:
|
| 352 |
+
stream = client.chat.completions.create(
|
| 353 |
+
model="zai-org/GLM-4.5",
|
| 354 |
+
messages=messages,
|
| 355 |
+
stream=True,
|
| 356 |
+
max_tokens=16384,
|
| 357 |
+
)
|
| 358 |
+
|
| 359 |
+
content = ""
|
| 360 |
+
for chunk in stream:
|
| 361 |
+
if chunk.choices[0].delta.content:
|
| 362 |
+
content += chunk.choices[0].delta.content
|
| 363 |
+
clean_code = remove_code_block(content)
|
| 364 |
+
|
| 365 |
+
yield {
|
| 366 |
+
'code_output': gr.update(value=clean_code, language=self._get_gradio_language(language)),
|
| 367 |
+
'sandbox': self._generate_preview(clean_code, language),
|
| 368 |
+
'status': 'streaming'
|
| 369 |
+
}
|
| 370 |
+
|
| 371 |
+
# Apply media generation
|
| 372 |
+
final_content = apply_generated_media_to_html(
|
| 373 |
+
clean_code, query, session_id=session_id, **media_options
|
| 374 |
+
)
|
| 375 |
+
|
| 376 |
+
yield {
|
| 377 |
+
'code_output': final_content,
|
| 378 |
+
'history': [(query, final_content)],
|
| 379 |
+
'sandbox': self._generate_preview(final_content, language),
|
| 380 |
+
'history_output': self._convert_history_to_messages([(query, final_content)]),
|
| 381 |
+
'status': 'completed'
|
| 382 |
+
}
|
| 383 |
+
|
| 384 |
+
except Exception as e:
|
| 385 |
+
raise Exception(f"GLM-4.5 generation failed: {str(e)}")
|
| 386 |
+
|
| 387 |
+
def _handle_glm_45v_generation(self, client, messages, language, query, session_id, media_options):
|
| 388 |
+
"""Handle GLM-4.5V multimodal generation"""
|
| 389 |
+
try:
|
| 390 |
+
# Enhanced system prompt for multimodal
|
| 391 |
+
enhanced_messages = [
|
| 392 |
+
{"role": "system", "content": """You are an expert web developer creating modern, responsive applications.
|
| 393 |
+
|
| 394 |
+
Output complete, standalone HTML documents that render directly in browsers.
|
| 395 |
+
- Include proper DOCTYPE, head, and body structure
|
| 396 |
+
- Use modern CSS frameworks and responsive design
|
| 397 |
+
- Ensure accessibility and mobile compatibility
|
| 398 |
+
- Output raw HTML without escape characters
|
| 399 |
+
|
| 400 |
+
Always output only the HTML code inside ```html ... ``` blocks."""}
|
| 401 |
+
] + messages[1:] # Skip original system message
|
| 402 |
+
|
| 403 |
+
stream = client.chat.completions.create(
|
| 404 |
+
model="zai-org/GLM-4.5V",
|
| 405 |
+
messages=enhanced_messages,
|
| 406 |
+
stream=True,
|
| 407 |
+
max_tokens=16384,
|
| 408 |
+
)
|
| 409 |
+
|
| 410 |
+
content = ""
|
| 411 |
+
for chunk in stream:
|
| 412 |
+
if hasattr(chunk, "choices") and chunk.choices and hasattr(chunk.choices[0], "delta"):
|
| 413 |
+
delta_content = getattr(chunk.choices[0].delta, "content", None)
|
| 414 |
+
if delta_content:
|
| 415 |
+
content += delta_content
|
| 416 |
+
clean_code = remove_code_block(content)
|
| 417 |
+
|
| 418 |
+
# Handle escaped characters
|
| 419 |
+
if "\\n" in clean_code:
|
| 420 |
+
clean_code = clean_code.replace("\\n", "\n")
|
| 421 |
+
if "\\t" in clean_code:
|
| 422 |
+
clean_code = clean_code.replace("\\t", "\t")
|
| 423 |
+
|
| 424 |
+
yield {
|
| 425 |
+
'code_output': gr.update(value=clean_code, language=self._get_gradio_language(language)),
|
| 426 |
+
'sandbox': self._generate_preview(clean_code, language),
|
| 427 |
+
'status': 'streaming'
|
| 428 |
+
}
|
| 429 |
+
|
| 430 |
+
# Clean final content
|
| 431 |
+
clean_code = remove_code_block(content)
|
| 432 |
+
if "\\n" in clean_code:
|
| 433 |
+
clean_code = clean_code.replace("\\n", "\n")
|
| 434 |
+
if "\\t" in clean_code:
|
| 435 |
+
clean_code = clean_code.replace("\\t", "\t")
|
| 436 |
+
|
| 437 |
+
yield {
|
| 438 |
+
'code_output': clean_code,
|
| 439 |
+
'history': [(query, clean_code)],
|
| 440 |
+
'sandbox': self._generate_preview(clean_code, language),
|
| 441 |
+
'history_output': self._convert_history_to_messages([(query, clean_code)]),
|
| 442 |
+
'status': 'completed'
|
| 443 |
+
}
|
| 444 |
+
|
| 445 |
+
except Exception as e:
|
| 446 |
+
raise Exception(f"GLM-4.5V generation failed: {str(e)}")
|
| 447 |
+
|
| 448 |
+
def _create_completion_stream(self, client, current_model, messages):
|
| 449 |
+
"""Create completion stream based on model type"""
|
| 450 |
+
if current_model["id"] in ("codestral-2508", "mistral-medium-2508"):
|
| 451 |
+
return client.chat.stream(
|
| 452 |
+
model=current_model["id"],
|
| 453 |
+
messages=messages,
|
| 454 |
+
max_tokens=16384
|
| 455 |
+
)
|
| 456 |
+
elif current_model["id"] in ("gpt-5", "grok-4", "claude-opus-4.1"):
|
| 457 |
+
model_name_map = {
|
| 458 |
+
"gpt-5": "GPT-5",
|
| 459 |
+
"grok-4": "Grok-4",
|
| 460 |
+
"claude-opus-4.1": "Claude-Opus-4.1"
|
| 461 |
+
}
|
| 462 |
+
return client.chat.completions.create(
|
| 463 |
+
model=model_name_map[current_model["id"]],
|
| 464 |
+
messages=messages,
|
| 465 |
+
stream=True,
|
| 466 |
+
max_tokens=16384
|
| 467 |
+
)
|
| 468 |
+
else:
|
| 469 |
+
return client.chat.completions.create(
|
| 470 |
+
model=current_model["id"],
|
| 471 |
+
messages=messages,
|
| 472 |
+
stream=True,
|
| 473 |
+
max_tokens=16384
|
| 474 |
+
)
|
| 475 |
+
|
| 476 |
+
def _extract_chunk_content(self, chunk, current_model) -> Optional[str]:
|
| 477 |
+
"""Extract content from stream chunk based on model format"""
|
| 478 |
+
try:
|
| 479 |
+
if current_model["id"] in ("codestral-2508", "mistral-medium-2508"):
|
| 480 |
+
# Mistral format
|
| 481 |
+
if (hasattr(chunk, "data") and chunk.data and
|
| 482 |
+
hasattr(chunk.data, "choices") and chunk.data.choices and
|
| 483 |
+
hasattr(chunk.data.choices[0], "delta") and
|
| 484 |
+
hasattr(chunk.data.choices[0].delta, "content")):
|
| 485 |
+
return chunk.data.choices[0].delta.content
|
| 486 |
+
else:
|
| 487 |
+
# OpenAI format
|
| 488 |
+
if (hasattr(chunk, "choices") and chunk.choices and
|
| 489 |
+
hasattr(chunk.choices[0], "delta") and
|
| 490 |
+
hasattr(chunk.choices[0].delta, "content")):
|
| 491 |
+
|
| 492 |
+
content = chunk.choices[0].delta.content
|
| 493 |
+
|
| 494 |
+
# Handle GPT-5 thinking placeholders
|
| 495 |
+
if current_model["id"] == "gpt-5" and content:
|
| 496 |
+
if self._is_placeholder_thinking_only(content):
|
| 497 |
+
return None # Skip placeholder content
|
| 498 |
+
return self._strip_placeholder_thinking(content)
|
| 499 |
+
|
| 500 |
+
return content
|
| 501 |
+
except Exception:
|
| 502 |
+
pass
|
| 503 |
+
|
| 504 |
+
return None
|
| 505 |
+
|
| 506 |
+
def _handle_transformers_streaming(self, content: str) -> Generator:
|
| 507 |
+
"""Handle streaming for transformers.js projects"""
|
| 508 |
+
files = parse_transformers_js_output(content)
|
| 509 |
+
has_all_files = all([files.get('index.html'), files.get('index.js'), files.get('style.css')])
|
| 510 |
+
|
| 511 |
+
if has_all_files:
|
| 512 |
+
merged_html = build_transformers_inline_html(files)
|
| 513 |
+
yield {
|
| 514 |
+
'code_output': gr.update(value=merged_html, language="html"),
|
| 515 |
+
'sandbox': self._send_transformers_to_sandbox(files),
|
| 516 |
+
'status': 'streaming'
|
| 517 |
+
}
|
| 518 |
+
else:
|
| 519 |
+
yield {
|
| 520 |
+
'code_output': gr.update(value=content, language="html"),
|
| 521 |
+
'sandbox': "<div style='padding:1em;text-align:center;color:#888;'>Generating transformers.js app...</div>",
|
| 522 |
+
'status': 'streaming'
|
| 523 |
+
}
|
| 524 |
+
|
| 525 |
+
def _handle_svelte_streaming(self, content: str) -> Generator:
|
| 526 |
+
"""Handle streaming for Svelte projects"""
|
| 527 |
+
yield {
|
| 528 |
+
'code_output': gr.update(value=content, language="html"),
|
| 529 |
+
'sandbox': "<div style='padding:1em;text-align:center;color:#888;'>Generating Svelte app...</div>",
|
| 530 |
+
'status': 'streaming'
|
| 531 |
+
}
|
| 532 |
+
|
| 533 |
+
def _handle_standard_streaming(self, content: str, language: str) -> Generator:
|
| 534 |
+
"""Handle streaming for standard projects"""
|
| 535 |
+
clean_code = remove_code_block(content)
|
| 536 |
+
preview = self._generate_preview(clean_code, language)
|
| 537 |
+
|
| 538 |
+
yield {
|
| 539 |
+
'code_output': gr.update(value=clean_code, language=self._get_gradio_language(language)),
|
| 540 |
+
'sandbox': preview,
|
| 541 |
+
'status': 'streaming'
|
| 542 |
+
}
|
| 543 |
+
|
| 544 |
+
def _finalize_content(self, content: str, language: str, query: str,
|
| 545 |
+
gen_image: Optional[gr.Image], session_id: str, media_options: Dict) -> str:
|
| 546 |
+
"""Finalize content with post-processing and media integration"""
|
| 547 |
+
final_content = remove_code_block(content)
|
| 548 |
+
|
| 549 |
+
# Apply media generation for HTML projects
|
| 550 |
+
if language == "html":
|
| 551 |
+
final_content = apply_generated_media_to_html(
|
| 552 |
+
final_content, query, session_id=session_id, **media_options
|
| 553 |
+
)
|
| 554 |
+
|
| 555 |
+
return final_content
|
| 556 |
+
|
| 557 |
+
def _generate_preview(self, content: str, language: str) -> str:
|
| 558 |
+
"""Generate preview HTML for different content types"""
|
| 559 |
+
if language == "html":
|
| 560 |
+
# Handle multi-page HTML
|
| 561 |
+
files = parse_multipage_html_output(content)
|
| 562 |
+
files = validate_and_autofix_files(files)
|
| 563 |
+
if files and files.get('index.html'):
|
| 564 |
+
merged = inline_multipage_into_single_preview(files)
|
| 565 |
+
return self._send_to_sandbox(merged)
|
| 566 |
+
return self._send_to_sandbox(content)
|
| 567 |
+
|
| 568 |
+
elif language == "streamlit":
|
| 569 |
+
if is_streamlit_code(content):
|
| 570 |
+
return self._send_streamlit_to_stlite(content)
|
| 571 |
+
return "<div style='padding:1em;color:#888;text-align:center;'>Add `import streamlit as st` to enable Streamlit preview.</div>"
|
| 572 |
+
|
| 573 |
+
elif language == "gradio":
|
| 574 |
+
if is_gradio_code(content):
|
| 575 |
+
return self._send_gradio_to_lite(content)
|
| 576 |
+
return "<div style='padding:1em;color:#888;text-align:center;'>Add `import gradio as gr` to enable Gradio preview.</div>"
|
| 577 |
+
|
| 578 |
+
elif language == "python":
|
| 579 |
+
if is_streamlit_code(content):
|
| 580 |
+
return self._send_streamlit_to_stlite(content)
|
| 581 |
+
elif is_gradio_code(content):
|
| 582 |
+
return self._send_gradio_to_lite(content)
|
| 583 |
+
return "<div style='padding:1em;color:#888;text-align:center;'>Preview available for Streamlit/Gradio apps. Add the appropriate import.</div>"
|
| 584 |
+
|
| 585 |
+
elif language == "transformers.js":
|
| 586 |
+
files = parse_transformers_js_output(content)
|
| 587 |
+
if files['index.html']:
|
| 588 |
+
return self._send_transformers_to_sandbox(files)
|
| 589 |
+
|
| 590 |
+
return "<div style='padding:1em;color:#888;text-align:center;'>Preview is only available for HTML, Streamlit, and Gradio applications.</div>"
|
| 591 |
+
|
| 592 |
+
def _send_to_sandbox(self, code: str) -> str:
|
| 593 |
+
"""Send HTML to sandboxed iframe with cache busting"""
|
| 594 |
+
import base64
|
| 595 |
+
import time
|
| 596 |
+
|
| 597 |
+
# Add cache-busting timestamp
|
| 598 |
+
timestamp = str(int(time.time() * 1000))
|
| 599 |
+
cache_bust_comment = f"<!-- refresh-{timestamp} -->"
|
| 600 |
+
html_doc = cache_bust_comment + (code or "").strip()
|
| 601 |
+
|
| 602 |
+
# Inline file URLs as data URIs for iframe compatibility
|
| 603 |
+
try:
|
| 604 |
+
html_doc = self._inline_file_urls_as_data_uris(html_doc)
|
| 605 |
+
except Exception:
|
| 606 |
+
pass
|
| 607 |
+
|
| 608 |
+
encoded_html = base64.b64encode(html_doc.encode('utf-8')).decode('utf-8')
|
| 609 |
+
data_uri = f"data:text/html;charset=utf-8;base64,{encoded_html}"
|
| 610 |
+
return f'<iframe src="{data_uri}" width="100%" height="920px" sandbox="allow-scripts allow-same-origin allow-forms allow-popups allow-modals allow-presentation" allow="display-capture" key="preview-{timestamp}"></iframe>'
|
| 611 |
+
|
| 612 |
+
def _send_streamlit_to_stlite(self, code: str) -> str:
|
| 613 |
+
"""Send Streamlit code to stlite preview"""
|
| 614 |
+
import base64
|
| 615 |
+
|
| 616 |
+
html_doc = f"""<!doctype html>
|
| 617 |
+
<html>
|
| 618 |
+
<head>
|
| 619 |
+
<meta charset="UTF-8" />
|
| 620 |
+
<meta name="viewport" content="width=device-width, initial-scale=1, shrink-to-fit=no" />
|
| 621 |
+
<title>Streamlit Preview</title>
|
| 622 |
+
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/@stlite/[email protected]/build/stlite.css" />
|
| 623 |
+
<style>html,body{{margin:0;padding:0;height:100%;}} streamlit-app{{display:block;height:100%;}}</style>
|
| 624 |
+
<script type="module" src="https://cdn.jsdelivr.net/npm/@stlite/[email protected]/build/stlite.js"></script>
|
| 625 |
+
</head>
|
| 626 |
+
<body>
|
| 627 |
+
<streamlit-app>{code or ""}</streamlit-app>
|
| 628 |
+
</body>
|
| 629 |
+
</html>"""
|
| 630 |
+
|
| 631 |
+
encoded_html = base64.b64encode(html_doc.encode('utf-8')).decode('utf-8')
|
| 632 |
+
data_uri = f"data:text/html;charset=utf-8;base64,{encoded_html}"
|
| 633 |
+
return f'<iframe src="{data_uri}" width="100%" height="920px" sandbox="allow-scripts allow-same-origin allow-forms allow-popups allow-modals allow-presentation" allow="display-capture"></iframe>'
|
| 634 |
+
|
| 635 |
+
def _send_gradio_to_lite(self, code: str) -> str:
|
| 636 |
+
"""Send Gradio code to gradio-lite preview"""
|
| 637 |
+
import base64
|
| 638 |
+
|
| 639 |
+
html_doc = f"""<!doctype html>
|
| 640 |
+
<html>
|
| 641 |
+
<head>
|
| 642 |
+
<meta charset="UTF-8" />
|
| 643 |
+
<meta name="viewport" content="width=device-width, initial-scale=1, shrink-to-fit=no" />
|
| 644 |
+
<title>Gradio Preview</title>
|
| 645 |
+
<script type="module" crossorigin src="https://cdn.jsdelivr.net/npm/@gradio/lite/dist/lite.js"></script>
|
| 646 |
+
<link rel="stylesheet" href="https://cdn.jsdelivr.net/npm/@gradio/lite/dist/lite.css" />
|
| 647 |
+
<style>html,body{{margin:0;padding:0;height:100%;}} gradio-lite{{display:block;height:100%;}}</style>
|
| 648 |
+
</head>
|
| 649 |
+
<body>
|
| 650 |
+
<gradio-lite>{code or ""}</gradio-lite>
|
| 651 |
+
</body>
|
| 652 |
+
</html>"""
|
| 653 |
+
|
| 654 |
+
encoded_html = base64.b64encode(html_doc.encode('utf-8')).decode('utf-8')
|
| 655 |
+
data_uri = f"data:text/html;charset=utf-8;base64,{encoded_html}"
|
| 656 |
+
return f'<iframe src="{data_uri}" width="100%" height="920px" sandbox="allow-scripts allow-same-origin allow-forms allow-popups allow-modals allow-presentation" allow="display-capture"></iframe>'
|
| 657 |
+
|
| 658 |
+
def _send_transformers_to_sandbox(self, files: Dict[str, str]) -> str:
|
| 659 |
+
"""Send transformers.js files to sandbox"""
|
| 660 |
+
merged_html = build_transformers_inline_html(files)
|
| 661 |
+
return self._send_to_sandbox(merged_html)
|
| 662 |
+
|
| 663 |
+
def _inline_file_urls_as_data_uris(self, html_doc: str) -> str:
|
| 664 |
+
"""Convert file:// URLs to data URIs for iframe compatibility"""
|
| 665 |
+
import base64
|
| 666 |
+
import mimetypes
|
| 667 |
+
import urllib.parse
|
| 668 |
+
|
| 669 |
+
def _file_url_to_data_uri(file_url: str) -> Optional[str]:
|
| 670 |
+
try:
|
| 671 |
+
parsed = urllib.parse.urlparse(file_url)
|
| 672 |
+
path = urllib.parse.unquote(parsed.path)
|
| 673 |
+
if not path:
|
| 674 |
+
return None
|
| 675 |
+
with open(path, 'rb') as f:
|
| 676 |
+
raw = f.read()
|
| 677 |
+
mime = mimetypes.guess_type(path)[0] or 'application/octet-stream'
|
| 678 |
+
b64 = base64.b64encode(raw).decode()
|
| 679 |
+
return f"data:{mime};base64,{b64}"
|
| 680 |
+
except Exception:
|
| 681 |
+
return None
|
| 682 |
+
|
| 683 |
+
def _replace_file_url(match):
|
| 684 |
+
url = match.group(1)
|
| 685 |
+
data_uri = _file_url_to_data_uri(url)
|
| 686 |
+
return f'src="{data_uri}"' if data_uri else match.group(0)
|
| 687 |
+
|
| 688 |
+
html_doc = re.sub(r'src="(file:[^"]+)"', _replace_file_url, html_doc)
|
| 689 |
+
html_doc = re.sub(r"src='(file:[^']+)'", _replace_file_url, html_doc)
|
| 690 |
+
|
| 691 |
+
return html_doc
|
| 692 |
+
|
| 693 |
+
def _apply_transformers_js_changes(self, original_content: str, changes_text: str) -> str:
|
| 694 |
+
"""Apply search/replace changes to transformers.js content"""
|
| 695 |
+
# Parse original content
|
| 696 |
+
files = parse_transformers_js_output(original_content)
|
| 697 |
+
|
| 698 |
+
# Apply changes to each file
|
| 699 |
+
blocks = self._parse_search_replace_blocks(changes_text)
|
| 700 |
+
|
| 701 |
+
for block in blocks:
|
| 702 |
+
search_text, replace_text = block
|
| 703 |
+
|
| 704 |
+
# Determine target file and apply changes
|
| 705 |
+
for file_key in ['index.html', 'index.js', 'style.css']:
|
| 706 |
+
if search_text in files[file_key]:
|
| 707 |
+
files[file_key] = files[file_key].replace(search_text, replace_text)
|
| 708 |
+
break
|
| 709 |
+
|
| 710 |
+
return format_transformers_js_output(files)
|
| 711 |
+
|
| 712 |
+
def _parse_search_replace_blocks(self, changes_text: str) -> List[Tuple[str, str]]:
|
| 713 |
+
"""Parse search/replace blocks from text"""
|
| 714 |
+
blocks = []
|
| 715 |
+
current_block = ""
|
| 716 |
+
lines = changes_text.split('\n')
|
| 717 |
+
|
| 718 |
+
for line in lines:
|
| 719 |
+
if line.strip() == SEARCH_START:
|
| 720 |
+
if current_block.strip():
|
| 721 |
+
blocks.append(current_block.strip())
|
| 722 |
+
current_block = line + '\n'
|
| 723 |
+
elif line.strip() == REPLACE_END:
|
| 724 |
+
current_block += line + '\n'
|
| 725 |
+
blocks.append(current_block.strip())
|
| 726 |
+
current_block = ""
|
| 727 |
+
else:
|
| 728 |
+
current_block += line + '\n'
|
| 729 |
+
|
| 730 |
+
if current_block.strip():
|
| 731 |
+
blocks.append(current_block.strip())
|
| 732 |
+
|
| 733 |
+
# Parse each block into search/replace pairs
|
| 734 |
+
parsed_blocks = []
|
| 735 |
+
for block in blocks:
|
| 736 |
+
lines = block.split('\n')
|
| 737 |
+
search_lines = []
|
| 738 |
+
replace_lines = []
|
| 739 |
+
in_search = False
|
| 740 |
+
in_replace = False
|
| 741 |
+
|
| 742 |
+
for line in lines:
|
| 743 |
+
if line.strip() == SEARCH_START:
|
| 744 |
+
in_search = True
|
| 745 |
+
in_replace = False
|
| 746 |
+
elif line.strip() == DIVIDER:
|
| 747 |
+
in_search = False
|
| 748 |
+
in_replace = True
|
| 749 |
+
elif line.strip() == REPLACE_END:
|
| 750 |
+
in_replace = False
|
| 751 |
+
elif in_search:
|
| 752 |
+
search_lines.append(line)
|
| 753 |
+
elif in_replace:
|
| 754 |
+
replace_lines.append(line)
|
| 755 |
+
|
| 756 |
+
if search_lines:
|
| 757 |
+
search_text = '\n'.join(search_lines).strip()
|
| 758 |
+
replace_text = '\n'.join(replace_lines).strip()
|
| 759 |
+
parsed_blocks.append((search_text, replace_text))
|
| 760 |
+
|
| 761 |
+
return parsed_blocks
|
| 762 |
+
|
| 763 |
+
def _call_llm(self, client, current_model: Dict, messages: List[Dict],
|
| 764 |
+
max_tokens: int = 4000, temperature: float = 0.7) -> Optional[str]:
|
| 765 |
+
"""Call LLM and return response content"""
|
| 766 |
+
try:
|
| 767 |
+
if current_model.get('type') == 'openai':
|
| 768 |
+
response = client.chat.completions.create(
|
| 769 |
+
model=current_model['id'],
|
| 770 |
+
messages=messages,
|
| 771 |
+
max_tokens=max_tokens,
|
| 772 |
+
temperature=temperature
|
| 773 |
+
)
|
| 774 |
+
return response.choices[0].message.content
|
| 775 |
+
elif current_model.get('type') == 'mistral':
|
| 776 |
+
response = client.chat.complete(
|
| 777 |
+
model=current_model['id'],
|
| 778 |
+
messages=messages,
|
| 779 |
+
max_tokens=max_tokens,
|
| 780 |
+
temperature=temperature
|
| 781 |
+
)
|
| 782 |
+
return response.choices[0].message.content
|
| 783 |
+
else:
|
| 784 |
+
completion = client.chat.completions.create(
|
| 785 |
+
model=current_model['id'],
|
| 786 |
+
messages=messages,
|
| 787 |
+
max_tokens=max_tokens,
|
| 788 |
+
temperature=temperature
|
| 789 |
+
)
|
| 790 |
+
return completion.choices[0].message.content
|
| 791 |
+
except Exception as e:
|
| 792 |
+
print(f"[GenerationEngine] LLM call failed: {e}")
|
| 793 |
+
return None
|
| 794 |
+
|
| 795 |
+
def _convert_history_to_messages(self, history: List[Tuple[str, str]]) -> List[Dict[str, str]]:
|
| 796 |
+
"""Convert history tuples to message format"""
|
| 797 |
+
messages = []
|
| 798 |
+
for user_msg, assistant_msg in history:
|
| 799 |
+
if isinstance(user_msg, list):
|
| 800 |
+
text_content = ""
|
| 801 |
+
for item in user_msg:
|
| 802 |
+
if isinstance(item, dict) and item.get("type") == "text":
|
| 803 |
+
text_content += item.get("text", "")
|
| 804 |
+
user_msg = text_content if text_content else str(user_msg)
|
| 805 |
+
|
| 806 |
+
messages.append({"role": "user", "content": user_msg})
|
| 807 |
+
messages.append({"role": "assistant", "content": assistant_msg})
|
| 808 |
+
return messages
|
| 809 |
+
|
| 810 |
+
def _get_gradio_language(self, language: str) -> Optional[str]:
|
| 811 |
+
"""Get appropriate Gradio language for syntax highlighting"""
|
| 812 |
+
from config import get_gradio_language
|
| 813 |
+
return get_gradio_language(language)
|
| 814 |
+
|
| 815 |
+
def _is_placeholder_thinking_only(self, text: str) -> bool:
|
| 816 |
+
"""Check if text contains only thinking placeholders"""
|
| 817 |
+
if not text:
|
| 818 |
+
return False
|
| 819 |
+
stripped = text.strip()
|
| 820 |
+
if not stripped:
|
| 821 |
+
return False
|
| 822 |
+
return re.fullmatch(r"(?s)(?:\s*Thinking\.\.\.(?:\s*\(\d+s elapsed\))?\s*)+", stripped) is not None
|
| 823 |
+
|
| 824 |
+
def _strip_placeholder_thinking(self, text: str) -> str:
|
| 825 |
+
"""Remove placeholder thinking status lines"""
|
| 826 |
+
if not text:
|
| 827 |
+
return text
|
| 828 |
+
return re.sub(r"(?mi)^[\t ]*Thinking\.\.\.(?:\s*\(\d+s elapsed\))?[\t ]*$\n?", "", text)
|
| 829 |
+
|
| 830 |
+
def _update_avg_response_time(self, elapsed_time: float):
|
| 831 |
+
"""Update average response time statistic"""
|
| 832 |
+
current_avg = self._generation_stats['avg_response_time']
|
| 833 |
+
total_successful = self._generation_stats['successful_generations']
|
| 834 |
+
|
| 835 |
+
if total_successful <= 1:
|
| 836 |
+
self._generation_stats['avg_response_time'] = elapsed_time
|
| 837 |
+
else:
|
| 838 |
+
# Weighted average
|
| 839 |
+
self._generation_stats['avg_response_time'] = (current_avg * (total_successful - 1) + elapsed_time) / total_successful
|
| 840 |
+
|
| 841 |
+
def get_generation_stats(self) -> Dict[str, Any]:
|
| 842 |
+
"""Get current generation statistics"""
|
| 843 |
+
return self._generation_stats.copy()
|
| 844 |
+
|
| 845 |
+
def get_active_generations(self) -> Dict[str, Dict[str, Any]]:
|
| 846 |
+
"""Get information about currently active generations"""
|
| 847 |
+
return self._active_generations.copy()
|
| 848 |
+
|
| 849 |
+
# Global generation engine instance
|
| 850 |
+
generation_engine = GenerationEngine()
|