import gradio as gr import os from random import randint from all_models import models from datetime import datetime from concurrent.futures import TimeoutError, ThreadPoolExecutor import numpy as np import time import requests import logging import traceback # For better error reporting os.environ["CUDA_VISIBLE_DEVICES"] = "-1" logging.basicConfig(level=logging.WARNING) now2 = 0 index_tracker = 0 # Index tracker for the current model model_scores = {model: 0 for model in models} # Dictionary to track scores for each model processed_models_count = 0 kii=" blonde mohawk femboy playing game with self at computer with programmer socks on, still a wip" combined_prompt = "" def get_current_time(): now = datetime.now() now2 = now current_time = now2.strftime("%Y-%m-%d %H:%M:%S") ki = f'{kii} {current_time}' return ki # Sanitize file names and truncate them def sanitize_file_name(file_name, max_length=100): """Shortens and removes unsafe characters from file name.""" file_name = file_name[:max_length] return file_name.replace(" ", "_").replace("/", "_") def load_fn(models): global models_load models_load = {} for model in models: if model not in models_load.keys(): try: m = gr.load(f'models/{model}') print(f"{m}\n"); models_load.update({model: m}) models_load[model] = m # Store in dictionary except Exception as error: print(f"Error loading model {model}: {error}\n") #m = gr.Interface(lambda _: None, inputs=gr.Textbox(), outputs=gr.Image(), queue=False) #models_load.update({model: m}) traceback.print_exc() # Prints full error stack trace for debugging #m = gr.Interface(fn=lambda _: None, inputs=gr.Textbox(), outputs=gr.Image(), queue=False) models_load[model] = None #return models_load # Return dictionary instead of using global load_fn(models) num_models = len(models) default_models = models[:num_models] def extend_choices(choices): return choices + (num_models - len(choices)) * ['NA'] def update_imgbox(choices): choices_plus = extend_choices(choices) return [gr.Image(None, label=m, visible=(m != 'NA')) for m in choices_plus] executor = ThreadPoolExecutor(max_workers=num_models) def gen_fn(model_str, prompt): global index_tracker, model_scores, processed_models_count if model_str == 'NA': return None try: index_tracker = (index_tracker + 1) % len(models) current_model_index = index_tracker current_model_name = models[current_model_index] max_prompt_length = 100 truncated_prompt = sanitize_file_name(prompt[:max_prompt_length]) combined_prompt = f"{truncated_prompt}_{randint(0, 9999)}" # Execute the model's processing with a timeout future = executor.submit(models_load[model_str], f"{combined_prompt}") response = future.result(timeout=150) # Wait for result with timeout if isinstance(response, gr.Image): return response elif isinstance(response, tuple): return None elif isinstance(response, str): if processed_models_count == 0: print(f"***a***********") # print(f"{prompt}") print(f"{prompt}") # print(f"{prompt}") print(f"***b***********") model_scores[current_model_name] += 1 print(f"OOO n:{processed_models_count} x:{current_model_index} r[{model_scores[current_model_name]}] {model_str}") processed_models_count += 1 if processed_models_count == len(models): print("\nCycle Complete! Updated Scores:") print(model_scores) processed_models_count = 0 return response except TimeoutError: print(f"TimeoutError: Model '{model_str}' did not respond within 150 seconds.") processed_models_count += 1 if processed_models_count == len(models): print("\nCycle Complete! Updated Scores:") print(model_scores) processed_models_count = 0 return None except Exception as e: if processed_models_count == 0: print(f"******c*******") # print(f"{prompt}") # print(f"{prompt}") # print(f"{prompt}") print(f"******d*******") print(f"--- n:{processed_models_count} x:{current_model_index} r[{model_scores[current_model_name]}] {model_str}") processed_models_count += 1 if processed_models_count == len(models): print("\nCycle Complete! Updated Scores:") print(model_scores) processed_models_count = 0 return None def make_me(): with gr.Row(): txt_input = gr.Textbox(lines=2, value=kii, label=None) gen_button = gr.Button('Generate images') stop_button = gr.Button('Stop', variant='secondary', interactive=False) gen_button.click(lambda _: gr.update(interactive=True), None, stop_button) gen_button.click(lambda _: gr.update(interactive=True), None) gr.HTML("""