Spaces:
Sleeping
Sleeping
File size: 2,342 Bytes
ac5a3a1 ac35688 ac5a3a1 ac35688 ac5a3a1 40b123c ac5a3a1 9f05f25 84c30f1 9f05f25 ac5a3a1 40b123c ac35688 9f05f25 f728ad5 ac5a3a1 40b3752 dfce113 92f3987 84c30f1 92f3987 9f05f25 ac5a3a1 ac35688 92f3987 40b3752 ac5a3a1 92f3987 40b123c 9470356 9f05f25 ac5a3a1 ac35688 ac5a3a1 92f3987 9f05f25 ac5a3a1 de81947 ac5a3a1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 |
# FILE: modules/config_settings_public.py (HF Demo - v8.0)
import os
import logging
from huggingface_hub import hf_hub_download
logger = logging.getLogger("ZOTHEOS_Config_HF")
# --- β
DEFINITIVE & VERIFIED WEB-OPTIMIZED MODEL SOURCES ---
MODEL_DEFINITIONS = {
"mistral": {
"repo_id": "TheBloke/Mistral-7B-Instruct-v0.2-GGUF",
"filename": "mistral-7b-instruct-v0.2.Q2_K.gguf"
},
"gemma": {
"repo_id": "google/gemma-2b-it-gguf", # β
Using the OFFICIAL Google repository
"filename": "gemma-2b-it.gguf" # β
This is the correct filename
},
"qwen": {
"repo_id": "Qwen/Qwen1.5-0.5B-Chat-GGUF",
"filename": "qwen1_5-0.5b-chat-q2_k.gguf"
}
}
MODEL_PATHS = {}
logger.info("β
β
β
RUNNING IN WEB DEMO MODE (True Fusion - CPU Survival) β
β
β
")
N_GPU_LAYERS_FALLBACK = 0 # Force CPU-only mode
for name, model_info in MODEL_DEFINITIONS.items():
logger.info(f"Downloading demo model: {name} from {model_info['repo_id']}...")
try:
MODEL_PATHS[name] = hf_hub_download(repo_id=model_info["repo_id"], filename=model_info["filename"])
logger.info(f"β
Successfully downloaded {name}")
except Exception as e:
logger.error(f"β FAILED to download {name}: {e}")
raise e
# --- β
CPU-OPTIMIZED MODEL PARAMETERS ---
MODEL_SPECIFIC_PARAMS = {
"_default": {
"n_gpu_layers": N_GPU_LAYERS_FALLBACK,
"n_ctx": 1024,
"n_batch": 256,
"n_threads": 4,
"verbose": True
},
"mistral": { "chat_format": "mistral-instruct" },
"gemma": { "chat_format": "gemma" },
"qwen": { "chat_format": "chatml" }
}
# --- TIER INFERENCE & PROMPTS ---
INFERENCE_PRESETS = {"balanced": {"temperature": 0.7, "max_tokens": 256}}
DEFAULT_INFERENCE_PRESET = "balanced"
DEFAULT_SYSTEM_PROMPT = "You are ZOTHEOS, an ethical AI. Respond concisely and only in English."
MODEL_ROLES = {"mistral": "analyst", "gemma": "humanist", "qwen": "skeptic"}
MODEL_ROLE_SYSTEM_PROMPTS = {
"analyst": "You are an analyst. Be logical. Respond only in English.",
"humanist": "You are a humanist. Focus on values. Respond only in English.",
"skeptic": "You are a skeptic. Challenge the premise. Respond only in English.",
}
logger.info("β
Hugging Face Demo (CPU Survival Mode) Configuration Loaded.") |