feat: Add VLM config to config.example.yaml with config.yaml support
This commit is contained in:
@@ -17,6 +17,20 @@ llm:
|
|||||||
# DeepSeek: base_url: "https://api.deepseek.com" model: "deepseek-chat"
|
# DeepSeek: base_url: "https://api.deepseek.com" model: "deepseek-chat"
|
||||||
# Ollama (Local): base_url: "http://localhost:11434/v1" model: "llama3.2"
|
# Ollama (Local): base_url: "http://localhost:11434/v1" model: "llama3.2"
|
||||||
|
|
||||||
|
# ==================== VLM Configuration (Vision Language Model) ====================
|
||||||
|
# Used for character analysis and image understanding
|
||||||
|
# If not configured, will try to use LLM config with vision model auto-detection
|
||||||
|
vlm:
|
||||||
|
provider: "qwen" # Options: qwen, glm, openai
|
||||||
|
api_key: "" # Leave empty to use DASHSCOPE_API_KEY or VLM_API_KEY env var
|
||||||
|
base_url: "" # Leave empty for auto-detection based on provider
|
||||||
|
model: "" # Leave empty for default model based on provider
|
||||||
|
|
||||||
|
# VLM Provider presets:
|
||||||
|
# Qwen (通义千问): provider: "qwen" model: "qwen-vl-plus" or "qwen-vl-max" or "qwen3-vl-plus"
|
||||||
|
# GLM (智谱): provider: "glm" model: "glm-4v-flash" or "glm-4v"
|
||||||
|
# OpenAI: provider: "openai" model: "gpt-4-vision-preview" or "gpt-4o"
|
||||||
|
|
||||||
# ==================== ComfyUI Configuration ====================
|
# ==================== ComfyUI Configuration ====================
|
||||||
comfyui:
|
comfyui:
|
||||||
# Global ComfyUI settings
|
# Global ComfyUI settings
|
||||||
|
|||||||
@@ -145,12 +145,18 @@ Output ONLY the JSON object, no additional text."""
|
|||||||
}
|
}
|
||||||
]
|
]
|
||||||
|
|
||||||
# Get VLM configuration from environment or fallback to LLM config
|
# Get VLM configuration
|
||||||
|
# Priority: Environment variables > config.yaml > defaults
|
||||||
import os
|
import os
|
||||||
vlm_provider = os.getenv("VLM_PROVIDER", "qwen") # qwen, glm, openai
|
from pixelle_video.config import config_manager
|
||||||
vlm_api_key = os.getenv("VLM_API_KEY") or os.getenv("DASHSCOPE_API_KEY")
|
|
||||||
vlm_base_url = os.getenv("VLM_BASE_URL")
|
# Try to get VLM config from config.yaml
|
||||||
vlm_model = os.getenv("VLM_MODEL")
|
vlm_config = getattr(config_manager.config, 'vlm', None)
|
||||||
|
|
||||||
|
vlm_provider = os.getenv("VLM_PROVIDER") or (vlm_config.provider if vlm_config and hasattr(vlm_config, 'provider') else "qwen")
|
||||||
|
vlm_api_key = os.getenv("VLM_API_KEY") or os.getenv("DASHSCOPE_API_KEY") or (vlm_config.api_key if vlm_config and hasattr(vlm_config, 'api_key') else None)
|
||||||
|
vlm_base_url = os.getenv("VLM_BASE_URL") or (vlm_config.base_url if vlm_config and hasattr(vlm_config, 'base_url') else None)
|
||||||
|
vlm_model = os.getenv("VLM_MODEL") or (vlm_config.model if vlm_config and hasattr(vlm_config, 'model') else None)
|
||||||
|
|
||||||
# Configure based on provider
|
# Configure based on provider
|
||||||
if vlm_provider == "qwen":
|
if vlm_provider == "qwen":
|
||||||
|
|||||||
Reference in New Issue
Block a user