""" LLM (Large Language Model) endpoints """ from fastapi import APIRouter, HTTPException from loguru import logger from api.dependencies import PixelleVideoDep from api.schemas.llm import LLMChatRequest, LLMChatResponse router = APIRouter(prefix="/llm", tags=["Basic Services"]) @router.post("/chat", response_model=LLMChatResponse) async def llm_chat( request: LLMChatRequest, pixelle_video: PixelleVideoDep ): """ LLM chat endpoint Generate text response using configured LLM. - **prompt**: User prompt/question - **temperature**: Creativity level (0.0-2.0, lower = more deterministic) - **max_tokens**: Maximum response length Returns generated text response. """ try: logger.info(f"LLM chat request: {request.prompt[:50]}...") # Call LLM service response = await pixelle_video.llm( prompt=request.prompt, temperature=request.temperature, max_tokens=request.max_tokens ) return LLMChatResponse( content=response, tokens_used=None # Can add token counting if needed ) except Exception as e: logger.error(f"LLM chat error: {e}") raise HTTPException(status_code=500, detail=str(e))