Files
HOA_Financial_Platform/backend
olsch01 36271585d9 Fix AI timeout and token limits for thinking model (kimi-k2.5)
- Increase max_tokens from 4096 to 16384 to accommodate reasoning tokens
- Increase timeout from 90s to 180s for thinking model latency
- Add logging for response diagnostics (content length, reasoning, finish reason)
- Better error message when model exhausts tokens on reasoning

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-25 15:50:53 -05:00
..