-
Notifications
You must be signed in to change notification settings - Fork 225
LLM update 202507 #8455
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LLM update 202507 #8455
Conversation
2b1a5c4
to
02295d5
Compare
02295d5
to
27bdd94
Compare
51375f8
to
fae9f42
Compare
fae9f42
to
de3e86c
Compare
77c44d7
to
1928b93
Compare
…mprove admin panel LLM test UI
1928b93
to
6bdcbfa
Compare
- Fix o1 models stream_options error by only including stream_options when streaming is enabled - Fix o1 models system role error by omitting system messages entirely (o1 models don't support system roles) - Update tests to use USE_NEWER_LC_IMPL flag to switch between legacy and unified LangChain implementations - Export USE_NEWER_LC_IMPL flag for test usage - All 22 LLM tests now pass including both o1 and o1-mini models 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <[email protected]>
- Add comprehensive test suite for user-defined LLMs - Test OpenAI, Google, Anthropic, Mistral, and custom OpenAI models - Create ephemeral test database account with proper user-defined LLM config storage - Use environment variables for API keys (COCALC_TEST_*_KEY) - Tests validate end-to-end functionality from database storage to LLM evaluation - Update Anthropic model to use claude-3-5-haiku-latest alias - All 5 user-defined LLM tests passing 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <[email protected]>
4b88bb5
to
db444e6
Compare
db444e6
to
f764dc0
Compare
@haraldschilly what is the release process? E.g., what's the current coupling between frontend backend for this? You usually do a good job explaining this. E.g., what happens if only the frontend is updated? Only the backend? do I need to force people to refresh? |
Hmm, good question. Since the backend reports to the frontend what LLMs are available, it is fine to update the frontend and then the backend. There was no change with the actual communication. Overall, both need to be updated. |
Thanks - that's optimal ! And of course it is all live right now. |
LLM Infrastructure Improvements
Key Updates:
maxReasoningTokens: 1024
for Gemini 2.5 modelsHighlights:
USE_NEWER_LC_IMPL
flag to support gradual migration to unified implementationAll 27 LLM tests now pass, including the new user-defined LLM functionality. The changes maintain backward compatibility while providing a foundation for future LLM improvements.