-
Notifications
You must be signed in to change notification settings - Fork 1.6k
Open
Description
Describe the bug
Sometimes when chatting the model with plenty of tokens will respond with summarization but it fails. And it turns out to be another error that eventually shows Input is too long for the requested model
after chatting again
To Reproduce
See attached screenshots
Expected behavior
If the error is indeed non recoverable with summarization it should show the error right away and not attempt summarization or show an option to summarize. Secondly how do we avoid hitting this Input is too long for the requested model
error in the first place or better recover from it.
Screenshots





Please provide following information:
- OS & Arch: Mac arm
- Interface: UI
- Version: v1.4.1
- Extensions enabled: redacted
- Provider & Model: Sonnet
Additional context
Add any other context about the problem here.
Metadata
Metadata
Assignees
Labels
No labels