Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Changing LLMs mid-conversation results in unexpected behavior #415

Open
jonathonbarton opened this issue Sep 15, 2024 · 0 comments
Open

Comments

@jonathonbarton
Copy link
Contributor

Part Bug Report
Part Feature Request

Sometimes a model will start spitting out 0 Token replies over and over and the model needs to be changed
Sometimes it would be nice to be able to switch from a Vision model to a Non Vision model mid conversation.

The web UI allows the selection of a new model, the new model is confirmed in the Python window.
image

A summary is apparently saved (though this summary ("OK") is really terrible, and appears to have been saved by Lumimaid) and then the summary ("OK") is loaded.

The next step is expected, but not ideal - We return to 'Waiting for player to select NPC'
Then the NPC is correctly selected - expected, but not ideal
Then I get an error message about the override file (that I do not get on initial load)

image

Then comes the repeatable bug.

TTS; Connecting to XTTS...

image

...and this becomes an infinite wait that ultimately forces a restart of Mantella from the MCM.

BUGFIX:
Changing Models should not hang Mantella indefinitely when using XTTS.

FEATURE REQUEST:
Changing Models should be seamless when using OpenRouter.
Current Behavior (post-bugfix): Close Conversation, Summarize, Change Model, Load Summary, Continue from summary
Proposed Behavior: Write Conversation JSON, No Summary, Change Model, Send the entire collected context from the other LLM to the new LLM (rely on OpenRouter's "Middle Out" culling for Context Overrun cases), including the entirety of the 'current' conversation from the Converation JSON log we just wrote, so the conversation is effectively seamless (though the reply 'style' and 'tone' will change if shifting from, say Anthropic Vision to Hermes No-Vision or even Lumimaid 8b from Hermes 405B...)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant