-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Another one to check out #9
Comments
Maybe this works now. I can't test though cause only 8GB VRAM. LoRAs seem to work at least Download the LoRA: https://huggingface.co/kaiokendev/superhot-13b-8k-no-rlhf-test - and set the path in the app (File>Parameters>Exllama) |
Seen the stuff going on with these long conext + exllama? https://huggingface.co/Panchovix https://huggingface.co/TheBloke "Please make sure you're using the latest version of text-generation-webui Click the Model tab. |
They should work here now with Exllama backend, if set context to 8192 and 'Compress pos embeddings' to 4 |
Extend context 6-8k on any model. *Have not tested
https://kaiokendev.github.io/til#extending-context-to-8k
The text was updated successfully, but these errors were encountered: