-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
V2 update plan #2
Comments
<= 67.2 ms on 7B with partial GPU offload (3060 Ti 8G) is a huge win win |
Hi,
Is it the compatibility issue mentioned here? |
Yup, the new rwkv.cpp - is now merged into v2 (publishing now) |
Merged in : 74655de This resolves all issues, EXCEPT support for world model tokenizer (needs a new JS tokenizer) |
update the docs please |
waiting for it ~~ |
Hi, I reinstalled the package, but when I run it, I get only following result:
Did I convert the model wrong? |
world models not support yet |
It is just a normal finetuned raven model |
Latest version of https://github.com/saharNooby/rwkv.cpp has new quantization format (breaking change?) and GPU offload (!!!)
Since this might be potentially breaking changes, its gonna be a v2 update.
The text was updated successfully, but these errors were encountered: