Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

i was told i could use these bindings with VICUNA model #9

Open
shubham8550 opened this issue Apr 6, 2023 · 6 comments
Open

i was told i could use these bindings with VICUNA model #9

shubham8550 opened this issue Apr 6, 2023 · 6 comments

Comments

@shubham8550
Copy link

what i pass to constructior i have quantized
file with me

  const gpt4all = new GPT4All('gpt4all-lora-unfiltered-quantized', true);
@shubham8550
Copy link
Author

So i installed VICUNA with this repo

https://github.com/mps256/autovicuna

They put everything in C:/VICUNA
(only 13B Vicuna CPU )

they use
llama.zip

and
ggml-vicuna-13b-4bit.bin

you can get these from

https://huggingface.co/eachadea/ggml-vicuna-13b-4bit/blob/main/ggml-vicuna-13b-4bit-rev1.bin

" ...rev1.bin" have some fixes but its not included in mps256/autovicuna ig

you can get 7B from and it can run on same llama files
https://huggingface.co/eachadea/ggml-vicuna-7b-4bit/tree/main

Heres Some SS running 13B VICUNA
image

and here 7B (its faster and gives answers better than LLAMA/ alpaca in my pov)
image

output sometimes end with
image
maybe because of these
Reverse prompt: '### Human:'
Reverse prompt: '### Instruction:

@shubham8550
Copy link
Author

C:\VICUNA\main.exe -i --interactive-first -r "### Human:" -t 8 --temp 0 -c 2048 -n -1 --ignore-eos --repeat_penalty 1.2 --instruct -m C:\VICUNA\ggml-vicuna-7b-4bit-rev1.bin

sample command

@jellydn
Copy link

jellydn commented Apr 8, 2023

Refer to the source code, it's NOT supported yet.

constructor(model: string = 'gpt4all-lora-quantized', forceDownload: boolean = false, decoderConfig: Record<string, any> = {}) {
this.model = model;
this.decoderConfig = decoderConfig;
/*
allowed models:
M1 Mac/OSX: cd chat;./gpt4all-lora-quantized-OSX-m1
Linux: cd chat;./gpt4all-lora-quantized-linux-x86
Windows (PowerShell): cd chat;./gpt4all-lora-quantized-win64.exe
Intel Mac/OSX: cd chat;./gpt4all-lora-quantized-OSX-intel
*/
if (
'gpt4all-lora-quantized' !== model &&
'gpt4all-lora-unfiltered-quantized' !== model
) {
throw new Error(`Model ${model} is not supported. Current models supported are:
gpt4all-lora-quantized
gpt4all-lora-unfiltered-quantized`
);
}

@shubham8550
Copy link
Author

Refer to the source code, it's NOT supported yet.

constructor(model: string = 'gpt4all-lora-quantized', forceDownload: boolean = false, decoderConfig: Record<string, any> = {}) {
this.model = model;
this.decoderConfig = decoderConfig;
/*
allowed models:
M1 Mac/OSX: cd chat;./gpt4all-lora-quantized-OSX-m1
Linux: cd chat;./gpt4all-lora-quantized-linux-x86
Windows (PowerShell): cd chat;./gpt4all-lora-quantized-win64.exe
Intel Mac/OSX: cd chat;./gpt4all-lora-quantized-OSX-intel
*/
if (
'gpt4all-lora-quantized' !== model &&
'gpt4all-lora-unfiltered-quantized' !== model
) {
throw new Error(`Model ${model} is not supported. Current models supported are:
gpt4all-lora-quantized
gpt4all-lora-unfiltered-quantized`
);
}

yes that's why i made one

https://www.npmjs.com/package/vicuna-ts

@jellydn
Copy link

jellydn commented Apr 9, 2023

Refer to the source code, it's NOT supported yet.

constructor(model: string = 'gpt4all-lora-quantized', forceDownload: boolean = false, decoderConfig: Record<string, any> = {}) {
this.model = model;
this.decoderConfig = decoderConfig;
/*
allowed models:
M1 Mac/OSX: cd chat;./gpt4all-lora-quantized-OSX-m1
Linux: cd chat;./gpt4all-lora-quantized-linux-x86
Windows (PowerShell): cd chat;./gpt4all-lora-quantized-win64.exe
Intel Mac/OSX: cd chat;./gpt4all-lora-quantized-OSX-intel
*/
if (
'gpt4all-lora-quantized' !== model &&
'gpt4all-lora-unfiltered-quantized' !== model
) {
throw new Error(`Model ${model} is not supported. Current models supported are:
gpt4all-lora-quantized
gpt4all-lora-unfiltered-quantized`
);
}

yes that's why i made one

https://www.npmjs.com/package/vicuna-ts

Nice 👍 Do you think that's good for community if your package is part of gpt4all-ts? Have you considered to send PR to gpt4all-ts? Thanks.

@shubham8550
Copy link
Author

originally thats what i wanted but gpt4all-ts s code would have become mess if i stuffed vicuna

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants