-
-
Notifications
You must be signed in to change notification settings - Fork 221
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama 2 Chat implementation #221
base: master
Are you sure you want to change the base?
Conversation
I'm not having a lot of luck with this.
This is with TheBloke/Llama-2-13B-chat-GPTQ which is otherwise coherent enough. |
Huh, that's super weird, in my testing it worked flawlessly with Llama 2 7b Chat Must have to do with the sequence_actual splitting, I will investigate with your prompt Did you supply bn and un? This is the 13b chat model and it works for anything else than Hi!? |
I only supplied the model, and it seems to fail with any prompt. |
That's extremely odd, that's my response (also works with 7b chat for me): |
System: You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature. If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.. User: Write me a NodeJS script that tells the time in human readable format.
Here's what's happening here:
When you run this script, it should output the current time in the format "hh:mm AM" (or PM, depending on the time of day). For example:
Note that this script uses the
|
@turboderp I am quite baffled by your output, it doesn't make any sense to me. I am not able to replicate your issue. Could you please print: I just personally printed them like this: Thank you very much |
The only other thing I could imagine making truuble is the version of sentence piece and torch |
Hi, I tested with 13B model is everything seems OK. Maybe it is related to some hyperparameters. I have changed some of them.
|
The ���s still hint at a problem of some sort. The 13B chat model I tried works fine when just sampled. I'll have to dig into it a little more, maybe with a 7B model? |
I have to add that I often got those (or single ones) with your normal example chatbot too with the 7B chat and 13B chat model sometimes, idk where they come from I think tomorrow I will be able to fix at least the sequence slicing problem |
Hey!
This is the correct LLama 2 Chat prompt formatting implementation into
example_llama2chat.py
.This PR uses #195 to copy the exact implementation of the original Llama repo
The format for Llama 2 looks like this:
To run it yourself:
python example_llama2chat.py -d ./your/model/path
Have fun!