-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
chat with bob example broken #12
Comments
Hi @nuance1979, You are welcome & thanks for reporting the bug. Could you please let me know what model are you using so I can debug using the same model ? |
@nuance1979 Oh, I have just noticed! it was just the |
@nuance1979, I think you just need to specify the exact parameters as that Please update from source and give it a try ? |
@nuance1979, that's weird, the model seems to be always hallucinating! on my end everything's working as expected (as you can see on my previous comment). yeah you are right, I really don't know why it is devised by half, usually it equals to the context size! have you tried other models? |
I tried I checked the SHA256SUM of my |
Yeah something is happening, but I honestly have no idea since I couldn't reproduce this issue on my end. |
Can you ask a third person to try it? Just to see whether it's a problem on my side. |
Sure .. let us try that, @ParisNeo is using -- Could you please let us know if someone on your repo has reported a similar problem to this issue ? Thank you! |
Hi, No, I didn't have any complaints about the pyllamacpp backend yet. If I have time tomorrow I'll try. I got to go. |
Thanks @ParisNeo, Let us know if you found any issues. |
hi @nuance1979, any news on this ? Are you still getting the same error ? If you know someone else who can test it then please send them a message! Otherwise I have tried to test it on colab as well, even though it is slow, but it worked as expected. |
Yes. Still non-sensical answers.
Sure. I'll ask my friend to test it.
All my tests were done with the original llama 7B model (quantized into q4_0.bin with llama.cpp). But you are testing |
Ok. I tried your notebook with llama-7b and it can reproduce what I saw: Again, I want to emphasize that the same model behaves correctly when I use You can try it yourself with this model link: https://huggingface.co/TheBloke/LLaMa-7B-GGML/resolve/main/llama-7b.ggmlv3.q4_0.bin |
Oh! are you using the original model? .. So maybe that's the source of the problem. I will try to test with the original model and see. |
I understand the difference between original llama and instruction-tuned variants. All I'm saying is that the fact that llama.cpp works under the same condition points to a potential bug in pyllamacpp and it would be great if you can fix it. |
@nuance1979, Yeah you are right. Sorry for that :( Let me know if you have any ideas .. any help would be appreciated! Thanks! |
if it helps someone, I tried it like this. edit cli.py and make following changes:
example output:
I am synced to commit 6d487b9 |
Thanks @siddhsql ! However, you are using |
Hi @abdeladim-s , thanks for the update!
I was trying to update to
pyllamacpp==2.4.0
but found that even the example on the README, which is similar tollama.cpp
's./examples/chat.sh
but not identical, is not working properly. For example, when I copied the example code into afoo.py
and run it, I got:If I go to
llama.cpp
, check out66874d4
thenmake clean && make && ./examples/chat.sh
, I got:I just want to get an equivalent of running
llama.cpp
'schat.sh
withpyllamacpp==2.4.0
, no more no less. How should I do it?The text was updated successfully, but these errors were encountered: