Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Follow-up answers are slow #408

Open
woheller69 opened this issue May 9, 2024 · 0 comments
Open

Follow-up answers are slow #408

woheller69 opened this issue May 9, 2024 · 0 comments

Comments

@woheller69
Copy link

I have a CPU only setup, so my system is quite slow.
I notice that llamafile is much slower than gpt4all for follow-up answers.

e.g. I ask (using Dolphin 2.7 Mixtral 8x7b with its lengthy system message): A farmer with a wolf, a goat, and a cabbage must cross a river by boat. The boat can carry only the farmer and a single item. If left unattended together, the wolf would eat the goat, or the goat would eat the cabbage. How can they cross the river without anything being eaten?

First prompt evaluation with llamafile is about the same as for gpt4all (about 80s)

But when I reply to the answer telling that it is wrong, llamafile (60s) takes about as long for prompt processing as for the first answer, while gpt4all (6s) answers almost immediately.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants