This repository was archived by the owner on Jun 24, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 372
Reporting model stats #46
Comments
Would be nice to see the stats, I'm curious why it feels a little slower than the C++ version. |
@brysgo oh really? From my local testing, the performance is about the same... do you have any metrics/results can share? |
Right now it is just qualitative, watching the words come up. Maybe it is I/O bound? I'm going to checkout your branch and run the stats. |
llama-rs llama.cpp
Model Name: MacBook Pro |
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hey, in llama.cpp, there are some useful output for telling the model predict / evaluation stats:
https://github.com/ggerganov/llama.cpp/blob/da5303c1ea68aa19db829c634f1e10d08d409680/main.cpp#L1086-L1095
Can we also export such information after the model run?
I have a local change which produces output like this:
The text was updated successfully, but these errors were encountered: