-
Notifications
You must be signed in to change notification settings - Fork 572
Add llama-cli example to llama.cpp snippets #1889
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
- Add llama-cli command alongside llama-server for all installation methods - Update llama-server comment to mention web UI - Build from source now compiles both llama-server and llama-cli targets
pcuenca
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Very nice, thanks a lot for taking the time to do this!
Wauplin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM 👍
julien-c
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nice!
cc @ngxson as well
ngxson
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Perfect, thanks!! 🚀
| const snippet = ["# Load and run the model:", `${binary} -hf ${model.id}${getQuantTag(filepath)}`]; | ||
| const serverCommand = (binary: string) => { | ||
| const snippet = [ | ||
| "# Start a local OpenAI-compatible server with a web UI:", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
small nits (not related to the current PR), can we modify the copy button such that it doesn't copy the line starts with # (I mean comment lines)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
yes good catch @copilot open an issue about it 👀
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@copilot please open an issue about this
Add
llama-clicommand alongsidellama-serverfor all installation methods.Changes
llama-clisnippet for direct terminal inferencellama-servercomment to mention web UI