About 24,000 results
Open links in new tab
  1. Pacha - A TUI Frontend for llama.cpp #2071 - GitHub

    Jul 2, 2023 · This frontend is not meant to be a chat UI or to replace anything, but rather a tool to quickly test a model, a prompt style and/or certain parameters. I think this might be a good first stop to test …

  2. Benchmarks for llama_cpp and other backends here - GitHub

    Mar 28, 2024 · Here's my initial testing. Feel free to contact me if you want the actual test scripts as I'm hesitant to past the entirety here! EDITED to include numbers from running 15 tests of all models …

  3. llama.cpp/examples/server/utils.hpp at master - GitHub

    If tmpl is empty, we take the template from model metadata inline std::string format_chat (const common_chat_template & tmpl, const std::vector<json> & messages) { …

  4. Share a simpler Cmake methd to compile and run GPU accelerated

    Jul 10, 2023 · I browse all issues and the official setup tutorial of compiling llama.cpp to GPU. But I found it is really confused by using MAKE tool and copy file from a src path to a dest …

  5. Sampling special tokens, not just <|im_end|> ? #9886 - GitHub

    Both llama-cli and llama-server can sample special token <tool_call> during inference. Sorry for inconvenience.

  6. llama.cpp/gguf-py/pyproject.toml at master - GitHub

    [tool.poetry] name = "gguf" version = "0.13.0" description = "Read and write ML models in GGUF for GGML" authors = ["GGML <[email protected]>"] packages = [ {include = "gguf"}, {include = …

  7. Tool call support (Llama 3.x, Functionary v3, Hermes 2 Pro, Mistral ...

    Tool call support (Llama 3.x, Functionary v3, Hermes 2 Pro, Mistral Nemo, generic) w/ lazy grammars & minimalist Jinja engine #719

  8. llama.cpp/examples/server/tests/tests.sh at master - GitHub

    # Slow tests for tool calls need quite a few models ahead of time to avoid timing out.

  9. Chinese LLaMA and Alpaca available, seamlessly work with llama.cpp

    Mar 28, 2023 · Thank you for providing such a wonderful tool for LLM deployment on CPU. We released Chinese LLaMA and Alpaca model (with Chinese vocabulary extension!), trained with Alpaca-LoRA.

  10. Refactor: GGUF my Repo tool on HF needs its scripts updated ... - GitHub

    Jun 18, 2024 · Background Description On HF, GGUF my Repo scripts seem to currently use the old one, and LlamaCPP probably got updated, leaving the whole tool broken. Possible Refactor …