Download Latest Version koboldcpp-1.111.2 source code.zip (64.1 MB)
Email in envelope

Get an email when there's a new version of KoboldCpp

Home / v1.11
Name Modified Size InfoDownloads / Week
Parent folder
koboldcpp.exe 2023-04-22 20.0 MB
koboldcpp-1.11 source code.tar.gz 2023-04-22 9.2 MB
koboldcpp-1.11 source code.zip 2023-04-22 9.2 MB
README.md 2023-04-22 1.7 kB
Totals: 4 Items   38.4 MB 0

koboldcpp-1.11

  • Now has GPT-NeoX / Pythia / StableLM support!
  • Try my special model, Pythia-70m-ChatSalad here: https://huggingface.co/concedo/pythia-70m-chatsalad-ggml/tree/main
  • Added upstream LORA file support for llama, use the --lora parameter.
  • Added limited fast-forwarding capabilities for RWKV, context can be reused if its completely unmodified.
  • Kobold Lite now supports using an additional custom stopping sequence, edit it in the Memory panel.
  • Updated Kobold Lite, and pulled llama improvements from upstream.
  • Improved OSX and Linux build support - now automatically builds all libraries with the requested flags, and you can select which ones to use at runtime. Example: do a make LLAMA_OPENBLAS=1 LLAMA_CLBLAST=1 and it will build both OpenBlas and CLBlast libraries on your platform, then you select clblast with --useclblast at runtime.

To use, download and run the koboldcpp.exe, which is a one-file pyinstaller. Alternatively, drag and drop a compatible ggml model on top of the .exe, or run it and manually select the model in the popup dialog.

and then once loaded, you can connect like this (or use the full koboldai client): http://localhost:5001

For more information, be sure to run the program with the --help flag.

Alternative Options: Non-AVX2 version now included in the same .exe file, enable with --noavx2 flags Big context too slow? Try the --smartcontext flag to reduce prompt processing frequency Run with your GPU using CLBlast, with --useclblast flag for a speedup

Disclaimer: This version has Cloudflare Insights in the Kobold Lite UI, which was subsequently removed in v1.17

Source: README.md, updated 2023-04-22