Download Latest Version koboldcpp-1.111.2 source code.zip (64.1 MB)
Email in envelope

Get an email when there's a new version of KoboldCpp

Home / v1.53
Name Modified Size InfoDownloads / Week
Parent folder
repacked-koboldcpp-1.53.zip 2023-12-28 283.5 MB
koboldcpp-linux-x64 2023-12-23 359.8 MB
koboldcpp.exe 2023-12-23 289.5 MB
koboldcpp_nocuda.exe 2023-12-23 26.4 MB
koboldcpp-1.53 source code.tar.gz 2023-12-23 12.4 MB
koboldcpp-1.53 source code.zip 2023-12-23 12.5 MB
README.md 2023-12-23 1.7 kB
Totals: 7 Items   984.1 MB 0

koboldcpp-1.53

  • Added support for SSL. You can now import your own SSL cert to use with KoboldCpp and serve it over HTTPS with --ssl [cert.pem] [key.pem] or via the GUI. The .pem files must be unencrypted, you can also generate them with OpenSSL, eg. openssl req -x509 -newkey rsa:4096 -keyout key.pem -out cert.pem -sha256 -days 365 -config openssl.cnf -nodes for your own self signed certificate.
  • Added support for presence penalty (alternative rep pen) over the KAI API and in Lite. If Presence Penalty is set over the OpenAI API, and rep_pen is not set, then rep_pen will be set to a default of 1.0 instead of 1.1. Both penalties can be used together, although this is probably not a good idea.
  • Added fixes for Broken Pipe error, thanks @mahou-shoujo.
  • Added fixes for aborting ongoing connections while streaming in SillyTavern.
  • Merged upstream support for Phi models and speedups for Mixtral
  • The default non-blas batch size for GGUF models is now increased from 8 to 32.
  • Merged HIPBlas fixes from @YellowRoseCx
  • Fixed an issue with building convert tools in 1.52

To use, download and run the koboldcpp.exe, which is a one-file pyinstaller. If you don't need CUDA, you can use koboldcpp_nocuda.exe which is much smaller. If you're using AMD, you can try koboldcpp_rocm at YellowRoseCx's fork here

Run it from the command line with the desired launch parameters (see --help), or manually select the model in the GUI. and then once loaded, you can connect like this (or use the full koboldai client): http://localhost:5001

For more information, be sure to run the program from command line with the --help flag.

Source: README.md, updated 2023-12-23