Skip to content

Releases: YellowRoseCx/koboldcpp-rocm

KoboldCPP-v1.83.1.yr1-ROCm

18 Feb 23:33
30e4827
Compare
Choose a tag to compare
p. sure I figured it out CMake Flag Fix

KoboldCPP-v1.82.4.yr0-ROCm

28 Jan 10:50
Compare
Choose a tag to compare

Apparently there's starting to be trouble again with the "unofficially supported" ROCm GPUs..I'm trying to look into it when I'm at home and able to. If the regular koboldcpp_rocm.exe doesn't work for you, try the rocm-5.7 version please

KoboldCPP-v1.82.1.yr0-ROCm

21 Jan 02:55
Compare
Choose a tag to compare
Merge remote-tracking branch 'upstream/concedo'

KoboldCPP-v1.82.yr0-ROCm

20 Jan 07:00
Compare
Choose a tag to compare
Pre-release
Merge remote-tracking branch 'upstream/concedo'

KoboldCPP-v1.81.1.yr0-ROCm

10 Jan 15:30
Compare
Choose a tag to compare
Pre-release
Merge remote-tracking branch 'upstream/concedo'

KoboldCPP-v1.80.3.yr0-ROCm

30 Dec 07:47
073fdee
Compare
Choose a tag to compare
Update cmake-rocm-windows.yml

KoboldCPP-v1.79.1.yr1-ROCm

05 Dec 03:43
d6949d6
Compare
Choose a tag to compare
attempt 6700xt fix for cmake-rocm-windows.yml

KoboldCPP-v1.79.1.yr0-ROCm

03 Dec 11:02
d6949d6
Compare
Choose a tag to compare
attempt 6700xt fix for cmake-rocm-windows.yml

KoboldCPP-v1.78.yr0-ROCm

19 Nov 06:36
Compare
Choose a tag to compare

koboldcpp-rocm-1.78

image

  • NEW: Added support for Flux and Stable Diffusion 3.5 models: Image generation has been updated with new arch support (thanks to stable-diffusion.cpp) with additional enhancements. You can use either fp16 or fp8 safetensor models, or the GGUF models. Supports all-in-one models (bundled T5XXL, Clip-L/G, VAE) or loading them individually.
  • Debug mode prints penalties for XTC
  • Added a new flag --nofastforward, this forces full prompt reprocessing on every request. It can potentially give more repeatable/reliable/consistent results in some cases.
  • CLBlast support is still retained, but has been further downgraded to "compatibility mode" and is no longer recommended (use Vulkan instead). CLBlast GPU offload must now maintain duplicate a copy of the layers in RAM as well, as it now piggybacks off the CPU backend.
  • Added common identity provider /.well-known/serviceinfo Haidra-Org/AI-Horde#466 PygmalionAI/aphrodite-engine#807 theroyallab/tabbyAPI#232
  • Reverted some changes that reduced speed in HIPBLAS.
  • Fixed a bug where bad logprobs JSON was output when logits were -Infinity
  • Updated Kobold Lite, multiple fixes and improvements
    • Added support for custom CSS styles
    • Added support for generating larger images (select BigSquare in image gen settings)
    • Fixed some streaming issues when connecting to Tabby backend
    • Better world info length limiting (capped at 50% of max context before appending to memory)
    • Added support for Clip Skip for local image generation.
  • Merged fixes and improvements from upstream

To use, download and run the koboldcpp_rocm.exe, which is a one-file pyinstaller.
If you're using Linux, clone the repo and build in terminal with make LLAMA_HIPBLAS=1 -j

Run it from the command line with the desired launch parameters (see --help), or manually select the model in the GUI.
and then once loaded, you can connect like this (or use the full koboldai client):
http://localhost:5001

For more information, be sure to run the program from command line with the --help flag.
Release notes from: https://github.com/LostRuins/koboldcpp/releases/tag/v1.78

KoboldCPP-v1.77.yr1-ROCm

06 Nov 18:57
Compare
Choose a tag to compare
  • Bring Speed Back

upstream llama.cpp introduced a change to calculate certain values in full 32 bit precision by default which introduced a major slow down for some users with AMD GPUs, this reverts that change until improvements are made