Skip to content

Releases: oobabooga/text-generation-webui

v1.4

24 Jul 19:42
a07d070

Choose a tag to compare

What's Changed

Bug fixes

  • Add checks for ROCm and unsupported architectures to llama_cpp_cuda loading by @jllllll in #3225

Extensions

  • [extensions/openai] Fixes for: embeddings, tokens, better errors. +Docs update, +Images, +logit_bias/logprobs, +more. by @matatonic in #3122

v1.3.1

19 Jul 14:22

Choose a tag to compare

Changes

  • Add missing EOS and BOS tokens to Llama-2 template
  • Bump transformers for better Llama-2 support
  • Bump llama-cpp-python for better unicode support (untested)

v1.3

18 Jul 20:33
3ef4939

Choose a tag to compare

Changes

  • Llama-v2: add instruction template, autodetect the truncation length, add conversion documentation
  • [GGML] Support for customizable RoPE by @randoentity in #3083
  • Optimize llamacpp_hf (a bit)
  • Add Airoboros-v1.2 template
  • Disable "Autoload the model" by default
  • Disable auto-loading at startup when only one model is available by @jllllll in #3187
  • Don't unset the LoRA menu when loading a model
  • Bump accelerate to 0.21.0
  • Bump bitsandbytes to 0.40.2 (Windows wheels provided by @jllllll in #3186)
  • Bump AutoGPTQ to 0.3.0 (loading LoRAs is now supported out of the box)
  • Update LLaMA-v1 documentation

Bug fixes

  • Use 'torch.backends.mps.is_available' to check if mps is supported by @appe233 in #3164

v1.2

16 Jul 05:44
9f08038

Choose a tag to compare

Changes

Bug fixes

  • Fix reload screen background color in dark mode

Extensions

v1.1.1

13 Jul 02:46

Choose a tag to compare

Bug fixes

  • Fix output path when downloading models through the UI

v1.1

12 Jul 18:56
6447b2e

Choose a tag to compare

Changes

  • Bump bitsandbytes Windows wheel by @jllllll in #3097 -- --load-in-4bit is now a lot faster
  • Add support low vram mode on llama.cpp module by @gabriel-pena in #3076
  • Add links/reference to new multimodal instructblip-pipeline in multimodal readme by @kjerk in #2947
  • Add token authorization for downloading model by @fahadh4ilyas in #3067
  • Add default environment variable values to docker compose file by @Josh-XT in #3102
  • models/config.yaml: +platypus/gplatty, +longchat, +vicuna-33b, +Redmond-Hermes-Coder, +wizardcoder, +more by @matatonic in #2928
  • Add context_instruct to API. Load default model instruction template … by @atriantafy in #2688
  • Chat history download creates more detailed file names by @UnskilledWolf in #3051
  • Disable wandb remote HTTP requests
  • Add Feature to Log Sample of Training Dataset for Inspection by @practicaldreamer in #1711
  • Add ability to load all text files from a subdirectory for training by @kizinfo in #1997
  • Add Tensorboard/Weights and biases integration for training by @kabachuha in #2624
  • Fix: Fixed the tokenization process of a raw dataset and improved its efficiency by @Nan-Do in #3035
  • More robust and error prone training by @FartyPants in #3058

Bug fixes

Extensions

v1.0

11 Jul 16:18
a81cdd1

Choose a tag to compare

Let's call it "version 1.0".

Windows/Linux/MacOS installers

09 Feb 16:16

Choose a tag to compare

1-click installers for Windows, Linux, MacOS, and WSL. Just download the zip, extract it, and double click on "start". The web UI and all its dependencies will be installed in the same folder.

The source codes and more information can be found at: https://github.com/oobabooga/one-click-installers

This is now obsolete! The one-click installers have been merged into the repository.