|
1 | 1 | # Change Log for SD.Next
|
2 | 2 |
|
| 3 | +## Update for 2024-06-23 |
| 4 | + |
| 5 | +### Highlights for 2024-06-23 |
| 6 | + |
| 7 | +Following zero-day **SD3** release, a 10 days later here's a refresh with 10+ improvements |
| 8 | +including full prompt attention, support for compressed weights, additional text-encoder quantization modes. |
| 9 | + |
| 10 | +But there's more than SD3: |
| 11 | +- support for quantized **T5** text encoder *FP16/FP8/FP4/INT8* in all models that use T5: SD3, PixArt-Σ, etc. |
| 12 | +- support for **PixArt-Sigma** in small/medium/large variants |
| 13 | +- support for **HunyuanDiT 1.1** |
| 14 | +- additional **NNCF weights compression** support: SD3, PixArt, ControlNet, Lora |
| 15 | +- integration of **MS Florence** VLM/VQA *Base* and *Large* models |
| 16 | +- (finally) new release of **Torch-DirectML** |
| 17 | +- additional efficiencies for users with low VRAM GPUs |
| 18 | +- over 20 overall fixes |
| 19 | + |
| 20 | +### Model Improvements |
| 21 | + |
| 22 | +- **SD3**: enable tiny-VAE (TAESD) preview and non-full quality mode |
| 23 | +- SD3: enable base LoRA support |
| 24 | +- SD3: add support for FP4 quantized T5 text encoder |
| 25 | + simply select in *settings -> model -> text encoder* |
| 26 | + *note* for SD3 with T5, set SD.Next to use FP16 precision, not BF16 precision |
| 27 | +- SD3: add support for INT8 quantized T5 text encoder, thanks @Disty0! |
| 28 | +- SD3: enable cpu-offloading for T5 text encoder, thanks @Disty0! |
| 29 | +- SD3: simplified loading of model in single-file safetensors format |
| 30 | + model load can now be performed fully offline |
| 31 | +- SD3: full support for prompt parsing and attention, thanks @AI-Casanova! |
| 32 | +- SD3: ability to target different prompts to each of text-encoders, thanks @AI-Casanova! |
| 33 | + example: `dog TE2: cat TE3: bird` |
| 34 | +- SD3: add support for sampler shift for Euler FlowMatch |
| 35 | + see *settings -> samplers*, also available as param in xyz grid |
| 36 | + higher shift means model will spend more time on structure and less on details |
| 37 | +- SD3: add support for selecting T5 text encoder variant in XYZ grid |
| 38 | +- **Pixart-Σ**: Add *small* (512px) and *large* (2k) variations, in addition to existing *medium* (1k) |
| 39 | +- Pixart-Σ: Add support for 4/8bit quantized t5 text encoder |
| 40 | + *note* by default pixart-Σ uses full fp16 t5 encoder with large memory footprint |
| 41 | + simply select in *settings -> model -> text encoder* before or after model load |
| 42 | +- **HunyuanDiT**: support for model version 1.1 |
| 43 | +- **MS Florence**: integration of Microsoft Florence VLM/VQA Base and Large models |
| 44 | + simply select in *process -> visual query*! |
| 45 | + |
| 46 | +### General Improvements |
| 47 | + |
| 48 | +- support FP4 quantized T5 text encoder, in addtion to existing FP8 and FP16 |
| 49 | +- support for T5 text-encoder loader in **all** models that use T5 |
| 50 | + *example*: load FP4 or FP8 quantized T5 text-encoder into PixArt Sigma! |
| 51 | +- support for `torch-directml` **0.2.2**, thanks @lshqqytiger! |
| 52 | + *note*: new directml is finally based on modern `torch` 2.3.1! |
| 53 | +- xyz grid: add support for LoRA selector |
| 54 | +- vae load: store original vae so it can be restored when set to none |
| 55 | +- extra networks: info display now contains link to source url if model if its known |
| 56 | + works for civitai and huggingface models |
| 57 | +- force gc for lowvram users and improve gc logging |
| 58 | +- improved google.colab support |
| 59 | +- css tweaks for standardui |
| 60 | +- css tweaks for modernui |
| 61 | +- additional torch gc checks, thanks @Disty0! |
| 62 | + |
| 63 | +**Improvements: NNCF**, thanks @Disty0! |
| 64 | +- SD3 and PixArt support |
| 65 | +- moved the first compression step to CPU |
| 66 | +- sequential cpu offload (lowvram) support |
| 67 | +- Lora support without reloading the model |
| 68 | +- ControlNet compression support |
| 69 | + |
| 70 | +### Fixes |
| 71 | + |
| 72 | +- fix unsaturated outputs, force apply vae config on model load |
| 73 | +- fix hidiffusion handling of non-square aspect ratios, thanks @ShenZhang-Shin! |
| 74 | +- fix control second pass resize |
| 75 | +- fix hunyuandit set attention processor |
| 76 | +- fix civitai download without name |
| 77 | +- fix compatibility with latest adetailer |
| 78 | +- fix invalid sampler warning |
| 79 | +- fix starting from non git repo |
| 80 | +- fix control api negative prompt handling |
| 81 | +- fix saving style without name provided |
| 82 | +- fix t2i-color adapter |
| 83 | +- fix sdxl "has been incorrectly initialized" |
| 84 | +- fix api face-hires |
| 85 | +- fix api ip-adapter |
| 86 | +- fix memory exceptions with ROCm, thanks @Disty0! |
| 87 | +- fix face-hires with lowvram, thanks @Disty0! |
| 88 | +- fix pag incorrectly resetting pipeline |
| 89 | +- cleanup image metadata |
| 90 | +- restructure api examples: `cli/api-*` |
| 91 | +- handle theme fallback when invalid theme is specified |
| 92 | +- remove obsolete training code leftovers |
| 93 | + |
3 | 94 | ## Update for 2024-06-13
|
4 | 95 |
|
5 | 96 | ### Highlights for 2024-06-13
|
|
0 commit comments