New Release #3175
vladmandic
announced in
Announcements
New Release
#3175
Replies: 1 comment 1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Update for 2024-05-28
Highlights for 2024-05-28
New SD.Next release has been baking in
dev
for a longer than usual, but changes are massive - about 350 commits for core and 300 for UI...Starting with the new UI - yup, this version ships with a preview of the new ModernUI
For details on how to enable and use it, see Home and WiKi
ModernUI is still in early development and not all features are available yet, please report issues and feedback
Thanks to @BinaryQuantumSoul for his hard work on this project!
What else?
New built-in features
List, preview, search through all your images and videos!
Yup, even popular DPM++2M can now run in 10 steps with quality equaling 30 steps using AYS presets
For details of above features and full list, see Changelog
New models
While still waiting for Stable Diffusion 3.0, there have been some significant models released in the meantime:
Note
SD.Next is no longer marked as a fork of A1111 and github project has been fully detached
Given huge number of changes with +3443/-3342 commits diff (at the time of fork detach) over the past year,
a completely different backend/engine and a change of focus, it is time to give credit to original author, and move on!
Full ChangeLog for 2024-05-28
Implemented as infinite-scroll with client-side-caching and lazy-loading while being fully async and non-blocking
Search or sort by path, name, size, width, height, mtime or any image metadata item, also with extended syntax like width > 1000
Settings: optional additional user-defined folders, thumbnails in fixed or variable aspect-ratio
Generate high-resolution images using your standard models without duplicates/distorsions AND improved performance
For example, SD15 can now go up to 2024x2048 and SDXL up to 4k natively
Simply enable checkbox in advanced menu and set desired resolution
Additional settings are available in settings -> inference settings -> hidiffusion
And can also be set and used via xyz grid
Note: HiDiffusion resolution sensitive, so if you get error, set resolution to be multiples of 128
PAG enhances sample quality by utilizing self-attention in formation of latent in addition to standard CFG scale
Simply set advanced -> attention guidance and advanced -> adaptive scaling
Additional options are available in settings -> inference settings -> pag
Note: PAG has replaced SAG as attention guidance method in SD.Next
Create transparent images with foreground-only being generated
Simply select from scripts -> apply to current model
All necessary files will be auto-downloaded on first use
Powerful method of using masking with ip-adapters
When combined with multiple ip-adapters, it allows for different inputs guidance for each segment of the input image
Hint: to create masks, you can use manually created masks or control->mask module with auto-segment to create masks and later upload them
Allows for more control over how each layer of ip-adapter is applied, requires a valid dict to be passed as input
See InstantStyle for details
As with all other compile engines, enable via settings -> compute settings -> compile
Newer alternative method to ToMe that can provide speed-up with minimal quality loss
Enable in settings -> inference settings -> token merging
Also available in XYZ grid
New method of outpainting that uses a combination of auto-masking and edge generation to create seamless transitions between original and generated image
Use on control tab:
"__*__"
in positive and negative prompts|
formatWorks similar to how VAE is selected and loaded: Set UNet folder and UNet model in settings
Can be replaced on-the-fly, not just during initial model load
Enables usage of fine-tunes such as DPO-SD15 or DPO-SDXL
Note: if there is a
JSON
file with the same name as the model it will be used as Unet config, otherwise Unet config from currently loaded model will be usedpixart-Σ is a high end diffusion Transformer model (DiT) with a T5 encoder/decoder capable of directly generating images at 4K resolution
to use, simply select from networks -> models -> reference -> PixArt-Σ
note: this is a very large model at ~22GB
set parameters: sampler: Default
sdxs is an extremely fast 1-step generation consistency model that also uses TAESD as quick VAE out-of-the-box
to use, simply select from networks -> models -> reference -> SDXS
set parameters: sampler: CMSI, steps: 1, cfg_scale: 0.0
sd15 and sdxl 1-step, 2-step, 4-step and 8-step optimized models using lora
set parameters: sampler: TCD or LCM, steps: 1/2/4/8, cfg_scale: 0.0
Standard (built-in themes), Modern (experimental nextgen ui), None (used for Gradio and Huggingface 3rd party themes)
Specifying a theme type updates list of available themes
For example, Gradio themes will not appear as available if theme type is set to Standard
/sdapi/v1/control
and CLI utilcli/simple-control.py
(in addition to previously added
/sdapi/v1/preprocessors
and/sdapi/v1/masking
)example:
/sdapi/v1/vqa
and CLI utilcli/simple-vqa.py
initial model load will now fetch config files required for the model
from the Huggingface site instead of using predefined YAML files
as both image-browser and controlnet have native built-in equivalents
both can still be installed by user if desired
Can improve memory utilization on compatible GPUs (RTX and newer)
You can enable/disable full torch profiling in settings top menu on-the-fly
Useful for batches as prompt processing is ~0.1sec on each pass
SD_PROMPT_DEBUG
to show actual tokens usedFor standalone, simply copy safetensors file to
models/control/controlnet
folderFor diffusers format, create folder with model name in
models/control/controlnet/
and copy
model.json
anddiffusion_pytorch_model.safetensors
to that foldernote: you can use other samplers as well with SDXL-Lightning models
Note: changing timestep spacing changes behavior of sampler and can help to make any sampler turbo/lightning compatibile
Additionally, presets from nVidias align-you-steps reasearch are provided
Result is that perfectly aligned steps can drastically reduce number of steps needed!
For example, AYS preset alows DPM++2M to run in ~10 steps with quality equallying ~30 steps!
requires removing the venv folder to update properly
IPEXRUN=True
if you want to useipexrun
--reinstall-zluda
(to download the latest ZLUDA)Beta Was this translation helpful? Give feedback.
All reactions