mirror of https://github.com/vladmandic/automatic
commit
fad3f41b49
96
CHANGELOG.md
96
CHANGELOG.md
|
|
@ -87,7 +87,7 @@
|
|||
|
||||
## Update for 2023-12-04
|
||||
|
||||
What's new? Native video in SD.Next via both **AnimateDiff** and **Stable-Video-Diffusion** - and including native MP4 encoding and smooth video outputs out-of-the-box, not just animated-GIFs.
|
||||
Whats new? Native video in SD.Next via both **AnimateDiff** and **Stable-Video-Diffusion** - and including native MP4 encoding and smooth video outputs out-of-the-box, not just animated-GIFs.
|
||||
Also new is support for **SDXL-Turbo** as well as new **Kandinsky 3** models and cool latent correction via **HDR controls** for any *txt2img* workflows, best-of-class **SDXL model merge** using full ReBasin methods and further mobile UI optimizations.
|
||||
|
||||
- **Diffusers**
|
||||
|
|
@ -106,7 +106,7 @@ Also new is support for **SDXL-Turbo** as well as new **Kandinsky 3** models and
|
|||
- for AnimateDiff support in *Original* backend, use standard *AnimateDiff* extension
|
||||
- **HDR latent control**, based on [article](https://huggingface.co/blog/TimothyAlexisVass/explaining-the-sdxl-latent-space#long-prompts-at-high-guidance-scales-becoming-possible)
|
||||
- in *Advanced* params
|
||||
- allows control of *latent clamping*, *color centering* and *range maximimization*
|
||||
- allows control of *latent clamping*, *color centering* and *range maximization*
|
||||
- supported by *XYZ grid*
|
||||
- [SD21 Turbo](https://huggingface.co/stabilityai/sd-turbo) and [SDXL Turbo](<https://huggingface.co/stabilityai/sdxl-turbo>) support
|
||||
- just set CFG scale (0.0-1.0) and steps (1-3) to a very low value
|
||||
|
|
@ -164,7 +164,7 @@ Also included are several other improvements and large number of hotfixes - see
|
|||
- Add `cli/lcm-convert.py` script to convert any SD 1.5 or SD-XL model to LCM model
|
||||
by baking in LORA and uploading to Huggingface, thanks @Disty0
|
||||
- Support for [Stable Fast](https://github.com/chengzeyi/stable-fast) model compile on *Windows/Linux/WSL2* with *CUDA*
|
||||
See [Wiki:Benchmark](https://github.com/vladmandic/automatic/wiki/Benchmark) for details and comparisment
|
||||
See [Wiki:Benchmark](https://github.com/vladmandic/automatic/wiki/Benchmark) for details and comparison
|
||||
of different backends, precision modes, advanced settings and compile modes
|
||||
*Hint*: **70+ it/s** is possible on *RTX4090* with no special tweaks
|
||||
- Add additional pipeline types for manual model loads when loading from `safetensors`
|
||||
|
|
@ -183,7 +183,7 @@ Also included are several other improvements and large number of hotfixes - see
|
|||
and heavily modified for additional functionality and fully integrated by @AI-Casanova (thanks!)
|
||||
- merge SD or SD-XL models using *simple merge* (12 methods),
|
||||
using one of *presets* (20 built-in presets) or custom block merge values
|
||||
- merge with ReBasin permuatations and/or clipping protection
|
||||
- merge with ReBasin permutations and/or clipping protection
|
||||
- fully multithreaded for fastest merge possible
|
||||
- **Model update**
|
||||
- under UI -> Models - Update
|
||||
|
|
@ -232,7 +232,7 @@ Also, [Wiki](https://github.com/vladmandic/automatic/wiki) has been updated with
|
|||
Some highlights: [OpenVINO](https://github.com/vladmandic/automatic/wiki/OpenVINO), [IntelArc](https://github.com/vladmandic/automatic/wiki/Intel-ARC), [DirectML](https://github.com/vladmandic/automatic/wiki/DirectML), [ONNX/Olive](https://github.com/vladmandic/automatic/wiki/ONNX-Olive)
|
||||
|
||||
- **Diffusers**
|
||||
- since now **SD.Next** supports **12** different model types, we've added reference model for each type in
|
||||
- since now **SD.Next** supports **12** different model types, weve added reference model for each type in
|
||||
*Extra networks -> Reference* for easier select & auto-download
|
||||
Models can still be downloaded manually, this is just a convenience feature & a showcase for supported models
|
||||
- new model type: [Segmind SSD-1B](https://huggingface.co/segmind/SSD-1B)
|
||||
|
|
@ -245,7 +245,7 @@ Some highlights: [OpenVINO](https://github.com/vladmandic/automatic/wiki/OpenVIN
|
|||
- new model type: [Pixart-α XL 2](https://github.com/PixArt-alpha/PixArt-alpha)
|
||||
in medium/512px and large/1024px variations
|
||||
comparable in quality to SD 1.5 and SD-XL, but with better text encoder and highly optimized training pipeline
|
||||
so finetunes can be done in as little as 10% compared to SD/SD-XL (note that due to much larger text encodeder, it is a large model)
|
||||
so finetunes can be done in as little as 10% compared to SD/SD-XL (note that due to much larger text encoder, it is a large model)
|
||||
select from extra networks -> reference or download using built-in **Huggingface** downloader: `PixArt-alpha/PixArt-XL-2-1024-MS`
|
||||
- new model type: [LCM: Latent Consistency Models](https://github.com/openai/consistency_models)
|
||||
trained at 512px, but with near-instant generate in a as little as 3 steps!
|
||||
|
|
@ -266,7 +266,7 @@ Some highlights: [OpenVINO](https://github.com/vladmandic/automatic/wiki/OpenVIN
|
|||
- extend support for [Free-U](https://github.com/ChenyangSi/FreeU)
|
||||
improve generations quality at no cost (other than finding params that work for you)
|
||||
- **General**
|
||||
- attempt to auto-fix invalid samples which occure due to math errors in lower precision
|
||||
- attempt to auto-fix invalid samples which occur due to math errors in lower precision
|
||||
example: `RuntimeWarning: invalid value encountered in cast: sample = sample.astype(np.uint8)`
|
||||
begone **black images** *(note: if it proves as working, this solution will need to be expanded to cover all scenarios)*
|
||||
- add **Lora OFT** support, thanks @antis0007 and @ai-casanova
|
||||
|
|
@ -326,8 +326,8 @@ Some highlights: [OpenVINO](https://github.com/vladmandic/automatic/wiki/OpenVIN
|
|||
- fix diffusers switch from invalid model
|
||||
- force second requirements check on startup
|
||||
- remove **lyco**, multiple_tqdm
|
||||
- enhance extension compatibility for exensions directly importing codeformers
|
||||
- enhance extension compatibility for exensions directly accessing processing params
|
||||
- enhance extension compatibility for extensions directly importing codeformers
|
||||
- enhance extension compatibility for extensions directly accessing processing params
|
||||
- **css** fixes
|
||||
- clearly mark external themes in ui
|
||||
- update `typing-extensions`
|
||||
|
|
@ -336,8 +336,8 @@ Some highlights: [OpenVINO](https://github.com/vladmandic/automatic/wiki/OpenVIN
|
|||
|
||||
This is a major release, with many changes and new functionality...
|
||||
|
||||
Changelog is massive, but do read through or you'll be missing on some very cool new functionality
|
||||
or even free speedups and quality improvements (regardless of which workflows you're using)!
|
||||
Changelog is massive, but do read through or youll be missing on some very cool new functionality
|
||||
or even free speedups and quality improvements (regardless of which workflows youre using)!
|
||||
|
||||
Note that for this release its recommended to perform a clean install (e.g. fresh `git clone`)
|
||||
Upgrades are still possible and supported, but clean install is recommended for best experience
|
||||
|
|
@ -391,7 +391,7 @@ Upgrades are still possible and supported, but clean install is recommended for
|
|||
- **General**
|
||||
- faster search, ability to show/hide/sort networks
|
||||
- refactored subfolder handling
|
||||
*note*: this will trigger model hash recaclulation on first model use
|
||||
*note*: this will trigger model hash recalculation on first model use
|
||||
- **Diffusers**:
|
||||
- better pipeline **auto-detect** when loading from safetensors
|
||||
- **SDXL Inpaint**
|
||||
|
|
@ -553,14 +553,14 @@ Upgrades are still possible and supported, but clean install is recommended for
|
|||
- add end-to-end example how to use API: `cli/simple-txt2img.js`
|
||||
covers txt2img, upscale, hires, refiner
|
||||
- **train.py**
|
||||
- wrapper script around built-in **kohya's lora** training script
|
||||
- wrapper script around built-in **kohyas lora** training script
|
||||
see `cli/train.py --help`
|
||||
new support for sd and sdxl, thanks @evshiron
|
||||
new support for full offline mode (without sdnext server running)
|
||||
- **Themes**
|
||||
- all built-in themes are fully supported:
|
||||
- *black-teal (default), light-teal, black-orange, invoked, amethyst-nightfall, midnight-barbie*
|
||||
- if you're using any **gradio default** themes or a **3rd party** theme or that are not optimized for SD.Next, you may experience issues
|
||||
- if youre using any **gradio default** themes or a **3rd party** theme or that are not optimized for SD.Next, you may experience issues
|
||||
default minimal style has been updated for compatibility, but actual styling is completely outside of SD.Next control
|
||||
|
||||
## Update for 2023-09-13
|
||||
|
|
@ -581,7 +581,7 @@ Major changes how **hires** works as well as support for a very interesting new
|
|||
but operations are really resource intensive as it includes: *base->decode->upscale->encode->hires->refine*
|
||||
- all combinations of: decode full/quick + upscale none/latent/non-latent + hires on/off + refiner on/off
|
||||
should be supported, but given the number of combinations, issues are possible
|
||||
- all operations are captured in image medata
|
||||
- all operations are captured in image metadata
|
||||
- diffusers:
|
||||
- allow loading of sd/sdxl models from safetensors without online connectivity
|
||||
- support for new model: [wuerstchen](https://huggingface.co/warp-ai/wuerstchen)
|
||||
|
|
@ -616,7 +616,7 @@ One week later, another large update!
|
|||
- system:
|
||||
- full **python 3.11** support
|
||||
note that changing python version does require reinstall
|
||||
and if you're already on python 3.10, really no need to upgrade
|
||||
and if youre already on python 3.10, really no need to upgrade
|
||||
- themes:
|
||||
- new default theme: **black-teal**
|
||||
- new light theme: **light-teal**
|
||||
|
|
@ -640,9 +640,9 @@ One week later, another large update!
|
|||
thanks @ai-casanova for porting compel/sdxl code
|
||||
- mix&match **base** and **refiner** models (*experimental*):
|
||||
most of those are "because why not" and can result in corrupt images, but some are actually useful
|
||||
also note that if you're not using actual refiner model, you need to bump refiner steps
|
||||
also note that if youre not using actual refiner model, you need to bump refiner steps
|
||||
as normal models are not designed to work with low step count
|
||||
and if you're having issues, try setting prompt parser to "fixed attention" as majority of problems
|
||||
and if youre having issues, try setting prompt parser to "fixed attention" as majority of problems
|
||||
are due to token mismatches when using prompt attention
|
||||
- any sd15 + any sd15
|
||||
- any sd15 + sdxl-refiner
|
||||
|
|
@ -743,18 +743,18 @@ Smaller update, but with some breaking changes (to prepare for future larger fun
|
|||
thanks @evshiron
|
||||
- fix **img2img** resizing (applies to *original, diffusers, hires*)
|
||||
- config change: main `config.json` no longer contains entire configuration
|
||||
but only differences from defaults (simmilar to recent change performed to `ui-config.json`)
|
||||
but only differences from defaults (similar to recent change performed to `ui-config.json`)
|
||||
- diffusers:
|
||||
- enable **batch img2img** workflows
|
||||
- original:
|
||||
- new samplers: **dpm++ 3M sde** (standard and karras variations)
|
||||
enable in *settings -> samplers -> show samplers*
|
||||
- expose always/never discard penultimage sigma
|
||||
- expose always/never discard penultimate sigma
|
||||
enable in *settings -> samplers*
|
||||
|
||||
## Update for 2023-08-11
|
||||
|
||||
This is a big one that's been cooking in `dev` for a while now, but finally ready for release...
|
||||
This is a big one thats been cooking in `dev` for a while now, but finally ready for release...
|
||||
|
||||
- diffusers:
|
||||
- **pipeline autodetect**
|
||||
|
|
@ -800,7 +800,7 @@ Another minor update, but it unlocks some cool new items...
|
|||
Smaller release, but IMO worth a post...
|
||||
|
||||
- diffusers:
|
||||
- sd-xl lora's are now supported!
|
||||
- sd-xl loras are now supported!
|
||||
- memory optimizations: Enhanced sequential CPU offloading, model CPU offload, FP16 VAE
|
||||
- significant impact if running SD-XL (for example, but applies to any model) with only 8GB VRAM
|
||||
- update packages
|
||||
|
|
@ -835,7 +835,7 @@ Actual changelog is:
|
|||
- add backend and operation info to metadata
|
||||
|
||||
- diffusers:
|
||||
- we're out of experimental phase and diffusers backend is considered stable
|
||||
- were out of experimental phase and diffusers backend is considered stable
|
||||
- sd-xl: support for **sd-xl 1.0** official model
|
||||
- sd-xl: loading vae now applies to both base and refiner and saves a bit of vram
|
||||
- sd-xl: denoising_start/denoising_end
|
||||
|
|
@ -850,7 +850,7 @@ Actual changelog is:
|
|||
- vae: enable loading of pure-safetensors vae files without config
|
||||
also enable *automatic* selection to work with diffusers
|
||||
- sd-xl: initial lora support
|
||||
right now this applies to official lora released by **stability-ai**, support for **kohya's** lora is expected soon
|
||||
right now this applies to official lora released by **stability-ai**, support for **kohyas** lora is expected soon
|
||||
- implement img2img and inpainting (experimental)
|
||||
actual support and quality depends on model
|
||||
it works as expected for sd 1.5, but not so much for sd-xl for now
|
||||
|
|
@ -868,7 +868,7 @@ Actual changelog is:
|
|||
|
||||
## Update for 2023-07-18
|
||||
|
||||
While we're waiting for official SD-XL release, here's another update with some fixes and enhancements...
|
||||
While were waiting for official SD-XL release, heres another update with some fixes and enhancements...
|
||||
|
||||
- **global**
|
||||
- image save: option to add invisible image watermark to all your generated images
|
||||
|
|
@ -1007,7 +1007,7 @@ A bit bigger update this time, but contained to specific areas...
|
|||
- **live server log monitoring** in ui
|
||||
configurable via settings -> live preview
|
||||
- new **extra networks interface**
|
||||
*note: if you're using a 3rd party ui extension for extra networks, it will likely need to be updated to work with new interface*
|
||||
*note: if youre using a 3rd party ui extension for extra networks, it will likely need to be updated to work with new interface*
|
||||
- display in front of main ui, inline with main ui or as a sidebar
|
||||
- lazy load thumbnails
|
||||
drastically reduces load times for large number of extra networks
|
||||
|
|
@ -1030,7 +1030,7 @@ A bit bigger update this time, but contained to specific areas...
|
|||
|
||||
## Update for 2023-06-20
|
||||
|
||||
This one is less relevant for standard users, but pretty major if you're running an actual server
|
||||
This one is less relevant for standard users, but pretty major if youre running an actual server
|
||||
But even if not, it still includes bunch of cumulative fixes since last release - and going by number of new issues, this is probably the most stable release so far...
|
||||
(next one is not going to be as stable, but it will be fun :) )
|
||||
|
||||
|
|
@ -1050,11 +1050,11 @@ Second stage of a jumbo merge from upstream plus few minor changes...
|
|||
- simplify token merging
|
||||
- reorganize some settings
|
||||
- all updates from upstream: **A1111** v1.3.2 [df004be] *(latest release)*
|
||||
pretty much nothing major that i haven't released in previous versions, but its still a long list of tiny changes
|
||||
pretty much nothing major that i havent released in previous versions, but its still a long list of tiny changes
|
||||
- skipped/did-not-port:
|
||||
add separate hires prompt: unnecessarily complicated and spread over large number of commits due to many regressions
|
||||
allow external scripts to add cross-optimization methods: dangerous and i don't see a use case for it so far
|
||||
load extension info in threads: unnecessary as other optimizations i've already put place perform equally good
|
||||
allow external scripts to add cross-optimization methods: dangerous and i dont see a use case for it so far
|
||||
load extension info in threads: unnecessary as other optimizations ive already put place perform equally good
|
||||
- broken/reverted:
|
||||
sub-quadratic optimization changes
|
||||
|
||||
|
|
@ -1089,7 +1089,7 @@ Both some **new functionality** as well as **massive merges** from upstream
|
|||
- updated **settings** tab functionality, thanks @gegell
|
||||
with real-time monitor for all new and/or updated settings
|
||||
- **launcher** will now warn if application owned files are modified
|
||||
you are free to add any user files, but do not modify app files unless you're sure in what you're doing
|
||||
you are free to add any user files, but do not modify app files unless youre sure in what youre doing
|
||||
- add more profiling for scripts/extensions so you can see what takes time
|
||||
this applies both to initial load as well as execution
|
||||
- experimental `sd_model_dict` setting which allows you to load model dictionary
|
||||
|
|
@ -1107,9 +1107,9 @@ that caused my phone to go crazy with notifications over the weekend...
|
|||
- new **vae decode** method to help with larger batch sizes, thanks @bigdog
|
||||
- new setting -> lora -> **use lycoris to handle all lora types**
|
||||
this is still experimental, but the goal is to obsolete old built-in lora module
|
||||
as it doesn't understand many new loras and built-in lyco module can handle it all
|
||||
as it doesnt understand many new loras and built-in lyco module can handle it all
|
||||
- somewhat optimize browser page loading
|
||||
still slower than i'd want, but gradio is pretty bad at this
|
||||
still slower than id want, but gradio is pretty bad at this
|
||||
- profiling of scripts/extensions callbacks
|
||||
you can now see how much or pre/post processing is done, not just how long generate takes
|
||||
- additional exception handling so bad exception does not crash main app
|
||||
|
|
@ -1165,17 +1165,17 @@ Some quality-of-life improvements...
|
|||
- `cuda` => Scaled-Dot-Product
|
||||
- `rocm` => Sub-quadratic
|
||||
- `directml` => Sub-quadratic
|
||||
- `ipex` => InvokeAI's
|
||||
- `mps` => Doggettx's
|
||||
- `cpu` => Doggettx's
|
||||
- `ipex` => invokeais
|
||||
- `mps` => Doggettxs
|
||||
- `cpu` => Doggettxs
|
||||
- optimize logging
|
||||
- optimize profiling
|
||||
now includes startup profiling as well as `cuda` profiling during generate
|
||||
- minor lightbox improvements
|
||||
- bugfixes...i don't recall when was a release with at least several of those
|
||||
- bugfixes...i dont recall when was a release with at least several of those
|
||||
|
||||
other than that - first stage of [Diffusers](https://github.com/huggingface/diffusers) integration is now in master branch
|
||||
i don't recommend anyone to try it (and dont even think reporting issues for it)
|
||||
i dont recommend anyone to try it (and dont even think reporting issues for it)
|
||||
but if anyone wants to contribute, take a look at [project page](https://github.com/users/vladmandic/projects/1/views/1)
|
||||
|
||||
## Update for 2023-05-23
|
||||
|
|
@ -1228,11 +1228,11 @@ but hopefully it will go ok...
|
|||
## Update for 2023-05-15
|
||||
|
||||
- major work on **prompt parsing**
|
||||
this can cause some differences in results compared to what you're used to, but its all about fixes & improvements
|
||||
this can cause some differences in results compared to what youre used to, but its all about fixes & improvements
|
||||
- prompt parser was adding commas and spaces as separate words and tokens and/or prefixes
|
||||
- negative prompt weight using `[word:weight]` was ignored, it was always `0.909`
|
||||
- bracket matching was anything but correct. complex nested attention brackets are now working.
|
||||
- btw, if you run with `--debug` flag, you'll now actually see parsed prompt & schedule
|
||||
- btw, if you run with `--debug` flag, youll now actually see parsed prompt & schedule
|
||||
- updated all scripts in `/cli`
|
||||
- add option in settings to force different **latent sampler** instead of using primary only
|
||||
- add **interrupt/skip** capabilities to process images
|
||||
|
|
@ -1283,7 +1283,7 @@ Again, few days later...
|
|||
- reviewed/ported **all** commits from **A1111** upstream
|
||||
some a few are not applicable as i already have alternative implementations
|
||||
and very few i choose not to implement (save/restore last-known-good-config is a bad hack)
|
||||
otherwise, we're fully up to date (its doesn't show on fork status as code merges were mostly manual due to conflicts)
|
||||
otherwise, were fully up to date (it doesnt show on fork status as code merges were mostly manual due to conflicts)
|
||||
but...due to sheer size of the updates, this may introduce some temporary issues
|
||||
- redesigned server restart function
|
||||
now available and working in ui
|
||||
|
|
@ -1310,19 +1310,19 @@ Been a bit quieter for last few days as changes were quite significant, but fina
|
|||
Maintained by <https://github.com/Linaqruf/sd-notebook-collection>
|
||||
- Fixes, fixes, fixes...
|
||||
|
||||
To take advantage of new out-of-the-box tunings, its recommended to delete your `config.json` so new defaults are applied. Its not necessary, but otherwise you may need to play with UI Settings to get the best of Intel ARC, TorchML, ROCm or Apple M1/M2.
|
||||
To take advantage of new out-of-the-box tunings, its recommended to delete your `config.json` so new defaults are applied. its not necessary, but otherwise you may need to play with UI Settings to get the best of Intel ARC, TorchML, ROCm or Apple M1/M2.
|
||||
|
||||
## Update for 2023-04-27
|
||||
|
||||
a bit shorter list as:
|
||||
|
||||
- i've been busy with bugfixing
|
||||
- ive been busy with bugfixing
|
||||
there are a lot of them, not going to list each here.
|
||||
but seems like critical issues backlog is quieting down and soon i can focus on new features development.
|
||||
- i've started collaboration with couple of major projects,
|
||||
- ive started collaboration with couple of major projects,
|
||||
hopefully this will accelerate future development.
|
||||
|
||||
what's new:
|
||||
whats new:
|
||||
|
||||
- ability to view/add/edit model description shown in extra networks cards
|
||||
- add option to specify fallback sampler if primary sampler is not compatible with desired operation
|
||||
|
|
@ -1383,7 +1383,7 @@ what's new:
|
|||
- merge tomesd token merging
|
||||
- merge 23 PRs pending from a1111 backlog (!!)
|
||||
|
||||
*expect shorter updates for the next few days as i'll be partially ooo*
|
||||
*expect shorter updates for the next few days as ill be partially ooo*
|
||||
|
||||
## Update for 2023-04-20
|
||||
|
||||
|
|
@ -1457,7 +1457,7 @@ note: if you previously had command line optimizations such as --no-half, those
|
|||
## Update for 2023-04-15
|
||||
|
||||
- update default values
|
||||
- remove `ui-config.json` from repo, its not fully under user control
|
||||
- remove `ui-config.json` from repo, its now fully under user control
|
||||
- updated extensions manager
|
||||
- updated locon/lycoris plugin
|
||||
- enable quick launch by default
|
||||
|
|
@ -1490,7 +1490,7 @@ This has been pending for a while, but finally uploaded some massive changes
|
|||
Platform specific wrapper scripts that starts `launch.py` in Python virtual environment
|
||||
*Note*: Server can run without virtual environment, but it is recommended to use it
|
||||
This is carry-over from original repo
|
||||
**If you're unsure which launcher to use, this is the one you want**
|
||||
**If youre unsure which launcher to use, this is the one you want**
|
||||
- `launch.py`:
|
||||
Main startup script
|
||||
Can be used directly to start server in manually activated `venv` or to run it without `venv`
|
||||
|
|
|
|||
|
|
@ -81,7 +81,7 @@ Also supported are modifiers such as:
|
|||
|
||||
> [!IMPORTANT]
|
||||
> - Loading any model other than standard SD 1.x / SD 2.x requires use of backend **Diffusers**
|
||||
> - Loading any other models using **Original** backend is not supproted
|
||||
> - Loading any other models using **Original** backend is not supported
|
||||
> - Loading manually download model `.safetensors` files is supported for SD 1.x / SD 2.x / SD-XL models only
|
||||
> - For all other model types, use backend **Diffusers** and use built in Model downloader or
|
||||
select model from Networks -> Models -> Reference list in which case it will be auto-downloaded and loaded
|
||||
|
|
@ -224,7 +224,7 @@ check [ChangeLog](CHANGELOG.md) for when feature was first introduced as it will
|
|||
### **Sponsors**
|
||||
|
||||
<div align="center">
|
||||
<!-- sponsors --><a href="https://github.com/Tillerz"><img src="https://github.com/Tillerz.png" width="60px" alt="Tillerz" /></a><a href="https://github.com/allangrant"><img src="https://github.com/allangrant.png" width="60px" alt="Allan Grant" /></a><a href="https://github.com/BrentOzar"><img src="https://github.com/BrentOzar.png" width="60px" alt="Brent Ozar" /></a><a href="https://github.com/inktomi"><img src="https://github.com/inktomi.png" width="60px" alt="Matthew Runo" /></a><a href="https://github.com/HELLO-WORLD-SAS"><img src="https://github.com/HELLO-WORLD-SAS.png" width="60px" alt="HELLO WORLD SAS" /></a><a href="https://github.com/4joeknight4"><img src="https://github.com/4joeknight4.png" width="60px" alt="" /></a><a href="https://github.com/SaladTechnologies"><img src="https://github.com/SaladTechnologies.png" width="60px" alt="Salad Technologies" /></a><a href="https://github.com/mantzaris"><img src="https://github.com/mantzaris.png" width="60px" alt="a.v.mantzaris" /></a><!-- sponsors -->
|
||||
<!-- sponsors --><a href="https://github.com/allangrant"><img src="https://github.com/allangrant.png" width="60px" alt="Allan Grant" /></a><a href="https://github.com/BrentOzar"><img src="https://github.com/BrentOzar.png" width="60px" alt="Brent Ozar" /></a><a href="https://github.com/inktomi"><img src="https://github.com/inktomi.png" width="60px" alt="Matthew Runo" /></a><a href="https://github.com/HELLO-WORLD-SAS"><img src="https://github.com/HELLO-WORLD-SAS.png" width="60px" alt="HELLO WORLD SAS" /></a><a href="https://github.com/4joeknight4"><img src="https://github.com/4joeknight4.png" width="60px" alt="" /></a><a href="https://github.com/SaladTechnologies"><img src="https://github.com/SaladTechnologies.png" width="60px" alt="Salad Technologies" /></a><a href="https://github.com/mantzaris"><img src="https://github.com/mantzaris.png" width="60px" alt="a.v.mantzaris" /></a><!-- sponsors -->
|
||||
</div>
|
||||
|
||||
<br>
|
||||
|
|
|
|||
Loading…
Reference in New Issue