Commit message (Collapse) | Author | Age | Files | Lines | |
---|---|---|---|---|---|
* | linting | Kohaku-Blueleaf | 2024-01-29 | 1 | -1/+0 |
| | |||||
* | Fix potential bugs | Kohaku-Blueleaf | 2024-01-29 | 1 | -2/+7 |
| | |||||
* | Avoid exceptions to be silenced | Kohaku-Blueleaf | 2024-01-20 | 1 | -6/+5 |
| | |||||
* | Avoid early disable | Kohaku-Blueleaf | 2024-01-20 | 1 | -0/+4 |
| | |||||
* | Fix nested manual cast | Kohaku-Blueleaf | 2024-01-18 | 1 | -1/+5 |
| | |||||
* | rearrange if-statements for cpu | Kohaku-Blueleaf | 2024-01-09 | 1 | -3/+3 |
| | |||||
* | Apply the correct behavior of precision='full' | Kohaku-Blueleaf | 2024-01-09 | 1 | -4/+7 |
| | |||||
* | Revert "Apply correct inference precision implementation" | Kohaku-Blueleaf | 2024-01-09 | 1 | -33/+9 |
| | | | | This reverts commit e00365962b17550a42235d1fbe2ad2c7cc4b8961. | ||||
* | Apply correct inference precision implementation | Kohaku-Blueleaf | 2024-01-09 | 1 | -9/+33 |
| | |||||
* | linting and debugs | Kohaku-Blueleaf | 2024-01-09 | 1 | -6/+6 |
| | |||||
* | Fix bugs when arg dtype doesn't match | KohakuBlueleaf | 2024-01-09 | 1 | -15/+10 |
| | |||||
* | improve efficiency and support more device | Kohaku-Blueleaf | 2024-01-09 | 1 | -17/+43 |
| | |||||
* | change import statements for #14478 | AUTOMATIC1111 | 2023-12-31 | 1 | -2/+2 |
| | |||||
* | Add utility to inspect a model's parameters (to get dtype/device) | Aarni Koskela | 2023-12-31 | 1 | -1/+2 |
| | |||||
* | Merge branch 'dev' into test-fp8 | Kohaku-Blueleaf | 2023-12-03 | 1 | -0/+13 |
|\ | |||||
| * | Merge pull request #14171 from Nuullll/ipex | AUTOMATIC1111 | 2023-12-02 | 1 | -0/+13 |
| |\ | | | | | | | Initial IPEX support for Intel Arc GPU | ||||
| | * | Disable ipex autocast due to its bad perf | Nuullll | 2023-12-02 | 1 | -7/+13 |
| | | | |||||
| | * | Initial IPEX support | Nuullll | 2023-11-30 | 1 | -2/+9 |
| | | | |||||
* | | | Merge branch 'dev' into test-fp8 | Kohaku-Blueleaf | 2023-12-02 | 1 | -1/+1 |
|\| | | |||||
| * | | Merge pull request #14131 from read-0nly/patch-1 | AUTOMATIC1111 | 2023-12-02 | 1 | -1/+1 |
| |\ \ | | |/ | |/| | Update devices.py - Make 'use-cpu all' actually apply to 'all' | ||||
| | * | Update devices.py | obsol | 2023-11-28 | 1 | -1/+1 |
| | | | | | | | | | | | | | | | fixes issue where "--use-cpu" all properly makes SD run on CPU but leaves ControlNet (and other extensions, I presume) pointed at GPU, causing a crash in ControlNet caused by a mismatch between devices between SD and CN https://github.com/AUTOMATIC1111/stable-diffusion-webui/issues/14097 | ||||
* | | | Better naming | Kohaku-Blueleaf | 2023-11-19 | 1 | -3/+3 |
| | | | |||||
* | | | Use options instead of cmd_args | Kohaku-Blueleaf | 2023-11-19 | 1 | -11/+14 |
| | | | |||||
* | | | Add MPS manual cast | KohakuBlueleaf | 2023-10-28 | 1 | -1/+5 |
| | | | |||||
* | | | ManualCast for 10/16 series gpu | Kohaku-Blueleaf | 2023-10-28 | 1 | -6/+51 |
| | | | |||||
* | | | Add CPU fp8 support | Kohaku-Blueleaf | 2023-10-23 | 1 | -1/+5 |
|/ / | | | | | | | | | | | Since norm layer need fp32, I only convert the linear operation layer(conv2d/linear) And TE have some pytorch function not support bf16 amp in CPU. I add a condition to indicate if the autocast is for unet. | ||||
* | | fix for crash when running #12924 without --device-id | AUTOMATIC1111 | 2023-09-09 | 1 | -1/+1 |
| | | |||||
* | | More accurate check for enabling cuDNN benchmark on 16XX cards | catboxanon | 2023-08-31 | 1 | -1/+2 |
|/ | |||||
* | split shared.py into multiple files; should resolve all circular reference ↵ | AUTOMATIC1111 | 2023-08-09 | 1 | -9/+1 |
| | | | | import errors related to shared.py | ||||
* | rework RNG to use generators instead of generating noises beforehand | AUTOMATIC1111 | 2023-08-09 | 1 | -79/+2 |
| | |||||
* | rework torchsde._brownian.brownian_interval replacement to use ↵ | AUTOMATIC1111 | 2023-08-03 | 1 | -6/+38 |
| | | | | device.randn_local and respect the NV setting. | ||||
* | add NV option for Random number generator source setting, which allows to ↵ | AUTOMATIC1111 | 2023-08-02 | 1 | -2/+37 |
| | | | | generate same pictures on CPU/AMD/Mac as on NVidia videocards. | ||||
* | Fix MPS cache cleanup | Aarni Koskela | 2023-07-11 | 1 | -2/+3 |
| | | | | Importing torch does not import torch.mps so the call failed. | ||||
* | added torch.mps.empty_cache() to torch_gc() | AUTOMATIC1111 | 2023-07-08 | 1 | -0/+3 |
| | | | | changed a bunch of places that use torch.cuda.empty_cache() to use torch_gc() instead | ||||
* | Remove a bunch of unused/vestigial code | Aarni Koskela | 2023-06-05 | 1 | -7/+0 |
| | | | | As found by Vulture and some eyes | ||||
* | run basic torch calculation at startup in parallel to reduce the performance ↵ | AUTOMATIC | 2023-05-21 | 1 | -0/+18 |
| | | | | impact of first generation | ||||
* | ruff auto fixes | AUTOMATIC | 2023-05-10 | 1 | -1/+1 |
| | |||||
* | rename CPU RNG to RNG source in settings, add infotext and parameters ↵ | AUTOMATIC | 2023-04-29 | 1 | -2/+2 |
| | | | | copypaste support to RNG source | ||||
* | Option to use CPU for random number generation. | Deciare | 2023-04-19 | 1 | -2/+6 |
| | | | | | | | Makes a given manual seed generate the same images across different platforms, independently of the GPU architecture in use. Fixes #9613. | ||||
* | Refactor Mac specific code to a separate file | brkirch | 2023-02-01 | 1 | -45/+7 |
| | | | | Move most Mac related code to a separate file, don't even load it unless web UI is run under macOS. | ||||
* | Refactor MPS fixes to CondFunc | brkirch | 2023-02-01 | 1 | -36/+14 |
| | |||||
* | MPS fix is still needed :( | brkirch | 2023-02-01 | 1 | -0/+3 |
| | | | | Apparently I did not test with large enough images to trigger the bug with torch.narrow on MPS | ||||
* | Merge pull request #7309 from brkirch/fix-embeddings | AUTOMATIC1111 | 2023-01-28 | 1 | -3/+8 |
|\ | | | | | Fix embeddings, upscalers, and refactor `--upcast-sampling` | ||||
| * | Remove MPS fix no longer needed for PyTorch | brkirch | 2023-01-28 | 1 | -3/+0 |
| | | | | | | | | The torch.narrow fix was required for nightly PyTorch builds for a while to prevent a hard crash, but newer nightly builds don't have this issue. | ||||
| * | Refactor conditional casting, fix upscalers | brkirch | 2023-01-28 | 1 | -0/+8 |
| | | |||||
* | | clarify the option to disable NaN check. | AUTOMATIC | 2023-01-27 | 1 | -0/+2 |
| | | |||||
* | | remove the need to place configs near models | AUTOMATIC | 2023-01-27 | 1 | -4/+8 |
|/ | |||||
* | Add UI setting for upcasting attention to float32 | brkirch | 2023-01-25 | 1 | -1/+5 |
| | | | | | | Adds "Upcast cross attention layer to float32" option in Stable Diffusion settings. This allows for generating images using SD 2.1 models without --no-half or xFormers. In order to make upcasting cross attention layer optimizations possible it is necessary to indent several sections of code in sd_hijack_optimizations.py so that a context manager can be used to disable autocast. Also, even though Stable Diffusion (and Diffusers) only upcast q and k, unfortunately my findings were that most of the cross attention layer optimizations could not function unless v is upcast also. | ||||
* | Add option for float32 sampling with float16 UNet | brkirch | 2023-01-25 | 1 | -0/+2 |
| | | | | This also handles type casting so that ROCm and MPS torch devices work correctly without --no-half. One cast is required for deepbooru in deepbooru_model.py, some explicit casting is required for img2img and inpainting. depth_model can't be converted to float16 or it won't work correctly on some systems (it's known to have issues on MPS) so in sd_models.py model.depth_model is removed for model.half(). | ||||
* | Merge pull request #6922 from brkirch/cumsum-fix | AUTOMATIC1111 | 2023-01-19 | 1 | -4/+7 |
|\ | | | | | Improve cumsum fix for MPS |