Commit message (Collapse) | Author | Age | Files | Lines | |
---|---|---|---|---|---|
* | Add numpy fix for MPS on PyTorch 1.12.1 | brkirch | 2022-12-17 | 1 | -0/+9 |
| | | | | | | | When saving training results with torch.save(), an exception is thrown: "RuntimeError: Can't call numpy() on Tensor that requires grad. Use tensor.detach().numpy() instead." So for MPS, check if Tensor.requires_grad and detach() if necessary. | ||||
* | add built-in extension system | AUTOMATIC | 2022-12-03 | 1 | -1/+10 |
| | | | | | add support for adding upscalers in extensions move LDSR, ScuNET and SwinIR to built-in extensions | ||||
* | add comment for #4407 and remove seemingly unnecessary cudnn.enabled | AUTOMATIC | 2022-12-03 | 1 | -1/+3 |
| | |||||
* | fix #4407 breaking UI entirely for card other than ones related to the PR | AUTOMATIC | 2022-12-03 | 1 | -4/+2 |
| | |||||
* | Merge pull request #4407 from yoinked-h/patch-1 | AUTOMATIC1111 | 2022-12-03 | 1 | -0/+7 |
|\ | | | | | Fix issue with 16xx cards | ||||
| * | actual better fix | pepe10-gpu | 2022-11-08 | 1 | -5/+2 |
| | | | | | | thanks C43H66N12O12S2 | ||||
| * | terrible hack | pepe10-gpu | 2022-11-08 | 1 | -2/+9 |
| | | |||||
| * | 16xx card fix | pepe10-gpu | 2022-11-07 | 1 | -0/+3 |
| | | | | | | cudnn | ||||
* | | Rework MPS randn fix, add randn_like fix | brkirch | 2022-11-30 | 1 | -12/+3 |
| | | | | | | | | torch.manual_seed() already sets a CPU generator, so there is no reason to create a CPU generator manually. torch.randn_like also needs a MPS fix for k-diffusion, but a torch hijack with randn_like already exists so it can also be used for that. | ||||
* | | Merge pull request #4918 from brkirch/pytorch-fixes | AUTOMATIC1111 | 2022-11-27 | 1 | -7/+24 |
|\ \ | | | | | | | Fixes for PyTorch 1.12.1 when using MPS | ||||
| * | | Add fixes for PyTorch 1.12.1 | brkirch | 2022-11-21 | 1 | -1/+27 |
| | | | | | | | | | | | | | | | | | | | | | | | | Fix typo "MasOS" -> "macOS" If MPS is available and PyTorch is an earlier version than 1.13: * Monkey patch torch.Tensor.to to ensure all tensors sent to MPS are contiguous * Monkey patch torch.nn.functional.layer_norm to ensure input tensor is contiguous (required for this program to work with MPS on unmodified PyTorch 1.12.1) | ||||
| * | | Revert "MPS Upscalers Fix" | brkirch | 2022-11-17 | 1 | -9/+0 |
| | | | | | | | | | | | | This reverts commit 768b95394a8500da639b947508f78296524f1836. | ||||
* | | | eliminate duplicated code from #5095 | AUTOMATIC | 2022-11-27 | 1 | -19/+11 |
| | | | |||||
* | | | torch.cuda.empty_cache() defaults to cuda:0 device unless explicitly set ↵ | Matthew McGoogan | 2022-11-26 | 1 | -2/+12 |
|/ / | | | | | | | otherwise first. Updating torch_gc() to use the device set by --device-id if specified to avoid OOM edge cases on multi-GPU systems. | ||||
* | | change formatting to match the main program in devices.py | AUTOMATIC | 2022-11-12 | 1 | -5/+16 |
| | | |||||
* | | Update devices.py | 源文雨 | 2022-11-12 | 1 | -1/+1 |
| | | |||||
* | | Fix wrong mps selection below MasOS 12.3 | 源文雨 | 2022-11-12 | 1 | -3/+10 |
|/ | |||||
* | MPS Upscalers Fix | brkirch | 2022-10-25 | 1 | -0/+4 |
| | | | | Get ESRGAN, SCUNet, and SwinIR working correctly on MPS by ensuring memory is contiguous for tensor views before sending to MPS device. | ||||
* | Remove BSRGAN from --use-cpu, add SwinIR | brkirch | 2022-10-25 | 1 | -1/+1 |
| | |||||
* | remove parsing command line from devices.py | AUTOMATIC | 2022-10-22 | 1 | -9/+5 |
| | |||||
* | implement CUDA device selection by ID | Extraltodeus | 2022-10-21 | 1 | -3/+18 |
| | |||||
* | Add 'interrogate' and 'all' choices to --use-cpu | brkirch | 2022-10-14 | 1 | -1/+1 |
| | | | | | * Add 'interrogate' and 'all' choices to --use-cpu * Change type for --use-cpu argument to str.lower, so that choices are case insensitive | ||||
* | --no-half-vae | AUTOMATIC | 2022-10-10 | 1 | -1/+5 |
| | |||||
* | Merge branch 'master' into cpu-cmdline-opt | brkirch | 2022-10-04 | 1 | -0/+10 |
|\ | |||||
| * | send all three of GFPGAN's and codeformer's models to CPU memory instead of ↵ | AUTOMATIC | 2022-10-04 | 1 | -0/+10 |
| | | | | | | | | just one for #1283 | ||||
* | | Add BSRGAN to --add-cpu | brkirch | 2022-10-04 | 1 | -1/+1 |
| | | |||||
* | | Add --use-cpu command line option | brkirch | 2022-10-04 | 1 | -3/+2 |
| | | | | | | | | Remove MPS detection to use CPU for GFPGAN / CodeFormer and add a --use-cpu command line option. | ||||
* | | Merge branch 'master' into master | brkirch | 2022-10-04 | 1 | -2/+1 |
|\| | |||||
| * | initial support for training textual inversion | AUTOMATIC | 2022-10-02 | 1 | -2/+1 |
| | | |||||
* | | When device is MPS, use CPU for GFPGAN instead | brkirch | 2022-10-01 | 1 | -1/+1 |
|/ | | | | GFPGAN will not work if the device is MPS, so default to CPU instead. | ||||
* | first attempt to produce crrect seeds in batch | AUTOMATIC | 2022-09-13 | 1 | -0/+10 |
| | |||||
* | changes for #294 | AUTOMATIC | 2022-09-12 | 1 | -0/+17 |
| | |||||
* | Allow TF32 in CUDA for increased performance #279 | AUTOMATIC | 2022-09-12 | 1 | -0/+11 |
| | |||||
* | add half() supporrt for CLIP interrogation | AUTOMATIC | 2022-09-11 | 1 | -0/+6 |
| | |||||
* | CLIP interrogator | AUTOMATIC | 2022-09-11 | 1 | -6/+10 |
| | |||||
* | Modular device management | Abdullah Barhoum | 2022-09-11 | 1 | -0/+12 |