aboutsummaryrefslogtreecommitdiffstats
path: root/modules/textual_inversion
diff options
context:
space:
mode:
authorAUTOMATIC1111 <16777216c@gmail.com>2023-04-29 14:42:57 +0000
committerGitHub <noreply@github.com>2023-04-29 14:42:57 +0000
commit88c7debb02caa5e9b9a2bd519085f1f9134c4b01 (patch)
tree049d07f32194c1eced8ae4436af819fbfe4f5316 /modules/textual_inversion
parent67955ca9e5cb6b3cc539333d0a7d9591009bc800 (diff)
parent97167a576880e4802de081db71372ddc8c16fd92 (diff)
downloadstable-diffusion-webui-gfx803-88c7debb02caa5e9b9a2bd519085f1f9134c4b01.tar.gz
stable-diffusion-webui-gfx803-88c7debb02caa5e9b9a2bd519085f1f9134c4b01.tar.bz2
stable-diffusion-webui-gfx803-88c7debb02caa5e9b9a2bd519085f1f9134c4b01.zip
Merge branch 'dev' into ui-config-tabs
Diffstat (limited to 'modules/textual_inversion')
-rw-r--r--modules/textual_inversion/preprocess.py4
-rw-r--r--modules/textual_inversion/textual_inversion.py6
2 files changed, 9 insertions, 1 deletions
diff --git a/modules/textual_inversion/preprocess.py b/modules/textual_inversion/preprocess.py
index 2239cb84..de1ddb59 100644
--- a/modules/textual_inversion/preprocess.py
+++ b/modules/textual_inversion/preprocess.py
@@ -161,7 +161,9 @@ def preprocess_work(process_src, process_dst, process_width, process_height, pre
params.subindex = 0
filename = os.path.join(src, imagefile)
try:
- img = Image.open(filename).convert("RGB")
+ img = Image.open(filename)
+ img = ImageOps.exif_transpose(img)
+ img = img.convert("RGB")
except Exception:
continue
diff --git a/modules/textual_inversion/textual_inversion.py b/modules/textual_inversion/textual_inversion.py
index d2e62e58..379df243 100644
--- a/modules/textual_inversion/textual_inversion.py
+++ b/modules/textual_inversion/textual_inversion.py
@@ -233,6 +233,12 @@ class EmbeddingDatabase:
self.load_from_dir(embdir)
embdir.update()
+ # re-sort word_embeddings because load_from_dir may not load in alphabetic order.
+ # using a temporary copy so we don't reinitialize self.word_embeddings in case other objects have a reference to it.
+ sorted_word_embeddings = {e.name: e for e in sorted(self.word_embeddings.values(), key=lambda e: e.name.lower())}
+ self.word_embeddings.clear()
+ self.word_embeddings.update(sorted_word_embeddings)
+
displayed_embeddings = (tuple(self.word_embeddings.keys()), tuple(self.skipped_embeddings.keys()))
if self.previously_displayed_embeddings != displayed_embeddings:
self.previously_displayed_embeddings = displayed_embeddings