0
4

Delete article

Deleted articles cannot be recovered.

Draft of this article would be also deleted.

Are you sure you want to delete this article?

More than 1 year has passed since last update.

M1 MacBookで画像生成AI「Stable Diffusion」を動かしてみた。

Posted at

はじめに

AI。流行ってますねー。
イラスト生成に、chatGPT、なんか本当にシンギュラリティとか来ちゃいそうですが。
で、画像生成AIの「Stable Diffusion」ですがWEBサイト経由やアプリなんかで利用できますが、
動かせる環境とかマシンがあればチャレジしたいじゃないですかー。

というわけで、やってみた。

まずベースですが、この記事に記載されている内容を補完するものです。

忙しい人向けに要点をまとめ

Stable Diffusionを動かすのに必要な要件は
Python 3.10以上です。

(1)Pythonをインストールしたら、Stable Diffusionのフォークをクローンします。

git clone -b apple-silicon-mps-support https://github.com/bfirsh/stable-diffusion.git
cd stable-diffusion
mkdir -p models/ldm/stable-diffusion-v1/

(2)依存関係を解消するためにVirtualenvをインストールして実行

pip install virtualenv
python3 -m virtualenv venv
source venv/bin/activate

(3)依存関係のインストール

pip install -r requirements.txt

(4)sd-v1-4.ckptのダウンロードと配置
下記リンクより「sd-v1-4.ckpt」を取得して(1)で作成したディレクトリに「model.ckpt」というファイル名で配置します。

以下のスクリーンショットのようになっていればOKです。
スクリーンショット 2023-02-20 13.59.46.png
(5)その他必要なパッケージのインストール
gigazineの解説記事では「model.ckpt」の配置の後で、すぐにtxt2img.py を実行するように記載されていますが
当然、このままでは動きません。
その為、必要なパッケージ群をインストールしていきます。
なおログでえらいことになりますので、必要なコマンドの羅列のみとします。

pip install diffusers
pip install opencv-python 
pip install torch torchvision torchaudio 
pip install omegaconf
pip install imwatermark
pip install invisible-watermark 
pip install einops
pip install pytorch_lightning
pip install transformers 
pip install taming-transformers-rom1504
pip install clip
pip install kornia

(6)実行してみる。
gigazineの記事にあったサンプルのおまじないを唱えてみます。
3行に見えてますが、バックスラッシュで折り返しているので実際のコマンドはワンライナーです。

stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms

「宇宙空間に惑星のように浮いている美味しそうな赤いリンゴ」と指定していますね。

実行中の様子です。
(少し時間がかかります)
スクリーンショット 2023-02-20 14.17.25.png
最終的にこのようなメッセージが出力されたら成功です。
スクリーンショット 2023-02-20 14.18.45.png
出力されたファイルがこちら「grid-0000.png」
スクリーンショット 2023-02-20 14.20.59.png
出力画像
スクリーンショット 2023-02-20 14.23.00.png
アップルマークが自己主張気味?

他の指定も試してみましょう。

stable-diffusion % python scripts/txt2img.py \
--prompt "grandma overtakes supercar" \                                                            
--n_samples 1 --n_iter 1 --plms

スーパーカーを追い抜くおばあちゃん

はたして?

出力画像
grid-0002.png
スーパーカーは出たけど、追い抜くおばあちゃんまでは無理か?

というわけで無事、Stable Diffusionを動かす事に成功しました。
良かったらお試し下さい。

最後にトライアンドエラーしたログを掲載して終わりたいと思います。
それでは、また〜。

おまけ(トライエラーの様子です)

~ % pip install diffusers
Collecting diffusers
  Using cached diffusers-0.13.0-py3-none-any.whl (716 kB)
Collecting regex!=2019.12.17
  Using cached regex-2022.10.31-cp310-cp310-macosx_10_9_x86_64.whl (293 kB)
Requirement already satisfied: filelock in ./.pyenv/versions/3.10.7/lib/python3.10/site-packages (from diffusers) (3.9.0)
Collecting Pillow
  Using cached Pillow-9.4.0-2-cp310-cp310-macosx_10_10_x86_64.whl (3.3 MB)
Collecting importlib-metadata
  Using cached importlib_metadata-6.0.0-py3-none-any.whl (21 kB)
Collecting huggingface-hub>=0.10.0
  Using cached huggingface_hub-0.12.1-py3-none-any.whl (190 kB)
Collecting requests
  Using cached requests-2.28.2-py3-none-any.whl (62 kB)
Requirement already satisfied: numpy in ./.pyenv/versions/3.10.7/lib/python3.10/site-packages (from diffusers) (1.23.5)
Collecting packaging>=20.9
  Using cached packaging-23.0-py3-none-any.whl (42 kB)
Collecting typing-extensions>=3.7.4.3
  Using cached typing_extensions-4.5.0-py3-none-any.whl (27 kB)
Collecting tqdm>=4.42.1
  Using cached tqdm-4.64.1-py2.py3-none-any.whl (78 kB)
Collecting pyyaml>=5.1
  Using cached PyYAML-6.0-cp310-cp310-macosx_10_9_x86_64.whl (197 kB)
Collecting zipp>=0.5
  Using cached zipp-3.14.0-py3-none-any.whl (6.7 kB)
Collecting urllib3<1.27,>=1.21.1
  Using cached urllib3-1.26.14-py2.py3-none-any.whl (140 kB)
Collecting idna<4,>=2.5
  Using cached idna-3.4-py3-none-any.whl (61 kB)
Collecting certifi>=2017.4.17
  Downloading certifi-2022.12.7-py3-none-any.whl (155 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 155.3/155.3 kB 5.7 MB/s eta 0:00:00
Collecting charset-normalizer<4,>=2
  Using cached charset_normalizer-3.0.1-cp310-cp310-macosx_10_9_x86_64.whl (124 kB)
Installing collected packages: charset-normalizer, zipp, urllib3, typing-extensions, tqdm, regex, pyyaml, Pillow, packaging, idna, certifi, requests, importlib-metadata, huggingface-hub, diffusers
Successfully installed Pillow-9.4.0 certifi-2022.12.7 charset-normalizer-3.0.1 diffusers-0.13.0 huggingface-hub-0.12.1 idna-3.4 importlib-metadata-6.0.0 packaging-23.0 pyyaml-6.0 regex-2022.10.31 requests-2.28.2 tqdm-4.64.1 typing-extensions-4.5.0 urllib3-1.26.14 zipp-3.14.0




stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 2, in <module>
    import cv2
ModuleNotFoundError: No module named 'cv2'



stable-diffusion % pip install opencv-python 
Collecting opencv-python
  Downloading opencv_python-4.7.0.68-cp37-abi3-macosx_10_13_x86_64.whl (51.7 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 51.7/51.7 MB 26.3 MB/s eta 0:00:00
Requirement already satisfied: numpy>=1.21.2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from opencv-python) (1.23.5)
Installing collected packages: opencv-python
Successfully installed opencv-python-4.7.0.68
stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 3, in <module>
    import torch
ModuleNotFoundError: No module named 'torch'



stable-diffusion % pip install torch torchvision torchaudio 
Collecting torch
  Downloading torch-1.13.1-cp310-none-macosx_10_9_x86_64.whl (135.3 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 135.3/135.3 MB 18.1 MB/s eta 0:00:00
Collecting torchvision
  Downloading torchvision-0.14.1-cp310-cp310-macosx_10_9_x86_64.whl (1.4 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.4/1.4 MB 31.2 MB/s eta 0:00:00
Collecting torchaudio
  Downloading torchaudio-0.13.1-cp310-cp310-macosx_10_9_x86_64.whl (3.3 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.3/3.3 MB 40.7 MB/s eta 0:00:00
Requirement already satisfied: typing-extensions in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torch) (4.5.0)
Requirement already satisfied: requests in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision) (2.28.2)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision) (9.4.0)
Requirement already satisfied: numpy in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision) (1.23.5)
Requirement already satisfied: certifi>=2017.4.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision) (2022.12.7)
Requirement already satisfied: urllib3<1.27,>=1.21.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision) (1.26.14)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision) (3.0.1)
Requirement already satisfied: idna<4,>=2.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision) (3.4)
Installing collected packages: torch, torchvision, torchaudio
Successfully installed torch-1.13.1 torchaudio-0.13.1 torchvision-0.14.1


stable-diffusion % python scripts/txt2img.py \             
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 5, in <module>
    from omegaconf import OmegaConf
ModuleNotFoundError: No module named 'omegaconf'


stable-diffusion % pip install omegaconf
Collecting omegaconf
  Downloading omegaconf-2.3.0-py3-none-any.whl (79 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 79.5/79.5 kB 4.0 MB/s eta 0:00:00
Requirement already satisfied: PyYAML>=5.1.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from omegaconf) (6.0)
Collecting antlr4-python3-runtime==4.9.*
  Downloading antlr4-python3-runtime-4.9.3.tar.gz (117 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 117.0/117.0 kB 14.0 MB/s eta 0:00:00
  Preparing metadata (setup.py) ... done
Installing collected packages: antlr4-python3-runtime, omegaconf
  DEPRECATION: antlr4-python3-runtime is being installed using the legacy 'setup.py install' method, because it does not have a 'pyproject.toml' and the 'wheel' package is not installed. pip 23.1 will enforce this behaviour change. A possible replacement is to enable the '--use-pep517' option. Discussion can be found at https://github.com/pypa/pip/issues/8559
  Running setup.py install for antlr4-python3-runtime ... done
Successfully installed antlr4-python3-runtime-4.9.3 omegaconf-2.3.0


stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 8, in <module>
    from imwatermark import WatermarkEncoder
ModuleNotFoundError: No module named 'imwatermark'


stable-diffusion % pip install imwatermark    
Collecting imwatermark
  Downloading imWatermark-0.0.2-py3-none-any.whl (2.8 kB)
Requirement already satisfied: numpy in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from imwatermark) (1.23.5)
Collecting imhist
  Downloading imhist-0.0.4-py3-none-any.whl (2.8 kB)
Installing collected packages: imhist, imwatermark
Successfully installed imhist-0.0.4 imwatermark-0.0.2


stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 8, in <module>
    from imwatermark import WatermarkEncoder
ModuleNotFoundError: No module named 'imwatermark'


stable-diffusion % pip install imwatermark           
Requirement already satisfied: imwatermark in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (0.0.2)
Requirement already satisfied: numpy in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from imwatermark) (1.23.5)
Requirement already satisfied: imhist in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from imwatermark) (0.0.4)


stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 8, in <module>
    from imwatermark import WatermarkEncoder
ModuleNotFoundError: No module named 'imwatermark'


stable-diffusion % pip install invisible-watermark                    
Collecting invisible-watermark
  Using cached invisible_watermark-0.1.5-py3-none-any.whl (1.6 MB)
Requirement already satisfied: Pillow>=6.0.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from invisible-watermark) (9.4.0)
Collecting PyWavelets>=1.1.1
  Using cached PyWavelets-1.4.1-cp310-cp310-macosx_10_13_x86_64.whl (4.4 MB)
Collecting onnxruntime
  Using cached onnxruntime-1.14.0-cp310-cp310-macosx_11_0_x86_64.whl (6.6 MB)
Requirement already satisfied: opencv-python>=4.1.0.25 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from invisible-watermark) (4.7.0.68)
Collecting onnx
  Using cached onnx-1.13.0-cp310-cp310-macosx_10_12_x86_64.whl (12.7 MB)
Requirement already satisfied: numpy>=1.17.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from invisible-watermark) (1.23.5)
Requirement already satisfied: torch in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from invisible-watermark) (1.13.1)
Requirement already satisfied: typing-extensions>=3.6.2.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from onnx->invisible-watermark) (4.5.0)
Collecting protobuf<4,>=3.20.2
  Using cached protobuf-3.20.3-py2.py3-none-any.whl (162 kB)
Requirement already satisfied: packaging in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from onnxruntime->invisible-watermark) (23.0)
Collecting coloredlogs
  Using cached coloredlogs-15.0.1-py2.py3-none-any.whl (46 kB)
Collecting flatbuffers
  Using cached flatbuffers-23.1.21-py2.py3-none-any.whl (26 kB)
Collecting sympy
  Downloading sympy-1.11.1-py3-none-any.whl (6.5 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.5/6.5 MB 36.2 MB/s eta 0:00:00
Collecting humanfriendly>=9.1
  Using cached humanfriendly-10.0-py2.py3-none-any.whl (86 kB)
Collecting mpmath>=0.19
  Downloading mpmath-1.2.1-py3-none-any.whl (532 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 532.6/532.6 kB 37.8 MB/s eta 0:00:00
Installing collected packages: mpmath, flatbuffers, sympy, PyWavelets, protobuf, humanfriendly, onnx, coloredlogs, onnxruntime, invisible-watermark
Successfully installed PyWavelets-1.4.1 coloredlogs-15.0.1 flatbuffers-23.1.21 humanfriendly-10.0 invisible-watermark-0.1.5 mpmath-1.2.1 onnx-1.13.0 onnxruntime-1.14.0 protobuf-3.20.3 sympy-1.11.1


stable-diffusion % python scripts/txt2img.py \                        
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 10, in <module>
    from einops import rearrange
ModuleNotFoundError: No module named 'einops'


stable-diffusion % pip install einops             
Collecting einops
  Downloading einops-0.6.0-py3-none-any.whl (41 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 41.6/41.6 kB 2.9 MB/s eta 0:00:00
Installing collected packages: einops
Successfully installed einops-0.6.0


stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 13, in <module>
    from pytorch_lightning import seed_everything
ModuleNotFoundError: No module named 'pytorch_lightning'


stable-diffusion % pip install pytorch_lightning
Collecting pytorch_lightning
  Downloading pytorch_lightning-1.9.2-py3-none-any.whl (826 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 826.2/826.2 kB 15.3 MB/s eta 0:00:00
Requirement already satisfied: tqdm>=4.57.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch_lightning) (4.64.1)
Collecting torchmetrics>=0.7.0
  Downloading torchmetrics-0.11.1-py3-none-any.whl (517 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 517.2/517.2 kB 23.6 MB/s eta 0:00:00
Requirement already satisfied: torch>=1.10.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch_lightning) (1.13.1)
Collecting fsspec[http]>2021.06.0
  Using cached fsspec-2023.1.0-py3-none-any.whl (143 kB)
Requirement already satisfied: numpy>=1.17.2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch_lightning) (1.23.5)
Collecting lightning-utilities>=0.6.0.post0
  Downloading lightning_utilities-0.6.0.post0-py3-none-any.whl (18 kB)
Requirement already satisfied: typing-extensions>=4.0.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch_lightning) (4.5.0)
Requirement already satisfied: PyYAML>=5.4 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch_lightning) (6.0)
Requirement already satisfied: packaging>=17.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch_lightning) (23.0)
Collecting aiohttp!=4.0.0a0,!=4.0.0a1
  Using cached aiohttp-3.8.4-cp310-cp310-macosx_10_9_x86_64.whl (358 kB)
Requirement already satisfied: requests in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from fsspec[http]>2021.06.0->pytorch_lightning) (2.28.2)
Collecting aiosignal>=1.1.2
  Using cached aiosignal-1.3.1-py3-none-any.whl (7.6 kB)
Collecting async-timeout<5.0,>=4.0.0a3
  Using cached async_timeout-4.0.2-py3-none-any.whl (5.8 kB)
Requirement already satisfied: charset-normalizer<4.0,>=2.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch_lightning) (3.0.1)
Collecting frozenlist>=1.1.1
  Using cached frozenlist-1.3.3-cp310-cp310-macosx_10_9_x86_64.whl (35 kB)
Collecting attrs>=17.3.0
  Using cached attrs-22.2.0-py3-none-any.whl (60 kB)
Collecting multidict<7.0,>=4.5
  Using cached multidict-6.0.4-cp310-cp310-macosx_10_9_x86_64.whl (29 kB)
Collecting yarl<2.0,>=1.0
  Using cached yarl-1.8.2-cp310-cp310-macosx_10_9_x86_64.whl (61 kB)
Requirement already satisfied: urllib3<1.27,>=1.21.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->fsspec[http]>2021.06.0->pytorch_lightning) (1.26.14)
Requirement already satisfied: idna<4,>=2.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->fsspec[http]>2021.06.0->pytorch_lightning) (3.4)
Requirement already satisfied: certifi>=2017.4.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->fsspec[http]>2021.06.0->pytorch_lightning) (2022.12.7)
Installing collected packages: multidict, lightning-utilities, fsspec, frozenlist, attrs, async-timeout, yarl, torchmetrics, aiosignal, aiohttp, pytorch_lightning
Successfully installed aiohttp-3.8.4 aiosignal-1.3.1 async-timeout-4.0.2 attrs-22.2.0 frozenlist-1.3.3 fsspec-2023.1.0 lightning-utilities-0.6.0.post0 multidict-6.0.4 pytorch_lightning-1.9.2 torchmetrics-0.11.1 yarl-1.8.2



stable-diffusion % python scripts/txt2img.py \  
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 32, in <module>
    from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker
  File "/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/diffusers/pipelines/stable_diffusion/safety_checker.py", line 18, in <module>
    from transformers import CLIPConfig, CLIPVisionModel, PreTrainedModel
ModuleNotFoundError: No module named 'transformers'



stable-diffusion % pip install transformers     
Collecting transformers
  Downloading transformers-4.26.1-py3-none-any.whl (6.3 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.3/6.3 MB 32.0 MB/s eta 0:00:00
Requirement already satisfied: huggingface-hub<1.0,>=0.11.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (0.12.1)
Requirement already satisfied: tqdm>=4.27 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (4.64.1)
Requirement already satisfied: filelock in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (3.9.0)
Requirement already satisfied: numpy>=1.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (1.23.5)
Requirement already satisfied: requests in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (2.28.2)
Requirement already satisfied: packaging>=20.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (23.0)
Collecting tokenizers!=0.11.3,<0.14,>=0.11.1
  Downloading tokenizers-0.13.2-cp310-cp310-macosx_10_11_x86_64.whl (3.8 MB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.8/3.8 MB 42.7 MB/s eta 0:00:00
Requirement already satisfied: pyyaml>=5.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (6.0)
Requirement already satisfied: regex!=2019.12.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from transformers) (2022.10.31)
Requirement already satisfied: typing-extensions>=3.7.4.3 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from huggingface-hub<1.0,>=0.11.0->transformers) (4.5.0)
Requirement already satisfied: certifi>=2017.4.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->transformers) (2022.12.7)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->transformers) (3.0.1)
Requirement already satisfied: idna<4,>=2.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->transformers) (3.4)
Requirement already satisfied: urllib3<1.27,>=1.21.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->transformers) (1.26.14)
Installing collected packages: tokenizers, transformers
Successfully installed tokenizers-0.13.2 transformers-4.26.1


stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
Downloading ()rocessor_config.json: 100%|███████████████████████████████████████████████████████████| 342/342 [00:00<00:00, 89.7kB/s]
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead.
  warnings.warn(
Downloading ()lve/main/config.json: 100%|███████████████████████████████████████████████████████| 4.55k/4.55k [00:00<00:00, 1.41MB/s]
Downloading ()"pytorch_model.bin";: 100%|███████████████████████████████████████████████████████| 1.22G/1.22G [00:19<00:00, 62.3MB/s]
Global seed set to 42
Loading model from models/ldm/stable-diffusion-v1/model.ckpt
Global Step: 470000
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 359, in <module>
    main()
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 251, in main
    model = load_model_from_config(config, f"{opt.ckpt}")
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 65, in load_model_from_config
    model = instantiate_from_config(config.model)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 85, in instantiate_from_config
    return get_obj_from_str(config["target"])(**config.get("params", dict()))
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 93, in get_obj_from_str
    return getattr(importlib.import_module(module, package=None), cls)
  File "/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/importlib/__init__.py", line 126, in import_module
    return _bootstrap._gcd_import(name[level:], package, level)
  File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
  File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
  File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
  File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
  File "<frozen importlib._bootstrap_external>", line 883, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/diffusion/ddpm.py", line 24, in <module>
    from ldm.models.autoencoder import VQModelInterface, IdentityFirstStage, AutoencoderKL
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/autoencoder.py", line 6, in <module>
    from taming.modules.vqvae.quantize import VectorQuantizer2 as VectorQuantizer
ModuleNotFoundError: No module named 'taming'


stable-diffusion % pip install taming-transformers
Collecting taming-transformers
  Downloading taming_transformers-0.0.1-py3-none-any.whl (45 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 45.6/45.6 kB 3.1 MB/s eta 0:00:00
Requirement already satisfied: torchvision in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers) (0.14.1)
Requirement already satisfied: torch in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers) (1.13.1)
Requirement already satisfied: pytorch-lightning>=1.0.8 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers) (1.9.2)
Requirement already satisfied: omegaconf>=2.0.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers) (2.3.0)
Requirement already satisfied: tqdm in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers) (4.64.1)
Requirement already satisfied: numpy in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers) (1.23.5)
Requirement already satisfied: PyYAML>=5.1.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from omegaconf>=2.0.0->taming-transformers) (6.0)
Requirement already satisfied: antlr4-python3-runtime==4.9.* in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from omegaconf>=2.0.0->taming-transformers) (4.9.3)
Requirement already satisfied: torchmetrics>=0.7.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers) (0.11.1)
Requirement already satisfied: typing-extensions>=4.0.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers) (4.5.0)
Requirement already satisfied: lightning-utilities>=0.6.0.post0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers) (0.6.0.post0)
Requirement already satisfied: packaging>=17.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers) (23.0)
Requirement already satisfied: fsspec[http]>2021.06.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers) (2023.1.0)
Requirement already satisfied: requests in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision->taming-transformers) (2.28.2)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision->taming-transformers) (9.4.0)
Requirement already satisfied: aiohttp!=4.0.0a0,!=4.0.0a1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (3.8.4)
Requirement already satisfied: urllib3<1.27,>=1.21.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers) (1.26.14)
Requirement already satisfied: idna<4,>=2.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers) (3.4)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers) (3.0.1)
Requirement already satisfied: certifi>=2017.4.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers) (2022.12.7)
Requirement already satisfied: async-timeout<5.0,>=4.0.0a3 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (4.0.2)
Requirement already satisfied: multidict<7.0,>=4.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (6.0.4)
Requirement already satisfied: yarl<2.0,>=1.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (1.8.2)
Requirement already satisfied: frozenlist>=1.1.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (1.3.3)
Requirement already satisfied: attrs>=17.3.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (22.2.0)
Requirement already satisfied: aiosignal>=1.1.2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers) (1.3.1)
Installing collected packages: taming-transformers
Successfully installed taming-transformers-0.0.1
stable-diffusion % python scripts/txt2img.py \    
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead.
  warnings.warn(
Global seed set to 42
Loading model from models/ldm/stable-diffusion-v1/model.ckpt
Global Step: 470000
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 359, in <module>
    main()
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 251, in main
    model = load_model_from_config(config, f"{opt.ckpt}")
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 65, in load_model_from_config
    model = instantiate_from_config(config.model)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 85, in instantiate_from_config
    return get_obj_from_str(config["target"])(**config.get("params", dict()))
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 93, in get_obj_from_str
    return getattr(importlib.import_module(module, package=None), cls)
  File "/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/importlib/__init__.py", line 126, in import_module
    return _bootstrap._gcd_import(name[level:], package, level)
  File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
  File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
  File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
  File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
  File "<frozen importlib._bootstrap_external>", line 883, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/diffusion/ddpm.py", line 24, in <module>
    from ldm.models.autoencoder import VQModelInterface, IdentityFirstStage, AutoencoderKL
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/autoencoder.py", line 6, in <module>
    from taming.modules.vqvae.quantize import VectorQuantizer2 as VectorQuantizer
ImportError: cannot import name 'VectorQuantizer2' from 'taming.modules.vqvae.quantize' (/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/taming/modules/vqvae/quantize.py)



stable-diffusion % pip install taming-transformers-rom1504 
Collecting taming-transformers-rom1504
  Downloading taming_transformers_rom1504-0.0.6-py3-none-any.whl (51 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 51.5/51.5 kB 3.5 MB/s eta 0:00:00
Requirement already satisfied: tqdm in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers-rom1504) (4.64.1)
Requirement already satisfied: numpy in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers-rom1504) (1.23.5)
Requirement already satisfied: torchvision in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers-rom1504) (0.14.1)
Requirement already satisfied: pytorch-lightning>=1.0.8 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers-rom1504) (1.9.2)
Requirement already satisfied: torch in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers-rom1504) (1.13.1)
Requirement already satisfied: omegaconf>=2.0.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from taming-transformers-rom1504) (2.3.0)
Requirement already satisfied: PyYAML>=5.1.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from omegaconf>=2.0.0->taming-transformers-rom1504) (6.0)
Requirement already satisfied: antlr4-python3-runtime==4.9.* in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from omegaconf>=2.0.0->taming-transformers-rom1504) (4.9.3)
Requirement already satisfied: lightning-utilities>=0.6.0.post0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers-rom1504) (0.6.0.post0)
Requirement already satisfied: torchmetrics>=0.7.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers-rom1504) (0.11.1)
Requirement already satisfied: typing-extensions>=4.0.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers-rom1504) (4.5.0)
Requirement already satisfied: fsspec[http]>2021.06.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers-rom1504) (2023.1.0)
Requirement already satisfied: packaging>=17.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from pytorch-lightning>=1.0.8->taming-transformers-rom1504) (23.0)
Requirement already satisfied: pillow!=8.3.*,>=5.3.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision->taming-transformers-rom1504) (9.4.0)
Requirement already satisfied: requests in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torchvision->taming-transformers-rom1504) (2.28.2)
Requirement already satisfied: aiohttp!=4.0.0a0,!=4.0.0a1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (3.8.4)
Requirement already satisfied: urllib3<1.27,>=1.21.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers-rom1504) (1.26.14)
Requirement already satisfied: idna<4,>=2.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers-rom1504) (3.4)
Requirement already satisfied: charset-normalizer<4,>=2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers-rom1504) (3.0.1)
Requirement already satisfied: certifi>=2017.4.17 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from requests->torchvision->taming-transformers-rom1504) (2022.12.7)
Requirement already satisfied: aiosignal>=1.1.2 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (1.3.1)
Requirement already satisfied: frozenlist>=1.1.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (1.3.3)
Requirement already satisfied: yarl<2.0,>=1.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (1.8.2)
Requirement already satisfied: attrs>=17.3.0 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (22.2.0)
Requirement already satisfied: async-timeout<5.0,>=4.0.0a3 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (4.0.2)
Requirement already satisfied: multidict<7.0,>=4.5 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from aiohttp!=4.0.0a0,!=4.0.0a1->fsspec[http]>2021.06.0->pytorch-lightning>=1.0.8->taming-transformers-rom1504) (6.0.4)
Installing collected packages: taming-transformers-rom1504
Successfully installed taming-transformers-rom1504-0.0.6



stable-diffusion % python scripts/txt2img.py \             
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead.
  warnings.warn(
Global seed set to 42
Loading model from models/ldm/stable-diffusion-v1/model.ckpt
Global Step: 470000
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/pytorch_lightning/utilities/distributed.py:258: LightningDeprecationWarning: `pytorch_lightning.utilities.distributed.rank_zero_only` has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from `pytorch_lightning.utilities` instead.
  rank_zero_deprecation(
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
making attention of type 'vanilla' with 512 in_channels
Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
making attention of type 'vanilla' with 512 in_channels
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 359, in <module>
    main()
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 251, in main
    model = load_model_from_config(config, f"{opt.ckpt}")
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 65, in load_model_from_config
    model = instantiate_from_config(config.model)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 85, in instantiate_from_config
    return get_obj_from_str(config["target"])(**config.get("params", dict()))
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/diffusion/ddpm.py", line 461, in __init__
    self.instantiate_cond_stage(cond_stage_config)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/diffusion/ddpm.py", line 519, in instantiate_cond_stage
    model = instantiate_from_config(config)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 85, in instantiate_from_config
    return get_obj_from_str(config["target"])(**config.get("params", dict()))
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 93, in get_obj_from_str
    return getattr(importlib.import_module(module, package=None), cls)
  File "/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/importlib/__init__.py", line 126, in import_module
    return _bootstrap._gcd_import(name[level:], package, level)
  File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
  File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
  File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
  File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
  File "<frozen importlib._bootstrap_external>", line 883, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/modules/encoders/modules.py", line 4, in <module>
    import clip
ModuleNotFoundError: No module named 'clip'



stable-diffusion % pip install clip                        
Collecting clip
  Downloading clip-0.2.0.tar.gz (5.5 kB)
  Preparing metadata (setup.py) ... done
Installing collected packages: clip
  DEPRECATION: clip is being installed using the legacy 'setup.py install' method, because it does not have a 'pyproject.toml' and the 'wheel' package is not installed. pip 23.1 will enforce this behaviour change. A possible replacement is to enable the '--use-pep517' option. Discussion can be found at https://github.com/pypa/pip/issues/8559
  Running setup.py install for clip ... done
Successfully installed clip-0.2.0



stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead.
  warnings.warn(
Global seed set to 42
Loading model from models/ldm/stable-diffusion-v1/model.ckpt
Global Step: 470000
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/pytorch_lightning/utilities/distributed.py:258: LightningDeprecationWarning: `pytorch_lightning.utilities.distributed.rank_zero_only` has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from `pytorch_lightning.utilities` instead.
  rank_zero_deprecation(
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
making attention of type 'vanilla' with 512 in_channels
Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
making attention of type 'vanilla' with 512 in_channels
Traceback (most recent call last):
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 359, in <module>
    main()
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 251, in main
    model = load_model_from_config(config, f"{opt.ckpt}")
  File "/Users/gatuwo/SD/stable-diffusion/scripts/txt2img.py", line 65, in load_model_from_config
    model = instantiate_from_config(config.model)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 85, in instantiate_from_config
    return get_obj_from_str(config["target"])(**config.get("params", dict()))
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/diffusion/ddpm.py", line 461, in __init__
    self.instantiate_cond_stage(cond_stage_config)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/models/diffusion/ddpm.py", line 519, in instantiate_cond_stage
    model = instantiate_from_config(config)
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 85, in instantiate_from_config
    return get_obj_from_str(config["target"])(**config.get("params", dict()))
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/util.py", line 93, in get_obj_from_str
    return getattr(importlib.import_module(module, package=None), cls)
  File "/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/importlib/__init__.py", line 126, in import_module
    return _bootstrap._gcd_import(name[level:], package, level)
  File "<frozen importlib._bootstrap>", line 1050, in _gcd_import
  File "<frozen importlib._bootstrap>", line 1027, in _find_and_load
  File "<frozen importlib._bootstrap>", line 1006, in _find_and_load_unlocked
  File "<frozen importlib._bootstrap>", line 688, in _load_unlocked
  File "<frozen importlib._bootstrap_external>", line 883, in exec_module
  File "<frozen importlib._bootstrap>", line 241, in _call_with_frames_removed
  File "/Users/gatuwo/SD/stable-diffusion/scripts/../ldm/modules/encoders/modules.py", line 7, in <module>
    import kornia
ModuleNotFoundError: No module named 'kornia'



stable-diffusion % pip install kornia         
Collecting kornia
  Downloading kornia-0.6.10-py2.py3-none-any.whl (612 kB)
     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 612.0/612.0 kB 10.2 MB/s eta 0:00:00
Requirement already satisfied: torch>=1.9.1 in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from kornia) (1.13.1)
Requirement already satisfied: packaging in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from kornia) (23.0)
Requirement already satisfied: typing-extensions in /Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages (from torch>=1.9.1->kornia) (4.5.0)
Installing collected packages: kornia
Successfully installed kornia-0.6.10



stable-diffusion % python scripts/txt2img.py \
--prompt "a red juicy apple floating in outer space, like a planet" \
--n_samples 1 --n_iter 1 --plms
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/transformers/models/clip/feature_extraction_clip.py:28: FutureWarning: The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please use CLIPImageProcessor instead.
  warnings.warn(
Global seed set to 42
Loading model from models/ldm/stable-diffusion-v1/model.ckpt
Global Step: 470000
/Users/gatuwo/.pyenv/versions/3.10.7/lib/python3.10/site-packages/pytorch_lightning/utilities/distributed.py:258: LightningDeprecationWarning: `pytorch_lightning.utilities.distributed.rank_zero_only` has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from `pytorch_lightning.utilities` instead.
  rank_zero_deprecation(
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
making attention of type 'vanilla' with 512 in_channels
Working with z of shape (1, 4, 32, 32) = 4096 dimensions.
making attention of type 'vanilla' with 512 in_channels
Downloading ()olve/main/vocab.json: 100%|█████████████████████████████████████████████████████████| 961k/961k [00:00<00:00, 1.02MB/s]
Downloading ()olve/main/merges.txt: 100%|██████████████████████████████████████████████████████████| 525k/525k [00:00<00:00, 593kB/s]
Downloading ()cial_tokens_map.json: 100%|████████████████████████████████████████████████████████████| 389/389 [00:00<00:00, 128kB/s]
Downloading ()okenizer_config.json: 100%|████████████████████████████████████████████████████████████| 905/905 [00:00<00:00, 221kB/s]
Downloading ()lve/main/config.json: 100%|███████████████████████████████████████████████████████| 4.52k/4.52k [00:00<00:00, 1.38MB/s]
Downloading ()"pytorch_model.bin";: 100%|███████████████████████████████████████████████████████| 1.71G/1.71G [00:26<00:00, 64.2MB/s]
Some weights of the model checkpoint at openai/clip-vit-large-patch14 were not used when initializing CLIPTextModel: ['vision_model.encoder.layers.22.self_attn.q_proj.bias', 'vision_model.encoder.layers.12.layer_norm1.bias', 'vision_model.encoder.layers.21.self_attn.out_proj.weight', 'vision_model.encoder.layers.4.self_attn.k_proj.weight', 'vision_model.encoder.layers.21.self_attn.q_proj.bias', 'vision_model.encoder.layers.11.mlp.fc1.bias', 'vision_model.encoder.layers.13.layer_norm1.weight', 'vision_model.encoder.layers.9.layer_norm2.weight', 'vision_model.encoder.layers.23.self_attn.v_proj.bias', 'vision_model.encoder.layers.12.self_attn.out_proj.weight', 'vision_model.encoder.layers.22.layer_norm1.weight', 'vision_model.encoder.layers.1.self_attn.v_proj.weight', 'vision_model.encoder.layers.6.layer_norm1.weight', 'vision_model.encoder.layers.18.self_attn.k_proj.weight', 'vision_model.encoder.layers.6.self_attn.out_proj.bias', 'vision_model.encoder.layers.1.self_attn.v_proj.bias', 'vision_model.encoder.layers.19.layer_norm2.weight', 'vision_model.encoder.layers.6.self_attn.k_proj.bias', 'vision_model.encoder.layers.9.layer_norm2.bias', 'vision_model.encoder.layers.10.mlp.fc2.bias', 'vision_model.encoder.layers.19.layer_norm1.bias', 'vision_model.encoder.layers.4.layer_norm2.weight', 'vision_model.encoder.layers.16.mlp.fc1.weight', 'vision_model.encoder.layers.1.self_attn.out_proj.weight', 'vision_model.encoder.layers.11.layer_norm2.weight', 'vision_model.encoder.layers.0.self_attn.v_proj.weight', 'vision_model.encoder.layers.8.layer_norm2.weight', 'vision_model.encoder.layers.7.self_attn.v_proj.bias', 'vision_model.encoder.layers.2.mlp.fc1.weight', 'vision_model.encoder.layers.23.mlp.fc2.bias', 'vision_model.encoder.layers.15.mlp.fc1.weight', 'vision_model.encoder.layers.22.self_attn.k_proj.bias', 'vision_model.encoder.layers.0.layer_norm1.bias', 'vision_model.encoder.layers.16.self_attn.out_proj.weight', 'vision_model.encoder.layers.15.mlp.fc2.weight', 'vision_model.encoder.layers.6.mlp.fc2.bias', 'vision_model.encoder.layers.1.mlp.fc1.weight', 'vision_model.encoder.layers.2.self_attn.v_proj.weight', 'vision_model.encoder.layers.19.self_attn.q_proj.bias', 'vision_model.encoder.layers.10.self_attn.q_proj.bias', 'vision_model.encoder.layers.1.mlp.fc1.bias', 'vision_model.encoder.layers.23.self_attn.k_proj.weight', 'vision_model.encoder.layers.6.layer_norm1.bias', 'vision_model.encoder.layers.21.layer_norm1.weight', 'vision_model.encoder.layers.11.self_attn.q_proj.weight', 'vision_model.encoder.layers.15.self_attn.out_proj.weight', 'vision_model.encoder.layers.2.self_attn.k_proj.weight', 'vision_model.encoder.layers.3.self_attn.v_proj.weight', 'vision_model.encoder.layers.12.self_attn.v_proj.weight', 'vision_model.encoder.layers.5.mlp.fc1.weight', 'vision_model.encoder.layers.13.self_attn.k_proj.weight', 'vision_model.encoder.layers.12.mlp.fc2.bias', 'vision_model.encoder.layers.19.layer_norm1.weight', 'logit_scale', 'vision_model.encoder.layers.12.self_attn.q_proj.weight', 'vision_model.encoder.layers.17.mlp.fc1.bias', 'vision_model.encoder.layers.9.mlp.fc1.weight', 'vision_model.encoder.layers.2.mlp.fc1.bias', 'vision_model.encoder.layers.9.mlp.fc1.bias', 'vision_model.encoder.layers.15.self_attn.q_proj.bias', 'vision_model.encoder.layers.21.self_attn.q_proj.weight', 'vision_model.encoder.layers.11.self_attn.k_proj.bias', 'vision_model.encoder.layers.17.mlp.fc1.weight', 'vision_model.encoder.layers.2.layer_norm2.bias', 'vision_model.encoder.layers.5.self_attn.k_proj.bias', 'vision_model.encoder.layers.16.self_attn.k_proj.weight', 'vision_model.encoder.layers.15.self_attn.k_proj.bias', 'vision_model.encoder.layers.21.self_attn.v_proj.bias', 'vision_model.encoder.layers.15.layer_norm2.weight', 'vision_model.embeddings.position_embedding.weight', 'vision_model.encoder.layers.4.mlp.fc1.weight', 'vision_model.encoder.layers.1.self_attn.q_proj.bias', 'vision_model.encoder.layers.14.self_attn.q_proj.weight', 'vision_model.encoder.layers.12.self_attn.out_proj.bias', 'vision_model.encoder.layers.17.layer_norm2.bias', 'vision_model.encoder.layers.8.self_attn.out_proj.bias', 'vision_model.encoder.layers.6.self_attn.q_proj.bias', 'vision_model.encoder.layers.0.self_attn.q_proj.bias', 'vision_model.encoder.layers.17.self_attn.q_proj.bias', 'vision_model.encoder.layers.22.mlp.fc2.weight', 'vision_model.encoder.layers.14.self_attn.out_proj.bias', 'vision_model.encoder.layers.1.self_attn.k_proj.weight', 'vision_model.encoder.layers.15.self_attn.v_proj.weight', 'vision_model.encoder.layers.19.self_attn.out_proj.weight', 'vision_model.embeddings.position_ids', 'vision_model.encoder.layers.12.mlp.fc1.bias', 'vision_model.encoder.layers.21.mlp.fc2.weight', 'vision_model.encoder.layers.20.self_attn.v_proj.bias', 'vision_model.encoder.layers.12.layer_norm2.bias', 'vision_model.encoder.layers.22.mlp.fc1.weight', 'vision_model.encoder.layers.5.self_attn.q_proj.weight', 'vision_model.encoder.layers.16.layer_norm2.bias', 'vision_model.encoder.layers.23.self_attn.out_proj.weight', 'vision_model.encoder.layers.10.mlp.fc2.weight', 'vision_model.encoder.layers.3.self_attn.q_proj.weight', 'vision_model.encoder.layers.13.mlp.fc1.weight', 'vision_model.encoder.layers.8.self_attn.k_proj.weight', 'vision_model.encoder.layers.20.layer_norm1.bias', 'vision_model.encoder.layers.5.mlp.fc2.weight', 'vision_model.encoder.layers.17.self_attn.q_proj.weight', 'vision_model.encoder.layers.13.self_attn.out_proj.bias', 'vision_model.encoder.layers.3.layer_norm2.weight', 'vision_model.encoder.layers.22.layer_norm1.bias', 'vision_model.encoder.layers.1.layer_norm1.bias', 'vision_model.encoder.layers.4.self_attn.v_proj.bias', 'vision_model.encoder.layers.15.self_attn.out_proj.bias', 'vision_model.encoder.layers.4.self_attn.v_proj.weight', 'vision_model.encoder.layers.3.self_attn.out_proj.bias', 'vision_model.encoder.layers.4.self_attn.q_proj.weight', 'vision_model.encoder.layers.12.self_attn.k_proj.bias', 'vision_model.encoder.layers.5.layer_norm2.weight', 'vision_model.encoder.layers.4.mlp.fc1.bias', 'vision_model.encoder.layers.11.self_attn.v_proj.weight', 'vision_model.encoder.layers.18.mlp.fc1.weight', 'vision_model.encoder.layers.3.self_attn.q_proj.bias', 'vision_model.encoder.layers.21.layer_norm2.weight', 'vision_model.encoder.layers.22.self_attn.k_proj.weight', 'vision_model.encoder.layers.19.self_attn.k_proj.bias', 'vision_model.encoder.layers.10.layer_norm2.bias', 'vision_model.encoder.layers.0.self_attn.q_proj.weight', 'vision_model.encoder.layers.19.self_attn.out_proj.bias', 'vision_model.encoder.layers.5.self_attn.q_proj.bias', 'vision_model.encoder.layers.9.self_attn.out_proj.bias', 'vision_model.encoder.layers.23.self_attn.k_proj.bias', 'vision_model.encoder.layers.8.self_attn.k_proj.bias', 'vision_model.encoder.layers.7.layer_norm2.weight', 'vision_model.encoder.layers.3.mlp.fc1.bias', 'vision_model.encoder.layers.5.self_attn.out_proj.weight', 'vision_model.encoder.layers.14.self_attn.v_proj.weight', 'vision_model.encoder.layers.16.layer_norm1.weight', 'vision_model.encoder.layers.6.layer_norm2.bias', 'vision_model.encoder.layers.14.self_attn.v_proj.bias', 'vision_model.encoder.layers.2.self_attn.out_proj.bias', 'vision_model.encoder.layers.4.mlp.fc2.weight', 'vision_model.encoder.layers.9.self_attn.k_proj.bias', 'vision_model.encoder.layers.10.self_attn.k_proj.bias', 'vision_model.encoder.layers.8.mlp.fc1.bias', 'vision_model.encoder.layers.3.layer_norm1.weight', 'vision_model.encoder.layers.14.self_attn.q_proj.bias', 'vision_model.encoder.layers.8.mlp.fc1.weight', 'vision_model.encoder.layers.18.layer_norm1.bias', 'vision_model.encoder.layers.14.self_attn.k_proj.bias', 'vision_model.encoder.layers.7.layer_norm1.weight', 'vision_model.encoder.layers.1.mlp.fc2.bias', 'vision_model.encoder.layers.21.mlp.fc2.bias', 'vision_model.encoder.layers.0.mlp.fc1.weight', 'vision_model.encoder.layers.20.mlp.fc1.weight', 'vision_model.encoder.layers.13.self_attn.out_proj.weight', 'vision_model.encoder.layers.13.mlp.fc2.bias', 'vision_model.encoder.layers.18.layer_norm2.bias', 'vision_model.encoder.layers.20.mlp.fc1.bias', 'vision_model.encoder.layers.9.mlp.fc2.bias', 'vision_model.encoder.layers.13.self_attn.q_proj.weight', 'vision_model.encoder.layers.20.self_attn.v_proj.weight', 'vision_model.encoder.layers.5.layer_norm2.bias', 'vision_model.encoder.layers.17.layer_norm2.weight', 'vision_model.encoder.layers.11.self_attn.out_proj.bias', 'vision_model.encoder.layers.5.mlp.fc1.bias', 'vision_model.encoder.layers.10.self_attn.v_proj.weight', 'vision_model.embeddings.patch_embedding.weight', 'vision_model.encoder.layers.21.layer_norm2.bias', 'vision_model.encoder.layers.4.self_attn.q_proj.bias', 'vision_model.encoder.layers.1.layer_norm1.weight', 'vision_model.encoder.layers.4.self_attn.k_proj.bias', 'vision_model.encoder.layers.14.mlp.fc1.bias', 'vision_model.encoder.layers.12.mlp.fc1.weight', 'vision_model.encoder.layers.6.self_attn.out_proj.weight', 'vision_model.encoder.layers.20.self_attn.out_proj.bias', 'vision_model.encoder.layers.22.self_attn.out_proj.bias', 'vision_model.encoder.layers.0.self_attn.k_proj.weight', 'vision_model.encoder.layers.20.layer_norm2.weight', 'vision_model.encoder.layers.16.self_attn.v_proj.weight', 'vision_model.encoder.layers.9.self_attn.q_proj.weight', 'vision_model.encoder.layers.14.layer_norm1.weight', 'vision_model.encoder.layers.14.mlp.fc2.bias', 'vision_model.encoder.layers.2.self_attn.k_proj.bias', 'vision_model.encoder.layers.19.mlp.fc2.bias', 'vision_model.encoder.layers.5.self_attn.v_proj.weight', 'vision_model.post_layernorm.bias', 'vision_model.encoder.layers.3.mlp.fc1.weight', 'vision_model.encoder.layers.13.self_attn.q_proj.bias', 'vision_model.encoder.layers.23.layer_norm2.weight', 'vision_model.encoder.layers.10.layer_norm1.bias', 'vision_model.encoder.layers.11.self_attn.out_proj.weight', 'vision_model.encoder.layers.8.mlp.fc2.weight', 'vision_model.encoder.layers.6.layer_norm2.weight', 'vision_model.encoder.layers.8.self_attn.q_proj.weight', 'vision_model.encoder.layers.3.self_attn.out_proj.weight', 'vision_model.encoder.layers.8.mlp.fc2.bias', 'vision_model.encoder.layers.7.self_attn.out_proj.weight', 'vision_model.encoder.layers.20.self_attn.out_proj.weight', 'vision_model.encoder.layers.7.self_attn.out_proj.bias', 'vision_model.encoder.layers.5.self_attn.out_proj.bias', 'vision_model.encoder.layers.11.layer_norm1.weight', 'vision_model.encoder.layers.14.layer_norm2.bias', 'vision_model.encoder.layers.12.mlp.fc2.weight', 'text_projection.weight', 'vision_model.encoder.layers.23.layer_norm1.weight', 'vision_model.encoder.layers.9.layer_norm1.bias', 'vision_model.encoder.layers.16.self_attn.q_proj.bias', 'vision_model.encoder.layers.16.self_attn.k_proj.bias', 'vision_model.encoder.layers.17.self_attn.v_proj.weight', 'vision_model.encoder.layers.21.mlp.fc1.weight', 'vision_model.encoder.layers.23.self_attn.q_proj.weight', 'vision_model.encoder.layers.17.layer_norm1.bias', 'vision_model.encoder.layers.2.layer_norm2.weight', 'vision_model.encoder.layers.21.self_attn.k_proj.weight', 'vision_model.encoder.layers.0.layer_norm2.bias', 'vision_model.encoder.layers.2.self_attn.v_proj.bias', 'vision_model.encoder.layers.7.mlp.fc1.weight', 'vision_model.encoder.layers.19.mlp.fc2.weight', 'vision_model.encoder.layers.14.self_attn.out_proj.weight', 'vision_model.encoder.layers.23.layer_norm1.bias', 'vision_model.encoder.layers.11.self_attn.v_proj.bias', 'vision_model.encoder.layers.15.mlp.fc1.bias', 'vision_model.encoder.layers.9.layer_norm1.weight', 'vision_model.encoder.layers.6.self_attn.k_proj.weight', 'vision_model.encoder.layers.8.layer_norm1.bias', 'vision_model.encoder.layers.20.layer_norm1.weight', 'vision_model.encoder.layers.22.layer_norm2.bias', 'vision_model.encoder.layers.8.self_attn.v_proj.bias', 'vision_model.encoder.layers.20.self_attn.k_proj.bias', 'vision_model.encoder.layers.20.mlp.fc2.bias', 'vision_model.encoder.layers.20.layer_norm2.bias', 'vision_model.encoder.layers.8.self_attn.v_proj.weight', 'vision_model.encoder.layers.8.self_attn.out_proj.weight', 'vision_model.encoder.layers.16.layer_norm1.bias', 'vision_model.encoder.layers.19.mlp.fc1.weight', 'vision_model.encoder.layers.2.self_attn.q_proj.bias', 'vision_model.encoder.layers.4.self_attn.out_proj.weight', 'vision_model.encoder.layers.17.self_attn.k_proj.weight', 'vision_model.encoder.layers.22.layer_norm2.weight', 'vision_model.encoder.layers.12.self_attn.v_proj.bias', 'vision_model.encoder.layers.0.self_attn.k_proj.bias', 'vision_model.encoder.layers.6.mlp.fc1.weight', 'vision_model.encoder.layers.23.mlp.fc1.weight', 'vision_model.encoder.layers.5.self_attn.k_proj.weight', 'vision_model.encoder.layers.5.mlp.fc2.bias', 'vision_model.encoder.layers.3.self_attn.k_proj.bias', 'vision_model.encoder.layers.18.mlp.fc2.weight', 'vision_model.encoder.layers.9.self_attn.out_proj.weight', 'vision_model.encoder.layers.8.layer_norm1.weight', 'vision_model.encoder.layers.18.self_attn.out_proj.bias', 'vision_model.encoder.layers.14.layer_norm1.bias', 'vision_model.post_layernorm.weight', 'vision_model.encoder.layers.13.self_attn.v_proj.bias', 'vision_model.encoder.layers.6.mlp.fc1.bias', 'vision_model.encoder.layers.20.mlp.fc2.weight', 'vision_model.encoder.layers.14.mlp.fc1.weight', 'vision_model.encoder.layers.11.mlp.fc1.weight', 'vision_model.encoder.layers.7.self_attn.k_proj.weight', 'vision_model.encoder.layers.22.mlp.fc2.bias', 'vision_model.encoder.layers.12.layer_norm2.weight', 'vision_model.encoder.layers.1.self_attn.k_proj.bias', 'vision_model.encoder.layers.15.layer_norm1.bias', 'vision_model.encoder.layers.17.self_attn.v_proj.bias', 'vision_model.encoder.layers.19.self_attn.v_proj.weight', 'vision_model.encoder.layers.13.layer_norm2.bias', 'vision_model.encoder.layers.0.mlp.fc1.bias', 'vision_model.encoder.layers.4.mlp.fc2.bias', 'vision_model.encoder.layers.11.layer_norm2.bias', 'vision_model.encoder.layers.15.self_attn.q_proj.weight', 'vision_model.pre_layrnorm.weight', 'vision_model.encoder.layers.0.self_attn.out_proj.bias', 'vision_model.encoder.layers.22.self_attn.q_proj.weight', 'vision_model.encoder.layers.4.layer_norm2.bias', 'vision_model.encoder.layers.10.self_attn.k_proj.weight', 'vision_model.encoder.layers.16.self_attn.q_proj.weight', 'vision_model.encoder.layers.7.mlp.fc2.bias', 'vision_model.encoder.layers.2.self_attn.q_proj.weight', 'vision_model.encoder.layers.0.layer_norm1.weight', 'vision_model.encoder.layers.5.self_attn.v_proj.bias', 'vision_model.encoder.layers.23.mlp.fc2.weight', 'vision_model.encoder.layers.16.mlp.fc2.bias', 'vision_model.encoder.layers.4.layer_norm1.weight', 'vision_model.encoder.layers.11.mlp.fc2.bias', 'vision_model.encoder.layers.7.mlp.fc2.weight', 'vision_model.encoder.layers.16.layer_norm2.weight', 'vision_model.encoder.layers.3.layer_norm2.bias', 'vision_model.encoder.layers.18.self_attn.v_proj.weight', 'vision_model.encoder.layers.2.mlp.fc2.weight', 'vision_model.encoder.layers.0.mlp.fc2.bias', 'vision_model.encoder.layers.23.self_attn.q_proj.bias', 'vision_model.encoder.layers.5.layer_norm1.bias', 'vision_model.encoder.layers.13.self_attn.k_proj.bias', 'vision_model.encoder.layers.23.self_attn.v_proj.weight', 'vision_model.encoder.layers.17.mlp.fc2.weight', 'vision_model.encoder.layers.18.mlp.fc1.bias', 'vision_model.encoder.layers.18.mlp.fc2.bias', 'vision_model.encoder.layers.17.layer_norm1.weight', 'vision_model.encoder.layers.13.layer_norm2.weight', 'vision_model.encoder.layers.23.mlp.fc1.bias', 'vision_model.encoder.layers.20.self_attn.q_proj.weight', 'vision_model.encoder.layers.7.self_attn.v_proj.weight', 'vision_model.encoder.layers.3.mlp.fc2.bias', 'vision_model.encoder.layers.2.layer_norm1.weight', 'vision_model.encoder.layers.0.mlp.fc2.weight', 'vision_model.encoder.layers.12.self_attn.q_proj.bias', 'vision_model.encoder.layers.0.self_attn.out_proj.weight', 'vision_model.encoder.layers.17.self_attn.k_proj.bias', 'vision_model.encoder.layers.18.self_attn.q_proj.bias', 'vision_model.encoder.layers.3.layer_norm1.bias', 'vision_model.encoder.layers.0.self_attn.v_proj.bias', 'vision_model.encoder.layers.1.layer_norm2.weight', 'vision_model.encoder.layers.10.layer_norm1.weight', 'vision_model.encoder.layers.23.self_attn.out_proj.bias', 'vision_model.encoder.layers.19.layer_norm2.bias', 'vision_model.encoder.layers.9.self_attn.v_proj.weight', 'vision_model.encoder.layers.18.self_attn.k_proj.bias', 'vision_model.encoder.layers.12.layer_norm1.weight', 'vision_model.encoder.layers.15.self_attn.k_proj.weight', 'vision_model.encoder.layers.21.mlp.fc1.bias', 'vision_model.encoder.layers.8.layer_norm2.bias', 'vision_model.encoder.layers.23.layer_norm2.bias', 'vision_model.encoder.layers.7.layer_norm2.bias', 'vision_model.encoder.layers.14.self_attn.k_proj.weight', 'vision_model.encoder.layers.21.self_attn.out_proj.bias', 'vision_model.encoder.layers.2.layer_norm1.bias', 'vision_model.encoder.layers.16.self_attn.out_proj.bias', 'vision_model.encoder.layers.9.self_attn.v_proj.bias', 'vision_model.pre_layrnorm.bias', 'vision_model.encoder.layers.19.mlp.fc1.bias', 'vision_model.encoder.layers.3.mlp.fc2.weight', 'vision_model.encoder.layers.1.self_attn.out_proj.bias', 'vision_model.encoder.layers.15.mlp.fc2.bias', 'vision_model.encoder.layers.1.layer_norm2.bias', 'vision_model.encoder.layers.5.layer_norm1.weight', 'vision_model.encoder.layers.18.layer_norm1.weight', 'vision_model.encoder.layers.9.self_attn.q_proj.bias', 'vision_model.encoder.layers.15.layer_norm1.weight', 'vision_model.encoder.layers.19.self_attn.v_proj.bias', 'vision_model.encoder.layers.10.mlp.fc1.bias', 'vision_model.encoder.layers.13.self_attn.v_proj.weight', 'vision_model.encoder.layers.13.mlp.fc1.bias', 'vision_model.encoder.layers.22.mlp.fc1.bias', 'vision_model.encoder.layers.6.mlp.fc2.weight', 'vision_model.encoder.layers.18.self_attn.v_proj.bias', 'visual_projection.weight', 'vision_model.encoder.layers.17.self_attn.out_proj.weight', 'vision_model.encoder.layers.0.layer_norm2.weight', 'vision_model.encoder.layers.10.self_attn.out_proj.weight', 'vision_model.encoder.layers.16.mlp.fc2.weight', 'vision_model.encoder.layers.7.self_attn.q_proj.bias', 'vision_model.encoder.layers.10.self_attn.v_proj.bias', 'vision_model.encoder.layers.7.self_attn.q_proj.weight', 'vision_model.encoder.layers.22.self_attn.v_proj.bias', 'vision_model.encoder.layers.3.self_attn.k_proj.weight', 'vision_model.encoder.layers.11.self_attn.q_proj.bias', 'vision_model.encoder.layers.18.self_attn.q_proj.weight', 'vision_model.embeddings.class_embedding', 'vision_model.encoder.layers.11.mlp.fc2.weight', 'vision_model.encoder.layers.15.self_attn.v_proj.bias', 'vision_model.encoder.layers.16.mlp.fc1.bias', 'vision_model.encoder.layers.6.self_attn.v_proj.weight', 'vision_model.encoder.layers.9.self_attn.k_proj.weight', 'vision_model.encoder.layers.21.self_attn.k_proj.bias', 'vision_model.encoder.layers.21.layer_norm1.bias', 'vision_model.encoder.layers.10.layer_norm2.weight', 'vision_model.encoder.layers.14.mlp.fc2.weight', 'vision_model.encoder.layers.4.self_attn.out_proj.bias', 'vision_model.encoder.layers.13.mlp.fc2.weight', 'vision_model.encoder.layers.2.self_attn.out_proj.weight', 'vision_model.encoder.layers.19.self_attn.k_proj.weight', 'vision_model.encoder.layers.6.self_attn.q_proj.weight', 'vision_model.encoder.layers.13.layer_norm1.bias', 'vision_model.encoder.layers.6.self_attn.v_proj.bias', 'vision_model.encoder.layers.17.mlp.fc2.bias', 'vision_model.encoder.layers.1.mlp.fc2.weight', 'vision_model.encoder.layers.18.self_attn.out_proj.weight', 'vision_model.encoder.layers.22.self_attn.v_proj.weight', 'vision_model.encoder.layers.10.self_attn.q_proj.weight', 'vision_model.encoder.layers.22.self_attn.out_proj.weight', 'vision_model.encoder.layers.20.self_attn.q_proj.bias', 'vision_model.encoder.layers.19.self_attn.q_proj.weight', 'vision_model.encoder.layers.18.layer_norm2.weight', 'vision_model.encoder.layers.12.self_attn.k_proj.weight', 'vision_model.encoder.layers.21.self_attn.v_proj.weight', 'vision_model.encoder.layers.1.self_attn.q_proj.weight', 'vision_model.encoder.layers.11.self_attn.k_proj.weight', 'vision_model.encoder.layers.9.mlp.fc2.weight', 'vision_model.encoder.layers.7.mlp.fc1.bias', 'vision_model.encoder.layers.8.self_attn.q_proj.bias', 'vision_model.encoder.layers.7.layer_norm1.bias', 'vision_model.encoder.layers.14.layer_norm2.weight', 'vision_model.encoder.layers.7.self_attn.k_proj.bias', 'vision_model.encoder.layers.10.mlp.fc1.weight', 'vision_model.encoder.layers.17.self_attn.out_proj.bias', 'vision_model.encoder.layers.20.self_attn.k_proj.weight', 'vision_model.encoder.layers.10.self_attn.out_proj.bias', 'vision_model.encoder.layers.15.layer_norm2.bias', 'vision_model.encoder.layers.2.mlp.fc2.bias', 'vision_model.encoder.layers.4.layer_norm1.bias', 'vision_model.encoder.layers.11.layer_norm1.bias', 'vision_model.encoder.layers.3.self_attn.v_proj.bias', 'vision_model.encoder.layers.16.self_attn.v_proj.bias']
- This IS expected if you are initializing CLIPTextModel from the checkpoint of a model trained on another task or with another architecture (e.g. initializing a BertForSequenceClassification model from a BertForPreTraining model).
- This IS NOT expected if you are initializing CLIPTextModel from the checkpoint of a model that you expect to be exactly identical (initializing a BertForSequenceClassification model from a BertForSequenceClassification model).
Creating invisible watermark encoder (see https://github.com/ShieldMnt/invisible-watermark)...
Sampling:   0%|                                                                                                 | 0/1 [00:00<?, ?it/sData shape for PLMS sampling is (1, 4, 64, 64)                                                                   | 0/1 [00:00<?, ?it/s]
Running PLMS Sampling with 50 timesteps
PLMS Sampler: 100%|███████████████████████████████████████████████████████████████████████████████████| 50/50 [04:11<00:00,  5.03s/it]
data: 100%|████████████████████████████████████████████████████████████████████████████████████████████| 1/1 [04:21<00:00, 261.23s/it]
Sampling: 100%|████████████████████████████████████████████████████████████████████████████████████████| 1/1 [04:21<00:00, 261.23s/it]
Your samples are ready and waiting for you here: 
outputs/txt2img-samples 
 
Enjoy.
stable-diffusion % 
0
4
0

Register as a new user and use Qiita more conveniently

  1. You get articles that match your needs
  2. You can efficiently read back useful information
  3. You can use dark theme
What you can do with signing up
0
4

Delete article

Deleted articles cannot be recovered.

Draft of this article would be also deleted.

Are you sure you want to delete this article?