教程:Stable Diffusion 安装过程中常见报错解决方法
Stable Diffusion|安装过程中常见报错解决方法
大家可能在安装部署 Stable Diffusion WebUI 过程中遇到各种各样的问题,本站已对所有问题进行汇总整理,而且教大家如何去查看报错。解决常见的 Installing gfpgan、BLIP 错误、显存不足等各种各样的花式问题,希望大家在看完此文章之后,都能够顺利解决各自所遇到的问题。
如何查看报错
在你安装时可能经常遇到各种各样的问题,但是对于一堆陌生的英文和各种各样的错误,大家可能经常无从下手,下面我将会教大家如何查看报错。
(base) C:\OpenAI.Wiki\stable-diffusion-webui>webui-user.bat
venv “C:\OpenAI.Wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash: 955df7751eef11bb7697e2d77f6b8a6226b21e13
Installing clip
Installing open_clip
Cloning Stable Diffusion into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai…
Cloning Taming Transformers into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\taming-transformers…
Cloning K-diffusion into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\k-diffusion…
Cloning CodeFormer into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer…
Traceback (most recent call last):
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 351, in
prepare_environment()
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 287, in prepare_environment
git_clone(codeformer_repo, repo_dir(‘CodeFormer’), “CodeFormer”, codeformer_commit_hash)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 151, in git_clone
run(f'”{git}” clone “{url}” “{dir}”‘, f”Cloning {name} into {dir}…”, f”Couldn’t clone {name}”)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t clone CodeFormer.
Command: “git” clone “https://github.com/sczhou/CodeFormer.git” “C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer”
Error code: 128
stdout:
stderr: Cloning into ‘C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer’…
我们先看一下第一行内容(base) C:\OpenAI.Wiki\stable-diffusion-webui>webui-user.bat
,前面的(base)是使用 conda 创建的虚拟环境名称,代表我们目前正在使用哪一个虚拟环境,很显然这与本站所提供的教程是不一致的,正常来说应该是(D:\openai.wiki\stable-diffusion-webui\automatic)
,这一串路径就是虚拟环境的名称。
第 5 行和第 6 行分别是Installing clip
和Installing open_clip
,这是两个依赖库。这里没有报错,因为后面是是 Cloning 一些其它的依赖库。
第 11 行Traceback (most recent call last):
,这就是代表出现问题了,我们在出现问题的前一行中可以看到内容为Cloning CodeFormer into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer…
,这就是代表是一步出错了。
网络问题
解决方法
本部分内容基本都是因为网络问题,大部分安装问题可以通过使用国内镜像源来解决。
例如:我们需要在执行安装 requirements.txt 文件时,我们可以尝试使用pip install -i https://mirrors.aliyun.com/pypi/simple/ -r D:/openai.wiki/stable-diffusion-webui/requirements.txt
。
这段代码可以理解为不使用官方下载地址,而是国内阿里云的镜像地址下载相关依赖组件。
如果还是不能解决,请自行搜索CMD魔法上网
,有一些魔法工具即便可以访问 Youtube、Google 等网站,也不代表它能够在 CMD 中是可以正常使用的,因为有一些魔法工具无法在 CMD 中被继承。
Stable-Diffusion-Stability-AI
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash: 22bcc7be428c94e9408f589966c2040187245d81
Cloning Stable Diffusion into D:\openai.wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai…
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 355, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 288, in prepare_environment
git_clone(stable_diffusion_repo, repo_dir(‘stable-diffusion-stability-ai’), “Stable Diffusion”, stable_diffusion_commit_hash)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 151, in git_clone
run(f'”{git}” clone “{url}” “{dir}”‘, f”Cloning {name} into {dir}…”, f”Couldn’t clone {name}”)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t clone Stable Diffusion.
Command: “git” clone “https://github.com/Stability-AI/stablediffusion.git” “D:\openai.wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai”
Error code: 128
stdout:
stderr: Cloning into ‘D:\openai.wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai’…
fatal: unable to access ‘https://github.com/Stability-AI/stablediffusion.git/’: OpenSSL SSL_connect: SSL_ERROR_SYSCALL in connection to github.com:443
脚本在执行git clone https://github.com/Stability-AI/stablediffusion.git
命令时无法下载Stable-Diffusion-Stability-AI
库。
PyPI 错误
note: This error originates from a subprocess, and is likely not a problem with pip.
error: metadata-generation-failed
Encountered error while generating package metadata.
See above for output.
note: This is an issue with the package mentioned above, not pip.
hint: See above for details.
PyPI 元数据包有问题,可以尝试执行pip cache purge
清理缓存。
哈希值不正确
(base) tianjihuideMBP:stable-diffusion-webui color$ ./webui.sh
################################################################
Install script for stable-diffusion + Web UI
Tested on Debian 11 (Bullseye)
################################################################
################################################################
Running on color user
################################################################
################################################################
Repo already cloned, using it as install directory
################################################################
################################################################
Create and activate python venv
################################################################
################################################################
Launching launch.py…
################################################################
Python 3.10.10 (v3.10.10:aad5f6a891, Feb 7 2023, 08:47:40) [Clang 13.0.0 (clang-1300.0.29.30)]
Commit hash:
Traceback (most recent call last):
File “/Users/color/stable-diffusion-webui/launch.py”, line 378, in
prepare_environment()
File “/Users/color/stable-diffusion-webui/launch.py”, line 315, in prepare_environment
git_clone(stable_diffusion_repo, repo_dir(‘stable-diffusion-stability-ai’), “Stable Diffusion”,
File “/Users/color/stable-diffusion-webui/launch.py”, line 152, in git_clone
current_hash = run(f'”{git}” -C “{dir}” rev-parse HEAD’, None,
File “/Users/color/stable-diffusion-webui/launch.py”, line 105, in run
raise RuntimeError(message)
RuntimeError: Couldn’t determine Stable Diffusion’s hash: 47b6b607fdd31875c9279cd2f4f16b92e4ea958e.
Command: “git” -C “repositories/stable-diffusion-stability-ai” rev-parse HEAD
Error code: 128
stdout: HEAD
stderr: fatal: ambiguous argument ‘HEAD’: unknown revision or path not in the working tree.
Use ‘–‘ to separate paths from revisions, like this:
‘git […] — […]’
(base) tianjihuideMBP:stable-diffusion-webui color$
这个错误似乎是在运行一个 Python 脚本时出现的。根据输出信息来看,该脚本尝试克隆一个名为 Stable Diffusion 的 git 仓库,但是它无法确定该仓库的哈希值。
建议删除整个项目,然后重新安装部署。
组件不匹配
ERROR: Could not find a version that satisfies the requirement opencv-contrib-python (from versions: none)
ERROR: No matching distribution found for opencv-contrib-python
这个错误信息意味着在当前使用的 Python 环境下,没有找到匹配的opencv-contrib-python
版本。这可能是因为在 PyPI 上没有发布所需的版本或版本不兼容。请尝试使用pip search
命令来搜索可用的版本,并尝试安装兼容的版本。
按道理来说,如果根据教程来安装是不会出现这个问题的,尽量删除后从头来过,严格遵守每一步。
CodeFormer
(base) C:\OpenAI.Wiki\stable-diffusion-webui>webui-user.bat
venv “C:\OpenAI.Wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash: 955df7751eef11bb7697e2d77f6b8a6226b21e13
Installing clip
Installing open_clip
Cloning Stable Diffusion into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai…
Cloning Taming Transformers into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\taming-transformers…
Cloning K-diffusion into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\k-diffusion…
Cloning CodeFormer into C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer…
Traceback (most recent call last):
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 351, in
prepare_environment()
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 287, in prepare_environment
git_clone(codeformer_repo, repo_dir(‘CodeFormer’), “CodeFormer”, codeformer_commit_hash)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 151, in git_clone
run(f'”{git}” clone “{url}” “{dir}”‘, f”Cloning {name} into {dir}…”, f”Couldn’t clone {name}”)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t clone CodeFormer.
Command: “git” clone “https://github.com/sczhou/CodeFormer.git” “C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer”
Error code: 128
stdout:
stderr: Cloning into ‘C:\OpenAI.Wiki\stable-diffusion-webui\repositories\CodeFormer’…
这里看起来是在克隆 CodeFormer 时出了问题。错误信息表明 Git 无法克隆该仓库。请确保您的计算机可以访问 https://github.com/sczhou/CodeFormer.git ,并且您有权访问该仓库。您可以尝试手动克隆该仓库以查看是否存在任何错误。
无法链接 SD 库
venv “C:\OpenAI.Wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash: 64da5c46ef0d68b9048747c2e0d46ce3495f9f29
Fetching updates for Stable Diffusion…
Traceback (most recent call last):
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 351, in
prepare_environment()
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 284, in prepare_environment
git_clone(stable_diffusion_repo, repo_dir(‘stable-diffusion-stability-ai’), “Stable Diffusion”, stable_diffusion_com
mit_hash)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 147, in git_clone
run(f'”{git}” -C “{dir}” fetch’, f”Fetching updates for {name}…”, f”Couldn’t fetch {name}”)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t fetch Stable Diffusion.
Command: “git” -C “C:\OpenAI.Wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai” fetch
Error code: 128
stdout:
stderr: fatal: unable to access ‘https://github.com/AUTOMATIC1111/stable-diffusion-webui.git/’: OpenSSL SSL_read: Connec
tion was reset, errno 10054
这个错误信息看起来是从 Github 下载 Stable Diffusion 的时候出错了,可能是网络连接问题。建议检查网络连接是否能够正常使用魔法,
SD 库不完整
venv “C:\OpenAI.Wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash: 64da5c46ef0d68b9048747c2e0d46ce3495f9f29
Fetching updates for Stable Diffusion…
Checking out commit for Stable Diffusion with hash: 47b6b607fdd31875c9279cd2f4f16b92e4ea958e…
Traceback (most recent call last):
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 351, in
prepare_environment()
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 284, in prepare_environment
git_clone(stable_diffusion_repo, repo_dir(‘stable-diffusion-stability-ai’), “Stable Diffusion”, stable_diffusion_com
mit_hash)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 148, in git_clone
run(f'”{git}” -C “{dir}” checkout {commithash}’, f”Checking out commit for {name} with hash: {commithash}…”, f”Cou
ldn’t checkout commit {commithash} for {name}”)
File “C:\OpenAI.Wiki\stable-diffusion-webui\launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t checkout commit 47b6b607fdd31875c9279cd2f4f16b92e4ea958e for Stable Diffusion.
Command: “git” -C “C:\OpenAI.Wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai” checkout 47b6b607fd
d31875c9279cd2f4f16b92e4ea958e
Error code: 128
stdout:
stderr: fatal: reference is not a tree: 47b6b607fdd31875c9279cd2f4f16b92e4ea958e
这个错误表示在检出 Stable Diffusion 时出现问题。可能是由于该库的某些文件或依赖项已更改或已移除,导致无法找到所需的提交。请尝试清除本地库并重新克隆 Stable Diffusion,然后再次运行代码。
未找到 pip
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 (tags/v3.10.10:aad5f6a, Feb 7 2023, 17:20:36) [MSC v.1929 64 bit (AMD64)]
Commit hash: 4c1ad743e3baf1246db0711aa0107debf036a12b
Installing torch and torchvision
D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe: No module named pip
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 351, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 253, in prepare_environment
run(f'”{python}” -m {torch_command}’, “Installing torch and torchvision”, “Couldn’t install torch”, live=True)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 81, in run
raise RuntimeError(f”””{errdesc or ‘Error running command’}.
RuntimeError: Couldn’t install torch.
Command: “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe” -m pip install torch==1.13.1+cu117 torchvision==0.14.1+cu117 –extra-index-url https://download.pytorch.org/whl/cu117
Error code: 1
根据错误日志,似乎在运行 launch.py
文件尝试安装 torch 和 torchvision。然而,这个命令在运行时出现了错误,错误信息是 No module named pip
。
这可能是因为 Python 环境中没有安装 pip,所以在运行 pip 命令时出现了问题。
你可以尝试先安装 pip,然后再运行 launch.py
。你可以使用以下命令来安装 pip:
python get-pip.py
这应该会在你的 Python 环境中安装 pip。然后你可以再次运行启动文件即可。
open_clip
F:\openai.wiki\stable-diffusion-webui>webui-user.bat
venv “F:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 8 2023, 10:42:25) [MSC v.1916 64 bit (AMD64)]
Commit hash: a9fed7c364061ae6efb37f797b6b522cb3cf7aa2
Installing open_clip
Traceback (most recent call last):
File “F:\openai.wiki\stable-diffusion-webui\launch.py”, line 380, in
prepare_environment()
File “F:\openai.wiki\stable-diffusion-webui\launch.py”, line 296, in prepare_environment
run_pip(f”install {openclip_package}”, “open_clip”)
File “F:\openai.wiki\stable-diffusion-webui\launch.py”, line 145, in run_pip
return run(f'”{python}” -m pip {args} –prefer-binary{index_url_line}’, desc=f”Installing {desc}”, errdesc=f”Couldn’t install {desc}”)
File “F:\openai.wiki\stable-diffusion-webui\launch.py”, line 113, in run
raise RuntimeError(message)
RuntimeError: Couldn’t install open_clip.
Command: “F:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe” -m pip install git+https://github.com/mlfoundations/open_clip.git@bb6e834e9c70d9c27d0dc3ecedeebeaeb1ffad6b –prefer-binary
Error code: 1
stdout: Collecting git+https://github.com/mlfoundations/open_clip.git@bb6e834e9c70d9c27d0dc3ecedeebeaeb1ffad6b
Cloning https://github.com/mlfoundations/open_clip.git (to revision bb6e834e9c70d9c27d0dc3ecedeebeaeb1ffad6b) to c:\users\administrator\appdata\local\temp\pip-req-build-0bs_j2f1
stderr: Running command git clone –filter=blob:none –quiet https://github.com/mlfoundations/open_clip.git ‘C:\Users\Administrator\AppData\Local\Temp\pip-req-build-0bs_j2f1’
error: RPC failed; curl 18 HTTP/2 stream 3 was not closed cleanly before end of the underlying stream
fatal: expected flush after ref listing
error: subprocess-exited-with-error
git clone –filter=blob:none –quiet https://github.com/mlfoundations/open_clip.git ‘C:\Users\Administrator\AppData\Local\Temp\pip-req-build-0bs_j2f1’ did not run successfully.
exit code: 128
See above for output.
note: This error originates from a subprocess, and is likely not a problem with pip.
error: subprocess-exited-with-error
git clone –filter=blob:none –quiet https://github.com/mlfoundations/open_clip.git ‘C:\Users\Administrator\AppData\Local\Temp\pip-req-build-0bs_j2f1’ did not run successfully.
exit code: 128
See above for output.
note: This error originates from a subprocess, and is likely not a problem with pip.
[notice] A new release of pip available: 22.3.1 -> 23.0.1
[notice] To update, run: F:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe -m pip install –upgrade pip
这个错误提示是在运行一个名为“webui-user.bat”的文件时出现的。错误提示表明程序试图安装名为“open_clip”的包,但是安装失败了。下面是详细的错误信息:
Running command git clone –filter=blob:none –quiet https://github.com/mlfoundations/open_clip.git ‘C:\Users\Administrator\AppData\Local\Temp\pip-req-build-0bs_j2f1’
error: RPC failed; curl 18 HTTP/2 stream 3 was not closed cleanly before end of the underlying stream
fatal: expected flush after ref listing
error: subprocess-exited-with-error
git clone –filter=blob:none –quiet https://github.com/mlfoundations/open_clip.git ‘C:\Users\Administrator\AppData\Local\Temp\pip-req-build-0bs_j2f1’ did not run successfully.
exit code: 128
See above for output.
note: This error originates from a subprocess, and is likely not a problem with pip.
根据错误信息分析,程序在安装“open_clip”包时,尝试从 GitHub 上克隆代码仓库,但是克隆失败了,可能是由于网络连接问题或 GitHub 服务器出现问题导致的。您可以尝试重新运行程序,或稍后再试,看看能否成功安装该包。
SD 库不正确
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Commit hash: a9fed7c364061ae6efb37f797b6b522cb3cf7aa2
Fetching updates for Stable Diffusion…
Checking out commit for Stable Diffusion with hash: 47b6b607fdd31875c9279cd2f4f16b92e4ea958e…
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 380, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 315, in prepare_environment
git_clone(stable_diffusion_repo, repo_dir(‘stable-diffusion-stability-ai’), “Stable Diffusion”, stable_diffusion_commit_hash)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 164, in git_clone
run(f'”{git}” -C “{dir}” checkout {commithash}’, f”Checking out commit for {name} with hash: {commithash}…”, f”Couldn’t checkout commit {commithash} for {name}”)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 113, in run
raise RuntimeError(message)
RuntimeError: Couldn’t checkout commit 47b6b607fdd31875c9279cd2f4f16b92e4ea958e for Stable Diffusion.
Command: “git” -C “D:\openai.wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai” checkout 47b6b607fdd31875c9279cd2f4f16b92e4ea958e
Error code: 128
stdout:
stderr: fatal: reference is not a tree: 47b6b607fdd31875c9279cd2f4f16b92e4ea958e
根据您提供的信息,看起来是在运行一个名为 launch.py
的 Python 脚本时发生了错误。根据错误信息,似乎出现了 Git 无法检出提交的问题。这可能是由于 Git 存储库的状态不正确或由于网络连接问题而导致的。您可以尝试运行以下命令来检查网络连接:
ping files.pythonhosted.org
PIP 超时
ERROR: Exception:
Traceback (most recent call last):
File “E:\conda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 437, in _error_catcher
yield
File “E:\conda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 560, in read
data = self._fp_read(amt) if not fp_closed else b””
File “E:\conda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 526, in _fp_read
return self._fp.read(amt) if amt is not None else self._fp.read()
File “E:\conda\lib\site-packages\pip\_vendor\cachecontrol\filewrapper.py”, line 90, in read
data = self.__fp.read(amt)
File “E:\conda\lib\http\client.py”, line 465, in read
s = self.fp.read(amt)
File “E:\conda\lib\socket.py”, line 705, in readinto
return self._sock.recv_into(b)
File “E:\conda\lib\ssl.py”, line 1274, in recv_into
return self.read(nbytes, buffer)
File “E:\conda\lib\ssl.py”, line 1130, in read
return self._sslobj.read(len, buffer)
TimeoutError: The read operation timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File “E:\conda\lib\site-packages\pip\_internal\cli\base_command.py”, line 160, in exc_logging_wrapper
status = run_func(*args)
File “E:\conda\lib\site-packages\pip\_internal\cli\req_command.py”, line 247, in wrapper
return func(self, options, args)
File “E:\conda\lib\site-packages\pip\_internal\commands\install.py”, line 400, in run
requirement_set = resolver.resolve(
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\resolver.py”, line 92, in resolve
result = self._result = resolver.resolve(
File “E:\conda\lib\site-packages\pip\_vendor\resolvelib\resolvers.py”, line 481, in resolve
state = resolution.resolve(requirements, max_rounds=max_rounds)
File “E:\conda\lib\site-packages\pip\_vendor\resolvelib\resolvers.py”, line 348, in resolve
self._add_to_criteria(self.state.criteria, r, parent=None)
File “E:\conda\lib\site-packages\pip\_vendor\resolvelib\resolvers.py”, line 172, in _add_to_criteria
if not criterion.candidates:
File “E:\conda\lib\site-packages\pip\_vendor\resolvelib\structs.py”, line 151, in bool
return bool(self._sequence)
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\found_candidates.py”, line 155, in bool
return any(self)
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\found_candidates.py”, line 143, in
return (c for c in iterator if id(c) not in self._incompatible_ids)
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\found_candidates.py”, line 47, in _iter_built
candidate = func()
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\factory.py”, line 206, in _make_candidate_from_link
self._link_candidate_cache[link] = LinkCandidate(
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 297, in init
super().__init__(
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 162, in init
self.dist = self._prepare()
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 231, in _prepare
dist = self._prepare_distribution()
File “E:\conda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 308, in _prepare_distribution
return preparer.prepare_linked_requirement(self._ireq, parallel_builds=True)
File “E:\conda\lib\site-packages\pip\_internal\operations\prepare.py”, line 491, in prepare_linked_requirement
return self._prepare_linked_requirement(req, parallel_builds)
File “E:\conda\lib\site-packages\pip\_internal\operations\prepare.py”, line 536, in _prepare_linked_requirement
local_file = unpack_url(
File “E:\conda\lib\site-packages\pip\_internal\operations\prepare.py”, line 166, in unpack_url
file = get_http_url(
File “E:\conda\lib\site-packages\pip\_internal\operations\prepare.py”, line 107, in get_http_url
from_path, content_type = download(link, temp_dir.path)
File “E:\conda\lib\site-packages\pip\_internal\network\download.py”, line 147, in call
for chunk in chunks:
File “E:\conda\lib\site-packages\pip\_internal\cli\progress_bars.py”, line 53, in _rich_progress_bar
for chunk in iterable:
File “E:\conda\lib\site-packages\pip\_internal\network\utils.py”, line 63, in response_chunks
for chunk in response.raw.stream(
File “E:\conda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 621, in stream
data = self.read(amt=amt, decode_content=decode_content)
File “E:\conda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 559, in read
with self._error_catcher():
File “E:\conda\lib\contextlib.py”, line 153, in exit
self.gen.throw(typ, value, traceback)
File “E:\conda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 442, in _error_catcher
raise ReadTimeoutError(self._pool, None, “Read timed out.”)
pip._vendor.urllib3.exceptions.ReadTimeoutError: HTTPSConnectionPool(host=’files.pythonhosted.org’, port=443): Read timed out.
这个错误通常是因为 pip 尝试从远程服务器下载软件包时连接超时。你可以尝试重新运行命令,或者尝试使用 pip 的“–default-timeout”选项来增加超时时间,如下所示:
pip install --default-timeout=1000 package_name
这里,超时时间为 1000 秒。如果仍然无法解决问题,你可以尝试更改你的网络连接。
requirements
卡在了 installing requirements for web ui 这一步
输出显示出现了一个 SSL 错误,它可能是由于网络问题导致的。你可以尝试使用其他网络环境或在网络良好的情况下重试安装。
此外,你还可以尝试在命令行中运行以下命令更新 pip,然后重新运行安装:python -m pip install --upgrade pip
xformers
No module ‘xformers’. Proceeding without it.
每次启动时显示如上内容,这并不算错误,新版 SD 不支持 xformers,可以手动更改为支持。
这个库的主要作用是可以加速生成图像,但有时候安装过后速度反而下降,所以建议保持即可,不用理会。
GFPGAN
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 8 2023, 10:42:25) [MSC v.1916 64 bit (AMD64)]
Commit hash:
Installing gfpgan
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 380, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 290, in prepare_environment
run_pip(f”install {gfpgan_package}”, “gfpgan”)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 145, in run_pip
return run(f'”{python}” -m pip {args} –prefer-binary{index_url_line}’, desc=f”Installing {desc}”, errdesc=f”Couldn’t install {desc}”)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 113, in run
raise RuntimeError(message)
RuntimeError: Couldn’t install gfpgan.
Command: “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe” -m pip install git+https://github.com/TencentARC/GFPGAN.git@8d2447a2d918f8eba5a4a01463fd48e45126a379 –prefer-binary
Error code: 1
stdout: Collecting git+https://github.com/TencentARC/GFPGAN.git@8d2447a2d918f8eba5a4a01463fd48e45126a379
Cloning https://github.com/TencentARC/GFPGAN.git (to revision 8d2447a2d918f8eba5a4a01463fd48e45126a379) to c:\users\17742\appdata\local\temp\pip-req-build-ocz06mwj
stderr: Running command git clone –filter=blob:none –quiet https://github.com/TencentARC/GFPGAN.git ‘C:\Users\17742\AppData\Local\Temp\pip-req-build-ocz06mwj’
fatal: unable to access ‘https://github.com/TencentARC/GFPGAN.git/’: Failed to connect to github.com port 443 after 21046 ms: Couldn’t connect to server
error: subprocess-exited-with-error
git clone –filter=blob:none –quiet https://github.com/TencentARC/GFPGAN.git ‘C:\Users\17742\AppData\Local\Temp\pip-req-build-ocz06mwj’ did not run successfully.
exit code: 128
See above for output.
note: This error originates from a subprocess, and is likely not a problem with pip.
error: subprocess-exited-with-error
git clone –filter=blob:none –quiet https://github.com/TencentARC/GFPGAN.git ‘C:\Users\17742\AppData\Local\Temp\pip-req-build-ocz06mwj’ did not run successfully.
exit code: 128
See above for output.
note: This error originates from a subprocess, and is likely not a problem with pip.
[notice] A new release of pip available: 22.3.1 -> 23.0.1
[notice] To update, run: D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe -m pip install –upgrade pip
这个错误显示在安装 gfpgan 时出现了问题。看起来尝试从 GitHub 克隆代码库失败了。
最有可能的原因是你的计算机无法连接到 GitHub,或者 GitHub 网站当前遇到了问题。
尝试检查你的网络连接并确保你可以连接到 GitHub,或者稍后再尝试安装。另外,你可能也可以尝试升级你的 pip 版本。
subprocess-exited-with-error
error: subprocess-exited-with-error
× python setup.py egg_info did not run successfully.
│ exit code: 1
╰─> [109 lines of output]
D:\Miniconda\lib\site-packages\setuptools\installer.py:27: SetuptoolsDeprecationWarning: setuptools.installer is deprecated. Requirements should be satisfied by a PEP 517 installer.
warnings.warn(
ERROR: Exception:
Traceback (most recent call last):
File “D:\Miniconda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 437, in _error_catcher
yield
File “D:\Miniconda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 560, in read
data = self._fp_read(amt) if not fp_closed else b””
File “D:\Miniconda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 526, in _fp_read
return self._fp.read(amt) if amt is not None else self._fp.read()
File “D:\Miniconda\lib\site-packages\pip\_vendor\cachecontrol\filewrapper.py”, line 90, in read
data = self.__fp.read(amt)
File “D:\Miniconda\lib\http\client.py”, line 465, in read
s = self.fp.read(amt)
File “D:\Miniconda\lib\socket.py”, line 705, in readinto
return self._sock.recv_into(b)
File “D:\Miniconda\lib\ssl.py”, line 1274, in recv_into
return self.read(nbytes, buffer)
File “D:\Miniconda\lib\ssl.py”, line 1130, in read
return self._sslobj.read(len, buffer)
TimeoutError: The read operation timed out
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File “D:\Miniconda\lib\site-packages\pip\_internal\cli\base_command.py”, line 160, in exc_logging_wrapper
status = run_func(*args)
File “D:\Miniconda\lib\site-packages\pip\_internal\cli\req_command.py”, line 247, in wrapper
return func(self, options, args)
File “D:\Miniconda\lib\site-packages\pip\_internal\commands\wheel.py”, line 170, in run
requirement_set = resolver.resolve(reqs, check_supported_wheels=True)
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\resolver.py”, line 92, in resolve
result = self._result = resolver.resolve(
File “D:\Miniconda\lib\site-packages\pip\_vendor\resolvelib\resolvers.py”, line 481, in resolve
state = resolution.resolve(requirements, max_rounds=max_rounds)
File “D:\Miniconda\lib\site-packages\pip\_vendor\resolvelib\resolvers.py”, line 348, in resolve
self._add_to_criteria(self.state.criteria, r, parent=None)
File “D:\Miniconda\lib\site-packages\pip\_vendor\resolvelib\resolvers.py”, line 172, in _add_to_criteria
if not criterion.candidates:
File “D:\Miniconda\lib\site-packages\pip\_vendor\resolvelib\structs.py”, line 151, in bool
return bool(self._sequence)
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\found_candidates.py”, line 155, in bool
return any(self)
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\found_candidates.py”, line 143, in
return (c for c in iterator if id(c) not in self._incompatible_ids)
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\found_candidates.py”, line 47, in _iter_built
candidate = func()
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\factory.py”, line 206, in _make_candidate_from_link
self._link_candidate_cache[link] = LinkCandidate(
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 297, in init
super().__init__(
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 162, in init
self.dist = self._prepare()
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 231, in _prepare
dist = self._prepare_distribution()
File “D:\Miniconda\lib\site-packages\pip\_internal\resolution\resolvelib\candidates.py”, line 308, in _prepare_distribution
return preparer.prepare_linked_requirement(self._ireq, parallel_builds=True)
File “D:\Miniconda\lib\site-packages\pip\_internal\operations\prepare.py”, line 491, in prepare_linked_requirement
return self._prepare_linked_requirement(req, parallel_builds)
File “D:\Miniconda\lib\site-packages\pip\_internal\operations\prepare.py”, line 536, in _prepare_linked_requirement
local_file = unpack_url(
File “D:\Miniconda\lib\site-packages\pip\_internal\operations\prepare.py”, line 166, in unpack_url
file = get_http_url(
File “D:\Miniconda\lib\site-packages\pip\_internal\operations\prepare.py”, line 107, in get_http_url
from_path, content_type = download(link, temp_dir.path)
File “D:\Miniconda\lib\site-packages\pip\_internal\network\download.py”, line 147, in call
for chunk in chunks:
File “D:\Miniconda\lib\site-packages\pip\_internal\network\utils.py”, line 63, in response_chunks
for chunk in response.raw.stream(
File “D:\Miniconda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 621, in stream
data = self.read(amt=amt, decode_content=decode_content)
File “D:\Miniconda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 559, in read
with self._error_catcher():
File “D:\Miniconda\lib\contextlib.py”, line 153, in exit
self.gen.throw(typ, value, traceback)
File “D:\Miniconda\lib\site-packages\pip\_vendor\urllib3\response.py”, line 442, in _error_catcher
raise ReadTimeoutError(self._pool, None, “Read timed out.”)
pip._vendor.urllib3.exceptions.ReadTimeoutError: HTTPSConnectionPool(host=’files.pythonhosted.org’, port=443): Read timed out.
Traceback (most recent call last):
File “D:\Miniconda\lib\site-packages\setuptools\installer.py”, line 82, in fetch_build_egg
subprocess.check_call(cmd)
File “D:\Miniconda\lib\subprocess.py”, line 369, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command ‘[‘D:\\Miniconda\\python.exe’, ‘-m’, ‘pip’, ‘–disable-pip-version-check’, ‘wheel’, ‘–no-deps’, ‘-w’, ‘C:\\Users\\23243\\AppData\\Local\\Temp\\tmpk82g_bqu’, ‘–quiet’, ‘torch’]’ returned non-zero exit status 2.
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File “”, line 2, in
File “”, line 34, in
File “C:\Users\23243\AppData\Local\Temp\pip-install-hauelb5m\basicsr_ff9fa51afe15482f8d5ed318666d9752\setup.py”, line 147, in
setup(
File “D:\Miniconda\lib\site-packages\setuptools\__init__.py”, line 86, in setup
_install_setup_requires(attrs)
File “D:\Miniconda\lib\site-packages\setuptools\__init__.py”, line 80, in _install_setup_requires
dist.fetch_build_eggs(dist.setup_requires)
File “D:\Miniconda\lib\site-packages\setuptools\dist.py”, line 874, in fetch_build_eggs
resolved_dists = pkg_resources.working_set.resolve(
File “D:\Miniconda\lib\site-packages\pkg_resources\__init__.py”, line 789, in resolve
dist = best[req.key] = env.best_match(
File “D:\Miniconda\lib\site-packages\pkg_resources\__init__.py”, line 1075, in best_match
return self.obtain(req, installer)
File “D:\Miniconda\lib\site-packages\pkg_resources\__init__.py”, line 1087, in obtain
return installer(requirement)
File “D:\Miniconda\lib\site-packages\setuptools\dist.py”, line 944, in fetch_build_egg
return fetch_build_egg(self, req)
File “D:\Miniconda\lib\site-packages\setuptools\installer.py”, line 84, in fetch_build_egg
raise DistutilsError(str(e)) from e
distutils.errors.DistutilsError: Command ‘[‘D:\\Miniconda\\python.exe’, ‘-m’, ‘pip’, ‘–disable-pip-version-check’, ‘wheel’, ‘–no-deps’, ‘-w’, ‘C:\\Users\\23243\\AppData\\Local\\Temp\\tmpk82g_bqu’, ‘–quiet’, ‘torch’]’ returned non-zero exit status 2.
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip. error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details.
这个错误可能是由于 pip 在下载软件包时遇到了网络问题,导致安装不成功。您可以尝试重新运行命令,或者检查您的网络连接是否正常。另外,您可以尝试使用国内的镜像源来安装软件包,例如使用清华大学的镜像源,命令如下:
pip install -i https://pypi.tuna.tsinghua.edu.cn/simple some-package
将 some-package
替换为您要安装的软件包的名称。
BLIP 错误
F:\openai.wiki\stable-diffusion-webui-master>webui-user.bat
venv “F:\openai.wiki\stable-diffusion-webui-master\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 8 2023, 10:42:25) [MSC v.1916 64 bit (AMD64)]
Commit hash:
Cloning Taming Transformers into F:\openai.wiki\stable-diffusion-webui-master\repositories\taming-transformers…
Cloning K-diffusion into F:\openai.wiki\stable-diffusion-webui-master\repositories\k-diffusion…
Cloning CodeFormer into F:\openai.wiki\stable-diffusion-webui-master\repositories\CodeFormer…
Cloning BLIP into F:\openai.wiki\stable-diffusion-webui-master\repositories\BLIP…
Traceback (most recent call last):
File “F:\openai.wiki\stable-diffusion-webui-master\launch.py”, line 380, in
prepare_environment()
File “F:\openai.wiki\stable-diffusion-webui-master\launch.py”, line 319, in prepare_environment
git_clone(blip_repo, repo_dir(‘BLIP’), “BLIP”, blip_commit_hash)
File “F:\openai.wiki\stable-diffusion-webui-master\launch.py”, line 167, in git_clone
run(f'”{git}” clone “{url}” “{dir}”‘, f”Cloning {name} into {dir}…”, f”Couldn’t clone {name}”)
File “F:\openai.wiki\stable-diffusion-webui-master\launch.py”, line 113, in run
raise RuntimeError(message)
RuntimeError: Couldn’t clone BLIP.
Command: “git” clone “https://github.com/salesforce/BLIP.git” “F:\openai.wiki\stable-diffusion-webui-master\repositories\BLIP”
Error code: 128
stdout:
stderr: Cloning into ‘F:\openai.wiki\stable-diffusion-webui-master\repositories\BLIP’…
fatal: unable to access ‘https://github.com/salesforce/BLIP.git/’: OpenSSL SSL_read: Connection was reset, errno 10054
这个错误信息意味着在尝试从 GitHub 克隆 BLIP 存储库时出现了一个连接错误。 这可能是由于网络问题或 GitHub 服务器的问题导致的。
taming-transformers
(D:\openai.wiki\stable-diffusion-webui\automatic) D:\openai.wiki\stable-diffusion-webui>webui-user.bat
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 8 2023, 10:42:25) [MSC v.1916 64 bit (AMD64)]
Commit hash: a9fed7c364061ae6efb37f797b6b522cb3cf7aa2
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 380, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 316, in prepare_environment
git_clone(taming_transformers_repo, repo_dir(‘taming-transformers’), “Taming Transformers”, taming_transformers_commit_hash)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 159, in git_clone
current_hash = run(f'”{git}” -C “{dir}” rev-parse HEAD’, None, f”Couldn’t determine {name}’s hash: {commithash}”).strip()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 113, in run
raise RuntimeError(message)
RuntimeError: Couldn’t determine Taming Transformers’s hash: 24268930bf1dce879235a7fddd0b2355b84d7ea6.
Command: “git” -C “D:\openai.wiki\stable-diffusion-webui\repositories\taming-transformers” rev-parse HEAD
Error code: 128
stdout: HEAD
stderr: fatal: ambiguous argument ‘HEAD’: unknown revision or path not in the working tree.
Use ‘–‘ to separate paths from revisions, like this:
‘git […] — […]’
这个错误显示无法克隆 taming-transformers
仓库并检索其特定的提交哈希值(commit hash),但是在尝试获取哈希值时出现了错误。错误信息中显示了一条有关“HEAD”的错误消息,表明 Git 无法识别此哈希值。建议删除 SD 安装目录.\openai.wiki\stable-diffusion-webui\repositories\
下的taming-transformers
文件夹。
PIP 超时
WARNING: Retrying (Retry(total=4, connect=None, read=None, redirect=None, status=None)) after connection broken by ‘ReadTimeoutError(“HTTPSConnectionPool(host=’files.pythonhosted.org’, port=443): Read timed out. (read timeout=15)”)’: /packages/bc/bf/58dbe1f382ecac2c0571c43b6e95028b14e159d67d75e49a00c26ef63d8f/lazy_loader-0.1-py3-none-any.whl
WARNING: Retrying (Retry(total=3, connect=None, read=None, redirect=None, status=None)) after connection broken by ‘ConnectTimeoutError(, ‘Connection to files.pythonhosted.org timed out. (connect timeout=15)’)’: /packages/bc/bf/58dbe1f382ecac2c0571c43b6e95028b14e159d67d75e49a00c26ef63d8f/lazy_loader-0.1-py3-none-any.whl
WARNING: Retrying (Retry(total=2, connect=None, read=None, redirect=None, status=None)) after connection broken by ‘ConnectTimeoutError(, ‘Connection to files.pythonhosted.org timed out. (connect timeout=15)’)’: /packages/bc/bf/58dbe1f382ecac2c0571c43b6e95028b14e159d67d75e49a00c26ef63d8f/lazy_loader-0.1-py3-none-any.whl
WARNING: Retrying (Retry(total=1, connect=None, read=None, redirect=None, status=None)) after connection broken by ‘ConnectTimeoutError(, ‘Connection to files.pythonhosted.org timed out. (connect timeout=15)’)’: /packages/bc/bf/58dbe1f382ecac2c0571c43b6e95028b14e159d67d75e49a00c26ef63d8f/lazy_loader-0.1-py3-none-any.whl
WARNING: Retrying (Retry(total=0, connect=None, read=None, redirect=None, status=None)) after connection broken by ‘ConnectTimeoutError(, ‘Connection to files.pythonhosted.org timed out. (connect timeout=15)’)’: /packages/bc/bf/58dbe1f382ecac2c0571c43b6e95028b14e159d67d75e49a00c26ef63d8f/lazy_loader-0.1-py3-none-any.whl
ERROR: Could not install packages due to an OSError: HTTPSConnectionPool(host=’files.pythonhosted.org’, port=443): Max retries exceeded with url: /packages/bc/bf/58dbe1f382ecac2c0571c43b6e95028b14e159d67d75e49a00c26ef63d8f/lazy_loader-0.1-py3-none-any.whl (Caused by ConnectTimeoutError(, ‘Connection to files.pythonhosted.org timed out. (connect timeout=15)’))
没有开启魔法上网,建议添加镜像源。
CondaHTTPError
CondaHTTPError: HTTP 000 CONNECTION FAILED for url
Elapsed: –
An HTTP error occurred when trying to retrieve this URL.
HTTP errors are often intermittent, and a simple retry will get you on your way.
‘https://conda.anaconda.org/pytorch/win-64’
这个错误提示表明在使用 Conda 时发生了 HTTP 错误,导致无法从指定的 URL 下载数据。可能的原因是网络连接问题或服务器故障。
如果你安装了 anaconda 请卸载,然后重新安装 miniconda。如果不是,请检查网络环境。
Python 问题
这类问题都是因为与 Python 相关的错误
PIP 升级
[notice] A new release of pip available: 22.3.1 -> 23.0.1
[notice] To update, run: D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe -m pip install –upgrade pip
PIP 需要升级,解决方法很简单,直接复制执行你的电脑所给出的代码即可。例如我们复制run:
后面的全部,然后在 CMD 中执行即可。
D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe -m pip install –upgrade pip
未找到 Python
Couldn’t launch python
exit code: 9009
当你看到“Couldn’t launch python exit code: 9009”这个错误时,这通常意味着系统无法找到 Python 解释器。这可能是因为 Python 未正确安装或未添加到系统环境变量中。
您可以检查是否已正确安装 Python 并将其添加到系统环境变量中。如果已经安装并添加到系统环境变量中,则可能需要重新启动终端或计算机以使更改生效。
Python 环境变量错误
Expecting value: line 1 column 1 (char 0)
可能是 Python 找不到系统变量了,你可以打开 CMD 窗口之后输入 Python,看看系统能否正常启动 Python,如果找不到 Python 的话,可以搜索一下 Python 环境变量修复的办法。
如果不想修复也能够正常使用,可以通过 miniconda 的终端内执行
conda activate 你的环境名称或路径
然后 CD 到你的 SD 根目录之后执行 webui-user.bat 即可。
未安装 PIP
(D:\openai.wiki\stable-diffusion-webui\automatic) C:\Users\86173>cd /d D:\openai.wiki\stable-diffusion-webui
(D:\openai.wiki\stable-diffusion-webui\automatic) D:\openai.wiki\stable-diffusion-webui>webui-user.bat
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash: 955df7751eef11bb7697e2d77f6b8a6226b21e13
Fetching updates for Taming Transformers…
Checking out commit for Taming Transformers with hash: 24268930bf1dce879235a7fddd0b2355b84d7ea6…
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 351, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 285, in prepare_environment
git_clone(taming_transformers_repo, repo_dir(‘taming-transformers’), “Taming Transformers”, taming_transformers_commit_hash)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 148, in git_clone
run(f'”{git}” -C “{dir}” checkout {commithash}’, f”Checking out commit for {name} with hash: {commithash}…”, f”Couldn’t checkout commit {commithash} for {name}”)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t checkout commit 24268930bf1dce879235a7fddd0b2355b84d7ea6 for Taming Transformers.
Command: “git” -C “D:\openai.wiki\stable-diffusion-webui\repositories\taming-transformers” checkout 24268930bf1dce879235a7fddd0b2355b84d7ea6
Error code: 128
stdout:
stderr: fatal: reference is not a tree: 24268930bf1dce879235a7fddd0b2355b84d7ea6
可能因为没有安装 pip 而导致的,建议执行python -m ensurepip
。
Python 版本不正确
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
==============================================================================================================
INCOMPATIBLE PYTHON VERSION
This program is tested with 3.10.6 Python, but you have 3.9.12.
If you encounter an error with “RuntimeError: Couldn’t install torch.” message,
or any other error regarding unsuccessful package (library) installation,
please downgrade (or upgrade) to the latest version of 3.10 Python
and delete current Python and “venv” folder in WebUI’s directory.
You can download 3.10 Python from here: https://www.python.org/downloads/release/python-3109/
Alternatively, use a binary release of WebUI: https://github.com/AUTOMATIC1111/stable-diffusion-webui/releases
Use –skip-python-version-check to suppress this warning.
==============================================================================================================
Python 3.9.12 (main, Apr 4 2022, 05:22:27) [MSC v.1916 64 bit (AMD64)]
Commit hash: 22bcc7be428c94e9408f589966c2040187245d81
Installing requirements for Web UI
Launching Web UI with arguments:
No module ‘xformers’. Proceeding without it.
Loading weights [a7529df023] from D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\final-pruned.ckpt
Creating model from config: D:\openai.wiki\stable-diffusion-webui\configs\v1-inference.yaml
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
Loading VAE weights found near the checkpoint: D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\final-pruned.vae.pt
loading stable diffusion model: OutOfMemoryError
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\webui.py”, line 139, in initialize
modules.sd_models.load_model()
File “D:\openai.wiki\stable-diffusion-webui\modules\sd_models.py”, line 449, in load_model
sd_model.to(shared.device)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\lightning_fabric\utilities\device_dtype_mixin.py”, line 54, in to
return super().to(*args, **kwargs)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 989, in to
return self._apply(convert)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 641, in _apply
module._apply(fn)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 641, in _apply
module._apply(fn)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 641, in _apply
module._apply(fn)
[Previous line repeated 2 more times]
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 664, in _apply
param_applied = fn(param)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 987, in convert
return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking)
torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 2.00 GiB total capacity; 1.66 GiB already allocated; 0 bytes free; 1.70 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Stable diffusion model failed to load, exiting
这个错误信息是由于 Web UI 需要 Python 3.10.6 版本,而你的 Python 版本是 3.9.12 导致的,所以建议你下载安装 3.10.6 版本的 Python。
或者删除当前 Python 和 Web UI 目录中的venv
文件夹,并升级到最新的 3.10 版本。
其它问题
未找到 launch.py 启动文件
(D:\openai.wiki\stable-diffusion-webui\automatic) D:\openai.wiki\stable-diffusion-webui>webui-user.bat
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
D:\openai.wiki\stable-diffusion-webui\automatic\python.exe: can’t open file ‘D:\\openai.wiki\\stable-diffusion-webui\\launch.py’: [Errno 2] No such file or directory
请按任意键继续。 . .
已经成功的进入到了 Conda 已激活的环境,但是在 D:\openai.wiki\stable-diffusion-webui 目录内没有 launch.py 这个文件,应该是 SD 没 Git 完整,或者目录错误。
GPU 显卡驱动较旧
[notice] A new release of pip available: 22.3.1 -> 23.0.1
[notice] To update, run: D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe -m pip install –upgrade pip
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 380, in
prepare_environment()
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 287, in prepare_environment
run_python(“import torch; assert torch.cuda.is_available(), ‘Torch is not able to use GPU; add –skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check'”)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 137, in run_python
return run(f'”{python}” -c “{code}”‘, desc, errdesc)
File “D:\openai.wiki\stable-diffusion-webui\launch.py”, line 113, in run
raise RuntimeError(message)
RuntimeError: Error running command.
Command: “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe” -c “import torch; assert torch.cuda.is_available(), ‘Torch is not able to use GPU; add –skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check'”
Error code: 1
stdout:
stderr: D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\cuda\__init__.py:88: UserWarning: CUDA initialization: The NVIDIA driver on your system is too old (found version 10020). Please update your GPU driver by downloading and installing a new version from the URL: http://www.nvidia.com/Download/index.aspx Alternatively, go to: https://pytorch.org to install a PyTorch version that has been compiled with your version of the CUDA driver. (Triggered internally at ..\c10\cuda\CUDAFunctions.cpp:109.)
return torch._C._cuda_getDeviceCount() > 0
Traceback (most recent call last):
File “”, line 1, in
AssertionError: Torch is not able to use GPU; add –skip-torch-cuda-test to COMMANDLINE_ARGS variable to disable this check
您的 GPU 驱动程序版本太旧,需要更新 GPU 驱动程序。更新 GPU 驱动程序可能会解决这些错误。
模型损坏
venv “G:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 8 2023, 10:42:25) [MSC v.1916 64 bit (AMD64)]
Commit hash: a9fed7c364061ae6efb37f797b6b522cb3cf7aa2
Installing requirements for Web UI
Launching Web UI with arguments:
No module ‘xformers’. Proceeding without it.
Loading weights [a586d5a51a] from G:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\v1-5-pruned-emaonly.safetensors
loading stable diffusion model: RuntimeError
Traceback (most recent call last):
File “G:\openai.wiki\stable-diffusion-webui\webui.py”, line 136, in initialize
modules.sd_models.load_model()
File “G:\openai.wiki\stable-diffusion-webui\modules\sd_models.py”, line 407, in load_model
state_dict = get_checkpoint_state_dict(checkpoint_info, timer)
File “G:\openai.wiki\stable-diffusion-webui\modules\sd_models.py”, line 262, in get_checkpoint_state_dict
res = read_state_dict(checkpoint_info.filename)
File “G:\openai.wiki\stable-diffusion-webui\modules\sd_models.py”, line 241, in read_state_dict
pl_sd = safetensors.torch.load_file(checkpoint_file, device=device)
File “G:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\safetensors\torch.py”, line 100, in load_file
result[k] = f.get_tensor(k)
RuntimeError: self.size(-1) must be divisible by 4 to view Byte as Float (different element sizes), but got 6545423
Stable diffusion model failed to load, exiting
请按任意键继续。 . .
根据错误信息,您的 Stable-diffusion 模型加载失败,原因是 self.size(-1)必须可以被 4 整除,但是实际上得到了 6545423,这是不可能被 4 整除的。这可能是由于模型参数文件已损坏或不完整导致的。您可以尝试重新下载模型文件并重新运行程序,以解决这个问题。
yaml 未找到
EnvironmentFileNotFound: ‘D:\openai.wiki\stable-diffusion-webui\environment-wsl2.yaml’ file not found
这个错误说明找不到指定路径下的environment-wsl2.yaml
文件,请确保路径和文件名正确,手动是否能够找到该文件。
无法切换模型
点击左上角的模型切换按钮后,一直加载中,无法正常使用。
解决办法:多点几次重新加载 UI 按钮,或者换一个浏览器。特殊情况下浏览器无法与 CMD 建立链接。
模型加载失败
(D:\openai.wiki\stable-diffusion-webui\automatic) D:\openai.wiki\stable-diffusion-webui>webui-user.bat
Creating venv in directory D:\openai.wiki\stable-diffusion-webui\venv using python “D:\openai.wiki\stable-diffusion-webui\automatic\python.exe”
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash:
Installing torch and torchvision
Looking in indexes: https://pypi.org/simple, https://download.pytorch.org/whl/cu117
Collecting torch==1.13.1+cu117
Downloading https://download.pytorch.org/whl/cu117/torch-1.13.1%2Bcu117-cp310-cp310-win_amd64.whl (2255.4 MB)
—————————————- 2.3/2.3 GB 1.0 MB/s eta 0:00:00
Collecting torchvision==0.14.1+cu117
Downloading https://download.pytorch.org/whl/cu117/torchvision-0.14.1%2Bcu117-cp310-cp310-win_amd64.whl (4.8 MB)
—————————————- 4.8/4.8 MB 8.1 MB/s eta 0:00:00
Collecting typing-extensions
Downloading typing_extensions-4.5.0-py3-none-any.whl (27 kB)
Collecting numpy
Downloading numpy-1.24.2-cp310-cp310-win_amd64.whl (14.8 MB)
—————————————- 14.8/14.8 MB 13.6 MB/s eta 0:00:00
Collecting requests
Downloading requests-2.28.2-py3-none-any.whl (62 kB)
—————————————- 62.8/62.8 kB 1.7 MB/s eta 0:00:00
Collecting pillow!=8.3.*,>=5.3.0
Downloading Pillow-9.4.0-cp310-cp310-win_amd64.whl (2.5 MB)
—————————————- 2.5/2.5 MB 19.7 MB/s eta 0:00:00
Collecting charset-normalizer=2
Downloading charset_normalizer-3.1.0-cp310-cp310-win_amd64.whl (97 kB)
—————————————- 97.1/97.1 kB 5.8 MB/s eta 0:00:00
Collecting idna=2.5
Downloading https://download.pytorch.org/whl/idna-3.4-py3-none-any.whl (61 kB)
—————————————- 61.5/61.5 kB 3.2 MB/s eta 0:00:00
Collecting urllib3=1.21.1
Downloading urllib3-1.26.15-py2.py3-none-any.whl (140 kB)
—————————————- 140.9/140.9 kB 8.7 MB/s eta 0:00:00
Collecting certifi>=2017.4.17
Downloading https://download.pytorch.org/whl/certifi-2022.12.7-py3-none-any.whl (155 kB)
—————————————- 155.3/155.3 kB 9.1 MB/s eta 0:00:00
Installing collected packages: urllib3, typing-extensions, pillow, numpy, idna, charset-normalizer, certifi, torch, requests, torchvision
Successfully installed certifi-2022.12.7 charset-normalizer-3.1.0 idna-3.4 numpy-1.24.2 pillow-9.4.0 requests-2.28.2 torch-1.13.1+cu117 torchvision-0.14.1+cu117 typing-extensions-4.5.0 urllib3-1.26.15
[notice] A new release of pip available: 22.3.1 -> 23.0.1
[notice] To update, run: D:\openai.wiki\stable-diffusion-webui\venv\Scripts\python.exe -m pip install –upgrade pip
Installing gfpgan
Installing clip
Installing open_clip
Cloning Stable Diffusion into D:\openai.wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai…
Cloning Taming Transformers into D:\openai.wiki\stable-diffusion-webui\repositories\taming-transformers…
Cloning K-diffusion into D:\openai.wiki\stable-diffusion-webui\repositories\k-diffusion…
Cloning CodeFormer into D:\openai.wiki\stable-diffusion-webui\repositories\CodeFormer…
Cloning BLIP into D:\openai.wiki\stable-diffusion-webui\repositories\BLIP…
Installing requirements for CodeFormer
Installing requirements for Web UI
Launching Web UI with arguments:
No module ‘xformers’. Proceeding without it.
Downloading: “https://huggingface.co/runwayml/stable-diffusion-v1-5/resolve/main/v1-5-pruned-emaonly.safetensors” to D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\v1-5-pruned-emaonly.safetensors
100%|█████████████████████████████████████████████████████████████████████████████| 3.97G/3.97G [05:34<00:00, 12.7MB/s]
Calculating sha256 for D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\v1-5-pruned-emaonly.safetensors: 6ce0161689b3853acaa03779ec93eafe75a02f4ced659bee03f50797806fa2fa
Loading weights [6ce0161689] from D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\v1-5-pruned-emaonly.safetensors
Creating model from config: D:\openai.wiki\stable-diffusion-webui\configs\v1-inference.yaml
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
Downloading (…)olve/main/vocab.json: 100%|███████████████████████████████████████████| 961k/961k [00:01<00:00, 862kB/s]
Downloading (…)olve/main/merges.txt: 100%|███████████████████████████████████████████| 525k/525k [00:00<00:00, 548kB/s]
Downloading (…)cial_tokens_map.json: 100%|█████████████████████████████████████████████| 389/389 [00:00<00:00, 282kB/s]
Downloading (…)okenizer_config.json: 100%|█████████████████████████████████████████████| 905/905 [00:00<00:00, 900kB/s]
Downloading (…)lve/main/config.json: 100%|████████████████████████████████████████| 4.52k/4.52k [00:00> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Stable diffusion model failed to load, exiting
这里出现了一个错误,似乎是 Stable Diffusion 模型加载失败导致的。建议检查一下网络连接,以及确保下载的模型文件完整无损。
你可以尝试重新运行这个命令,或者重新下载并安装 Stable Diffusion 模型文件。
目录不正确
(E:\ai\stable-diffusion-webui\automatic) E:\ai\stable-diffusion-webui>webui-user.bat
fatal: not a git repository (or any of the parent directories): .git
venv “E:\ai\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash:
Installing requirements for Web UI
Launching Web UI with arguments: –xformers
Loading weights [fe4efff1e1] from E:\ai\stable-diffusion-webui\models\Stable-diffusion\Model.ckpt
Creating model from config: E:\ai\stable-diffusion-webui\configs\v1-inference.yaml
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
Applying xformers cross attention optimization.
Textual inversion embeddings loaded(0):
Model loaded in 4.9s (load weights from disk: 1.7s, create model: 0.4s, apply weights to model: 0.5s, apply half(): 0.7s, move model to device: 0.7s, load textual inversion embeddings: 0.9s).
Traceback (most recent call last):
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\adapters.py”, line 412, in send
conn = self.get_connection(request.url, proxies)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\adapters.py”, line 305, in get_connection
proxy_url = parse_url(proxy)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\urllib3\util\url.py”, line 397, in parse_url
return six.raise_from(LocationParseError(source_url), None)
File “”, line 3, in raise_from
urllib3.exceptions.LocationParseError: Failed to parse: http://127.0.0.1:7890
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File “E:\ai\stable-diffusion-webui\launch.py”, line 352, in
start()
File “E:\ai\stable-diffusion-webui\launch.py”, line 347, in start
webui.webui()
File “E:\ai\stable-diffusion-webui\webui.py”, line 257, in webui
app, local_url, share_url = shared.demo.launch(
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\gradio\blocks.py”, line 1483, in launch
requests.get(f”{self.local_url}startup-events”)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\api.py”, line 76, in get
return request(‘get’, url, params=params, **kwargs)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\api.py”, line 61, in request
return session.request(method=method, url=url, **kwargs)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\sessions.py”, line 542, in request
resp = self.send(prep, **send_kwargs)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\sessions.py”, line 655, in send
r = adapter.send(request, **kwargs)
File “E:\ai\stable-diffusion-webui\venv\lib\site-packages\requests\adapters.py”, line 414, in send
raise InvalidURL(e, request=request)
requests.exceptions.InvalidURL: Failed to parse: http://127.0.0.1:7890
请按任意键继续。 . .
据错误消息,你的命令行工作目录可能不在 Git 仓库中,所以 Git 执行 git rev-parse HEAD 命令时出现了 fatal: not a git repository (or any of the parent directories): 。git 错误。你可以在命令行中切换到仓库目录,或者在启动脚本中指定 Git 仓库的路径。
加载时间较长
(base) D:\openai.wiki\stable-diffusion-webui>webui-user.bat
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 1 2023, 18:18:15) [MSC v.1916 64 bit (AMD64)]
Commit hash: 955df7751eef11bb7697e2d77f6b8a6226b21e13
Installing requirements for Web UI
Launching Web UI with arguments:
No module ‘xformers’. Proceeding without it.
Calculating sha256 for D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\final-pruned.ckpt: a7529df02340e5b4c3870c894c1ae84f22ea7b37fd0633e5bacfad9618228032
Loading weights [a7529df023] from D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\final-pruned.ckpt
Creating model from config: D:\openai.wiki\stable-diffusion-webui\configs\v1-inference.yaml
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
Applying cross attention optimization (Doggettx).
Textual inversion embeddings loaded(0):
Model loaded in 16.0s (calculate hash: 5.9s, load weights from disk: 3.8s, create model: 0.5s, apply weights to model: 2.0s, apply half(): 0.9s, move model to device: 1.0s, load textual inversion embeddings: 1.8s).
电脑配置差,多等一会。
端口被占用
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.10 | packaged by Anaconda, Inc. | (main, Mar 21 2023, 18:39:17) [MSC v.1916 64 bit (AMD64)]
Commit hash:
Installing requirements for Web UI
Launching Web UI with arguments:
No module ‘xformers’. Proceeding without it.
Loading weights [cc6cb27103] from D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\v1-5-pruned-emaonly.ckpt
Creating model from config: D:\openai.wiki\stable-diffusion-webui\configs\v1-inference.yaml
LatentDiffusion: Running in eps-prediction mode
DiffusionWrapper has 859.52 M params.
Applying cross attention optimization (Doggettx).
Textual inversion embeddings loaded(0):
Model loaded in 4.5s (load weights from disk: 1.3s, create model: 0.4s, apply weights to model: 0.5s, apply half(): 0.7s, move model to device: 0.7s, load textual inversion embeddings: 0.9s).
Running on local URL: http://127.0.0.1:7861
大佬好,在安装以后运行 webui-user.bat 的时候出现一下,安装过程中弹过 error 但是自己重试安装成功了,之后运行网址都没有反应,就是空白的,请您帮忙看看什么原因呢?
To create a public link, set share=
True
` in launch
()
`.
Startup time: 10.4s (import torch: 1.4s, import gradio: 1.0s, import ldm: 0.5s, other imports: 1.0s, setup codeformer: 0.2s, load scripts: 1.0s, load SD checkpoint: 4.8s, create ui: 0.4s, gradio launch: 0.1s).
正常来说 给出的应该是 7860,而你的是 http://127.0.0.1:7861,那代表你可能在 CMD 中开启了代理,或者其它应用占用了 7860 端口号,建议排查一下端口号的问题。
最便捷的解决办法就是关掉是这个 CMD,然后去 SD 的项目根目录内运行webui-user.bat
文件夹即可。
非 Windows 系统
Python 3.10.9 (main, Jan 11 2023, 15:21:40) [GCC 11.2.0]
Commit hash: 22bcc7be428c94e9408f589966c2040187245d81
Traceback (most recent call last):
File “/home/mark/Desktop/123/stable-diffusion-webui_23-03-10/launch.py”, line 355, in
prepare_environment()
File “/home/mark/Desktop/123/stable-diffusion-webui_23-03-10/launch.py”, line 288, in prepare_environment
git_clone(stable_diffusion_repo, repo_dir(‘stable-diffusion-stability-ai’), “Stable Diffusion”, stable_diffusion_commit_hash)
File “/home/mark/Desktop/123/stable-diffusion-webui_23-03-10/launch.py”, line 143, in git_clone
current_hash = run(f'”{git}” -C “{dir}” rev-parse HEAD’, None, f”Couldn’t determine {name}’s hash: {commithash}”).strip()
File “/home/mark/Desktop/123/stable-diffusion-webui_23-03-10/launch.py”, line 97, in run
raise RuntimeError(message)
RuntimeError: Couldn’t determine Stable Diffusion’s hash: cf1d67a6fd5ea1aa600c4df58e5b47da45f6bdbf.
Command: “git” -C “/home/mark/Desktop/123/stable-diffusion-webui_23-03-10/repositories/stable-diffusion-stability-ai” rev-parse HEAD
Error code: 129
stdout:
stderr: Unknown option: -C
usage: git [–version] [–help] [-c name=value]
[–exec-path[=]] [–html-path] [–man-path] [–info-path]
[-p|–paginate|–no-pager] [–no-replace-objects] [–bare]
[–git-dir=] [–work-tree=] [–namespace=]
[]
这个错误提示是 git 命令在运行时出了问题,它在执行“git -C”命令时出现了错误,因为“-C”是一个未知的选项。
可能是因为并非 Windows 系统而导致的,本教程仅针对 Windows,对 Liunx 暂不讨论。
语法错误
标准错误:错误:Error [WinError 2] 系统找不到指定的文件,执行命令 git version 错误:找不到命令’git’ – 你是否已安装并将其加入了环境变量中?
set-Location:找不到接受实际参数的“D:openai.wiki\stable-diffusion-webui”的位置参数所在的位置 行:1 字符:1
+cd、d D: \openai.wiki\stable-diffusion-webui
这个错误提示是在尝试改变当前目录(CD)到“D:\openai.wiki\stable-diffusion-webui”时出现的。错误提示表明系统找不到这个目录。
可能的原因之一是指定的路径不正确,或者该目录不存在。您可以确认一下指定的路径是否正确,并且该目录是否存在。如果不存在,请确保您输入的路径正确,并且该目录已创建。
另外一个问题就是cd 、d 路径
这句命令行,应该是cd /d 路径
,请注意中英文符号问题。
显存不足
(base) D:\openai.wiki\stable-diffusion-webui>webui-user.bat
venv “D:\openai.wiki\stable-diffusion-webui\venv\Scripts\Python.exe”
Python 3.10.9 | packaged by Anaconda, Inc. | (main, Mar 8 2023, 10:42:25) [MSC v.1916 64 bit (AMD64)]
Commit hash: a9fed7c364061ae6efb37f797b6b522cb3cf7aa2
Installing requirements for Web UI
Launching Web UI with arguments:
No module ‘xformers’. Proceeding without it.
Loading weights [ad2a33c361] from D:\openai.wiki\stable-diffusion-webui\models\Stable-diffusion\v2-1_768-ema-pruned.ckptCreating model from config: D:\openai.wiki\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\configs\stable-diffusion\v2-inference-v.yaml
LatentDiffusion: Running in v-prediction mode
DiffusionWrapper has 865.91 M params.
loading stable diffusion model: OutOfMemoryError
Traceback (most recent call last):
File “D:\openai.wiki\stable-diffusion-webui\webui.py”, line 136, in initialize
modules.sd_models.load_model()
File “D:\openai.wiki\stable-diffusion-webui\modules\sd_models.py”, line 441, in load_model
sd_model.to(shared.device)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\pytorch_lightning\core\mixins\device_dtype_mixin.py”, line 113, in to
return super().to(*args, **kwargs)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 989, in to
return self._apply(convert)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 641, in _apply
module._apply(fn)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 641, in _apply
module._apply(fn)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 641, in _apply
module._apply(fn)
[Previous line repeated 4 more times]
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 664, in _apply
param_applied = fn(param)
File “D:\openai.wiki\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py”, line 987, in convert
return t.to(device, dtype if t.is_floating_point() or t.is_complex() else None, non_blocking)
torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB (GPU 0; 2.00 GiB total capacity; 1.66 GiB already allocated; 0 bytes free; 1.71 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Stable diffusion model failed to load, exiting
请按任意键继续。 . .
根据错误信息显示,Stable Diffusion 模型加载时遇到了 CUDA 显存不足的错误,简单点来说就是显卡不行,没有足够的显存空间分配给该模型。
解决方法 1:
换个好一些的显卡,至少英伟达 2060 以上。
解决方法 2:
开启低显存模式,开启低显存模式的方法如下。
在 stable-diffusion-webui 文件夹下找到 webui-user.bat,用文本或代码编辑器打开该文件夹,可以看到如下内容。
@echo off
set PYTHON=
set GIT=
set VENV_DIR=
set COMMANDLINE_ARGS=
call webui.bat
我们只需要修改set COMMANDLINE_ARGS=
部分即可,该部分内容是启动参数。
低于 3G 显存
如果你的显卡显存不足 3G,可以在set COMMANDLINE_ARGS=
的后面添加参数–lowvram –always-batch-cond-uncond
之后保存即可,变更之后的文件内容如下。
@echo off
set PYTHON=
set GIT=
set VENV_DIR=
set COMMANDLINE_ARGS=–lowvram –always-batch-cond-uncond
call webui.bat
4G 显存
如果你的显卡只有 4G 显存,可以在set COMMANDLINE_ARGS=
的后面添加参数–precision full –no-half –lowvram –always-batch-cond-uncond
之后保存即可,变更之后的文件内容如下。
@echo off
set PYTHON=
set GIT=
set VENV_DIR=
set COMMANDLINE_ARGS=–precision full –no-half –lowvram –always-batch-cond-uncond
call webui.bat
低于 5G 显存
如果你的显卡显存不足 5G,可以在set COMMANDLINE_ARGS=
的后面添加参数–medvram
之后保存即可,变更之后的文件内容如下。
@echo off
set PYTHON=
set GIT=
set VENV_DIR=
set COMMANDLINE_ARGS=–medvram
call webui.bat
6G 显存
如果你的显卡只有 6G 显存,可以在set COMMANDLINE_ARGS=
的后面添加参数–precision full –no-half –medvram
之后保存即可,变更之后的文件内容如下。
@echo off
set PYTHON=
set GIT=
set VENV_DIR=
set COMMANDLINE_ARGS=–precision full –no-half –medvram
call webui.bat