forked from AUTOMATIC1111/stable-diffusion-webui
-
Notifications
You must be signed in to change notification settings - Fork 236
Open
Description
Checklist
- The issue exists after disabling all extensions
- The issue exists on a clean installation of webui
- The issue is caused by an extension, but I believe it is caused by a bug in the webui
- The issue exists in the current version of the webui
- The issue has not been reported before recently
- The issue has been reported before but has not been fixed yet
What happened?
I tried generating an image with the same setting as another image that i had already generated and it ran out of memory. Even if i remove the Hires, fix it still runs out of memory when generating a 832X1216 image. It doesn't seems to happen when its 512X512.
Steps to reproduce the problem
- Generate image higher than 512X512 (i tried with 832X1216 as its the same resolution as a previously generated image and 1024X1024)
- Generation stops at the last step where it stays for a while until it runs out of memory.
What should have happened?
It should generate images bigger than 512X512 without running out of memory.
What browsers do you use to access the UI ?
Mozilla Firefox
Sysinfo
Console logs
From https://github.com/lshqqytiger/stable-diffusion-webui-directml
* branch HEAD -> FETCH_HEAD
Already up to date.
venv "E:\stable-diffusion-webui-amdgpu\venv\Scripts\Python.exe"
WARNING: ZLUDA works best with SD.Next. Please consider migrating to SD.Next.
Python 3.10.6 (tags/v3.10.6:9c7b4bd, Aug 1 2022, 21:53:49) [MSC v.1932 64 bit (AMD64)]
Version: v1.10.1-amd-43-g1ad6edf1
Commit hash: 1ad6edf170c2c4307e0d2400f760a149e621dc38
ROCm: agents=['gfx1100', 'gfx1036']
ROCm: version=6.2, using agent gfx1100
ZLUDA support: experimental
ZLUDA load: path='E:\stable-diffusion-webui-amdgpu\.zluda' nightly=False
Skipping onnxruntime installation.
W0903 13:31:26.367725 17452 venv\Lib\site-packages\torch\distributed\elastic\multiprocessing\redirects.py:29] NOTE: Redirects are currently not supported in Windows or MacOs.
E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\timm\models\layers\__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers
warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning)
no module 'xformers'. Processing without...
no module 'xformers'. Processing without...
No module 'xformers'. Proceeding without it.
E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\pytorch_lightning\utilities\distributed.py:258: LightningDeprecationWarning: `pytorch_lightning.utilities.distributed.rank_zero_only` has been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it from `pytorch_lightning.utilities` instead.
rank_zero_deprecation(
Launching Web UI with arguments: --use-zluda --upcast-sampling --skip-ort
Loading weights [bdb59bac77] from E:\stable-diffusion-webui-amdgpu\models\Stable-diffusion\waiNSFWIllustrious_v140.safetensors
Creating model from config: E:\stable-diffusion-webui-amdgpu\repositories\generative-models\configs\inference\sd_xl_base.yaml
Running on local URL: http://127.0.0.1:7860
To create a public link, set `share=True` in `launch()`.
Startup time: 22.2s (prepare environment: 33.3s, initialize shared: 0.9s, other imports: 0.4s, load scripts: 0.7s, create ui: 0.6s, gradio launch: 0.4s).
Applying attention optimization: Doggettx... done.
Model loaded in 20.6s (load weights from disk: 0.6s, create model: 0.8s, apply weights to model: 17.8s, load textual inversion embeddings: 0.2s, calculate empty prompt: 1.0s).
100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:09<00:00, 2.16it/s]
Compilation is in progress. Please wait...████████████████▋ | 20/35 [00:10<00:05, 2.57it/s]
100%|██████████████████████████████████████████████████████████████████████████████████| 15/15 [00:09<00:00, 1.67it/s]
*** Error completing request███████████████████████████████████████████████████████████| 35/35 [00:21<00:00, 1.57it/s]
*** Arguments: ('task(f5g1rvnfpa4juy7)', <gradio.routes.Request object at 0x0000028636D87AF0>, '8k,best quality,masterpiece,(ultra-detailed),(high detailed skin),((artist:reoen),(artist:fuzichoco),(artist:atdan),(artist:torino_aqua),year 2024:0.8),[artist:wlop],[artist:ningen_mame],artist:ciloranko,[[artist:rhasta]],artist:tidsean,colorful,<lora:c_chasca (genshin impact)_nbe11_xl:1>, chasca \\(genshin impact\\),pointy ears,red hair,hat,jewelry,blue eyes,braid,earrings,black headwear,red scarf, open mouth,smirk,naughty_face, detailed dreamy background, outdoors', 'anilingus,rimjob,huge ass,thick thighs,2boys,mosaic censoring,monochrome,ai-generated,lowres,(worst quality, bad quality:1.2),bad anatomy,sketch,jpeg artifacts,signature,watermark,aged down,chibi,censored,simple background,nose,nostrils,colorized,text,english text,shiny shirt,mismatched pupils,skinny,futanari,ugly,distorted,blurry,deformed,twisted,watermark,logo,text,username', [], 1, 1, 7, 1216, 832, True, 0.3, 1.2, 'Latent', 15, 0, 0, 'Use same checkpoint', 'Use same sampler', 'Use same scheduler', '', '', [], 0, 20, 'Euler a', 'Automatic', False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
Traceback (most recent call last):
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 74, in f
res = list(func(*args, **kwargs))
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 53, in f
res = func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 37, in f
res = func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\modules\txt2img.py", line 109, in txt2img
processed = processing.process_images(p)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 849, in process_images
res = process_images_inner(p)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 1083, in process_images_inner
samples_ddim = p.sample(conditioning=p.c, unconditional_conditioning=p.uc, seeds=p.seeds, subseeds=p.subseeds, subseed_strength=p.subseed_strength, prompts=p.prompts)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 1457, in sample
return self.sample_hr_pass(samples, decoded_samples, seeds, subseeds, subseed_strength, prompts)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 1556, in sample_hr_pass
decoded_samples = decode_latent_batch(self.sd_model, samples, target_device=devices.cpu, check_for_nans=True)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 633, in decode_latent_batch
sample = decode_first_stage(model, batch[i:i + 1])[0]
File "E:\stable-diffusion-webui-amdgpu\modules\sd_samplers_common.py", line 76, in decode_first_stage
return samples_to_images_tensor(x, approx_index, model)
File "E:\stable-diffusion-webui-amdgpu\modules\sd_samplers_common.py", line 58, in samples_to_images_tensor
x_sample = model.decode_first_stage(sample.to(model.first_stage_model.dtype))
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\diffusion.py", line 121, in decode_first_stage
out = self.first_stage_model.decode(z)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\autoencoder.py", line 315, in decode
dec = self.decoder(z, **decoder_kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\diffusionmodules\model.py", line 728, in forward
h = self.up[i_level].block[i_block](h, temb, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\diffusionmodules\model.py", line 132, in forward
h = self.conv1(h)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\extensions-builtin\Lora\networks.py", line 599, in network_Conv2d_forward
return originals.Conv2d_forward(self, input)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 554, in forward
return self._conv_forward(input, self.weight, self.bias)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 549, in _conv_forward
return F.conv2d(
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 3.10 GiB. GPU 0 has a total capacity of 19.98 GiB of which 8.50 GiB is free. Of the allocated memory 7.45 GiB is allocated by PyTorch, and 3.63 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
---
100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:07<00:00, 2.69it/s]
*** Error completing request 2.06it/s]
*** Arguments: ('task(zl9veevrl90oy5p)', <gradio.routes.Request object at 0x000002866C3C2A10>, '8k,best quality,masterpiece,(ultra-detailed),(high detailed skin),((artist:reoen),(artist:fuzichoco),(artist:atdan),(artist:torino_aqua),year 2024:0.8),[artist:wlop],[artist:ningen_mame],artist:ciloranko,[[artist:rhasta]],artist:tidsean,colorful,<lora:c_chasca (genshin impact)_nbe11_xl:1>, chasca \\(genshin impact\\),pointy ears,red hair,hat,jewelry,blue eyes,braid,earrings,black headwear,red scarf, open mouth,smirk,naughty_face, detailed dreamy background, outdoors', 'anilingus,rimjob,huge ass,thick thighs,2boys,mosaic censoring,monochrome,ai-generated,lowres,(worst quality, bad quality:1.2),bad anatomy,sketch,jpeg artifacts,signature,watermark,aged down,chibi,censored,simple background,nose,nostrils,colorized,text,english text,shiny shirt,mismatched pupils,skinny,futanari,ugly,distorted,blurry,deformed,twisted,watermark,logo,text,username', [], 1, 1, 7, 1216, 832, False, 0.3, 1.2, 'Latent', 15, 0, 0, 'Use same checkpoint', 'Use same sampler', 'Use same scheduler', '', '', [], 0, 20, 'Euler a', 'Automatic', False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
Traceback (most recent call last):
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 74, in f
res = list(func(*args, **kwargs))
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 53, in f
res = func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 37, in f
res = func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\modules\txt2img.py", line 109, in txt2img
processed = processing.process_images(p)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 849, in process_images
res = process_images_inner(p)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 1097, in process_images_inner
x_samples_ddim = decode_latent_batch(p.sd_model, samples_ddim, target_device=devices.cpu, check_for_nans=True)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 633, in decode_latent_batch
sample = decode_first_stage(model, batch[i:i + 1])[0]
File "E:\stable-diffusion-webui-amdgpu\modules\sd_samplers_common.py", line 76, in decode_first_stage
return samples_to_images_tensor(x, approx_index, model)
File "E:\stable-diffusion-webui-amdgpu\modules\sd_samplers_common.py", line 58, in samples_to_images_tensor
x_sample = model.decode_first_stage(sample.to(model.first_stage_model.dtype))
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\diffusion.py", line 121, in decode_first_stage
out = self.first_stage_model.decode(z)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\autoencoder.py", line 315, in decode
dec = self.decoder(z, **decoder_kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\diffusionmodules\model.py", line 732, in forward
h = self.up[i_level].upsample(h)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\diffusionmodules\model.py", line 67, in forward
x = self.conv(x)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\extensions-builtin\Lora\networks.py", line 599, in network_Conv2d_forward
return originals.Conv2d_forward(self, input)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 554, in forward
return self._conv_forward(input, self.weight, self.bias)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 549, in _conv_forward
return F.conv2d(
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.34 GiB. GPU 0 has a total capacity of 19.98 GiB of which 10.99 GiB is free. Of the allocated memory 7.68 GiB is allocated by PyTorch, and 927.77 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
---
100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:01<00:00, 10.59it/s]
Total progress: 75it [06:37, 5.30s/it]
100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:07<00:00, 2.58it/s]
*** Error completing request███████████████████████████████████████████████████████████| 20/20 [00:07<00:00, 2.48it/s]
*** Arguments: ('task(2y7ef7e0o709u5z)', <gradio.routes.Request object at 0x000002866C3C65C0>, '8k,best quality,masterpiece,(ultra-detailed),(high detailed skin),((artist:reoen),(artist:fuzichoco),(artist:atdan),(artist:torino_aqua),year 2024:0.8),[artist:wlop],[artist:ningen_mame],artist:ciloranko,[[artist:rhasta]],artist:tidsean,colorful,<lora:c_chasca (genshin impact)_nbe11_xl:1>, chasca \\(genshin impact\\),pointy ears,red hair,hat,jewelry,blue eyes,braid,earrings,black headwear,red scarf, open mouth,smirk,naughty_face, detailed dreamy background, outdoors', 'anilingus,rimjob,huge ass,thick thighs,2boys,mosaic censoring,monochrome,ai-generated,lowres,(worst quality, bad quality:1.2),bad anatomy,sketch,jpeg artifacts,signature,watermark,aged down,chibi,censored,simple background,nose,nostrils,colorized,text,english text,shiny shirt,mismatched pupils,skinny,futanari,ugly,distorted,blurry,deformed,twisted,watermark,logo,text,username', [], 1, 1, 7, 1024, 1024, False, 0.3, 1.2, 'Latent', 15, 0, 0, 'Use same checkpoint', 'Use same sampler', 'Use same scheduler', '', '', [], 0, 20, 'Euler a', 'Automatic', False, '', 0.8, -1, False, -1, 0, 0, 0, False, False, 'positive', 'comma', 0, False, False, 'start', '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, False, False, False, 0, False) {}
Traceback (most recent call last):
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 74, in f
res = list(func(*args, **kwargs))
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 53, in f
res = func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\modules\call_queue.py", line 37, in f
res = func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\modules\txt2img.py", line 109, in txt2img
processed = processing.process_images(p)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 849, in process_images
res = process_images_inner(p)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 1097, in process_images_inner
x_samples_ddim = decode_latent_batch(p.sd_model, samples_ddim, target_device=devices.cpu, check_for_nans=True)
File "E:\stable-diffusion-webui-amdgpu\modules\processing.py", line 633, in decode_latent_batch
sample = decode_first_stage(model, batch[i:i + 1])[0]
File "E:\stable-diffusion-webui-amdgpu\modules\sd_samplers_common.py", line 76, in decode_first_stage
return samples_to_images_tensor(x, approx_index, model)
File "E:\stable-diffusion-webui-amdgpu\modules\sd_samplers_common.py", line 58, in samples_to_images_tensor
x_sample = model.decode_first_stage(sample.to(model.first_stage_model.dtype))
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\diffusion.py", line 121, in decode_first_stage
out = self.first_stage_model.decode(z)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\models\autoencoder.py", line 315, in decode
dec = self.decoder(z, **decoder_kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\diffusionmodules\model.py", line 732, in forward
h = self.up[i_level].upsample(h)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\repositories\generative-models\sgm\modules\diffusionmodules\model.py", line 67, in forward
x = self.conv(x)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl
return forward_call(*args, **kwargs)
File "E:\stable-diffusion-webui-amdgpu\extensions-builtin\Lora\networks.py", line 599, in network_Conv2d_forward
return originals.Conv2d_forward(self, input)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 554, in forward
return self._conv_forward(input, self.weight, self.bias)
File "E:\stable-diffusion-webui-amdgpu\venv\lib\site-packages\torch\nn\modules\conv.py", line 549, in _conv_forward
return F.conv2d(
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.50 GiB. GPU 0 has a total capacity of 19.98 GiB of which 10.87 GiB is free. Of the allocated memory 7.72 GiB is allocated by PyTorch, and 1013.90 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
---
Refresh index: 100% (371/371), done.Additional information
I'm running a 7900XT with 20GB of vram witch i think should be plenty.
Seems to be happening after the last webui update.
Metadata
Metadata
Assignees
Labels
No labels