Skip to content

Can svdquant comfyui be used with flux pulid? #69

Open
@pyoliu

Description

@pyoliu

model_type FLUX
[2024-12-25 14:49:43.765] [info] Loading partial weights from /home/ubuntu/.cache/huggingface/hub/models--mit-han-lab--svdquant-models/snapshots/0684a6e2693230e1aa0c54827a90e64bf082c95d/svdq-flux.1-dev-lora-realism.safetensors
[2024-12-25 14:49:43.765] [warning] Unable to pin memory: operation not supported
[2024-12-25 14:49:43.765] [info] Try MAP_PRIVATE
[2024-12-25 14:49:43.861] [info] Done.
[2024-12-25 14:49:43.871] [info] Set lora scale to 1 (skip 32 ranks)
clip missing: ['text_projection.weight']
Requested to load FluxClipModel_
loaded completely 9.5367431640625e+25 9319.23095703125 True
Requested to load Flux
loaded completely 9.5367431640625e+25 922.6993408203125 True
0%| | 0/25 [00:00<?, ?it/s]Passing txt_ids 3d torch.Tensor is deprecated.Please remove the batch dimension and pass it as a 2d torch Tensor
Passing img_ids 3d torch.Tensor is deprecated.Please remove the batch dimension and pass it as a 2d torch Tensor
0%| | 0/25 [00:00<?, ?it/s]
!!! Exception during processing !!!
Traceback (most recent call last):
File "/home/ubuntu/16T/lsm/ComfyUI/execution.py", line 328, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "/home/ubuntu/16T/lsm/ComfyUI/execution.py", line 203, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
File "/home/ubuntu/16T/lsm/ComfyUI/execution.py", line 174, in _map_node_over_list
process_inputs(input_dict, i)
File "/home/ubuntu/16T/lsm/ComfyUI/execution.py", line 163, in process_inputs
results.append(getattr(obj, func)(**inputs))
File "/home/ubuntu/16T/lsm/ComfyUI/comfy_extras/nodes_custom_sampler.py", line 633, in sample
samples = guider.sample(noise.generate_noise(latent), latent_image, sampler, sigmas, denoise_mask=noise_mask, callback=callback, disable_pbar=disable_pbar, seed=noise.seed)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 897, in sample
output = executor.execute(noise, latent_image, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/patcher_extension.py", line 110, in execute
return self.original(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 866, in outer_sample
output = self.inner_sample(noise, latent_image, device, sampler, sigmas, denoise_mask, callback, disable_pbar, seed)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 850, in inner_sample
samples = executor.execute(self, sigmas, extra_args, callback, noise, latent_image, denoise_mask, disable_pbar)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/patcher_extension.py", line 110, in execute
return self.original(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 707, in sample
samples = self.sampler_function(model_k, noise, sigmas, extra_args=extra_args, callback=k_callback, disable=disable_pbar, **self.extra_options)
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
return func(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/k_diffusion/sampling.py", line 1098, in sample_deis
denoised = model(x_cur, t_cur * s_in, **extra_args)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 379, in call
out = self.inner_model(x, sigma, model_options=model_options, seed=seed)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 832, in call
return self.predict_noise(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 835, in predict_noise
return sampling_function(self.inner_model, x, timestep, self.conds.get("negative", None), self.conds.get("positive", None), self.cfg, model_options=model_options, seed=seed)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 359, in sampling_function
out = calc_cond_batch(model, conds, x, timestep, model_options)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 195, in calc_cond_batch
return executor.execute(model, conds, x_in, timestep, model_options)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/patcher_extension.py", line 110, in execute
return self.original(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/samplers.py", line 308, in calc_cond_batch
output = model.apply_model(input_x, timestep
, **c).chunk(batch_chunks)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/model_base.py", line 130, in apply_model
return comfy.patcher_extension.WrapperExecutor.new_class_executor(
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/patcher_extension.py", line 110, in execute
return self.original(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/comfy/model_base.py", line 159, in _apply_model
model_output = self.diffusion_model(xc, t, context=context, control=control, transformer_options=transformer_options, **extra_conds).float()
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ubuntu/16T/lsm/ComfyUI/custom_nodes/svdquant/nodes.py", line 47, in forward
out = self.model(
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/diffusers/models/transformers/transformer_flux.py", line 522, in forward
encoder_hidden_states, hidden_states = block(
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1736, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1747, in _call_impl
return forward_call(*args, **kwargs)
File "/home/ubuntu/miniconda3/envs/flux/lib/python3.10/site-packages/nunchaku-0.0.2b0-py3.10-linux-x86_64.egg/nunchaku/models/transformer_flux.py", line 49, in forward
assert image_rotary_emb.shape[2] == batch_size * (txt_tokens + img_tokens)
AssertionError

Prompt executed in 44.93 seconds

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions