Skip to content

Unable to convert safetensors #6

@LeonSolisPedro

Description

@LeonSolisPedro

Hi I downloaded 3 models from civitai, and none of them work, I don't know what am I doing wrong

Specs: M1 macbook air
Sonoma 14.5
Xcode 15.4


Starting python converter
scikit-learn version 1.3.1 is not supported. Minimum required version: 0.17. Maximum required version: 1.1.2. Disabling scikit-learn conversion API.
Initializing StableDiffusionPipeline from /Users/travis/Downloads/ohmenToontastic_ohmenToontasticV2.safetensors..
Traceback (most recent call last):
  File "transformers/utils/hub.py", line 430, in cached_file
  File "huggingface_hub/utils/_validators.py", line 118, in _inner_fn
  File "huggingface_hub/file_download.py", line 1340, in hf_hub_download
huggingface_hub.utils._errors.LocalEntryNotFoundError: Cannot find the requested files in the disk cache and outgoing traffic has been disabled. To enable hf.co look-ups and downloads online, set 'local_files_only' to False.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 791, in convert_ldm_clip_checkpoint
File "transformers/models/clip/configuration_clip.py", line 141, in from_pretrained
  File "transformers/configuration_utils.py", line 622, in get_config_dict
  File "transformers/configuration_utils.py", line 677, in _get_config_dict
  File "transformers/utils/hub.py", line 470, in cached_file
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like openai/clip-vit-large-patch14 is not the path to a directory containing a file named config.json.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "guernikatools/torch2coreml.py", line 150, in main
File "diffusers/loaders.py", line 2822, in from_single_file
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 1633, in download_from_original_stable_diffusion_ckpt
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 793, in convert_ldm_clip_checkpoint
ValueError: With local_files_only set to True, you must first locally save the configuration in the following path: 'openai/clip-vit-large-patch14'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "transformers/utils/hub.py", line 430, in cached_file
File "huggingface_hub/utils/_validators.py", line 118, in _inner_fn
  File "huggingface_hub/file_download.py", line 1340, in hf_hub_download
huggingface_hub.utils._errors.LocalEntryNotFoundError: Cannot find the requested files in the disk cache and outgoing traffic has been disabled. To enable hf.co look-ups and downloads online, set 'local_files_only' to False.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 791, in convert_ldm_clip_checkpoint
  File "transformers/models/clip/configuration_clip.py", line 141, in from_pretrained
File "transformers/configuration_utils.py", line 622, in get_config_dict
File "transformers/configuration_utils.py", line 677, in _get_config_dict
  File "transformers/utils/hub.py", line 470, in cached_file
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like openai/clip-vit-large-patch14 is not the path to a directory containing a file named config.json.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "guernikatools/torch2coreml.py", line 153, in main
  File "diffusers/loaders.py", line 2822, in from_single_file
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 1633, in download_from_original_stable_diffusion_ckpt
File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 793, in convert_ldm_clip_checkpoint
ValueError: With local_files_only set to True, you must first locally save the configuration in the following path: 'openai/clip-vit-large-patch14'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "guernikatools/torch2coreml.py", line 500, in <module>
File "guernikatools/torch2coreml.py", line 155, in main
  File "diffusers/loaders.py", line 2822, in from_single_file
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 1442, in download_from_original_stable_diffusion_ckpt
  File "accelerate/utils/modeling.py", line 285, in set_module_tensor_to_device
ValueError: Trying to set a tensor of shape torch.Size([320, 4, 3, 3]) in "weight" (which has shape torch.Size([320, 9, 3, 3])), this look incorrect.
[5541] Failed to execute script 'torch2coreml' due to unhandled exception: Trying to set a tensor of shape torch.Size([320, 4, 3, 3]) in "weight" (which has shape torch.Size([320, 9, 3, 3])), this look incorrect.
[5541] Traceback:
Traceback (most recent call last):
  File "transformers/utils/hub.py", line 430, in cached_file
    resolved_file = hf_hub_download(
                    ^^^^^^^^^^^^^^^^
  File "huggingface_hub/utils/_validators.py", line 118, in _inner_fn
  File "huggingface_hub/file_download.py", line 1340, in hf_hub_download
huggingface_hub.utils._errors.LocalEntryNotFoundError: Cannot find the requested files in the disk cache and outgoing traffic has been disabled. To enable hf.co look-ups and downloads online, set 'local_files_only' to False.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 791, in convert_ldm_clip_checkpoint
    config = CLIPTextConfig.from_pretrained(config_name, local_files_only=local_files_only)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "transformers/models/clip/configuration_clip.py", line 141, in from_pretrained
    config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "transformers/configuration_utils.py", line 622, in get_config_dict
    config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "transformers/configuration_utils.py", line 677, in _get_config_dict
    resolved_config_file = cached_file(
                           ^^^^^^^^^^^^
  File "transformers/utils/hub.py", line 470, in cached_file
    raise EnvironmentError(
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like openai/clip-vit-large-patch14 is not the path to a directory containing a file named config.json.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "guernikatools/torch2coreml.py", line 150, in main
  File "diffusers/loaders.py", line 2822, in from_single_file
    pipe = download_from_original_stable_diffusion_ckpt(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 1633, in download_from_original_stable_diffusion_ckpt
    text_model = convert_ldm_clip_checkpoint(
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 793, in convert_ldm_clip_checkpoint
    raise ValueError(
ValueError: With local_files_only set to True, you must first locally save the configuration in the following path: 'openai/clip-vit-large-patch14'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "transformers/utils/hub.py", line 430, in cached_file
    resolved_file = hf_hub_download(
                    ^^^^^^^^^^^^^^^^
  File "huggingface_hub/utils/_validators.py", line 118, in _inner_fn
  File "huggingface_hub/file_download.py", line 1340, in hf_hub_download
huggingface_hub.utils._errors.LocalEntryNotFoundError: Cannot find the requested files in the disk cache and outgoing traffic has been disabled. To enable hf.co look-ups and downloads online, set 'local_files_only' to False.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 791, in convert_ldm_clip_checkpoint
    config = CLIPTextConfig.from_pretrained(config_name, local_files_only=local_files_only)
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "transformers/models/clip/configuration_clip.py", line 141, in from_pretrained
    config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs)
                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "transformers/configuration_utils.py", line 622, in get_config_dict
    config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs)
                          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "transformers/configuration_utils.py", line 677, in _get_config_dict
    resolved_config_file = cached_file(
                           ^^^^^^^^^^^^
  File "transformers/utils/hub.py", line 470, in cached_file
    raise EnvironmentError(
OSError: We couldn't connect to 'https://huggingface.co' to load this file, couldn't find it in the cached files and it looks like openai/clip-vit-large-patch14 is not the path to a directory containing a file named config.json.
Checkout your internet connection or see how to run the library in offline mode at 'https://huggingface.co/docs/transformers/installation#offline-mode'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "guernikatools/torch2coreml.py", line 153, in main
  File "diffusers/loaders.py", line 2822, in from_single_file
    pipe = download_from_original_stable_diffusion_ckpt(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 1633, in download_from_original_stable_diffusion_ckpt
    text_model = convert_ldm_clip_checkpoint(
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 793, in convert_ldm_clip_checkpoint
    raise ValueError(
ValueError: With local_files_only set to True, you must first locally save the configuration in the following path: 'openai/clip-vit-large-patch14'.

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
  File "guernikatools/torch2coreml.py", line 500, in <module>
  File "guernikatools/torch2coreml.py", line 155, in main
  File "diffusers/loaders.py", line 2822, in from_single_file
    pipe = download_from_original_stable_diffusion_ckpt(
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "diffusers/pipelines/stable_diffusion/convert_from_ckpt.py", line 1442, in download_from_original_stable_diffusion_ckpt
    set_module_tensor_to_device(unet, param_name, "cpu", value=param)
  File "accelerate/utils/modeling.py", line 285, in set_module_tensor_to_device
ValueError: Trying to set a tensor of shape torch.Size([320, 4, 3, 3]) in "weight" (which has shape torch.Size([320, 9, 3, 3])), this look incorrect.


Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions