diff --git a/docs/contributing/MODEL_MANAGER.md b/docs/contributing/MODEL_MANAGER.md index e8b6f13e316..627a3e42ec4 100644 --- a/docs/contributing/MODEL_MANAGER.md +++ b/docs/contributing/MODEL_MANAGER.md @@ -265,7 +265,7 @@ If the key is unrecognized, this call raises an #### exists(key) -> AnyModelConfig -Returns True if a model with the given key exists in the databsae. +Returns True if a model with the given key exists in the database. #### search_by_path(path) -> AnyModelConfig @@ -718,7 +718,7 @@ When downloading remote models is implemented, additional configuration information, such as list of trigger terms, will be retrieved from the HuggingFace and Civitai model repositories. -The probed values can be overriden by providing a dictionary in the +The probed values can be overridden by providing a dictionary in the optional `config` argument passed to `import_model()`. You may provide overriding values for any of the model's configuration attributes. Here is an example of setting the @@ -841,7 +841,7 @@ variable. #### installer.start(invoker) -The `start` method is called by the API intialization routines when +The `start` method is called by the API initialization routines when the API starts up. Its effect is to call `sync_to_config()` to synchronize the model record store database with what's currently on disk. diff --git a/docs/contributing/contributors.md b/docs/contributing/contributors.md index 5df11933cba..d5691a9064f 100644 --- a/docs/contributing/contributors.md +++ b/docs/contributing/contributors.md @@ -16,7 +16,7 @@ We thank [all contributors](https://github.com/invoke-ai/InvokeAI/graphs/contrib - @psychedelicious (Spencer Mabrito) - Web Team Leader - @joshistoast (Josh Corbett) - Web Development - @cheerio (Mary Rogers) - Lead Engineer & Web App Development -- @ebr (Eugene Brodsky) - Cloud/DevOps/Sofware engineer; your friendly neighbourhood cluster-autoscaler +- @ebr (Eugene Brodsky) - Cloud/DevOps/Software engineer; your friendly neighbourhood cluster-autoscaler - @sunija - Standalone version - @brandon (Brandon Rising) - Platform, Infrastructure, Backend Systems - @ryanjdick (Ryan Dick) - Machine Learning & Training diff --git a/invokeai/backend/quantization/scripts/load_flux_model_bnb_llm_int8.py b/invokeai/backend/quantization/scripts/load_flux_model_bnb_llm_int8.py index 804336e0007..045ebbbf2c4 100644 --- a/invokeai/backend/quantization/scripts/load_flux_model_bnb_llm_int8.py +++ b/invokeai/backend/quantization/scripts/load_flux_model_bnb_llm_int8.py @@ -20,7 +20,7 @@ def main(): "/data/invokeai/models/.download_cache/https__huggingface.co_black-forest-labs_flux.1-schnell_resolve_main_flux1-schnell.safetensors/flux1-schnell.safetensors" ) - with log_time("Intialize FLUX transformer on meta device"): + with log_time("Initialize FLUX transformer on meta device"): # TODO(ryand): Determine if this is a schnell model or a dev model and load the appropriate config. p = params["flux-schnell"] diff --git a/invokeai/backend/quantization/scripts/load_flux_model_bnb_nf4.py b/invokeai/backend/quantization/scripts/load_flux_model_bnb_nf4.py index f1621dbc6dd..c8802b9e49e 100644 --- a/invokeai/backend/quantization/scripts/load_flux_model_bnb_nf4.py +++ b/invokeai/backend/quantization/scripts/load_flux_model_bnb_nf4.py @@ -33,7 +33,7 @@ def main(): ) # inference_dtype = torch.bfloat16 - with log_time("Intialize FLUX transformer on meta device"): + with log_time("Initialize FLUX transformer on meta device"): # TODO(ryand): Determine if this is a schnell model or a dev model and load the appropriate config. p = params["flux-schnell"] diff --git a/invokeai/backend/quantization/scripts/quantize_t5_xxl_bnb_llm_int8.py b/invokeai/backend/quantization/scripts/quantize_t5_xxl_bnb_llm_int8.py index fc681e8fc57..2e610404cdc 100644 --- a/invokeai/backend/quantization/scripts/quantize_t5_xxl_bnb_llm_int8.py +++ b/invokeai/backend/quantization/scripts/quantize_t5_xxl_bnb_llm_int8.py @@ -27,7 +27,7 @@ def main(): """ model_path = Path("/data/misc/text_encoder_2") - with log_time("Intialize T5 on meta device"): + with log_time("Initialize T5 on meta device"): model_config = AutoConfig.from_pretrained(model_path) with accelerate.init_empty_weights(): model = AutoModelForTextEncoding.from_config(model_config) diff --git a/scripts/get_external_contributions.py b/scripts/get_external_contributions.py index c799b6f75a1..3d798e91342 100644 --- a/scripts/get_external_contributions.py +++ b/scripts/get_external_contributions.py @@ -46,7 +46,7 @@ def fetch_commits_between_tags( commit_info: list[CommitInfo] = [] headers = {"Authorization": f"token {token}"} if token else None - # Get the total number of pages w/ an intial request - a bit hacky but it works... + # Get the total number of pages w/ an initial request - a bit hacky but it works... response = requests.get( f"https://api.github.com/repos/{org_name}/{repo_name}/compare/{from_ref}...{to_ref}?page=1&per_page=100", headers=headers,