diff --git a/kohya_gui/class_anima.py b/kohya_gui/class_anima.py index 5fafe6a..9f710a9 100644 --- a/kohya_gui/class_anima.py +++ b/kohya_gui/class_anima.py @@ -180,12 +180,6 @@ class animaTraining: info="Offload activations to CPU RAM using async non-blocking transfers. Faster than cpu_offload_checkpointing. Cannot combine with blocks_to_swap.", interactive=True, ) - self.anima_torch_compile = gr.Checkbox( - label="torch.compile", - value=self.config.get("anima.anima_torch_compile", False), - info="JIT-compile DiT with torch.compile (inductor backend). Can speed up training ~10-30%. Incompatible with Unsloth Offload Checkpointing.", - interactive=True, - ) self.anima_disable_mmap_load_safetensors = gr.Checkbox( label="Disable mmap Load", value=self.config.get("anima.anima_disable_mmap_load_safetensors", False), diff --git a/kohya_gui/lora_gui.py b/kohya_gui/lora_gui.py index 46fa066..0da10ea 100644 --- a/kohya_gui/lora_gui.py +++ b/kohya_gui/lora_gui.py @@ -336,7 +336,6 @@ def save_configuration( anima_cache_text_encoder_outputs_to_disk, anima_blocks_to_swap, anima_unsloth_offload_checkpointing, - anima_torch_compile, anima_disable_mmap_load_safetensors, anima_vae_chunk_size, anima_vae_disable_cache, @@ -643,7 +642,6 @@ def open_configuration( anima_cache_text_encoder_outputs_to_disk, anima_blocks_to_swap, anima_unsloth_offload_checkpointing, - anima_torch_compile, anima_disable_mmap_load_safetensors, anima_vae_chunk_size, anima_vae_disable_cache, @@ -1041,7 +1039,6 @@ def train_model( anima_cache_text_encoder_outputs_to_disk, anima_blocks_to_swap, anima_unsloth_offload_checkpointing, - anima_torch_compile, anima_disable_mmap_load_safetensors, anima_vae_chunk_size, anima_vae_disable_cache, @@ -1651,7 +1648,7 @@ def train_model( "color_aug": color_aug, "dataset_config": dataset_config, "debiased_estimation_loss": debiased_estimation_loss, - "dynamo_backend": "inductor" if (anima_checkbox and anima_torch_compile and dynamo_backend == "no") else dynamo_backend, + "dynamo_backend": dynamo_backend, "dim_from_weights": dim_from_weights, "disable_mmap_load_safetensors": disable_mmap_load_safetensors_value, "enable_bucket": enable_bucket, @@ -1873,7 +1870,6 @@ def train_model( "t5_max_token_length": int(anima_t5_max_token_length) if anima_checkbox else None, "split_attn": anima_split_attn if anima_checkbox else None, "unsloth_offload_checkpointing": anima_unsloth_offload_checkpointing if anima_checkbox else None, - "torch_compile": True if (anima_checkbox and anima_torch_compile) else None, "vae_chunk_size": int(anima_vae_chunk_size) if anima_checkbox and anima_vae_chunk_size else None, "vae_disable_cache": anima_vae_disable_cache if anima_checkbox else None, } @@ -3170,7 +3166,6 @@ def lora_tab( anima_training.anima_cache_text_encoder_outputs_to_disk, anima_training.anima_blocks_to_swap, anima_training.anima_unsloth_offload_checkpointing, - anima_training.anima_torch_compile, anima_training.anima_disable_mmap_load_safetensors, anima_training.vae_chunk_size, anima_training.vae_disable_cache,