-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[LoRA] make set_adapters()
robust on silent failures.
#9618
base: main
Are you sure you want to change the base?
Changes from all commits
2c250e7
45c15d3
ce4fcbf
a6f7c90
473bbad
2714043
47ef802
e75a283
5338275
1989c71
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -156,6 +156,10 @@ def test_with_alpha_in_state_dict(self): | |
) | ||
self.assertFalse(np.allclose(images_lora_with_alpha, images_lora, atol=1e-3, rtol=1e-3)) | ||
|
||
@unittest.skip("Not supported in SD3.") | ||
def test_simple_inference_with_text_denoiser_block_scale(self): | ||
pass | ||
|
||
@unittest.skip("Not supported in Flux.") | ||
def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(self): | ||
pass | ||
|
@@ -164,6 +168,10 @@ def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(se | |
def test_modify_padding_mode(self): | ||
pass | ||
|
||
@unittest.skip("Not supported in Flux.") | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. |
||
def test_simple_inference_with_text_denoiser_multi_adapter_block_lora(self): | ||
pass | ||
|
||
|
||
@slow | ||
@nightly | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -29,6 +29,7 @@ | |
from diffusers.utils import load_image | ||
from diffusers.utils.import_utils import is_accelerate_available | ||
from diffusers.utils.testing_utils import ( | ||
is_flaky, | ||
is_peft_available, | ||
numpy_cosine_similarity_distance, | ||
require_peft_backend, | ||
|
@@ -129,6 +130,10 @@ def test_simple_inference_with_text_denoiser_block_scale_for_all_dict_options(se | |
def test_modify_padding_mode(self): | ||
pass | ||
|
||
@is_flaky | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Same as #9618 (comment). |
||
def test_multiple_wrong_adapter_name_raises_error(self): | ||
super().test_multiple_wrong_adapter_name_raises_error() | ||
|
||
|
||
@require_torch_gpu | ||
@require_peft_backend | ||
|
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -37,6 +37,7 @@ | |
from diffusers.utils.import_utils import is_accelerate_available | ||
from diffusers.utils.testing_utils import ( | ||
CaptureLogger, | ||
is_flaky, | ||
load_image, | ||
nightly, | ||
numpy_cosine_similarity_distance, | ||
|
@@ -111,6 +112,10 @@ def tearDown(self): | |
gc.collect() | ||
torch.cuda.empty_cache() | ||
|
||
@is_flaky | ||
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. There's absolutely no reason for it to be flaky but I think okay for now. |
||
def test_multiple_wrong_adapter_name_raises_error(self): | ||
super().test_multiple_wrong_adapter_name_raises_error() | ||
|
||
|
||
@slow | ||
@nightly | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Now that we're catching the error appropriately within the code, we should skip this test for unsupported models.