-
Notifications
You must be signed in to change notification settings - Fork 5.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pipeline has no attribute '_execution_device' #9180
Comments
hi @choidaedae would you be able to provide a minimum reproducible code example? i.e. a script I can run on my end to get the same error that you're getting? it would really help us understand the problem :) |
hi @yiyixuxu |
same question. |
Same question |
I had a similar problem, and it happened randomly so it was hard to reproduce, and I have now abandoned this attribute |
so can we assume this issue is resolved? OP also didn't provide any code to reproduce the error. |
I am sorry for not providing any code, but it occurs randomly and actually I'm not working with this pipeline now, so I can't provide it. If anyone still has this problem, providing the code can be userful for huggingface. Thank you for your hard work : ) |
think we can try to catch the exception for that property - so maybe in the future it happens we will have some visibility into the problem |
Same problem. This bug will disappear if I run it again. |
Same problem, occurs randomly. |
Same problem, occurs about half of the time with the same script |
Again, having a minimal reproducible snippet would be really amazing! |
I am re-opening this due to 3 comments in two weeks because it seems like a legitimate problem now. Could any of you please provide a minimal script for what you're trying that causes this error? I don't see enough information here to be able to start debugging. If it happens on a single GPU as well, as mentioned above, I think it is safe to rule out race conditions, but maybe not since there is mentions that it happens randomly. It would also be nice if you could check with latest diffusers release |
@a-r-r-o-w, see ,my comment here #9180 (comment) |
Yes, me too, but always and reproducible:
So I tried to make you a dummy model + dummy embeds minimal example -- but found myself flummoxed as I suddenly didn't get that error anymore - but got the error I actually expected, informing me about what is expected from me (as I had hoped!):
So now I am saving the dummies and loading them from file, and I get the error maybe 20% of the time (primary drive). However, if I save + load the dummies from HDD, and not from fast NVMe SSD, I can get the I should also note that I am loading Flux from a secondary not-as-fast internal SSD, so this may help in erroring out. I remember very occasionally getting AttributeError: So yeah, as it appears to me, this new issue is again related to speed of loading the weights for some weird reason. Maybe the two different errors are even related? I've had that happen since "many versions" of PyTorch and even a version of Python ago (the one about 'device' attribute). Still, current:
Here's the code:
PS: If you have a tip for constructing Either way, I hope this helps your quest in squashing this bug - thanks in advance for your work! 👍 |
Thank you so much!
You can compute the text embeddings needed to run the from diffusers import FluxPipeline
import torch
ckpt_id = "black-forest-labs/FLUX.1-dev"
prompt = "a photo of a dog with cat-like look"
pipeline = FluxPipeline.from_pretrained(
ckpt_id,
vae=None,
torch_type=torch.bfloat16
).to("cuda")
with torch.no_grad():
print("Encoding prompts.")
prompt_embeds, pooled_prompt_embeds, text_ids = pipeline.encode_prompt(
prompt=prompt, prompt_2=None, max_sequence_length=512
) Does this help? |
for me it happens as well, but even simpler, pipe = StableDiffusionPipeline.from_pretrained(..)
print(pipe.dtype) sometimes work, and sometimes Only as part of a script and not in iPython, for example. I think it's related maybe to accelerate and offloading? |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
@HilaManor Hi, that example code is too simple, I can execute that 100 times and all 100 times it will work for me, probably there's something in your env that it's making this to happen. IMO this should be treated as an issue or bug if it happens with our pipelines and code, if someone is using part of it or doing a custom pipeline, this error probably means that there's something wrong in the code or the environment, but it would be ideal to also properly catch it and give a better explanation of the error. |
It occurs randomly, so I think it's a race condition. When I captured that and tried to debug this it disappeared, that's part of the reason I think it's a race thing.. |
if that would be the case it should happen at least once for me in a 100 times experiment no? What I mean is that we probably need more information about what you're doing, for example, your environment and hardware. |
@asomoza it happened on 5 different computers, which have 3 different types of GPUs (an A6000 server, an 2080Ti server and on a personal computer, a 3090Ti on a personal computer), with different CPUs and generally different hardware, The env is roughly the same (up to the fact that the nvidia drivers is different on each server) (below). Please let me know what more info could help.
|
thanks for the detailed answer, just to be clear, you are getting pipe = StableDiffusionPipeline.from_pretrained(..)
print(pipe.dtype) Ideally we should have a fully reproducible code snippet that gives the error all the time, since you're saying this is random, I'll try to use an AWS instance to test this for a longer time. As I told you, I constantly use diffusers (hundreds of times a day) with linux and windows, with a 3090, A100 and a mobile 4090 and never, not once, I had this error, so it seems this is going to be a really hard bug to catch. You can guess this also because this issue has a low number of active users, if this was a recurrent problem, we would have a lot of people/issues reporting this. |
Clarifying - for the code you wrote I get |
oh, I look at a lot of issues so I usually go by the title of the issue, that's why is not a good idea to use another issue to add yours. Probably this should go in a new one since it's totally unrelated to the one from this discussion. What happens to you it's even more weird, since the You're using just plain unmodified diffusers for this right? You seem to have solved it anyway but if you want to keep digging into this, please open a new issue with your problem and also we don't need the full library list of your environment, you can just use |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
closing this since there's no more updates from the main issue, assuming it was resolved or was a problem in the env. |
Intriguing. The bug is still present today but, for me, only with LoRAs. And it happens half of the times with exactly the same script. |
confirm @asomoza |
Hi, in case this is not that clear yet, we don't need a confirmation or "I have the same problem", what we need to be able to help you is a snippet of code that reproduces the error, also we can't support custom pipelines or custom code because we don't have the bandwidth, so this needs to be with just diffusers code. Sadly this is something than never happens to us or the majority of the people using diffusers, so we need your help to be able to help you. |
This happened to me when I was trying to trick diffuser's DDIMPipeline into working with a non-diffusers diffusion model (a bit unrelated to OP's original use case, but it may still be useful to someone). The findings is that the import torch
import diffusers
class DummyNoDevice(torch.nn.Module):
def __init__(self):
super().__init__()
def forward(self, sample, timestep):
return (sample,)
class DummyWithDevice(torch.nn.Module):
def __init__(self):
super().__init__()
self.device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
def forward(self, sample, timestep):
return (sample,)
noise_scheduler = diffusers.DDPMScheduler(1000)
pipe_no_device = diffusers.DDIMPipeline(unet=DummyNoDevice(), scheduler=noise_scheduler)
pipe_with_device = diffusers.DDIMPipeline(unet=DummyWithDevice(), scheduler=noise_scheduler)
try:
print(f"Success: W/ device: {pipe_with_device._execution_device}")
except Exception as e:
print(f"Failed: W/ device: {e}")
try:
print(f"Success: no device: {pipe_no_device._execution_device}")
except Exception as e:
print(f"Failed: no device: {e}") Result: (diffusers 0.32.2)
|
thanks a lot @mashrurmorshed @yiyixuxu wee have a reproducible snippet of code now |
Describe the bug
Hello, I implemented my own custom pipeline referring StableDiffusionPipeline (RepDiffusionPipeline), but there are some issues
I called "accelerator.prepare" properly, and mapped the models on device (with "to.(accelerator.device)")
But when I call pipeline and the 'call' function is called, sometimes I met the error
It is not only problem in using multi-gpu, it occurs when I use single gpu.
For example, I defined my pipeline for my validation in training code like this:
then, when I called 'val_pipe' like this:
At that time, the error "RepDiffusionPipeline has no attribute '_execution_device'" occurs. (Not always, just randomly)
How can I solve this issue, or what part of my code can be doubted and fixed?
Thank you for reading:)
Reproduction
It occurs randomly, so there is no option to reproduce...
But when I call the defined pipeline, it occurs randomly.
Logs
RepDiffusionPipeline has no attribute '_execution_device'
System Info
I tried to test in various diffusers & python versions, but the problem still occurs.
In now, I am running my code in diffusers 0.27.2, python 3.10.14.
WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for:
PyTorch 2.2.2+cu121 with CUDA 1201 (you have 2.2.2+cu118)
Python 3.10.14 (you have 3.10.14)
Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers)
Memory-efficient attention, SwiGLU, sparse and more won't be available.
Set XFORMERS_MORE_DETAILS=1 for more details
Copy-and-paste the text below in your GitHub issue and FILL OUT the two last points.
diffusers
version: 0.27.2Who can help?
@sayakpaul @yiyixuxu
The text was updated successfully, but these errors were encountered: