Skip to content

using from_pipe and sequential / model offload #10015

Answered by asomoza
paparico asked this question in Q&A
Discussion options

You must be logged in to vote

AFAIK the model offloading is per pipeline and the auto pipeline just transfers the different modules, so you'll still need to enable it in all the pipelines you will use for inference but if you just re-use the same one, you don't need to.

To clarify the other points that were mentioned:

  • You should not enable any kind of offloading even if your CPU is idle, this will always make inference slower, you should only use it when you don't have enough VRAM to load all the models at the same time. The only other use case when you can use this is when you want to load multiple models in the RAM and just transfer them to VRAM when needed, but this must be implemented manually and is not in the s…

Replies: 3 comments 1 reply

Comment options

You must be logged in to vote
1 reply
@paparico
Comment options

Comment options

You must be logged in to vote
0 replies
Answer selected by paparico
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
4 participants