KEMBAR78
Set multimodal status during Model Loading by altoiddealer · Pull Request #7199 · oobabooga/text-generation-webui · GitHub
Skip to content

Conversation

altoiddealer
Copy link
Contributor

@altoiddealer altoiddealer commented Aug 13, 2025

[ X ] I have read the Contributing guidelines.

I'm proposing this change to make it easier for future expansion (if other loaders may support multimodal, etc).

Instead of analyzing if the model is multimodal capable on every generation, in load_model():

  • Assert False by default
  • if Exl3 or llama.cpp - check once and store the result in shared.
  • Suggestion: Maybe use a placeholder is_multimodal() -> False in all other loader modules.

@altoiddealer
Copy link
Contributor Author

Tested with ExLlamav3 model with vision, and llama.cpp using the example model scenario in your tutorial.

@altoiddealer
Copy link
Contributor Author

I realized that I did not commit the correct version of models.py. I resolved this now.

@oobabooga
Copy link
Owner

Good idea, it's indeed better to check if the model is multimodal just once instead of every call.

Suggestion: Maybe use a placeholder is_multimodal() -> False in all other loader modules.

I can see the merit in this. To keep things simple, I'd prefer not to add it for now for loaders where it can't be True.

Thanks for the excellent refactor!

@oobabooga oobabooga changed the base branch from main to dev August 13, 2025 19:47
@oobabooga oobabooga merged commit 57f6e9a into oobabooga:dev Aug 13, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants