Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat(vllm): add support for image-to-text and video-to-text #3729

Merged
merged 7 commits into from
Oct 4, 2024

Conversation

mudler
Copy link
Owner

@mudler mudler commented Oct 4, 2024

Description

Depends on #3728

Closes: #2318
Closes: #3670

Also covers #3669 as should enable also Vision with Llama3.2 models

Notes for Reviewers

not tested yet on a GPU - just made sure it does not break generation by running vLLM in CPU-mode.

Signed commits

  • Yes, I signed my commits.

Related to #3670

Signed-off-by: Ettore Di Giacinto <[email protected]>
@mudler mudler added the enhancement New feature or request label Oct 4, 2024
@mudler mudler marked this pull request as draft October 4, 2024 15:30
Copy link

netlify bot commented Oct 4, 2024

Deploy Preview for localai ready!

Name Link
🔨 Latest commit d68e7ae
🔍 Latest deploy log https://app.netlify.com/sites/localai/deploys/6700274f04703500083d8b3f
😎 Deploy Preview https://deploy-preview-3729--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify site configuration.

@mudler mudler changed the title feat(vllm): add support for image-to-text feat(vllm): add support for image-to-text and video-to-text Oct 4, 2024
@mudler
Copy link
Owner Author

mudler commented Oct 4, 2024

Gotta try images on a GPU card. On CPU seems that generation is not broken at least, so should be good to merge and test on master

@mudler mudler marked this pull request as ready for review October 4, 2024 17:29
Signed-off-by: Ettore Di Giacinto <[email protected]>
backend/python/vllm/backend.py Outdated Show resolved Hide resolved
backend/python/vllm/install.sh Outdated Show resolved Hide resolved
Signed-off-by: Ettore Di Giacinto <[email protected]>
@mudler mudler merged commit 2553de0 into master Oct 4, 2024
33 of 34 checks passed
@mudler mudler deleted the feat/vllm_multimodal branch October 4, 2024 21:42
siddimore pushed a commit to siddimore/LocalAI that referenced this pull request Oct 6, 2024
)

* feat(vllm): add support for image-to-text

Related to mudler#3670

Signed-off-by: Ettore Di Giacinto <[email protected]>

* feat(vllm): add support for video-to-text

Closes: mudler#2318

Signed-off-by: Ettore Di Giacinto <[email protected]>

* feat(vllm): support CPU installations

Signed-off-by: Ettore Di Giacinto <[email protected]>

* feat(vllm): add bnb

Signed-off-by: Ettore Di Giacinto <[email protected]>

* chore: add docs reference

Signed-off-by: Ettore Di Giacinto <[email protected]>

* Apply suggestions from code review

Signed-off-by: Ettore Di Giacinto <[email protected]>

---------

Signed-off-by: Ettore Di Giacinto <[email protected]>
Signed-off-by: Ettore Di Giacinto <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Support multimodals models with vLLM feat(multimodal): Video understanding
1 participant