Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] GPU support for GPT-J #2435

Closed
3Simplex opened this issue Jun 12, 2024 · 3 comments
Closed

[Feature] GPU support for GPT-J #2435

3Simplex opened this issue Jun 12, 2024 · 3 comments
Labels
backend gpt4all-backend issues enhancement New feature or request

Comments

@3Simplex
Copy link
Collaborator

Bug Report

When using Mac set to use metal, gpt-j model fails to fallback to CPU.

Steps to Reproduce

  1. With a Mac set application device to use metal.
  2. Try to load a gpt-j architecture model.
  3. With a Mac set application device to use CPU.
  4. Load the same gpt-j architecture model.

Expected Behavior

The gpt-j model has no GPU support should fallback to CPU.

Your Environment

  • GPT4All version: 2.8.0
  • Operating System: Mac
  • Chat model used (if applicable): https://huggingface.co/EleutherAI/gpt-j-6b converted to gguf.

Works in windows as expected with Vulkan.

@3Simplex 3Simplex added bug-unconfirmed chat gpt4all-chat issues labels Jun 12, 2024
@3Simplex
Copy link
Collaborator Author

From Discord: qui — Today at 7:13 PM
Yes. Happened when set to auto as well - in MAC
Also failed when set to CUDA on Windows.

@cebtenzzre
Copy link
Member

GPU support for GPT-J would basically require upstreaming it into llama.cpp. It's the one architecture that we still support using an old implementation from the original ggml repo.

@cebtenzzre cebtenzzre changed the title GPT-J model failing to fallback to CPU [Feature] GPU support for GPT-J Jun 13, 2024
@cebtenzzre cebtenzzre added enhancement New feature or request backend gpt4all-backend issues and removed chat gpt4all-chat issues bug-unconfirmed labels Jun 13, 2024
@cebtenzzre
Copy link
Member

Support for GPT-J models was removed in #2676. If someone makes a proper llama.cpp implementation of GPT-J, it should get GPU support automatically.

@cebtenzzre cebtenzzre closed this as not planned Won't fix, can't repro, duplicate, stale Aug 4, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
backend gpt4all-backend issues enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants