Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

llama2 #459

Open
1 task
irthomasthomas opened this issue Jan 28, 2024 · 0 comments
Open
1 task

llama2 #459

irthomasthomas opened this issue Jan 28, 2024 · 0 comments
Labels
AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models base-model llm base models not finetuned for chat llm-quantization All about Quantized LLM models and serving openai OpenAI APIs, LLMs, Recipes and Evals

Comments

@irthomasthomas
Copy link
Owner

Llama 2

The most popular model for general use.

265.8K Pulls
Updated 4 weeks ago

Overview

Llama 2 is released by Meta Platforms, Inc. This model is trained on 2 trillion tokens, and by default supports a context length of 4096. Llama 2 Chat models are fine-tuned on over 1 million human annotations, and are made for chat.

CLI

Open the terminal and run

ollama run llama2

API

Example using curl:

curl -X POST http://localhost:11434/api/generate -d '{
  "model": "llama2",
  "prompt":"Why is the sky blue?"
 }'

API documentation

Memory requirements

  • 7b models generally require at least 8GB of RAM
  • 13b models generally require at least 16GB of RAM
  • 70b models generally require at least 64GB of RAM

If you run into issues with higher quantization levels, try using the q4 model or shut down any other programs that are using a lot of memory.

Model variants

  • Chat: fine-tuned for chat/dialogue use cases. These are the default in Ollama, and for models tagged with -chat in the tags tab.

    Example: ollama run llama2

  • Pre-trained: without the chat fine-tuning. This is tagged as -text in the tags tab.

    Example: ollama run llama2:text

By default, Ollama uses 4-bit quantization. To try other quantization levels, please use the other tags. The number after the q represents the number of bits used for quantization (i.e. q4 means 4-bit quantization). The higher the number, the more accurate the model is, but the slower it runs, and the more memory it requires.

References

Suggested labels

{ "label-name": "llama2-model", "description": "A powerful text model for chat, dialogue, and general use.", "repo": "ollama.ai/library/llama2", "confidence": 91.74 }

@irthomasthomas irthomasthomas added AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models base-model llm base models not finetuned for chat llama llm-quantization All about Quantized LLM models and serving New-Label Choose this option if the existing labels are insufficient to describe the content accurately openai OpenAI APIs, LLMs, Recipes and Evals and removed New-Label Choose this option if the existing labels are insufficient to describe the content accurately labels Jan 28, 2024
@ShellLM ShellLM removed the llama label May 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
AI-Chatbots Topics related to advanced chatbot platforms integrating multiple AI models base-model llm base models not finetuned for chat llm-quantization All about Quantized LLM models and serving openai OpenAI APIs, LLMs, Recipes and Evals
Projects
None yet
Development

No branches or pull requests

2 participants