Skip to content

Commit

Permalink
Update docs
Browse files Browse the repository at this point in the history
  • Loading branch information
marella committed Aug 12, 2023
1 parent d5ea825 commit 4c7a1a5
Show file tree
Hide file tree
Showing 2 changed files with 13 additions and 15 deletions.
26 changes: 12 additions & 14 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,17 +16,17 @@ Python bindings for the Transformer models implemented in C/C++ using [GGML](htt

## Supported Models

| Models | Model Type |
| :-------------------- | ----------- |
| GPT-2 | `gpt2` |
| GPT-J, GPT4All-J | `gptj` |
| GPT-NeoX, StableLM | `gpt_neox` |
| LLaMA, LLaMA 2 | `llama` |
| MPT | `mpt` |
| Dolly V2 | `dolly-v2` |
| Replit | `replit` |
| StarCoder, StarChat | `starcoder` |
| Falcon (Experimental) | `falcon` |
| Models | Model Type | CUDA | Metal |
| :------------------ | ------------- | :--: | :---: |
| GPT-2 | `gpt2` | | |
| GPT-J, GPT4All-J | `gptj` | | |
| GPT-NeoX, StableLM | `gpt_neox` | | |
| Falcon | `falcon` || |
| LLaMA, LLaMA 2 | `llama` || |
| MPT | `mpt` || |
| StarCoder, StarChat | `gpt_bigcode` || |
| Dolly V2 | `dolly-v2` | | |
| Replit | `replit` | | |

## Installation

Expand Down Expand Up @@ -108,8 +108,6 @@ It is integrated into LangChain. See [LangChain docs](https://python.langchain.c

### GPU

> **Note:** Currently only LLaMA, MPT and Falcon models have GPU support.
To run some of the model layers on GPU, set the `gpu_layers` parameter:

```py
Expand Down Expand Up @@ -179,7 +177,7 @@ It can also be used with LangChain. Low-level APIs are not fully supported.
| `context_length` | `int` | The maximum context length to use. | `-1` |
| `gpu_layers` | `int` | The number of layers to run on GPU. | `0` |

> **Note:** Currently only LLaMA, MPT and Falcon models support the `context_length` and `gpu_layers` parameters.
> **Note:** Currently only LLaMA, MPT and Falcon models support the `context_length` parameter.
### <kbd>class</kbd> `AutoModelForCausalLM`

Expand Down
2 changes: 1 addition & 1 deletion scripts/docs.py
Original file line number Diff line number Diff line change
Expand Up @@ -29,7 +29,7 @@
default = getattr(Config, param)
docs += f"| `{param}` | `{type_}` | {description} | `{default}` |\n"
docs += """
> **Note:** Currently only LLaMA, MPT and Falcon models support the `context_length` and `gpu_layers` parameters.
> **Note:** Currently only LLaMA, MPT and Falcon models support the `context_length` parameter.
"""

# Class Docs
Expand Down

0 comments on commit 4c7a1a5

Please sign in to comment.