Skip to content

Commit

Permalink
Merge pull request #33131 from apache/users/damccorm/typo
Browse files Browse the repository at this point in the history
Fix vllm typo
  • Loading branch information
liferoad authored Nov 15, 2024
2 parents 5bfacc9 + cc57bde commit bc28ea4
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion examples/notebooks/beam-ml/run_inference_vllm.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@
{
"cell_type": "markdown",
"source": [
"[vLLM](https://github.com/vllm-project/vllm) is a fast and user-frienly library for LLM inference and serving. vLLM optimizes LLM inference with mechanisms like PagedAttention for memory management and continuous batching for increasing throughput. For popular models, vLLM has been shown to increase throughput by a multiple of 2 to 4. With Apache Beam, you can serve models with vLLM and scale that serving with just a few lines of code.\n",
"[vLLM](https://github.com/vllm-project/vllm) is a fast and user-friendly library for LLM inference and serving. vLLM optimizes LLM inference with mechanisms like PagedAttention for memory management and continuous batching for increasing throughput. For popular models, vLLM has been shown to increase throughput by a multiple of 2 to 4. With Apache Beam, you can serve models with vLLM and scale that serving with just a few lines of code.\n",
"\n",
"This notebook demonstrates how to run machine learning inference by using vLLM and GPUs in three ways:\n",
"\n",
Expand Down

0 comments on commit bc28ea4

Please sign in to comment.