Skip to content

Releases: ahyatt/llm

Fix for various byte compilation issues, with real Vertex issue

05 May 23:34
Compare
Choose a tag to compare

The 0.14.0 was released incorrectly, and had a few byte compilation issues, which are now fixed.

Corrected incorrect functions in Vertex and the request library to functions that no longer exist.

Change to way prompts are created

05 May 21:34
Compare
Choose a tag to compare

What's Changed

  • llm-make-chat-prompt is introduced and obsoletes other ways of creating chat prompts.
  • Improved Vertex error handling
  • fix: OpenAI API keys passed as multibyte strings by @hraban in #44

New Contributors

Full Changelog: 0.13.0...0.14.0

Claude function calling, provider rewrite, better cleanup and errors

08 Apr 01:23
Compare
Choose a tag to compare
  • Add Claude's new support for function calling.
  • Refactor of providers to centralize embedding and chat logic.
  • Remove connection buffers after use.
  • Fixes to provider more specific error messages for most providers.

Warning improvements

31 Mar 05:19
Compare
Choose a tag to compare
  • Refactor of warn-non-nonfree methods.
  • Add non-free warnings for Gemini and Claude.

Error and streaming fixes

25 Mar 04:29
Compare
Choose a tag to compare
  • Send connection issues to error callbacks, and fix an error handling issue in Ollama.
  • Fix issue where, in some cases, streaming does not work the first time attempted.

Ollama and Open AI-compatibility fixes

22 Mar 05:04
Compare
Choose a tag to compare
  • Fix issue in llm-ollama with not using provider host for sync embeddings.
  • Fix issue in llm-openai where were incompatible with some Open AI-compatible backends due to assumptions about inconsequential JSON details.

Claude support

17 Mar 05:29
Compare
Choose a tag to compare

This release adds support for Anthropic's Claude models.

Function calling plus minor fixes

02 Mar 20:08
Compare
Choose a tag to compare

What's Changed

  • Added function calling, currently supported just by Open AI and Gemini (and, frankly, Gemini's support has problems)
  • Introduce llm-capabilities, which indicates which non-standard functionality each provider supports.
  • Fixed issue of logging when we weren't supposed to.
  • Change to use Ollama chat endpoint by @tquartus in #16
  • Fix emacs 28 incompatibility llm-fake by @s-kostyaev in #21
  • Add CI by @s-kostyaev in #24
  • Fix ollama mentioned instead of llama.cpp by @SmallAndSoft in #25

New Contributors

Full Changelog: 0.9.1...0.10.0

Fixes for Gemini, new default for Open AI embedding model

03 Feb 23:17
Compare
Choose a tag to compare
  • Default to the new "text-embedding-3-small" model for Open AI. Important: Anyone who has stored embeddings should either regenerate embeddings (recommended) or hard-code the old embedding model ("text-embedding-ada-002").
  • Fix response breaking when prompts run afoul of Gemini / Vertex's safety checks.
  • Change Gemini streaming to be the correct URL. This doesn't seem to have an effect on behavior.

Token limits

21 Jan 02:17
Compare
Choose a tag to compare
  • Add llm-chat-token-limit to find the token limit based on the model.
  • Add request timeout customization.