v0.2.2
- feat/fix: adding deepseek model by @makyinmars in #16
- Implement prefill for some models by @chottolabs in #17
New Contributors
- @makyinmars made their first contribution in #16
Full Changelog: v0.2.1...v0.2.2
Note:
At this point, only anthropic, deepseek, vllm have working prefill implementations - however I have 0 credits for them so I will have fun with deepseek 🥲
The other thing is all their implementations are different
- anthropic just automatically assumes if your last message is from assistant, then it's prefill
- groq expects stop but doesn't seem to support prefill at all (in my testing, it tries to start with
```
and immediately terminates) - lambda/vllm expects stop_token_ids (not even sure if it works properly, i think it's just the
add_generation_prompt
condition in the jinja the template) - deepseek expects
stop = ```
, but also requires you to addprefix = true
to the final assistant message