File tree 2 files changed +15
-1
lines changed
2 files changed +15
-1
lines changed Original file line number Diff line number Diff line change @@ -7,6 +7,20 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
7
7
8
8
## [ Unreleased]
9
9
10
+ ## [ 0.3.3]
11
+
12
+ - feat: Update llama.cpp to ggerganov/llama.cpp@ce8784bdb153ff7794dde5a50b0ebfa51baa6171
13
+ - fix: chat API logprobs format by @domdomegg in #1788
14
+ - feat: Add support for CUDA 12.6, fix CUDA 12.5 by @Smartappli in #1775
15
+ - fix: Make content not required in ChatCompletionRequestAssistantMessage by @feloy in #1807
16
+ - fix: Fix pickling of Llama class by setting seed from _ seed member by @abetlen in 2523472c3eccb9ab9277117cc4ff705212b6888a
17
+ - fix: Fix logit-bias type hint by @ddh0 in #1802
18
+ - fix(server): Avoid thread starvation on many concurrent requests by making use of asyncio to lock llama_proxy context by @gjpower in #1798
19
+ - fix(server): Added missing exit_stack.close() to /v1/chat/completions by @Ian321 in #1796
20
+ - fix(examples): Refactor Batching notebook to use new sampler chain API by @lukestanley in #1793
21
+ - fix(docs): Update development instructions by @Florents-Tselai in #1833
22
+ - fix(docs): Remove ref to llama_eval in llama_cpp.py docs by @richdougherty in #1819
23
+
10
24
## [ 0.3.2]
11
25
12
26
- feat: Update llama.cpp to ggerganov/llama.cpp@74d73dc85cc2057446bf63cc37ff649ae7cebd80
Original file line number Diff line number Diff line change 1
1
from .llama_cpp import *
2
2
from .llama import *
3
3
4
- __version__ = "0.3.2 "
4
+ __version__ = "0.3.3 "
You can’t perform that action at this time.
0 commit comments