Skip to content

Commit b2fda18

Browse files
committed
llama: increase MEM_REQ_EVAL for MODEL_3B
It avoids crashing for quantized weights on CPU. Better ways to calculate the required buffer size would be better.
1 parent c0bee23 commit b2fda18

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

llama.cpp

+1-1
Original file line numberDiff line numberDiff line change
@@ -122,7 +122,7 @@ static const std::map<e_model, size_t> & MEM_REQ_KV_SELF()
122122
static const std::map<e_model, size_t> & MEM_REQ_EVAL()
123123
{
124124
static std::map<e_model, size_t> k_sizes = {
125-
{ MODEL_3B, 512ull * MB },
125+
{ MODEL_3B, 640ull * MB },
126126
{ MODEL_7B, 768ull * MB },
127127
{ MODEL_13B, 1024ull * MB },
128128
{ MODEL_30B, 1280ull * MB },

0 commit comments

Comments
 (0)