File tree 1 file changed +4
-4
lines changed
1 file changed +4
-4
lines changed Original file line number Diff line number Diff line change @@ -287,16 +287,16 @@ function! s:ring_update()
287
287
288
288
" no samplers needed here
289
289
let l: request = json_encode ({
290
- \ ' prompt' : " " ,
291
290
\ ' input_prefix' : " " ,
292
291
\ ' input_suffix' : " " ,
292
+ \ ' input_extra' : l: extra_context ,
293
+ \ ' prompt' : " " ,
293
294
\ ' n_predict' : 1 ,
294
295
\ ' penalty_last_n' : 0 ,
295
296
\ ' temperature' : 0.0 ,
296
297
\ ' stream' : v: false ,
297
298
\ ' samplers' : [" temperature" ],
298
299
\ ' cache_prompt' : v: true ,
299
- \ ' extra_context' : l: extra_context ,
300
300
\ ' t_max_prompt_ms' : 1 ,
301
301
\ ' t_max_predict_ms' : 1
302
302
\ })
@@ -379,16 +379,16 @@ function! llama#fim(is_auto, on_hold) abort
379
379
380
380
let l: request = json_encode ({
381
381
\ ' input_prefix' : l: prefix ,
382
- \ ' prompt' : l: prompt ,
383
382
\ ' input_suffix' : l: suffix ,
383
+ \ ' input_extra' : l: extra_context ,
384
+ \ ' prompt' : l: prompt ,
384
385
\ ' n_predict' : g: llama_config .n_predict,
385
386
\ ' penalty_last_n' : 0 ,
386
387
\ ' top_k' : 40 ,
387
388
\ ' top_p' : 0.99 ,
388
389
\ ' stream' : v: false ,
389
390
\ ' samplers' : [" top_k" , " top_p" , " infill" ],
390
391
\ ' cache_prompt' : v: true ,
391
- \ ' extra_context' : l: extra_context ,
392
392
\ ' t_max_prompt_ms' : g: llama_config .t_max_prompt_ms,
393
393
\ ' t_max_predict_ms' : g: llama_config .t_max_predict_ms
394
394
\ })
You can’t perform that action at this time.
0 commit comments