Skip to content

Commit

Permalink
llama.vim : update infill API params [no ci]
Browse files Browse the repository at this point in the history
  • Loading branch information
ggerganov committed Oct 15, 2024
1 parent 36d7b6d commit 5155b68
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions examples/llama.vim
Original file line number Diff line number Diff line change
Expand Up @@ -287,16 +287,16 @@ function! s:ring_update()

" no samplers needed here
let l:request = json_encode({
\ 'prompt': "",
\ 'input_prefix': "",
\ 'input_suffix': "",
\ 'input_extra': l:extra_context,
\ 'prompt': "",
\ 'n_predict': 1,
\ 'penalty_last_n': 0,
\ 'temperature': 0.0,
\ 'stream': v:false,
\ 'samplers': ["temperature"],
\ 'cache_prompt': v:true,
\ 'extra_context': l:extra_context,
\ 't_max_prompt_ms': 1,
\ 't_max_predict_ms': 1
\ })
Expand Down Expand Up @@ -379,16 +379,16 @@ function! llama#fim(is_auto, on_hold) abort

let l:request = json_encode({
\ 'input_prefix': l:prefix,
\ 'prompt': l:prompt,
\ 'input_suffix': l:suffix,
\ 'input_extra': l:extra_context,
\ 'prompt': l:prompt,
\ 'n_predict': g:llama_config.n_predict,
\ 'penalty_last_n': 0,
\ 'top_k': 40,
\ 'top_p': 0.99,
\ 'stream': v:false,
\ 'samplers': ["top_k", "top_p", "infill"],
\ 'cache_prompt': v:true,
\ 'extra_context': l:extra_context,
\ 't_max_prompt_ms': g:llama_config.t_max_prompt_ms,
\ 't_max_predict_ms': g:llama_config.t_max_predict_ms
\ })
Expand Down

0 comments on commit 5155b68

Please sign in to comment.