Skip to content

Commit

Permalink
add tokenizer for codellama 70b
Browse files Browse the repository at this point in the history
  • Loading branch information
mmoskal committed Feb 23, 2024
1 parent 1b5691c commit ae9c35d
Show file tree
Hide file tree
Showing 2 changed files with 10 additions and 0 deletions.
6 changes: 6 additions & 0 deletions aicirt/src/bintokens.rs
Original file line number Diff line number Diff line change
Expand Up @@ -43,6 +43,12 @@ pub fn tokenizers() -> Vec<TokenizerInfo> {
hf_model: "codellama/CodeLlama-13b-Instruct-hf",
model_ids: "codellama-13b",
},
TokenizerInfo {
name: "llama70",
description: "used by codellama-70b; with <step> token",
hf_model: "codellama/CodeLlama-70b-Instruct-hf",
model_ids: "codellama-70b",
},
TokenizerInfo {
name: "orca",
description: "llama",
Expand Down
4 changes: 4 additions & 0 deletions rllm/rllm-cuda/server.sh
Original file line number Diff line number Diff line change
Expand Up @@ -69,6 +69,9 @@ if [ "$CPP" = 1 ] ; then
mixtral )
ARGS="-m https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF/blob/main/mixtral-8x7b-instruct-v0.1.Q6_K.gguf"
;;
code70 )
ARGS="-m https://huggingface.co/TheBloke/CodeLlama-70B-Instruct-GGUF/blob/main/codellama-70b-instruct.Q5_K_M.gguf"
;;
https* )
ARGS="-m $1"
;;
Expand All @@ -86,6 +89,7 @@ model_name can a HuggingFace URL pointing to a .gguf file, or one of the followi
orca https://huggingface.co/TheBloke/Orca-2-13B-GGUF/blob/main/orca-2-13b.Q8_0.gguf
mistral https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GGUF/blob/main/mistral-7b-instruct-v0.2.Q5_K_M.gguf
mixtral https://huggingface.co/TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF/blob/main/mixtral-8x7b-instruct-v0.1.Q6_K.gguf
code70 https://huggingface.co/TheBloke/CodeLlama-70B-Instruct-GGUF/blob/main/codellama-70b-instruct.Q5_K_M.gguf
Additionally, "$SELF build" will just build the server, and not run a model.
Expand Down

0 comments on commit ae9c35d

Please sign in to comment.