From 7398f795e3fde21b8c1a6a40cd67c0b1854ed60c Mon Sep 17 00:00:00 2001 From: Peter Tripp Date: Wed, 25 Sep 2024 22:01:12 +0000 Subject: [PATCH] Ollama llama3.2 default context size (#18366) Release Notes: - Ollama: Added llama3.2 support --- crates/ollama/src/ollama.rs | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/crates/ollama/src/ollama.rs b/crates/ollama/src/ollama.rs index e592bfa177..84404afce1 100644 --- a/crates/ollama/src/ollama.rs +++ b/crates/ollama/src/ollama.rs @@ -83,7 +83,7 @@ fn get_max_tokens(name: &str) -> usize { "codellama" | "starcoder2" => 16384, "mistral" | "codestral" | "mixstral" | "llava" | "qwen2" | "dolphin-mixtral" => 32768, "llama3.1" | "phi3" | "phi3.5" | "command-r" | "deepseek-coder-v2" | "yi-coder" - | "qwen2.5-coder" => 128000, + | "llama3.2" | "qwen2.5-coder" => 128000, _ => DEFAULT_TOKENS, } .clamp(1, MAXIMUM_TOKENS)