chore(ai): Update Ollama AI model list
This commit is contained in:
@@ -16,15 +16,24 @@
|
||||
gptel-backend (gptel-make-ollama "Ollama"
|
||||
:host "localhost:11434"
|
||||
:stream t
|
||||
:models '("mistral:latest"
|
||||
"deepseek-r1:1.5b"
|
||||
"gemma3n:e2b"
|
||||
:models '("cajina/gemma4_e2b-Q4_k_s:v01"
|
||||
"su_robin/gemma-4-E4B-it-Q4_K_M:latest"
|
||||
"dagbs/qwen2.5-coder-0.5b-instruct-abliterated:q4_k_l"
|
||||
"jewelzufo/Qwen2.5-Coder-0.5B-Instruct-GGUF-Assistant:latest"
|
||||
"jaahas/qwen3.5-uncensored:4b"
|
||||
"jaahas/qwen3.5-uncensored:2b"
|
||||
"qwen3.5:4b"
|
||||
"qwen3.5:2b"
|
||||
"gemma4:e2b"
|
||||
"cajina/gemma4_e2b-q2_k_xl:v01"
|
||||
"deepcoder"
|
||||
"glm-5.1:cloud"
|
||||
"gemma4:latest"
|
||||
"gemini-3-flash-preview:cloud")))
|
||||
"gemma4:31b-cloud"
|
||||
"minimax-m2.5:cloud"
|
||||
"gpt-oss:120b-cloud"
|
||||
"gemini-3-flash-preview:cloud"
|
||||
"glm-5:cloud"
|
||||
"qwen3-coder-next:cloud"
|
||||
"qwen3-coder:480b-cloud"
|
||||
"embeddinggemma:latest")))
|
||||
(when (getenv "GEMINI_KEY")
|
||||
(setq gptel-model 'gemini-2.5-flash
|
||||
gptel-backend (gptel-make-gemini "Gemini"
|
||||
|
||||
Reference in New Issue
Block a user