Fix(ai): Use gemma3n:e2b model by default for ollama gptel
This commit is contained in:
@@ -12,7 +12,7 @@
|
|||||||
;; Cliente LLM (ollama, chatgpt, gemini, etc.)
|
;; Cliente LLM (ollama, chatgpt, gemini, etc.)
|
||||||
(use-package gptel
|
(use-package gptel
|
||||||
:config
|
:config
|
||||||
(setq gptel-model 'gemma4:e2b
|
(setq gptel-model 'gemma3n:e2b
|
||||||
gptel-backend (gptel-make-ollama "Ollama"
|
gptel-backend (gptel-make-ollama "Ollama"
|
||||||
:host "localhost:11434"
|
:host "localhost:11434"
|
||||||
:stream t
|
:stream t
|
||||||
|
|||||||
Reference in New Issue
Block a user