Fix(ai): Use gemma3n:e2b model by default for ollama gptel
This commit is contained in:
@@ -12,7 +12,7 @@
|
||||
;; Cliente LLM (ollama, chatgpt, gemini, etc.)
|
||||
(use-package gptel
|
||||
:config
|
||||
(setq gptel-model 'gemma4:e2b
|
||||
(setq gptel-model 'gemma3n:e2b
|
||||
gptel-backend (gptel-make-ollama "Ollama"
|
||||
:host "localhost:11434"
|
||||
:stream t
|
||||
|
||||
Reference in New Issue
Block a user