feat(ai): Add gptel switch and allow use ollama alongside gemini
This commit is contained in:
@@ -12,26 +12,24 @@
|
|||||||
;; Cliente LLM (ollama, chatgpt, gemini, etc.)
|
;; Cliente LLM (ollama, chatgpt, gemini, etc.)
|
||||||
(use-package gptel
|
(use-package gptel
|
||||||
:config
|
:config
|
||||||
(if (getenv "GEMINI_KEY")
|
(setq gptel-model 'gemma4:e2b
|
||||||
(setq gptel-model 'gemini-2.5-flash
|
gptel-backend (gptel-make-ollama "Ollama"
|
||||||
gptel-backend (gptel-make-gemini "Gemini"
|
:host "localhost:11434"
|
||||||
:key (getenv "GEMINI_KEY")
|
:stream t
|
||||||
:stream t))
|
:models '("mistral:latest"
|
||||||
(setq gptel-model 'cajina/gemma4_e2b-q2_k_xl:v01
|
"deepseek-r1:1.5b"
|
||||||
gptel-backend (gptel-make-ollama "Ollama"
|
"gemma3n:e2b"
|
||||||
:host "localhost:11434"
|
"gemma4:e2b"
|
||||||
:request-params '(:think nil)
|
"cajina/gemma4_e2b-q2_k_xl:v01"
|
||||||
:stream t
|
"deepcoder"
|
||||||
:models '("mistral:latest"
|
"glm-5.1:cloud"
|
||||||
"deepseek-r1:1.5b"
|
"gemma4:latest"
|
||||||
"gemma3n:e2b"
|
"gemini-3-flash-preview:cloud")))
|
||||||
"gemma4:e2b"
|
(when (getenv "GEMINI_KEY")
|
||||||
"cajina/gemma4_e2b-q2_k_xl:v01"
|
(setq gptel-model 'gemini-2.5-flash
|
||||||
"deepcoder"
|
gptel-backend (gptel-make-gemini "Gemini"
|
||||||
"glm-5.1:cloud"
|
:key (getenv "GEMINI_KEY")
|
||||||
"gemma4:latest"
|
:stream t)))
|
||||||
"gemini-3-flash-preview:cloud")))
|
|
||||||
)
|
|
||||||
(setq gptel-default-mode 'org-mode
|
(setq gptel-default-mode 'org-mode
|
||||||
gptel-prompt-prefix-alist
|
gptel-prompt-prefix-alist
|
||||||
'((markdown-mode . "# User\n\n")
|
'((markdown-mode . "# User\n\n")
|
||||||
@@ -44,7 +42,33 @@
|
|||||||
gptel-directives
|
gptel-directives
|
||||||
'((default . "You are a large language model living in Emacs and a helpful assistant. Respond concisely in Spanish."))
|
'((default . "You are a large language model living in Emacs and a helpful assistant. Respond concisely in Spanish."))
|
||||||
)
|
)
|
||||||
(setopt gptel-include-reasoning nil)) ;; Mantener hasta resolver: https://github.com/ragnard/gptel-magit/issues/8
|
(setopt gptel-include-reasoning nil) ;; Mantener hasta resolver: https://github.com/ragnard/gptel-magit/issues/8
|
||||||
|
|
||||||
|
(defun gptel-switch+model ()
|
||||||
|
"Switch to gptel backend and model in a single completion prompt."
|
||||||
|
(interactive)
|
||||||
|
(let (choices)
|
||||||
|
(dolist (pair gptel--known-backends)
|
||||||
|
(let* ((backend-name (car pair))
|
||||||
|
(backend (cdr pair))
|
||||||
|
(models
|
||||||
|
(and (fboundp 'gptel-backend-models)
|
||||||
|
(gptel-backend-models backend))))
|
||||||
|
(when models
|
||||||
|
(dolist (model models)
|
||||||
|
(push (cons ; (format "%s:%s" backend-name model)
|
||||||
|
(format "%s → %s"
|
||||||
|
(propertize backend-name 'face 'font-lock-keyword-face)
|
||||||
|
(propertize (symbol-name model) 'face 'font-lock-function-name-face))
|
||||||
|
(cons backend-name model))
|
||||||
|
choices)))))
|
||||||
|
(let* ((choice
|
||||||
|
(completing-read "Model: " (mapcar #'car choices) nil t))
|
||||||
|
(sel (cdr (assoc choice choices))))
|
||||||
|
(setq gptel-backend (cdr (assoc (car sel) gptel--known-backends))
|
||||||
|
gptel-model (cdr sel))
|
||||||
|
(message "gptel set to %s:%s" (car sel) (cdr sel)))))
|
||||||
|
)
|
||||||
|
|
||||||
;; (use-package copilot
|
;; (use-package copilot
|
||||||
;; :hook (prog-mode . copilot-mode)
|
;; :hook (prog-mode . copilot-mode)
|
||||||
|
|||||||
Reference in New Issue
Block a user