Hi,

I've been using Emacs for decades and notmuch for years. I just want to 
recommend that notmuch users check out ellama with a local ollama installation.

It has been working like a charm using notmuch to have my emails 
grammar-checked by a local LLM installation. The only configuration change I 
made was to override the default content-type "flowed", which was causing 
issues for recipients.

I also recommend using unidecode because ollama returns text with unicode 
characters, and I want to send pure ASCII emails.

;; settings in my emacs config

;; prevent format=flowed
(setq message-default-headers "Content-Type: text/plain; charset=utf-8")

;; setup ellama
(use-package ellama
  :init
  (setopt ellama-language "English")
  (require 'llm-ollama)
  (setopt ellama-provider
                  (make-llm-ollama
                   :chat-model "gpt-oss" :embedding-model "gpt-oss" :host 
"my-gpu-machine-running-ollama")))

(global-set-key [f9] 'ellama-chat)
(global-set-key (kbd "<S-f9>") 'ellama-proofread)



Here're the links to the projects:

ellama: https://github.com/s-kostyaev/ellama
unidecode: https://github.com/sindikat/unidecode
ollama: https://github.com/ollama/ollama


Have fun,
  Sebastian
_______________________________________________
notmuch mailing list -- notmuch@notmuchmail.org
To unsubscribe send an email to notmuch-le...@notmuchmail.org

Reply via email to