branch: elpa/gptel
commit 4c5f7e6953c3e4dac36b4c9cf4b4a2c854febfad
Author: Karthik Chikmagalur <[email protected]>
Commit: Karthik Chikmagalur <[email protected]>
gptel: Bump version to v0.9.8
* gptel.el: Update version and package description.
---
gptel.el | 23 ++++++++++++++---------
1 file changed, 14 insertions(+), 9 deletions(-)
diff --git a/gptel.el b/gptel.el
index d3e77ce1c3..e8f5e1406a 100644
--- a/gptel.el
+++ b/gptel.el
@@ -3,7 +3,7 @@
;; Copyright (C) 2023 Karthik Chikmagalur
;; Author: Karthik Chikmagalur <[email protected]>
-;; Version: 0.9.7
+;; Version: 0.9.8
;; Package-Requires: ((emacs "27.1") (transient "0.7.4") (compat "29.1.4.1"))
;; Keywords: convenience, tools
;; URL: https://github.com/karthink/gptel
@@ -36,7 +36,7 @@
;;
;; - The services ChatGPT, Azure, Gemini, Anthropic AI, Anyscale, Together.ai,
;; Perplexity, Anyscale, OpenRouter, Groq, PrivateGPT, DeepSeek, Cerebras,
-;; Github Models, xAI and Kagi (FastGPT & Summarizer).
+;; Github Models, Novita AI, xAI and Kagi (FastGPT & Summarizer).
;; - Local models via Ollama, Llama.cpp, Llamafiles or GPT4All
;;
;; Additionally, any LLM service (local or remote) that provides an
@@ -51,6 +51,7 @@
;; - Supports conversations and multiple independent sessions.
;; - Supports tool-use to equip LLMs with agentic capabilities.
;; - Supports multi-modal models (send images, documents).
+;; - Supports "reasoning" content in LLM responses.
;; - Save chats as regular Markdown/Org/Text files and resume them later.
;; - You can go back and edit your previous prompts or LLM responses when
;; continuing a conversation. These will be fed back to the model.
@@ -70,9 +71,11 @@
;; - For Gemini: define a gptel-backend with `gptel-make-gemini', which see.
;; - For Anthropic (Claude): define a gptel-backend with
`gptel-make-anthropic',
;; which see.
-;; - For Together.ai, Anyscale, Perplexity, Groq, OpenRouter, DeepSeek,
Cerebras or
+;; - For Together.ai, Anyscale, Groq, OpenRouter, DeepSeek, Cerebras or
;; Github Models: define a gptel-backend with `gptel-make-openai', which see.
;; - For PrivateGPT: define a backend with `gptel-make-privategpt', which see.
+;; - For Perplexity: define a backend with `gptel-make-perplexity', which see.
+;; - For Deepseek: define a backend with `gptel-make-deepseek', which see.
;; - For Kagi: define a gptel-backend with `gptel-make-kagi', which see.
;;
;; For local models using Ollama, Llama.cpp or GPT4All:
@@ -125,11 +128,12 @@
;; Include more context with requests:
;;
;; If you want to provide the LLM with more context, you can add arbitrary
-;; regions, buffers or files to the query with `gptel-add'. To add text or
-;; media files, call `gptel-add' in Dired or use the dedicated
`gptel-add-file'.
+;; regions, buffers, files or directories to the query with `gptel-add'. To
add
+;; text or media files, call `gptel-add' in Dired or use the dedicated
+;; `gptel-add-file'.
;;
-;; You can also add context from gptel's menu instead (gptel-send with a prefix
-;; arg), as well as examine or modify context.
+;; You can also add context from gptel's menu instead (`gptel-send' with a
+;; prefix arg), as well as examine or modify context.
;;
;; When context is available, gptel will include it with each LLM query.
;;
@@ -156,8 +160,9 @@
;; will always use these settings, allowing you to create mostly reproducible
;; LLM chat notebooks.
;;
-;; Finally, gptel offers a general purpose API for writing LLM ineractions
-;; that suit your workflow, see `gptel-request'.
+;; Finally, gptel offers a general purpose API for writing LLM ineractions that
+;; suit your workflow. See `gptel-request', and `gptel-fsm' for more advanced
+;; usage.
;;; Code:
(declare-function markdown-mode "markdown-mode")