* gptel.el (gptel--convert-markdown->org, gptel--stream-convert-markdown->org): Don't touch underscores in the source markdown. This will turn some emphasis blocks into underlines in Org, but we can live with that. Fix #40. |
||
---|---|---|
img | ||
.gitignore | ||
gptel-curl.el | ||
gptel-gemini.el | ||
gptel-kagi.el | ||
gptel-ollama.el | ||
gptel-openai.el | ||
gptel-transient.el | ||
gptel.el | ||
LICENSE | ||
README.org |
GPTel: A simple LLM client for Emacs
- Contents
- Installation
- Setup
- Usage
- FAQ
- I want the window to scroll automatically as the response is inserted
- I want the cursor to move to the next prompt after the response is inserted
- I want to change the formatting of the prompt and LLM response
- I want the transient menu options to be saved so I only need to set them once
- I want to use gptel in a way that's not supported by
gptel-send
or the options menu - (Doom Emacs) Sending a query from the gptel menu fails because of a key conflict with Org mode
- Why another LLM client?
- Additional Configuration
- COMMENT Will you add feature X?
- Alternatives
- Breaking Changes
- Acknowledgments
GPTel is a simple Large Language Model chat client for Emacs, with support for multiple models and backends.
LLM Backend | Supports | Requires |
---|---|---|
ChatGPT | ✓ | API key |
Azure | ✓ | Deployment and API key |
Ollama | ✓ | Ollama running locally |
GPT4All | ✓ | GPT4All running locally |
Gemini | ✓ | API key |
Llama.cpp | ✓ | Llama.cpp running locally |
Llamafile | ✓ | Local Llamafile server |
Kagi FastGPT | ✓ | API key |
Kagi Summarizer | ✓ | API key |
together.ai | ✓ | API key |
Anyscale | ✓ | API key |
General usage: (YouTube Demo)
https://user-images.githubusercontent.com/8607532/230516812-86510a09-a2fb-4cbd-b53f-cc2522d05a13.mp4
https://user-images.githubusercontent.com/8607532/230516816-ae4a613a-4d01-4073-ad3f-b66fa73c6e45.mp4
Multi-LLM support demo:
- It's async and fast, streams responses.
- Interact with LLMs from anywhere in Emacs (any buffer, shell, minibuffer, wherever)
- LLM responses are in Markdown or Org markup.
- Supports conversations and multiple independent sessions.
- Save chats as regular Markdown/Org/Text files and resume them later.
- You can go back and edit your previous prompts or LLM responses when continuing a conversation. These will be fed back to the model.
- Don't like gptel's workflow? Use it to create your own for any supported model/backend with a simple API.
GPTel uses Curl if available, but falls back to url-retrieve to work without external dependencies.
Contents toc
-
- I want the window to scroll automatically as the response is inserted
- I want the cursor to move to the next prompt after the response is inserted
- I want to change the formatting of the prompt and LLM response
- I want the transient menu options to be saved so I only need to set them once
- I want to use gptel in a way that's not supported by
gptel-send
or the options menu - (Doom Emacs) Sending a query from the gptel menu fails because of a key conflict with Org mode
- Why another LLM client?
- Additional Configuration
- Breaking Changes
- Acknowledgments
Installation
GPTel is on MELPA. Ensure that MELPA is in your list of sources, then install gptel with M-x package-install⏎
gptel
.
(Optional: Install markdown-mode
.)
Straight
Straight
(straight-use-package 'gptel)
Installing the markdown-mode
package is optional.
Manual
Clone or download this repository and run M-x package-install-file⏎
on the repository directory.
Installing the markdown-mode
package is optional.
Doom Emacs
In packages.el
(package! gptel)
In config.el
(use-package! gptel
:config
(setq! gptel-api-key "your key"))
"your key" can be the API key itself, or (safer) a function that returns the key. Setting gptel-api-key
is optional, you will be asked for a key if it's not found.
Spacemacs
After installation with M-x package-install⏎
gptel
- Add
gptel
todotspacemacs-additional-packages
- Add
(require 'gptel)
todotspacemacs/user-config
Setup
ChatGPT
Procure an OpenAI API key.
Optional: Set gptel-api-key
to the key. Alternatively, you may choose a more secure method such as:
-
Storing in
~/.authinfo
. By default, "api.openai.com" is used as HOST and "apikey" as USER.machine api.openai.com login apikey password TOKEN
- Setting it to a function that returns the key.
Other LLM backends
Azure
Azure
Register a backend with
(gptel-make-azure "Azure-1" ;Name, whatever you'd like
:protocol "https" ;Optional -- https is the default
:host "YOUR_RESOURCE_NAME.openai.azure.com"
:endpoint "/openai/deployments/YOUR_DEPLOYMENT_NAME/chat/completions?api-version=2023-05-15" ;or equivalent
:stream t ;Enable streaming responses
:key #'gptel-api-key
:models '("gpt-3.5-turbo" "gpt-4"))
Refer to the documentation of gptel-make-azure
to set more parameters.
You can pick this backend from the menu when using gptel. (see Usage)
If you want it to be the default, set it as the default value of gptel-backend
:
(setq-default gptel-backend (gptel-make-azure "Azure-1" ...)
gptel-model "gpt-3.5-turbo")
GPT4All
Register a backend with
(gptel-make-gpt4all "GPT4All" ;Name of your choosing
:protocol "http"
:host "localhost:4891" ;Where it's running
:models '("mistral-7b-openorca.Q4_0.gguf")) ;Available models
These are the required parameters, refer to the documentation of gptel-make-gpt4all
for more.
You can pick this backend from the menu when using gptel (see Usage), or set this as the default value of gptel-backend
. Additionally you may want to increase the response token size since GPT4All uses very short (often truncated) responses by default:
;; OPTIONAL configuration
(setq-default gptel-model "mistral-7b-openorca.Q4_0.gguf" ;Pick your default model
gptel-backend (gptel-make-gpt4all "GPT4All" :protocol ...))
(setq-default gptel-max-tokens 500)
Ollama
Register a backend with
(gptel-make-ollama "Ollama" ;Any name of your choosing
:host "localhost:11434" ;Where it's running
:stream t ;Stream responses
:models '("mistral:latest")) ;List of models
These are the required parameters, refer to the documentation of gptel-make-ollama
for more.
You can pick this backend from the menu when using gptel (see Usage), or set this as the default value of gptel-backend
:
;; OPTIONAL configuration
(setq-default gptel-model "mistral:latest" ;Pick your default model
gptel-backend (gptel-make-ollama "Ollama" :host ...))
Gemini
Register a backend with
;; :key can be a function that returns the API key.
(gptel-make-gemini "Gemini"
:key "YOUR_GEMINI_API_KEY"
:stream t)
These are the required parameters, refer to the documentation of gptel-make-gemini
for more.
You can pick this backend from the menu when using gptel (see Usage), or set this as the default value of gptel-backend
:
;; OPTIONAL configuration
(setq-default gptel-model "gemini-pro" ;Pick your default model
gptel-backend (gptel-make-gemini "Gemini" :host ...))
Llama.cpp or Llamafile
(If using a llamafile, run a server llamafile instead of a "command-line llamafile", and a model that supports text generation.)
Register a backend with
;; Llama.cpp offers an OpenAI compatible API
(gptel-make-openai "llama-cpp" ;Any name
:stream t ;Stream responses
:protocol "http"
:host "localhost:8000" ;Llama.cpp server location
:models '("test")) ;Any names, doesn't matter for Llama
These are the required parameters, refer to the documentation of gptel-make-openai
for more.
You can pick this backend from the menu when using gptel (see Usage), or set this as the default value of gptel-backend
:
;; OPTIONAL configuration
(setq-default gptel-backend (gptel-make-openai "llama-cpp" ...)
gptel-model "test")
Kagi (FastGPT & Summarizer)
Kagi's FastGPT model and the Universal Summarizer are both supported. A couple of notes:
- Universal Summarizer: If there is a URL at point, the summarizer will summarize the contents of the URL. Otherwise the context sent to the model is the same as always: the buffer text upto point, or the contents of the region if the region is active.
- Kagi models do not support multi-turn conversations, interactions are "one-shot". They also do not support streaming responses.
Register a backend with
(gptel-make-kagi "Kagi" ;any name
:key "YOUR_KAGI_API_KEY") ;can be a function that returns the key
These are the required parameters, refer to the documentation of gptel-make-kagi
for more.
You can pick this backend and the model (fastgpt/summarizer) from the transient menu when using gptel. Alternatively you can set this as the default value of gptel-backend
:
;; OPTIONAL configuration
(setq-default gptel-model "fastgpt"
gptel-backend (gptel-make-kagi "Kagi" :key ...))
The alternatives to fastgpt
include summarize:cecil
, summarize:agnes
, summarize:daphne
and summarize:muriel
. The difference between the summarizer engines is documented here.
together.ai
Register a backend with
;; Together.ai offers an OpenAI compatible API
(gptel-make-openai "TogetherAI" ;Any name you want
:host "api.together.xyz"
:key "your-api-key" ;can be a function that returns the key
:stream t
:models '(;; has many more, check together.ai
"mistralai/Mixtral-8x7B-Instruct-v0.1"
"codellama/CodeLlama-13b-Instruct-hf"
"codellama/CodeLlama-34b-Instruct-hf"))
You can pick this backend from the menu when using gptel (see Usage), or set this as the default value of gptel-backend
:
;; OPTIONAL configuration
(setq-default gptel-backend (gptel-make-openai "TogetherAI" ...)
gptel-model "mistralai/Mixtral-8x7B-Instruct-v0.1")
Anyscale
Register a backend with
;; Anyscale offers an OpenAI compatible API
(gptel-make-openai "Anyscale" ;Any name you want
:host "api.endpoints.anyscale.com"
:key "your-api-key" ;can be a function that returns the key
:models '(;; has many more, check anyscale
"mistralai/Mixtral-8x7B-Instruct-v0.1"))
You can pick this backend from the menu when using gptel (see Usage), or set this as the default value of gptel-backend
:
;; OPTIONAL configuration
(setq-default gptel-backend (gptel-make-openai "Anyscale" ...)
gptel-model "mistralai/Mixtral-8x7B-Instruct-v0.1")
Usage
(This is also a video demo showing various uses of gptel.)
Command | Description |
gptel-send |
Send conversation up to (point) , or selection if region is active. Works anywhere in Emacs. |
gptel |
Create a new dedicated chat buffer. Not required to use gptel. |
C-u gptel-send |
Transient menu for preferences, input/output redirection etc. |
gptel-menu |
(Same) |
gptel-set-topic |
(Org-mode only) Limit conversation context to an Org heading |
In any buffer:
- Call
M-x gptel-send
to send the text up to the cursor. The response will be inserted below. Continue the conversation by typing below the response. - If a region is selected, the conversation will be limited to its contents.
- Call
M-x gptel-send
with a prefix argument to
- set chat parameters (GPT model, directives etc) for this buffer,
- to read the prompt from elsewhere or redirect the response elsewhere,
- or to replace the prompt with the response.
With a region selected, you can also rewrite prose or refactor code from here:
Code:
Prose:
In a dedicated chat buffer:
- Run
M-x gptel
to start or switch to the chat buffer. It will ask you for the key if you skipped the previous step. Run it with a prefix-arg (C-u M-x gptel
) to start a new session. - In the gptel buffer, send your prompt with
M-x gptel-send
, bound toC-c RET
. - Set chat parameters (LLM provider, model, directives etc) for the session by calling
gptel-send
with a prefix argument (C-u C-c RET
):
That's it. You can go back and edit previous prompts and responses if you want.
The default mode is markdown-mode
if available, else text-mode
. You can set gptel-default-mode
to org-mode
if desired.
Save and restore your chat sessions
Saving the file will save the state of the conversation as well. To resume the chat, open the file and turn on gptel-mode
before editing the buffer.
FAQ
I want the window to scroll automatically as the response is inserted
I want the window to scroll automatically as the response is inserted
To be minimally annoying, GPTel does not move the cursor by default. Add the following to your configuration to enable auto-scrolling.
(add-hook 'gptel-post-stream-hook 'gptel-auto-scroll)
I want the cursor to move to the next prompt after the response is inserted
To be minimally annoying, GPTel does not move the cursor by default. Add the following to your configuration to move the cursor:
(add-hook 'gptel-post-response-functions 'gptel-end-of-response)
You can also call gptel-end-of-response
as a command at any time.
I want to change the formatting of the prompt and LLM response
For dedicated chat buffers: customize gptel-prompt-prefix-alist
and gptel-response-prefix-alist
. You can set a different pair for each major-mode.
Anywhere in Emacs: Use gptel-pre-response-hook
and gptel-post-response-functions
, which see.
I want the transient menu options to be saved so I only need to set them once
Any model options you set are saved for the current buffer. But the redirection options in the menu are set for the next query only:
https://github.com/karthink/gptel/assets/8607532/2ecc6be9-aa52-4287-a739-ba06e1369ec2
You can make them persistent across this Emacs session by pressing C-x C-s
:
https://github.com/karthink/gptel/assets/8607532/b8bcb6ad-c974-41e1-9336-fdba0098a2fe
(You can also cycle through presets you've saved with C-x p
and C-x n
.)
Now these will be enabled whenever you send a query from the transient menu. If you want to use these options without invoking the transient menu, you can use a keyboard macro:
;; Replace with your key to invoke the transient menu:
(keymap-global-set "<f6>" "C-u C-c <return> <return>")
See this comment by Tianshu Wang for an Elisp solution.
I want to use gptel in a way that's not supported by gptel-send
or the options menu
GPTel's default usage pattern is simple, and will stay this way: Read input in any buffer and insert the response below it. Some custom behavior is possible with the transient menu (C-u M-x gptel-send
).
For more programmable usage, gptel provides a general gptel-request
function that accepts a custom prompt and a callback to act on the response. You can use this to build custom workflows not supported by gptel-send
. See the documentation of gptel-request
, and the wiki for examples.
(Doom Emacs) Sending a query from the gptel menu fails because of a key conflict with Org mode
Doom binds RET
in Org mode to +org/dwim-at-point
, which appears to conflict with gptel's transient menu bindings for some reason.
Two solutions:
- Press
C-m
instead of the return key. -
Change the send key from return to a key of your choice:
(transient-suffix-put 'gptel-menu (kbd "RET") :key "<f8>")
Why another LLM client?
Other Emacs clients for LLMs prescribe the format of the interaction (a comint shell, org-babel blocks, etc). I wanted:
- Something that is as free-form as possible: query the model using any text in any buffer, and redirect the response as required. Using a dedicated
gptel
buffer just adds some visual flair to the interaction. - Integration with org-mode, not using a walled-off org-babel block, but as regular text. This way the model can generate code blocks that I can run.
Additional Configuration
Connection options | |
gptel-use-curl |
Use Curl (default), fallback to Emacs' built-in url . |
gptel-proxy |
Proxy server for requests, passed to curl via --proxy . |
gptel-api-key |
Variable/function that returns the API key for the active backend. |
LLM options | (Note: not supported uniformly across LLMs) |
gptel-backend |
Default LLM Backend. |
gptel-model |
Default model to use, depends on the backend. |
gptel-stream |
Enable streaming responses, if the backend supports it. |
gptel-directives |
Alist of system directives, can switch on the fly. |
gptel-max-tokens |
Maximum token count (in query + response). |
gptel-temperature |
Randomness in response text, 0 to 2. |
Chat UI options | |
gptel-default-mode |
Major mode for dedicated chat buffers. |
gptel-prompt-prefix-alist |
Text inserted before queries. |
gptel-response-prefix-alist |
Text inserted before responses. |
gptel-use-header-line |
Display status messages in header-line (default) or minibuffer |
COMMENT Will you add feature X?
Maybe, I'd like to experiment a bit more first. Features added since the inception of this package include
- Curl support (
gptel-use-curl
) - Streaming responses (
gptel-stream
) - Cancelling requests in progress (
gptel-abort
) - General API for writing your own commands (
gptel-request
, wiki) - Dispatch menus using Transient (
gptel-send
with a prefix arg) - Specifying the conversation context size
- GPT-4 support
- Response redirection (to the echo area, another buffer, etc)
- A built-in refactor/rewrite prompt
- Limiting conversation context to Org headings using properties (#58)
- Saving and restoring chats (#17)
- Support for local LLMs.
Features being considered or in the pipeline:
- Fully stateless design (#17)
Alternatives
Other Emacs clients for LLMs include
- chatgpt-shell: comint-shell based interaction with ChatGPT. Also supports DALL-E, executable code blocks in the responses, and more.
- org-ai: Interaction through special
#+begin_ai ... #+end_ai
Org-mode blocks. Also supports DALL-E, querying ChatGPT with the contents of project files, and more.
There are several more: chatgpt-arcana, leafy-mode, chat.el
Extensions using GPTel
These are packages that depend on GPTel to provide additional functionality
- gptel-extensions: Extra utility functions for GPTel.
- ai-blog.el: Streamline generation of blog posts in Hugo.
Breaking Changes
gptel-post-response-hook
has been renamed togptel-post-response-functions
, and functions in this hook are now called with two arguments: the start and end buffer positions of the response. This should make it easy to act on the response text without having to locate it first.- Possible breakage, see #120: If streaming responses stop working for you after upgrading to v0.5, try reinstalling gptel and deleting its native comp eln cache in
native-comp-eln-load-path
. - The user option
gptel-host
is deprecated. If the defaults don't work for you, usegptel-make-openai
(which see) to customize server settings. gptel-api-key-from-auth-source
now searches for the API key using the host address for the active LLM backend, i.e. "api.openai.com" when using ChatGPT. You may need to update your~/.authinfo
.
Acknowledgments
- Alexis Gallagher and Diego Alvarez for fixing a nasty multi-byte bug with
url-retrieve
. - Jonas Bernoulli for the Transient library.