README: Tweak instructions for local LLMs, mention #120
This commit is contained in:
parent
63027083cd
commit
50a2498259
1 changed files with 29 additions and 14 deletions
43
README.org
43
README.org
|
@ -4,14 +4,14 @@
|
||||||
|
|
||||||
GPTel is a simple Large Language Model chat client for Emacs, with support for multiple models/backends.
|
GPTel is a simple Large Language Model chat client for Emacs, with support for multiple models/backends.
|
||||||
|
|
||||||
| LLM Backend | Supports | Requires |
|
| LLM Backend | Supports | Requires |
|
||||||
|-------------+----------+------------------------|
|
|-------------+----------+-------------------------|
|
||||||
| ChatGPT | ✓ | [[https://platform.openai.com/account/api-keys][API key]] |
|
| ChatGPT | ✓ | [[https://platform.openai.com/account/api-keys][API key]] |
|
||||||
| Azure | ✓ | Deployment and API key |
|
| Azure | ✓ | Deployment and API key |
|
||||||
| Ollama | ✓ | An LLM running locally |
|
| Ollama | ✓ | [[https://ollama.ai/][Ollama running locally]] |
|
||||||
| GPT4All | ✓ | An LLM running locally |
|
| GPT4All | ✓ | [[https://gpt4all.io/index.html][GPT4All running locally]] |
|
||||||
| PrivateGPT | Planned | - |
|
| PrivateGPT | Planned | - |
|
||||||
| Llama.cpp | Planned | - |
|
| Llama.cpp | Planned | - |
|
||||||
|
|
||||||
*General usage*:
|
*General usage*:
|
||||||
|
|
||||||
|
@ -59,6 +59,8 @@ GPTel uses Curl if available, but falls back to url-retrieve to work without ext
|
||||||
|
|
||||||
** Breaking Changes
|
** Breaking Changes
|
||||||
|
|
||||||
|
- Possible breakage, see #120: If streaming responses stop working for you after upgrading to v0.5, try reinstalling gptel and deleting its native comp eln cache in =native-comp-eln-load-path=.
|
||||||
|
|
||||||
- The user option =gptel-host= is deprecated. If the defaults don't work for you, use =gptel-make-openai= (which see) to customize server settings.
|
- The user option =gptel-host= is deprecated. If the defaults don't work for you, use =gptel-make-openai= (which see) to customize server settings.
|
||||||
|
|
||||||
- =gptel-api-key-from-auth-source= now searches for the API key using the host address for the active LLM backend, /i.e./ "api.openai.com" when using ChatGPT. You may need to update your =~/.authinfo=.
|
- =gptel-api-key-from-auth-source= now searches for the API key using the host address for the active LLM backend, /i.e./ "api.openai.com" when using ChatGPT. You may need to update your =~/.authinfo=.
|
||||||
|
@ -162,7 +164,14 @@ Register a backend with
|
||||||
#+end_src
|
#+end_src
|
||||||
These are the required parameters, refer to the documentation of =gptel-make-gpt4all= for more.
|
These are the required parameters, refer to the documentation of =gptel-make-gpt4all= for more.
|
||||||
|
|
||||||
You can pick this backend from the transient menu when using gptel (see usage), or set this as the default value of =gptel-backend=.
|
You can pick this backend from the transient menu when using gptel (see usage), or set this as the default value of =gptel-backend=. Additionally you may want to increase the response token size since GPT4All uses very short (often truncated) responses by default:
|
||||||
|
|
||||||
|
#+begin_src emacs-lisp
|
||||||
|
;; OPTIONAL configuration
|
||||||
|
(setq-default gptel-model "mistral-7b-openorca.Q4_0.gguf" ;Pick your default model
|
||||||
|
gptel-backend (gptel-make-gpt4all "GPT4All" :protocol ...))
|
||||||
|
(setq-default gptel-max-tokens 500)
|
||||||
|
#+end_src
|
||||||
|
|
||||||
#+html: </details>
|
#+html: </details>
|
||||||
|
|
||||||
|
@ -178,19 +187,25 @@ Register a backend with
|
||||||
:models '("mistral:latest") ;Installed models
|
:models '("mistral:latest") ;Installed models
|
||||||
:stream t) ;Stream responses
|
:stream t) ;Stream responses
|
||||||
#+end_src
|
#+end_src
|
||||||
These are the required parameters, refer to the documentation of =gptel-make-gpt4all= for more.
|
These are the required parameters, refer to the documentation of =gptel-make-ollama= for more.
|
||||||
|
|
||||||
You can pick this backend from the transient menu when using gptel (see usage), or set this as the default value of =gptel-backend=.
|
You can pick this backend from the transient menu when using gptel (see Usage), or set this as the default value of =gptel-backend=:
|
||||||
|
|
||||||
|
#+begin_src emacs-lisp
|
||||||
|
;; OPTIONAL configuration
|
||||||
|
(setq-default gptel-model "mistral:latest" ;Pick your default model
|
||||||
|
gptel-backend (gptel-make-ollama "Ollama" :host ...))
|
||||||
|
#+end_src
|
||||||
|
|
||||||
#+html: </details>
|
#+html: </details>
|
||||||
|
|
||||||
** Usage
|
** Usage
|
||||||
|
|
||||||
|-------------------+-------------------------------------------------------------------------|
|
|-------------------+-------------------------------------------------------------------------|
|
||||||
| *Commands* | Description |
|
| *Command* | Description |
|
||||||
|-------------------+-------------------------------------------------------------------------|
|
|-------------------+-------------------------------------------------------------------------|
|
||||||
| =gptel= | Create a new dedicated chat buffer. Not required, gptel works anywhere. |
|
| =gptel= | Create a new dedicated chat buffer. (Not required, gptel works anywhere.) |
|
||||||
| =gptel-send= | Send selection, or conversation up to =(point)=. Works anywhere in Emacs. |
|
| =gptel-send= | Send selection, or conversation up to =(point)=. (Works anywhere in Emacs.) |
|
||||||
| =C-u= =gptel-send= | Transient menu for preferenes, input/output redirection etc. |
|
| =C-u= =gptel-send= | Transient menu for preferenes, input/output redirection etc. |
|
||||||
| =gptel-menu= | /(Same)/ |
|
| =gptel-menu= | /(Same)/ |
|
||||||
|-------------------+-------------------------------------------------------------------------|
|
|-------------------+-------------------------------------------------------------------------|
|
||||||
|
|
Loading…
Add table
Reference in a new issue