-
Notifications
You must be signed in to change notification settings - Fork 162
/
gptel-ollama.el
277 lines (233 loc) · 10.5 KB
/
gptel-ollama.el
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
;;; gptel-ollama.el --- Ollama support for gptel -*- lexical-binding: t; -*-
;; Copyright (C) 2023 Karthik Chikmagalur
;; Author: Karthik Chikmagalur <[email protected]>
;; Keywords: hypermedia
;; This program is free software; you can redistribute it and/or modify
;; it under the terms of the GNU General Public License as published by
;; the Free Software Foundation, either version 3 of the License, or
;; (at your option) any later version.
;; This program is distributed in the hope that it will be useful,
;; but WITHOUT ANY WARRANTY; without even the implied warranty of
;; MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
;; GNU General Public License for more details.
;; You should have received a copy of the GNU General Public License
;; along with this program. If not, see <https://www.gnu.org/licenses/>.
;;; Commentary:
;; This file adds support for the Ollama LLM API to gptel
;;; Code:
(require 'gptel)
(require 'cl-generic)
(declare-function json-read "json" ())
(declare-function gptel-context--wrap "gptel-context")
(declare-function gptel-context--collect-media "gptel-context")
(defvar json-object-type)
;;; Ollama
(cl-defstruct (gptel-ollama (:constructor gptel--make-ollama)
(:copier nil)
(:include gptel-backend)))
(defvar-local gptel--ollama-token-count 0
"Token count for ollama conversations.
This variable holds the total token count for conversations with
Ollama models.
Intended for internal use only.")
(cl-defmethod gptel-curl--parse-stream ((_backend gptel-ollama) info)
"Parse response stream for the Ollama API."
(when (and (bobp) (re-search-forward "^{" nil t))
(forward-line 0))
(let* ((content-strs) (content) (pt (point)))
(condition-case nil
(while (setq content (gptel--json-read))
(setq pt (point))
(let ((done (map-elt content :done))
(response (map-nested-elt content '(:message :content))))
(push response content-strs)
(unless (eq done :json-false)
(with-current-buffer (plist-get info :buffer)
(cl-incf gptel--ollama-token-count
(+ (or (map-elt content :prompt_eval_count) 0)
(or (map-elt content :eval_count) 0))))
(goto-char (point-max)))))
(error (goto-char pt)))
(apply #'concat (nreverse content-strs))))
(cl-defmethod gptel--parse-response ((_backend gptel-ollama) response info)
"Parse a one-shot RESPONSE from the Ollama API."
(when-let ((context
(+ (or (map-elt response :prompt_eval_count) 0)
(or (map-elt response :eval_count) 0))))
(with-current-buffer (plist-get info :buffer)
(cl-incf gptel--ollama-token-count context)))
(map-nested-elt response '(:message :content)))
(cl-defmethod gptel--request-data ((_backend gptel-ollama) prompts)
"JSON encode PROMPTS for sending to ChatGPT."
(when (and gptel--system-message
(not (gptel--model-capable-p 'nosystem)))
(push (list :role "system"
:content gptel--system-message)
prompts))
(let* ((prompts-plist
(gptel--merge-plists
`(:model ,(gptel--model-name gptel-model)
:messages [,@prompts]
:stream ,(or (and gptel-stream gptel-use-curl
(gptel-backend-stream gptel-backend))
:json-false))
(gptel-backend-request-params gptel-backend)
(gptel--model-request-params gptel-model)))
;; the initial options (if any) from request params
(options-plist (plist-get prompts-plist :options)))
;; if the temperature and max-tokens aren't set as
;; backend/model-specific, use the global settings
(when (and gptel-temperature (not (plist-get options-plist :temperature)))
(setq options-plist
(plist-put options-plist :temperature gptel-temperature)))
(when (and gptel-max-tokens (not (plist-get options-plist :num_predict)))
(setq options-plist
(plist-put options-plist :num_predict gptel-max-tokens)))
(plist-put prompts-plist :options options-plist)))
(cl-defmethod gptel--parse-list ((_backend gptel-ollama) prompt-list)
(cl-loop for text in prompt-list
for role = t then (not role)
if text collect
(list :role (if role "user" "assistant") :content text)))
(cl-defmethod gptel--parse-buffer ((_backend gptel-ollama) &optional max-entries)
(let ((prompts) (prop)
(include-media (and gptel-track-media (or (gptel--model-capable-p 'media)
(gptel--model-capable-p 'url)))))
(if (or gptel-mode gptel-track-response)
(while (and
(or (not max-entries) (>= max-entries 0))
(setq prop (text-property-search-backward
'gptel 'response
(when (get-char-property (max (point-min) (1- (point)))
'gptel)
t))))
(if (prop-match-value prop) ;assistant role
(push (list :role "assistant"
:content (buffer-substring-no-properties (prop-match-beginning prop)
(prop-match-end prop)))
prompts)
(if include-media
(push (append '(:role "user")
(gptel--ollama-parse-multipart
(gptel--parse-media-links
major-mode (prop-match-beginning prop) (prop-match-end prop))))
prompts)
(push (list :role "user"
:content
(gptel--trim-prefixes
(buffer-substring-no-properties (prop-match-beginning prop)
(prop-match-end prop))))
prompts)))
(and max-entries (cl-decf max-entries)))
(push (list :role "user"
:content
(string-trim (buffer-substring-no-properties (point-min) (point-max))))
prompts))
prompts))
(defun gptel--ollama-parse-multipart (parts)
"Convert a multipart prompt PARTS to the Ollama API format.
The input is an alist of the form
((:text \"some text\")
(:media \"/path/to/media.png\" :mime \"image/png\")
(:text \"More text\")).
The output is a vector of entries in a backend-appropriate
format."
(cl-loop
for part in parts
for n upfrom 1
with last = (length parts)
for text = (plist-get part :text)
for media = (plist-get part :media)
if text do
(and (or (= n 1) (= n last)) (setq text (gptel--trim-prefixes text))) and
unless (string-empty-p text)
collect text into text-array end
else if media
collect (gptel--base64-encode media) into media-array end
finally return
`(,@(and text-array (list :content (mapconcat #'identity text-array " ")))
,@(and media-array (list :images (vconcat media-array))))))
(cl-defmethod gptel--wrap-user-prompt ((_backend gptel-ollama) prompts
&optional inject-media)
"Wrap the last user prompt in PROMPTS with the context string.
If INJECT-MEDIA is non-nil wrap it with base64-encoded media files in the context."
(if inject-media
;; Wrap the first user prompt with included media files/contexts
(when-let* ((media-list (gptel-context--collect-media))
(media-processed (gptel--ollama-parse-multipart media-list)))
(cl-callf (lambda (images)
(vconcat (plist-get media-processed :images)
images))
(plist-get (cadr prompts) :images)))
;; Wrap the last user prompt with included text contexts
(cl-callf gptel-context--wrap (plist-get (car (last prompts)) :content))))
;;;###autoload
(cl-defun gptel-make-ollama
(name &key curl-args header key models stream request-params
(host "localhost:11434")
(protocol "http")
(endpoint "/api/chat"))
"Register an Ollama backend for gptel with NAME.
Keyword arguments:
CURL-ARGS (optional) is a list of additional Curl arguments.
HOST is where Ollama runs (with port), defaults to localhost:11434
MODELS is a list of available model names, as symbols.
Additionally, you can specify supported LLM capabilities like
vision or tool-use by appending a plist to the model with more
information, in the form
(model-name . plist)
Currently recognized plist keys are :description, :capabilities
and :mime-types. An example of a model specification including
both kinds of specs:
:models
\\='(mistral:latest ;Simple specs
openhermes:latest
(llava:13b ;Full spec
:description
\"Llava 1.6: Large Lanuage and Vision Assistant\"
:capabilities (media)
:mime-types (\"image/jpeg\" \"image/png\")))
STREAM is a boolean to toggle streaming responses, defaults to
false.
PROTOCOL (optional) specifies the protocol, http by default.
ENDPOINT (optional) is the API endpoint for completions, defaults to
\"/api/generate\".
HEADER (optional) is for additional headers to send with each
request. It should be an alist or a function that retuns an
alist, like:
((\"Content-Type\" . \"application/json\"))
KEY (optional) is a variable whose value is the API key, or
function that returns the key. This is typically not required
for local models like Ollama.
REQUEST-PARAMS (optional) is a plist of additional HTTP request
parameters (as plist keys) and values supported by the API. Use
these to set parameters that gptel does not provide user options
for.
Example:
-------
(gptel-make-ollama
\"Ollama\"
:host \"localhost:11434\"
:models \\='(mistral:latest)
:stream t)"
(declare (indent 1))
(let ((backend (gptel--make-ollama
:curl-args curl-args
:name name
:host host
:header header
:key key
:models (gptel--process-models models)
:protocol protocol
:endpoint endpoint
:stream stream
:request-params request-params
:url (if protocol
(concat protocol "://" host endpoint)
(concat host endpoint)))))
(prog1 backend
(setf (alist-get name gptel--known-backends
nil nil #'equal)
backend))))
(provide 'gptel-ollama)
;;; gptel-ollama.el ends here