|
1 | 1 | local curl = require("plenary.curl")
|
2 |
| -local Render = require("codegpt.template_render") |
3 | 2 | local Utils = require("codegpt.utils")
|
4 | 3 | local Api = require("codegpt.api")
|
5 | 4 | local Config = require("codegpt.config")
|
6 | 5 | local tokens = require("codegpt.tokens")
|
7 | 6 | local errors = require("codegpt.errors")
|
| 7 | +local Messages = require("codegpt.providers.messages") |
8 | 8 |
|
9 | 9 | local M = {}
|
10 | 10 |
|
11 |
| ----@param command string |
12 |
| ----@param cmd_opts codegpt.CommandOpts |
13 |
| ----@param command_args string[] |
14 |
| -local function generate_messages(command, cmd_opts, command_args, text_selection) |
15 |
| - local system_message = |
16 |
| - Render.render(command, cmd_opts.system_message_template, command_args, text_selection, cmd_opts) |
17 |
| - local user_message = Render.render(command, cmd_opts.user_message_template, command_args, text_selection, cmd_opts) |
18 |
| - if cmd_opts.append_string then |
19 |
| - user_message = user_message .. " " .. cmd_opts.append_string |
20 |
| - end |
21 |
| - |
22 |
| - local messages = {} |
23 |
| - if system_message ~= nil and system_message ~= "" then |
24 |
| - table.insert(messages, { role = "system", content = system_message }) |
25 |
| - end |
26 |
| - |
27 |
| - if user_message ~= nil and user_message ~= "" then |
28 |
| - table.insert(messages, { role = "user", content = user_message }) |
29 |
| - end |
30 |
| - |
31 |
| - return messages |
32 |
| -end |
33 |
| - |
34 | 11 | local function get_max_tokens(max_tokens, prompt)
|
35 | 12 | local total_length = tokens.get_tokens(prompt)
|
36 | 13 |
|
|
48 | 25 | ---@param is_stream? boolean
|
49 | 26 | function M.make_request(command, cmd_opts, command_args, text_selection, is_stream)
|
50 | 27 | local models = require("codegpt.models")
|
51 |
| - local messages = generate_messages(command, cmd_opts, command_args, text_selection) |
| 28 | + local messages = Messages.generate_messages(command, cmd_opts, command_args, text_selection) |
52 | 29 |
|
53 | 30 | -- max # of tokens to generate
|
54 | 31 | local max_tokens = get_max_tokens(cmd_opts.max_tokens, messages)
|
|
0 commit comments