From ea83fdc59b64843f570c94d34d97c3131663d3e8 Mon Sep 17 00:00:00 2001 From: Martin Bielik Date: Sun, 8 Dec 2024 21:15:53 +0100 Subject: print prompt in debug mode --- py/chat.py | 1 + py/complete.py | 2 ++ 2 files changed, 3 insertions(+) diff --git a/py/chat.py b/py/chat.py index a3729f1..e1bbc13 100644 --- a/py/chat.py +++ b/py/chat.py @@ -50,6 +50,7 @@ initial_prompt = '\n'.join(options.get('initial_prompt', [])) initial_messages = parse_chat_messages(initial_prompt) chat_content = vim.eval('trim(join(getline(1, "$"), "\n"))') +printDebug("[chat] text:\n" + chat_content) chat_messages = parse_chat_messages(chat_content) is_selection = vim.eval("l:is_selection") diff --git a/py/complete.py b/py/complete.py index d7040a3..0580257 100644 --- a/py/complete.py +++ b/py/complete.py @@ -14,6 +14,7 @@ is_selection = vim.eval("l:is_selection") def complete_engine(prompt): openai_options = make_openai_options(config_options) http_options = make_http_options(config_options) + printDebug("[engine-complete] text:\n" + prompt) request = { 'prompt': prompt, @@ -33,6 +34,7 @@ def chat_engine(prompt): initial_prompt = '\n'.join(initial_prompt) chat_content = f"{initial_prompt}\n\n>>> user\n\n{prompt}".strip() messages = parse_chat_messages(chat_content) + printDebug("[engine-chat] text:\n" + chat_content) return make_chat_text_chunks(messages, config_options) engines = {"chat": chat_engine, "complete": complete_engine} -- cgit v1.2.3