From 2d644e05545be6cd699eb5e834d6ba4468f12b41 Mon Sep 17 00:00:00 2001 From: Martin Bielik Date: Tue, 11 Apr 2023 17:30:16 +0200 Subject: added debug logging --- py/complete.py | 26 ++++++++++++++++++++++---- 1 file changed, 22 insertions(+), 4 deletions(-) (limited to 'py/complete.py') diff --git a/py/complete.py b/py/complete.py index b0b0dc5..eea2add 100644 --- a/py/complete.py +++ b/py/complete.py @@ -13,8 +13,17 @@ prompt = vim.eval("prompt").strip() openai.api_key = load_api_key() def complete_engine(prompt): - response = openai.Completion.create(stream=True, prompt=prompt, **request_options) - text_chunks = map(lambda resp: resp['choices'][0].get('text', ''), response) + request = { + 'stream': True, + 'prompt': prompt, + **request_options + } + printDebug("[engine-complete] request: {}", request) + response = openai.Completion.create(**request) + def map_chunk(resp): + printDebug("[engine-complete] response: {}", resp) + return resp['choices'][0].get('text', '') + text_chunks = map(map_chunk, response) return text_chunks def chat_engine(prompt): @@ -22,8 +31,17 @@ def chat_engine(prompt): initial_prompt = '\n'.join(initial_prompt) chat_content = f"{initial_prompt}\n\n>>> user\n\n{prompt}".strip() messages = parse_chat_messages(chat_content) - response = openai.ChatCompletion.create(messages=messages, stream=True, **request_options) - text_chunks = map(lambda resp: resp['choices'][0]['delta'].get('content', ''), response) + request = { + 'stream': True, + 'messages': messages, + **request_options + } + printDebug("[engine-chat] request: {}", request) + response = openai.ChatCompletion.create(**request) + def map_chunk(resp): + printDebug("[engine-chat] response: {}", resp) + return resp['choices'][0]['delta'].get('content', '') + text_chunks = map(map_chunk, response) return text_chunks engines = {"chat": chat_engine, "complete": complete_engine} -- cgit v1.2.3