From 63e0fe926634a152f56586b63434c4739819a601 Mon Sep 17 00:00:00 2001 From: Martin Bielik Date: Fri, 31 May 2024 09:01:35 +0200 Subject: define required max_tokens for turbo-instruct model --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) (limited to 'README.md') diff --git a/README.md b/README.md index d0af65d..34fdf9d 100644 --- a/README.md +++ b/README.md @@ -257,7 +257,7 @@ generate a paragraph of lorem ipsum ``` Below are listed all available configuration options, along with their default values. -Please note that there isn't any token limit imposed, though 1000 is recommended. +Please note that there isn't any token limit imposed on chat model. ```vim " :AI @@ -272,7 +272,7 @@ let g:vim_ai_complete = { \ "options": { \ "model": "gpt-3.5-turbo-instruct", \ "endpoint_url": "https://api.openai.com/v1/completions", -\ "max_tokens": 0, +\ "max_tokens": 1000, \ "temperature": 0.1, \ "request_timeout": 20, \ "enable_auth": 1, @@ -295,7 +295,7 @@ let g:vim_ai_edit = { \ "options": { \ "model": "gpt-3.5-turbo-instruct", \ "endpoint_url": "https://api.openai.com/v1/completions", -\ "max_tokens": 0, +\ "max_tokens": 1000, \ "temperature": 0.1, \ "request_timeout": 20, \ "enable_auth": 1, -- cgit v1.2.3