File tree Expand file tree Collapse file tree 2 files changed +17
-2
lines changed Expand file tree Collapse file tree 2 files changed +17
-2
lines changed Original file line number Diff line number Diff line change 2
2
from ..utils .merge_deltas import merge_deltas
3
3
from ..utils .parse_partial_json import parse_partial_json
4
4
from ..utils .convert_to_openai_messages import convert_to_openai_messages
5
+ from ..utils .display_markdown_message import display_markdown_message
5
6
import tokentrim as tt
6
7
7
8
@@ -47,7 +48,17 @@ def coding_llm(messages):
47
48
messages = messages [1 :]
48
49
49
50
# Trim messages, preserving the system_message
50
- messages = tt .trim (messages = messages , system_message = system_message , model = interpreter .model )
51
+ try :
52
+ messages = tt .trim (messages = messages , system_message = system_message , model = interpreter .model )
53
+ except :
54
+ if interpreter .context_window :
55
+ messages = tt .trim (messages = messages , system_message = system_message , max_tokens = interpreter .context_window )
56
+ else :
57
+ display_markdown_message ("""
58
+ **We were unable to determine the context window of this model.** Defaulting to 3000.
59
+ If your model can handle more, run `interpreter --context_window {token limit}` or `interpreter.context_window = {token limit}`.
60
+ """ )
61
+ messages = tt .trim (messages = messages , system_message = system_message , max_tokens = 3000 )
51
62
52
63
if interpreter .debug_mode :
53
64
print ("Sending this to the OpenAI LLM:" , messages )
Original file line number Diff line number Diff line change @@ -79,7 +79,11 @@ def base_llm(messages):
79
79
try :
80
80
messages = tt .trim (messages , system_message = system_message , model = interpreter .model )
81
81
except :
82
- # If we don't know the model, just do 3000.
82
+ display_markdown_message ("""
83
+ **We were unable to determine the context window of this model.** Defaulting to 3000.
84
+ If your model can handle more, run `interpreter --context_window {token limit}` or `interpreter.context_window = {token limit}`.
85
+ Also, please set max_tokens: `interpreter --max_tokens {max tokens per response}` or `interpreter.max_tokens = {max tokens per response}`
86
+ """ )
83
87
messages = tt .trim (messages , system_message = system_message , max_tokens = 3000 )
84
88
85
89
if interpreter .debug_mode :
You can’t perform that action at this time.
0 commit comments