Skip to content

Commit

Permalink
Merge pull request #1244 from awwaawwa/fix_gpt_35_16k_maxtoken
Browse files Browse the repository at this point in the history
修改 gpt-3.5-turbo-16k 系列模型 max_token 为 16385
  • Loading branch information
binary-husky authored Nov 10, 2023
2 parents b49b272 + 8d94564 commit caf45ef
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions request_llms/bridge_all.py
Original file line number Diff line number Diff line change
Expand Up @@ -94,7 +94,7 @@ def decode(self, *args, **kwargs):
"fn_with_ui": chatgpt_ui,
"fn_without_ui": chatgpt_noui,
"endpoint": openai_endpoint,
"max_token": 1024*16,
"max_token": 16385,
"tokenizer": tokenizer_gpt35,
"token_cnt": get_token_num_gpt35,
},
Expand All @@ -112,7 +112,7 @@ def decode(self, *args, **kwargs):
"fn_with_ui": chatgpt_ui,
"fn_without_ui": chatgpt_noui,
"endpoint": openai_endpoint,
"max_token": 1024 * 16,
"max_token": 16385,
"tokenizer": tokenizer_gpt35,
"token_cnt": get_token_num_gpt35,
},
Expand Down Expand Up @@ -204,7 +204,7 @@ def decode(self, *args, **kwargs):
"fn_with_ui": chatgpt_ui,
"fn_without_ui": chatgpt_noui,
"endpoint": api2d_endpoint,
"max_token": 1024*16,
"max_token": 16385,
"tokenizer": tokenizer_gpt35,
"token_cnt": get_token_num_gpt35,
},
Expand Down

0 comments on commit caf45ef

Please sign in to comment.