From 1e856e45479bc83f8afe6dcac4035be157d5120f Mon Sep 17 00:00:00 2001 From: Jeff Epler Date: Mon, 6 Nov 2023 21:25:57 +0100 Subject: [PATCH] Add gpt-4-1106-preview (gpt-4-turbo) to model list --- src/chap/backends/openai_chatgpt.py | 12 +++++++++--- 1 file changed, 9 insertions(+), 3 deletions(-) diff --git a/src/chap/backends/openai_chatgpt.py b/src/chap/backends/openai_chatgpt.py index a6d8fc7..f2096db 100644 --- a/src/chap/backends/openai_chatgpt.py +++ b/src/chap/backends/openai_chatgpt.py @@ -4,6 +4,7 @@ import functools import json +import warnings from dataclasses import dataclass from typing import AsyncGenerator, cast @@ -36,16 +37,20 @@ class EncodingMeta: @functools.cache def from_model(cls, model: str) -> "EncodingMeta": if model == "gpt-3.5-turbo": - # print("Warning: gpt-3.5-turbo may update over time. Returning num tokens assuming gpt-3.5-turbo-0613.") + warnings.warn( + "Warning: gpt-3.5-turbo may update over time. Returning num tokens assuming gpt-3.5-turbo-0613." + ) model = "gpt-3.5-turbo-0613" if model == "gpt-4": - # print("Warning: gpt-4 may update over time. Returning num tokens assuming gpt-4-0613.") + warnings.warn( + "Warning: gpt-4 may update over time. Returning num tokens assuming gpt-4-0613." + ) model = "gpt-4-0613" try: encoding = tiktoken.encoding_for_model(model) except KeyError: - print("Warning: model not found. Using cl100k_base encoding.") + warnings.warn("Warning: model not found. Using cl100k_base encoding.") encoding = tiktoken.get_encoding("cl100k_base") if model in { @@ -55,6 +60,7 @@ class EncodingMeta: "gpt-4-32k-0314", "gpt-4-0613", "gpt-4-32k-0613", + "gpt-4-1106-preview", }: tokens_per_message = 3 tokens_per_name = 1