From 2ef85ac794945c81fa6aa795d50f15d020cf0d32 Mon Sep 17 00:00:00 2001 From: Jeremy D <115047575+bmosaicml@users.noreply.github.com> Date: Mon, 18 Sep 2023 16:11:21 -0400 Subject: [PATCH] add comment (#606) --- llmfoundry/models/inference_api_wrapper/openai_causal_lm.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/llmfoundry/models/inference_api_wrapper/openai_causal_lm.py b/llmfoundry/models/inference_api_wrapper/openai_causal_lm.py index 14228134d2..72dfd9db76 100644 --- a/llmfoundry/models/inference_api_wrapper/openai_causal_lm.py +++ b/llmfoundry/models/inference_api_wrapper/openai_causal_lm.py @@ -30,7 +30,7 @@ class OpenAITokenizerWrapper(AutoTokenizer): - + # this API is experimental and for evaluation only. It is subject to change as we add support for training def __init__(self, name: str) -> None: try: import tiktoken