From d4567b5a4ae34700407e666a7f7847ff5bb33749 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=88=98=E5=AE=87?= <506610466@qq.com> Date: Sat, 7 Dec 2024 16:27:04 +0800 Subject: [PATCH] =?UTF-8?q?=E6=9B=B4=E6=96=B0openai=E7=89=88=E6=9C=AC?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- pyproject.toml | 2 +- requirements2.txt | 881 ++++++++++++++++++++++++++++++++++++++++++++++ setup.py | 1 - uv.lock | 12 +- 4 files changed, 888 insertions(+), 8 deletions(-) create mode 100644 requirements2.txt diff --git a/pyproject.toml b/pyproject.toml index 65bbb26..50f5a2d 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -15,7 +15,7 @@ dependencies = [ "infinity-emb[all]==0.0.53", "lmdeploy==0.6.2", "loguru>=0.7.2", - "openai==1.45.0", + "openai==1.55.3", "setuptools==75.2.0", "streamlit==1.39.0", "torch==2.5.1", diff --git a/requirements2.txt b/requirements2.txt new file mode 100644 index 0000000..b6a1414 --- /dev/null +++ b/requirements2.txt @@ -0,0 +1,881 @@ +# This file was autogenerated by uv via the following command: +# uv pip compile pyproject.toml +absl-py==2.1.0 + # via + # evalscope + # rouge-score +accelerate==1.2.0 + # via + # gpt-server (pyproject.toml) + # evalscope + # lmdeploy + # peft +addict==2.4.0 + # via + # mmengine-lite + # modelscope +aiofiles==23.2.1 + # via gradio +aiohappyeyeballs==2.4.4 + # via aiohttp +aiohttp==3.11.10 + # via + # datasets + # evalscope + # fschat + # fsspec + # vllm +aiosignal==1.3.1 + # via + # aiohttp + # ray +aliyun-python-sdk-core==2.16.0 + # via + # aliyun-python-sdk-kms + # oss2 +aliyun-python-sdk-kms==2.16.5 + # via oss2 +altair==5.5.0 + # via + # gradio + # streamlit +annotated-types==0.7.0 + # via pydantic +anyio==4.7.0 + # via + # httpx + # openai + # sse-starlette + # starlette + # watchfiles +async-timeout==5.0.1 + # via aiohttp +attrs==24.2.0 + # via + # aiohttp + # jsonlines + # jsonschema + # modelscope + # referencing +av==14.0.1 + # via qwen-vl-utils +blinker==1.9.0 + # via streamlit +cachetools==5.5.0 + # via + # evalscope + # streamlit +certifi==2024.8.30 + # via + # httpcore + # httpx + # requests +cffi==1.17.1 + # via cryptography +charset-normalizer==3.4.0 + # via requests +click==8.1.7 + # via + # nltk + # ray + # streamlit + # typer + # uvicorn +cloudpickle==3.1.0 + # via outlines +colorama==0.4.6 + # via + # sacrebleu + # typer +coloredlogs==15.0.1 + # via + # onnxruntime + # optimum +compressed-tensors==0.8.0 + # via vllm +contourpy==1.3.1 + # via matplotlib +crcmod==1.7 + # via oss2 +cryptography==44.0.0 + # via aliyun-python-sdk-core +ctranslate2==4.5.0 + # via infinity-emb +cycler==0.12.1 + # via matplotlib +datasets==3.0.1 + # via + # evalscope + # evaluate + # modelscope + # optimum + # outlines +dill==0.3.8 + # via + # datasets + # evaluate + # multiprocess +diskcache==5.6.3 + # via + # infinity-emb + # outlines +distro==1.9.0 + # via openai +editdistance==0.8.1 + # via evalscope +einops==0.8.0 + # via + # infinity-emb + # lmdeploy + # modelscope + # vllm +evalscope==0.7.0 + # via gpt-server (pyproject.toml) +evaluate==0.4.3 + # via optimum +exceptiongroup==1.2.2 + # via anyio +fastapi==0.114.1 + # via + # gpt-server (pyproject.toml) + # evalscope + # fschat + # gradio + # infinity-emb + # lmdeploy + # vllm +ffmpy==0.4.0 + # via + # gpt-server (pyproject.toml) + # gradio +filelock==3.16.1 + # via + # datasets + # huggingface-hub + # ray + # torch + # transformers + # triton + # vllm +fire==0.7.0 + # via lmdeploy +flatbuffers==24.3.25 + # via onnxruntime +fonttools==4.55.2 + # via matplotlib +frozenlist==1.5.0 + # via + # aiohttp + # aiosignal + # ray +fschat==0.2.36 + # via gpt-server (pyproject.toml) +fsspec==2024.6.1 + # via + # datasets + # evaluate + # gradio-client + # huggingface-hub + # torch +gguf==0.10.0 + # via vllm +gitdb==4.0.11 + # via gitpython +gitpython==3.1.43 + # via streamlit +gradio==4.26.0 + # via gpt-server (pyproject.toml) +gradio-client==0.15.1 + # via gradio +h11==0.14.0 + # via + # httpcore + # uvicorn +hf-transfer==0.1.8 + # via infinity-emb +httpcore==1.0.7 + # via httpx +httptools==0.6.4 + # via uvicorn +httpx==0.28.1 + # via + # fschat + # gradio + # gradio-client + # openai +huggingface-hub==0.26.5 + # via + # accelerate + # datasets + # evaluate + # gradio + # gradio-client + # infinity-emb + # optimum + # peft + # sentence-transformers + # timm + # tokenizers + # transformers +humanfriendly==10.0 + # via coloredlogs +idna==3.10 + # via + # anyio + # httpx + # requests + # yarl +importlib-metadata==8.5.0 + # via vllm +importlib-resources==6.4.5 + # via gradio +infinity-emb==0.0.53 + # via gpt-server (pyproject.toml) +interegular==0.3.3 + # via + # lm-format-enforcer + # outlines +jieba==0.42.1 + # via evalscope +jinja2==3.1.4 + # via + # altair + # gradio + # outlines + # pydeck + # torch +jiter==0.8.0 + # via openai +jmespath==0.10.0 + # via aliyun-python-sdk-core +joblib==1.4.2 + # via + # nltk + # scikit-learn +jsonlines==4.0.0 + # via evalscope +jsonschema==4.23.0 + # via + # altair + # mistral-common + # outlines + # ray +jsonschema-specifications==2024.10.1 + # via jsonschema +kiwisolver==1.4.7 + # via matplotlib +lark==1.2.2 + # via outlines +latex2mathml==3.77.0 + # via markdown2 +llvmlite==0.43.0 + # via numba +lm-format-enforcer==0.10.9 + # via vllm +lmdeploy==0.6.2 + # via gpt-server (pyproject.toml) +loguru==0.7.3 + # via gpt-server (pyproject.toml) +lxml==5.3.0 + # via sacrebleu +markdown-it-py==3.0.0 + # via rich +markdown2==2.5.1 + # via fschat +markupsafe==2.1.5 + # via + # gradio + # jinja2 +matplotlib==3.9.3 + # via + # evalscope + # gradio + # seaborn +mdurl==0.1.2 + # via markdown-it-py +mistral-common==1.5.1 + # via vllm +mmengine-lite==0.10.5 + # via lmdeploy +modelscope==1.20.1 + # via + # gpt-server (pyproject.toml) + # evalscope +mpmath==1.3.0 + # via sympy +msgpack==1.1.0 + # via ray +msgspec==0.18.6 + # via vllm +multidict==6.1.0 + # via + # aiohttp + # yarl +multiprocess==0.70.16 + # via + # datasets + # evaluate +narwhals==1.15.2 + # via altair +nest-asyncio==1.6.0 + # via outlines +networkx==3.4.2 + # via torch +nh3==0.2.19 + # via fschat +nltk==3.9.1 + # via + # evalscope + # rouge-score +numba==0.60.0 + # via outlines +numpy==1.26.4 + # via + # accelerate + # contourpy + # ctranslate2 + # datasets + # evalscope + # evaluate + # fschat + # gguf + # gradio + # infinity-emb + # lmdeploy + # matplotlib + # mistral-common + # mmengine-lite + # numba + # onnx + # onnxruntime + # opencv-python-headless + # optimum + # outlines + # pandas + # peft + # pyarrow + # pydeck + # rouge-score + # sacrebleu + # scikit-learn + # scipy + # seaborn + # streamlit + # torchvision + # transformers + # vllm + # xformers +nvidia-cublas-cu12==12.4.5.8 + # via + # lmdeploy + # nvidia-cudnn-cu12 + # nvidia-cusolver-cu12 + # torch +nvidia-cuda-cupti-cu12==12.4.127 + # via torch +nvidia-cuda-nvrtc-cu12==12.4.127 + # via torch +nvidia-cuda-runtime-cu12==12.4.127 + # via + # lmdeploy + # torch +nvidia-cudnn-cu12==9.1.0.70 + # via torch +nvidia-cufft-cu12==11.2.1.3 + # via torch +nvidia-curand-cu12==10.3.5.147 + # via + # lmdeploy + # torch +nvidia-cusolver-cu12==11.6.1.9 + # via torch +nvidia-cusparse-cu12==12.3.1.170 + # via + # nvidia-cusolver-cu12 + # torch +nvidia-ml-py==12.560.30 + # via + # pynvml + # vllm +nvidia-nccl-cu12==2.21.5 + # via + # lmdeploy + # torch +nvidia-nvjitlink-cu12==12.4.127 + # via + # nvidia-cusolver-cu12 + # nvidia-cusparse-cu12 + # torch +nvidia-nvtx-cu12==12.4.127 + # via torch +onnx==1.17.0 + # via optimum +onnxruntime==1.20.1 + # via optimum +openai==1.55.3 + # via + # gpt-server (pyproject.toml) + # evalscope + # lmdeploy + # vllm +opencv-python-headless==4.10.0.84 + # via mistral-common +optimum==1.23.3 + # via infinity-emb +orjson==3.10.12 + # via + # gradio + # infinity-emb +oss2==2.19.1 + # via modelscope +outlines==0.0.46 + # via + # lmdeploy + # vllm +packaging==24.2 + # via + # accelerate + # altair + # datasets + # evaluate + # gradio + # gradio-client + # huggingface-hub + # lm-format-enforcer + # matplotlib + # onnxruntime + # optimum + # peft + # plotly + # qwen-vl-utils + # ray + # streamlit + # transformers +pandas==2.2.3 + # via + # datasets + # evalscope + # evaluate + # gradio + # seaborn + # streamlit +partial-json-parser==0.2.1.1.post4 + # via vllm +peft==0.11.1 + # via lmdeploy +pillow==10.4.0 + # via + # gradio + # infinity-emb + # lmdeploy + # matplotlib + # mistral-common + # modelscope + # qwen-vl-utils + # sentence-transformers + # streamlit + # torchvision + # vllm +platformdirs==4.3.6 + # via yapf +plotly==5.24.1 + # via evalscope +ply==3.11 + # via simple-ddl-parser +portalocker==3.0.0 + # via sacrebleu +prometheus-client==0.21.1 + # via + # prometheus-fastapi-instrumentator + # vllm +prometheus-fastapi-instrumentator==7.0.0 + # via + # infinity-emb + # vllm +prompt-toolkit==3.0.48 + # via fschat +propcache==0.2.1 + # via + # aiohttp + # yarl +protobuf==5.29.1 + # via + # lmdeploy + # onnx + # onnxruntime + # optimum + # ray + # streamlit + # vllm +psutil==6.1.0 + # via + # accelerate + # peft + # vllm +py-cpuinfo==9.0.0 + # via vllm +pyairports==2.1.1 + # via outlines +pyarrow==17.0.0 + # via + # datasets + # evalscope + # streamlit +pycountry==24.6.1 + # via outlines +pycparser==2.22 + # via cffi +pycryptodome==3.21.0 + # via oss2 +pydantic==2.10.3 + # via + # compressed-tensors + # fastapi + # fschat + # gradio + # infinity-emb + # lm-format-enforcer + # lmdeploy + # mistral-common + # openai + # outlines + # vllm +pydantic-core==2.27.1 + # via pydantic +pydeck==0.9.1 + # via streamlit +pydub==0.25.1 + # via gradio +pygments==2.18.0 + # via + # markdown2 + # rich +pympler==1.1 + # via evalscope +pynvml==12.0.0 + # via lmdeploy +pyparsing==3.2.0 + # via matplotlib +python-dateutil==2.9.0.post0 + # via + # matplotlib + # modelscope + # pandas +python-dotenv==1.0.1 + # via uvicorn +python-multipart==0.0.19 + # via gradio +pytz==2024.2 + # via pandas +pyyaml==6.0.2 + # via + # accelerate + # ctranslate2 + # datasets + # evalscope + # gguf + # gradio + # huggingface-hub + # lm-format-enforcer + # mmengine-lite + # peft + # ray + # timm + # transformers + # uvicorn + # vllm + # wavedrom +pyzmq==26.2.0 + # via vllm +qwen-vl-utils==0.0.8 + # via gpt-server (pyproject.toml) +ray==2.40.0 + # via vllm +referencing==0.35.1 + # via + # jsonschema + # jsonschema-specifications + # outlines +regex==2024.11.6 + # via + # evalscope + # nltk + # sacrebleu + # tiktoken + # transformers +requests==2.32.3 + # via + # datasets + # evalscope + # evaluate + # fschat + # huggingface-hub + # mistral-common + # modelscope + # oss2 + # outlines + # qwen-vl-utils + # ray + # requests-toolbelt + # streamlit + # tiktoken + # transformers + # vllm +requests-toolbelt==1.0.0 + # via evalscope +rich==13.9.4 + # via + # fschat + # infinity-emb + # mmengine-lite + # streamlit + # typer +rouge-chinese==1.0.3 + # via evalscope +rouge-score==0.1.2 + # via evalscope +rpds-py==0.22.3 + # via + # jsonschema + # referencing +ruff==0.8.2 + # via gradio +sacrebleu==2.4.3 + # via evalscope +safetensors==0.4.5 + # via + # accelerate + # lmdeploy + # peft + # timm + # transformers +scikit-learn==1.5.2 + # via + # evalscope + # sentence-transformers +scipy==1.14.1 + # via + # modelscope + # scikit-learn + # sentence-transformers +seaborn==0.13.2 + # via evalscope +semantic-version==2.10.0 + # via gradio +sentence-transformers==3.3.1 + # via infinity-emb +sentencepiece==0.2.0 + # via + # evalscope + # lmdeploy + # mistral-common + # vllm +setuptools==75.2.0 + # via + # --override (workspace) + # gpt-server (pyproject.toml) + # ctranslate2 + # modelscope + # torch + # vllm +shellingham==1.5.4 + # via typer +shortuuid==1.0.13 + # via + # fschat + # lmdeploy +simple-ddl-parser==1.7.1 + # via evalscope +simplejson==3.19.3 + # via modelscope +six==1.17.0 + # via + # oss2 + # python-dateutil + # rouge-chinese + # rouge-score + # wavedrom +smmap==5.0.1 + # via gitdb +sniffio==1.3.1 + # via + # anyio + # openai +sortedcontainers==2.4.0 + # via modelscope +sse-starlette==2.1.3 + # via evalscope +starlette==0.38.6 + # via + # fastapi + # prometheus-fastapi-instrumentator + # sse-starlette +streamlit==1.39.0 + # via gpt-server (pyproject.toml) +svgwrite==1.4.3 + # via wavedrom +sympy==1.13.1 + # via + # onnxruntime + # optimum + # torch +tabulate==0.9.0 + # via + # evalscope + # sacrebleu +tenacity==9.0.0 + # via + # plotly + # streamlit +termcolor==2.5.0 + # via + # fire + # mmengine-lite +threadpoolctl==3.5.0 + # via scikit-learn +tiktoken==0.7.0 + # via + # evalscope + # fschat + # lmdeploy + # mistral-common + # vllm +timm==1.0.12 + # via infinity-emb +tokenizers==0.20.3 + # via + # transformers + # vllm +toml==0.10.2 + # via streamlit +tomli==2.2.1 + # via yapf +tomlkit==0.12.0 + # via gradio +torch==2.5.1 + # via + # --override (workspace) + # gpt-server (pyproject.toml) + # accelerate + # compressed-tensors + # evalscope + # infinity-emb + # lmdeploy + # optimum + # peft + # sentence-transformers + # timm + # torchvision + # vllm + # xformers +torchvision==0.20.1 + # via + # --override (workspace) + # gpt-server (pyproject.toml) + # lmdeploy + # timm + # vllm +tornado==6.4.2 + # via streamlit +tqdm==4.67.1 + # via + # datasets + # evalscope + # evaluate + # gguf + # huggingface-hub + # modelscope + # nltk + # openai + # outlines + # peft + # sentence-transformers + # transformers + # vllm +transformers==4.45.2 + # via + # gpt-server (pyproject.toml) + # compressed-tensors + # evalscope + # lmdeploy + # modelscope + # optimum + # peft + # sentence-transformers + # transformers-stream-generator + # vllm +transformers-stream-generator==0.0.5 + # via evalscope +triton==3.1.0 + # via + # --override (workspace) + # lmdeploy + # torch +typer==0.9.4 + # via + # gradio + # infinity-emb +typing-extensions==4.12.2 + # via + # altair + # anyio + # fastapi + # gradio + # gradio-client + # huggingface-hub + # mistral-common + # multidict + # openai + # outlines + # pydantic + # pydantic-core + # rich + # streamlit + # torch + # typer + # uvicorn + # vllm +tzdata==2024.2 + # via pandas +unicorn==2.1.1 + # via evalscope +urllib3==2.2.3 + # via + # modelscope + # requests +uvicorn==0.23.2 + # via + # fschat + # gradio + # infinity-emb + # lmdeploy + # sse-starlette + # vllm +uvloop==0.21.0 + # via uvicorn +vllm==0.6.4.post1 + # via gpt-server (pyproject.toml) +watchdog==5.0.3 + # via streamlit +watchfiles==1.0.0 + # via uvicorn +wavedrom==2.0.3.post3 + # via markdown2 +wcwidth==0.2.13 + # via prompt-toolkit +websockets==11.0.3 + # via + # gradio-client + # uvicorn +xformers==0.0.28.post3 + # via vllm +xxhash==3.5.0 + # via + # datasets + # evaluate +yapf==0.43.0 + # via mmengine-lite +yarl==1.18.3 + # via aiohttp +zipp==3.21.0 + # via importlib-metadata diff --git a/setup.py b/setup.py index 51b8286..861d388 100644 --- a/setup.py +++ b/setup.py @@ -1,6 +1,5 @@ import os from setuptools import setup, find_packages -from setuptools.command.install import install pwd = os.path.dirname(__file__) diff --git a/uv.lock b/uv.lock index 5a656de..a643f7b 100644 --- a/uv.lock +++ b/uv.lock @@ -1089,7 +1089,7 @@ requires-dist = [ { name = "lmdeploy", specifier = "==0.6.2" }, { name = "loguru", specifier = ">=0.7.2" }, { name = "modelscope", specifier = "==1.20.1" }, - { name = "openai", specifier = "==1.45.0" }, + { name = "openai", specifier = "==1.55.3" }, { name = "qwen-vl-utils" }, { name = "setuptools", specifier = "==75.2.0" }, { name = "streamlit", specifier = "==1.39.0" }, @@ -1681,9 +1681,9 @@ dependencies = [ { name = "colorama", marker = "sys_platform == 'win32'" }, { name = "win32-setctime", marker = "sys_platform == 'win32'" }, ] -sdist = { url = "https://pypi.tuna.tsinghua.edu.cn/packages/9e/30/d87a423766b24db416a46e9335b9602b054a72b96a88a241f2b09b560fa8/loguru-0.7.2.tar.gz", hash = "sha256:e671a53522515f34fd406340ee968cb9ecafbc4b36c679da03c18fd8d0bd51ac" } +sdist = { url = "https://pypi.tuna.tsinghua.edu.cn/packages/9e/30/d87a423766b24db416a46e9335b9602b054a72b96a88a241f2b09b560fa8/loguru-0.7.2.tar.gz", hash = "sha256:e671a53522515f34fd406340ee968cb9ecafbc4b36c679da03c18fd8d0bd51ac", size = 145103 } wheels = [ - { url = "https://pypi.tuna.tsinghua.edu.cn/packages/03/0a/4f6fed21aa246c6b49b561ca55facacc2a44b87d65b8b92362a8e99ba202/loguru-0.7.2-py3-none-any.whl", hash = "sha256:003d71e3d3ed35f0f8984898359d65b79e5b21943f78af86aa5491210429b8eb" }, + { url = "https://pypi.tuna.tsinghua.edu.cn/packages/03/0a/4f6fed21aa246c6b49b561ca55facacc2a44b87d65b8b92362a8e99ba202/loguru-0.7.2-py3-none-any.whl", hash = "sha256:003d71e3d3ed35f0f8984898359d65b79e5b21943f78af86aa5491210429b8eb", size = 62549 }, ] [[package]] @@ -2458,7 +2458,7 @@ wheels = [ [[package]] name = "openai" -version = "1.45.0" +version = "1.55.3" source = { registry = "https://pypi.tuna.tsinghua.edu.cn/simple" } dependencies = [ { name = "anyio" }, @@ -2470,9 +2470,9 @@ dependencies = [ { name = "tqdm" }, { name = "typing-extensions" }, ] -sdist = { url = "https://pypi.tuna.tsinghua.edu.cn/packages/70/cd/5ec65b9a56999370c032af7933433143f78239d44a8c03a5ba34159af945/openai-1.45.0.tar.gz", hash = "sha256:731207d10637335413aa3c0955f8f8df30d7636a4a0f9c381f2209d32cf8de97", size = 295638 } +sdist = { url = "https://pypi.tuna.tsinghua.edu.cn/packages/1e/39/d4859d897da053b61b84403f67dbef1abd075e441cb354892ff14f98e2c7/openai-1.55.3.tar.gz", hash = "sha256:547e85b94535469f137a779d8770c8c5adebd507c2cc6340ca401a7c4d5d16f0", size = 314571 } wheels = [ - { url = "https://pypi.tuna.tsinghua.edu.cn/packages/d4/2a/97e80a4551346efc9cd937e11adb640207acc5045fdf4e06786eac55bfb1/openai-1.45.0-py3-none-any.whl", hash = "sha256:2f1f7b7cf90f038a9f1c24f0d26c0f1790c102ec5acd07ffd70a9b7feac1ff4e", size = 374133 }, + { url = "https://pypi.tuna.tsinghua.edu.cn/packages/10/06/691ef3f0112ecf0d7420d0bf35b5d16cf81554141f4b4913a9831031013d/openai-1.55.3-py3-none-any.whl", hash = "sha256:2a235d0e1e312cd982f561b18c27692e253852f4e5fb6ccf08cb13540a9bdaa1", size = 389558 }, ] [[package]]