Skip to content

Commit

Permalink
update multimodal doc and requirement (#47)
Browse files Browse the repository at this point in the history
Co-authored-by: Xiaotong Chen <“[email protected]”>
  • Loading branch information
x574chen and Xiaotong Chen authored Dec 18, 2024
1 parent 1b2a6ad commit 97108ec
Show file tree
Hide file tree
Showing 4 changed files with 10 additions and 5 deletions.
10 changes: 7 additions & 3 deletions docs/sphinx/vlm/vlm_offline_inference_en.rst
Original file line number Diff line number Diff line change
Expand Up @@ -107,15 +107,19 @@ Launching with CLI
You can also opt to install dashinfer-vlm locally and use command line to launch server.

1. Pull dashinfer docker image (see :ref:`docker-label`)
2. Download and extract the TensorRT GA build
2. Install TensorRT Python package, and download TensorRT GA build from NVIDIA Developer Zone.

Example: TensorRT 10.6.0.26 for CUDA 12.6, Linux x86_64

.. code-block:: bash
pip install tensorrt
wget https://developer.nvidia.com/downloads/compute/machine-learning/tensorrt/10.6.0/tars/TensorRT-10.6.0.26.Linux.x86_64-gnu.cuda-12.6.tar.gz
tar -xvzf TensorRT-10.6.0.26.Linux.x86_64-gnu.cuda-12.6.tar.gz
export TRT_LIBPATH=`pwd`/TensorRT-10.6.0.26
export LD_LIBRARY_PATH=`pwd`/TensorRT-10.6.0.26/lib
3. Install ``dashinfer-vlm``: ``pip install dashinfer-vlm``.
3. Install dashinfer Python Package from `release <https://github.com/modelscope/dash-infer/releases>`_
4. Install dashinfer-vlm: ``pip install dashinfer-vlm``.

Now you can launch server with command line:

Expand Down
1 change: 1 addition & 0 deletions multimodal/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -6,6 +6,7 @@ RUN mkdir /root/code/
COPY ./dashinfer_vlm /root/code/dashinfer_vlm
COPY ./setup.py code/
COPY ./requirements.txt /root/code/requirements.txt
RUN python3 -m pip install https://github.com/modelscope/dash-infer/releases/download/v2.0.0-rc2/dashinfer-2.0.0rc2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
RUN python3 -m pip install -r /root/code/requirements.txt --index-url=http://mirrors.aliyun.com/pypi/simple/ --trusted-host=mirrors.aliyun.com

RUN python3 -m pip install -e /root/code/
Expand Down
4 changes: 2 additions & 2 deletions multimodal/requirements.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,3 @@
dashinfer
av
numpy==1.24.3
requests==2.32.3
Expand All @@ -12,7 +11,8 @@ shortuuid
fastapi
pydantic_settings
uvicorn
cmake==3.22.6
cmake==3.22.6
modelscope
aiohttp
onnx
torchvision
Binary file modified multimodal/resource/dashinfer-vlm-arch.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.

0 comments on commit 97108ec

Please sign in to comment.