Skip to content

Commit

Permalink
install numactl to enable fastsafetensors
Browse files Browse the repository at this point in the history
Signed-off-by: Jefferson Fialho <[email protected]>
  • Loading branch information
fialhocoelho committed Dec 13, 2024
1 parent 97a6423 commit 11267fc
Showing 1 changed file with 25 additions and 3 deletions.
28 changes: 25 additions & 3 deletions Dockerfile.ubi
Original file line number Diff line number Diff line change
Expand Up @@ -42,15 +42,35 @@ FROM python-install as cuda-base
RUN curl -Lo /etc/yum.repos.d/cuda-rhel9.repo \
https://developer.download.nvidia.com/compute/cuda/repos/rhel9/x86_64/cuda-rhel9.repo

# Install necessary packages to build numactl from source
RUN microdnf install -y autoconf automake libtool make rpm-build

# Download the numactl source RPM to /root directory
WORKDIR /root
RUN microdnf download --enablerepo=ubi-9-baseos-source --source numactl.src

# Get NUMACTL_V (just the version part of the RPM file)
RUN NUMACTL_V=$(ls /root/numactl-* | sed -r 's|/root/numactl-(.+)\.el9\.src\.rpm|\1|') && \
echo "########################################### NUMACTL_V is set to $NUMACTL_V" && \
rpm -i /root/numactl-${NUMACTL_V}.el9.src.rpm && \
rpmbuild -ba /root/rpmbuild/SPECS/numactl.spec && \
rpm -i /root/rpmbuild/RPMS/x86_64/{numactl-libs-${NUMACTL_V}.el9.x86_64.rpm,numactl-${NUMACTL_V}.el9.x86_64.rpm,numactl-devel-${NUMACTL_V}.el9.x86_64.rpm}

RUN microdnf install -y \
cuda-nvcc-12-4 cuda-nvtx-12-4 cuda-libraries-devel-12-4 && \
cuda-nvcc-12-4 cuda-nvtx-12-4 cuda-libraries-devel-12-4 && \
microdnf clean all

# Search for numa.h in common locations
RUN echo "### Searching for numa.h in common directories ###" && \
find /usr /usr/local /opt /lib -name numa.h | tee /dev/stderr && \
echo "### END OF NUMA.H SEARCH ###"

ENV CUDA_HOME="/usr/local/cuda" \
PATH="${CUDA_HOME}/bin:${PATH}" \
LD_LIBRARY_PATH="${CUDA_HOME}/lib64:${CUDA_HOME}/extras/CUPTI/lib64:${LD_LIBRARY_PATH}"

## Python cuda base #################################################################

## Python cuda base ############################################################
FROM cuda-base AS python-cuda-base

ENV VIRTUAL_ENV=/opt/vllm
Expand Down Expand Up @@ -201,11 +221,13 @@ WORKDIR /home/vllm

ENTRYPOINT ["python3", "-m", "vllm.entrypoints.openai.api_server"]


FROM vllm-openai as vllm-grpc-adapter

USER root

RUN --mount=type=bind,from=cuda-base,source=/root/rpmbuild/RPMS/x86_64,target=/tmp/rpm/ \
rpm -i /tmp/rpm/numactl-libs-*.el9.x86_64.rpm

RUN --mount=type=cache,target=/root/.cache/pip \
--mount=type=cache,target=/root/.cache/uv \
--mount=type=bind,from=build,src=/workspace/dist,target=/workspace/dist \
Expand Down

0 comments on commit 11267fc

Please sign in to comment.