diff --git a/.github/workflows/build-image.yml b/.github/workflows/build-image.yml index 78efa7800..05e4dd447 100644 --- a/.github/workflows/build-image.yml +++ b/.github/workflows/build-image.yml @@ -40,16 +40,16 @@ jobs: tags: superbench/main:cuda11.1.1,superbench/superbench:latest runner: ubuntu-latest build_args: "NUM_MAKE_JOBS=8" - - name: rocm5.7 - dockerfile: rocm5.7.x - tags: superbench/main:rocm5.7 - runner: [self-hosted, rocm-build] - build_args: "NUM_MAKE_JOBS=64" - name: rocm6.0 dockerfile: rocm6.0.x tags: superbench/main:rocm6.0 runner: [self-hosted, rocm-build] - build_args: "NUM_MAKE_JOBS=64" + build_args: "NUM_MAKE_JOBS=16" + - name: rocm6.2 + dockerfile: rocm6.2.x + tags: superbench/main:rocm6.2 + runner: [self-hosted, rocm-build] + build_args: "NUM_MAKE_JOBS=16" steps: - name: Checkout uses: actions/checkout@v2 @@ -68,6 +68,8 @@ jobs: else echo "No Docker images found with the specified references." fi + sudo docker ps -q | grep build | xargs -r sudo docker stop + echo y | sudo docker system prune -a --volumes df -h - name: Prepare metadata id: metadata diff --git a/README.md b/README.md index 7706fc472..8d4e0a8f1 100644 --- a/README.md +++ b/README.md @@ -15,7 +15,7 @@ __SuperBench__ is a validation and profiling tool for AI infrastructure. -📢 [v0.10.0](https://github.com/microsoft/superbenchmark/releases/tag/v0.10.0) has been released! +📢 [v0.11.0](https://github.com/microsoft/superbenchmark/releases/tag/v0.11.0) has been released! ## _Check [aka.ms/superbench](https://aka.ms/superbench) for more details._ diff --git a/dockerfile/cuda11.1.1.dockerfile b/dockerfile/cuda11.1.1.dockerfile index 65df9d6e5..54577fdad 100644 --- a/dockerfile/cuda11.1.1.dockerfile +++ b/dockerfile/cuda11.1.1.dockerfile @@ -83,11 +83,13 @@ RUN cd /tmp && \ rm -rf /tmp/MLNX_OFED_LINUX-${OFED_VERSION}* # Install HPC-X +ENV HPCX_VERSION=v2.9.0 RUN cd /opt && \ - wget -q https://azhpcstor.blob.core.windows.net/azhpc-images-store/hpcx-v2.8.3-gcc-MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu20.04-x86_64.tbz && \ - tar xf hpcx-v2.8.3-gcc-MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu20.04-x86_64.tbz && \ - ln -s hpcx-v2.8.3-gcc-MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu20.04-x86_64 hpcx && \ - rm hpcx-v2.8.3-gcc-MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu20.04-x86_64.tbz + rm -rf hpcx && \ + wget -q https://content.mellanox.com/hpc/hpc-x/${HPCX_VERSION}/hpcx-${HPCX_VERSION}-gcc-inbox-ubuntu20.04-x86_64.tbz -O hpcx.tbz && \ + tar xf hpcx.tbz && \ + mv hpcx-${HPCX_VERSION}-gcc-inbox-ubuntu20.04-x86_64 hpcx && \ + rm hpcx.tbz # Install NCCL RDMA SHARP plugins RUN cd /tmp && \ diff --git a/dockerfile/cuda12.4.dockerfile b/dockerfile/cuda12.4.dockerfile index 984c6ee01..3ec82f39d 100644 --- a/dockerfile/cuda12.4.dockerfile +++ b/dockerfile/cuda12.4.dockerfile @@ -8,7 +8,7 @@ FROM nvcr.io/nvidia/pytorch:24.03-py3 # - CUDA: 12.4.0 # - cuDNN: 9.0.0.306 # - cuBLAS: 12.4.2.65 -# - NCCL: v2.20 +# - NCCL: v2.23.4-1 # - TransformerEngine 1.4 # Mellanox: # - OFED: 23.07-0.5.1.2 @@ -115,6 +115,23 @@ RUN cd /tmp && \ mv amd-blis /opt/AMD && \ rm -rf aocl-blis-linux-aocc-4.0.tar.gz +# Install NCCL 2.23.4 +RUN cd /tmp && \ + git clone -b v2.23.4-1 https://github.com/NVIDIA/nccl.git && \ + cd nccl && \ + make -j ${NUM_MAKE_JOBS} src.build && \ + make install && \ + rm -rf /tmp/nccl + +# Install UCX v1.16.0 with multi-threading support +RUN cd /tmp && \ + wget https://github.com/openucx/ucx/releases/download/v1.16.0/ucx-1.16.0.tar.gz && \ + tar xzf ucx-1.16.0.tar.gz && \ + cd ucx-1.16.0 && \ + ./contrib/configure-release-mt --prefix=/usr/local && \ + make -j ${NUM_MAKE_JOBS} && \ + make install + ENV PATH="${PATH}" \ LD_LIBRARY_PATH="/usr/local/lib:/usr/local/mpi/lib:${LD_LIBRARY_PATH}" \ SB_HOME=/opt/superbench \ diff --git a/dockerfile/rocm6.0.x.dockerfile b/dockerfile/rocm6.0.x.dockerfile index bd33e289a..ce5736e29 100644 --- a/dockerfile/rocm6.0.x.dockerfile +++ b/dockerfile/rocm6.0.x.dockerfile @@ -173,6 +173,11 @@ RUN make RCCL_HOME=/opt/rccl/build/ ROCBLAS_BRANCH=release/rocm-rel-6.0 HIPBLASL RUN cd third_party/Megatron/Megatron-DeepSpeed && \ git apply ../megatron_deepspeed_rocm6.patch +# Install AMD SMI Python Library +RUN apt install amd-smi-lib -y && \ + cd /opt/rocm/share/amd_smi && \ + python3 -m pip install . + ADD . . ENV USE_HIP_DATATYPE=1 ENV USE_HIPBLAS_COMPUTETYPE=1 diff --git a/dockerfile/rocm6.2.x.dockerfile b/dockerfile/rocm6.2.x.dockerfile new file mode 100644 index 000000000..31506d351 --- /dev/null +++ b/dockerfile/rocm6.2.x.dockerfile @@ -0,0 +1,193 @@ +ARG BASE_IMAGE=rocm/pytorch:rocm6.2_ubuntu20.04_py3.9_pytorch_release_2.3.0 + +FROM ${BASE_IMAGE} + +# OS: +# - Ubuntu: 22.04 +# - Docker Client: 20.10.8 +# ROCm: +# - ROCm: 6.2 +# Lib: +# - torch: 2.3.0 +# - rccl: 2.18.3+hip6.0 develop:7e1cbb4 +# - hipblaslt: release-staging/rocm-rel-6.2 +# - rocblas: release-staging/rocm-rel-6.2 +# - openmpi: 4.1.x +# Intel: +# - mlc: v3.11 + +LABEL maintainer="SuperBench" + +ENV DEBIAN_FRONTEND=noninteractive +RUN apt-get update && \ + apt-get -q install -y --no-install-recommends \ + autoconf \ + automake \ + bc \ + build-essential \ + curl \ + dmidecode \ + git \ + hipify-clang \ + iproute2 \ + jq \ + libaio-dev \ + libboost-program-options-dev \ + libcap2 \ + libcurl4-openssl-dev \ + libnuma-dev \ + libpci-dev \ + libssl-dev \ + libtinfo5 \ + libtool \ + lshw \ + net-tools \ + numactl \ + openssh-client \ + openssh-server \ + pciutils \ + python3-mpi4py \ + rsync \ + sudo \ + util-linux \ + vim \ + wget \ + && \ + rm -rf /tmp/* + +ARG NUM_MAKE_JOBS=64 + +# Check if CMake is installed and its version +RUN cmake_version=$(cmake --version 2>/dev/null | grep -oP "(?<=cmake version )(\d+\.\d+)" || echo "0.0") && \ + required_version="3.24.1" && \ + if [ "$(printf "%s\n" "$required_version" "$cmake_version" | sort -V | head -n 1)" != "$required_version" ]; then \ + echo "existing cmake version is ${cmake_version}" && \ + cd /tmp && \ + wget -q https://github.com/Kitware/CMake/releases/download/v${required_version}/cmake-${required_version}.tar.gz && \ + tar xzf cmake-${required_version}.tar.gz && \ + cd cmake-${required_version} && \ + ./bootstrap --prefix=/usr --no-system-curl --parallel=16 && \ + make -j ${NUM_MAKE_JOBS} && \ + make install && \ + rm -rf /tmp/cmake-${required_version}* \ + else \ + echo "CMake version is greater than or equal to 3.24.1"; \ + fi + +# Install Docker +ENV DOCKER_VERSION=20.10.8 +RUN cd /tmp && \ + wget -q https://download.docker.com/linux/static/stable/x86_64/docker-${DOCKER_VERSION}.tgz -O docker.tgz && \ + tar --extract --file docker.tgz --strip-components 1 --directory /usr/local/bin/ && \ + rm docker.tgz + +# Update system config +RUN mkdir -p /root/.ssh && \ + touch /root/.ssh/authorized_keys && \ + mkdir -p /var/run/sshd && \ + sed -i "s/[# ]*PermitRootLogin prohibit-password/PermitRootLogin yes/" /etc/ssh/sshd_config && \ + sed -i "s/[# ]*PermitUserEnvironment no/PermitUserEnvironment yes/" /etc/ssh/sshd_config && \ + sed -i "s/[# ]*Port.*/Port 22/" /etc/ssh/sshd_config && \ + echo "* soft nofile 1048576\n* hard nofile 1048576" >> /etc/security/limits.conf && \ + echo "root soft nofile 1048576\nroot hard nofile 1048576" >> /etc/security/limits.conf + + +# Get Ubuntu version and set as an environment variable +RUN export UBUNTU_VERSION=$(lsb_release -r -s) +RUN echo "Ubuntu version: $UBUNTU_VERSION" +ENV UBUNTU_VERSION=${UBUNTU_VERSION} + +# Install OFED +ENV OFED_VERSION=5.9-0.5.6.0 +# Check if ofed_info is present and has a version +RUN if ! command -v ofed_info >/dev/null 2>&1; then \ + echo "OFED not found. Installing OFED..."; \ + cd /tmp && \ + wget -q http://content.mellanox.com/ofed/MLNX_OFED-${OFED_VERSION}/MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu${UBUNTU_VERSION}-x86_64.tgz && \ + tar xzf MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu${UBUNTU_VERSION}-x86_64.tgz && \ + PATH=/usr/bin:${PATH} MLNX_OFED_LINUX-${OFED_VERSION}-ubuntu${UBUNTU_VERSION}-x86_64/mlnxofedinstall --user-space-only --without-fw-update --force --all && \ + rm -rf MLNX_OFED_LINUX-${OFED_VERSION}* ; \ + fi + +ENV ROCM_PATH=/opt/rocm + +# Install OpenMPI +ENV OPENMPI_VERSION=4.1.x +ENV MPI_HOME=/usr/local/mpi +# Check if Open MPI is installed +RUN cd /tmp && \ + git clone --recursive https://github.com/open-mpi/ompi.git -b v${OPENMPI_VERSION} && \ + cd ompi && \ + ./autogen.pl && \ + mkdir build && \ + cd build && \ + ../configure --prefix=/usr/local/mpi --enable-orterun-prefix-by-default --enable-mpirun-prefix-by-default --enable-prte-prefix-by-default --with-rocm=/opt/rocm && \ + make -j $(nproc) && \ + make -j $(nproc) install && \ + ldconfig && \ + cd / && \ + rm -rf /tmp/openmpi-${OPENMPI_VERSION}* + +# Install Intel MLC +RUN cd /tmp && \ + wget -q https://downloadmirror.intel.com/763324/mlc_v3.10.tgz -O mlc.tgz && \ + tar xzf mlc.tgz Linux/mlc && \ + cp ./Linux/mlc /usr/local/bin/ && \ + rm -rf ./Linux mlc.tgz + +# Install RCCL +RUN cd /opt/ && \ + git clone -b release/rocm-rel-6.2 https://github.com/ROCmSoftwarePlatform/rccl.git && \ + cd rccl && \ + mkdir build && \ + cd build && \ + CXX=/opt/rocm/bin/hipcc cmake -DHIP_COMPILER=clang -DCMAKE_BUILD_TYPE=Release -DCMAKE_VERBOSE_MAKEFILE=1 \ + -DCMAKE_PREFIX_PATH="${ROCM_PATH}/hsa;${ROCM_PATH}/hip;${ROCM_PATH}/share/rocm/cmake/;${ROCM_PATH}" \ + .. && \ + make -j${NUM_MAKE_JOBS} + +# Install AMD SMI Python Library +RUN apt install amd-smi-lib -y && \ + cd /opt/rocm/share/amd_smi && \ + python3 -m pip install . + +ENV PATH="/usr/local/mpi/bin:/opt/superbench/bin:/usr/local/bin/:/opt/rocm/hip/bin/:/opt/rocm/bin/:${PATH}" \ + LD_PRELOAD="/opt/rccl/build/librccl.so:$LD_PRELOAD" \ + LD_LIBRARY_PATH="/usr/local/mpi/lib:/usr/lib/x86_64-linux-gnu/:/usr/local/lib/:/opt/rocm/lib:${LD_LIBRARY_PATH}" \ + SB_HOME=/opt/superbench \ + SB_MICRO_PATH=/opt/superbench \ + ANSIBLE_DEPRECATION_WARNINGS=FALSE \ + ANSIBLE_COLLECTIONS_PATH=/usr/share/ansible/collections + +RUN echo PATH="$PATH" > /etc/environment && \ + echo LD_LIBRARY_PATH="$LD_LIBRARY_PATH" >> /etc/environment && \ + echo SB_MICRO_PATH="$SB_MICRO_PATH" >> /etc/environment + +RUN apt install rocm-cmake -y && \ + python3 -m pip install --upgrade pip wheel setuptools==65.7 + +WORKDIR ${SB_HOME} + +ADD third_party third_party +# Apply patch +RUN cd third_party/perftest && \ + git apply ../perftest_rocm6.patch +RUN make RCCL_HOME=/opt/rccl/build/ ROCBLAS_BRANCH=release-staging/rocm-rel-6.2 HIPBLASLT_BRANCH=release-staging/rocm-rel-6.2 ROCM_VER=rocm-5.5.0 -C third_party rocm -o cpu_hpl -o cpu_stream -o megatron_lm +RUN cp -r /opt/superbench/third_party/hipBLASLt/build/release/hipblaslt-install/lib/* /opt/rocm/lib/ && \ + cp -r /opt/superbench/third_party/hipBLASLt/build/release/hipblaslt-install/include/* /opt/rocm/include/ +RUN cd third_party/Megatron/Megatron-DeepSpeed && \ + git apply ../megatron_deepspeed_rocm6.patch + +# Install transformer_engine +RUN git clone --recursive https://github.com/ROCm/TransformerEngine.git && \ + cd TransformerEngine && \ + export NVTE_FRAMEWORK=pytorch && \ + pip install . + +ADD . . +ENV USE_HIP_DATATYPE=1 +ENV USE_HIPBLAS_COMPUTETYPE=1 +RUN python3 -m pip install .[amdworker] && \ + CXX=/opt/rocm/bin/hipcc make cppbuild && \ + make postinstall + diff --git a/docs/getting-started/installation.mdx b/docs/getting-started/installation.mdx index cf48c4caa..30fdee829 100644 --- a/docs/getting-started/installation.mdx +++ b/docs/getting-started/installation.mdx @@ -61,7 +61,7 @@ You can clone the source from GitHub and build it. :::note Note You should checkout corresponding tag to use release version, for example, -`git clone -b v0.10.0 https://github.com/microsoft/superbenchmark` +`git clone -b v0.11.0 https://github.com/microsoft/superbenchmark` ::: ```bash diff --git a/docs/getting-started/run-superbench.md b/docs/getting-started/run-superbench.md index a6bb3bc1e..73f00c9c0 100644 --- a/docs/getting-started/run-superbench.md +++ b/docs/getting-started/run-superbench.md @@ -27,7 +27,7 @@ sb deploy -f remote.ini --host-password [password] :::note Note You should deploy corresponding Docker image to use release version, for example, -`sb deploy -f local.ini -i superbench/superbench:v0.10.0-cuda12.2` +`sb deploy -f local.ini -i superbench/superbench:v0.11.0-cuda12.4` You should note that version of git repo only determines version of sb CLI, and not the sb container. You should define the container version even if you specified a release version for the git clone. diff --git a/docs/superbench-config.mdx b/docs/superbench-config.mdx index b8ad058fa..102b8d69f 100644 --- a/docs/superbench-config.mdx +++ b/docs/superbench-config.mdx @@ -70,7 +70,7 @@ superbench: ```yaml -version: v0.10 +version: v0.11 superbench: enable: benchmark_1 monitor: diff --git a/docs/user-tutorial/container-images.mdx b/docs/user-tutorial/container-images.mdx index ffca22796..58a8079d7 100644 --- a/docs/user-tutorial/container-images.mdx +++ b/docs/user-tutorial/container-images.mdx @@ -30,6 +30,9 @@ available tags are listed below for all stable versions. | Tag | Description | |--------------------|-------------------------------------| +| v0.11.0-cuda12.4 | SuperBench v0.11.0 with CUDA 12.4 | +| v0.11.0-cuda12.2 | SuperBench v0.11.0 with CUDA 12.2 | +| v0.11.0-cuda11.1.1 | SuperBench v0.11.0 with CUDA 11.1.1 | | v0.10.0-cuda12.2 | SuperBench v0.10.0 with CUDA 12.2 | | v0.10.0-cuda11.1.1 | SuperBench v0.10.0 with CUDA 11.1.1 | | v0.9.0-cuda12.1 | SuperBench v0.9.0 with CUDA 12.1 | @@ -50,6 +53,9 @@ available tags are listed below for all stable versions. | Tag | Description | |-------------------------------|--------------------------------------------------| +| v0.11.0-rocm6.2 | SuperBench v0.11.0 with ROCm 6.2 | +| v0.11.0-rocm6.0 | SuperBench v0.11.0 with ROCm 6.0 | +| v0.10.0-rocm6.0 | SuperBench v0.10.0 with ROCm 6.0 | | v0.10.0-rocm5.7 | SuperBench v0.10.0 with ROCm 5.7 | | v0.9.0-rocm5.1.3 | SuperBench v0.9.0 with ROCm 5.1.3 | | v0.9.0-rocm5.1.1 | SuperBench v0.9.0 with ROCm 5.1.1 | diff --git a/docs/user-tutorial/data-diagnosis.md b/docs/user-tutorial/data-diagnosis.md index c2f0e3369..80278e7da 100644 --- a/docs/user-tutorial/data-diagnosis.md +++ b/docs/user-tutorial/data-diagnosis.md @@ -65,7 +65,7 @@ superbench: example: ```yaml # SuperBench rules -version: v0.10 +version: v0.11 superbench: rules: failure-rule: @@ -83,8 +83,8 @@ superbench: criteria: lambda x:x>0.05 categories: KernelLaunch metrics: - - kernel-launch/event_overhead:\d+ - - kernel-launch/wall_overhead:\d+ + - kernel-launch/event_time:\d+ + - kernel-launch/wall_time:\d+ rule1: # Rule 1: If H2D_Mem_BW or D2H_Mem_BW test suffers > 5% downgrade, label it as defective function: variance diff --git a/docs/user-tutorial/result-summary.md b/docs/user-tutorial/result-summary.md index dffee2514..8893f559c 100644 --- a/docs/user-tutorial/result-summary.md +++ b/docs/user-tutorial/result-summary.md @@ -58,7 +58,7 @@ superbench: ```yaml title="Example" # SuperBench rules -version: v0.10 +version: v0.11 superbench: rules: kernel_launch: @@ -70,8 +70,8 @@ superbench: aggregate: True categories: KernelLaunch metrics: - - kernel-launch/event_overhead - - kernel-launch/wall_overhead + - kernel-launch/event_time + - kernel-launch/wall_time nccl: statistics: mean categories: NCCL diff --git a/setup.py b/setup.py index ac9f2d266..686bef0b9 100644 --- a/setup.py +++ b/setup.py @@ -168,6 +168,7 @@ def run(self): 'openpyxl>=3.0.7', 'packaging>=21.0', 'pandas>=1.1.5', + 'protobuf<=3.20.3', 'pssh @ git+https://github.com/lilydjwg/pssh.git@v2.3.4', 'pyyaml>=5.3', 'requests>=2.27.1', @@ -218,7 +219,7 @@ def run(self): 'onnxruntime-gpu; python_version>="3.10"', ], 'nvidia': ['py3nvml>=0.2.6'], - 'amd': ['pyrsmi>=1.0.1'], + 'amd': ['amdsmi'], } ), include_package_data=True, diff --git a/superbench/__init__.py b/superbench/__init__.py index e1f4234fd..cbd47c45e 100644 --- a/superbench/__init__.py +++ b/superbench/__init__.py @@ -6,5 +6,5 @@ Provide hardware and software benchmarks for AI systems. """ -__version__ = '0.10.0' +__version__ = '0.11.0' __author__ = 'Microsoft' diff --git a/superbench/common/utils/device_manager.py b/superbench/common/utils/device_manager.py index 18bed8dce..2a594fef0 100644 --- a/superbench/common/utils/device_manager.py +++ b/superbench/common/utils/device_manager.py @@ -179,11 +179,11 @@ def get_device_temperature(self, idx): Return: temp (int): the temperature of device, None means failed to get the data. """ + temp = None try: temp = nvml.nvmlDeviceGetTemperature(self._device_handlers[idx], nvml.NVML_TEMPERATURE_GPU) except Exception as err: logger.warning('Get device temperature failed: {}'.format(str(err))) - temp = None return temp def get_device_power(self, idx): @@ -367,6 +367,7 @@ def get_device_temperature(self, idx): Return: temp (int): the temperature of device, None means failed to get the data. """ + temp = None try: temp = rocml.amdsmi_get_temp_metric( self._device_handlers[idx], rocml.AmdSmiTemperatureType.EDGE, rocml.AmdSmiTemperatureMetric.CURRENT @@ -375,7 +376,6 @@ def get_device_temperature(self, idx): pass except Exception as err: logger.warning('Get device temperature failed: {}'.format(str(err))) - temp = None return temp def get_device_power(self, idx): diff --git a/superbench/config/amd_mi100_hpe.yaml b/superbench/config/amd_mi100_hpe.yaml index 9aec785f5..0388fcda8 100644 --- a/superbench/config/amd_mi100_hpe.yaml +++ b/superbench/config/amd_mi100_hpe.yaml @@ -3,7 +3,7 @@ # Server: # - Product: HPE Apollo 6500 -version: v0.10 +version: v0.11 superbench: enable: null var: diff --git a/superbench/config/amd_mi100_z53.yaml b/superbench/config/amd_mi100_z53.yaml index 7e56d1a33..4ed8addf6 100644 --- a/superbench/config/amd_mi100_z53.yaml +++ b/superbench/config/amd_mi100_z53.yaml @@ -4,7 +4,7 @@ # - Product: G482-Z53 # - Link: https://www.gigabyte.cn/FileUpload/Global/MicroSite/553/G482-Z53.html -version: v0.10 +version: v0.11 superbench: enable: null var: diff --git a/superbench/config/amd_mi300.yaml b/superbench/config/amd_mi300.yaml new file mode 100644 index 000000000..b7aefba63 --- /dev/null +++ b/superbench/config/amd_mi300.yaml @@ -0,0 +1,232 @@ +# SuperBench Config +version: v0.11 +superbench: + enable: null + var: + default_local_mode: &default_local_mode + enable: true + modes: + - name: local + proc_num: 8 + prefix: HIP_VISIBLE_DEVICES={proc_rank} + parallel: yes + default_pytorch_mode: &default_pytorch_mode + enable: true + modes: + - name: torch.distributed + proc_num: 8 + node_num: 1 + frameworks: + - pytorch + common_model_config: &common_model_config + model_ddp_parameter: &model_ddp_param + duration: 0 + num_warmup: 128 + num_steps: 512 + sample_count: 8192 + batch_size: 128 + precision: [float32, float16] + model_action: [train] + pin_memory: yes + num_workers: 0 + benchmarks: + kernel-launch: + <<: *default_local_mode + gemm-flops: + <<: *default_local_mode + parameters: + m: 7680 + n: 8192 + k: 8192 + hipblaslt-gemm: + enable: true + modes: + - name: local + proc_num: 8 + prefix: HIP_VISIBLE_DEVICES={proc_rank} + parallel: yes + parameters: + in_types: ["fp32", "fp16", "bf16", 'fp8'] + tolerant_fail: yes + num_warmup: 100 + num_steps: 1000 + shapes: + - 4096,4096,4096 + - 8192,8192,8192 + - 16384,16384,16384 + rccl-bw: + enable: true + modes: + - name: mpi + proc_num: 8 + node_num: 1 + mca: + pml: ob1 + btl: ^openib + btl_tcp_if_exclude: lo,docker0 + coll_hcoll_enable: 0 + parameters: + maxbytes: 16G + ngpus: 1 + operation: allreduce + cpu-memory-bw-latency: + enable: false + modes: + - name: local + proc_num: 1 + parallel: no + parameters: + tests: + - bandwidth_matrix + - latency_matrix + - max_bandwidth + mem-bw: + enable: true + modes: + - name: local + proc_num: 8 + prefix: HIP_VISIBLE_DEVICES={proc_rank} numactl -N $(({proc_rank}/4)) + parallel: no + ib-loopback: + enable: true + modes: + - name: local + proc_num: 16 + prefix: PROC_RANK={proc_rank} IB_DEVICES=0,1,2,3,4,5,6,7,0,1,2,3,4,5,6,7 numactl -N $(({proc_rank}/8)) -m $(({proc_rank}/8)) + parallel: no + parameters: + msg_size: 8388608 + disk-benchmark: + enable: false + modes: + - name: local + proc_num: 1 + parallel: no + parameters: + block_devices: [] + gpu-copy-bw:correctness: + enable: true + modes: + - name: local + parallel: no + parameters: + mem_type: [htod, dtoh, dtod, one_to_all, all_to_one, all_to_all] + copy_type: [sm, dma] + size: 4096 + num_warm_up: 0 + num_loops: 1 + check_data: true + gpu-copy-bw:perf: + enable: true + modes: + - name: local + parallel: no + parameters: + mem_type: [htod, dtoh, dtod, one_to_all, all_to_one, all_to_all] + copy_type: [sm, dma] + ib-traffic: + enable: false + modes: + - name: mpi + proc_num: 1 + mca: + btl: tcp,self + pml: ob1 + btl_tcp_if_include: ens17f0 + gpcnet-network-test: + enable: false + modes: + - name: mpi + proc_num: 1 + mca: + pml: ucx + btl: ^uct + btl_tcp_if_include: ens17f0 + tcp-connectivity: + enable: false + modes: + - name: local + parallel: no + parameters: + port: 22 + dist-inference: + modes: + - name: mpi + proc_num: 8 + node_num: 1 + mca: + pml: ob1 + btl: ^openib + btl_tcp_if_exclude: lo,docker0 + coll_hcoll_enable: 0 + frameworks: + - pytorch + parameters: + num_layers: 50 + num_warmup: 20 + num_steps: 100 + use_cuda_graph: true + precision: float16 + hidden_size: 128 + input_size: 128 + batch_size: 1024 + model-benchmarks:gpt: + enable: true + <<: *default_pytorch_mode + models: + - gpt2-small + - gpt2-large + parameters: + <<: *model_ddp_param + precision: [float32, float16, fp8_hybrid] + batch_size: 32 + seq_len: 224 + model-benchmarks:bert: + enable: true + <<: *default_pytorch_mode + models: + - bert-base + - bert-large + parameters: + <<: *model_ddp_param + precision: [float32, float16, fp8_hybrid] + seq_len: 224 + model-benchmarks:lstm: + enable: true + <<: *default_pytorch_mode + models: + - lstm + parameters: + <<: *model_ddp_param + batch_size: 1024 + input_size: 224 + hidden_size: 1000 + seq_len: 32 + model-benchmarks:resnet: + enable: true + <<: *default_pytorch_mode + models: + - resnet50 + - resnet101 + - resnet152 + parameters: + <<: *model_ddp_param + batch_size: 384 + model-benchmarks:densenet: + enable: true + <<: *default_pytorch_mode + models: + - densenet169 + - densenet201 + parameters: + <<: *model_ddp_param + model-benchmarks:vgg: + enable: true + <<: *default_pytorch_mode + models: + - vgg11 + - vgg13 + - vgg16 + - vgg19 + parameters: + <<: *model_ddp_param diff --git a/superbench/config/azure/inference/standard_nc64as_t4_v3.yaml b/superbench/config/azure/inference/standard_nc64as_t4_v3.yaml index 7624a86de..b9d57bfbc 100644 --- a/superbench/config/azure/inference/standard_nc64as_t4_v3.yaml +++ b/superbench/config/azure/inference/standard_nc64as_t4_v3.yaml @@ -1,4 +1,4 @@ -version: v0.10 +version: v0.11 superbench: enable: null monitor: diff --git a/superbench/config/azure/inference/standard_nc96ads_a100_v4.yaml b/superbench/config/azure/inference/standard_nc96ads_a100_v4.yaml index befcd1783..224508e0d 100644 --- a/superbench/config/azure/inference/standard_nc96ads_a100_v4.yaml +++ b/superbench/config/azure/inference/standard_nc96ads_a100_v4.yaml @@ -1,4 +1,4 @@ -version: v0.10 +version: v0.11 superbench: enable: null monitor: diff --git a/superbench/config/azure/inference/standard_nv18ads_a10_v5.yaml b/superbench/config/azure/inference/standard_nv18ads_a10_v5.yaml index af19e0a22..e44510b27 100644 --- a/superbench/config/azure/inference/standard_nv18ads_a10_v5.yaml +++ b/superbench/config/azure/inference/standard_nv18ads_a10_v5.yaml @@ -1,4 +1,4 @@ -version: v0.10 +version: v0.11 superbench: enable: null monitor: diff --git a/superbench/config/azure_ndmv4.yaml b/superbench/config/azure_ndmv4.yaml index 3ef0c399a..7d7a8f185 100644 --- a/superbench/config/azure_ndmv4.yaml +++ b/superbench/config/azure_ndmv4.yaml @@ -3,7 +3,7 @@ # Azure NDm A100 v4 # reference: https://docs.microsoft.com/en-us/azure/virtual-machines/ndm-a100-v4-series -version: v0.10 +version: v0.11 superbench: enable: null monitor: diff --git a/superbench/config/azure_ndv4.yaml b/superbench/config/azure_ndv4.yaml index 921a446b8..b095d5c23 100644 --- a/superbench/config/azure_ndv4.yaml +++ b/superbench/config/azure_ndv4.yaml @@ -1,5 +1,5 @@ # SuperBench Config -version: v0.10 +version: v0.11 superbench: enable: null monitor: diff --git a/superbench/config/azure_ndv5.yaml b/superbench/config/azure_ndv5.yaml new file mode 100644 index 000000000..d4e030b4c --- /dev/null +++ b/superbench/config/azure_ndv5.yaml @@ -0,0 +1,305 @@ +# SuperBench Config +version: v0.11 +superbench: + enable: + monitor: + enable: true + sample_duration: 1 + sample_interval: 10 + var: + default_local_mode: &default_local_mode + enable: true + modes: + - name: local + proc_num: 8 + prefix: CUDA_VISIBLE_DEVICES={proc_rank} + parallel: yes + default_pytorch_mode: &default_pytorch_mode + enable: true + modes: + - name: torch.distributed + proc_num: 8 + node_num: 1 + frameworks: + - pytorch + common_model_config: &common_model_config + duration: 0 + num_warmup: 128 + num_steps: 512 + sample_count: 8192 + batch_size: 128 + precision: [float32, float16] + model_action: [train] + pin_memory: yes + num_workers: 0 + benchmarks: + kernel-launch: + <<: *default_local_mode + gemm-flops: + <<: *default_local_mode + parameters: + precision: ["fp64", "fp32", "fp16", "fp64_tc","tf32_tc", "bf16_tc", "fp16_tc", "int8_tc"] + cublaslt-gemm: + <<: *default_local_mode + parameters: + in_types: ['fp8e4m3', 'fp8e5m2', 'fp64', 'fp32', 'fp16', 'bf16', 'int8'] + shapes: + - 4096,4096,4096 + - 8192,8192,8192 + - 16384,16384,16384 + gpu-burn: + enable: false + modes: + - name: local + proc_num: 1 + parallel: no + parameters: + time: 900 + doubles: true + tensor_core: true + nccl-bw:default: + enable: true + modes: + - name: local + proc_num: 1 + parallel: no + parameters: + ngpus: 8 + nccl-bw:gdr-only: + enable: true + modes: + - name: local + proc_num: 1 + parallel: no + env: + NCCL_IB_PCI_RELAXED_ORDERING: '1' + NCCL_NET_GDR_LEVEL: '5' + NCCL_P2P_DISABLE: '1' + NCCL_SHM_DISABLE: '1' + NCCL_MIN_NCHANNELS: '16' + NCCL_IB_DISABLE: '0' + parameters: + ngpus: 8 + nccl-lat:default: + enable: true + modes: + - name: mpi + proc_num: 8 + node_num: 1 + parameters: + maxbytes: 16M + warmup_iters: 20 + iters: 1000 + graph_iters: 1 + ib-loopback: + timeout: *default_timeout + modes: + - name: local + proc_num: 4 + prefix: PROC_RANK={proc_rank} IB_DEVICES=0,2,4,6 NUMA_NODES=0,0,1,1 + parallel: yes + - name: local + proc_num: 4 + prefix: PROC_RANK={proc_rank} IB_DEVICES=1,3,5,7 NUMA_NODES=0,0,1,1 + parallel: yes + cpu-memory-bw-latency: + enable: false + modes: + - name: local + proc_num: 1 + parallel: no + parameters: + tests: + - bandwidth_matrix + - latency_matrix + - max_bandwidth + mem-bw: + enable: true + modes: + - name: local + proc_num: 8 + prefix: CUDA_VISIBLE_DEVICES={proc_rank} numactl -N $(({proc_rank}/2)) + parallel: no + disk-benchmark: + enable: false + modes: + - name: local + proc_num: 1 + parallel: no + parameters: + block_devices: + - /dev/nvme0n1 + - /dev/nvme1n1 + - /dev/nvme2n1 + - /dev/nvme3n1 + - /dev/nvme4n1 + - /dev/nvme5n1 + - /dev/nvme6n1 + - /dev/nvme7n1 + seq_read_runtime: 60 + seq_write_runtime: 60 + seq_readwrite_runtime: 60 + rand_read_runtime: 60 + rand_write_runtime: 60 + rand_readwrite_runtime: 60 + gpu-copy-bw:correctness: + enable: true + modes: + - name: local + parallel: no + parameters: + mem_type: [htod, dtoh, dtod, one_to_all, all_to_one, all_to_all] + copy_type: [sm, dma] + size: 4096 + num_warm_up: 0 + num_loops: 1 + check_data: true + gpu-copy-bw:perf: + enable: true + modes: + - name: local + parallel: no + parameters: + mem_type: [htod, dtoh, dtod, one_to_all, all_to_one, all_to_all] + copy_type: [sm, dma] + cudnn-function: + <<: *default_local_mode + cublas-function: + <<: *default_local_mode + matmul: + <<: *default_local_mode + frameworks: + - pytorch + sharding-matmul: + <<: *default_pytorch_mode + computation-communication-overlap: + <<: *default_pytorch_mode + dist-inference: + enable: true + timeout: 600 + modes: + - name: mpi + proc_num: 8 + node_num: 1 + env: + NCCL_TOPO_FILE: '/opt/microsoft/ndv5-topo.xml' + frameworks: + - pytorch + parameters: + num_layers: 50 + num_warmup: 20 + num_steps: 100 + use_cuda_graph: true + precision: float16 + hidden_size: 128 + input_size: 128 + batch_size: 1024 + ib-traffic: + enable: false + modes: + - name: mpi + proc_num: 8 + parameters: + msg_size: 8388608 + ib_dev: mlx5_$LOCAL_RANK + gpu_dev: $LOCAL_RANK + numa_dev: $((LOCAL_RANK/2)) + gpcnet-network-test: + enable: false + modes: + - name: mpi + proc_num: 1 + mca: + pml: ucx + btl: ^uct + btl_tcp_if_include: eth0 + gpcnet-network-load-test: + enable: false + modes: + - name: mpi + proc_num: 1 + mca: + pml: ucx + btl: ^uct + btl_tcp_if_include: eth0 + tcp-connectivity: + enable: false + modes: + - name: local + parallel: no + parameters: + port: 22 + ort-inference: + <<: *default_local_mode + tensorrt-inference: + <<: *default_local_mode + parameters: + pytorch_models: + - resnet50 + - resnet101 + - resnet152 + - densenet169 + - densenet201 + - bert-base + - bert-large + seq_length: 224 + batch_size: 32 + precision: int8 + model-benchmarks:gpt: + <<: *default_pytorch_mode + models: + - gpt2-small + - gpt2-large + parameters: + <<: *common_model_config + precision: [float32, float16, fp8_hybrid] + batch_size: 32 + seq_len: 224 + model-benchmarks:bert: + <<: *default_pytorch_mode + models: + - bert-base + - bert-large + parameters: + <<: *common_model_config + precision: [float32, float16, fp8_hybrid] + seq_len: 224 + model-benchmarks:lstm: + <<: *default_pytorch_mode + models: + - lstm + parameters: + <<: *common_model_config + batch_size: 1024 + input_size: 224 + hidden_size: 1000 + seq_len: 32 + pin_memory: no + model-benchmarks:resnet: + <<: *default_pytorch_mode + models: + - resnet50 + - resnet101 + - resnet152 + parameters: + <<: *common_model_config + batch_size: 384 + num_steps: 512 + model-benchmarks:densenet: + <<: *default_pytorch_mode + models: + - densenet169 + - densenet201 + parameters: + <<: *common_model_config + pin_memory: no + model-benchmarks:vgg: + <<: *default_pytorch_mode + models: + - vgg11 + - vgg13 + - vgg16 + - vgg19 + parameters: + <<: *common_model_config + pin_memory: no diff --git a/superbench/config/default.yaml b/superbench/config/default.yaml index 9533806cd..601136e9f 100644 --- a/superbench/config/default.yaml +++ b/superbench/config/default.yaml @@ -1,5 +1,5 @@ # SuperBench Config -version: v0.10 +version: v0.11 superbench: enable: null monitor: diff --git a/website/blog/2024-09-20-release-0-11.md b/website/blog/2024-09-20-release-0-11.md new file mode 100644 index 000000000..308e78dc1 --- /dev/null +++ b/website/blog/2024-09-20-release-0-11.md @@ -0,0 +1,44 @@ +--- +slug: release-sb-v0.11 +title: Releasing SuperBench v0.11 +author: Peng Cheng +author_title: SuperBench Team +author_url: https://github.com/cp5555 +author_image_url: https://github.com/cp5555.png +tags: [superbench, announcement, release] +--- + +We are very happy to announce that **SuperBench 0.11.0 version** is officially released today! + +You can install and try superbench by following [Getting Started Tutorial](https://microsoft.github.io/superbenchmark/docs/getting-started/installation). + +## SuperBench 0.11.0 Release Notes + +### SuperBench Improvements + +- Add CUDA 12.4 dockerfile. +- Upgrade nccl version to v2.23.4 and install ucx v1.16.0 in cuda 12.4 dockefile. +- Fix MSCCL build error in CUDA12.4 docker build pipeline. +- Add ROCm6.2 dockerfile. +- Update hpcx link in cuda11.1 dockerfile to fix docker build failure. +- Improve document (Fix metrics name and typos in user tutorial, add BibTeX in README and repo). +- Limit protobuf version to be 3.20.x to fix onnxruntime dependency error. +- Update omegaconf version to 2.3.0 and fix issues caused by omegaconf version update. +- Fix MSCCL build error in CUDA12.4 docker build pipeline. +- Update Docker Exec Command for Persistent HPCX Environment. +- Fix cuda 12.2 dockerfile LD_LIBRARY_PATH issue. +- Use types-setuptools to replace types-pkg_resources. +- Add configuration for NDv5 H100 and AMD MI300x. + +### Micro-benchmark Improvements + +- Add hipblasLt tuning to dist-inference cpp implementation. +- Add support for NVIDIA L4/L40/L40s GPUs in gemm-flops. +- Upgrade mlc to v3.11. + +## Model-benchmark Improvements + +- Support FP8 transformer model training in ROCm6.2 dockerfile. + +### Result Analysis +- Fix bug of failure test and warning of pandas in data diagnosis. diff --git a/website/docusaurus.config.js b/website/docusaurus.config.js index a533084e3..41ed52665 100644 --- a/website/docusaurus.config.js +++ b/website/docusaurus.config.js @@ -101,7 +101,7 @@ module.exports = { announcementBar: { id: 'supportus', content: - '📢 v0.10.0 has been released! ' + + '📢 v0.11.0 has been released! ' + '⭐️ If you like SuperBench, give it a star on GitHub! ⭐️', }, algolia: { diff --git a/website/package-lock.json b/website/package-lock.json index b4652de12..ba0c41478 100644 --- a/website/package-lock.json +++ b/website/package-lock.json @@ -1,6 +1,6 @@ { "name": "superbench-website", - "version": "0.10.0", + "version": "0.11.0", "lockfileVersion": 1, "requires": true, "dependencies": { diff --git a/website/package.json b/website/package.json index f2bb9ed76..8aa4c78b8 100644 --- a/website/package.json +++ b/website/package.json @@ -1,6 +1,6 @@ { "name": "superbench-website", - "version": "0.10.0", + "version": "0.11.0", "private": true, "scripts": { "docusaurus": "docusaurus", @@ -38,4 +38,4 @@ "last 1 safari version" ] } -} \ No newline at end of file +}