Skip to content
This repository has been archived by the owner on Aug 30, 2024. It is now read-only.

[Neural Speed] Support continuous batching + beam search inference in LLAMA #325

[Neural Speed] Support continuous batching + beam search inference in LLAMA

[Neural Speed] Support continuous batching + beam search inference in LLAMA #325

Triggered via pull request February 29, 2024 09:18
Status Cancelled
Total duration 3m 43s
Artifacts

unit-test-llmruntime.yml

on: pull_request
Fit to window
Zoom out
Zoom in

Annotations

2 errors
unit-test
Canceling since a higher priority waiting request for 'Python Unit Test-145' exists
unit-test
The operation was canceled.