Skip to content
This repository has been archived by the owner on Aug 30, 2024. It is now read-only.

Actions: intel/neural-speed

Format Scan

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
348 workflow run results
348 workflow run results

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

[LLM Runtime] Support 3bits & 4bits GPTQ for gpt-j
Format Scan #278: Pull request #100 synchronize by Zhenzhong1
January 31, 2024 05:35 11m 8s zhenzhong/gptq-gptj
January 31, 2024 05:35 11m 8s
[LLM Runtime] Support 3bits & 4bits GPTQ for gpt-j
Format Scan #276: Pull request #100 synchronize by Zhenzhong1
January 30, 2024 10:13 11m 17s zhenzhong/gptq-gptj
January 30, 2024 10:13 11m 17s
[LLM Runtime] Support 3bits & 4bits GPTQ for gpt-j
Format Scan #275: Pull request #100 synchronize by Zhenzhong1
January 30, 2024 09:13 11m 33s zhenzhong/gptq-gptj
January 30, 2024 09:13 11m 33s
[LLM Runtime] Support 3bits & 4bits GPTQ for gpt-j
Format Scan #274: Pull request #100 synchronize by Zhenzhong1
January 30, 2024 07:02 11m 8s zhenzhong/gptq-gptj
January 30, 2024 07:02 11m 8s
[LLM Runtime] Support 3bits & 4bits GPTQ for gpt-j
Format Scan #273: Pull request #100 synchronize by Zhenzhong1
January 30, 2024 04:41 11m 10s zhenzhong/gptq-gptj
January 30, 2024 04:41 11m 10s
[LLM Runtime] Support 3bits & 4bits GPTQ for gpt-j
Format Scan #272: Pull request #100 opened by Zhenzhong1
January 30, 2024 03:54 11m 2s zhenzhong/gptq-gptj
January 30, 2024 03:54 11m 2s
[LLM Runtime] Support loading models from HF directly.
Format Scan #268: Pull request #93 synchronize by Zhenzhong1
January 25, 2024 09:25 11m 25s zhenzhong/online_load
January 25, 2024 09:25 11m 25s
[LLM Runtime] Support loading models from HF directly.
Format Scan #267: Pull request #93 synchronize by Zhenzhong1
January 25, 2024 09:08 10m 50s zhenzhong/online_load
January 25, 2024 09:08 10m 50s
[LLM Runtime] Support loading models from HF directly.
Format Scan #266: Pull request #93 opened by Zhenzhong1
January 25, 2024 07:52 11m 19s zhenzhong/online_load
January 25, 2024 07:52 11m 19s
Cont batching mha++
Format Scan #265: Pull request #89 synchronize by DDEle
January 25, 2024 05:57 10m 47s cont_batching_mha++
January 25, 2024 05:57 10m 47s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #264: Pull request #78 synchronize by intellinjun
January 25, 2024 02:11 10m 38s phi2
January 25, 2024 02:11 10m 38s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #263: Pull request #78 synchronize by intellinjun
January 25, 2024 02:08 3m 35s phi2
January 25, 2024 02:08 3m 35s
[Neural Speed] Fix a blocker on Windows platforms
Format Scan #262: Pull request #92 opened by aahouzi
January 24, 2024 14:57 9m 48s aahouzi:bug-fix-windows
January 24, 2024 14:57 9m 48s
[Neural Speed] Improvements to run.py script
Format Scan #261: Pull request #87 synchronize by aahouzi
January 24, 2024 09:47 9m 50s aahouzi:main
January 24, 2024 09:47 9m 50s
3bit weight in BesTLA
Format Scan #260: Pull request #85 synchronize by zhewang1-intc
January 24, 2024 09:28 3m 18s 3bit_wei
January 24, 2024 09:28 3m 18s
[Neural Speed] Cont Batching in Offline and Server (GPT-J & Beam Search First)
Format Scan #259: Pull request #69 synchronize by DDEle
January 24, 2024 09:01 10m 31s cont_batching
January 24, 2024 09:01 10m 31s
3bit weight in BesTLA
Format Scan #258: Pull request #85 synchronize by zhewang1-intc
January 24, 2024 08:34 2m 3s 3bit_wei
January 24, 2024 08:34 2m 3s
3bit weight in BesTLA
Format Scan #257: Pull request #85 synchronize by zhewang1-intc
January 24, 2024 08:30 2m 3s 3bit_wei
January 24, 2024 08:30 2m 3s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #256: Pull request #78 synchronize by intellinjun
January 24, 2024 08:13 10m 12s phi2
January 24, 2024 08:13 10m 12s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #255: Pull request #78 synchronize by intellinjun
January 24, 2024 07:59 9m 57s phi2
January 24, 2024 07:59 9m 57s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #254: Pull request #78 synchronize by intellinjun
January 24, 2024 07:45 10m 13s phi2
January 24, 2024 07:45 10m 13s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #253: Pull request #78 synchronize by intellinjun
January 24, 2024 07:32 10m 24s phi2
January 24, 2024 07:32 10m 24s
[Neural Speed] Cont Batching in Offline and Server (GPT-J & Beam Search First)
Format Scan #252: Pull request #69 synchronize by zhentaoyu
January 24, 2024 07:19 12m 27s cont_batching
January 24, 2024 07:19 12m 27s
[LLM Runtime] Enable phi-2&phi-1.5&phi-1
Format Scan #251: Pull request #78 synchronize by intellinjun
January 24, 2024 07:05 24m 12s phi2
January 24, 2024 07:05 24m 12s
Cont batching mha++
Format Scan #250: Pull request #89 synchronize by DDEle
January 24, 2024 06:44 10m 48s cont_batching_mha++
January 24, 2024 06:44 10m 48s