Skip to content

Actions: intel/llm-on-ray

Merge

Actions

Loading...
Loading

Show workflow options

Create status badge

Loading
139 workflow runs
139 workflow runs

Filter by Event

Filter by Status

Filter by Branch

Filter by Actor

fix vllm fork version (#237)
Merge #114: Commit e113b8c pushed by carsonwang
May 29, 2024 01:44 58m 10s main
May 29, 2024 01:44 58m 10s
[Inference] Enable vllm on HPU (#232)
Merge #113: Commit 0e2b8f8 pushed by carsonwang
May 24, 2024 07:24 1h 33m 3s main
May 24, 2024 07:24 1h 33m 3s
add ignore_eos=False (#235)
Merge #112: Commit b37f275 pushed by carsonwang
May 24, 2024 05:38 38m 18s main
May 24, 2024 05:38 38m 18s
[CI] Change miniconda to miniforge (#230)
Merge #111: Commit a0bc380 pushed by xwu99
May 24, 2024 05:38 13s main
May 24, 2024 05:38 13s
May 24, 2024 05:17 22m 17s
Fix max_new_tokens for HPU predictcor (#226)
Merge #109: Commit 30b3204 pushed by carsonwang
May 20, 2024 07:01 41m 51s main
May 20, 2024 07:01 41m 51s
[Inference ] Integrate chat template in llm-on-ray (#199)
Merge #108: Commit 620800f pushed by carsonwang
May 16, 2024 03:42 50m 50s main
May 16, 2024 03:42 50m 50s
[Inference] Remove simple mode and use OpenAI API (#220)
Merge #107: Commit 9142112 pushed by carsonwang
May 15, 2024 08:08 37m 52s main
May 15, 2024 08:08 37m 52s
Refactor non-streaming generate (#209)
Merge #106: Commit 73ae86c pushed by xwu99
May 14, 2024 09:16 32m 0s main
May 14, 2024 09:16 32m 0s
May 13, 2024 05:56 43m 19s
May 13, 2024 03:00 44m 32s
May 11, 2024 01:18 34m 47s
[CI] Add llama2-70b inference workflow (#208)
Merge #102: Commit cc1556d pushed by Deegue
May 10, 2024 09:08 1h 14m 31s main
May 10, 2024 09:08 1h 14m 31s
Fix openai response for vLLM (#213)
Merge #101: Commit e093eb8 pushed by xwu99
May 10, 2024 07:46 1h 22m 20s main
May 10, 2024 07:46 1h 22m 20s
[Workflow] Unify Docker operations into bash (#123)
Merge #100: Commit 93689e0 pushed by xwu99
May 10, 2024 07:06 41m 28s main
May 10, 2024 07:06 41m 28s
Update VLLM installation script and documentation (#212)
Merge #99: Commit 1466eee pushed by KepingYan
May 10, 2024 01:31 34m 14s main
May 10, 2024 01:31 34m 14s
Fix CI proxy issue (#210)
Merge #98: Commit ae27555 pushed by xwu99
May 9, 2024 03:55 13m 33s main
May 9, 2024 03:55 13m 33s
Add support for HPU quantization and torch_compile (#207)
Merge #97: Commit 466abf3 pushed by carsonwang
May 8, 2024 03:36 21h 51m 22s main
May 8, 2024 03:36 21h 51m 22s
April 29, 2024 08:10 31m 46s
[CI] Add Gaudi2 inference workflow (#185)
Merge #95: Commit 64c1c1f pushed by Deegue
April 29, 2024 07:31 39m 38s main
April 29, 2024 07:31 39m 38s
Add config files for Llama3 on HPU (#201)
Merge #94: Commit 49dc7ee pushed by carsonwang
April 29, 2024 02:36 42m 49s main
April 29, 2024 02:36 42m 49s
Update vllm to use latest upstream to support CPU (#179)
Merge #93: Commit 4e81eb2 pushed by carsonwang
April 28, 2024 07:15 36m 15s main
April 28, 2024 07:15 36m 15s
[TEST]Add Openai protocol test (#167)
Merge #92: Commit 83803da pushed by carsonwang
April 28, 2024 01:27 34m 13s main
April 28, 2024 01:27 34m 13s
fix ci ipex version (#206)
Merge #91: Commit 971fd4f pushed by kira-lin
April 26, 2024 06:40 35m 47s main
April 26, 2024 06:40 35m 47s
fix bug: deal with string in config (#200)
Merge #90: Commit 2bd19c5 pushed by KepingYan
April 23, 2024 06:50 35m 24s main
April 23, 2024 06:50 35m 24s