mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2026-03-17 16:44:07 +00:00
23
.github/workflows/build-self-hosted.yml
vendored
23
.github/workflows/build-self-hosted.yml
vendored
@@ -97,19 +97,21 @@ jobs:
|
|||||||
vulkaninfo --summary
|
vulkaninfo --summary
|
||||||
GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
|
GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
|
||||||
|
|
||||||
ggml-ci-cpu-amx:
|
# TODO: provision AMX-compatible machine
|
||||||
runs-on: [self-hosted, Linux, CPU, AMX]
|
#ggml-ci-cpu-amx:
|
||||||
|
# runs-on: [self-hosted, Linux, CPU, AMX]
|
||||||
|
|
||||||
steps:
|
# steps:
|
||||||
- name: Clone
|
# - name: Clone
|
||||||
id: checkout
|
# id: checkout
|
||||||
uses: actions/checkout@v6
|
# uses: actions/checkout@v6
|
||||||
|
|
||||||
- name: Test
|
# - name: Test
|
||||||
id: ggml-ci
|
# id: ggml-ci
|
||||||
run: |
|
# run: |
|
||||||
bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
|
# bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
|
||||||
|
|
||||||
|
# TODO: provision AMD GPU machine
|
||||||
# ggml-ci-amd-vulkan:
|
# ggml-ci-amd-vulkan:
|
||||||
# runs-on: [self-hosted, Linux, AMD]
|
# runs-on: [self-hosted, Linux, AMD]
|
||||||
|
|
||||||
@@ -124,6 +126,7 @@ jobs:
|
|||||||
# vulkaninfo --summary
|
# vulkaninfo --summary
|
||||||
# GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
|
# GG_BUILD_VULKAN=1 bash ./ci/run.sh ~/results/llama.cpp /mnt/llama.cpp
|
||||||
|
|
||||||
|
# TODO: provision AMD GPU machine
|
||||||
# ggml-ci-amd-rocm:
|
# ggml-ci-amd-rocm:
|
||||||
# runs-on: [self-hosted, Linux, AMD]
|
# runs-on: [self-hosted, Linux, AMD]
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user