Skip to content

Commit 4ff72d9

Browse files
dm4hydai
authored andcommitted
[CI] use Q2_K Command-R model to reduce disk usage on CI
Signed-off-by: dm4 <dm4@secondstate.io>
1 parent fcd5a45 commit 4ff72d9

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

‎.github/workflows/llama.yml‎

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -119,11 +119,11 @@ jobs:
119119
run: |
120120
test -f ~/.wasmedge/env && source ~/.wasmedge/env
121121
cd wasmedge-ggml/command-r
122-
curl -LO https://huggingface.co/andrewcanis/c4ai-command-r-v01-GGUF/resolve/main/c4ai-command-r-v01-Q5_K_M.gguf
122+
curl -LO https://huggingface.co/andrewcanis/c4ai-command-r-v01-GGUF/resolve/main/c4ai-command-r-v01-Q2_K.gguf
123123
cargo build --target wasm32-wasi --release
124124
time wasmedge --dir .:. \
125125
--env n_gpu_layers="$NGL" \
126-
--nn-preload default:GGML:AUTO:c4ai-command-r-v01-Q5_K_M.gguf \
126+
--nn-preload default:GGML:AUTO:c4ai-command-r-v01-Q2_K.gguf \
127127
target/wasm32-wasi/release/wasmedge-ggml-command-r.wasm \
128128
default \
129129
'<|START_OF_TURN_TOKEN|><|USER_TOKEN|>What is the capital of the United States?<|END_OF_TURN_TOKEN|><|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>'

0 commit comments

Comments
 (0)