mirror of https://github.com/ggml-org/llama.cpp
51 lines
1.3 KiB
Bash
Executable File
51 lines
1.3 KiB
Bash
Executable File
#!/bin/bash
|
|
set -euo pipefail
|
|
this=$(realpath "$0"); readonly this
|
|
cd "$(dirname "$this")"
|
|
shellcheck "$this"
|
|
|
|
if (( $# != 1 && $# != 2 )); then
|
|
cat >&2 <<'EOF'
|
|
usage:
|
|
ci-run.sh <tmp_dir> [<cache_dir>]
|
|
|
|
This script wraps ci/run.sh:
|
|
* If <tmp_dir> is a ramdisk, you can reduce writes to your SSD. If <tmp_dir> is not a ramdisk, keep in mind that total writes will increase by the size of <cache_dir>.
|
|
(openllama_3b_v2: quantized models are about 30GB)
|
|
* Persistent model and data files are synced to and from <cache_dir>,
|
|
excluding generated .gguf files.
|
|
(openllama_3b_v2: persistent files are about 6.6GB)
|
|
* <cache_dir> defaults to ~/.cache/llama.cpp
|
|
EOF
|
|
exit 1
|
|
fi
|
|
|
|
cd .. # => llama.cpp repo root
|
|
|
|
tmp="$1"
|
|
mkdir -p "$tmp"
|
|
tmp=$(realpath "$tmp")
|
|
echo >&2 "Using tmp=$tmp"
|
|
|
|
cache="${2-$HOME/.cache/llama.cpp}"
|
|
mkdir -p "$cache"
|
|
cache=$(realpath "$cache")
|
|
echo >&2 "Using cache=$cache"
|
|
|
|
_sync() {
|
|
local from="$1"; shift
|
|
local to="$1"; shift
|
|
|
|
echo >&2 "Syncing from $from to $to"
|
|
mkdir -p "$from" "$to"
|
|
rsync -a "$from" "$to" --delete-during "$@"
|
|
}
|
|
|
|
_sync "$(realpath .)/" "$tmp/llama.cpp"
|
|
_sync "$cache/ci-mnt/models/" "$tmp/llama.cpp/ci-mnt/models/"
|
|
|
|
cd "$tmp/llama.cpp"
|
|
bash ci/run.sh ci-out ci-mnt
|
|
|
|
_sync 'ci-mnt/models/' "$cache/ci-mnt/models/" --exclude='*.gguf' -P
|