This commit is contained in:
Codex
2026-03-31 19:46:14 -06:00
parent a50172594b
commit d2c7941aaa
8 changed files with 424 additions and 12 deletions
+6 -9
View File
@@ -72,20 +72,17 @@ courseware_white_rabbit_repo: "bartowski/WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-GGU
courseware_white_rabbit_variants:
- ollama_model: "WhiteRabbitNeo"
quant: "Q4_K_M"
filename: "WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
alias_of_default: true
- ollama_model: "WhiteRabbitNeo-BF16"
quant: "BF16"
filename: "WhiteRabbitNeo-V3-7B-bf16.gguf"
- ollama_model: "WhiteRabbitNeo-Q8"
quant: "Q8_0"
filename: "WhiteRabbitNeo-V3-7B-Q8_0.gguf"
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-Q8_0.gguf"
- ollama_model: "WhiteRabbitNeo-Q4"
quant: "Q4_K_M"
filename: "WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
- ollama_model: "WhiteRabbitNeo-Q2"
quant: "Q2_K"
filename: "WhiteRabbitNeo-V3-7B-Q2_K.gguf"
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
- ollama_model: "WhiteRabbitNeo-IQ2"
quant: "IQ2_M"
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-IQ2_M.gguf"
courseware_ollama_models:
- "llama3.2"
- "qwen3.5:4b"
@@ -41,7 +41,7 @@ usage() {
cat <<'EOF'
Usage: ./download_whiterabbitneo-gguf.sh [--download-only]
Downloads the WhiteRabbitNeo GGUF variants used in lab 2 with git + git-lfs.
Downloads the WhiteRabbitNeo `Q4_K_M`, `Q8_0`, and `IQ2_M` GGUF variants used in lab 2 with git + git-lfs.
By default it also registers local Ollama aliases after the files are present.
Options: