v1.1
This commit is contained in:
@@ -72,20 +72,17 @@ courseware_white_rabbit_repo: "bartowski/WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-GGU
|
||||
courseware_white_rabbit_variants:
|
||||
- ollama_model: "WhiteRabbitNeo"
|
||||
quant: "Q4_K_M"
|
||||
filename: "WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
|
||||
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
|
||||
alias_of_default: true
|
||||
- ollama_model: "WhiteRabbitNeo-BF16"
|
||||
quant: "BF16"
|
||||
filename: "WhiteRabbitNeo-V3-7B-bf16.gguf"
|
||||
- ollama_model: "WhiteRabbitNeo-Q8"
|
||||
quant: "Q8_0"
|
||||
filename: "WhiteRabbitNeo-V3-7B-Q8_0.gguf"
|
||||
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-Q8_0.gguf"
|
||||
- ollama_model: "WhiteRabbitNeo-Q4"
|
||||
quant: "Q4_K_M"
|
||||
filename: "WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
|
||||
- ollama_model: "WhiteRabbitNeo-Q2"
|
||||
quant: "Q2_K"
|
||||
filename: "WhiteRabbitNeo-V3-7B-Q2_K.gguf"
|
||||
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-Q4_K_M.gguf"
|
||||
- ollama_model: "WhiteRabbitNeo-IQ2"
|
||||
quant: "IQ2_M"
|
||||
filename: "WhiteRabbitNeo_WhiteRabbitNeo-V3-7B-IQ2_M.gguf"
|
||||
courseware_ollama_models:
|
||||
- "llama3.2"
|
||||
- "qwen3.5:4b"
|
||||
|
||||
@@ -41,7 +41,7 @@ usage() {
|
||||
cat <<'EOF'
|
||||
Usage: ./download_whiterabbitneo-gguf.sh [--download-only]
|
||||
|
||||
Downloads the WhiteRabbitNeo GGUF variants used in lab 2 with git + git-lfs.
|
||||
Downloads the WhiteRabbitNeo `Q4_K_M`, `Q8_0`, and `IQ2_M` GGUF variants used in lab 2 with git + git-lfs.
|
||||
By default it also registers local Ollama aliases after the files are present.
|
||||
|
||||
Options:
|
||||
|
||||
Reference in New Issue
Block a user