Focus local lab deployment on Linux and WSL
This commit is contained in:
@@ -78,7 +78,7 @@
|
||||
|
||||
- name: Set llama.cpp backend flag
|
||||
set_fact:
|
||||
courseware_llama_backend_flag: "{{ '-DGGML_METAL=ON' if ansible_system == 'Darwin' else '-DGGML_CUDA=ON' }}"
|
||||
courseware_llama_backend_flag: "-DGGML_CUDA=ON"
|
||||
|
||||
- name: Set llama.cpp build parallelism
|
||||
set_fact:
|
||||
|
||||
Reference in New Issue
Block a user