Focus local lab deployment on Linux and WSL
This commit is contained in:
@@ -23,18 +23,6 @@
|
||||
gpu_type: "{{ 'nvidia' if nvidia_smi_output.rc == 0 else 'none' }}"
|
||||
when: is_wsl | default(false) or ansible_os_family == "Debian"
|
||||
|
||||
- name: Check for Metal GPU on macOS
|
||||
ansible.builtin.command: system_profiler SPDisplaysDataType
|
||||
register: metal_check
|
||||
changed_when: false
|
||||
failed_when: false
|
||||
when: ansible_os_family == "Darwin"
|
||||
|
||||
- name: Set GPU type for macOS
|
||||
ansible.builtin.set_fact:
|
||||
gpu_type: "metal"
|
||||
when: ansible_os_family == "Darwin" and metal_check.rc == 0
|
||||
|
||||
- name: Display detected GPU type
|
||||
ansible.builtin.debug:
|
||||
msg: "llama.cpp GPU type: {{ gpu_type | default('none') }}"
|
||||
@@ -58,7 +46,6 @@
|
||||
{{
|
||||
not llama_cpp_stat.stat.exists or
|
||||
(gpu_type == 'nvidia' and existing_gpu_check.stdout != 'cuda') or
|
||||
(gpu_type == 'metal' and existing_gpu_check.stdout != 'metal') or
|
||||
(gpu_type == 'amd' and existing_gpu_check.stdout != 'amd')
|
||||
}}
|
||||
|
||||
@@ -120,19 +107,6 @@
|
||||
when: gpu_type == 'amd' and cmake_configured.rc != 0
|
||||
become: no
|
||||
|
||||
- name: Configure llama.cpp for Metal (macOS)
|
||||
ansible.builtin.command:
|
||||
argv:
|
||||
- cmake
|
||||
- ..
|
||||
- -G Ninja
|
||||
- -DCMAKE_BUILD_TYPE=Release
|
||||
- -DGGML_METAL=on
|
||||
args:
|
||||
chdir: "{{ llmlab_base }}/lab2/llama.cpp/build"
|
||||
when: gpu_type == 'metal' and cmake_configured.rc != 0
|
||||
become: no
|
||||
|
||||
- name: Configure llama.cpp for CPU only
|
||||
ansible.builtin.command:
|
||||
argv:
|
||||
|
||||
Reference in New Issue
Block a user