We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent 2a4bcba commit 11dc109Copy full SHA for 11dc109
examples/llava/llava.cpp
@@ -79,7 +79,13 @@ int main(int argc, char ** argv) {
79
80
llama_backend_init(params.numa);
81
82
- llama_model_params model_params = llama_model_default_params();
+ llama_model_params model_params = llama_model_default_params();
83
+ model_params.n_gpu_layers = params.n_gpu_layers;
84
+ model_params.main_gpu = params.main_gpu;
85
+ model_params.tensor_split = params.tensor_split;
86
+ model_params.use_mmap = params.use_mmap;
87
+ model_params.use_mlock = params.use_mlock;
88
+
89
llama_model * model = llama_load_model_from_file(params.model.c_str(), model_params);
90
if (model == NULL) {
91
fprintf(stderr , "%s: error: unable to load model\n" , __func__);
0 commit comments