mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2026-05-13 12:34:05 +00:00
handle models that need to be downloaded before estimation
This commit is contained in:
@@ -83,6 +83,11 @@ int main(int argc, char ** argv) {
|
||||
return 1;
|
||||
}
|
||||
|
||||
if (params.download_only) {
|
||||
LOG_INF("%s: model downloaded successfully, exiting\n", __func__);
|
||||
return 0;
|
||||
}
|
||||
|
||||
// validate batch size for embeddings
|
||||
// embeddings require all tokens to be processed in a single ubatch
|
||||
// see https://github.com/ggml-org/llama.cpp/issues/12836
|
||||
|
||||
Reference in New Issue
Block a user