mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2026-05-13 12:34:05 +00:00
cont : process images throught the draft context
This commit is contained in:
@@ -2732,6 +2732,13 @@ private:
|
||||
continue;
|
||||
}
|
||||
|
||||
if (ctx_drft) {
|
||||
res = input_tokens.process_chunk(ctx_main, mctx, slot.prompt.n_tokens(), slot.prompt.tokens.pos_next(), slot.id, n_tokens_out);
|
||||
if (res != 0) {
|
||||
GGML_ABORT("failed to process multi-modal data on draft context\n");
|
||||
}
|
||||
}
|
||||
|
||||
slot.n_prompt_tokens_processed += n_tokens_out;
|
||||
|
||||
// add the image chunk to cache
|
||||
|
||||
Reference in New Issue
Block a user