mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2026-05-01 14:44:05 +00:00
webui : store reasoning_content so it is sent back in subsequent requests (#21249)
This commit is contained in:
File diff suppressed because one or more lines are too long
Binary file not shown.
@@ -474,6 +474,7 @@ class AgenticStore {
|
||||
sessionMessages.push({
|
||||
role: MessageRole.ASSISTANT,
|
||||
content: turnContent || undefined,
|
||||
reasoning_content: turnReasoningContent || undefined,
|
||||
tool_calls: normalizedCalls
|
||||
});
|
||||
|
||||
|
||||
@@ -41,6 +41,7 @@ export type AgenticMessage =
|
||||
| {
|
||||
role: MessageRole.ASSISTANT;
|
||||
content?: string | ApiChatMessageContentPart[];
|
||||
reasoning_content?: string;
|
||||
tool_calls?: AgenticToolCallPayload[];
|
||||
}
|
||||
| {
|
||||
|
||||
Reference in New Issue
Block a user