Files
llama.cpp/CODEOWNERS
Line: 4: incorrect codeowner user: ngxson Line: 4: no users/groups matched Line: 5: incorrect codeowner user: CISC Line: 5: no users/groups matched Line: 6: incorrect codeowner user: CISC Line: 6: no users/groups matched Line: 7: incorrect codeowner user: ggerganov Line: 7: no users/groups matched Line: 8: incorrect codeowner user: ggerganov Line: 8: no users/groups matched Line: 9: incorrect codeowner user: ggerganov Line: 9: no users/groups matched Line: 10: incorrect codeowner user: ggerganov Line: 10: no users/groups matched Line: 11: incorrect codeowner user: ggerganov Line: 11: no users/groups matched Line: 12: incorrect codeowner user: ggerganov Line: 12: no users/groups matched Line: 13: incorrect codeowner user: pwilkin Line: 13: no users/groups matched Line: 14: incorrect codeowner user: pwilkin Line: 14: no users/groups matched Line: 15: incorrect codeowner user: pwilkin Line: 15: no users/groups matched Line: 16: incorrect codeowner user: aldehir Line: 16: no users/groups matched Line: 17: incorrect codeowner user: ggerganov Line: 17: no users/groups matched Line: 18: incorrect codeowner user: ggerganov Line: 18: no users/groups matched Line: 19: incorrect codeowner user: angt Line: 19: no users/groups matched Line: 20: incorrect codeowner user: ngxson Line: 20: incorrect codeowner user: CISC Line: 20: incorrect codeowner user: aldehir Line: 20: no users/groups matched Line: 21: incorrect codeowner user: ggerganov Line: 21: no users/groups matched Line: 22: incorrect codeowner user: ggerganov Line: 22: no users/groups matched Line: 23: incorrect codeowner user: srogmann Line: 23: no users/groups matched Line: 24: incorrect codeowner user: aldehir Line: 24: no users/groups matched Line: 25: incorrect codeowner user: ggerganov Line: 25: no users/groups matched Line: 26: incorrect codeowner user: ggerganov Line: 26: no users/groups matched Line: 27: incorrect codeowner user: aldehir Line: 27: no users/groups matched Line: 28: incorrect codeowner user: CISC Line: 28: no users/groups matched Line: 29: incorrect codeowner user: ggerganov Line: 29: no users/groups matched Line: 30: incorrect codeowner user: ggerganov Line: 30: no users/groups matched Line: 31: incorrect codeowner user: ggerganov Line: 31: no users/groups matched Line: 32: incorrect codeowner user: danbev Line: 32: incorrect codeowner user: pwilkin Line: 32: no users/groups matched Line: 33: incorrect codeowner user: ggerganov Line: 33: no users/groups matched Line: 34: incorrect codeowner user: am17an Line: 34: no users/groups matched Line: 35: incorrect codeowner user: ggerganov Line: 35: no users/groups matched Line: 36: incorrect codeowner user: ggerganov Line: 36: no users/groups matched Line: 37: incorrect codeowner user: ggerganov Line: 37: no users/groups matched Line: 38: incorrect codeowner user: ggerganov Line: 38: no users/groups matched Line: 39: incorrect codeowner user: ggerganov Line: 39: no users/groups matched Line: 40: incorrect codeowner user: ggerganov Line: 40: incorrect codeowner user: hanyin-arm Line: 40: incorrect codeowner user: naco-siren Line: 40: no users/groups matched Line: 41: incorrect codeowner user: ggerganov Line: 41: no users/groups matched Line: 42: incorrect codeowner user: ggerganov Line: 42: no users/groups matched Line: 43: incorrect codeowner user: ggerganov Line: 43: no users/groups matched Line: 44: incorrect codeowner user: JohannesGaessler Line: 44: no users/groups matched Line: 45: incorrect codeowner user: danbev Line: 45: no users/groups matched Line: 46: incorrect codeowner user: ggerganov Line: 46: no users/groups matched Line: 47: incorrect codeowner user: ggerganov Line: 47: no users/groups matched Line: 48: incorrect codeowner user: ggerganov Line: 48: no users/groups matched Line: 49: incorrect codeowner user: ggerganov Line: 49: no users/groups matched Line: 50: incorrect codeowner user: ggerganov Line: 50: no users/groups matched Line: 51: incorrect codeowner user: ggerganov Line: 51: no users/groups matched Line: 52: incorrect codeowner user: ggerganov Line: 52: no users/groups matched Line: 53: incorrect codeowner user: ggerganov Line: 53: no users/groups matched Line: 54: incorrect codeowner user: ggerganov Line: 54: no users/groups matched Line: 55: incorrect codeowner user: ggerganov Line: 55: no users/groups matched Line: 56: incorrect codeowner user: alex-spacemit Line: 56: no users/groups matched Line: 57: incorrect codeowner user: JohannesGaessler Line: 57: no users/groups matched Line: 58: incorrect codeowner user: JohannesGaessler Line: 58: incorrect codeowner user: am17an Line: 58: no users/groups matched Line: 59: incorrect codeowner user: JohannesGaessler Line: 59: no users/groups matched Line: 60: incorrect codeowner user: JohannesGaessler Line: 60: no users/groups matched Line: 61: incorrect codeowner user: JohannesGaessler Line: 61: no users/groups matched Line: 62: incorrect codeowner user: IMbackK Line: 62: no users/groups matched Line: 63: incorrect codeowner user: IMbackK Line: 63: no users/groups matched Line: 64: incorrect codeowner user: IMbackK Line: 64: no users/groups matched Line: 65: incorrect codeowner user: ggerganov Line: 65: no users/groups matched Line: 66: incorrect codeowner user: ggerganov Line: 66: no users/groups matched Line: 67: incorrect codeowner user: lhez Line: 67: incorrect codeowner user: max-krasnyansky Line: 67: no users/groups matched Line: 68: incorrect codeowner user: max-krasnyansky Line: 68: incorrect codeowner user: lhez Line: 68: no users/groups matched Line: 69: incorrect codeowner user: JohannesGaessler Line: 69: no users/groups matched Line: 70: incorrect codeowner user: ggerganov Line: 70: no users/groups matched Line: 71: incorrect codeowner user: rgerganov Line: 71: no users/groups matched Line: 72: incorrect codeowner user: ggerganov Line: 72: no users/groups matched Line: 73: incorrect codeowner user: 0cc4m Line: 73: no users/groups matched Line: 74: incorrect codeowner user: kpouget Line: 74: no users/groups matched Line: 75: incorrect codeowner user: reeselevine Line: 75: no users/groups matched Line: 76: incorrect codeowner user: taronaeo Line: 76: incorrect codeowner user: Andreas-Krebbel Line: 76: incorrect codeowner user: AlekseiNikiforovIBM Line: 76: no users/groups matched Line: 77: incorrect codeowner user: cavusmustafa Line: 77: incorrect codeowner user: wine99 Line: 77: no users/groups matched Line: 78: incorrect codeowner user: ggerganov Line: 78: no users/groups matched Line: 79: incorrect codeowner user: ggerganov Line: 79: no users/groups matched Line: 80: incorrect codeowner user: JohannesGaessler Line: 80: incorrect codeowner user: Green-Sky Line: 80: no users/groups matched Line: 81: incorrect codeowner user: CISC Line: 81: no users/groups matched Line: 82: incorrect codeowner user: ggerganov Line: 82: no users/groups matched Line: 83: incorrect codeowner user: ggerganov Line: 83: no users/groups matched Line: 84: incorrect codeowner user: ggerganov Line: 84: no users/groups matched Line: 85: incorrect codeowner user: ggerganov Line: 85: no users/groups matched Line: 86: incorrect codeowner user: ggerganov Line: 86: no users/groups matched Line: 87: incorrect codeowner user: CISC Line: 87: no users/groups matched Line: 88: incorrect codeowner user: CISC Line: 88: no users/groups matched Line: 89: incorrect codeowner user: ngxson Line: 89: no users/groups matched Line: 90: incorrect codeowner user: CISC Line: 90: no users/groups matched Line: 91: incorrect codeowner user: CISC Line: 91: no users/groups matched Line: 92: incorrect codeowner user: CISC Line: 92: no users/groups matched Line: 93: incorrect codeowner user: CISC Line: 93: no users/groups matched Line: 94: incorrect codeowner user: ggerganov Line: 94: no users/groups matched Line: 95: incorrect codeowner user: pwilkin Line: 95: no users/groups matched Line: 96: incorrect codeowner user: ggerganov Line: 96: no users/groups matched Line: 97: incorrect codeowner user: ngxson Line: 97: no users/groups matched Line: 98: incorrect codeowner user: ggerganov Line: 98: no users/groups matched Line: 99: incorrect codeowner user: ngxson Line: 99: no users/groups matched Line: 100: incorrect codeowner user: ggerganov Line: 100: no users/groups matched Line: 101: incorrect codeowner user: pwilkin Line: 101: no users/groups matched Line: 102: incorrect codeowner user: ggerganov Line: 102: no users/groups matched Line: 103: incorrect codeowner user: rgerganov Line: 103: no users/groups matched Line: 104: incorrect codeowner user: ngxson Line: 104: incorrect codeowner user: ggerganov Line: 104: no users/groups matched Line: 105: incorrect codeowner user: allozaur Line: 105: no users/groups matched Line: 106: incorrect codeowner user: ggerganov Line: 106: no users/groups matched Line: 107: incorrect codeowner user: ggerganov Line: 107: no users/groups matched Line: 108: incorrect codeowner user: ggerganov Line: 108: no users/groups matched Line: 109: incorrect codeowner user: ggerganov Line: 109: no users/groups matched Line: 110: incorrect codeowner user: ggerganov Line: 110: no users/groups matched Line: 111: incorrect codeowner user: ggerganov Line: 111: no users/groups matched Line: 112: incorrect codeowner user: ggerganov Line: 112: no users/groups matched Line: 113: incorrect codeowner user: ggerganov Line: 113: no users/groups matched Line: 114: incorrect codeowner user: ggerganov Line: 114: no users/groups matched Line: 115: incorrect codeowner user: danbev Line: 115: no users/groups matched Line: 116: incorrect codeowner user: CISC Line: 116: no users/groups matched
Zijun Yu 9789c4ecdc ggml : add OpenVINO backend (#15307)
* Update build doc

* Add cgraph tensor output name to OV op name

* Update openvino build instructions

* Add initial NPU support

* draft NPU support version 2: prefill + kvcache

* NPU support version 2: prefill + kvcache

* Change due to ggml cgraph changes, not correct yet

* Change due to ggml cgraph changes, llama-3.2 CPU work

* Add AMD64 to CMakeLists

* Change due to ggml cgraph changes, all device work

* Refactor: clean, fix warning

* Update clang-format

* Statful transformation for CPU GPU

* Add SwiGLU

* Fuse to SDPA

* Replace Concat with Broadcast in MulMat for GQA

* Pull out indices creation for kv cache update

* Refactor: remove past_token_len from extra_inputs

* Fix Phi3 SwiGLU and SoftMax

* Pull out sin cos from rope

* Reduce memory: free ov weights node after graph conversion

* Fix CPY due to cgraph change

* Added OpenVINO CI/CD. Updated docs

* Fix llama-cli

* Fix Phi3 ROPE; Add test-backend-ops

* Fix NPU

* Fix llama-bench; Clang-format

* Fix llama-perplexity

* temp. changes for mark decomp

* matmul in fp32

* mulmat input conversion fix

* mulmat type conversion update

* add mark decomp pass

* Revert changes in fuse_to_sdpa

* Update build.md

* Fix test-backend-ops

* Skip test-thread-safety; Run ctest only in ci/run.sh

* Use CiD for NPU

* Optimize tensor conversion, improve TTFT

* Support op SET_ROWS

* Fix NPU

* Remove CPY

* Fix test-backend-ops

* Minor updates for raising PR

* Perf: RMS fused to OV internal RMS op

* Fix after rebasing

- Layout of cache k and cache v are unified: [seq, n_head, head_size]
- Add CPY and FLASH_ATTN_EXT, flash attn is not used yet
- Skip test-backend-ops due to flash attn test crash
- Add mutex around graph conversion to avoid test-thread-safety fali in the future
- Update NPU config
- Update GPU config to disable SDPA opt to make phi-3 run

* Change openvino device_type to GPU; Enable flash_attn

* Update supports_buft and supports_op for quantized models

* Add quant weight conversion functions from genai gguf reader

* Quant models run with accuracy issue

* Fix accuracy: disable cpu_repack

* Fix CI; Disable test-backend-ops

* Fix Q4_1

* Fix test-backend-ops: Treat quantized tensors as weights

* Add NPU Q4_0 support

* NPU perf: eliminate zp

* Dequantize q4_1 q4_k q6_k for NPU

* Add custom quant type: q8_1_c, q4_0_128

* Set m_is_static=false as default in decoder

* Simpilfy translation of get_rows

* Fix after rebasing

* Improve debug util; Eliminate nop ReshapeReshape

* STYLE: make get_types_to_requant a function

* Support BF16 model

* Fix NPU compile

* WA for npu 1st token acc issue

* Apply EliminateZP only for npu

* Add GeGLU

* Fix Hunyuan

* Support iSWA

* Fix NPU accuracy

* Fix ROPE accuracy when freq_scale != 1

* Minor: not add attention_size_swa for non-swa model

* Minor refactor

* Add Q5_K to support phi-3-q4_k_m

* Requantize Q6_K (gs16) to gs32 on GPU

* Fix after rebasing

* Always apply Eliminate_ZP to fix GPU compile issue on some platforms

* kvcachefusion support

* env variable GGML_OPENVINO_DISABLE_SDPA_OPTIMIZATION added

* Fix for Phi3

* Fix llama-cli (need to run with --no-warmup)

* Fix add_sliced_mask; Revert mulmat, softmax; Remove input attention_size, iSWA model not working

* fix after rebasing

* Fix llama-3-8b and phi3-mini q4_0 NPU

* Update to OV-2025.3 and CMakeLists.txt

* Add OV CI cache

* Apply CISC review and update CI to OV2025.3

* Update CI to run OV dep install before build

* Update OV dockerfile to use OV2025.3 and update build docs

* Style: use switch in supports_ops

* Style: middle ptr and ref align, omit optional struct keyword

* NPU Unify PD (#14)

* Stateless. Fix llama-cli llama-server

* Simplify broadcast op in attention

* Replace get_output_tensor+memcpy with set_output_tensor

* NPU unify PD. Unify dynamic and static dims

* Clean placeholders in ggml-openvino.cpp

* NPU unify PD (handled internally)

* change graph to 4d, support multi sequences

* Fix llama-bench

* Fix NPU

* Update ggml-decoder.cpp

Hitting error while compiling on windows:

error C3861: 'unsetenv': identifier not found

Reason: unsetenv() is a POSIX function; it doesn’t exist on Windows. Visual Studio (MSVC) won’t recognize it.

Proposed fix: Use _putenv_s() (Windows equivalent)
This is supported by MSVC and achieves the same effect: it removes the environment variable from the process environment.

This keeps cross-platform compatibility.

* Update ggml-decoder.cpp

* Update ggml-decoder.cpp

* Update ggml-decoder.cpp

* Update ggml-decoder.cpp

* Update ggml-decoder.cpp

* Remove the second decoder for node. Moving the function into the model decoder

* Fix error for naive

* NPU prefill chunking

* NPU fix llama-bench

* fallback naive run with accuracy issue

* NPU support llma-perplexity -b 512 --no-warmup

* Refactor: split ov_graph_compute for dynamic and static

* remove unused API GgmlOvDecoder::get_output_stride(const std::string & name)

* minor update due to ov 2025.4

* remove unused API GgmlOvDecoder::get_output_names()

* remove unused API get_output_shape(const std::string & name)

* Modified API GgmlOvDecoder::get_output_type(const std::string & name)

* Removed API GgmlOvDecoder::get_output_op_params(const std::string & name)

* Removed API get_output_ggml_tensor(const std::string & name)

* Removed API m_outputs

* Removed m_output_names

* Removed API GgmlOvDecoder::get_input_names()

* Removed API GgmlOvDecoder::get_input_stride(const std::string& name)

* Removed API get_input_type

* Removed API get_input_type

* Removed API GgmlOvDecoder::get_input_shape(const std::string & name)

* Removed API GgmlOvDecoder::get_input_op_params(const std::string & name)

* Fix error for decoder cache

* Reuse cached decoder

* GPU remove Q6_K requantization

* NPU fix wrong model output shape

* NPU fix q4 perf regression

* Remove unused variable nodes

* Fix decoder can_reuse for llama-bench

* Update build.md for Windows

* backend buffer: allocate on host

* Use shared_buffer for GPU NPU; Refactor

* Add ov_backend_host_buffer; Use cached remote context

* Put kvcache on GPU

* Use ggml_aligned_malloc

* only use remote tensor for kvcache

* only use remote tensor for kvcache for GPU

* FIX: use remote tensor from singleton

* Update build.md to include OpenCL

* NPU always requant to q4_0_128

* Optimize symmetric quant weight extraction: use single zp

* Use Q8_0_C in token embd, lm_head, and for 5 and 6 bits quant

* Update build.md

* Support -ctk f32

* Initial stateful graph support

* Update ggml/src/ggml-openvino/ggml-decoder.cpp

Co-authored-by: Yamini Nimmagadda <yamini.nimmagadda@intel.com>

* code cleanup

* npu perf fix

* requant to f16 for Q6 embed on NPU

* Update ggml/src/ggml-openvino/ggml-decoder.cpp

* Update ggml/src/ggml-openvino/ggml-openvino-extra.cpp

* Create OPENVINO.md in llama.cpp backend docs

* Update OPENVINO.md

* Update OPENVINO.md

* Update OPENVINO.md

* Update build.md

* Update OPENVINO.md

* Update OPENVINO.md

* Update OPENVINO.md

* kq_mask naming fix

* Syntax correction for workflows build file

* Change ov backend buffer is_host to false

* Fix llama-bench -p -n where p<=256

* Fix --direct-io 0

* Don't put kvcache on GPU in stateful mode

* Remove hardcode names

* Fix stateful shapes

* Simplification for stateful and update output shape processing

* Remove hardcode names

* Avoid re-compilation in llama-bench

* Extract zp directly instead of bias

* Refactor weight tensor processing

* create_weight_node accept non-ov backend buffer

* remove changes in llama-graph.cpp

* stateful masking fix (#38)

Fix for stateful accuracy issues and cl_out_of_resources error in stateful GPU with larger context sizes.

* Fix test-backend-ops crash glu, get_rows, scale, rms_norm, add

* hardcoded name handling for rope_freqs.weight

* Suppress logging and add error handling to allow test-backend-ops to complete

* Fix MUL_MAT with broadcast; Add unsupported MUL_MAT FLASH_ATTN cases

* Use bias instead of zp in test-backend-ops

* Update OV in CI, Add OV CI Tests in GH Actions

* Temp fix for multithreading bug

* Update OV CI, fix review suggestions.

* fix editorconfig-checker, update docs

* Fix tabs to spaces for editorconfig-checker

* fix editorconfig-checker

* Update docs

* updated model link to be GGUF model links

* Remove GGML_CPU_REPACK=OFF

* Skip permuted ADD and MUL

* Removed static variables from utils.cpp

* Removed initializing non-existing variable

* Remove unused structs

* Fix test-backend-ops for OV GPU

* unify api calling

* Update utils.cpp

* When the dim is dynamic, throw an error, need to is stastic forst

* Add interface compute_model_outputs(), which get the model output through computing the node use count & status in the cgraph to avoid the flag using

* No need to return

* Fix test-backend-ops for OV GPU LNL

* Fix test-thread-safety

* use the shape from infer request of output tensor create to avoid issue

* fix dynamic output shape  issue

* fix issue for the unused node in tests

* Remove unused lock

* Add comment

* Update openvino docs

* update to OV release version 2026.0

* add ci ov-gpu self hosted runner

* fix editorconfig

* Fix perplexity

* Rewrite the model inputs finding mechanism  (#54)

* Rewrite the model inputs finding logistic

* Put stateful shape handle in get input shape

* Put the iteration logistic in func

* Added ggml-ci-intel-openvino-gpu and doc update

* .hpp files converted to .h

* fix ggml-ci-x64-intel-openvino-gpu

* Fix for stateful execution bug in llama-bench

* Minor updates after stateful llama-bench fix

* Update ggml/src/ggml-openvino/utils.cpp

Co-authored-by: Yamini Nimmagadda <yamini.nimmagadda@intel.com>

* Remove multiple get_shape calls

* Bring back mutex into compute

* Fix VIEW op, which slice the input node

* Added token_len_per_seq existence check before slicing masks and moved node retrieval inside guarded block to prevent missing-key access

* Temp. fix for test requant errors

* Update to OV ggml-ci to low-perf

* ci : temporary disable "test-llama-archs"

* ci : cache v4 -> v5, checkout v4 -> v6, fix runner tag

* docs : update url

* Fix OV link in docker and Update docs

---------

Co-authored-by: Ravi Panchumarthy <ravi.panchumarthy@intel.com>
Co-authored-by: Cavus Mustafa <mustafa.cavus@intel.com>
Co-authored-by: Arshath <arshath.ramzan@intel.com>
Co-authored-by: XuejunZhai <Xuejun.Zhai@intel.com>
Co-authored-by: Yamini Nimmagadda <yamini.nimmagadda@intel.com>
Co-authored-by: Xuejun Zhai <Xuejun.Zhai@intel>
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2026-03-14 07:56:55 +02:00

117 lines
5.9 KiB
Plaintext

# collaborators can optionally add themselves here to indicate their availability for reviewing related PRs
# multiplie collaborators per item can be specified
/.devops/*.Dockerfile @ngxson
/.github/actions/ @CISC
/.github/workflows/ @CISC
/ci/ @ggerganov
/cmake/ @ggerganov
/common/CMakeLists.txt @ggerganov
/common/arg.* @ggerganov
/common/base64.hpp.* @ggerganov
/common/build-info.* @ggerganov
/common/chat.* @pwilkin
/common/chat-auto*.* @pwilkin
/common/chat-diff-analyzer.* @pwilkin
/common/chat-peg-parser.* @aldehir
/common/common.* @ggerganov
/common/console.* @ggerganov
/common/http.* @angt
/common/jinja/ @ngxson @CISC @aldehir
/common/llguidance.* @ggerganov
/common/log.* @ggerganov
/common/ngram-map.* @srogmann
/common/peg-parser.* @aldehir
/common/sampling.* @ggerganov
/common/speculative.* @ggerganov
/common/unicode.* @aldehir
/convert_*.py @CISC
/examples/batched.swift/ @ggerganov
/examples/batched/ @ggerganov
/examples/convert-llama2c-to-ggml/ @ggerganov
/examples/debug/ @danbev @pwilkin
/examples/deprecation-warning/ @ggerganov
/examples/diffusion/ @am17an
/examples/embedding/ @ggerganov
/examples/eval-callback/ @ggerganov
/examples/export-docs/ @ggerganov
/examples/gen-docs/ @ggerganov
/examples/gguf/ @ggerganov
/examples/llama.android/ @ggerganov @hanyin-arm @naco-siren
/examples/llama.swiftui/ @ggerganov
/examples/llama.vim @ggerganov
/examples/lookahead/ @ggerganov
/examples/lookup/ @JohannesGaessler
/examples/model-conversion/ @danbev
/examples/parallel/ @ggerganov
/examples/passkey/ @ggerganov
/examples/retrieval/ @ggerganov
/examples/save-load-state/ @ggerganov
/examples/speculative-simple/ @ggerganov
/examples/speculative/ @ggerganov
/ggml/cmake/ @ggerganov
/ggml/include/ @ggerganov
/ggml/src/ggml-common.h @ggerganov
/ggml/src/ggml-cpu/ @ggerganov
/ggml/src/ggml-cpu/spacemit/ @alex-spacemit
/ggml/src/ggml-cuda/fattn* @JohannesGaessler
/ggml/src/ggml-cuda/mmf.* @JohannesGaessler @am17an
/ggml/src/ggml-cuda/mmq.* @JohannesGaessler
/ggml/src/ggml-cuda/mmvf.* @JohannesGaessler
/ggml/src/ggml-cuda/mmvq.* @JohannesGaessler
/ggml/src/ggml-cuda/fattn-wmma* @IMbackK
/ggml/src/ggml-hip/ @IMbackK
/ggml/src/ggml-cuda/vendors/hip.h @IMbackK
/ggml/src/ggml-impl.h @ggerganov
/ggml/src/ggml-metal/ @ggerganov
/ggml/src/ggml-opencl/ @lhez @max-krasnyansky
/ggml/src/ggml-hexagon/ @max-krasnyansky @lhez
/ggml/src/ggml-opt.cpp @JohannesGaessler
/ggml/src/ggml-quants.* @ggerganov
/ggml/src/ggml-rpc/ @rgerganov
/ggml/src/ggml-threading.* @ggerganov
/ggml/src/ggml-vulkan/ @0cc4m
/ggml/src/ggml-virtgpu/ @kpouget
/ggml/src/ggml-webgpu/ @reeselevine
/ggml/src/ggml-zdnn/ @taronaeo @Andreas-Krebbel @AlekseiNikiforovIBM
/ggml/src/ggml-openvino/ @cavusmustafa @wine99
/ggml/src/ggml.c @ggerganov
/ggml/src/ggml.cpp @ggerganov
/ggml/src/gguf.cpp @JohannesGaessler @Green-Sky
/gguf-py/ @CISC
/media/ @ggerganov
/scripts/gen* @ggerganov
/scripts/get* @ggerganov
/scripts/sync* @ggerganov
/src/ @ggerganov
/src/llama-adapter.* @CISC
/src/llama-arch.* @CISC
/src/llama-chat.* @ngxson
/src/llama-graph.* @CISC
/src/llama-model.* @CISC
/src/llama-vocab.* @CISC
/src/models/ @CISC
/tests/ @ggerganov
/tests/test-chat.* @pwilkin
/tools/batched-bench/ @ggerganov
/tools/cli/ @ngxson
/tools/completion/ @ggerganov
/tools/mtmd/ @ngxson
/tools/perplexity/ @ggerganov
/tools/parser/ @pwilkin
/tools/quantize/ @ggerganov
/tools/rpc/ @rgerganov
/tools/server/* @ngxson @ggerganov # no subdir
/tools/server/webui/ @allozaur
/tools/tokenize/ @ggerganov
/tools/tts/ @ggerganov
/vendor/ @ggerganov
/AUTHORS @ggerganov
/CMakeLists.txt @ggerganov
/CONTRIBUTING.md @ggerganov
/LICENSE @ggerganov
/README.md @ggerganov
/SECURITY.md @ggerganov
/build-xcframework.sh @danbev
requirements*.txt @CISC