mirror of
https://github.com/ggml-org/llama.cpp.git
synced 2026-05-14 21:14:10 +00:00
convert-hf : fix Refact conversion
This commit is contained in:
@@ -999,7 +999,7 @@ class RefactModel(Model):
|
||||
ff_dim = multiple_of * ((hidden_dim + multiple_of - 1) // multiple_of)
|
||||
n_head = self.hparams["n_head"]
|
||||
n_head_kv = 1
|
||||
head_dim = hidden_dim // n_head
|
||||
head_dim = self.hparams["n_embd"] // n_head
|
||||
|
||||
tensors: list[tuple[str, Tensor]] = []
|
||||
|
||||
|
||||
Reference in New Issue
Block a user