@@ -2079,6 +2079,8 @@ def set_gguf_parameters(self):
2079
2079
self .gguf_writer .add_file_type (self .ftype )
2080
2080
2081
2081
def modify_tensors (self , data_torch : Tensor , name : str , bid : int | None ) -> Iterable [tuple [str , Tensor ]]:
2082
+ del bid # unused
2083
+
2082
2084
# lm_head is not used in llama.cpp, while autoawq will include this tensor in model
2083
2085
# To prevent errors, skip loading lm_head.weight.
2084
2086
if name == "lm_head.weight" :
@@ -2089,7 +2091,7 @@ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iter
2089
2091
if name .endswith ("norm.weight" ):
2090
2092
data_torch = data_torch + 1
2091
2093
2092
- return super (). modify_tensors ( data_torch , name , bid )
2094
+ return [( self . map_tensor_name ( name ), data_torch )]
2093
2095
2094
2096
2095
2097
@Model .register ("Starcoder2ForCausalLM" )
@@ -2277,7 +2279,7 @@ def __init__(self, *, meta: Tensor, data: Tensor | None = None, args: tuple = ()
2277
2279
2278
2280
@staticmethod
2279
2281
def _recurse_apply (o : Any , fn : Callable [[Any ], Any ]) -> Any :
2280
- # TODO: dicts
2282
+ # TODO: dict and set
2281
2283
if isinstance (o , (list , tuple )):
2282
2284
L = []
2283
2285
for item in o :
@@ -2379,7 +2381,7 @@ def __neg__(self, *args): # mamba
2379
2381
def __add__ (self , * args ): # gemma
2380
2382
return self ._wrap_fn (torch .Tensor .__add__ )(self , * args )
2381
2383
2382
- def __getitem__ (self , * args ): # bloom falcon internlm2
2384
+ def __getitem__ (self , * args ): # bloom falcon refact internlm2
2383
2385
return self ._wrap_fn (torch .Tensor .__getitem__ )(self , * args )
2384
2386
2385
2387
0 commit comments