@@ -153,8 +153,7 @@ def write_tensors(self):
153
153
# map tensor names
154
154
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
155
155
if new_name is None :
156
- logger .error (f"Can not map tensor { name !r} " )
157
- sys .exit ()
156
+ raise ValueError (f"Can not map tensor { name !r} " )
158
157
159
158
n_dims = len (data .shape )
160
159
data_dtype = data .dtype
@@ -486,8 +485,7 @@ def write_tensors(self):
486
485
# map tensor names
487
486
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
488
487
if new_name is None :
489
- logger .error (f"Can not map tensor { name !r} " )
490
- sys .exit ()
488
+ raise ValueError (f"Can not map tensor { name !r} " )
491
489
492
490
n_dims = len (data .shape )
493
491
data_dtype = data .dtype
@@ -570,8 +568,7 @@ def write_tensors(self):
570
568
else :
571
569
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
572
570
if new_name is None :
573
- logger .error (f"Can not map tensor { name !r} " )
574
- sys .exit ()
571
+ raise ValueError (f"Can not map tensor { name !r} " )
575
572
576
573
n_dims = len (data .shape )
577
574
data_dtype = data .dtype
@@ -614,8 +611,7 @@ def set_gguf_parameters(self):
614
611
elif "model_max_length" in self .hparams :
615
612
ctx_length = self .hparams ["model_max_length" ]
616
613
else :
617
- logger .error ("gguf: can not find ctx length parameter." )
618
- sys .exit ()
614
+ raise ValueError ("gguf: can not find ctx length parameter." )
619
615
620
616
self .gguf_writer .add_file_type (self .ftype )
621
617
self .gguf_writer .add_name (self .dir_model .name )
@@ -653,8 +649,7 @@ def write_tensors(self):
653
649
# map tensor names
654
650
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
655
651
if new_name is None :
656
- logger .error (f"Can not map tensor { name !r} " )
657
- sys .exit ()
652
+ raise ValueError (f"Can not map tensor { name !r} " )
658
653
659
654
n_dims = len (data .shape )
660
655
data_dtype = data .dtype
@@ -696,8 +691,7 @@ def set_gguf_parameters(self):
696
691
elif "model_max_length" in self .hparams :
697
692
ctx_length = self .hparams ["model_max_length" ]
698
693
else :
699
- logger .error ("gguf: can not find ctx length parameter." )
700
- sys .exit ()
694
+ raise ValueError ("gguf: can not find ctx length parameter." )
701
695
702
696
self .gguf_writer .add_name (self .dir_model .name )
703
697
self .gguf_writer .add_source_hf_repo (hf_repo )
@@ -751,8 +745,7 @@ def write_tensors(self):
751
745
# map tensor names
752
746
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
753
747
if new_name is None :
754
- logger .error (f"Can not map tensor { name !r} " )
755
- sys .exit ()
748
+ raise ValueError (f"Can not map tensor { name !r} " )
756
749
757
750
n_dims = len (data .shape )
758
751
data_dtype = data .dtype
@@ -853,8 +846,7 @@ def set_gguf_parameters(self):
853
846
elif "model_max_length" in self .hparams :
854
847
ctx_length = self .hparams ["model_max_length" ]
855
848
else :
856
- logger .error ("gguf: can not find ctx length parameter." )
857
- sys .exit ()
849
+ raise ValueError ("gguf: can not find ctx length parameter." )
858
850
859
851
self .gguf_writer .add_name (self .dir_model .name )
860
852
self .gguf_writer .add_source_hf_repo (hf_repo )
@@ -903,8 +895,7 @@ def write_tensors(self):
903
895
# map tensor names
904
896
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
905
897
if new_name is None :
906
- logger .error (f"Can not map tensor { name !r} " )
907
- sys .exit ()
898
+ raise ValueError (f"Can not map tensor { name !r} " )
908
899
909
900
n_dims = len (data .shape )
910
901
data_dtype = data .dtype
@@ -1008,8 +999,7 @@ def write_tensors(self):
1008
999
# map tensor names
1009
1000
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1010
1001
if new_name is None :
1011
- logger .error (f"Can not map tensor { name !r} " )
1012
- sys .exit ()
1002
+ raise ValueError (f"Can not map tensor { name !r} " )
1013
1003
1014
1004
n_dims = len (data .shape )
1015
1005
data_dtype = data .dtype
@@ -1111,10 +1101,9 @@ def write_tensors(self):
1111
1101
data = data_torch .squeeze ().numpy ()
1112
1102
1113
1103
# map tensor names
1114
- new_name = tensor_map .get_name (name , try_suffixes = (".weight" ,))
1104
+ new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1115
1105
if new_name is None :
1116
- logger .error (f"Can not map tensor { name !r} " )
1117
- sys .exit ()
1106
+ raise ValueError (f"Can not map tensor { name !r} " )
1118
1107
1119
1108
n_dims = len (data .shape )
1120
1109
data_dtype = data .dtype
@@ -1180,8 +1169,7 @@ def write_tensors(self):
1180
1169
data = data_torch .to (torch .float32 ).squeeze ().numpy ()
1181
1170
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1182
1171
if new_name is None :
1183
- logger .error (f"Can not map tensor { name !r} " )
1184
- sys .exit ()
1172
+ raise ValueError (f"Can not map tensor { name !r} " )
1185
1173
n_dims = len (data .shape )
1186
1174
logger .info (f"{ new_name } , n_dims = { n_dims } , { old_dtype } --> { data .dtype } " )
1187
1175
self .gguf_writer .add_tensor (new_name , data )
@@ -1293,8 +1281,7 @@ def write_tensors(self):
1293
1281
1294
1282
new_name = tensor_map .get_name (merged_name , try_suffixes = (".weight" , ".bias" ))
1295
1283
if new_name is None :
1296
- logger .error (f"Can not map tensor { name !r} " )
1297
- sys .exit ()
1284
+ raise ValueError (f"Can not map tensor { name !r} " )
1298
1285
1299
1286
logger .info (f"{ new_name } , n_dims = { len (data .shape )} , shape = { data .shape } --> { data .dtype } " )
1300
1287
@@ -1304,8 +1291,7 @@ def write_tensors(self):
1304
1291
# map tensor names
1305
1292
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1306
1293
if new_name is None :
1307
- logger .error (f"Can not map tensor { name !r} " )
1308
- sys .exit ()
1294
+ raise ValueError (f"Can not map tensor { name !r} " )
1309
1295
1310
1296
n_dims = len (data .shape )
1311
1297
data_dtype = data .dtype
@@ -1397,8 +1383,7 @@ def write_tensors(self):
1397
1383
1398
1384
new_name = tensor_map .get_name (merged_name , try_suffixes = (".weight" , ".bias" ))
1399
1385
if new_name is None :
1400
- logger .error (f"Can not map tensor { name !r} " )
1401
- sys .exit ()
1386
+ raise ValueError (f"Can not map tensor { name !r} " )
1402
1387
1403
1388
logger .info (f"{ new_name } , n_dims = { len (data .shape )} , shape = { data .shape } --> { data .dtype } " )
1404
1389
@@ -1408,8 +1393,7 @@ def write_tensors(self):
1408
1393
# map tensor names
1409
1394
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1410
1395
if new_name is None :
1411
- logger .error (f"Can not map tensor { name !r} " )
1412
- sys .exit ()
1396
+ raise ValueError (f"Can not map tensor { name !r} " )
1413
1397
1414
1398
n_dims = len (data .shape )
1415
1399
data_dtype = data .dtype
@@ -1502,17 +1486,15 @@ def write_tensors(self):
1502
1486
# https://huggingface.co/databricks/dbrx-instruct/blob/main/model.safetensors.index.json#L15
1503
1487
new_name = tensor_map .get_name (name if not experts else name + ".weight" , try_suffixes = (".weight" ,))
1504
1488
if new_name is None :
1505
- logger .error (f"Can not map tensor { name !r} " )
1506
- sys .exit ()
1489
+ raise ValueError (f"Can not map tensor { name !r} " )
1507
1490
1508
1491
n_dims = len (data .shape )
1509
1492
data_dtype = data .dtype
1510
1493
1511
1494
# Most of the codebase that takes in 1D tensors only handles F32 tensors
1512
1495
# and most of the outputs tensors are F32.
1513
1496
if data_dtype != np .float32 and n_dims == 1 :
1514
- logger .error (f"Can not map tensor { name !r} : all 1D tensors must be F32" )
1515
- sys .exit ()
1497
+ raise ValueError (f"Can not map tensor { name !r} : all 1D tensors must be F32" )
1516
1498
1517
1499
# if f32 desired, convert any float16 to float32
1518
1500
if self .ftype == 0 and data_dtype == np .float16 :
@@ -1584,8 +1566,7 @@ def write_tensors(self):
1584
1566
# map tensor names
1585
1567
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1586
1568
if new_name is None :
1587
- logger .error (f"Can not map tensor { name !r} " )
1588
- sys .exit ()
1569
+ raise ValueError (f"Can not map tensor { name !r} " )
1589
1570
1590
1571
n_dims = len (data .shape )
1591
1572
data_dtype = data .dtype
@@ -1668,8 +1649,7 @@ def write_tensors(self):
1668
1649
# map tensor names
1669
1650
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1670
1651
if new_name is None :
1671
- logger .error (f"Can not map tensor { name !r} " )
1672
- sys .exit ()
1652
+ raise ValueError (f"Can not map tensor { name !r} " )
1673
1653
1674
1654
n_dims = len (data .shape )
1675
1655
data_dtype = data .dtype
@@ -1732,8 +1712,7 @@ def write_tensors(self):
1732
1712
# map tensor names
1733
1713
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1734
1714
if new_name is None :
1735
- logger .error (f"Can not map tensor { name !r} " )
1736
- sys .exit ()
1715
+ raise ValueError (f"Can not map tensor { name !r} " )
1737
1716
1738
1717
n_dims = len (data .shape )
1739
1718
data_dtype = data .dtype
@@ -1830,8 +1809,7 @@ def write_tensors(self):
1830
1809
# map tensor names
1831
1810
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1832
1811
if new_name is None :
1833
- logger .error (f"Can not map tensor { name !r} " )
1834
- sys .exit ()
1812
+ raise ValueError (f"Can not map tensor { name !r} " )
1835
1813
1836
1814
# shuffle for broadcasting of gqa in ggml_mul_mat
1837
1815
if new_name .endswith ("attn_q.weight" ):
@@ -1908,8 +1886,7 @@ def write_tensors(self):
1908
1886
# map tensor names
1909
1887
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
1910
1888
if new_name is None :
1911
- logger .error (f"Can not map tensor { name !r} " )
1912
- sys .exit ()
1889
+ raise ValueError (f"Can not map tensor { name !r} " )
1913
1890
1914
1891
n_dims = len (data .shape )
1915
1892
data_dtype = data .dtype
@@ -2054,8 +2031,7 @@ def post_write_tensors(self, tensor_map, name, data_torch):
2054
2031
# map tensor names
2055
2032
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2056
2033
if new_name is None :
2057
- logger .error (f"Can not map tensor { name !r} " )
2058
- sys .exit ()
2034
+ raise ValueError (f"Can not map tensor { name !r} " )
2059
2035
2060
2036
n_dims = len (data .shape )
2061
2037
data_dtype = data .dtype
@@ -2182,8 +2158,7 @@ def write_tensors(self):
2182
2158
# map tensor names
2183
2159
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2184
2160
if new_name is None :
2185
- logger .error (f"Can not map tensor { name !r} " )
2186
- sys .exit ()
2161
+ raise ValueError (f"Can not map tensor { name !r} " )
2187
2162
2188
2163
data = data_torch .squeeze ().numpy ()
2189
2164
n_dims = len (data .shape )
@@ -2279,8 +2254,7 @@ def write_tensors(self):
2279
2254
# map tensor names
2280
2255
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2281
2256
if new_name is None :
2282
- logger .error (f"Can not map tensor { name !r} " )
2283
- sys .exit ()
2257
+ raise ValueError (f"Can not map tensor { name !r} " )
2284
2258
2285
2259
n_dims = len (data .shape )
2286
2260
data_dtype = data .dtype
@@ -2382,8 +2356,7 @@ def write_tensors(self):
2382
2356
# map tensor names
2383
2357
new_name = tensor_map .get_name (name , try_suffixes = (".weight" , ".bias" ))
2384
2358
if new_name is None :
2385
- logger .error (f"Can not map tensor { name !r} " )
2386
- sys .exit ()
2359
+ raise ValueError (f"Can not map tensor { name !r} " )
2387
2360
2388
2361
if name .endswith (".A_log" ):
2389
2362
logger .debug ("A_log --> A ==> " + new_name )
0 commit comments