@@ -1764,25 +1764,19 @@ def set_vocab(self):
1764
1764
self .gguf_writer .add_token_list (tokens )
1765
1765
self .gguf_writer .add_token_types (toktypes )
1766
1766
1767
- special_vocab = gguf .SpecialVocab (
1768
- self .dir_model , load_merges = True ,
1769
- special_token_types = ['bos' , 'eos' , 'eom' , 'eot' ]
1770
- )
1771
- special_vocab ._set_special_token ("bos" , 128000 )
1772
- special_vocab ._set_special_token ("eos" , 128001 )
1773
- special_vocab ._set_special_token ("eom" , 128008 )
1774
- special_vocab ._set_special_token ("eot" , 128009 )
1767
+ special_vocab = gguf .SpecialVocab (self .dir_model , load_merges = True )
1775
1768
special_vocab .add_to_gguf (self .gguf_writer )
1776
1769
else :
1777
1770
# DeciLM-7B
1778
1771
self ._set_vocab_llama_hf ()
1779
- # self._set_vocab_gpt2()
1780
1772
1781
1773
def set_gguf_parameters (self ):
1782
1774
if "block_configs" in self .hparams : # Llama-3_1-Nemotron-51B
1783
1775
assert self .block_count == len (self ._num_kv_heads )
1784
1776
assert self .block_count == len (self ._num_heads )
1785
1777
assert self .block_count == len (self ._ffn_dims )
1778
+ if (rope_theta := self .hparams .get ("rope_theta" )) is not None :
1779
+ self .gguf_writer .add_rope_freq_base (rope_theta )
1786
1780
self .gguf_writer .add_head_count_kv (self ._num_kv_heads )
1787
1781
self .gguf_writer .add_head_count (self ._num_heads )
1788
1782
self .gguf_writer .add_feed_forward_length (self ._ffn_dims )
0 commit comments