@@ -2647,13 +2647,15 @@ def set_vocab(self):
2647
2647
2648
2648
text = piece .encode ("utf-8" )
2649
2649
score = 0.0
2650
- if len (piece ) != 0 and token_id < 64789 :
2650
+ # Referencing the tokenizer Python implementation(https://huggingface.co/THUDM/chatglm3-6b/blob/main/tokenization_chatglm.py),
2651
+ # it is only valid if it is less than tokenizer.tokenizer.sp_model.vocab_size()
2652
+ if len (piece ) != 0 and token_id < tokenizer .tokenizer .sp_model .vocab_size ():
2651
2653
score = tokenizer .tokenizer .sp_model .get_score (token_id )
2652
2654
2653
2655
if len (piece ) == 0 :
2654
2656
text = f"[PAD{ token_id } ]" .encode ("utf-8" )
2655
2657
2656
- if token_id >= 64789 :
2658
+ if token_id >= tokenizer . tokenizer . sp_model . vocab_size () :
2657
2659
toktype = SentencePieceTokenTypes .UNKNOWN
2658
2660
tokens .append (text )
2659
2661
scores .append (score )
@@ -2683,7 +2685,7 @@ def set_vocab(self):
2683
2685
special_vocab .add_to_gguf (self .gguf_writer )
2684
2686
2685
2687
def set_gguf_parameters (self ):
2686
- self .gguf_writer .add_name ("ChatGLM-6b-chat" )
2688
+ self .gguf_writer .add_name (self . dir_model . name )
2687
2689
n_embed = self .hparams .get ("hidden_size" , self .hparams .get ("n_embed" ))
2688
2690
n_head = self .hparams .get ("n_head" , self .hparams .get ("num_attention_heads" ))
2689
2691
n_head_kv = self .hparams .get ("multi_query_group_num" , n_head )
@@ -2699,16 +2701,12 @@ def set_gguf_parameters(self):
2699
2701
self .gguf_writer .add_add_bos_token (False )
2700
2702
2701
2703
def modify_tensors (self , data_torch : Tensor , name : str , bid : int | None ) -> Iterable [tuple [str , Tensor ]]:
2702
- if name .endswith (".rotary_pos_emb.inv_freq" ):
2703
- return []
2704
-
2705
2704
del bid # unused
2706
2705
2707
- name = re .sub (r'transformer\.' , '' , name )
2708
-
2709
- if name == "word_embeddings.weight" :
2710
- assert self .tensor_names is not None
2706
+ if name .endswith (".rotary_pos_emb.inv_freq" ):
2707
+ return []
2711
2708
2709
+ name = name .removeprefix ("transformer." )
2712
2710
return [(self .map_tensor_name (name ), data_torch )]
2713
2711
2714
2712
0 commit comments