Skip to content

Commit ac7ac87

Browse files
authored
Fix bos and eos token ids in the model configuration (#3)
1 parent 2126d1c commit ac7ac87

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

src/transformers/models/gemma/configuration_gemma.py

+4-4
Original file line numberDiff line numberDiff line change
@@ -74,9 +74,9 @@ class GemmaConfig(PretrainedConfig):
7474
relevant if `config.is_decoder=True`.
7575
pad_token_id (`int`, *optional*):
7676
Padding token id.
77-
bos_token_id (`int`, *optional*, defaults to 1):
77+
bos_token_id (`int`, *optional*, defaults to 2):
7878
Beginning of stream token id.
79-
eos_token_id (`int`, *optional*, defaults to 2):
79+
eos_token_id (`int`, *optional*, defaults to 1):
8080
End of stream token id.
8181
pretraining_tp (`int`, *optional*, defaults to 1):
8282
Experimental feature. Tensor parallelism rank used during pretraining. Please refer to [this
@@ -131,8 +131,8 @@ def __init__(
131131
rms_norm_eps=1e-6,
132132
use_cache=True,
133133
pad_token_id=None,
134-
bos_token_id=1,
135-
eos_token_id=2,
134+
bos_token_id=2,
135+
eos_token_id=1,
136136
tie_word_embeddings=True,
137137
rope_theta=10000.0,
138138
rope_scaling=None,

0 commit comments

Comments
 (0)