Skip to content

Commit

Permalink
Fix bos and eos token ids in the model configuration (#3)
Browse files Browse the repository at this point in the history
  • Loading branch information
pcuenca authored Feb 13, 2024
1 parent 2126d1c commit ac7ac87
Showing 1 changed file with 4 additions and 4 deletions.
8 changes: 4 additions & 4 deletions src/transformers/models/gemma/configuration_gemma.py
Original file line number Diff line number Diff line change
Expand Up @@ -74,9 +74,9 @@ class GemmaConfig(PretrainedConfig):
relevant if `config.is_decoder=True`.
pad_token_id (`int`, *optional*):
Padding token id.
bos_token_id (`int`, *optional*, defaults to 1):
bos_token_id (`int`, *optional*, defaults to 2):
Beginning of stream token id.
eos_token_id (`int`, *optional*, defaults to 2):
eos_token_id (`int`, *optional*, defaults to 1):
End of stream token id.
pretraining_tp (`int`, *optional*, defaults to 1):
Experimental feature. Tensor parallelism rank used during pretraining. Please refer to [this
Expand Down Expand Up @@ -131,8 +131,8 @@ def __init__(
rms_norm_eps=1e-6,
use_cache=True,
pad_token_id=None,
bos_token_id=1,
eos_token_id=2,
bos_token_id=2,
eos_token_id=1,
tie_word_embeddings=True,
rope_theta=10000.0,
rope_scaling=None,
Expand Down

0 comments on commit ac7ac87

Please sign in to comment.