You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The config for large model specifies a vocab size of 51200, is there a separate tokenizer file for it? Weirdly vocab falls back down to 32 for xlarge which makes me think typo?
The tokenizer file specifies a vocab_size of 30, while the config for base and small specifies 32. Is this rounding to a power of two for efficiency?
The text was updated successfully, but these errors were encountered:
The text was updated successfully, but these errors were encountered: