Skip to content

Commit

Permalink
Add tokenizer tests
Browse files Browse the repository at this point in the history
  • Loading branch information
NielsRogge committed Nov 21, 2023
1 parent 7dfaceb commit 7b7608f
Show file tree
Hide file tree
Showing 2 changed files with 564 additions and 0 deletions.
1 change: 1 addition & 0 deletions src/transformers/models/siglip/tokenization_siglip.py
Original file line number Diff line number Diff line change
Expand Up @@ -121,6 +121,7 @@ def __init__(
split_special_tokens=True,
**kwargs,
) -> None:
print("Pad token:", pad_token)
pad_token = AddedToken(pad_token, rstrip=True, lstrip=True)
unk_token = AddedToken(unk_token, rstrip=True, lstrip=True)
eos_token = AddedToken(eos_token, rstrip=True, lstrip=True)
Expand Down
Loading

0 comments on commit 7b7608f

Please sign in to comment.