llm-foundry
Fix error when decoding a token in the id gap (or out of range) in a tiktoken tokenizer
#841
Merged

Fix error when decoding a token in the id gap (or out of range) in a tiktoken tokenizer #841

dakinggg merged 5 commits into mosaicml:main from dakinggg:tiktoken-gap
dakinggg
fix gap token errors
23760568
change approach
c6e716b0
update comment
46fba400
dakinggg dakinggg requested a review from irenedea irenedea 2 years ago
dakinggg dakinggg requested a review from rajammanabrolu rajammanabrolu 2 years ago
dakinggg dakinggg requested a review from sashaDoubov sashaDoubov 2 years ago
dakinggg dakinggg marked this pull request as ready for review 2 years ago
add encoding
8100f63b
another one
991a80eb
sashaDoubov
sashaDoubov approved these changes on 2024-01-08
irenedea
irenedea approved these changes on 2024-01-08
dakinggg dakinggg merged 5b994884 into main 2 years ago
dakinggg dakinggg deleted the tiktoken-gap branch 1 year ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone