Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Added a vocabulary_size argument to UnicodeCharacterTokenizer #163

Merged
merged 8 commits into from
May 5, 2022

Conversation

aflah02
Copy link
Collaborator

@aflah02 aflah02 commented May 3, 2022

Fixes #155
I've also added new tests and fixed old tests by modifying the config files with the new parameter

@aflah02
Copy link
Collaborator Author

aflah02 commented May 3, 2022

@mattdangerw I think the PR is ready for review now. Here's a demo

Copy link
Member

@mattdangerw mattdangerw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks! This looks great. Just a few comments then this is ready to go.

keras_nlp/tokenizers/unicode_character_tokenizer.py Outdated Show resolved Hide resolved
keras_nlp/tokenizers/unicode_character_tokenizer_test.py Outdated Show resolved Hide resolved
@aflah02 aflah02 requested a review from mattdangerw May 4, 2022 05:51
Copy link
Member

@mattdangerw mattdangerw left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks good! Thanks. Added a little more to the example to show how it could be used.

@mattdangerw mattdangerw merged commit a2c6067 into keras-team:master May 5, 2022
aflah02 added a commit to aflah02/keras-nlp that referenced this pull request May 14, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Add a vocabulary_size argument to UnicodeCharacterTokenizer
2 participants