Articles → NLP → Sub-Word Tokenizer In NLP

Sub-Word Tokenizer In NLP






Purpose





Install Libraries




pip install transformers



Example


from transformers import AutoTokenizer

# Load a pretrained tokenizer (e.g., BERT's WordPiece tokenizer)
tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased")

text = "Playfulness makes NLP interesting"

# Tokenize
tokens = tokenizer.tokenize(text)
print("Subword Tokens:", tokens)



Output


Picture showing the output of sub-word tokenizer in nlp



Posted By  -  Karan Gupta
 
Posted On  -  Monday, September 15, 2025

Query/Feedback


Your Email Id
 
Subject
 
Query/FeedbackCharacters remaining 250