Tokenization in Natural Language Processing
Manage episode 311353554 series 3111581
In this episode we discuss about tokenization in Natural Language Processing. As discussed in previous episode, tokenisation is an important step in data cleaning and it entails dividing a large piece of text into smaller chunks. In this episode we discuss some of the basic tokenizers available from nltk.tokenize in nltk.
If you liked this episode, do follow and do connect with me on twitter @sarvesh0829
follow my blog at www.stacklearn.org.
If you sell something locally, do it using BagUp app available at play store, It would help a lot.
--- Send in a voice message: https://podcasters.spotify.com/pod/show/sarvesh-bhatnagar/message22 episodios