Natural language processing, NLP, is the convergence between linguistics, computer science and artificial intelligence. It mainly aims for the interconnection between natural languages and computers that means how to analyse and model a high volume of natural language data. For understanding NLP in detail, click the link.
As you already know, computers are able to understand numbers well rather than understanding words. There is a lot of research and development happening in the domain of NLP everyday.
There are huge amounts of applications that are today working because of NLP. Like, Alexa or Apple’s Siri both are able to understand whenever someone raises a question and are even able to answer those whether to ask for playing music or taking the weather updates. Spam mail filtering is also an example of NLP.
Being in the Data science world, we can use NLP for text classification, sentimental analysis (classifying sentiments as positive or negative), text summarizations and all other classification models amid other applications.
(Must read: Top NLP python libraries)
Generally, if the data is scrapped or data is given for analyzing it would always be in its natural human format of sentences, or paragraphs etc. Before doing an analysis on that we need to transform that language and clean it so that the computer is able to understand that language in the desired format.
Data preprocessing is a fundamental step while building a machine learning model. If the data is fairly pre-processed the results would be reliable. In NLP, the first step before building the machine learning model is to pre-process the data.
(Suggested read: Top NLP trends)
Let’s see the various different steps that are followed while preprocessing the data also used for dimensionality reduction.
Stop words removal
Each term is the axis in the vector space model. In muti-dimensional space, the text or document are constituted as vectors. The number of different words represents the number of dimensions.
The python library that is used to do the preprocessing tasks in nlp is nltk. You can install the nltk package using “pip install nltk”.
It is a method in which sentences are converted into words.
import nltk from nltk.tokenize import word_tokenize token = word_tokenize("My Email address is: firstname.lastname@example.org") token
(Read also: Sentiment Analysis of YouTube Comments)
the tokenized words into lower case format. (NLU -> nlu). Words having the same meaning like nlp and NLP if they are not converted into lowercase then these both will constitute as non-identical words in the vector space model.
Lowercase =  for lowercase in token: Lowercase.append(lowercase.lower()) Lowercase
These are the most often used that do not have any significance while determining the two different documents like (a, an, the, etc.) so they are to be removed. Check the below image wherefrom the sentence “Introduction to Natural Language Processing” the “to” word is removed.
from nltk.corpus import stopwords stop_words = stopwords.words('english') from string import punctuation punct = list(punctuation) print(dataset['quote']) tokens = word_tokenize(dataset['quote']) len(tokens)
Without removing Stopwords
We got to see 50 tokens without removing stopwords, Now we shall remove stopwords.
cleaned_tokens = [token for token in tokens if token not in stop_words and token not in punctuation] len(cleaned_tokens)
By cleaning the stopwords we got the length of the dataset as 24.
(Referred blog: What is SqueezeBERT in NLP?)
It is the process in which the words are converted to its base from. Check the below code implementation where the words of the sentence are converted to the base form.
from nltk.stem import PorterStemmer ps = PorterStemmer() print(ps.stem('jumping')) print(ps.stem('lately')) print(ps.stem('assess')) print(ps.stem('ran'))
Different from stemming, lemmatization lowers the words to word in the present language for example check the below image where word has and is are changed to ha and be respectively.
from nltk import WordNetLemmatizer lemmatizer = WordNetLemmatizer() print(lemmatizer.lemmatize('ran', 'v')) print(lemmatizer.lemmatize('better', 'a'))
In this blog, I have discussed the basic preprocessing steps that are required before building models in natural language processing that are fundamentals. These include tokenization, lowercasing the text, stop word removal, stemming, and lemmatization. Although the range with which natural language processing could be implemented is wide, much research has been going on in this particular topic.
6 Major Branches of Artificial Intelligence (AI)READ MORE
Reliance Jio and JioMart: Marketing Strategy, SWOT Analysis, and Working EcosystemREAD MORE
8 Most Popular Business Analysis Techniques used by Business AnalystREAD MORE
Top 10 Big Data TechnologiesREAD MORE
Elasticity of Demand and its TypesREAD MORE
What is PESTLE Analysis? Everything you need to know about itREAD MORE
An Overview of Descriptive AnalysisREAD MORE
5 Factors Affecting the Price Elasticity of Demand (PED)READ MORE
Dijkstra’s Algorithm: The Shortest Path AlgorithmREAD MORE
What Are Recommendation Systems in Machine Learning?READ MORE