Professional Documents
Culture Documents
IT(AI&ML)-III
P83A1NLP: Natural Language Processing
Practical List
1. Write a python program to explain various methods of the OS Module.
Code:-
import os
print(os.name)
os.mkdir("D:\\New_folder")
print(os.getcwd())
os.chdir("D:\\")
os.rmdir("D:\\New_folder")
fw = os.popen("D:\\02file.txt", 'w')
fw.write("This is awesome")
os.rename("D:\\02file.txt",'Python1.txt')
Ouput:-
output:-
Output:-
df['tokenized']=df['text'].apply(word_tokenize)
df['lower'] = df['tokenized'].apply(lambda x: [word.lower() for word in x])
stop_words=set(stopwords.words('english'))
df['stopwords_removed']= df['lower'].apply(lambda x:
[word for word in x if word not in stop_words])
wnl = WordNetLemmatizer()
df['lemmatized'] = df['stopwords_removed'].apply(lambda x:
[wnl.lemmatize(word) for word in x])
Output:-
Output:-
Output:-
df['tokenized']=df['text'].apply(word_tokenize)
df['lower'] = df['tokenized'].apply(lambda x: [word.lower() for word in x])
stop_words=set(stopwords.words('english'))
df['stopwords_removed']= df['lower'].apply(lambda x:
[word for word in x if word not in stop_words])
nltk.download('averaged_perceptron_tagger')
df['pos_tags'] = df['stopwords_removed'].apply(nltk.tag.pos_tag)
Output:-
Code:-
ps = PorterStemmer()
wnl = WordNetLemmatizer()
df['lemmatized'] = df['stopwords_removed'].apply(lambda x:
[wnl.lemmatize(word) for word in x])
df.head()
Yash Amin 21084341001 10
Output:-
NER = spacy.load("en_core_web_sm")
Output:-
10. Write a complete NLP task for cleaning and pre-processing text using
NLTK.
Code:-
import nltk
from nltk.tokenize import word_tokenize
from nltk.corpus import stopwords,wordnet
from nltk.stem import WordNetLemmatizer
from nltk.stem import PorterStemmer
from nltk import bigrams,trigrams,ngrams
from nltk.probability import FreqDist
df['tokenized']=df['text'].apply(word_tokenize)
wnl = WordNetLemmatizer()
df['lemmatized'] = df['stopwords_removed'].apply(lambda x:
[wnl.lemmatize(word) for word in x])
Output:-