Not a member of Pastebin yet?
Sign Up,
it unlocks many cool features!
- def word_count(self,List):
- > dictionary_text={}
- > for word in List:
- > if word in dictionary_text:
- > dictionary_text[word]+=1
- > else:
- > dictionary_text[word]=1
- >
- > > def text_process(self,string):
- > > from nltk.stem import WordNetLemmatizer
- > > from nltk.tokenize import word_tokenize
- > > lemmatizer = WordNetLemmatizer()
- > > New_text=word_tokenize(string)
- > > New_text_lem=[lemmatizer.lemmatize(words) for words in New_text]
- > > dictionary_text=self.word_count(New_text_lem)
- > > return dictionary_text
- def main(self,Text):
- text_1=self.text_process(Text[0])
- text_2=self.text_process(Text[1])
- text_3=self.text_process(Text[2])
- final_dic={}
- for key in set(text_1.keys() + text_2.keys() + text_3.keys()):
- try:
- final_dic.setdefault(key,[]).append(text_1[key])
- except KeyError:
- pass
- try:
- final_dic.setdefault(key,[]).append(text_2[key])
- except KeyError:
- pass
- try:
- final_dic.setdefault(key,[]).append(text_3[key])
- except KeyError:
- pass
Add Comment
Please, Sign In to add comment