Advertisement
Guest User

Untitled

a guest
Aug 14th, 2018
55
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 4.03 KB | None | 0 0
  1. Sentiment Classification using Machine Learning Techniques
  2. Pranjal Vachaspati
  3. pranjal@mit.edu
  4. Cathy Wu
  5. cathywu@mit.edu
  6. Abstract
  7. We implement a series of classifiers (Naive Bayes, Max-
  8. imum Entropy, and SVM) to distinguish positive and nega-
  9. tive sentiment in critic and user reviews. We apply various
  10. processing methods, including negation tagging, part-of-
  11. speech tagging, and position tagging to achieve maximum
  12. accuracy. We test our classifiers on an external dataset to
  13. see how well they generalize. Finally, we use a majority-
  14. voting technique to combine classifiers and achieve accu-
  15. racy of close to 90% in 3-fold cross-validation, far outper-
  16. forming Pang’s 2002 work [7].
  17. 1. Introduction
  18. Sentiment analysis, broadly speaking, is the set of tech-
  19. niques that allows detection of emotional content in text.
  20. This has a variety of applications: it is commonly used by
  21. trading algorithms to process news articles, as well as by
  22. corporations to better respond to consumer service needs.
  23. Similar techniques can also be applied to other text analysis
  24. problems, like spam filtering.
  25. The source code described in this paper is available at
  26. https://github.com/cathywu/Sentiment-Analysis.
  27. 2. Previous Work
  28. We set out to replicate Pang’s work [7] from 2002 on
  29. using classical knowledge-free supervised machine learn-
  30. ing techniques to perform sentiment classification. They
  31. used the machine learning methods (Naive Bayes, maxi-
  32. mum entropy classification, and support vector machines),
  33. methods commonly used for topic classification, to explore
  34. the difference between and sentiment classification in doc-
  35. uments. Pang cited a number of related works, but they
  36. mostly pertain to classifying documents on criteria weakly
  37. tied to sentiment or using knowledge-based sentiment clas-
  38. sification methods. We used a similar dataset, as released
  39. by the authors, and made efforts to use the same libraries
  40. and pre-processing techniques.
  41. In addition to replicating Pang’s work as closely as we
  42. could, we extended the work by exploring an additional
  43. dataset, additional preprocessing techniques, and combin-
  44. ing classifiers. We tested how well classifiers trained on
  45. Pang’s dataset extended to reviews in another domain. Al-
  46. though Pang limited many of his tests to use only the
  47. 16165 most common ngrams, advanced processors have
  48. lifted this computational constraint, and so we addition-
  49. ally tested on all ngrams. We used a newer parameter es-
  50. timation algorithm called Limited-Memory Variable Met-
  51. ric (L-BFGS)[5] for maximum entropy classification. Pang
  52. used the Improved Iterative Scaling method. We also imple-
  53. mented and tested the effect of term frequency-inver docu-
  54. ment frequency (TF-IDF) on classification results.
  55. 3. The User Review Domain
  56. For our experiments, we worked with movie re-
  57. views. Our data source was Pang’s released dataset
  58. (http://www.cs.cornell.edu/people/pabo/movie-review-
  59. data/) from their 2004 publication. The dataset contains
  60. 1000 positive reviews and 1000 negative reviews, each
  61. labeled with their true sentiment. The original data source
  62. was the Internet Movie Database (IMDb).
  63. Pang applied the bag-of-words method to positive and
  64. negative sentiment classification, but the same method can
  65. be extended to various other domains, including topic clas-
  66. sification. We additionally chose to work with a set of 5000
  67. Yelp reviews, 1000 for each of their five star rating. Yelp
  68. is a popular online urban city guide that houses reviews
  69. of restaurants, shopping areas, and businesses. Although
  70. a movie review and a Yelp review will differ in specialized
  71. vocabulary, audience, tone, etc., the ways that people con-
  72. vey sentiment (e.g. I loved it!) may not differ entirely. We
  73. wished to explore how training classifiers in one domain
  74. might generalize to neighbor domains.
  75. The domain of reviews is experimentally convenient be-
  76. cause there are largely available on-line and because re-
  77. viewers often summarize their overall sentiment with a
  78. machine-extractable rating indicator; hence, there was no
  79. need for hand-labeling of data.
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement