A unified latent variable model for contrastive opinion mining

Ebuka IBEKE, Chenghua LIN, Adam WYNER, Mohamad Hardyman BARAWI

PDF(444 KB)
PDF(444 KB)
Front. Comput. Sci. ›› 2020, Vol. 14 ›› Issue (2) : 404-416. DOI: 10.1007/s11704-018-7073-5
RESEARCH ARTICLE

A unified latent variable model for contrastive opinion mining

Author information +
History +

Abstract

There are large and growing textual corpora in which people express contrastive opinions about the same topic. This has led to an increasing number of studies about contrastive opinion mining. However, there are several notable issues with the existing studies. They mostly focus on mining contrastive opinions from multiple data collections, which need to be separated into their respective collections beforehand. In addition, existing models are opaque in terms of the relationship between topics that are extracted and the sentences in the corpus which express the topics; this opacity does not help us understand the opinions expressed in the corpus. Finally, contrastive opinion is mostly analysed qualitatively rather than quantitatively. This paper addresses these matters and proposes a novel unified latent variable model (contraLDA), which: mines contrastive opinions from both single and multiple data collections, extracts the sentences that project the contrastive opinion, and measures the strength of opinion contrastiveness towards the extracted topics. Experimental results show the effectiveness of our model in mining contrasted opinions, which outperformed our baselines in extracting coherent and informative sentiment-bearing topics. We further show the accuracy of our model in classifying topics and sentiments of textual data, and we compared our results to five strong baselines.

Keywords

contrastive opinion mining / sentiment analysis / topic modelling

Cite this article

Download citation ▾
Ebuka IBEKE, Chenghua LIN, Adam WYNER, Mohamad Hardyman BARAWI. A unified latent variable model for contrastive opinion mining. Front. Comput. Sci., 2020, 14(2): 404‒416 https://doi.org/10.1007/s11704-018-7073-5

References

[1]
Fang Y, Si L, Somasundaram N, Yu Z. Mining contrastive opinions on political texts using cross-perspective topic model. In: Proceedings of the International Conference on Web Search and Data Mining. 2012, 63–72
CrossRef Google scholar
[2]
Trabelsi A, Zaïane O R. A joint topic viewpoint model for contention analysis. In: Proceedings of the International Conference on Applications of Natural Language to Data Bases/Information Systems. 2014, 114–125
CrossRef Google scholar
[3]
Lerman K, McDonald R. Contrastive summarization: an experiment with consumer reviews. In: Proceedings of the HLT Annual Conference of the North American Chapter of the Association for Computational Linguistics. 2009, 113–116
CrossRef Google scholar
[4]
Paul M, Girju R. Cross-cultural analysis of blogs and forums with mixed-collection topic models. In: Proceedings of the ACL Conference on Empirical Methods in Natural Language Processing. 2009, 1408–1417
CrossRef Google scholar
[5]
Elahi M F, Monachesi P. An examination of cross-cultural similarities and differences from social media data with respect to language use. In: Proceedings of the 8th International Conference on Language Resources and Evaluation. 2012, 4080–4086
[6]
Ibeke E, Lin C, Wyner A, Barawi M H. Extracting and understanding contrastive opinion through topic relevant sentences. In: Proceedings of the 8th International Joint Conference on Natural Language Processing. 2017, 395–400
[7]
Barawi M H, Lin C, Siddharthan A. Automatically labelling sentimentbearing topics with descriptive sentence labels. In: Proceedings of the 22nd International Conference on Natural Language and Information Systems. 2017, 299–312
CrossRef Google scholar
[8]
Zhai C, Velivelli A, Yu B. A cross-collection mixture model for comparative text mining. In: Proceedings of the International Conference on Knowledge Discovery and Data Mining. 2004, 743–748
CrossRef Google scholar
[9]
Ahmed A, Xing E P. Staying informed: supervised and semisupervised multi-view topical analysis of ideological perspective. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing. 2010, 1140–1150
[10]
Mukherjee A, Liu B. Mining contentions from discussions and debates. In: Proceedings of the International Conference on Knowledge Discovery and Data Mining. 2012, 841–849
CrossRef Google scholar
[11]
Thonet T, Cabanac G, Boughanem M, Pinel-Sauvagnat K. VODUM: a topic model unifying viewpoint, topic and opinion discovery. In: Proceedings of the European Conference on Information Retrieval. 2016, 533–545
CrossRef Google scholar
[12]
Paul M J, Zhai C, Girju R. Summarizing contrastive viewpoints in opinionated text. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing. 2010, 66–76
[13]
Paul M J, Girju R. A two-dimensional topic-aspect model for discovering multi-faceted topics. Urbana, 2010, 51(61801): 36
[14]
Guo J, Lu Y, Mori T, Blake C. Expert-guided contrastive opinion summarization for controversial issues. In: Proceedings of the 24th International Conference on World Wide Web Companion. 2015, 1105–1110
CrossRef Google scholar
[15]
Ren Z, de Rijke M. Summarizing contrastive themes via hierarchical non-parametric processes. In: Proceedings of the 38th SIGIR International Conference on Research and Development in Information Retrieval. 2015, 93–102
CrossRef Google scholar
[16]
He L, Li W, Zhuge H. Exploring differential topic models for comparative summarization of scientific papers. In: Proceedings of COLING International Conference on Computational Linguistics. 2016, 1028–1038
[17]
Nakasaki H, Kawaba M, Utsuro T, Fukuhara T. Mining crosslingual/cross-cultural differences in concerns and opinions in blogs. In: Proceedings of the 22nd International Conferee on Computer Processing of Oriental Languages. Language Technology for the Knowledgebased Economy. 2009, 213–224
CrossRef Google scholar
[18]
Guo H, Zhu H, Guo Z, Zhang X, Su Z. OpinionIt: a text mining system for cross-lingual opinion analysis. In: Proceedings of the 19th ACM International Conference on Information and Knowledge Management. 2010, 1199–1208
CrossRef Google scholar
[19]
Gutiérrez E D, Shutova E, Lichtenstein P, deMelo G, Gilardi L. Detecting cross-cultural differences using a multilingual topic model. Journal of Transactions of the Association for Computational Linguistics, 2016, 4: 47–60
CrossRef Google scholar
[20]
Lin C, He Y. Joint sentiment/topic model for sentiment analysis. In: Proceedings of the 18th ACM conference on Information and Knowledge Management. 2009, 375–384
CrossRef Google scholar
[21]
Lin C, He Y, Everson R, Ruger S. Weakly supervised joint sentimenttopic detection from text. Journal of IEEE Transactions on Knowledge and Data Engineering, 2012, 24(6): 1134–1145
CrossRef Google scholar
[22]
Wallach H, Mimno D, McCallum A. Rethinking LDA: why priors matter. Advances in Neural Information Processing Systems, 2009, 22: 1973–1981
[23]
Minka T. Estimating a dirichlet distribution. Technical Report, 2003
[24]
Ibeke E, Lin C, Coe C, Wyner A, Liu D, Barawi M H, Yusof N F A. A curated corpus for sentiment-topic analysis. In: Proceedings of the LREC 2016 Workshop on Emotion and Sentiment Analysis. 2016, 32–39
[25]
Chang J, Gerrish S, Wang C, Boyd-Graber J L, Jordan L, Blei D M. Reading tea leaves: how humans interpret topic models. In: Proceedings of the 22nd Conference on Neural Information Processing Systems. 2009, 288–296
[26]
Mimno D, Wallach H M, Talley E, Leenders M, McCallum A. Optimizing semantic coherence in topic models. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing. 2011, 262–272
[27]
Bouma G. Normalized (pointwise) mutual information in collocation extraction. In: Proceedings of the German Society for Computational Linguistics and Language Technology. 2009, 31–40
[28]
Newman D, Lau J H, Grieser K, Baldwin T. Automatic evaluation of topic coherence. In: Proceedings of the HLT Annual Conference of the North American Chapter of the Association for Computational Linguistics. 2010, 100–108
[29]
Aletras N, Stevenson M. Evaluating topic coherence using distributional semantics. In: Proceedings of the 10th International Conference on Computational Semantics. 2013, 13–22
[30]
Steyvers M, Griffiths T. Probabilistic topic models. Handbook of Latent Semantic Analysis, 2007, 427(7): 424–440
[31]
Cano A B, He Y, Xu R. Automatic labelling of topic models learned from Twitter by summarisation. In: Proceedings of the 52nd Annual Meeting of Association for Computational Linguists. 2014, 618–624
[32]
Ramage D, Hall D, Nallapati R, Manning C D. Labeled LDA: a supervised topic model for credit attribution in multi-labeled corpora. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing. 2009, 248–256
CrossRef Google scholar
[33]
Mcauliffe J D, Blei D M. Supervised topic models. In: Proceedings of the 21st Annual Conference on Neural Information Processing Systems. 2008, 121–128
[34]
Kim Y. Convolutional neural networks for sentence classification. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing. 2014, 1746–1751
CrossRef Google scholar
[35]
Ganu G, Marian A, Elhadad N. URSA-user review structure analysis: understanding online reviewing trends. DCS Technical Report: Citeseer, 2010

RIGHTS & PERMISSIONS

2019 Higher Education Press and Springer-Verlag GmbH Germany, part of Springer Nature
AI Summary AI Mindmap
PDF(444 KB)

Accesses

Citations

Detail

Sections
Recommended

/