Skip to main content
Log in

Entity-level sentiment prediction in Danmaku video interaction

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

Sentiment analysis in Danmaku video interaction aims at measuring public mood in respect of the video, which is helpful for the potential applications in behavioral science. Once these sentiments are discovered, this feedback can help video creators improve the video quality and greatly enhance online users’ watching experience. Predicting these entity-level sentiments is challenging because there is no publicly available dataset about entity-level sentiment analysis of Danmaku-enabled video comments. Furthermore, the targeted entity with skewed unbalance distribution in real-world scenarios, making the task more challenging, especially when the target entity only has positive (negative) emotional comments. In this case, applying previous aspect-level sentiment analysis models directly will introduce entity bias. In this paper, we propose a large-scale Chinese video comments dataset containing time-sync Danmaku comments and traditional video comments, targeting multiple entities and sentiments associated with each entity from popular video websites. We also propose a framework of entity-level sentiment analysis with two de-biasing models: hard-masking de-bias model and soft-masking de-bias model. This framework is defined by parallel neural networks to learn the representation of comments sentences. Based on the representations, our model learns a masking strategy for entity words to avoid overfitting and mitigate the bias. Our experiments on Danmaku-enabled video datasets show that the soft-masking model significantly outperforms comparable baselines, with a relative F1-score improvement of 9.33% compared to AEN-BERT and a relative F1-score improvement of 45.77% compared to Td-LSTM. Furthermore, experiments on different distribution bias of entity demonstrate that our proposed model can achieve competitive performances. The findings of this research have implications for measuring public sentiment for entities mentioned in a specific video domain. It can also be used as a benchmark dataset for aspect entity sentiment detection methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8

Similar content being viewed by others

Notes

  1. https://github.com/adableau/DanSentiment.

  2. http://alt.qcri.org/semeval2014/task4/.

  3. http://alt.qcri.org/semeval2015/task12/.

  4. http://alt.qcri.org/semeval2016/task5/.

  5. http://www.m-mitchell.com/code/index.html.

  6. https://www.imdb.com/.

  7. https://iqiyi.com.

  8. https://github.com/PaddlePaddle/PaddleHub.

  9. https://pypi.org/project/snownlp/.

  10. https://github.com/ymcui/Chinese-BERT-wwm.

References

  1. Alarifi A, Tolba A, Al-Makhadmeh Z, Said W (2020) A big data approach to sentiment analysis using greedy feature selection with cat swarm optimization-based long short-term memory neural networks. J Supercomput 76(6):4414–4429. https://doi.org/10.1007/s11227-018-2398-2

    Article  Google Scholar 

  2. Bai Q, Hu Q, Fang F, He L (2018) Topic detection with danmaku: a time-sync joint NMF approach. DEXA 11030:428–435

    Google Scholar 

  3. Cai Y, Wan X (2019a) Multi-domain sentiment classification based on domain-aware embedding and attention. In: IJCAI-19, IJCAI, pp 4904–4910

  4. Cai Y, Wan X (2019b) Multi-domain sentiment classification based on domain-aware embedding and attention. In: Kraus S (ed) IJCAI. ijcai.org, pp 4904–4910

  5. Cao Y, Xu H (2020) Satnet: Symmetric adversarial transfer network based on two-level alignment strategy towards cross-domain sentiment classification (student abstract). In: AAAI, pp 13763–13764

  6. Chen Y, Gao Q, Rau PL (2017b) Watching a movie alone yet together: understanding reasons for watching Danmaku videos. In: International Journal of Human–Computer Interaction

  7. Chen X, Zhang Y, Ai Q, Xu H, Yan J, Qin Z (2017a) Personalized key frame recommendation. In: ACM SIGIR, 2017. ACM, pp 315–324

  8. Cho K, van Merrienboer B, Gülçehre Ç, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder–decoder for statistical machine translation. In: EMNLP, pp 1724–1734

  9. Cui Y, Che W, Liu T, Qin B, Yang Z, Wang S, Hu G (2019) Pre-training with whole word masking for chinese bert. arXiv preprint arXiv:190608101

  10. Devlin J, Chang MW, Lee K, Toutanova K (2018) Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:181004805

  11. Dong L, Wei F, Tan C, Tang D, Zhou M, Xu K (2014) Adaptive recursive neural network for target-dependent twitter sentiment classification. In: Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014, June 22–27, 2014, Baltimore, MD, USA, Volume 2: Short Papers. The Association for Computer Linguistics, pp 49–54. https://doi.org/10.3115/v1/p14-2009

  12. Felsenthal DS, Machover M (2001) The treaty of nice and qualified majority voting. Soc Choice Welf 18(3):431–464

    Article  Google Scholar 

  13. Gonen H, Goldberg Y (2019) Lipstick on a pig: Debiasing methods cover up systematic gender biases in word embeddings but do not remove them. In: NAACL HLT 2019–2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies—Proceedings of the Conference, pp 609–614

  14. He M, Ge Y, Wu L, Chen E, Tan C (2016) Predicting the popularity of DanMu—enabled videos: a multi-factor view. Springer, Berlin

    Google Scholar 

  15. Khot T, Clark P, Guerquin M, Jansen P, Sabharwal A (2020) QASC: a dataset for question answering via sentence composition. In: The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, the Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7–12, pp 8082–8090

  16. Kingma DP, Ba J (2015) Adam: a method for stochastic optimization. In: ICLR 2015

  17. Kurita K, Vyas N, Pareek A, Black AW, Tsvetkov Y (2019) Measuring bias in contextualized word representations. CoRR abs/1906.07337. arxiv:1906.07337

  18. Lei J, Zhang Q, Wang J, Luo H (2019) BERT based hierarchical sequence classification for context-aware microblog sentiment analysis. In: Gedeon T, Wong KW, Lee M (eds) ICONIP, Lecture Notes in Computer Science, vol 11955, pp 376–386

  19. Li X, Bing L, Lam W, Shi B (2018) Transformation networks for target-oriented sentiment classification. arXiv preprint. arXiv:180501086

  20. Lin C, Zhao S, Meng L, Chua T (2020) Multi-source domain adaptation for visual sentiment classification. In: AAAI. AAAI Press, pp 2661–2668

  21. Lv G, Tong X, Chen E, Yi Z, Yi Z (2016) Reading the videos: temporal labeling for crowdsourced time-sync videos based on semantic embedding. In: AAAI, pp 3000–3006

  22. Maas AL, Daly RE, Pham PT, Huang D, Potts C (2011) Learning word vectors for sentiment analysis. In: Meeting of the Association for Computational Linguistics: Human Language Technologies

  23. Ma X, Cao N (2017) Video-based evanescent, anonymous, asynchronous social interaction: motivation and adaption to medium. In: ACM CSCW, pp 770–782

  24. Ma S, Cui L, Dai D, Wei F, Sun X (2019) Livebot: generating live video comments based on visual and textual contexts. In: AAAI

  25. Maudslay RH, Gonen H, Cotterell R, Teufel S (2019) It’s all in the name: mitigating gender bias with name-based counterfactual data substitution. EMNLP

  26. Maudslay RH, Gonen H, Cotterell R, Teufel S (2020) It’s all in the name: mitigating gender bias with name-based counterfactual data substitution. EMNLP, pp 5267–5275

  27. Mehrabi N, Morstatter F, Saxena N, Lerman K, Galstyan A (2019) A survey on bias and fairness in machine learning. CoRR abs/1908.09635. arxiv:1908.09635

  28. Qian X, Liu X, Ma X, Lu D, Xu C (2016) What is happening in the video? Annotate video by sentence. IEEE Trans Circuits Syst Video Technol 26(9):1746–1757

    Article  Google Scholar 

  29. Ruan S, Zhang K, Wang Y, Tao H, He W, Lv G, Chen E (2020) Context-aware generation-based net for multi-label visual emotion recognition. In: ICME, pp 1–6

  30. Saeidi M, Bouchard G, Liakata M, Riedel S (2016) Sentihood: Targeted aspect based sentiment analysis dataset for urban neighbourhoods. In: Calzolari N, Matsumoto Y, Prasad R (eds) COLING 2016, 26th International Conference on Computational Linguistics, Proceedings of the Conference: Technical Papers, December 11–16, 2016, Osaka, Japan, ACL, pp 1546–1556. https://www.aclweb.org/anthology/C16-1146/

  31. Song Y, Shi S, Li J, Zhang H (2018) Directional skip-gram: explicitly distinguishing left and right context for word embeddings. In: ACL, pp 175–180

  32. Song Y, Wang J, Jiang T, Liu Z, Rao Y (2019) Attentional encoder network for targeted sentiment classification. arXiv preprint. arXiv:190209314

  33. Su Y, Hu W, Jiang J, Su R (2020) A novel LMAEB-CNN model for Chinese microblog sentiment analysis. J Supercomput 76(11):9127–9141. https://doi.org/10.1007/s11227-020-03198-x

    Article  Google Scholar 

  34. Tang D, Qin B, Feng X, Liu T (2015) Effective lstms for target-dependent sentiment classification. arXiv preprint. arXiv:151201100

  35. Tang D, Qin B, Liu T (2016) Aspect level sentiment classification with deep memory network. arXiv preprint arXiv:160508900

  36. Wang B, Yao T, Zhang Q, Xu J, Wang X (2020) Reco: A large scale Chinese reading comprehension dataset on opinion. In: AAAI, pp 9146–9153

  37. Xian Y, Li J, Zhang C, Liao Z (2015) Video highlight shot extraction with time-sync comment. In: International Workshop on Hot Topics in Planet-Scale Mobile Computing and Online Social Networking, pp 31–36

  38. Yang X, Binglu W, Junjie H, Shuwen L (2017b) Natural language processing in “bullet screen” application. In: ICSSSM. IEEE, pp 1–6

  39. Yang W, Ruan N, Gao W, Wang K, Ran W, Jia W (2017a) Crowdsourced time-sync video tagging using semantic association graph. In: ICME, 2017. IEEE, pp 547–552

  40. Yao Y, Bort J, Huang Y (2017) Understanding Danmaku’s potential in online video learning. CHI 2017:3034–3040

    Google Scholar 

  41. Yu J, Jiang J (2019) Adapting bert for target-oriented multimodal sentiment classification. In: Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, IJCAI, pp 5408–5414

  42. Zeng B, Yang H, Xu R, Zhou W, Han X (2019) LCF: A Local context focus mechanism for aspect-based sentiment classification. Applied Sciences (Switzerland) 9(16), https://doi.org/10.3390/app9163389

  43. Zhao Y, Peng X, Tang J, Song S (2017) Understanding young people’s we-intention to contribute in Danmaku websites: motivational, social, and subculture influence. In: Conference 2017 Proceedings

  44. Zhou J, Chen Q, Huang JX, Hu QV, He L (2020a) Position-aware hierarchical transfer model for aspect-level sentiment classification. Inf Sci 513:1–16

    Article  Google Scholar 

  45. Zhou J, Huang JX, Hu QV, He L (2020b) Is position important? Deep multi-task learning for aspect-based sentiment analysis. Appl Intell 50(10):3367–3378

    Article  Google Scholar 

  46. Zhou J, Huang JX, Hu QV, He L (2020d) SK-GCN: modeling syntax and knowledge via graph convolutional network for aspect-level sentiment classification. Knowl Based Syst 205:106292

    Article  Google Scholar 

  47. Zhou J, Huang JX, Hu QV, He L (2020c) Modeling multi-aspect relationship with joint learning for aspect-level sentiment classification. In: DASFAA, pp 786–802

  48. Zhou J, Tian J, Wang R, Wu Y, Xiao W, He L (2020e) Sentix: A sentiment-aware pre-trained model for cross-domain sentiment analysis. In: Proceedings of the 28th International Conference on Computational Linguistics, COLING 2020, Barcelona, Spain (Online), December 8–13

Download references

Acknowledgements

The computation is performed in ECNU Multifunctional Platform for Innovation (001). The authors would like to thank Shijun Zhang and Kai Song for excellent technical support and data annotation support. The authors would also like to thank the editors and the reviewers for their high-quality and constructive comments. These suggestions and comments are important and useful.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Qingchun Bai.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Bai, Q., Wei, K., Zhou, J. et al. Entity-level sentiment prediction in Danmaku video interaction. J Supercomput 77, 9474–9493 (2021). https://doi.org/10.1007/s11227-021-03652-4

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-021-03652-4

Keywords

Navigation