{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,6,18]],"date-time":"2025-06-18T04:17:25Z","timestamp":1750220245941,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":21,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,4,25]],"date-time":"2022-04-25T00:00:00Z","timestamp":1650844800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2022,4,25]]},"DOI":"10.1145\/3487553.3524219","type":"proceedings-article","created":{"date-parts":[[2022,8,16]],"date-time":"2022-08-16T22:41:30Z","timestamp":1660689690000},"page":"131-135","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Deriving Customer Experience Implicitly from Social Media"],"prefix":"10.1145","author":[{"given":"Aditya","family":"Kumar","sequence":"first","affiliation":[{"name":"Flipkart, India"}]},{"given":"Sneh","family":"Gupta","sequence":"additional","affiliation":[{"name":"Flipkart, India"}]},{"given":"Ankit","family":"Sahu","sequence":"additional","affiliation":[{"name":"Flipkart, India"}]},{"given":"Mayank","family":"Kant","sequence":"additional","affiliation":[{"name":"Flipkart, India"}]}],"member":"320","published-online":{"date-parts":[[2022,8,16]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Tom\u00a0B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel\u00a0M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. arxiv:2005.14165\u00a0[cs.CL]  Tom\u00a0B. Brown Benjamin Mann Nick Ryder Melanie Subbiah Jared Kaplan Prafulla Dhariwal Arvind Neelakantan Pranav Shyam Girish Sastry Amanda Askell Sandhini Agarwal Ariel Herbert-Voss Gretchen Krueger Tom Henighan Rewon Child Aditya Ramesh Daniel\u00a0M. Ziegler Jeffrey Wu Clemens Winter Christopher Hesse Mark Chen Eric Sigler Mateusz Litwin Scott Gray Benjamin Chess Jack Clark Christopher Berner Sam McCandlish Alec Radford Ilya Sutskever and Dario Amodei. 2020. Language Models are Few-Shot Learners. arxiv:2005.14165\u00a0[cs.CL]"},{"key":"e_1_3_2_1_2_1","volume-title":"A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks 106 (Oct","author":"Buda Mateusz","year":"2018","unstructured":"Mateusz Buda , Atsuto Maki , and Maciej\u00a0 A. Mazurowski . 2018. A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks 106 (Oct 2018 ), 249\u2013259. https:\/\/doi.org\/10.1016\/j.neunet.2018.07.011 10.1016\/j.neunet.2018.07.011 Mateusz Buda, Atsuto Maki, and Maciej\u00a0A. Mazurowski. 2018. A systematic study of the class imbalance problem in convolutional neural networks. Neural Networks 106 (Oct 2018), 249\u2013259. https:\/\/doi.org\/10.1016\/j.neunet.2018.07.011"},{"key":"e_1_3_2_1_3_1","volume-title":"XGBoost. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (Aug","author":"Chen Tianqi","year":"2016","unstructured":"Tianqi Chen and Carlos Guestrin . 2016 . XGBoost. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (Aug 2016). https:\/\/doi.org\/10.1145\/2939672.2939785 10.1145\/2939672.2939785 Tianqi Chen and Carlos Guestrin. 2016. XGBoost. Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (Aug 2016). https:\/\/doi.org\/10.1145\/2939672.2939785"},{"key":"e_1_3_2_1_4_1","volume-title":"Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805(2018).","author":"Devlin Jacob","year":"2018","unstructured":"Jacob Devlin , Ming-Wei Chang , Kenton Lee , and Kristina Toutanova . 2018 . Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805(2018). Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805(2018)."},{"key":"e_1_3_2_1_5_1","unstructured":"Katherine Lee Daphne Ippolito Andrew Nystrom Chiyuan Zhang Douglas Eck Chris Callison-Burch and Nicholas Carlini. 2021. Deduplicating Training Data Makes Language Models Better. arXiv preprint arXiv:2107.06499(2021).  Katherine Lee Daphne Ippolito Andrew Nystrom Chiyuan Zhang Douglas Eck Chris Callison-Burch and Nicholas Carlini. 2021. Deduplicating Training Data Makes Language Models Better. arXiv preprint arXiv:2107.06499(2021)."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Xin Li Lidong Bing Wenxuan Zhang and Wai Lam. 2019. Exploiting BERT for end-to-end aspect-based sentiment analysis. arXiv preprint arXiv:1910.00883(2019).  Xin Li Lidong Bing Wenxuan Zhang and Wai Lam. 2019. Exploiting BERT for end-to-end aspect-based sentiment analysis. arXiv preprint arXiv:1910.00883(2019).","DOI":"10.18653\/v1\/D19-5505"},{"key":"e_1_3_2_1_7_1","volume-title":"Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692(2019).","author":"Liu Yinhan","year":"2019","unstructured":"Yinhan Liu , Myle Ott , Naman Goyal , Jingfei Du , Mandar Joshi , Danqi Chen , Omer Levy , Mike Lewis , Luke Zettlemoyer , and Veselin Stoyanov . 2019 . Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692(2019). Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, and Veselin Stoyanov. 2019. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692(2019)."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"crossref","unstructured":"Clifton Poth Jonas Pfeiffer Andreas R\u00fcckl\u00e9 and Iryna Gurevych. 2021. What to Pre-Train on? Efficient Intermediate Task Selection. arxiv:2104.08247\u00a0[cs.CL]  Clifton Poth Jonas Pfeiffer Andreas R\u00fcckl\u00e9 and Iryna Gurevych. 2021. What to Pre-Train on? Efficient Intermediate Task Selection. arxiv:2104.08247\u00a0[cs.CL]","DOI":"10.18653\/v1\/2021.emnlp-main.827"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"crossref","unstructured":"Yada Pruksachatkun Jason Phang Haokun Liu Phu\u00a0Mon Htut Xiaoyi Zhang Richard\u00a0Yuanzhe Pang Clara Vania Katharina Kann and Samuel\u00a0R. Bowman. 2020. Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?arxiv:2005.00628\u00a0[cs.CL]  Yada Pruksachatkun Jason Phang Haokun Liu Phu\u00a0Mon Htut Xiaoyi Zhang Richard\u00a0Yuanzhe Pang Clara Vania Katharina Kann and Samuel\u00a0R. Bowman. 2020. Intermediate-Task Transfer Learning with Pretrained Models for Natural Language Understanding: When and Why Does It Work?arxiv:2005.00628\u00a0[cs.CL]","DOI":"10.18653\/v1\/2020.acl-main.467"},{"key":"e_1_3_2_1_10_1","volume-title":"Induction of decision trees. Machine learning 1, 1","author":"Quinlan Ross","year":"1986","unstructured":"J.\u00a0 Ross Quinlan . 1986. Induction of decision trees. Machine learning 1, 1 ( 1986 ), 81\u2013106. J.\u00a0Ross Quinlan. 1986. Induction of decision trees. Machine learning 1, 1 (1986), 81\u2013106."},{"key":"e_1_3_2_1_11_1","volume":"201","author":"Raffel Colin","unstructured":"Colin Raffel , Noam Shazeer , Adam Roberts , Katherine Lee , Sharan Narang , Michael Matena , Yanqi Zhou , Wei Li , and Peter\u00a0 J Liu. 201 9. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683(2019). Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, and Peter\u00a0J Liu. 2019. Exploring the limits of transfer learning with a unified text-to-text transformer. arXiv preprint arXiv:1910.10683(2019).","journal-title":"J Liu."},{"key":"e_1_3_2_1_12_1","volume-title":"The one number you need to grow. Harvard business review 81, 12","author":"Reichheld F","year":"2003","unstructured":"Frederick\u00a0 F Reichheld . 2003. The one number you need to grow. Harvard business review 81, 12 ( 2003 ), 46\u201355. Frederick\u00a0F Reichheld. 2003. The one number you need to grow. Harvard business review 81, 12 (2003), 46\u201355."},{"key":"e_1_3_2_1_13_1","unstructured":"Emanuel\u00a0H Silva and Ricardo\u00a0M Marcacini. [n. d.]. Aspect-based Sentiment Analysis using BERT with Disentangled Attention. ([n. d.]).  Emanuel\u00a0H Silva and Ricardo\u00a0M Marcacini. [n. d.]. Aspect-based Sentiment Analysis using BERT with Disentangled Attention. ([n. d.])."},{"key":"e_1_3_2_1_14_1","unstructured":"Youwei Song Jiahai Wang Zhiwei Liang Zhiyue Liu and Tao Jiang. 2020. Utilizing BERT intermediate layers for aspect based sentiment analysis and natural language inference. arXiv preprint arXiv:2002.04815(2020).  Youwei Song Jiahai Wang Zhiwei Liang Zhiyue Liu and Tao Jiang. 2020. Utilizing BERT intermediate layers for aspect based sentiment analysis and natural language inference. arXiv preprint arXiv:2002.04815(2020)."},{"key":"e_1_3_2_1_15_1","unstructured":"Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998\u20136008.  Ashish Vaswani Noam Shazeer Niki Parmar Jakob Uszkoreit Llion Jones Aidan\u00a0N Gomez \u0141ukasz Kaiser and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998\u20136008."},{"key":"e_1_3_2_1_16_1","volume-title":"Superglue: A stickier benchmark for general-purpose language understanding systems. arXiv preprint arXiv:1905.00537(2019).","author":"Wang Alex","year":"2019","unstructured":"Alex Wang , Yada Pruksachatkun , Nikita Nangia , Amanpreet Singh , Julian Michael , Felix Hill , Omer Levy , and Samuel\u00a0 R Bowman . 2019 . Superglue: A stickier benchmark for general-purpose language understanding systems. arXiv preprint arXiv:1905.00537(2019). Alex Wang, Yada Pruksachatkun, Nikita Nangia, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel\u00a0R Bowman. 2019. Superglue: A stickier benchmark for general-purpose language understanding systems. arXiv preprint arXiv:1905.00537(2019)."},{"key":"e_1_3_2_1_17_1","volume-title":"GLUE: A multi-task benchmark and analysis platform for natural language understanding. arXiv preprint arXiv:1804.07461(2018).","author":"Wang Alex","year":"2018","unstructured":"Alex Wang , Amanpreet Singh , Julian Michael , Felix Hill , Omer Levy , and Samuel\u00a0 R Bowman . 2018 . GLUE: A multi-task benchmark and analysis platform for natural language understanding. arXiv preprint arXiv:1804.07461(2018). Alex Wang, Amanpreet Singh, Julian Michael, Felix Hill, Omer Levy, and Samuel\u00a0R Bowman. 2018. GLUE: A multi-task benchmark and analysis platform for natural language understanding. arXiv preprint arXiv:1804.07461(2018)."},{"key":"e_1_3_2_1_18_1","volume-title":"The Free Encyclopedia. https:\/\/en.wikipedia.org\/w\/index.php?title=Gradient_boosting&oldid=1055694616 [Online","author":"Wikipedia","year":"2021","unstructured":"Wikipedia contributors. 2021. Gradient boosting \u2014 Wikipedia , The Free Encyclopedia. https:\/\/en.wikipedia.org\/w\/index.php?title=Gradient_boosting&oldid=1055694616 [Online ; accessed 18- November - 2021 ]. Wikipedia contributors. 2021. Gradient boosting \u2014 Wikipedia, The Free Encyclopedia. https:\/\/en.wikipedia.org\/w\/index.php?title=Gradient_boosting&oldid=1055694616 [Online; accessed 18-November-2021]."},{"key":"e_1_3_2_1_19_1","volume-title":"The Free Encyclopedia. https:\/\/en.wikipedia.org\/w\/index.php?title=Net_promoter_score&oldid=1054323643 [Online","author":"Wikipedia","year":"2021","unstructured":"Wikipedia contributors. 2021. Net promoter score \u2014 Wikipedia , The Free Encyclopedia. https:\/\/en.wikipedia.org\/w\/index.php?title=Net_promoter_score&oldid=1054323643 [Online ; accessed 17- November - 2021 ]. Wikipedia contributors. 2021. Net promoter score \u2014 Wikipedia, The Free Encyclopedia. https:\/\/en.wikipedia.org\/w\/index.php?title=Net_promoter_score&oldid=1054323643 [Online; accessed 17-November-2021]."},{"key":"e_1_3_2_1_20_1","volume-title":"Context-guided bert for targeted aspect-based sentiment analysis","author":"Wu Zhengxuan","year":"2020","unstructured":"Zhengxuan Wu and Desmond\u00a0 C Ong . 2020. Context-guided bert for targeted aspect-based sentiment analysis . Association for the Advancement of Artificial Intelligence ( 2020 ), 1\u20139. Zhengxuan Wu and Desmond\u00a0C Ong. 2020. Context-guided bert for targeted aspect-based sentiment analysis. Association for the Advancement of Artificial Intelligence (2020), 1\u20139."},{"key":"e_1_3_2_1_21_1","unstructured":"Hu Xu Bing Liu Lei Shu and Philip\u00a0S Yu. 2019. BERT post-training for review reading comprehension and aspect-based sentiment analysis. arXiv preprint arXiv:1904.02232(2019).  Hu Xu Bing Liu Lei Shu and Philip\u00a0S Yu. 2019. BERT post-training for review reading comprehension and aspect-based sentiment analysis. arXiv preprint arXiv:1904.02232(2019)."}],"event":{"name":"WWW '22: The ACM Web Conference 2022","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"],"location":"Virtual Event, Lyon France","acronym":"WWW '22"},"container-title":["Companion Proceedings of the Web Conference 2022"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3487553.3524219","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3487553.3524219","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T19:30:33Z","timestamp":1750188633000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3487553.3524219"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,25]]},"references-count":21,"alternative-id":["10.1145\/3487553.3524219","10.1145\/3487553"],"URL":"https:\/\/doi.org\/10.1145\/3487553.3524219","relation":{},"subject":[],"published":{"date-parts":[[2022,4,25]]},"assertion":[{"value":"2022-08-16","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}