Corpus Analysis and Annotation for Helpful Sentences in Product Reviews


  •  Hana Almagrabi    
  •  Areej Malibari    
  •  John McNaught    

Abstract

For the last two decades, various studies on determining the quality of online product reviews have been concerned with the classification of complete documents into helpful or unhelpful classes using supervised learning methods. As in any supervised machine-learning task, a manually annotated corpus is required to train a model. Corpora annotated for helpful product reviews are an important resource for the understanding of what makes online product reviews helpful and of how to rank them according to their quality. However, most corpora for helpfulness are annotated on the document level: the full review. Little attention has been paid to carrying out a deeper analysis of helpful comments in reviews. In this article, a new annotation scheme is proposed to identify helpful sentences from each product review in the dataset. The annotation scheme, guidelines and the inter-annotator agreement scores are presented and discussed. A high level of inter-annotator agreement is obtained, indicating that the annotated corpus is suitable to support subsequent research.



This work is licensed under a Creative Commons Attribution 4.0 License.
  • Issn(Print): 1913-8989
  • Issn(Onlne): 1913-8997
  • Started: 2008
  • Frequency: quarterly

Journal Metrics

(The data was calculated based on Google Scholar Citations)

Google-based Impact Factor (2018): 18.20

h-index (January 2018): 23

i10-index (January 2018): 90

h5-index (January 2018): 11

h5-median(January 2018):17

Contact