skip to main content
Tipo de recurso Mostra resultados com: Mostra resultados com: Índice

On the automatic classification of app reviews.(Report)

Maalej, Walid ; Kurtanovic, Zijad ; Nabil, Hadeer ; Stanik, Christoph

Requirements Engineering, 2016, Vol.21(3), p.311(21) [Periódico revisado por pares]

Texto completo disponível

Citações Citado por
  • Título:
    On the automatic classification of app reviews.(Report)
  • Autor: Maalej, Walid ; Kurtanovic, Zijad ; Nabil, Hadeer ; Stanik, Christoph
  • Assuntos: Metadata – Analysis ; Machine Learning – Analysis ; Natural Language Processing – Analysis
  • É parte de: Requirements Engineering, 2016, Vol.21(3), p.311(21)
  • Descrição: To access, purchase, authenticate, or subscribe to the full-text of this article, please visit this link: Byline: Walid Maalej (1), Zijad Kurtanovic (1), Hadeer Nabil (2), Christoph Stanik (1) Keywords: User feedback; Review analytics; Software analytics; Machine learning; Natural language processing; Data-driven requirements engineering Abstract: App stores like Google Play and Apple AppStore have over 3 million apps covering nearly every kind of software and service. Billions of users regularly download, use, and review these apps. Recent studies have shown that reviews written by the users represent a rich source of information for the app vendors and the developers, as they include information about bugs, ideas for new features, or documentation of released features. The majority of the reviews, however, is rather non-informative just praising the app and repeating to the star ratings in words. This paper introduces several probabilistic techniques to classify app reviews into four types: bug reports, feature requests, user experiences, and text ratings. For this, we use review metadata such as the star rating and the tense, as well as, text classification, natural language processing, and sentiment analysis techniques. We conducted a series of experiments to compare the accuracy of the techniques and compared them with simple string matching. We found that metadata alone results in a poor classification accuracy. When combined with simple text classification and natural language preprocessing of the text--particularly with bigrams and lemmatization--the classification precision for all review types got up to 88--92 % and the recall up to 90--99 %. Multiple binary classifiers outperformed single multiclass classifiers. Our results inspired the design of a review analytics tool, which should help app vendors and developers deal with the large amount of reviews, filter critical reviews, and assign them to the appropriate stakeholders. We describe the tool main features and summarize nine interviews with practitioners on how review analytics tools including ours could be used in practice. Author Affiliation: (1) Department of Informatics, University of Hamburg, Hamburg, Germany (2) German University of Cairo, Cairo, Egypt Article History: Registration Date: 26/04/2016 Received Date: 14/11/2015 Accepted Date: 26/04/2016 Online Date: 14/05/2016
  • Idioma: Inglês

Buscando em bases de dados remotas. Favor aguardar.