Filtering Useful App Reviews Using Naïve Bayes—Which Naïve Bayes?

App reviews provide crucial feedback for software maintenance and evolution, but manually extracting useful reviews from vast volumes is time-consuming and challenging. This study investigates the effectiveness of six Naïve Bayes variants for automatically filtering useful app reviews. We evaluated...

Full description

Saved in:
Bibliographic Details
Main Authors: Pouya Ataei, Sri Regula, Daniel Staegemann, Saurabh Malgaonkar
Format: Article
Language:English
Published: MDPI AG 2024-11-01
Series:AI
Subjects:
Online Access:https://www.mdpi.com/2673-2688/5/4/110
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:App reviews provide crucial feedback for software maintenance and evolution, but manually extracting useful reviews from vast volumes is time-consuming and challenging. This study investigates the effectiveness of six Naïve Bayes variants for automatically filtering useful app reviews. We evaluated these variants on datasets from five popular apps, comparing their performance in terms of accuracy, precision, recall, F-measure, and processing time. Our results show that Expectation Maximization-Multinomial Naïve Bayes with Laplace smoothing performed best overall, achieving up to 89.2% accuracy and 0.89 F-measure. Complement Naïve Bayes with Laplace smoothing demonstrated particular effectiveness for imbalanced datasets. Generally, incorporating Laplace smoothing and Expectation Maximization improved performance, albeit with increased processing time. This study also examined the impact of data imbalance on classification performance. Our findings suggest that these advanced Naïve Bayes variants hold promise for filtering useful app reviews, especially when dealing with limited labeled data or imbalanced datasets. This research contributes to the body of evidence around app review mining and provides insights for enhancing software maintenance and evolution processes.
ISSN:2673-2688