How Effectively Do LLMs Extract Feature-Sentiment Pairs from App Reviews?
[Motivation] Automatic analysis of user reviews to understand user sentiments toward app functionality (i.e. app feature) helps align development efforts with user expectations and demands. Recent advances in Large Language Models (LLMs) such as ChatGPT have shown impressive performance on several new tasks without updating the model’s parameters i.e., using zero or a few labeled examples, but the capabilities of LLMs are yet unexplored.[Problem] The goal of our study is to explore the capabilities of LLMs to perform feature-level sentiment analysis of user reviews.[Method] This study compares the performance of state-of-the-art LLMs, including GPT-4, ChatGPT, and different variants of Llama-2 chat, against previous approaches for extracting app features and associated sentiments in 0-shot, 1-shot, and 5-shot scenarios.[Result]The results indicate that GPT-4 outperforms the rule-based SAFE by 17% in f1-score for extracting app features in the 0-shot scenario, with 5-shot further improving it by 6%. However, the fine-tuned RE-BERT exceeds GPT-4 by 6% in f1-score. For predicting positive and neutral sentiments, GPT-4 achieves f1-scores of 76% and 45% in the 0-shot setting, which improve by 7% and 23% in the 5-shot setting, respectively.[Contribution] Our study conducts a thorough evaluation of both proprietary and open-source LLMs to provide an objective assessment of their performance in extracting feature-sentiment pairs.