• Chỉ mục bởi
  • Năm xuất bản
LIÊN KẾT WEBSITE

Deceptive reviews and sentiment polarity: Effective link by exploiting BERT

Catelli Institute for High Performance Computing and Networking (ICAR), National Research Council (CNR), Naples, Italy|
Massimo (25926920500) | Giuseppe (6508247659); Esposito i-somet Incorporation association, Morioka, Japan| Hamido (35611951900); De Pietro National Taipei University of Technology, Taipei, Taiwan| Rosario (57210848004); Fujita Faculty of Information Technology, Ho Chi Minh City University of Technology (HUTECH), Ho Chi Minh City, Viet Nam|

Expert Systems with Applications Số , năm 2022 (Tập 209, trang -)

ISSN: 9574174

ISSN: 9574174

DOI:

Tài liệu thuộc danh mục:

Article

English

Từ khóa: Classification (of information); Computational linguistics; Deep learning; Fake detection; Learning systems; Modeling languages; Advertizing media; BERT; Customer spending; Deceptive review; Deep learning; Feedback tool; Language model; Multi-labels; Neural language model; Sentiment; Sales
Tóm tắt tiếng anh
Today, reviews are the advertising medium par excellence through which companies are able to influence customers’ spending decisions. Although the initial purpose of reviews was to provide companies with a feedback tool to improve products and services based on customer needs, they soon became a way to climb the sales rankings, often illegally. In fact, deceptive and fake reviews have managed to evade the often non-existent means of validation of online platforms, proliferating a new business. To combat this phenomenon, several classification methods have been developed to train automated tools in the arduous task of distinguishing between genuine and misleading reviews, the most recent based on machine and deep learning techniques. This paper proposes a multi-label classification methodology based on the Google BERT neural language model to build a deceptive review detector aided by its sentiment awareness: improved modeling of the link between sentiment polarity and deceptiveness during the fine-tuning phase by exploiting the Binary Cross Entropy with Logits loss function adds to the advantages provided by pre-trained contextual models, which are able to capture word polysemy through word embeddings and benefit from pre-training on huge corpora. Tests were performed on the Deceptive Opinion Spam Corpus and Yelp New York City datasets, providing a quantitative and qualitative analysis of the results which, when compared with the state of the art available in the literature, showed an encouraging increase in performance. © 2022 Elsevier Ltd

Xem chi tiết