Open Access

Effective Opinion Spam Detection: A Study on Review Metadata Versus Content


Cite

Figure 1

Framework of proposed research methodology.
Framework of proposed research methodology.

Figure 2

Tripartite network of reviewers (U), reviews (R) and products (P).
Tripartite network of reviewers (U), reviews (R) and products (P).

Figure 3

Overall process from dataset to model training through feature extraction and balancing for any of the reviewer, review and product-centric settings.
Overall process from dataset to model training through feature extraction and balancing for any of the reviewer, review and product-centric settings.

Figure 4

ROC curves of different classifiers trained using behavioral and textual features on YelpZip over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.
ROC curves of different classifiers trained using behavioral and textual features on YelpZip over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.

Figure 5

ROC curves of different classifiers trained using behavioral and textual features on YelpNYC over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.
ROC curves of different classifiers trained using behavioral and textual features on YelpNYC over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.

Figure 6

Performance of behavioral, textual and hybrid features using different classifiers on YelpZip over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.
Performance of behavioral, textual and hybrid features using different classifiers on YelpZip over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.

Figure 7

Performance of behavioral, textual and hybrid features using different classifiers on YelpNYC over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.
Performance of behavioral, textual and hybrid features using different classifiers on YelpNYC over (a–c) reviewer-centric, (d–f) review-centric, and (g–i) product-centric setting.

Figure 8

Performance comparison of different features used in different works on YelpZip over (a–b) reviewer-centric, (c–d) review-centric, and (e–f) product-centric setting.
Performance comparison of different features used in different works on YelpZip over (a–b) reviewer-centric, (c–d) review-centric, and (e–f) product-centric setting.

Figure 9

Performance comparison of different features used in different works on YelpNYC over (a–b) reviewer-centric, (c–d) review-centric, and (e–f) product-centric setting.
Performance comparison of different features used in different works on YelpNYC over (a–b) reviewer-centric, (c–d) review-centric, and (e–f) product-centric setting.

Figure 10

Computation time analysis of behavioral and textual feature extraction on (a) YelpZip and (b) YelpNYC dataset.
Computation time analysis of behavioral and textual feature extraction on (a) YelpZip and (b) YelpNYC dataset.

Classifiers performance on YelpNYC dataset using both behavioral and textual features over all three settings.

SVMLRMLPNB
(a) Reviewer-centric
BehavioralTextualBehavioralTextualBehavioralTextualBehavioralTextual
AP0.80010.68660.80220.69990.81360.70290.75940.6860
Recall0.66380.59450.68590.70110.70450.68780.55030.3103
F1 (Macro)0.71400.62500.71480.64950.72240.64720.67490.5462
F1 (Micro)0.71430.62590.71490.65170.72260.64910.67910.5759
(b) Review-centric
AP0.73130.65660.73110.64440.74610.67080.68110.6640
Recall0.71890.52700.74750.50730.75480.60890.34620.3611
F1 (Macro)0.65730.60520.67280.58970.68220.61730.56210.5655
F1 (Micro)0.65980.60690.67580.59170.68520.61790.58490.5851
(c) Product-centric
AP0.88390.83450.88760.83670.88960.83450.89090.8357
Recall0.84740.33960.82820.68440.81860.67700.84190.7006
F1 (Macro)0.78650.61770.80160.73690.79740.70480.80660.7332
F1 (Micro)0.78800.66010.80240.73850.79830.71090.80730.7344

Brief summary of features used by comparing methods under reviewer-centric, review-centric and product-centric settings.

Mukherjee et al. (2013a) FeaturesMukherjee et al. (2013c) FeaturesRayana & Akoglu (2015) Features
Reviewer-centric and Product-centricReview-centricReviewer-centric and Product-centricReview-centricReviewer-centric and Product-centricReview-centric
CSDUPMNRMNRRank
MNREXTPRPRRD
BSTDEVRLNREXT
RFRETFRDavgRDDEV
MCSWRDETF
BSTPCW
ERDPC
ETGL
RLPP1
ACSRES
MCSSW
OW
DLu
DLb

Dataset statistics after preprocessing (for YelpZip and YelpNYC).

Dataset# Reviews (spam%)# Reviewers (spammer%)# Products (restaurants)
YelpZip (Preprocessed)356,766 (4.66%)49,841 (9.21%)3,975
YelpNYC (Preprocessed)90,906 (7.58%)15,351 (10.67%)873

Classifiers performance on YelpZip dataset using both behavioral and textual features over all three settings.

SVMLRMLPNB
(a) Reviewer-centric
BehavioralTextualBehavioralTextualBehavioralTextualBehavioralTextual
AP0.73420.66820.73770.67170.74170.67830.69340.6558
Recall0.53010.60630.59070.65370.63950.64970.59020.3140
F1 (Macro)0.67000.62600.68410.63400.69430.63430.66810.5491
F1 (Micro)0.67670.62680.68680.63430.69520.63530.67010.5808
(b) Review-centric
AP0.68730.64610.68260.62320.69940.65810.64010.6478
Recall0.78210.43480.74130.37750.71210.59470.69070.3612
F1 (Macro)0.63940.58880.65440.56550.66370.61800.62330.5663
F1 (Micro)0.64710.59980.65740.58300.66500.61870.62590.5876
(c) Product-centric
AP0.86920.84400.87170.84210.87410.84990.86910.8432
Recall0.82180.77950.81010.77740.79260.77310.90041.0000
F1 (Macro)0.74880.72790.75260.72930.75690.73470.67580.3537
F1 (Micro)0.75410.73210.75690.73330.75980.73840.70200.5472

Algorithm for balancing the feature set.

Algorithm 1: Balancing Algorithm for Feature Set
  Input: Unbalanced feature set F.
  Output:k balanced partitions each containing nearly equal number of instances from both the classes.
1. Randomly shuffle the instances in F;
2. Divide F into two sets S1 and S2 representing majority class and minority class, respectively;
3.S1 ← minority class instances;
4.S2 ← majority class instances;
5.p ← count(S1);
6.q ← count(S2);
7.kqpk \leftarrow \left\lceil {{q \over p}} \right\rceil ;
8.S3 ← Divide S2 into k nearly equal size bins;
9. foreach bin zS3do
10. Combine S1 with z to get a balanced partition;
11. end
12. returnk balanced partitions for unbalanced feature set F;

Dataset statistics (for YelpZip and YelpNYC).

Dataset# Reviews (spam %)# Reviewers (spammer %)# Products (restaurants)
YelpZip608,598 (13.22%)260,277 (23.91%)5,044
YelpNYC359,052 (10.2 7%)160,225 (17.79%)923

Brief description of behavioral and textual features employed under reviewer-centric, review-centric and product-centric settings.

SettingFeaturetypeFeatureDescription
Reviewer-centric and Product-centricBehavioralARDAverage rating deviation (Fei et al., 2013)
WRDWeighted rating deviation (Rayana and Akoglu, 2015)
MRD*Maximum rating deviation
BSTBurstiness (Mukherjee et al., 2013a)
ERR*Early review ratio
MNRMaximum number of reviews (Mukherjee et al., 2013a)
RPRRatio of positive reviews (Rayana and Akoglu, 2015)
RNRRatio of negative reviews (Rayana and Akoglu, 2015)
FRRFirst review ratio (Mukherjee et al., 2013a)
EXRR*Extreme rating ratio
TRRR*Top ranked reviews ratio
BRRR*Bottom ranked reviews ratio
TextualMCSMaximum content similarity (Mukherjee et al., 2013a)
ACSAverage content similarity (Lim et al., 2010)
AFPP*Average first-person pronouns ratio
ASPP*Average second-person pronouns ratio
AFTAPP*Average first-and-third-person to all-person pronouns ratio
ASAPP*Average second-person to all-person pronouns ratio
ASW*Average subjective words ratio
AOW*Average objective words ratio
AInW*Average informative words ratio
AImW*Average imaginative words ratio
ARLAverage review length (Rayana and Akoglu, 2015)
Review-centricBehavioralRDRating deviation (Mukherjee et al., 2013a)
ERD*Early rating deviation
ETFEarly time frame (Mukherjee et al., 2013a)
EXTExtreme rating (Mukherjee et al., 2013a)
TRR*Top ranked review
BRR*Bottom ranked review
RRReview rank (Rayana and Akoglu, 2015)
RLReview length (Mukherjee et al., 2013c)
TextualRPWRatio of positive words (Li et al., 2011)
RNWRatio of negative words (Li et al., 2011)
RFPPRatio of first-person pronouns (Li et al., 2011)
RSPPRatio of second-person pronouns (Li et al., 2011)
RFTAPP*Ratio of first-and-third-person to all-person pronouns
RSAPP*Ratio of second-person to all-person pronouns
RSWRatio of subjective words (Li et al., 2011)
ROWRatio of objective words (Li et al., 2011)
RInWRatio of informative words (Ott et al., 2011)
RImWRatio of imaginative words (Ott et al., 2011)

Statistical significance of results obtained on behavioral and textual features using Z-test analysis.

Reviewer-centricReview-centricProduct-centric
Z-test statisticP-valueZ-test statisticP-valueZ-test statisticP-value
YelpZipROC-AUC30.03~ 0.053.400.03.140.0016
Avg. Precision27.69~ 0.037.58~ 0.03.310.0009
F1-Score (micro)20.88~ 0.048.910.02.070.0377
YelpNYCROC-AUC23.02~ 0.047.440.04.59~ 0.0
Avg. Precision23.35~ 0.033.48~ 0.03.860.0001
F1-Score (micro)22.41~ 0.030.17~ 0.08.73~ 0.0
eISSN:
2543-683X
Language:
English
Publication timeframe:
4 times per year
Journal Subjects:
Computer Sciences, Information Technology, Project Management, Databases and Data Mining