Abstract
There is a growing concern that online reviews are targets of systematic manipulation, and manipulated reviews serve purposes other than informing consumers. In this article, we report a cross-site comparison of the aggregate-level manipulation using Benford’s law to detect anomalies. Benford’s law states that digits in naturally occurring data follow a logarithmic distribution. Deviation from such distribution is considered as a sign of systematic manipulation. We empirically examine word-count distributions of reviews on a Chinese food delivery service platform (FDS), Dianping, Yelp, and Amazon. Our empirical analysis suggests, in general, word counts of online review contents do not obey Benford’s law, although Benford’s law holds among high-quality reviews. Deviation from Benford’s law is larger in emerging markets compared with mature online marketplaces. Further analyses reveal that positive reviews, especially positive and extreme reviews, exhibit more deviation from Benford’s law.




Similar content being viewed by others
Notes
For example, Amazon invests heavily in manual systems and automatic algorithms to identify fraudulent reviews. https://www.buzzfeednews.com/article/nicolenguyen/amazon-fake-review-problem.
For example, Amazon prohibits third-party sellers from offering free discounts to consumers for exchanging reviews. https://www.digitalcommerce360.com/2016/10/07/amazon-tells-sellers-stop-offering-free-products-reviews/.
Platforms set limits to the number of words contained in a review. For example, the FDS platform that we study has a limit of 500 characters, and Amazon sets 5000 as the ceiling of review word counts. Only 0.02% of FDS reviews in our dataset have more than 350 word counts, and 0.05% of Amazon reviews have more than 1000 word counts. The regulation aims to assure the function of the review systems rather than constrain review writing.
To form a comparable sample with the FDS review data that covers reviews posted between July 2016 and September 2016, we randomly chose the latest reviews (submitted in August 2015) in the Amazon dataset.
Consumers can click and access the detailed information of these suspected reviews from a separate page.
No ground truth or algorithm can assure a review is truly manipulated or not in observational data. Reviews with helpful votes and invited reviews have been repeatedly used as high-quality reviews. Hence, we believe that these reviews are not subject to systematic manipulation. We further admit that it is better to conduct experiments. However, since the main purpose of the paper is to conduct the cross-sample analysis of the aggregate-level systematic manipulation, we leave this as a future research direction.
These non-compliant promotional reviews are five-star reviews that violate Amazon’s guidelines. The guideline can be found on Amazon’s website.
Due to the space limit, we present full tables in the appendix.
Considering the smallest sample size is 6,930 for negative reviews on Dianping, we chose 6,500 as the sample size for all categories on the four sites.
The number of review observations is greater than 900 on both sites in and after 2010, while less than 300 before 2010.
For those years in which we have less than 3,000 observations, we randomly drew samples of 3,000 reviews with replacement. For the other years, we drew random samples without replacement.
We thank an anonymous reviewer for pointing out that cultural differences could influence the effectiveness of manipulation detection methods. Benford’s law has been tested in different cultures, which strengthens its credibility in cross-cultural studies. However, it is an interesting extension to conduct more comprehensive empirical tests of Benford’s law applicability in different cultures. We do not include such tests in the paper due to data availability limitations.
References
Kwark, Y., Chen, J., & Raghunathan, S. (2014). Online product reviews: Implications for retailers and competing manufacturers. Information Systems Research, 25(1), 93–110.
Duan, W., Gu, B., & Whinston, A. B. (2008). Do online reviews matter? An empirical investigation of panel data. Decision Support Systems, 45(4), 1007–1016.
Chevalier, J. A., & Mayzlin, D. (2006). The effect of word of mouth on sales: Online book reviews. Journal of Marketing Research, 43(3), 345–354.
Forman, C., Ghose, A., & Wiesenfeld, B. (2008). Examining the relationship between reviews and sales: The role of reviewer identity disclosure in electronic markets. Information Systems Research, 19(3), 291–313.
Wang, C., Zhang, X., & Hann, I.-H. (2018). Socially nudged: A quasi-experimental study of friends’ social influence in online product ratings. Information Systems Research, 29(3), 641–655.
Luca, M., & Zervas, G. (2016). Fake it till you make it: Reputation, competition, and Yelp review fraud. Management Science, 62(12), 3412–3427.
Mayzlin, D., Dover, Y., & Chevalier, J. (2014). Promotional reviews: An empirical investigation of online review manipulation. American Economic Review, 104(8), 2421–2455.
Ghose, A., & Ipeirotis, P. G. (2011). Estimating the helpfulness and economic impact of product reviews: Mining text and reviewer characteristics. IEEE Transactions on Knowledge and Data Engineering, 23(10), 1498–1512.
Susan, M. M., & David, S. (2010). What makes a helpful online review? A study of customer reviews on Amazon.com. MIS Quarterly, 34(1), 185–200.
Ott, M., Choi, Y., Cardie, C., & Hancock, J. T. (2011). Finding deceptive opinion spam by any stretch of the imagination. In Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, 1, 309–319.
Kumar, N., Venugopal, D., Qiu, L., & Kumar, S. (2019). Detecting anomalous online reviewers: An unsupervised approach using mixture models. Journal of Management Information Systems, 36(4), 1313–1346.
Mukherjee, A., Kumar, A., Liu, B., Wang, J., Hsu, M., Castellanos, M., & Ghosh, R. (2013). Spotting opinion spammers using behavioral footprints. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 19, 632–640.
Liang, N., Biros, D. P., & Luse, A. (2016). An empirical validation of malicious insider characteristics. Journal of Management Information Systems, 33(2), 361–392.
Jindal, N., & Liu, B. Opinion spam and analysis. (2008). In Proceedings of the 2008 International Conference on Web Search and Data Mining, 219–230: ACM.
Akoglu, L., Chandy, R., & Faloutsos, C. (2013). Opinion fraud detection in online reviews by network effects. In Proceedings of the 7th international AAAI Conference on Weblogs and Social Media, 7, 2–11.
Rayana, S., & Akoglu, L. (2015). Collective opinion spam detection: Bridging review networks and metadata. In Proceedings of the 21st ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, 21, 985–994.
Wang, G., Xie, S., Liu, B., & Philip, S. Y. (2011). Review graph based online store review spammer detection. In Proceedings of the 11th IEEE International Conference on Data Mining, 11, 1242–1247.
Judge, G., & Schechter, L. (2009). Detecting problems in survey data using Benford’s law. Journal of Human Resources, 44(1), 1–24.
Skousen, C. J., Guan, L., & Wetzel, T. S. (2004). Anomalies and unusual patterns in reported earnings: Japanese managers round earnings. Journal of International Financial Management & Accounting, 15(3), 212–234.
Rodriguez, R. J. (2004). First significant digit patterns from mixtures of uniform distributions. The American Statistician, 58(1), 64–71.
Durtschi, C., Hillison, W., & Pacini, C. (2004). The effective use of Benford’s law to assist in detecting fraud in accounting data. Journal of Forensic Accounting, 5(1), 17–34.
Guan, L. M., Lin, F. Y., & Fang, W. C. (2008). Goal-oriented earnings management: Evidence from Taiwanese firms. Emerging Markets Finance and Trade, 44(4), 19–32.
Lin, F., Zhao, L., & Guan, L. (2014). Window dressing in reported earnings: A comparison of high-tech and low-tech companies. Emerging Markets Finance and Trade, 50(sup1), 254–264.
Lin, F., Lin, L.-J., Yeh, C.-C., & Wang, T.-S. (2018). Does the board of directors as Fat Cats exert more earnings management? Evidence from Benford’s law. The Quarterly Review of Economics and Finance, 68, 158–170.
Chen, Z., & Lurie, N. H. (2013). Temporal contiguity and negativity bias in the impact of online word of mouth. Journal of Marketing Research, 50(4), 463–476.
Yin, D., Mitra, S., & Zhang, H. (2016). When do consumers value positive vs. negative reviews? An empirical investigation of confirmation bias in online word of mouth. Information Systems Research, 27(1), 131–144.
Shihab, M. R., & Putri, A. P. (2019). Negative online reviews of popular products: Understanding the effects of review proportion and quality on consumers’ attitude and intention to buy. Electronic Commerce Research, 19(1), 159–187.
Godes, D., & Mayzlin, D. (2004). Using online conversations to study word-of-mouth communication. Marketing Science, 23(4), 545–560.
Chintagunta, P. K., Gopinath, S., & Venkataraman, S. (2010). The effects of online user reviews on movie box office performance: Accounting for sequential rollout and aggregation across local markets. Marketing Science, 29(5), 944–957.
Dellarocas, C., Zhang, X. M., & Awad, N. F. (2007). Exploring the value of online product reviews in forecasting sales: The case of motion pictures. Journal of Interactive Marketing, 21(4), 23–45.
Duan, W., Gu, B., & Whinston, A. B. (2008). The dynamics of online word-of-mouth and product sales—An empirical investigation of the movie industry. Journal of Retailing, 84(2), 233–242.
Gu, B., Park, J., & Konana, P. (2012). The impact of external word-of-mouth sources on retailer sales of high-involvement products. Information Systems Research, 23(1), 182–196.
Zimmermann, S., Herrmann, P., Kundisch, D., & Nault, B. R. (2018). Decomposing the variance of consumer ratings and the impact on price and demand. Information Systems Research, 29(4), 984–1002.
Cheng, L.-C., & Huang, C.-L. (2019). Exploring contextual factors from consumer reviews affecting movie sales: An opinion mining approach. Electronic Commerce Research, 1–26.
Chen, Y., & Xie, J. (2008). Online consumer review: Word-of-mouth as a new element of marketing communication mix. Management Science, 54(3), 477–491.
Hollenbeck, B., Moorthy, S., & Proserpio, D. (2019). Advertising strategy in the presence of reviews: An empirical analysis. Marketing Science, 38(5), 793–811.
Lovett, M. J., Peres, R., & Shachar, R. (2013). On brands and word of mouth. Journal of Marketing Research, 50(4), 427–444.
Li, X., & Hitt, L. M. (2008). Self-selection and information role of online product reviews. Information Systems Research, 19(4), 456–474.
Dellarocas, C., & Wood, C. A. (2008). The sound of silence in online feedback: Estimating trading risks in the presence of reporting bias. Management Science, 54(3), 460–476.
Hu, N., Pavlou, P. A., & Zhang, J. (2009). Overcoming the J-shaped distribution of product reviews. Communications of the ACM, 52(10), 144–147.
Godes, D., & Silva, J. C. (2012). Sequential and temporal dynamics of online opinion. Marketing Science, 31(3), 448–473.
Moe, W. W., & Trusov, M. (2011). The value of social dynamics in online product ratings forums. Journal of Marketing Research, 48(3), 444–456.
Goes, P. B., Lin, M., & Au Yeung, C. (2014). “Popularity effect” in user-generated content: Evidence from online product reviews. Information Systems Research, 25(2), 222–238.
Lee, Y.-J., Hosanagar, K., & Tan, Y. (2015). Do I follow my friends or the crowd? Information cascades in online movie ratings. Management Science, 61(9), 2241–2258.
Burtch, G., Hong, Y., Bapna, R., & Griskevicius, V. (2017). Stimulating online reviews by combining financial incentives and social norms. Management Science, 64(5), 2065–2082.
Huang, N., Hong, Y., & Burtch, G. (2016). Social network integration and user content generation: Evidence from natural experiments. MIS Quarterly, 17–001.
Ho, Y.-C., Wu, J., & Tan, Y. (2017). Disconfirmation effect on online rating behavior: A structural model. Information Systems Research, 28(3), 626–642.
Chen, P.-Y., Hong, Y., & Liu, Y. (2017). The value of multidimensional rating systems: Evidence from a natural experiment and randomized experiments. Management Science, 64(10), 4629–4647.
Wang, H., Du, R., Li, J., & Fan, W. (2018). Subdivided or aggregated online review systems: Which is better for online takeaway vendors? Electronic Commerce Research, 1–30.
Kumar, N., Qiu, L., & Kumar, S. (2018). Exit, voice, and response on digital platforms: An empirical investigation of online management response strategies. Information Systems Research, 29(4), 849–870.
Lappas, T., Sabnis, G., & Valkanas, G. (2016). The impact of fake reviews on online visibility: A vulnerability assessment of the hotel industry. Information Systems Research, 27(4), 940–961.
Zhao, Y., Yang, S., Narayan, V., & Zhao, Y. (2013). Modeling consumer learning from online product reviews. Marketing Science, 32(1), 153–169.
Khern-am-nuai, W., Kannan, K., & Ghasemkhani, H. (2018). Extrinsic versus intrinsic rewards for contributing reviews in an online platform. Information Systems Research, 29(4), 871–892.
Sun, Y., Dong, X., & McIntyre, S. (2017). Motivation of user-generated content: Social connectedness moderates the effects of monetary rewards. Marketing Science, 36(3), 329–337.
Lin, Z., Zhang, Y., & Tan, Y. (2019). An empirical study of free product sampling and rating bias. Information Systems Research, 30(1), 260–275.
Nunamaker, J. F., Jr., Burgoon, J. K., & Giboney, J. S. (2016). Information systems for deception detection. Journal of Management Information Systems, 33(2), 327–331.
Zhang, D., Zhou, L., Kehoe, J. L., & Kilic, I. Y. (2016). What online reviewer behaviors really matter? Effects of verbal and nonverbal behaviors on detection of fake online reviews. Journal of Management Information Systems, 33(2), 456–481.
Kumar, N., Venugopal, D., Qiu, L., & Kumar, S. (2018). Detecting review manipulation on online platforms with hierarchical supervised learning. Journal of Management Information Systems, 35(1), 350–380.
Hu, N., Bose, I., Koh, N. S., & Liu, L. (2012). Manipulation of online reviews: An analysis of ratings, readability, and sentiments. Decision Support Systems, 52(3), 674–684.
Saumya, S., & Singh, J. P. (2020). Spam review detection using LSTM autoencoder: An unsupervised approach. Electronic Commerce Research, 1–21.
Feng, S., Xing, L., Gogar, A., & Choi, Y. (2012). Distributional footprints of deceptive product reviews. In Proceedings of the 6th International AAAI Conference on Weblogs and Social Media, 12(98), 105.
Newcomb, S. (1881). Note on the frequency of use of the different digits in natural numbers. American Journal of Mathematics, 4(1), 39–40.
Benford, F. (1938). The law of anomalous numbers. Proceedings of the American Philosophical Society, 551–572.
Boyle, J. (1994). An application of Fourier series to the most significant digit problem. The American Mathematical Monthly, 101(9), 879–886.
Hill, T. P. (1995). A statistical derivation of the significant-digit law. Statistical Science, 10(4), 354–363.
Hill, T. P. (1995). The significant-digit phenomenon. The American Mathematical Monthly, 102(4), 322–327.
Nigrini, M. J. (1996). A taxpayer compliance application of Benford’s law. The Journal of the American Taxation Association, 18(1), 72.
Nigrini, M. J., & Mittermaier, L. J. (1997). The use of Benford’s law as an aid in analytical procedures. Auditing-a Journal of Practice & Theory, 16(2), 52–67.
Jiang, L. C., Bazarova, N. N., & Hancock, J. T. (2013). From perception to behavior: Disclosure reciprocity and the intensification of intimacy in computer-mediated communication. Communication Research, 40(1), 125–143.
Cao, Z., Hui, K.-L., & Xu, H. (2018). When discounts hurt sales: The case of daily-deal markets. Information Systems Research, 29(3), 567–591.
Xiao, S., Tan, X., Dong, M., & Qi, J. (2014). How to design your project in the online crowdfunding market? Evidence from Kickstarter. In Proceedings of the 35th International Conference on Information Systems.
Herbig, P. A., & Kramer, H. (1994). The effect of information overload on the innovation choice process. Journal of Consumer Marketing., 11(2), 45–54.
Nigrini, M. J. (2012). Benford’s law: Applications for forensic accounting, auditing and fraud detection (Vol. 586). Hoboken: Wiley.
Giles, D. E. (2007). Benford’s law and naturally occurring prices in certain eBay auctions. Applied Economics Letters, 14(3), 157–161.
Nelson, P. (1970). Information and consumer behavior. Journal of Political Economy, 78(2), 311–329.
Nelson, P. (1974). Advertising as information. Journal of Political Economy, 82(4), 729–754.
Pu, J., Han, S., Kwark, Y., Gu, B., & Ye, Q. (2017). The double-edged sword of expert reviewer programs: The effects of offering expert reviewer status on review generation. In Proceedings of the 38th International Conference on Information Systems.
Hong, Y., Huang, N., Burtch, G., & Li, C. (2016). Culture, conformity, and emotional suppression in online reviews. Journal of the Association for Information Systems, 17(11), 737–758.
Koh, N. S., Hu, N., & Clemons, E. K. (2010). Do online reviews reflect a product’s true perceived quality? An investigation of online movie reviews across cultures. Electronic Commerce Research and Applications, 9(5), 374–385.
Wang, Y., Wang, Z., Zhang, D., & Zhang, R. (2019). Discovering cultural differences in online consumer product reviews. Journal of Electronic Commerce Research, 20(3), 169–183.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
On behalf of all authors, the corresponding author states that there is no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1
The following tables show the types, impacts, and detection methods of online review manipulation.
See Table
9. Table 9 summarizes different types of review manipulation attempts and discusses their impacts.
See Table
10. Table 10 presents a summary of the existing methods for review manipulation detection.
Appendix 2
The following figures show the distributions of review ratings and word counts in the four datasets.
See Fig.
5.
See Fig.
6.
Appendix 3
Tables report test results using character counts to measure review length.
See Table 12. Table 12 reports Benford’s law tests for high-quality and low-quality reviews on Amazon, corresponding to Table 6.
See Table 13. Table 13 reports review manipulation in different categories, corresponding to Table 7.
See Table 14. Table 14 reports the interaction between extremity and valence results, corresponding to Table 8.
See Table 11. Table 11 reports review manipulation on different sites, corresponding to Table 5.
Appendix 4
Tables report all four statistical tests for Benford’s law violation.
See Table 15. Table 15 reports review manipulation in different categories, corresponding to Table 7.
See Table 16. Table 16 reports the interaction between extremity and valence results, corresponding to Table 8.
See Table 17. Table 17 reports review manipulation in different categories using character counts as the review length measure, corresponding to Table 13.
See Table 18. Table 18 reports the interaction between extremity and valence results using character counts as the review length measure, corresponding to Table 14.
Appendix 5
The table reports the result using an alternative measure of review negativity.
See Table 19. Table 1919 reports review manipulation in different categories using an alternative measure of review negativity, corresponding to Table 7.
Appendix 6
The table reports the result of controlling for the same sample size among different sub-categories.
See Table 20. Table 20 reports review manipulation in different categories using the same sample size, corresponding to Table 7.
Rights and permissions
About this article
Cite this article
Zhao, C., Wang, C.A. A cross-site comparison of online review manipulation using Benford’s law. Electron Commer Res 23, 365–406 (2023). https://doi.org/10.1007/s10660-020-09455-8
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10660-020-09455-8