Skip to main content

Detecting Professional Spam Reviewers

  • Conference paper
Advanced Data Mining and Applications (ADMA 2013)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 8347))

Included in the following conference series:

Abstract

Spam reviewers are becoming more professional. The common approach in spam reviewer detection is mainly based on the similarities among reviews or ratings on the same products. Applying this approach to professional spammer detection has some difficulties. First, some of the review systems start to set some limitations, e.g., duplicate submissions from a same id on one product are forbidden. Second, the professional spammers also greatly improve their writing skills. They are consciously trying to use diverse expressions in reviews. In this paper, we present a novel model for detecting professional spam reviewers, which combines posting frequency and text sentiment strength by analyzing the writing and behavior styles. Specifically, we first introduce an approach for counting posting frequency based on a sliding window. We then evaluate the sentiment strength by calculating the sentimental words in the text. Finally, we present a linear combination model. Experimental results on a real dataset from Dianping.com demonstrate the effectiveness of the proposed method.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Cohen, J.: A coefficient of agreement for nominal scales. Educational and Psychological Measurement 20(1), 37–46 (1960)

    Article  Google Scholar 

  2. Cohen, J.: Weighted kappa: Nominal scale agreement with provision for scaled disagreement or partial credit. Psychological Bulletin 70(4), 213–220 (1968)

    Article  Google Scholar 

  3. Dong, Z., Dong, Q.: http://www.keenage.com/html/c_index.html

  4. Eason, G., Noble, B., Sneddon, I.N.: On certain integrals of Lipschitz-Hankel type involving products of Bessel functions. Phil. Trans. Roy. Soc. London A247, 529–551 (1955)

    Article  MathSciNet  Google Scholar 

  5. Feng, S., Xing, L., Gogar, A., Choi, Y.: Distributional Footprints of Deceptive Product Reviews. In: ICWSM (2012)

    Google Scholar 

  6. Gilbert, E., Karahalios, K.: Understanding Deja Reviewers. In: Proc. of ACM CSCW, pp. 225–228. ACM, New York (2010)

    Google Scholar 

  7. Hu, M., Liu, B.: Mining and summarizing customer reviews. In: Proc. of KDD, pp. 168–177 (2004)

    Google Scholar 

  8. Jindal, N., Liu, B.: Review spam detection. In: Proc. of WWW (Poster), pp. 1189–1190. ACM (2007)

    Google Scholar 

  9. Jindal, N., Liu, B.: Opinion spam and analysis. In: Proc. of WSDM, pp. 219–230. ACM (2008)

    Google Scholar 

  10. Jindal, N., Liu, B., Lim, E.-P.: Finding Unusual Review Patterns Using Unexpected Rules. In: Proc. of CIKM (2010)

    Google Scholar 

  11. Järvelin, K., Kekäläinen, J.: IR evaluation methods for retrieving highly relevant documents. In: Proc. of SIGIR, pp. 41–48. ACM, New York (2000)

    Google Scholar 

  12. Landis, J.R., Koch, G.G.: The measurement of observer agreement for categorical data. Biometrics 33(1), 159–174 (1977)

    Article  MATH  MathSciNet  Google Scholar 

  13. Li, F., Huang, M., Yang, Y., Zhu, X.: Learning to Identify Review Spam. In: Proc. of IJCAI, pp. 2488–2493 (2011)

    Google Scholar 

  14. Lim, E.P., Nguyen, V.A., Jindal, N., et al.: Detecting Product Review Spammers Using Rating Behaviors. In: Proc. of the 19th CIKM, pp. 939–948. ACM, New York (2010)

    Google Scholar 

  15. Mukherjee, A., Liu, B., Glance, N.: Spotting Fake Reviewer Groups in Consumer Reviews. In: Proc. of WWW, pp. 191–200 (2012)

    Google Scholar 

  16. Ott, M., Cardie, C., Hancock, J.: Estimating the prevalence of deception in online review communities. In: Proc. of WWW (2012)

    Google Scholar 

  17. Ott, M., Choi, Y., Cardie, C., Hancock, J.T.: Finding Deceptive Opinion Spam by Any Stretch of the Imagination. In: Proc. of ACL, pp. 309–319 (2011)

    Google Scholar 

  18. Wang, G., Xie, S., Liu, B., Yu, P.S.: Review Graph based Online Store Review Spammer Detection. In: Proc. of ICDM (2011)

    Google Scholar 

  19. Xie, S., Wang, G., Lin, S., Yu, P.S.: Review spam detection via temporal pattern discovery. In: Proc. of KDD (2012)

    Google Scholar 

  20. Yoo, K.H., Gretzel, U.: Comparison of Deceptive and Truthful Travel Reviews. In: Information and Communication Technologies in Tourism, pp. 37–47 (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Huang, J., Qian, T., He, G., Zhong, M., Peng, Q. (2013). Detecting Professional Spam Reviewers. In: Motoda, H., Wu, Z., Cao, L., Zaiane, O., Yao, M., Wang, W. (eds) Advanced Data Mining and Applications. ADMA 2013. Lecture Notes in Computer Science(), vol 8347. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-53917-6_26

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-53917-6_26

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-53916-9

  • Online ISBN: 978-3-642-53917-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics