Skip to main content

A Click Prediction Model Based on Residual Unit with Inception Module

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 11671))

Abstract

The explosion in online advertisement urges to better estimate the click prediction of ads. For click prediction on single ad impression, we have access to pairwise relevance among elements in an impression, but not to global interaction among key features of elements. Moreover, the existing method on sequential click prediction treats propagation unchangeable for different time intervals. In this work, we propose a novel model, Convolutional Click Prediction Model (RES-IN), based on residual unit with inception module. RES-IN can extract local-global key feature interactions from an input instance with varied elements, which can be implemented for not only single ad impression but also sequential ad impression. Experiment results on three public large-scale datasets indicate that RES-IN is effective on click prediction.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Chen, Y., Shu, L., Wang, L.: Poster abstract: traffic flow prediction with big data: a deep learning based time series model. In: Computer Communications Workshops, pp. 1010–1011. IEEE (2017)

    Google Scholar 

  2. Abadi, M., Agarwal, A., Barham, P., et al.: TensorFlow: large-scale machine learning on heterogeneous distributed systems (2016)

    Google Scholar 

  3. Cucchiara, A.: Applied logistic regression. Technometrics 34(3), 2 (2012)

    Google Scholar 

  4. Rendle S.: Factorization machines. In: IEEE International Conference on Data Mining (2011)

    Google Scholar 

  5. Yu, D., Eversole, A., Seltzer, M., et al.: An introduction to computational networks and the computational network toolkit. Microsoft Research (2014)

    Google Scholar 

  6. Chen, K., Huo, Q.: Scalable training of deep learning machines by incremental block training with intra-block parallel optimization and blockwise model-update filtering. In: IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 5880–5884. IEEE (2016)

    Google Scholar 

  7. Collobert, R., Weston, J., Bottou, L., Karlen, M., Kavukcuoglu, K., Kuksa, P.: Natural language processing (almost) from scratch. J. Mach. Learn. Res. 12, 2493–2537 (2011)

    MATH  Google Scholar 

  8. Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. In: International Conference on Neural Information Processing Systems, pp. 1097–1105. Curran Associates Inc. (2012)

    Google Scholar 

  9. Huang, P.S., He, X., Gao, J., et al.: Learning deep structured semantic models for web search using click through data. In: ACM International Conference on Conference on Information & Knowledge Management, pp. 2333–2338. ACM (2013)

    Google Scholar 

  10. Schmidhuber, J.: Deep learning in neural networks: an overview. Neural Netw. 61, 85–117 (2015)

    Article  Google Scholar 

  11. He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. arXiv preprint arXiv:1512.03385 (2015)

  12. Arora, S., Bhaskara, A., Ge, R., et al.: Provable bounds for learning some deep representations, pp. 584–592 (2013)

    Google Scholar 

  13. Seide, F., Li, G., Chen, X., et al.: Feature engineering in Context-Dependent Deep Neural Networks for conversational speech transcription. In: Automatic Speech Recognition and Understanding, pp. 24–29. IEEE (2011)

    Google Scholar 

  14. Jia, Y., Shelhamer, E. et al.: Caffe: convolutional architecture for fast feature embedding, pp. 675–678 (2014)

    Google Scholar 

  15. Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8689, pp. 818–833. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10590-1_53

    Chapter  Google Scholar 

  16. Srivastava, R.K., Greff, K., Schmidhuber, J.: Training very deep networks. Comput. Sci. (2015)

    Google Scholar 

Download references

Acknowledgements

This work was supported by Postgraduate Research and Practice Innovation Program of Jiangsu Province (KYCX18_2511).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhiwen Ni .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Ni, Z., Ma, X., Sun, X., Bian, L. (2019). A Click Prediction Model Based on Residual Unit with Inception Module. In: Nayak, A., Sharma, A. (eds) PRICAI 2019: Trends in Artificial Intelligence. PRICAI 2019. Lecture Notes in Computer Science(), vol 11671. Springer, Cham. https://doi.org/10.1007/978-3-030-29911-8_30

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-29911-8_30

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-29910-1

  • Online ISBN: 978-3-030-29911-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics