Skip to main content

A Heterogeneous Domain Adversarial Neural Network for Trans-Domain Behavioral Targeting

  • Conference paper
  • First Online:
Trends and Applications in Knowledge Discovery and Data Mining (PAKDD 2019)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 11607))

Included in the following conference series:

Abstract

To realize trans-domain behavioral targeting, which targets interested potential users of a source domain (e.g. E-Commerce) based on their behaviors in a target domain (e.g. Ad-Network), heterogeneous transfer learning (HeTL) is a promising technique for modeling behavior linkage between domains. It is required for HeTL to learn three functionalities: representation alignment, distribution alignment, and classification. In our previous work, we prototyped and evaluated two typical transfer learning algorithms, but neither of them jointly learns the three desired functionalities. Recent advances in transfer learning include a domain-adversarial neural network (DANN), which jointly learns distribution alignment and classification. In this paper, we extended DANN to be able to learn representation alignment by simply replacing its shared encoder with domain-specific types, so that it jointly learns the three desired functionalities. We evaluated the effectiveness of the joint learning of the three functionalities using real-world data of two domains: E-Commerce, which is set as the source domain, and Ad Network, which is set as the target domain.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. The promise of first-party data. Econsultancy. https://cdn2.hubspot.net/hubfs/370829/Campaigns_and_Emails/Archived_Emails/2015_Emails/2015_Econsultancy_Promise_of_First_Party_Data_June/The_Promise_of_First_Party_Data_Signal_Econsultancy_Report.pdf. Accessed 15 Oct 2018

  2. Shi, X., Liu, Q., Fan, W., Yang, Q., Yu, P.S.: Predictive modeling with heterogeneous sources. In: Proceedings of the 2010 SIAM International Conference on Data Mining, pp. 814–825 (2010)

    Google Scholar 

  3. Yamada, M., Suzuki, T., Kanamori, T., Hachiya, H., Sugiyama, M.: Relative density-ratio estimation for robust distribution comparison. In: Advances in Neural Information Processing Systems, vol. 24, pp. 594–602 (2011)

    Google Scholar 

  4. Ganin, Y., et al.: Domain-adversarial training of neural networks. J. Mach. Learn. Res. 17, 1–35 (2016)

    MathSciNet  MATH  Google Scholar 

  5. Kulis, B., Saenko, K., Darrell, T.: What you saw is not what you get: domain adaptation using asymmetric kernel transforms. In: Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, pp. 1785–1792 (2011)

    Google Scholar 

  6. Feuz, K.D., Cook, D.J.: Transfer learning across feature-rich heterogeneous feature spaces via feature-space remapping (FSR). ACM Trans. Intell. Syst. Technol. 6, 1–27 (2015)

    Article  Google Scholar 

  7. Shi, X., Liu, Q., Fan, W., Yu, P.S., Zhu, R.: Transfer learning on heterogenous feature spaces via spectral transformation. In: Proceedings of the 10th IEEE International Conference on Data Mining, pp. 1049–1054 (2010)

    Google Scholar 

  8. Duan, L., Tsang, I.W.: Learning with augmented features for heterogeneous domain. In: Proceedings of the 29th International Conference on Machine Learning (2012)

    Google Scholar 

  9. Li, W., Duan, L., Xu, D., Tsang, I.W.: Learning with augmented features for supervised and semi-supervised heterogeneous domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 36, 1134–1148 (2014)

    Article  Google Scholar 

  10. Chen, T., Guestrin, C.: Xgboost: a scalable tree boosting system. In: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data mining, pp. 785–794 (2016)

    Google Scholar 

  11. Rehurek, R., Sojka, P.: Software framework for topic modelling with large corpora. In: LREC 2010 Workshop on New Challenges for NLP Frameworks (2010)

    Google Scholar 

  12. Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. In: Proceedings of the International Conference on Learning Representations (2014)

    Google Scholar 

  13. Kurokawa, M., et al.: Virtual touch-point: trans-domain behavioral targeting via transfer learning. In: Workshop on Big Data Transfer Learning in conjunction with IEEE International Conference on Big Data (2018)

    Google Scholar 

Download references

Acknowledgment

This research was partially supported by JST CREST Grant Number J181401085, Japan.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kei Yonekawa .

Editor information

Editors and Affiliations

Appendix

Appendix

  • PM1: HeMap + HEGS + Xgboost

Modeling Behavior Linkage.

HeMap learns a common latent space from the source and target features. The optimization objective related to the common latent space is as follows (the same equation appears in [7]):

$$ \mathop{\min}\nolimits_{{{\text{B}}_{{\rm T}}^{\text{T}} {{\rm B}}_{\text{T}} = {{\rm I}}, {\text{B}}_{{\rm S}}^{\text{T}} {{\rm B}}_{\text{S}} = {\rm {I}}}} \left\| {{\text{X}}_{{\rm S}} - {\text{B}}_{{\rm S}} {\text{P}}_{{\rm S}} } \right\|^{2} + \left\| {{\text{X}}_{{\rm T}} - {\text{B}}_{{\rm T}} {\text{P}}_{{\rm T}} } \right\|^{2} + \beta \left( {\frac{1}{2}\left\| {{{\rm X}}_{\text{T}} - {{\rm B}}_{\text{S}} {{\rm P}}_{\text{T}} } \right\|^{2} + \frac{1}{2}\left\| {{{\rm X}}_{\text{S}} - {{\rm B}}_{\text{T}} {{\rm P}}_{\text{S}} } \right\|^{2} } \right) $$
(2)

where, \( {\text{B}}_{\text{S}} \), \( {\text{B}}_{\text{T}} \) are the projected source and target instances onto the common latent space respectively, and \( {\text{P}}_{\text{S}} \), \( {\text{P}}_{\text{T}} \) are the projection matrices from the common latent space onto the source and target space, respectively.

Then, HEGS selects the source instances in \( {\text{B}}_{\text{S}} \) similar to the target instances in \( {\text{B}}_{\text{T}} \). Since HEGS is originally a domain adaptation method for regression, we modified HEGS by hiring logistic regression to unify the labels in both domains. Finally, Xgboost learns the binary classification model \( F_{Xgb} \left( \cdot \right) \) using the selected instances and labels. Hyper-parameters of HeMap, HEGS, and Xgboost are tuned using cross validation. The resulting behavior linkage model \( {\mathbf{\mathcal{M}}} = \left( {{\text{P'}}_{\text{T}} ,F_{Xgb} } \right) \), where \( {\text{P'}}_{\text{T}} \) is a pseudo-inverse of the projection matrix \( {\text{P}}_{\text{T}} \) obtained by HeMap.

Applying Behavior Linkage Model.

We compute the predictive probability of conversion \( {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) = F_{Xgb} \left( {{\text{P'}}_{\text{T}} \varvec{ x}_{j}^{{\left( {\text{T}} \right)}} } \right) \) for each user \( j \) in the target domain. The resulting target users are defined by \( \left\{ {j; {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) > \theta } \right\} \) with an arbitrary threshold \( 0 < \theta < 1 \).

  • PM2: HFA

Modeling Behavior Linkage.

HFA learns a multiple kernel classifier \( F_{HFA} \left( \cdot \right) \) on a common latent space using the source and target features and the labels. The projection matrices \( {\text{P}}, {\text{Q}} \) from the source and target space onto the common latent space are coupled as \( {\text{H}} = \left[ {{\text{P}}, {\text{Q}}} \right] '\left[ {{\text{P}}, {\text{Q}}} \right] \), and then kernel matrices are computed and optimized based on \( {\text{H}} \). Hyper-parameters of HFA are tuned using cross validation. The resulting behavior linkage model \( {\mathbf{\mathcal{M}}} = \left( {F_{HFA} } \right) \).

Applying Behavior Linkage Model.

Similar to PM1, we compute the predictive probability of conversion \( {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) = F_{HFA} \left( {\varvec{ x}_{j}^{{\left( {\text{T}} \right)}} } \right) \) for each user \( j \) in the target domain. The resulting target users are defined by \( \left\{ {j; {\text{P}}\left( {y_{j} |\varvec{x}_{j}^{{\left( {\text{T}} \right)}} } \right) > \theta } \right\} \) with an arbitrary threshold \( 0 < \theta < 1 \).

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Yonekawa, K. et al. (2019). A Heterogeneous Domain Adversarial Neural Network for Trans-Domain Behavioral Targeting. In: U., L., Lauw, H. (eds) Trends and Applications in Knowledge Discovery and Data Mining. PAKDD 2019. Lecture Notes in Computer Science(), vol 11607. Springer, Cham. https://doi.org/10.1007/978-3-030-26142-9_24

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-26142-9_24

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-26141-2

  • Online ISBN: 978-3-030-26142-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics