Abstract
Most real-world events relate to historical cases scholastically and chronologically. Some increase the probability of the next event in the future, while some do not. Many approaches leverage temporal point processes with explicitly defined probability intensity functions concerning time t and history \({\mathscr{H}}\), such as Poisson Process and Hawkes Process, to measure the time-varying probability. However, fixed-form intensity functions can limit the performance of temporal point process models, owing to the lack of prior knowledge of required intensity functions and the complexity of the real world. Neural networks’ ability to approximate functions makes the neural temporal point process a promising approach compared with the traditional temporal point process. In the paper, we identify several drawbacks the previous works have in meeting the mathematical constraints before designing a matrix-multiplication-based model to tackle these drawbacks. The experimental results reveal that our model is superior to other neural temporal point processes with better mathematical interpretability and extrapolation capability.
Similar content being viewed by others
References
Du N, Dai H, Trivedi R, Upadhyay U, Gomez-Rodriguez M, Song L (2016) Recurrent marked temporal point processes. In: Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining. ACM, New York, USA, pp 1555–1564
Mei H, Eisner J (2017) The neural hawkes process: A neurally self-modulating multivariate point process. In: Advances in neural information processing systems 30: Annual conference on neural information processing systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp 6754–6764
Zuo S, Jiang H, Li Z, Zhao T, Zha H (2020) Transformer Hawkes process. In: Proceedings of the 37th international conference on machine learning, Proceedings of machine learning research. PMLR, vol 119, pp 11692–11702
Li S, Xiao S, Zhu S, Du N, Xie Y, Song L (2018) Learning temporal point processes via reinforcement learning. In: Advances in neural information processing systems 31: Annual conference on neural information processing systems 2018, NeurIPS 2018, December 3-8, 2018, Montréal Canada, pp 10804–10814
Guo R, Li J, Liu H (2018) INITIATOR: noise-contrastive estimation for marked temporal point process. In: Proceedings of the twenty-seventh international joint conference on artificial intelligence, IJCAI 2018, July 13-19, 2018. ijcai.org, Stockholm, Sweden, pp 2191–2197
Okawa M, Iwata T, Kurashima T, Tanaka Y, Toda H, Ueda N (2019) Deep mixture point processes: Spatio-temporal event prediction with rich contextual information. In: Proceedings of the 25th ACM SIGKDD international conference on knowledge discovery & data mining, KDD 2019, August 4-8, 2019. ACM, Anchorage, AK, USA, pp 373–383
Xiao S, Yan J, Farajtabar M, Song L, Yang X, Zha H (2019) Learning time series associated event sequences with recurrent point process networks. IEEE Trans Neural Netw Learn Syst 30(10):3124–3136
Omi T, Ueda N, Aihara K (2019) Fully neural network based model for general temporal point processes. In: Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp 2120–2129
Shchur O, Bilos M, Günnemann S (2020) Intensity-free learning of temporal point processes. In: 8th international conference on learning representations, ICLR 2020, April 26-30, 2020. OpenReview.net, Addis Ababa, Ethiopia
Enguehard J, Busbridge D, Bozson A, Woodcock C, Hammerla N (2020) Neural temporal point processes for modelling electronic health records. In: Proceedings of the machine learning for health NeurIPS workshop, Proceedings of machine learning research. PMLR, vol 136, pp 85–113
Mei H, Wan T, Eisner J (2020) Noise-contrastive estimation for multivariate point processes. In: Advances in neural information processing systems, vol 33. Curran Associates Inc, pp 5204– 5214
Sütfeld LR, Brieger F, Finger H, Füllhase S, Pipa G (2020) Adaptive blending units: trainable activation functions for deep neural networks. In: Intelligent computing - proceedings of the 2020 computing conference, volume 3, advances in intelligent systems and computing, vol 1230. Springer, pp 37–50
Rubanova Y, Chen TQ, Duvenaud D (2019) Latent ordinary differential equations for irregularly-sampled time series. In: Advances in neural information processing systems 32: annual conference on neural information processing systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC Canada, pp 5321–5331
Jia J, Benson AR (2019) Neural jump stochastic differential equations. In: Advances in neural information processing systems 32: annual conference on neural information processing systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC Canada, pp 9843–9854
Chen RTQ, Amos B, Nickel M (2021) Learning neural event functions for ordinary differential equations. In: 9th international conference on learning representations, ICLR 2021, virtual only, May 3–7, 2021
Chen RTQ, Amos B, Nickel M (2021) Neural spatio-temporal point processes. In: 9th international conference on learning representations, ICLR 2021, virtual only, May 3–7, 2021
Montavon G, Samek W, Müller KR (2018) Methods for interpreting and understanding deep neural networks. Digit Sig Process 73:1–15
Jacot A, Hongler C, Gabriel F (2018) Neural tangent kernel: Convergence and generalization in neural networks. In: Advances in neural information processing systems 31: annual conference on neural information processing systems 2018, NeurIPS 2018, December 3-8, 2018, Montréal, Canada, pp 8580–8589
Chen M, Jiang H, Liao W, Zhao T (2019) Efficient approximation of deep relu networks for functions on low dimensional manifolds. In: Advances in neural information processing systems 32: annual conference on neural information processing systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp 8172–8182
Xu K, Zhang M, Li J, Du SS, Kawarabayashi KI, Jegelka S (2021) How neural networks extrapolate: From feedforward to graph neural networks. In: 9th international conference on learning representations, ICLR 2021, virtual only, May 3–7, 2021
Nguyen T, Raghu M, Kornblith S (2021) Do wide and deep networks learn the same things? Uncovering how neural network representations vary with width and depth. In: 9th international conference on learning representations, ICLR 2021, virtual only, May, 3–7, 2021
Shokri R, Stronati M, Song C, Shmatikov V (2017) Membership inference attacks against machine learning models. In: 2017 IEEE symposium on security and privacy (SP), pp 3–18
Yeom S, Giacomelli I, Fredrikson M, Jha S (2018) Privacy risk in machine learning: Analyzing the connection to overfitting. In: 2018 IEEE 31st computer security foundations symposium (CSF), pp 268–282
Ramsauer H, Schäfl B, Lehner J, Seidl P, Widrich M, Gruber L, Holzleitner M, Adler T, Kreil D, Kopp MK, Klambauer G, Brandstetter J, Hochreiter S (2021) Hopfield networks is all you need. In: 9th international conference on learning representations, ICLR 2021, virtual only May 3–7, 2021
Mildenhall B, Srinivasan PP, Tancik M, Barron JT, Ramamoorthi R, Ng R (2020) NeRF: Representing scenes as neural radiance fields for view synthesis. In: Computer vision - ECCV 2020 - 16th european conference, Glasgow, UK, August 23-28, 2020, Proceedings, Part I, vol 12346, pp 405–421
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems 30: Annual conference on neural information processing systems 2017, December 4-9, 2017, Long Beach, CA USA, pp 5998–6008
Devlin J, Chang MW, Lee K, Toutanova K (2019) BERT: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 conference of the north american chapter of the association for computational linguistics: Human language technologies, vol 1 (Long and Short Papers), pp 4171–4186
Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, Uszkoreit J, Houlsby N (2021) An image is worth 16x16 words: transformers for image recognition at scale. In: 9th international conference on learning representations, ICLR 2021, Virtual Only May 3–7, 2021
Sill J (1998) Monotonic networks. In: Advances in neural information processing systems, vol 10, pp 661–667
You S, Ding D, Canini KR, Pfeifer J, Gupta MR (2017) Deep lattice networks and partial monotonic functions. In: Advances in neural information processing systems 30: Annual conference on neural information processing systems 2017, December 4-9, 2017, Long Beach, CA, USA, pp 2981–2989
Wehenkel A, Louppe G (2019) Unconstrained monotonic neural networks. In: Advances in neural information processing systems 32: Annual conference on neural information processing systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pp 1543–1553
Zhuang T, Zhang Z, Huang Y, Zeng X, Shuang K, Li X (2020) Neuron-level structured pruning using polarization regularizer. Adv Neural Inf Process Syst, vol 33
Mescheder LM, Geiger A, Nowozin S (2018) Which training methods for gans do actually converge?. In: Proceedings of the 35th international conference on machine learning, ICML 2018, Stockholmsmässan, July 10-15, 2018, Proceedings of machine learning research. PMLR, vol 80, Stockholm, Sweden, pp 3478–3487
Loshchilov I, Hutter F (2019) Decoupled weight decay regularization. In: 7th international conference on learning representations, ICLR 2019, May 6–9, 2019, New Orleans, LA, USA
Lobov SA, Mikhaylov AN, Shamshin M, Makarov VA, Kazantsev VB (2020) Spatial properties of stdp in a self-learning spiking neural network enable controlling a mobile robot. Front. Neurosci. 14:88
Naveros F, Luque NR, Ros E, Arleo A (2020) Vor adaptation on a humanoid icub robot using a spiking cerebellar model. IEEE Trans Cybern 50(11):4744–4757
Tang G, Shah A, Michmizos KP (2019) Spiking neural network on neuromorphic hardware for energy-efficient unidimensional slam. In: 2019 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp 4176–4181
Mozafari M, Kheradpisheh SR, Masquelier T, Nowzari-Dalini A, Ganjtabesh M (2018) First-spike-based visual categorization using reward-modulated stdp. IEEE Trans Neural Netw Learn Syst 29(12):6178–6190
Zhu L, Dong S, Li J, Huang T, Tian Y (2020) Retina-like visual image reconstruction via spiking neural model. In: 2020 IEEE/CVF conference on computer vision and pattern recognition (CVPR), pp 1435–1443
Yu Q, Yan R, Tang H, Tan KC, Li H (2016) A spiking neural network system for robust sequence recognition. IEEE Trans Neural Netw Learn Syst 27:621–635
Song Z, Xiang S, Ren Z, Han G, Hao Y (2020) Spike sequence learning in a photonic spiking neural network consisting of vcsels-sa with supervised training. IEEE J Sel Top Quant Electron 26(5):1–9
Maciag PS, Kryszkiewicz M, Bembenik RL, Lobo J, Del Ser J (2021) Unsupervised anomaly detection in stream data with online evolving spiking neural networks. Neural Netw 139:118–139
Zhang M, Qu H, Belatreche A, Chen Y, Yi Z (2019) A highly effective and robust membrane potential-driven supervised learning method for spiking neurons. IEEE Trans Neural Netw Learn Syst 30(1):123–137
Wang X, Lin X, Dang X (2020) Supervised learning in spiking neural networks: A review of algorithms and evaluations. Neural Netw 125:258–280
Yang S, Gao T, Wang J, Deng B, Lansdell B, Linares-Barranco B (2021) Efficient spikedriven learning with dendritic event-based processing. Front Neurosci, vol 15. Publisher: Frontiers
Huh D, Sejnowski TJ (2018) Gradient descent for spiking neural networks. In: Advances in neural information processing systems 31: Annual conference on neural information processing systems 2018, NeurIPS 2018, December 3-8, 2018, Montréal, Canada, pp 1440–1450
Yang S, Wang J, Zhang N, Deng B, Pang Y, Azghadi MR (2021) CerebelluMorphic: large-scale neuromorphic model and architecture for supervised motor learning. IEEE Trans Neural Netw Learn Syst, pp 1–15
Yang S, Deng B, Wang J, Liu C, Li H, Lin Q, Fietkiewicz C, Loparo KA (2019) Design of hidden-property-based variable universe fuzzy control for movement disorders and its efficient reconfigurable implementation. IEEE Trans Fuzzy Syst 27(2):304–318
Yang S, Deng B, Wang J, Li H, Lu M, Che Y, Wei X, Loparo KA (2020) Scalable digital neuromorphic architecture for large-scale biophysically meaningful neural network with multi-compartment neurons. IEEE Trans Neural Netw Learn Syst 31(1):148–162
Zhao Q, Erdogdu MA, He HY, Rajaraman A, Leskovec J (2015) SEISMIC: A self-exciting point process model for predicting tweet popularity. In: Proceedings of the 21th ACM SIGKDD international conference on knowledge discovery and data mining, August 10-13, 2015. ACM, Sydney, NSW, Australia, pp 1513–1522
Acknowledgements
This work is funded by the National Natural Science Foundation of China (ID:61877051) and the National Science Foundation Project of CQ (ID: cstc2018jscx-msyb1042 and cstc2018jscx-msybX0273). We thank Senlin Shu and Xiaofei Xu for all the inspiring discussions about the design of the dynamic weight layer. Additionally, we acknowledge all the developers and researchers for developing useful machine learning tools that enable our experiments.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Rights and permissions
About this article
Cite this article
Liu, S., Li, L. Learning general temporal point processes based on dynamic weight generation. Appl Intell 52, 3678–3690 (2022). https://doi.org/10.1007/s10489-021-02590-1
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-02590-1