Abstract
Linear data regression is a fundamental mathematical tool in engineering and applied sciences. However, for complex nonlinear phenomena, the standard linear least-squares regression may prove ineffective, hence calling for more involved data modeling techniques. The current research work investigates in particular nonlinear statistical regression of bivariate data that do not exhibit a monotonic dependency. The current contribution proposes a neural-network-based data processing method, termed data monotonization, followed by neural isotonic statistical regression. Such data monotonization processing is performed by means of an adaptive neural network that learns its nonlinear transfer function from the training set. The artificial neural system that performs data monotonization is implemented through a lookup table (LUT), which entails few computationally inexpensive algebraic operations to adapt and to compute the output from the input data-set. A number of learning rules to adapt such LUT-based neural system are introduced and compared in order to elucidate their relative merits and drawbacks.






Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
If there exist multiple records with an identical \(x_i\)-value, low-valued random displacements may be added to the x-value of such records, in such a way that those records may be sorted.
References
Abdullah A, Hussain A. A cognitively inspired approach to two-way cluster extraction from one-way clustered data. Cognit Comput. 2014;7:161–82.
Ahuja S, Lakshminarayana A, Shukla SK. Statistical regression based power models. In: Low power design with high-level power estimation and power-aware synthesis. New York: Springer; 2012. p. 59 – 70
Angelov S, Harb B, Kannan S, Wang L-S. Weighted isotonic regression under the \(\text{ L }_1\) norm. In: Proceedings of the Seventeenth Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), 2006. p. 783 – 791.
Blanco JL, Hernández LA, Fernández R, Ramos D. Improving automatic detection of obstructive sleep apnea through nonlinear analysis of sustained speech. Cognit Comput. 2013;5:458–72.
Breiman L. Statistical modeling: the two cultures. Stat Sci. 2001;16(3):199–231.
Brunk HD. Maximum likelihood estimates of monotone parameters. Ann Math Stat. 1955;26:607–16.
Burnham KP, Anderson DR. Multimodel inference: understanding AIC and BIC in model selection. Sociol Methods Res. 2004;33(2):261–304.
Chakravarti N. Sensitivity analysis in isotonic regression. Discrete Appl Math. 1993;45(3):183–96.
Calzada Defez Á, Socoró Carrie JC. Voice quality modification using a harmonics plus noise model. Cognit Comput. 2013;5:473–82.
Domínguez-Menchero JS, González-Rodríguez G. Analyzing an extension of the isotonic regression problem. Metrika. 2007;66(1):19–30.
Eckerle K. Circular interference transmittance study. Report of the National Institute of Standards and Technology (NIST), US Department of Commerce, USA. Unpublished report, 1979. Data publicly available from the repository http://www.itl.nist.gov/div898/strd/nls/data/eckerle4.shtml.
Ezeiza A, López de Ipiña K, Hernández C, Barroso N. Enhancing the feature extraction process for automatic speech recognition with fractal dimensions. Cognit Comput. 2013;5:545–50.
Fiori S. Hybrid independent component analysis by adaptive LUT activation function neurons. Neural Netw. 2002;15(1):85–94.
Fiori S. Statistical nonparametric bivariate isotonic regression by lookup-table-based neural networks. In: Proceedings of the 2011 International Conference on Neural Information Processing (ICONIP 2011, Shanghai (China), November 14–17, 2011), B.-L. Lu, L. Zhang, and J. Kwok (Eds.), Part III, LNCS 7064, Heidelberg: Springer; 2011. p. 365 – 372.
Fiori S. Fast statistical regression in presence of a dominant independent variable. Neural Computing and Applications. (Special issue of the 2011 International Conference on Neural Information Processing - ICONIP’2011), 2013; 22(7): 1367 – 1378.
Fiori S. An isotonic trivariate statistical regression method. Adv Data Anal Classif. 2013;7(2):209–35.
Fiori S. A two-dimensional Poisson equation formulation of non-parametric statistical non-linear modeling. Comput Math Appl. 2014;67(5):1171–85.
Forrest DR, Hetland RD, DiMarco SF. Multivariable statistical regression models of the areal extent of hypoxia over the Texas-Louisiana continental shelf. Environ Res Lett. 2011;6(4):045002 (10 pp).
Friedman J, Tibshirani R. The monotone smoothing of scatterplots. Technometrics. 1984;26(3):243–50.
Gómez-Vilda P, Rodellar-Biarge V, Nieto-Lluis V, Munöz-Mulas C, Mazaira-Fernández LM, Martínez-Olalla R, Álvarez-Marquina A, Ramírez-Calvo C, Fernández-Fernández M. Characterizing neurological disease from voice quality biomechanical analysis. Cognit Comput. 2013;5:399-42.
Hall P, Huang L-S. Nonparametric kernel regression subject to monotonicity constraints. Ann Stat. 2001;29(3):624–47.
Jaqaman K, Danuser G. Linking data to models: data regression. Nat Rev Mol Cell Biol. 2006;7:813–9.
Kulkarni MA, Patil S, Rama GV, Sen PN. Wind speed prediction using statistical regression and neural network. J Earth Syst Sci. 2008;117(4):457–63.
Li X, Liu H-Z. Statistical regression for efficient high-dimensional modeling of analog and mixed-signal performance variations. In: Proceedings of the \(45^{\rm th}\) ACM/IEEE Design Automation Conference (DAC 2008, Anaheim Convention Center, California, USA, June 9–13, 2008), p. 38 – 43, June 2008.
Liu S, Gao RX, He Q, Staudenmayer J, Freedson P. Development of statistical regression models for ventilation estimation. In: Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC, September 3–6, 2009), 2009. p. 1266 – 1269.
Liu J, Li H. Application research of a statistical regression algorithm in the IVR system. In: Proceedings of the 2010 International Conference on Educational and Network Technology (ICENT, Qinhuangdao (China), June 25–27, 2010), 2010. p. 358 – 360.
Naji M, Firoozabadi M, Azadfallah P. Classification of music-induced emotions based on information fusion of forehead biosignals and electrocardiogram. Cognit Comput. 2014;6:241–52.
Papoulis A. Probability and statistics. Upper Saddle River: Prentice Hall; 1996.
Punera K, Gosh J. Enhanced hierarchical classification via isotonic smoothing. In: Proceedings of the \(17^{\rm th}\) International Conference on World Wide Web (April 21–25, 2008, Beijing–China), 2008. p. 151–160.
Ramsay JO. Monotone regression splines in action. Stat Sci. 1988;3(4):425–41.
Siegert I, Philippou-Hübner D, Hartmann K, Böck R, Wendemuth A. Investigation of speaker group-dependent modelling for recognition of affective states from speech. Cognit Comput. 2014;6:892–913.
Specht DE. A general regression neural network. IEEE Trans Neural Netw. 1991;2(6):568–76.
Tam DN. Computation in emotional processing: quantitative confirmation of proportionality hypothesis for angry unhappy emotional intensity to perceived loss. Cognit Comput. 2011;3:394–415.
Tran TP, Ong E, Hodges AP, Paternostro G, Piermarocchi C. Prediction of kinase inhibitor response using activity profiling, in vitro screening, and elastic net regression. BMC Syst Biol. 2014;8(74):1–10.
Yan X. Dissociated emergent response system and fine-processing system in human neural network and a heuristic neural architecture for autonomous humanoid robots. Cognit Comput. 2011;3:367–73.
Zhu X, Welling M, Jin F, Lowengrub J. Predicting simulation parameters of biological systems using a Gaussian process model. Stat Anal Data Min: ASA Data Sci J (Special Issue: “Best Papers from the SLDM Competition”). 2012;5(6):509 – 522.
Zou L, Xu S, Ma Z, Lu J, Su W. Automatic removal of artifacts from attention deficit hyperactivity disorder electroencephalograms based on independent component analysis. Cognit Comput. 2013;5:225–33.
Author information
Authors and Affiliations
Corresponding author
Appendix: Bivariate Isotonic Regression Method
Appendix: Bivariate Isotonic Regression Method
The regression model to infer has a nonlinear structure described by relationship \(y=f(x)\), where \(x\in {\mathcal {X}}\subseteq {{{\mathcal {R}}}}\) denotes the input random variable, having probability density function \(p_x(\cdot )\), and \(y\in {\mathcal {Y}}\subseteq {{{\mathcal {R}}}}\) denotes the output random variable, having probability density function \(p_y(\cdot )\).
With the assumption that the regression model be strictly monotonic, namely \(f'(x)>0\) or \(f'(x)<0\), \(\forall x\in {\mathcal {X}}\), the input–output distributions and the regression model may be shown to stay in the relationship:
-
Positive-slope regression model: \(f(x)=P_y^{-1}(P_x(x))\),
-
Negative-slope regression model: \(f(x)=P_y^{-1}(1-P_x(x))\),
where the symbol \(P_y^{-1}(\cdot )\) denotes the inverse of the cumulative distribution function \(P_y(\cdot )\) pertaining to the y random variable and the symbol \(P_x(\cdot )\) denotes the cumulative distribution function pertaining to the x random variable. In effect, the above transformations make sure that the distribution \(p_x(\cdot )\) is transformed into the distribution \(p_y(\cdot )\) according to the law of measure invariance of probability density functions [28].
The above regression method does not depend explicitly on the data, but on the probability density/cumulative functions obtained from each data-set separately.
To develop a fully numerical bivariate isotonic statistical regression method, the following ingredients are of use: a suitable numerical estimation method for cumulative density functions and a suitable format for function representation/handling (with particular emphasis on numerical function inversion). In order to put the above regression equations into effect, a representation of the quantities of interest based on lookup tables was chosen.
A real-valued lookup table with N entries is represented by a pair \({\mathrm {LUT}}=({x},{y})\), where \({x}\in {{{\mathcal {R}}}}^{N}\) and \({y}\in {{{\mathcal {R}}}}^{N}\). The entries \(x_k\) of vector \({x}\) and the entries \(y_k\) of vector \({y}\), with \(k\in \{1,\ \ldots ,\ N\}\), are paired and provide a point-wise description of an arbitrarily shaped function. In order to handle the lookup tables for statistical regression purpose, the following operations are of use:
-
Histogram computation: In order to numerically approximate the probability distribution of a data-set \({\mathcal {D}}\), a histogram operator is made use of. The histogram-computation operation is denoted by \(({x},{y})=\mathrm {hist}({\mathcal {D}})\). The constructed lookup table is built up as follows: \(x_k\) equals the value of the k th bin center, and \(y_k\) equals the number of data points falling in the k th bin.
-
Cumulative-sum computation: On the basis of a lookup table \(({x},{y})\), a new lookup table \(({x},{v})=\mathrm {csum}({x},{y})\) is constructed, where array \({v}\) contains the cumulative sum of the entries of array \({y}\), possibly normalized in order to approximate the numerical integration of the function represented by the lookup table \(({x},{y})\). In its unnormalized version, the cumulative sum is described by \(v_1=y_1\) and \(v_k=v_{k-1}+y_k\) for \(2\le k\le N\).
-
Function/table interpolation: Interpolation may be invoked to make computations with lookup tables on other points in the domain. In the present context, it is necessary to preserve the monotonicity of an approximated function; therefore, linear interpolation only is made use of. Denote by \({\mathcal {D}}\) the x-coordinate point-set, where the function represented by a lookup table \(({x},{y})\) needs to be interpolated. The interpolation operation may be denoted by \({\mathcal {I}}=\mathrm {interp}({x},{y},{\mathcal {D}})\), where the set \({\mathcal {I}}\) contains the interpolated y-values corresponding to the x-values in the set \({\mathcal {D}}\). Because of the hypotheses of monotonicity of the model underlying the data, a set-type representation is equivalent to an ordered-list representation.
-
Function/table inversion: If a function is given a point-wise representation by the help of a lookup table \(({x},{y})\), then its inverse function may be easily given a point-wise representation by the lookup table \(({y},{x})\), namely function inversion is equivalent to swapping lookup table’s arguments. Apparently, therefore, function inversion in the context of lookup table representation is a computationally costless operation.
The bivariate isotonic regression method proposed in [15] consists of the following steps.
First, it is necessary to estimate the probability density functions of data within \({\mathcal {X}}\) and \({\mathcal {Y}}\) data-sets. They may be numerically estimated—up to scale factors—by:
The numerical cumulative distribution functions of the \({\mathcal {X}}\) and \({\mathcal {Y}}\) data-sets are estimated by numerical integration of the numerical probability density functions, which is achieved by the help of the cumulative-sum operator applied to lookup tables \(({x},{p}_x)\) and \(({y},{p}_y)\):
If the statistical model has negative slope, the lookup table \(({x},{P}_x)\) should be replaced by \(({x},1-{P}_x)\) in what follows.
For regression purpose, the regression model needs to be evaluated on an ordered set of x-points denoted here as an array \(\hat{{x}}\). The array \(\hat{{x}}\) consists of R points equally spaced over the range of interest for the x-variable.
The last step consists in numerically evaluating the quantity \(P_x(\cdot )\) over the points in \(\hat{{x}}\) and then in evaluating the function \(P_y^{-1}(\cdot )\) over the values of \(P_x(\hat{{x}})\), namely:
In the second of equations (4.3), the inverse function \(P_y^{-1}(\cdot )\) appears through the swapped lookup table of function \(P_y(\cdot )\).
The pair \((\hat{{x}},\hat{{y}})\) provides a lookup table representation for the nonlinear bivariate isotonic regression model \(f(\cdot )\).
Full details about the above-summarized statistical isotonic regression technique are available in [15].
Rights and permissions
About this article
Cite this article
Fiori, S., Gong, T. & Lee, H.K. Bivariate Nonisotonic Statistical Regression by a Lookup Table Neural System. Cogn Comput 7, 715–730 (2015). https://doi.org/10.1007/s12559-015-9353-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12559-015-9353-9