A regression tree approach using mathematical programming
Technology
Science & Technology
330
Operations Research & Management Science
PREDICTING SOIL PROPERTIES
Mathematical programming
PERFORMANCE
Regression tree
01 natural sciences
CLASSIFICATION
Surrogate model
004
Computer Science Applications
Engineering
Artificial Intelligence
SURROGATE MODELS
Computer Science
Electrical & Electronic
Optimisation
0101 mathematics
Regression analysis
Engineering(all)
DOI:
10.1016/j.eswa.2017.02.013
Publication Date:
2017-02-09T10:01:38Z
AUTHORS (4)
ABSTRACT
This work proposes a novel tree model for multivariate regression analysis.Both node splitting and regression coefficients are optimised in this model.The proposed method achieves improved prediction accuracy than literature methods.The resultant trees are at least as simple as the ones from previous methods. Regression analysis is a machine learning approach that aims to accurately predict the value of continuous output variables from certain independent input variables, via automatic estimation of their latent relationship from data. Tree-based regression models are popular in literature due to their flexibility to model higher order non-linearity and great interpretability. Conventionally, regression tree models are trained in a two-stage procedure, i.e. recursive binary partitioning is employed to produce a tree structure, followed by a pruning process of removing insignificant leaves, with the possibility of assigning multivariate functions to terminal leaves to improve generalisation. This work introduces a novel methodology of node partitioning which, in a single optimisation model, simultaneously performs the two tasks of identifying the break-point of a binary split and assignment of multivariate functions to either leaf, thus leading to an efficient regression tree model. Using six real world benchmark problems, we demonstrate that the proposed method consistently outperforms a number of state-of-the-art regression tree models and methods based on other techniques, with an average improvement of 760% on the mean absolute errors (MAE) of the predictions.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (46)
CITATIONS (104)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....