pradachan's picture
Upload folder using huggingface_hub
f71c233 verified
raw
history blame
39.1 kB
# SYNTHETIC REDUCED NEAREST NEIGHBOR MODEL
## FOR REGRESSION
Anonymous authors
Paper under double-blind review
ABSTRACT
Nearest neighbor models are among the most established and accurate approaches
to machine learning. In this paper, we investigate Synthetic Reduced Nearest
Neighbor (SRNN) as a novel approach to regression tasks. Existing prototype
nearest neighbor models are initialized by training a k-means model over each
class. However, such initialization is only applicable to classification tasks. In this
work, we propose a novel initialization and expectation maximization approach
for enabling the application of SRNN to regression. The proposed initialization
approach is based on applying the k-means algorithm on the target responses of
samples to create various clusters of targets. This is proceeded by learning several
centroids in the input space for each cluster found over the targets. Essentially, the
initialization consists of finding target clusters and running k-means in the space
of feature vectors for the corresponding target cluster. The optimization procedure
consists of applying an expectation maximization approach similar to the k-means
algorithm that optimizes the centroids in the input space. This algorithm is comprised of two steps: (1) The assignment step, where assignments of the samples to
each centroid is found and the target response (i.e., prediction) of each centroid is
determined; and (2) the update/centroid step, where each centroid is updated such
that the loss function of the entire model is minimized. We will show that the centroid step operates over all samples via solving a weighted binary classification.
However, the centroid step is NP-hard and no surrogate objective function exists
for solving this problem. Therefore, a new surrogate is proposed to approximate
the solution for the centroid step. Furthermore, we consider the consistency of
the model, and show that the model is consistent under mild assumptions. The
bias-variance relationship in this model is also discussed. We report the empirical evaluation of the proposed SRNN regression model in comparison to several
state-of-the-art techniques.
1 INTRODUCTION
One of the main topics of research in Machine Learning is the relation between the features and output responses Hastie et al. (2009); Santosa & Symes (1986); Tibshirani (1996); Criminisi & Shotton
(2013). Synthetic Reduced Nearest Neighbor (SRNN) models are shown to be an effective tool in
determining the relationships between features of the inputs and the sub-clusters of each class in
supervised learning tasks Tavallali et al. (2020b). However, existing prototype nearest neighbor
models such as SRNN are constrained to classification problems, and to the best of our knowledge,
there remains a gap in extending these algorithms towards regression tasks. Such regression reduced
nearest neighbor models may find extensive applications in epidemiological studies Tavallali et al.
(2020a); Cisneros et al. (2021), medical studies Criminisi & Shotton (2013); Graf et al. (2011b;a),
and other applied regression tasks in general Tibshirani (1996). To address this gap in the state of
the art, we propose a novel algorithm for the optimization and construction of Regression Synthetic
Reduced Nearest Neighbor (Reg-SRNN) models.
The proposed Reg-SRNN is capable of discovering various modalities of the input data, and relates
those to the modalities of the output responses. The Reg-SRNN algorithm is designed to handle both
single-response and multi-response regression. The multi-response regression consists of learning
the relation between input samples and several ground-truth output responses. Reg-SRNN partitions
the input space into piecewise constant regions, where each region is represented by a centroid and
-----
its output response. From this perspective, Reg-SRNN is similar to other piecewise constant models,
such as Li & Martin (2017); Begon et al. (2017); Bertsimas & Dunn (2017); Tavallali et al. (2019;
2020c). Reg-SRNN is capable of learning an accurate relation between each cluster of the data and
its corresponding output responses. Therefore, Reg-SRNN can also provide enhanced interpretablity
by reducing the information content of clusters into a compressed representation manifested in their
centroids.
The technical contributions of this paper include the proposal of a novel initialization that by itself is
competitive to other existing regression models. This is proceeded by an expectation maximization
algorithm for directly minimizing the least squares error of the mode. The proposed optimization
algorithm is provably convergent, and it is shown that it monotonically decreases the loss function.
Therefore, the algorithm has a convergence guarantee on minimizing the loss function and achieving a local optimum.It is also worth mentioning that the algorithm does not cycle. The proposed
optimization algorithm consists of two steps and is inspired by K-means algorithm Lloyd (1982).
One step is the assignment step and is composed of finding samples assignments and proper output
response of the centroid. Second step is the update step where the centroid is optimized such that
the loss function is decreased. The centroid step is affected by all the samples and we will show
that this update step is a kind of NP-hard weighted binary classification problem. The update step
is computed through a surrogate objective function that is similar to SVM. We establish that the
algorithm is efficient because of its linear computational complexity. Finally, the model is evaluated
on various datasets with various sizes and dimensionalities, the results of which demonstrate that
Reg-SRNN is capable of competing and even over-matching similar regression models.
Accordingly, the main contributions of this paper are as follows:
- We propose a novel algorithm for initialization of SRNN models to extend their application
to regression tasks.
- We develop an optimization algorithm for regression SRNN models with guarantees on
convergence.
- Through experimental evaluation, we demonstrated the feasibility of our proposed regression SRNN model in filling the gap between more complex models (such as random forests)
and basic and interpretable models such as linear regression and decision trees.
2 RELATED WORK
A regression task consists of learning the relation between samples of the input space and a numerical output space. More specifically, regression is a supervised learning task of mapping inputs
(independent variable X) to the output Y, which is a continuous vector (Y ∈ R[d]). If the dimensionality of the output d ≥ 2, the task is known as multi-response regression. Regression has been
the workhorse of numerous fields Tai (2021), and various regression models have been developed
and expanded fundamentally over the recent decades Hastie et al. (2009). This expansion has been
so rampant such that listing all such models and their relationships is a difficult task and is out of
the scope of this work. However, a brief review of the recent models is presented in this paper. A
common objective function for regression is to minimize the least squares error:
||Y[ˆ] − Y ||[2] (1)
Where Y[ˆ] is the prediction. According to the Gauss-Markov theorem Gauss (1823), the least squares
error can be an unbiased linear model of minimum variance of the data under certain assumptions.
Ordinary least squares may fail to properly predict outcomes if it is applied to settings where the
Gauss-Markov assumptions are not held. Therefore, it is important to understand the assumptions and occasionally apply the proper changes to the objective function of equation 1 to modify the model Tai (2021). Manifestations of such changes include imposing regulations or constraints over the objective function. The literature on ordinary least squares estimation has extensively dealt with some of the well-known concerns that might violate the assumptions, such as
Ridge Hoerl & Kennard (1970a;b), Lasso Tibshirani (1996), Elastic Net Zou & Hastie (2005), trees
Quinlan (2014), forest Breiman (2001), boosting B¨uhlmann & Yu (2003) and others.
Common regression models include bagging, boosting, random forest Criminisi & Shotton (2013),
oblique trees Murthy et al. (1994); Norouzi et al. (2015); Heath et al. (1993), and regression SVM
-----
Drucker et al. (1997). In the context of regression trees, various approaches of inducing a tree are
presented in the literature. Most decision tree induction methods are concentrated on the splitting
criterion used at the growing phase of the tree Ikonomovska et al. (2011); Levati´c et al. (2014). Application of decision tree algorithms to multi-response regression has been previously considered in
the literature Breiman et al. (1984); De’Ath (2002). In Breiman et al. (1984); Quinlan (1986), authors consider training a decision tree for each individual output response. However, such approach
constructs a large model specially if the number of output responses are high. Another approach proposed in De’Ath (2002) consists of constructing a single decision tree for all the output responses. In
other words, the model predicts all the output values simultaneously through a single decision tree.
However, a model for all the outputs might not be sufficient Kocev et al. (2009) because they train
model for single response rather than the true problem which is a multi-response regression. Authors in Kocev et al. (2009) have explored two approaches to the multi-response regression problem
by comparing learning a model for each output separately (i.e., multiple regression trees), and learning one model for all outputs simultaneously (i.e., a single multi-target regression tree). In order
to improve predictive performance, Kocev et al. (2013) has also considered two ensemble learning techniques, namely, bagging Breiman (1996); Liang et al. (2011) and random forests Breiman
(2001) for regression trees and multi-target regression trees. The results showed that multi-target
regression trees created more accurate and compact models.
A related topic to the problem of this paper is nearest neighbor regression. Nearest neighbor regression and local estimators are well-established methods in the literature of ordinary univariate
location estimators (Benedetti (1977); Stone (1974); Tukey et al. (1977). However, as per our extensive search, there remains a gap in prototype nearest neighbor approaches to regression. The
only work that considered a similar model and optimization to SRNN was Huang (2017). However,
the proposed algorithm does not have guarantee of convergence or achieving some sort of optimum
solution.
3 PROPOSED METHOD
3.1 PRELIMINARIES
Assume a dataset consisting of tuples (xi, yi) where x, y and i represent input features, output
responses and index number. Each tuple represent a data xi and its corresponding output response
yi. Here, xi ∈ R[D] and yi ∈ R[d]. The Regression Synthetic Reduced Nearest Neighbor (Reg.
SRNN) consists of K tuples of synthetically produced centroids/prototypes (cj, ˆyj) where c, ˆyj and
j represent the centroid’s point in the input space, output prediction and index. At the inference
time, the Reg. SRNN operates like a nearest neighbor model where the centroids are used as the
samples. The problem of training Reg. SRNN is as follows:
N
{(cjmin, ˆyj)}[K]1 Xi=1 ||yi − yˆji[∗] [||][2] (2)
s.t ji[∗] [=][ argmin] d(xi − cj)
{j}[K]1
where d(.) is a distance metric. Through this paper we use the l-2 norm as the distance metric:
||xi − cj||[2] (3)
d(xi − cj) =
Essentially the prediction of the model consists of the output prediction of closest centroid to the
input sample. Officially, we define Reg. SRNN as follows:
NN (x) =
yjI(x ∈ Rj) (4)
Xj=1
Where, NN (.) represents a nearest neighbor function of the K centroids. I(.) is an indicator function that produces 1 if the input x is in the region of Rj. Rj represents the region where the closest
centroid to the points in that region is cj.
-----
3.2 INITIALIZATION
Numerical optimization algorithms require initialization (cite num opt). In this paper, we propose a
novel initialization for the regression SRNN. Previous work on the initialization of SRNN models
consisted of learning a K-means model for each class of the data. For example, in case of M classes
and K centroids, M[K] [centroids are learned for each class as initialization of the SRNN Kusner et al.]
(2014); Wang et al. (2016); Zhong et al. (2017); Tavallali et al. (2020b). However, such approach is
not applicable to the regression. This initialization also have close ties with naive Bayes and density
estimation Silverman (2018). Here, we expand this initialization to the case of Reg. SRNN.
Intuitively, the output responses can consist of several modalities. In other words, it is possible
that the output responses are generated from several distributions. The clusters of such distributions
can be approximated by running a K-means over the output space (M centroids). Assume that Sm
represents the set of samples assigned to each output cluster. Next step consists of learning |[KN]Sm|
centroids over the input features of the Sm for all M clusters. In other words, we learn centroids
over the input features of each output cluster relative to the population of that cluster. The found
centroids at the second step are used as initialization for the Reg. SRNN. At this step, ˆyj is found
using the following formula:
yˆj = mean(yi ∈ Sj) (5)
where Sj represents the set of samples that are assigned to j[th] centroid. Sj essentially consists of
samples where j[th] centroid is the closest centroid to them. mean(.) represents the average of its
input set. Note that Sj ∈ Rj.
3.3 CONSISTENCY
Here we discuss the consistency of the Reg. SRNN. We show that Reg. SRNN is a nonparametric
and consistent under mild assumptions for continuous features. Assume an independent identically
distributed (iid) dataset where it is being generated from f (x). f (x) represents the true function
for relation between the inputs and outputs. Also assume that f (x) is a piecewise constant function.
Over a set of N observations, consistency of a nonparametric estimator f[ˆ]N (x) (such as Reg. SRNN)
is shown using the following formula Parzen (1962):
( f[ˆ]N (x) − f (x))[2]dx = 0) = 1 (6)
Pr( lim
N →∞
The proof of consistency of Reg. SRNN is similar to proof of consistency for regression trees
Breiman et al. (1984) and density estimators in Ram & Gray (2011).
Theorem 1 (Consistency of Reg. SRNN). The estimator defined in equation 4 satisfies equation
equation 6.
Proof. Assume B and d1 denote the collection of all sets t ⊂ X and a fixed positive integer, respectively. Assume that B describes the sets that are the solution to a system of d1 inequalities b[T] x ≤ c
where b ∈ R[d] and c ∈ R. Every region in the Reg. SRNN in formula equation 4 can be seen as a
solution of a system of d1 inequalities of the form b[T] x ≤ c where b is a hot-one-vector (only one
element of b is 1 and the rest are 0). Therefore, Reg.SRNN ⊂ B.
Assume a random point Xn from function f on X, (n ≥ 1). F[ˆ]N represents the empirical function
learned by Reg. SRNN over Xn For N ≥ 1, 1 ≤ n ≤ N, and defined on a set t ⊂ X by
FˆN (t) = [1]
nX=1 ynI(Xn ∈ t) = mean(yn ∈ Rt) =
fˆN (x)dx (7)
where yn = f (xn) and f[ˆ](x) is the estimator presented in equation 4. Using a general version of
Glivenko-Cantelli theorem Vapnik & Chervonenkis (2015)
Pr( lim |F[ˆ]N (t) −
N →∞[sup]t∈B
f (x)dx| = 0) = 1 (8)
-----
By replacing equation equation 7 in equation 8, we have
Pr( lim |
N →∞[sup]t∈B Z
Pr( lim
N →∞[sup]t∈B
fˆN (x)dx −
f (x)dx| = 0) = 1 (9)
then:
|f[ˆ]N (x) − f (x)|dx ≥ 0) = 1 (10)
Further, by assuming that the region of t leans toward 0 as N →∞ (Pr(N lim→∞ t [dx][ = 0) = 1][).]
R
Rest of the steps will follow similar to theorem 1 in Ram & Gray (2011) and we get
( f[ˆ]N (x) − f (x))[2]dx = 0) = 1 (11)
Pr( lim
N →∞
Please note that the steps are similarly done in Ram & Gray (2011) except the step of equation 7
which is different.
The consistency shows that as the number of samples go to infinity and as |SKj| [→] [0][, then the Reg.]
SRNN is consistent. This is provable thanks to the assumption that the true function that relates the
inputs to the outputs is a piecewise constant function. This essentially means that the SRNN itself is
consistent also for classification tasks.
3.4 THE EXPECTATION MAXIMIZATION OF REG. SRNN
The problem equation 2 represents the training problem of Reg. SRNN. SRNN generally is known
as prototype nearest neighbor in other papers of literature and centroids are also called prototypes.
Problem equation 2 resembles K-means Lloyd (1982) problem except that the loss function is ||xi −
cj||. The expectation maximization in this paper follows same approach as to K-means algorithm
and the approach in Tavallali et al. (2020b). The optimization consists of two steps, assignment step
and the update step.
3.4.1 ASSIGNMENT STEP
The assignment step consists of calculating the assignment of the samples to each centroid and
finding the optimum output prediction of each centroid. The problem of this step is as follows:
N
min ||yi − yˆj ∗||2.
{ ˆyj}[K]1 Xi=1 (12)
s.t ji[∗] [=][ argmin] d(xi − cj).
{j}[K]1
Note that problem equation 12 only tends to optimize over ˆyj for j = 1...K. Using sets Sj, problem
equation 12 can be simplified to:
{i|(xiX,yi)∈Sj} ||yi − yˆj||[2]. (13)
min
{ ˆyj}[K]1
j=1
The problem equation 13 can be separated over each centroid and its corresponding set. This can be
done since the regions are distinct and samples can not be shared among the regions. The prediction
for each region is the label of j[th] centroid ˆyj (the centroid that represents the region). As a result,
the problem of optimizing label for each region is
{i|(xiX,yi)∈Sj } ||yi − yˆj||[2]. (14)
min
yˆj
whose minimum is presented in formula equation 5. In other words, the optimum of ˆyj is the mean
of response of samples in the Rj.
As a result, in the assignment step, the Sj and ˆyj have to be calculated for all the samples.
-----
3.4.2 UPDATE STEP
The update step consists of updating the position of centroids such that the objective function in
equation 2 is minimized Lloyd (1982). At this step the output prediction of the centroids are kept
constant. The update step is affected by all the samples in the dataset because by changing the
position of centroid, the assignments of the samples can get changed and as a result the prediction
for each train sample gets changed. Further it will be shown that finding optimum of the problem
in this step is NP-hard. Therefore, the centroid problem is approximated through a novel surrogate
objective function.
Each centroid is optimized individually Tavallali et al. (2020b). The optimization problem for k[th]
centroid consists of moving the k[th] centroid such that the samples’ assignment are changed in favor
of decreasing the objective function equation 2. The centroid problem is
N
min ||yi − yˆj ∗||2.
ck
Xi=1 (15)
s.t ji[∗] [=][ argmin] d(xi − cj).
{j}[K]1
Note that the optimization is only over ck. We rewrite the problem equation 15 over the assignment
of a sample to k[th] centroid or to the rest of centroids. For simplicity we introduce the notation
rij = d(xi − cj). The problem is
N
mincj ||yi − yˆk||[2]U (riji[∗] [−] [r][ik][) +][ ||][y][i][ −] y[ˆ]ji[∗] [||][2][U] [(][r][ik][ −] [r][ij]i[∗] [)][.]
Xi=1 (16)
s.t ji[∗] [=][ argmin] d(xi − cj).
{j}[K]1 [,j][̸][=][k]
In problem equation 16, U (.) is a step function where it outputs 1 if the input is larger than 0 and
otherwise it will produce 0. Note that the input arguments of the step functions are negative of each
other. This means a sample has to either get assigned to the k[th] centroid or the rest of centroids. The
sample can not get assigned to both terms of equation 16. The assignment of i[th] sample to each term
will produce a continuous error. This essentially means that the problem equation 16 is a weighted
binary classification problem. The problem equation 16 encourages the sample to get assigned to the
side that produces lower error. For simplicity we introduce the sets Sc and Sc[′]k [. The][ S][c][k][ represents]
the set of samples that produce lower error if assigned to the ck centroid and Sc[′]k [represents the set]
of other samples. let ti = abs(||yi − yˆk||[2] −||yi − yˆji[∗] [||][2][)][ where][ abs][(][.][)][ returns the absolute value]
of its input. The problem of equation 16 is equivalent to
min
cj
(xi,yXi)∈Sc tiU (riji[∗] [−] [r][ik][) +]
tiU (rik − riji[∗] [)] (17)
(xi,yi)∈Sc′
The problem equation 17 is a NP-hard problem Nguyen & Sanner (2013). Therefore, inspired by the
SVM, we approximate the solution to problem equation 17 using the following surrogate objective
function
c[∗]k[(][µ][) =][ argmin]
ck
xXi∈Sc tirik +
tirelu(µrij∗ − rik) (18)
xi∈Sc′
where µ is a penalty coefficient. Intuitively, the surrogate objective function encourages the k[th]
centroid to stay close to samples of Sc while staying away from samples of Sc[′]k [.][ µ][ is increased from]
0 to 1 and along this path, the c[∗]k[(][µ][)][ that produces the smallest error for equation 17 is selected.]
This surrogate is a modified version of surrogate objective function in Tavallali et al. (2020b). The
µ acts similar to slack variables in a SVM problem. The objective function of problem equation 18
is a continuous function; thus, a local optimum of the problem can be found using gradient-based
algorithms.
3.5 RELATION TO EM ALGORITHM
The proposed algorithm is originally inspired by the EM algorithm used for K-means. The assignment step consists of finding the samples assigned to each centroid and finding the optimal output
-----
prediction of each centroid. The assignment of samples is the same as calculating the prior probabilities in an EM algorithm. Finding optimum output prediction of each centroid can also be considered
as a part of maximization step.
In the update step, the centroids have to be updated. At this step, the outcome of assigning each
sample to each centroid is evaluated (||yi − yj||[2][K]j=1[∀][i][ = 1][,][ 2][, ..., N] [). This evaluation is equivalent]
to calculating the posterior probabilities. Then the centroid problem is approximated using these
outcomes which is the maximization step.
3.6 RELATION OF SURROGATE TO SVM
The problem of finding the best centroid that is closer to samples of Sc than any other centroid can
be cast as a feasibility problem.
find ck
(19)
s.t rik < riji[∗] ∀(xi, yi) ∈ Sc
rik > riji[∗] ∀(xi, yi) ∈ Sc[′]
However, this feasibility problem is NP-hard since the the second set of constraints are concave
Sahni (1974). These constraints make the problem different from similar SVM problems where the
constraints are convex and global solution can be found in efficient time.
3.7 PROPERTIES OF THE ALGORITHM
computational complexity Optimizing the centroid problem takes O(ND) since it uses a gradient based algorithm for solving the surrogate objective. All the K centroids have to optimized at
each iteration. Therefore, the computational complexity of the algorithm is O(αNDK) where α is
the number of iterations.
Convergence The convergence to a local optimum is similar to that of the K-means algorithm
Lloyd (1982); Tavallali et al. (2020b). At each iteration, the error decreases and the objective function is bounded by 0 from bellow. Further, the different combination of assignment of the samples
to the centroids is finite. Therefore, the algorithm stops after a finite number of iterations.
4 EXPERIMENTAL RESULTS
In this section, the experimental results are presented to demonstrate the merits of proposed proposed algorithm. Various datasets are downloaded and used for evaluation from UCI repository
Dheeru & Karra Taniskidou (2017). We partitioned the datasets using the following setups: 1- if
the dataset contained a separate test set, the dataset was used as provided in the repository. In case
cross-validation was needed (for specific models) and validation set was not provided then the we
partitioned the trainset to 80% train and 20% validation sets. 2- if the test set was not provided we
partitioned the dataset to 80% train and 20% test set. For models that needed validationset, we divided the trainset to 80% train and 20% validation set. 3- if all sets were provided by the repository,
then the sets were used as provided. The setup for Reg. SRNN is as follows: at the initialization
phase, we used K = 4 for the output clusters for slice localization data and for the rest of datasets
K = 2 was used for the output clusters. The number of input centroids used for each output cluster
were proportional to the population of each cluster. The centroid problem was optimized using batch
stochastic gradient descent. We used various setups for each dataset to achieve the smallest possible
train error.
Various models are used for comparison with Reg. SRNN. The basis of comparison is based on
the number of base models used in each model. All the models comparable to the Reg. SRNN
can be seen as a kind of ensemble model that consists of several base models. In case of forest
and boosting, each tree is a base model. For prototype models such as Reg. SRNN and K-means,
each centroid is a base model. For Radial Basis Function (RBF) models, each RBF is considered
as a base model. We also compared our model with linear regression and Ridge regression. The
two models are presented with straight lines. Figures 1, and 2 present performance of each model
-----
100 **slice localization data** 100 **CPU** 35 **Airfoil** 0.02 **mg**
0.019
80 80 30 0.018
60 60 25 0.017
MSE MSE MSE MSE 0.016
40 40 20 0.015
20 15 0.014
20 0.013
0 100 200 300 400 500 10 20 30 40 50 60 10 10 20 30 40 50 60 0.012 10 20 30 40 50 60
#base models #base models #base models #base models
**slice localization data**
100
80
60
40
20
|Col1|SRNN SRNN0 RF K-means regboost|
|---|---|
|bagging RBF+lin-reg lin-reg ridgeCV|bagging RBF+lin-reg lin-reg ridgeCV|
SRNN
SRNN0
RF
K-means
regboost
bagging
RBF+lin-reg
lin-reg
ridgeCV
100 **CPU** 35 **Airfoil** 0.02 **mg**
0.019
80 30 0.018
60 25 0.017
MSE MSE MSE 0.016
40 20 0.015
15 0.014
20 0.013
10 20 30 40 50 60 10 10 20 30 40 50 60 0.012 10 20 30 40 50 60
#base models #base models #base models
Figure 1: Horizontal axis represents the number of base models(e.g., centroids, RBFs, trees and
etc.). The vertical axis represents the mean squared error. The first and second row curves show the
train errors and test errors, respectively. Reg. SRNN is compared with similar models.
over various datasets based on mean squared error versus number of base models. Reg. SRNN was
compared with Random Forest (RF), K-means, Regression Boosting (regboost), bagging, Radial
Basis Function+ linear regression (RBF+lin-reg), linear regression (lin-reg), and ridge regression
(ridgeCV). For the RF, each tree was trained using 70% of the trainset and 0.7 of features were
randomly selected and used at the split nodes. For the K-means, a K-means model with desired
number of centroids were trained over trainset and the assignment step of Reg. SRNN was applied
to the model. For the regboost, we used trees of depth 3. For bagging, similar setup to RF was used
except that no feature randomization is applied. RBF+lin-reg consists of first training RBFs (same
number of RBFs as the centroids in other models) and the training linear regression over the outputs
of RBFs. The width of RBFs were selected by cross-validation. The penalty coefficient of ridgeCV
was selected by cross-validation.
As can be observed from figures 1, and 2, the SRNN-Reg was able to achieve better or comparable
train and test errors to other models.
5 LIMITATIONS:
The model presented in this paper is new in the sense that it has never been proposed previously.
However, since its nature is a nearest neighbor model we expect that the model presents some limitations similar to those of a nearest neighbor model. Similar to nearest neighbor for regression that
might have high test error in high dimensional datasets Hastie et al. (2009). Therefore, we suspect
that the performance of Reg-SRNN might deteriorate with higher dimensional dataset. However,
in our experiments, we noticed that the Reg-SRNN performed very well in slice localization data
which is a high dimensional dataset (384 features).
6 CONCLUSION
In this paper, as per our search, we have proposed the first regression synthetic reduced nearest
neighbor. The consistency of the algorithm was proved and its properties were explored. We showed
that the algorithm is computationally efficient and can converge to a local optimum in the sense than
no move can improve the model any further. The approach is essentially the same type of EM
algorithm used for K-means and it is extended for the case of regression synthetic reduced nearest
neighbor. The update step was an NP-hard weighted binary classification problem. The optimum of
such problems are typically approximated using a surrogate objective function, such as hinge loss in
SVM for 0-1 binary classification problem. Therefore, we approximated the solution to the update
step through a novel surrogate objective function. Further, we analyzed the relation of the update
-----
train error test error
**eunite** **eunite**
1000 1000
900
800
800
MSE 600 MSE 700600
400 500
400
200
300
10 20 30 40 50 60 10 20 30 40 50 60
#base models #base models
Figure 2: Horizontal axis represents the number of base models(e.g., centroids, RBFs, trees and
etc.). The vertical axis represents the mean squared error. Reg. SRNN is compared with similar
models.
step with SVM. Experimentally, we showed that the Reg. SRNN performs better or competitive to
other similar models in the literature such as ensembles and centroid based models.
7 REPRODUCIBILITY
The experiment codes (currently attached as supplementary materials) will be published on a Github
repository to facilitate reproduction and extension of our results. Furthermore, the datasets used in
our experiments are publicly available via the UCI repository.
REFERENCES
Jean-Michel Begon, Arnaud Joly, and Pierre Geurts. Globally induced forest: A prepruning compression scheme. In International Conference on Machine Learning, pp. 420–428, 2017.
Jacqueline K Benedetti. On the nonparametric estimation of regression functions. Journal of the
Royal Statistical Society: Series B (Methodological), 39(2):248–253, 1977.
Dimitris Bertsimas and Jack Dunn. Optimal classification trees. Machine Learning, 106(7):1039–
1082, 2017.
Leo Breiman. Bagging predictors. Machine learning, 24(2):123–140, 1996.
Leo Breiman. Random forests. Machine learning, 45(1):5–32, 2001.
Leo J. Breiman, Jerome H. Friedman, R. A. Olshen, and Charles J. Stone. Classification and Regression Trees. Wadsworth, Belmont, Calif., 1984.
Peter B¨uhlmann and Bin Yu. Boosting with the l 2 loss: regression and classification. Journal of the
American Statistical Association, 98(462):324–339, 2003.
Ricardo Cisneros, Hamed Gharibi, Marcela R Entwistle, Pooya Tavallali, Mukesh Singhal, and
Donald Schweizer. Nitrogen dioxide and asthma emergency department visits in california, usa
during cold season (november to february) of 2005 to 2015: A time-stratified case-crossover
analysis. Science of The Total Environment, 754:142089, 2021.
Antonio Criminisi and Jamie Shotton. Decision Forests for Computer Vision and Medical Image
Analysis. Advances in Computer Vision and Pattern Recognition. Springer-Verlag, 2013.
Glenn De’Ath. Multivariate regression trees: a new technique for modeling species–environment
relationships. Ecology, 83(4):1105–1117, 2002.
Dua Dheeru and Efi Karra Taniskidou. UCI machine learning repository, 2017. URL
[http://archive.ics.uci.edu/ml.](http://archive.ics.uci.edu/ml)
Harris Drucker, Chris JC Burges, Linda Kaufman, Alex Smola, Vladimir Vapnik, et al. Support vector regression machines. Advances in neural information processing systems, 9:155–161, 1997.
-----
Carl Friedrich Gauss. Theoria combinationis observationum erroribus minimis obnoxiae, volume 2.
H. Dieterich, 1823.
Franz Graf, Hans-Peter Kriegel, Sebastian P¨olsterl, Matthias Schubert, and Alexander Cavallaro.
Position prediction in ct volume scans. In Proceedings of the 28th International Conference on
Machine Learning (ICML) Workshop on Learning for Global Challenges, Bellevue, Washington,
WA, 2011a.
Franz Graf, Hans-Peter Kriegel, Matthias Schubert, Sebastian P¨olsterl, and Alexander Cavallaro. 2d
image registration in ct images using radial image descriptors. In International Conference on
Medical Image Computing and Computer-Assisted Intervention, pp. 607–614. Springer, 2011b.
Trevor Hastie, Robert Tibshirani, and Jerome H. Friedman. The elements of statistical learning:
data mining, inference, and prediction. New York, NY: Springer, second edition, 2009.
David Heath, Simon Kasif, and Steven Salzberg. Induction of oblique decision trees. In IJCAI,
volume 1993, pp. 1002–1007, 1993.
Arthur E Hoerl and Robert W Kennard. Ridge regression: Biased estimation for nonorthogonal
problems. Technometrics, 12(1):55–67, 1970a.
Arthur E Hoerl and Robert W Kennard. Ridge regression: applications to nonorthogonal problems.
Technometrics, 12(1):69–82, 1970b.
Hanwen Huang. Regression in heterogeneous problems. Statistica Sinica, pp. 71–88, 2017.
Elena Ikonomovska, Jo˜ao Gama, and Saˇso Dˇzeroski. Incremental multi-target model trees for data
streams. In Proceedings of the 2011 ACM symposium on applied computing, pp. 988–993. ACM,
2011.
Dragi Kocev, Saˇso Dˇzeroski, Matt D White, Graeme R Newell, and Peter Griffioen. Using single-and
multi-target regression trees and ensembles to model a compound index of vegetation condition.
Ecological Modelling, 220(8):1159–1168, 2009.
Dragi Kocev, Celine Vens, Jan Struyf, and Saˇso Dˇzeroski. Tree ensembles for predicting structured
outputs. Pattern Recognition, 46(3):817–833, 2013.
Matt Kusner, Stephen Tyree, Kilian Weinberger, and Kunal Agrawal. Stochastic neighbor compression. In International Conference on Machine Learning, pp. 622–630, 2014.
Jurica Levati´c, Michelangelo Ceci, Dragi Kocev, and Saˇso Dˇzeroski. Semi-supervised learning for
multi-target regression. In International Workshop on New Frontiers in Mining Complex Patterns,
pp. 3–18. Springer, 2014.
Alexander Hanbo Li and Andrew Martin. Forest-type regression with general losses and robust
forest. In International Conference on Machine Learning, pp. 2091–2100, 2017.
Guohua Liang, Xingquan Zhu, and Chengqi Zhang. An empirical study of bagging predictors for
different learning algorithms. In Proceedings of the AAAI Conference on Artificial Intelligence,
volume 25, 2011.
Stuart Lloyd. Least squares quantization in pcm. IEEE transactions on information theory, 28(2):
129–137, 1982.
Sreerama K Murthy, Simon Kasif, and Steven Salzberg. A system for induction of oblique decision
trees. Journal of artificial intelligence research, 2:1–32, 1994.
Tan Nguyen and Scott Sanner. Algorithms for direct 0–1 loss optimization in binary classification.
In International Conference on Machine Learning, pp. 1085–1093, 2013.
Mohammad Norouzi, Maxwell D Collins, David J Fleet, and Pushmeet Kohli. Co2 forest: Improved
random forest by continuous optimization of oblique splits. arXiv preprint arXiv:1506.06155,
2015.
-----
Emanuel Parzen. On estimation of a probability density function and mode. The annals of mathematical statistics, 33(3):1065–1076, 1962.
J. Ross Quinlan. Induction of decision trees. Machine learning, 1(1):81–106, 1986.
J Ross Quinlan. C4. 5: programs for machine learning. Elsevier, 2014.
Parikshit Ram and Alexander G Gray. Density estimation trees. In Proceedings of the 17th ACM
SIGKDD international conference on Knowledge discovery and data mining, pp. 627–635. ACM,
2011.
Sartaj Sahni. Computationally related problems. SIAM Journal on computing, 3(4):262–279, 1974.
Fadil Santosa and William W Symes. Linear inversion of band-limited reflection seismograms.
SIAM Journal on Scientific and Statistical Computing, 7(4):1307–1330, 1986.
Bernard W Silverman. Density estimation for statistics and data analysis. Routledge, 2018.
Mervyn Stone. Cross-validatory choice and assessment of statistical predictions. Journal of the
Royal Statistical Society: Series B (Methodological), 36(2):111–133, 1974.
Yunpeng Tai. A survey of regression algorithms and connections with deep learning. arXiv preprint
arXiv:2104.12647, 2021.
Pooya Tavallali, Mehran Yazdi, and Mohammad Reza Khosravi. Robust cascaded skin detector
based on adaboost. Multimedia Tools and Applications, 78(2):2599–2620, 2019.
Pooya Tavallali, Hamed Gharibi, Mukesh Singhal, Donald Schweizer, and Ricardo Cisneros. A
multi-pollutant model: a method suitable for studying complex relationships in environmental
epidemiology. Air Quality, Atmosphere & Health, 13:645–657, 2020a.
Pooya Tavallali, Peyman Tavallali, Mohammad Reza Khosravi, and Mukesh Singhal. Interpretable
synthetic reduced nearest neighbor: An expectation maximization approach. In 2020 IEEE International Conference on Image Processing (ICIP), pp. 1921–1925. IEEE, 2020b.
Pooya Tavallali, Mehran Yazdi, and Mohammad R Khosravi. A systematic training procedure for
viola-jones face detector in heterogeneous computing architecture. Journal of Grid Computing,
pp. 1–16, 2020c.
Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical
Society: Series B (Methodological), 58(1):267–288, 1996.
John W Tukey et al. Exploratory data analysis, volume 2. Reading, Mass., 1977.
Vladimir N Vapnik and A Ya Chervonenkis. On the uniform convergence of relative frequencies of
events to their probabilities. In Measures of complexity, pp. 11–30. Springer, 2015.
Wenlin Wang, Changyou Chen, Wenlin Chen, Piyush Rai, and Lawrence Carin. Deep metric learning with data summarization. In Joint European Conference on Machine Learning and Knowledge
Discovery in Databases, pp. 777–794. Springer, 2016.
Kai Zhong, Ruiqi Guo, Sanjiv Kumar, Bowei Yan, David Simcha, and Inderjit Dhillon. Fast classification with binary prototypes. In Artificial Intelligence and Statistics, pp. 1255–1263, 2017.
Hui Zou and Trevor Hastie. Regularization and variable selection via the elastic net. Journal of the
royal statistical society: series B (statistical methodology), 67(2):301–320, 2005.
-----