Professional Documents
Culture Documents
Article
Reproducing Kernel Hilbert Space Approach to Multiresponse
Smoothing Spline Regression Function
Budi Lestari 1,2 , Nur Chamidah 2,3, * , Dursun Aydin 4 and Ersin Yilmaz 4
1 Department of Mathematics, Faculty of Mathematics and Natural Sciences, The University of Jember,
Jember 68121, Indonesia
2 Research Group of Statistical Modeling in Life Sciences, Faculty of Science and Technology, Airlangga
University, Surabaya 60115, Indonesia
3 Department of Mathematics, Faculty of Science and Technology, Airlangga University,
Surabaya 60115, Indonesia
4 Department of Statistics, Faculty of Science, Muğla Sıtkı Koçman University, Muğla 48000, Turkey
* Correspondence: nur-c@fst.unair.ac.id
Abstract: In statistical analyses, especially those using a multiresponse regression model approach,
a mathematical model that describes a functional relationship between more than one response
variables and one or more predictor variables is often involved. The relationship between these
variables is expressed by a regression function. In the multiresponse nonparametric regression (MNR)
model that is part of the multiresponse regression model, estimating the regression function becomes
the main problem, as there is a correlation between the responses such that it is necessary to include
a symmetric weight matrix into a penalized weighted least square (PWLS) optimization during the
estimation process. This is, of course, very complicated mathematically. In this study, to estimate
the regression function of the MNR model, we developed a PWLS optimization method for the
MNR model proposed by a previous researcher, and used a reproducing kernel Hilbert space (RKHS)
approach based on a smoothing spline to obtain the solution to the developed PWLS optimization.
Citation: Lestari, B.; Chamidah, N.; Additionally, we determined the symmetric weight matrix and optimal smoothing parameter, and
Aydin, D.; Yilmaz, E. Reproducing investigated the consistency of the regression function estimator. We provide an illustration of the
Kernel Hilbert Space Approach to effects of the smoothing parameters for the estimation results using simulation data. In the future, the
Multiresponse Smoothing Spline theory generated from this study can be developed within the scope of statistical inference, especially
Regression Function. Symmetry 2022, for the purpose of testing hypotheses involving multiresponse nonparametric regression models and
14, 2227. https://doi.org/10.3390/ multiresponse semiparametric regression models, and can be used to estimate the nonparametric
sym14112227
component of a multiresponse semiparametric regression model used to model Indonesian toddlers’
Academic Editor: Calogero Vetro standard growth charts.
multiresponse nonparametric regression (MNR) model; and Ref. [33] discussed smoothing
techniques, namely, kernel and spline, to estimate the coefficient of a rates model.
In regression modeling, a common problem involves more than one response variable
observed at several values of predictor variables and between responses that are correlated
with each other. The multiresponse nonparametric regression (MNR) model approach
is appropriate for modeling the functions which represent the relationship between the
response variable and predictor variable with correlated responses. In this model there is a
correlation between the responses. Because of this correlation, it is necessary to construct
a matrix called a weight matrix. Constructing the weight matrix is one of the things that
distinguishes the MNR model approach from a classical model approach, that is, a para-
metric regression model or uniresponse nonparametric regression model approach. Thus,
in the estimation process the regression function requires a weight matrix in the form of a
symmetric matrix, especially a diagonal matrix. Furthermore, in the MNR model there are
several smoothing techniques which can be used to estimate the regression function. One
of these smoothing techniques is the smoothing spline approach. In recent years, studies on
smoothing splines have attracted a great deal of attention and the methodology has been
widely used in many areas of research, for example, for estimating regression functions
of nonparametric regression models, in [34,35] used smoothing spline, mixed smoothing
spline, and Fourier series; estimating regression functions were conducted by [36,37] for a
semiparametric nonlinear regression model and a semiparametric regression model; and
smoothing spline in an ANOVA model was discussed by [38]. Smoothing spline estimator,
with its powerful and flexible properties, is one of the most popular estimators used for
estimating regression function of the nonparametric regression model. Although the re-
searchers mentioned above have previously discussed splines for estimating regression
functions in many cases, none of them have used a reproducing kernel Hilbert space
(RKHS) approach to estimate the regression function of the MNR model. On the other hand,
even though there are studies, as mentioned above, that have used the RKHS approach for
estimating regression functions, those researchers used the RKHS for estimating regression
functions of single–response or uniresponse linear regression models only. This means
that RKHS approaches were not used for estimating the regression function of the MNR
model based on a smoothing spline estimator. In addition, although [34] used the RKHS
approach to estimate the regression function of the MNR model, and also discussed it in a
special case involving a simulation study; but Ref. [34] assumed that the three responses of
the MNR model have the same smoothing parameter values, which in real life situation
is a difficult assumption to fulfill. In addition, Ref. [34] did not discuss the consistency of
the smoothing spline regression function estimator. Therefore, in this study we provide a
theoretical discussion on estimating the smoothing spline regression function of the MNR
model in case of unequal values of the smoothing parameters using the RKHS approach.
In other words, in this study we discuss it for the more general case.
where yri is the observation value of the response variable on the r th response and the ith
observation; f r (·) represents an unknown nonparametric regression function of r th response
which is assumed to be smooth in the sense that it is contained in a Sobolev space; tri is
the observation value of a predictor variable on the r th response and the ith observation;
and ε ri represents the value of the random error on the r th response and the ith observation,
Symmetry 2022, 14, 2227 4 of 22
which is assumed to have zero mean and variance σri2 (heteroscedastic). In this study, we
ρr f or r = s
assume that the correlation between responses is ρrs = .
0 f or r 6= s
In general, the main problem in MNR modelling is how we estimate the MNR model,
which in this case is equivalent to the problem of estimating the regression function of the
MNR model. There are many smoothing techniques that can be used to estimate the MNR
model presented in (1), for example, kernel, local linear, splines, and local polynomial. One
of these smoothing techniques is the spline approach, in which the smoothing spline is
the most flexible estimator for estimating fluctuating data on sub-intervals. The following
briefly presents the estimation method using the smoothing spline estimator. Further
details related to the smoothing spline estimator can be found in [20].
p
where N = ∑r=1 nr ; W1 , . . . , Wp are symmetric weight matrices, λ1 , . . . , λ p are smooth-
ing parameters, and f1 , f2 , . . . , fp are unknown regression functions in a Sobolev space
W2m [ ar , br ], where the Sobolev space W2m [ ar , br ] is defined as follows:
n
W2m [ ar , br ] = f f (v) , v = 0, 1, 2, . . . , m − 1 are absolutely continuous on [ ar , br ] and
Furthermore, to obtain the solution to the PWLS provided in (2), we used the repro-
ducing kernel Hilbert space (RKHS) approach. In the following section, we provide a brief
review of RKHS. Further details related to RKHS can be found in [39], a paper concerning
the theory of RKHS, and in [40], a textbook which discusses the use of RKHS in probability
and statistics.
A Hilbert space H is called an RKHS on a set X over field F if the following conditions
are met [1,39]:
(i) H is a vector subspace of F (X, F), where F (X, F) is a vector space over F;
(ii) H is endowed with an inner product h, i, making it into a Hilbert space;
(iii) the linear evaluation functional Ey : H → F , defined by Ey ( f ) = f (y), is bounded,
for every y ∈ X.
Furthermore, if H is an RKHS on X, then
forevery y ∈ X there exists a unique vector
k y ∈ H such that for every f ∈ H, f (y) = f , k y . This is because every bounded linear
functional is provided by the inner product with a unique vector in H. The function k y is
called a reproducing kernel (RK) for point y. The reproducing kernel (RK) for H is
a two–
variable function defined byK ( x, y) = k y ( x ). Hence, we have K ( x, y) = k y ( x ) = k y , k x
and k Ey k2 = kk y k2 = k y , k y = K (y, y).
2.4. Simulation
The simulation in this study consists of a simulation to determine the optimal smooth-
ing spline based on a generalized cross-validation (GCV) criterion to obtain the best esti-
mated MNR model and a simulation to describe the effect of the smoothing parameters
on the estimation results of the regression function of the MNR model based on minimal
GCV value. We generate samples sized n = 100 from the MNR model and provide an
illustration of the effects of the smoothing parameters in order to estimate the results of the
MNR model by comparing three kinds of different smoothing parameter values, namely,
small, optimal, and large smoothing parameters values.
In the following section, we provide the results and discussion of this study covering
estimation of the regression function of the MNR model using the RKHS approach by
estimating the symmetric weight matrix and optimal smoothing parameters, a simulation
study, and investigating the consistency of the smoothing spline regression function estimator.
3.1. Estimating the Regression Function of the MNR Model Using the RKHS Approach
The MNR model presented in (1) can be expressed in matrix notation as follows:
y = f+ε (3)
T T T T
where y = y1 , y2 , . . . , yp , f = f1 , f2 , . . . , fp , t = t1 , t2 , . . . , tp , ε = ε1 , ε2 , . . . , εp ,
yr = (yr1 , yr2 , . . . , yrnr )T , fr = ( f r (tr1 ), f r (tr2 ), . . . , f r (trnr ))T , tr = (tr1 , tr2 , . . . , trnr ) T , and
εr = (ε r1 , ε r2 , . . . , ε rnr )T .
We assume that ε is a zero mean random error with covariance W−1 . In this case, the
covariance matrix W−1 is a symmetrical matrix, that is, it is a diagonal matrix which can be
expressed as follows:
W−1 = diag W1−1 , W2−1 , . . . , Wp−1 (4)
To determine the regression function of the MNR model (1) using the RKHS approach,
we first express the MNR model in a general smoothing spline regression model [20].
Therefore, we can express the MNR model (1) as follows:
Fr = Gr ⊕ Kr . (6)
where f r ∈ Fr and h·, ·i denote an inner product. Next, by considering Equations (8) and (9)
and applying the properties of the inner product, the function fr (tri ) can be written as follows:
D E
f r (tri ) = δri , αrT br + βrT cr
D E (10)
= δri , αrT br + δri , βrT cr
Then, based on Equation (10), we can obtain the regression functions f r (tri ) for
r = 1, 2, . . . , p, which are the regresion functions for the first response, the second re-
sponse, . . . , and the pth response, as follows:
D E
For r = 1, we have : f 1 (t1i ) = δ1i , α1T b1 + δ1i , β1T c1 , i = 1, 2, . . . , n1
D E
T T
For r = 2, we have : f 2 (t2i ) = δ2i , α2 b2 + δ2i , β2 c2 , i = 1, 2, . . . , n2
(11)
..
.
D E D E
= δ , αT b + δ , βT c , i = 1, 2, . . . , n
For r = p, we have : f t p pi pi p p pi p p p
Hence, following Equation (11), we obtain the regression function for i = 1 as follows:
T
f1 (t1 ) = f 1 (t11 ), f 1 (t12 ), . . . , f 1 t1n1 = A1 b1 + C1 d1 (12)
For 𝑟 = 1, we have: 𝑓 (𝑡 ) = 〈𝛿 , 𝛂 𝐛 〉 +follows:
〈𝛿 , 𝛃 𝐜 〉, 𝑖 = 1,2, … , 𝑛
⎫
For 𝑟 = 2, we have: 𝑓 (𝑡 ) = 〈𝛿 , 𝛂 𝐛 〉 + 〈𝛿 , 𝛃 𝐜 〉, 𝑖 = 1,2, … , 𝑛
⋮ 𝐟𝟏 (𝑡 ) = 𝑓(11)
(𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 ) =
⎬
For 𝑟 = 𝑝, we have: 𝑓 𝑡 = 〈𝛿 , 𝛂 𝐛 〉 + 〈𝛿 , 𝛃 𝐜 〉, 𝑖 = 1,2, … , 𝑛 ⎭
where 𝐛𝟏 = 𝑏 , 𝑏 , … , 𝑏 ; 𝐝𝟏 = 𝑑 , 𝑑 , … , 𝑑 ;
Hence, following Equation (11), we obtain the regression function for 𝑖 = 1 as
Symmetry 2022, 14, 2227 follows: 〈𝛿 , 𝛼 〉 〈𝛿 , 𝛼 〉 ⋯ 〈𝛿 , 𝛼 7 of〉 22
⎛ 〈𝛿 , 𝛼 〉 〈𝛿 , 𝛼 〉 ⋯ 〈𝛿 , 𝛼 〉 ⎞
𝐟𝟏 (𝑡 ) = 𝑓 (𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 ) 𝐀= 𝐀𝟏⎜𝐛𝟏 + 𝐂𝟏 𝐝𝟏
𝟏 =
(12)
⎟; and
⋮ ⋮ ⋱ ⋮
where 𝐛𝟏 = 𝑏 , 𝑏 , … , 𝑏 ; 𝐝𝟏 = 𝑑 , 𝑑 , … , 𝑑 T ; 〈𝛿 , 𝛼 〉 〈𝛿 , 𝛼 〉 ⋯ 〈𝛿 , 𝛼 〉
⎝ ⎠T
where b1 = b11 , b12 , . . . , b1m1 ; d1 = d11 , d12 , . . . , d1n1 ;
〈𝛿 , 𝛼 〉 〈𝛿 , 𝛼 〉 ⋯ 〈𝛿 , 𝛼 〉 〈𝛿 , 𝛽 〉 〈𝛿 , 𝛽 〉 ⋯ 〈𝛿 , 𝛽 〉
⎛ 〈𝛿 , 𝛼 〉 〈𝛿 , 𝛼 〉 ⋯ 〈𝛿 , 𝛼 〉⎞ ⎛ 〈𝛿 , 𝛽 〉 〈𝛿 , 𝛽 〉 ⋯ 〈𝛿 , 𝛽 〉 ⎞
𝐀𝟏 = ⎜ and 𝐂𝟏 = ⎜
⎟; and ⎟.
⋮ ⋮ ⋱ ⋮ ⋮ ⋮ ⋱ ⋮
〈𝛿 , 𝛼 〉 〈𝛿 , 𝛼 〉 ⋯ 〈𝛿 , 𝛼 〉 〈𝛿 , 𝛽 〉 〈𝛿 , 𝛽 〉 ⋯ 〈𝛿 , 𝛽 〉
⎝ ⎠ ⎝ ⎠
Similarly, we for i = 2, . . . , n1 , which are f2 (t2 ),
obtain the regression functionsSimilarly,
f3 (t3 ), 〈𝛿
. . . ,,𝛽fp 〉 t p 〈𝛿 〉 ⋯ 〈𝛿 , 𝛽 〉
, as, 𝛽follows: we obtain the regression functions for 𝑖 =
⎛ 〈𝛿 , 𝛽 〉 〈𝛿 , 𝛽 〉 ⋯ 〈𝛿 , 𝛽 〉 ⎞ 𝐟𝟑 (𝑡 ), …, 𝐟𝒑 𝑡 , as follows:
𝐂𝟏 = ⎜ ⎟.
⋮ ⋮ ⋱ ⋮ T
f 2 ( t 2 ) = f 2 ( t 11 ) , f 2 ( t 12 ) , . . . , f 2 t 1n = A 2 b2𝐟𝟐+(𝑡C)2= d2 𝑓 (𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 ) =𝐀
〈𝛿 , 𝛽 〉 〈𝛿 , 𝛽 〉 ⋯ 〈𝛿 , 𝛽 〉 2
⎝ ⎠ T
f3 (t3 ) = f 3 (t11 ), f 3 (t12 ), . . . , f 3 t1n3 = A3 b3𝐟𝟑+ (𝑡 C) 3=d3𝑓 (𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 ) =𝐀
Similarly, we obtain the regression functions .
.. for 𝑖 = 2, … , 𝑛 , which are 𝐟𝟐 (𝑡 ), ⋮(13)
𝐟𝟑 (𝑡 ), …, 𝐟𝒑 𝑡 , as follows: 𝐟 𝑡 = 𝑓 (𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 ) =𝐀
T 𝒑
fp t p = f p (t11 ), f p (t12 ), . . . , f p t1n p = Ap bp + Cp dp
𝐟𝟐 (𝑡 ) = 𝑓 (𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 )
= 𝐀Hence,
𝟐 𝐛𝟐 + 𝐂based
𝟐 𝐝𝟐 ⎫ on Equations (12) and (13), we obtain th
the MNR model ⎪
as follows:
Hence, based 𝐟𝟑 (𝑡on
)= 𝑓 (𝑡 ), 𝑓 (𝑡
Equations (12)),and
… , 𝑓 (13),
(𝑡 )we=obtain
𝐀𝟑 𝐛𝟑 +the𝐂𝟑 𝐝regression
𝟑 function f(t) of the
(13)
⋮ = 𝐟 (𝑡 ), 𝐟 (𝑡 ), … , 𝐟 (𝑡⎬) 𝑻 = 𝐀 𝐛 , 𝐀 𝐛 , … , 𝐀 𝐛 𝑻 + 𝐂 𝐝 , 𝐂 𝐝
𝐟(𝑡)
MNR model as follows: 𝟏 𝟐 𝟐 𝒑
⎪
𝟏 𝟏 𝟐 𝟐 𝒑 𝒑 𝟏 𝟏 𝟐
𝐟𝒑 𝑡 = 𝑓 (𝑡 ), 𝑓 (𝑡 ), … , 𝑓 (𝑡 ) = 𝐀𝒑 𝐛𝒑 + 𝐂𝒑 𝐝𝒑 ⎭
T T = 𝑑𝑖𝑎𝑔 𝐀𝟏 , 𝐀𝟐 , … , 𝐀𝒑 𝐛T𝟏 , 𝐛𝟐 , … , 𝐛𝒑 +
f( t ) = f1 (t1 ), f2 (t2Hence, p t p on Equations
), . . . , fbased = A1 b1(12)
, A2 and
b2 , .(13),
. . , Awe
p bobtain
p + the d1 , C2 d2 ,function
C1regression . . . , Cp d𝐟(t)
p of
the MNR model as follows:
T
𝑑𝑖𝑎𝑔 𝐂 , 𝐂 , … , 𝐂 𝐝 , 𝐝 , … , 𝐝 = 𝐀𝐛 + 𝐂𝐝
= diag A1 , A2 , . . . , Ap b1 , b2 , . . . , bp + 𝟏 𝟐 𝒑 𝟏 𝟐 𝒑 (14)
𝑻 𝑻 𝑻
𝐟(𝑡) = 𝐟𝟏 (𝑡 ), 𝐟𝟐 (𝑡𝟐 ),
… , 𝐟𝒑 (𝑡 ) = 𝐀𝟏 𝐛𝟏, 𝐀𝟐 𝐛𝟐 , … , 𝐀𝒑 𝐛𝒑 +T 𝐂𝟏 𝐝𝟏 , 𝐂𝟐 𝐝𝟐Thus,
, … , 𝐂𝒑we
𝐝𝒑 can express the MNR model presented in (1)
diag C1 , C2 , . . . , Cp d1 , d2 , . . . , dp = Ab + Cd
= 𝑑𝑖𝑎𝑔 𝐀𝟏 , 𝐀𝟐 , … , 𝐀𝒑 𝐛𝟏 , 𝐛𝟐 , … , 𝐛𝒑 + (14) 𝐲 = 𝐀𝐛 + 𝐂𝐝 + 𝛆
Thus, we can express the MNR model presented in (1) in matrix notation as follows:
where 𝐀 = 𝑑𝑖𝑎𝑔(𝐀𝟏 , 𝐀𝟐 , … , 𝐀𝐩 ) is an (𝑁 × 𝑀) diagonal m
𝑑𝑖𝑎𝑔 𝐂𝟏 , 𝐂𝟐 , … , 𝐂𝒑 𝐝𝟏 , 𝐝𝟐 , … , 𝐝𝒑 = 𝐀𝐛 + 𝐂𝐝 𝐓
y = Ab + Cd ∑ +𝑚ε ; 𝐛 = 𝐛𝐓𝟏 , 𝐛𝐓𝟐 , … , 𝐛𝐓𝐩 is an (15) (𝑀 × 1) vec
Thus, we can express the MNR model presented in𝑑𝑖𝑎𝑔 𝟏 , 𝐂𝟐 , … notation
(1) in𝐂matrix , 𝐂𝐩 is an (𝑁 × 𝑁) diagonal matrix; and 𝐝
as follows:
1) vector of parameters. p
where A = diag A1 , A2 , . . . , A𝐲p= 𝐀𝐛 is + an𝐂𝐝(+N𝛆× M )Now, (15) N = ∑ nr ,
diagonal
we canmatrix
determinewithan estimated smoothing spli
r =1
where 𝐀 = 𝑑𝑖𝑎𝑔(𝐀𝟏 , 𝐀𝟐 , … , 𝐀𝐩 ) is an (𝑁 × 𝑀) diagonal T MNR matrix
model 𝑁 = ∑ in
withpresented 𝑛 (1)
, 𝑀using
= the RKHS approach b
p T T T
M =𝑚∑r=
∑ ; 1m ; 𝐛b𝐓𝟏 , 𝐛𝐓𝟐 , =
𝐛 r=
𝐓 , b2 an
… , 𝐛𝐓𝐩 b1 is , . . . , (𝑀
bp × 1) is M ×parameters;
an (optimization:
following
vector of 1) vector 𝐂of = parameters;
𝑁) diagonal matrix; and 𝐝 = 𝐝𝐓𝟏 , 𝐝𝐓𝟐 , … , 𝐝𝐓𝐩 is an (𝑁 𝐓 T
𝑑𝑖𝑎𝑔 𝐂𝟏 , 𝐂𝟐 , … , 𝐂𝐩 is an (𝑁 × ×T
C = diag C1 , C2 , . . . , Cp is an ( N × N ) diagonal matrix; and d = Min dT T
1 , d2𝐖, . .𝛆. , dp
= Minis 𝐖 (
1) vector of parameters. ∈ ∈
N × 1we
an (Now, ) vector of parameters.
can determine an estimated smoothing spline regression function of the , ,…, , ,…,
MNR Now,
model we can determine
presented in (1) usinganthe
estimated smoothing
RKHS approach
withby spline
taking
a constraint the regression
( )
solution
𝑓 (𝑡of) the function
𝑑𝑡 < 𝛾 , whereof the𝛾 ≥ 0 and
following optimization:
MNR model presented in (1) using the RKHS approach by taking the solution of the
following optimization:
Min 𝐖 𝛆 = Min 𝐖 (𝐲 − 𝐟) (16)
∈ ∈
, ,…, 1 2 , ,…, 1
Min {k W ε k } = 2 Min {k W (y − f) k 2 } 2 (16)
( )
with a constraint f r 𝑓∈ F(𝑡r ) 𝑑𝑡 < 𝛾 , where 𝛾 ≥ 0 fand
r ∈F 𝑟 r= 1,2, … , 𝑝.
r = 1, 2, . . . , p r = 1, 2, . . . , p
R b (m) 2
with a constraint arr f r (tr ) dtr < γr , where γr ≥ 0 and r = 1, 2, . . . , p.
Note that determining the solution to Equation (16) is equivalent to determining the
solution to the following PWLS optimization:
Z br 2
p
(y − f) W(y − f) + ∑
−1 T (m)
Min N r =1
λr fr ( tr ) dtr (17)
f r ∈ W2m [ ar , br ] ar
r = 1, 2, . . . , p
p
where N = ∑r=1 nr ; N −1 (y − f)T W(y − f) is a weighted least square that represents the
R b (m) 2
p
goodness of fit, ∑r=1 λr arr f r (tr ) dtr represents a penalty that measures smoothness,
and λr , r = 1, 2, . . . , p represents a smoothing parameter which controls the trade-off
between the goodness of fit and the penalty.
Symmetry 2022, 14, 2227 8 of 22
Based on Equations (19) and (20), we can express the PWLS optimization presented in
(17) as follows:
The solution to (21) can be obtained by taking the partial diferentiation Q(b, d) with
respect to b and d. In this step, we obtain the estimations of b and d as follows:
^ −1 ^
−1
b = AT D−1 WA AT D−1 Wy, and d = D−1 W I − A AT D−1 WA AT D − 1 W y
where D = WC + NΛI.
From this step, we obtain the estimation of the smoothing spline regression function
of the MNR model presented in (1) or (15) as follows:
^ ^
f̂ = Ab + Cd = Hy (22)
−1 −1
where H = A AT D−1 WA AT D−1 W + CD−1 W I − A AT D−1 WA AT D − 1 W ,
D = WC + NΛI, Λ = diag λ1 In1 , λ2 In2 , . . . , λ p In p , I is an identity matrix with dimension
p
N, and N = ∑r=1 nr .
p
N = ∑r=1 nr ) normally distributed random sample with mean f and covariance W−1 , we
have the following likelihood function:
1
1 T
n
L( f, W|y) = ∏ j=1 exp − y − f j W y − f j (23)
N
1 2 j j
(2π ) 2 W−1 2
p
Because N = ∑r=1 nr and W = diag W1 , W2 , . . . , Wp , the likelihood function pre-
Next, based on (24), the estimated weight matrix can be obtained by carrying out the
following optimization:
( )
T
L( f, W|y) = Max nn1
1
n exp − 12 ∑nj=1 y1j − f1j W1 , . . . y1j − f1j ×
W1 (2π ) 2 (W1−1 ) 2
( )
T
Max nn2
1
n exp − 12 ∑nj=1 y2j − f2j W2 , . . . y2j − f2j ×···× (25)
W2 (2π ) 2 (W2−1 ) 2
( )
T
Max nn p
1
n exp − 12 ∑nj=1 ypj − fpj Wp ypj − fpj
Wp (2π ) 2 (W−p 1 ) 2
According to [45], the maximum value of each component of the likelihood function
in Equation (25) can be determined using the following equations:
T T T
^T
^ ^T y1 − f̂1 y1 − f̂1 ^ ^T y2 − f̂2 y2 − f̂2 ^
yp − f̂p yp − f̂p
^ ε1 ε1 ^ ε2 ε2 ^ εp εp
W1 = = , W2 = = , . . . , Wp = = .
N N N N N N
We may express the estimated smoothing spline regression function presented in (22)
as follows:
f̂ λ, σ2 = H λ, σ2 y (26)
T T
where λ = λ1 , λ2 , . . . , λ p , and σ2 = σ12 , σ22 , . . . , σp2 .
Hence, the maximum likelihood estimator for the weight matrix W is provided by:
T T T
^
^ ^ ^
(y1 −f̂1 )(y1 −f̂1 ) (y2 −f̂2 )(y2 −f̂2 ) (yp −f̂p )(yp −f̂p )
W = diag W1 , W2 , . . . , Wp = diag N , N , ..., N
T T T
(In1 −H(λ1 ,σ21 ))y1 yT1 (In1 −H(λ1 ,σ21 )) (In1 −H(λ1 ,σ21 ))y1 yT1 (In1 −H(λ1 ,σ21 )) (Inp −H(λp ,σ2p ))yp yTp (Inp −H(λp ,σ2p ))
= diag N , N , . . . , N .
This shows that the estimated weight matrix obtained above is a symmetric matrix,
specifically, a diagonal matrix the main diagonal components of which are the estimated
weight matrices of the first response, second response, etc., up to the p-th response.
T T
where λ = λ1 , λ2 , . . . , λ p , σ2 = σ12 , σ22 , . . . , σp2 . The mean squared error (MSE) of the
estimated smoothing spline regression function presented in (26) is provided by
T T
(y−f̂(λ,σ2 )) W(y−f̂(λ,σ2 )) (y−f̂(λ,σ2 )) W(y−f̂(λ,σ2 ))
MSE = p = N
∑ r =1 nr
T
((I−H(λ,σ2 ))y) W(I−H(λ,σ2 ))y
= N
1 2
kW 2 (I−H(λ,σ2 ))yk
= N
Therefore, based on (27), we can obtain the optimal smoothing parameter value,
T
λopt = λ1;opt , λ2;opt , . . . , λ p;opt , by taking the solution of the following optimization:
Gopt λopt = Min { G (λ)}
λ1 ,...,λ p ∈R+
2
(28)
( 1
)
N −1 kW 2 (I−H(λ,σ2 ))yk
= Min 1 2
λ1 ,...,λ p ∈R+ ( N trace(IN −H(λ,σ2 )))
p
where R+ represents a positive real number set and N = ∑r=1 nr .
T
Thus, the optimal smoothing parameter value λopt = λ1;opt , λ2;opt , . . . , λ p;opt is
obtained from the minimizing process of the function G (λ) in (27). The function G (λ) in
(27) is called the generalized cross-validation function [1].
where n = 100 and with correlations ρ12 = 0.6, ρ13 = 0.7, ρ23 = 0.8 and variances σ12 = 2,
σ22 = 3, σ32 = 4. Based on the results of this simulation, we obtain a minimum generalized
cross-validation (GCV) value of 2.526286 and three optimal smoothing parameter values,
Symmetry 2022, 14, 2227 11 of 22
which are λ1(opt) = 2.146156 × 10−7 (for the first response), λ2(opt) = 1.084013 × 10−7 (for
the second response), and λ3(opt) = 5.930101 × 10−8 (for the third response).
Next, we present an illustration of the effects of the smoothing parameters on the estima-
tion results of the MNR Model by comparing three kinds of different smoothing parameter val-
ues, namely, λ1(small ) = 10−10 , λ2(small ) = 2 × 10−10 , and λ3(small ) = 3 × 10−10 , which repre-
sent small smoothing parameter values; λ1(opt) = 2.146156 × 10−7 , λ2(opt) = 1.084013 × 10−7 ,
and λ3(opt) = 5.930101 × 10−8 , which represent optimal smoothing parameter values; and
λ1(large) = 10−5 , λ2(large) = 2 × 10−5 , and λ3(large) = 3 × 10−5 , which represent large
smoothing parameter values. In the following table and figures, we provide the results of this
simulation study.
Table 1 shows that the smoothing parameter values of 2.146156 × 10−7 , 1.084013 × 10−7 ,
and 5.930101 × 10−8 are the optimal smoothing parameter values, as these smoothing pa-
rameters have the lowest GCV value (2.526286) among all the others. Thus, according to
(28), these smoothing parameter values are the optimal smoothing parameters. We can write
them as λ1(opt) = 2.146156 × 10−7 , λ2(opt) = 1.084013 × 10−7 , and λ3(opt) = 5.930101 × 10−8 .
The optimal smoothing parameters provide the best estimation results for the MNR model
presented in (29).
Table 1. Comparison estimation results of MNR Model in (29) for three kinds of smoothing parame-
ter values.
n = 100; ρ12 = 0.6; ρ13 = 0.7; ρ23 = 0.8; σ12 = 2; σ22 = 3; σ32 = 4
Smoothing Parameters Minimum Values of GCV Results of Estimation
λ1(small ) = 10−10
λ2(small ) = 2 × 10−10 (Small Values) 4.763234 The estimation results are too rough.
λ3(small ) = 3 × 10−10
λ1(opt) = 2.146156 × 10−7
λ2(opt) = 1.084013 × 10−7 (Optimal Values) 2.526286 The estimation results are the best.
= 5.930101 × 10−8
λ3(opt)
λ1(large) = 10−5
λ2(large) = 2 × 10−5 ( Large Values) 4.617504 The estimation results are too smooth.
λ3(large) = 3 × 10−5
The plots of the estimated regression function of the MNR model presented in (29) for
the three different smoothing parameters are shown in Figures 1–3.
Figure 1 shows that for all responses the small smoothing parameter values provide
estimates of the regression functions of the MNR model presented in (29) that are too
rough, namely, (y1 ) for the first response, (y2 ) for the second response, and (y3 ) for the
third response.
Figure 2 shows that the optimal smoothing parameter values provide the best estimates
of the regression functions of the MNR model presented in (29) for all responses, namely,
(y1 ) for the first response, (y2 ) for the second response, and (y3 ) for the third response.
Figure 3 shows that for all responses the large smoothing parameter values provide
estimates of the regression functions of the MNR model presented in (29) that are too
smooth, namely, (y1 ) for the first response, (y2 ) for the second response, and (y3 ) for the
third response.
𝜆 ( = 10
)
𝜆 ( ) = 2 × 10 (𝐿𝑎𝑟𝑔𝑒 𝑉𝑎𝑙𝑢𝑒𝑠) 4.617504 The estimation results are too smooth.
𝜆 ( ) = 3 × 10
Symmetry 2022, 14, 2227 The plots of the estimated regression function of the MNR model presented 12
inof(29)
22
for the three different smoothing parameters are shown in Figures 1–3.
Plotting T1 v er sus Y1
8
6
y1
4
2
t1
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Plotting T2 v er sus Y2
8
6
4
y2
2
0
0.0 0.5 1. 0 1. 5 2. 0
t2
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Plotting T3 v er sus Y3
6
4
2
y3
0
-2
t3
Symmetry 2022, 14, 2227 Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e 12 of 21
Figure 1. Plots of estimated MNR Models in (29) for small smoothing parameters.
Figure 1. Plots of estimated MNR Models in (29) for small smoothing parameters.
Plotting T1 versus Y1
8
6
y1
4
2
t1
Point:Observation Point dan Solid Line: Fitted Curve
Plotting T2 versus Y2
5
3
y2
1
-1
t2
Point:Observation Point dan Solid Line: Fitted Curve
Plotting T3 versus Y3
4
2
y3
0
-2
t3
Point:Observation Point dan Solid Line: Fitted Curve
Figure 2. Plots of estimated MNR Model in (29) for optimal smoothing parameters.
Figure 2. Plots of estimated MNR Model in (29) for optimal smoothing parameters.
Plotting T1 v er sus Y1
7 .0
6 .0
y1
5 .0
4 .0
t1
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Plotting T2 v er sus Y2
5 .0
4 .0
y2
3 .0
2 .0
0.0 0.5 1. 0 1. 5 2. 0
t2
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Plotting T3 v er sus Y3
2 .5
1 .5
y3
0 .5
0
-2
0.0 0.5 1.0 1.5 2.0
t3
Symmetry 2022, 14, 2227 Point:Observation Point dan Solid Line: Fitted Curve 13 of 22
Figure 2. Plots of estimated MNR Model in (29) for optimal smoothing parameters.
Plotting T1 v er sus Y1
7 .0
6 .0
y1
5 .0
4 .0
t1
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Plotting T2 v er sus Y2
5 .0
4 .0
y2
3 .0
2 .0
0.0 0.5 1. 0 1. 5 2. 0
t2
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Plotting T3 v er sus Y3
2 .5
1 .5
y3
0 .5
t3
Point :Obs ervat ion Point dan S olid Line: Fit ted C urv e
Figure 3. Plots of estimated MNR Models in (29) for large smoothing parameters.
Figure 3. Plots of estimated MNR Models in (29) for large smoothing parameters.
Figure 1 shows
3.5. Investigating that for allofresponses
the Consistency the Smoothing the small Splinesmoothing
Regressionparameter values provide
Function Estimator
estimates of the regression functions of the MNR model presented in (29) that are too
We first investigate the asymptotic properties of the smoothing spline regression
rough, namely,
function ) for the
estimator(𝑦f̂ based firstintegrated
on the response,mean (𝑦 ) for squarethe error second response,
(IMSE) andWe
criterion. (𝑦 develop
) for the
third response.
the IMSE proposed by [14] from the uniresponse case to the multiresponse case. Suppose
that weFigure 2 showsthe
decompose that
IMSEthe into
optimal smoothing parameter
two components, bias2 (λ) and values
Varprovide the best esti-
(λ), as follows:
mates of the regression functions of the MNR model presented in (29) for all responses,
namely, (𝑦 ) for the first response, (𝑦
Z b
) Tfor
the second response, and (𝑦 ) for the third re-
sponse. I MSE ( λ ) = E f ( t ) − f̂ ( t ) W f ( t ) − f̂ ( t ) dt = bias2 (λ) + Var (λ) (30)
a
Figure 3 shows that for all responses the large smoothing parameter values provide
estimates of the
2 (λregression functions of R btheMNR modelTpresented
inf̂((29)
that areand
too
where bias ) = a E [ f ( t ) − E f̂ ( t ) W f ( t ) − E t ) ]dt,
smooth, namely,
Rb
(𝑦 ) for thefirstresponse,
T
(𝑦 )
for the second response, and (𝑦 ) for the
third
Var = a E Ef̂(t) − f̂(t) W Ef̂(t) − f̂(t) dt. Furthermore, in order to investigate
(λ)response.
the asymptotic property of bias2 (λ), we assign the solution to PWLS optimization in the
following theorem.
Theorem 1. If f̂(t) is the solution that minimizes the following penalized weighted least
square (PWLS):
Z br 2
p
(y − g(t)) W(y − g(t)) + ∑
−1 T (m)
N r =1
λr gr ( tr ) dtr (31)
ar
then the solution that minimizes the following penalized weighted least square (PWLS):
Z br 2
p
N −1 (f(t) − g(t))T W(f(t) − g(t)) + ∑r=1 λr
(m)
gr ( tr ) dtr (32)
ar
Proof of Theorem 1. In Section 3.1, we obtained the solution to the PWLS in (31), that
^ ^ −1
is, f̂ = Ab + Cd = Hy, where as provided in (22), H = A ATD−1WA ATD−1W +
−1
CD−1W I − A ATD−1WA ATD−1W , D = WC + NΛ Λ, Λ = diag λ1In1 , λ2In2 , . . . , λ p Inp ,
Symmetry 2022, 14, 2227 14 of 22
p
N = ∑r=1 nr , and I is an identity matrix. Next, if we substitute y = f(t) into Equation (31), we
find that the value that minimizes PWLS (32) is
−1 −1
ĝ∗ (t) = A AT D−1 WA AT D−1 W + CD−1 W I − A AT D−1 WA AT D−1 W y = Ef̂(t)
Furthermore, we investigate the asymptotic property of bias2 (λ). For this purpose, we
first provide the following assumptions.
Assumptions (A):
(A1) For every r = 1, 2, . . . , p, nr = n and λr = λ.
−1 .
(A2) For every i = 1, 2, . . . , n, ti = 2i2n
(A3) For any continuous function g and 0 < Wi = W < ∞, i = 1, 2, . . . , n, the following
statements are satisfied [48,49]:
Rb
(a) n−1 ∑in=1 g(ti ) → a g(t)dt , as n → ∞ .
Rb
(b) n−1 ∑in=1 Wi g(ti ) → a Wg(t) dt, as n → ∞ .
R b (m) 2 Rb
(c) n−1 ∑in=1 λi a i gi (ti ) dti → λ a g(m) (t)dt , as n → ∞ .
i
Next, given Assumptions (A), the asymptotic property of bias2 (λ) is provided in the
Theorem 2.
Proof of Theorem 2. Suppose ĝ(t) is the value which minimizes the following penalized
weighted least square (PWLS):
Z b Z br 2
p
(f(t) − g(t))T W(f(t) − g(t))dt + ∑r=1 λr
(m)
gr ( tr ) dtr
a ar
Hence, we obtain ĝ∗ (t) = Ef̂(t) ≈ ĝ(t). Thus, for every g ∈ W2m [ a, b] we have
Rb T
Bias2 (λ) = a E[ f(t) − Ef̂(t) W f(t) − Ef̂(t) ]dt
T
Rb p R b (m)
≤ a E[ f(t) − Ef̂(t) W f(t) − Ef̂(t) ]dt + ∑r=1 λr a ĝr (tr )dtr
where O(·) represents “big oh”. Details about “big oh” can be found in [14,50].
Thus, Theorem 2 is proved.
Furthermore, the asymptotic property of Var (λ) is provided in Theorem 3.
!
1
Theorem 3. If Assumptions (A) hold and r = 1, 2, . . . , p, then Var (λr ) ≤ O 1 .
n λr2m
Proof of Theorem 3. To investigate the asymptotic property of Var (λ), we first define the
following function:
p
ω f̂, h, α = R f̂ + αh + ∑r=1 λr Jr f̂ + αh
R bh i2
(m)
where R(g) = n−1 (y − g(t)) T W(y − g(t)); Jr (g) = a gr (tr ) dtr ; h ∈ W2m [ a, b] and
α ∈ R.
∂ω(f̂,h,α)
Next, by taking the solution to ∂α = 0, for α = 0 we have
Z br 2
p
∑
(m) (m)
n−1 (y − f(t)) T Wg(t) = r =1
λr fr ( t r ) gr ( t r ) dtr
ar
n 1
f̂(t) = γ (t)β̂ = ∑ j=1 1 + nλθ j γ Tj y γ j (t) (38)
Symmetry 2022, 14, 2227 16 of 22
= ∑nj=1 1
1+nλr θrj ∑nk=1 γrj (trk )yrk γrj (trk )
Theorem 4. If f̂(t) is a smoothing spline estimator for regression function f(t) of the MNR model
presented in (1), then f̂(t) is a consistent estimator for f(t) based on the integrated mean square
error (IMSE) criterion.
Theorem 4. If 𝐟(𝐭) is a smoothing spline estimator for regression function 𝐟(𝐭) of the MNR
Symmetry 2022, 14, 2227 17 of 22
model presented in (1), then 𝐟(𝐭) is a consistent estimator for 𝐟(𝐭) based on the integrated mean
square error (IMSE) criterion.
Because P𝑃(|𝐼𝑀𝑆𝐸(𝛌)|
Because > δ𝛿)
(| I MSE(λ)| > )== 11−−𝑃(|𝐼𝑀𝑆𝐸(𝛌)|
P(| I MSE(λ)|≤ ≤
𝛿),δ)based onon
, based Equation (40)(40)
Equation andand
ap-
plying the probability properties we have
applying the probability properties we have
Lim(1 − 𝑃(|𝐼𝑀𝑆𝐸(𝛌)| ≤ 𝛿)) = 0 ⇔ 1 − Lim 𝑃(|𝐼𝑀𝑆𝐸(𝛌)| ≤ 𝛿) = 0
Lim (1→− P(| I MSE(λ)| ≤ δ)) = 0 ⇔ 1 − Lim
→ P (| I MSE (λ)| ≤ δ ) = 0
n→∞ n→∞ (41)
(41)
⇔⇔LimLim 𝑃(|𝐼𝑀𝑆𝐸(𝛌)|
P(| I MSE(λ)| ≤≤δ𝛿)) ==1.1.
→
n→∞
Equation (41) means that the smoothing spline regression function estimator of the
Equation (41)
multiresponse means that the
nonparametric smoothing
regression (MNR)spline regression
model function
is a consistent estimator
estimator of the
based on
multiresponse nonparametric regression (MNR) model is a consistent estimator
the integrated mean square error (IMSE) criterion. Thus, Theorem 4 is proved. □ based on
the integrated mean square error (IMSE) criterion. Thus, Theorem 4 is proved.
3.6. Illustration of Theorems
3.6. Illustration of Theorems
observation (𝑦 ) follows the multiresponse nonparametric re-
Supposeaapaired
Suppose pairedobservation
(yri , ,t𝑡ri ) follows the multiresponse nonparametric re-
gression(MNR)
gression (MNR)model:
model:
𝑦 = 𝑓 (𝑡 ) + 𝜀 , 𝑡 ∈ [0,1], 𝑟 = 1,2, … , 𝑝, 𝑖 = 1,2, … , 𝑛 . (42)
yri = f r (tri ) + ε ri , tri ∈ [0, 1], r = 1, 2, . . . , p, i = 1, 2, . . . , nr . (42)
Next, for every 𝑟 = 1,2, … , 𝑝, we assume 𝑓 = 𝑓 = ⋯ = 𝑓 = 𝑓 and 𝜀 = 𝜀 = ⋯ =
𝜀 = 𝜀 such
Next, that 𝑦r =
for every = 1,
𝑦 2,=. .⋯.,= p, 𝑦
we=assume𝑦 . Hence, f 1 = wef 2 =have
. . . =af pnonparametric
= f and ε 1 = εregression
2 = ... =
=
εmodel
p ε such that y
as follows: 1 = y 2 = . . . = y p = y. Hence, we have a nonparametric regression
model as follows:
yi =𝑦 f= ) + )ε i+, t𝜀i ,∈𝑡 [∈
(ti𝑓(𝑡 0, [0,1],
1], i = 𝑖=1, 1,2,
2, . .…. ,,n.
𝑛. (43)
(43)
Basedon
Based onthe
themodel
modelpresented
presentedinin(43), (43),we wepresent
presentan anillustration
illustrationrelated
relatedtotothe
thefour
four
theoremsininSection
theorems Section3.53.5through
through a simulation
a simulation study
study withwith sample
sample size of nof=𝒏200.
size = 𝟐𝟎𝟎. Based
Based on
𝟑 𝟑
on this model, let 𝒇(𝒕) =3 𝒔𝒊𝒏 (𝟐𝝅𝒕
3 ), where 𝒕 is generated
this model, let f(t) = sin 2πt , where t is generated from a uniform (0,1) distribution from a uniform (0,1) distribu-
tionεand
and 𝜺 is generated
is generated from a from
standard a standard
normal normal distribution.
distribution. The firstThe stepfirst
is tostep is to
create create
plots of
plots of the observation values (𝒕, 𝒚) and
the observation values (t, y) and f(t), as shown in Figure 4. 𝒇(𝒕), as shown in Figure 4.
Figure4.4.Plots
Figure Plotsofofobservation values(t,(𝒕,y)𝒚)and
observationvalues f(t𝒇(𝒕).
and ).
It can be seen from Figure 4 that there is a tendency towards y variance inequality. For
larger values of t, the y variance tends to be larger. Next, the data model is approximated
by a weighted spline with a weight of wi = 1/t2i , i = 1, 2, . . . , 200. The next step is to
determine the order, number, and location of the knot points. Here, we use a weighted
cubic spline model with two knot points, namely, 0.5 and 0.785. This weighted cubic spline
model can be written as follows:
Figure5.5.Plots
Plots of(a)(a)weighted
weightedcubic
cubic splinewith
withtwo
twoknot
knotpoints
pointsand curveofoff(𝒇(𝒕).
and (b)curve t).
Figure
Figure 5. Plots ofof
(a) weighted cubic splinespline
with two knot points and (b) curve of(b)
𝒇(𝒕).
Figure
Figure6.6.Plots
Figure ofofresiduals
6 shows
Plots and
that with
residuals weight
and 𝒘𝒊 = 𝟏
estimation
estimation ⁄𝒕𝟐𝒊 , the
values
values ofofweighted
variance ofcubic
weighted the spline.
response
cubic spline.variable y
tends to be the same. Meanwhile, Figure 7 provides a residual normality plot for the
shows𝟐2ino
weight w
Figure in Figurei 7=
Figure66shows shows that
that with
with weight 𝒘𝒊 = 𝟏1/t⁄𝒕𝒊 ,, the
the variance of the
the responsevariable
variabley
weighted cubic spline model. The plot indication towards deviation
variance of response
Symmetry 2022, 14, 2227 yfrom
tends
theto
tends normal
to bebe the
distribution.
the same. Meanwhile,
same. Meanwhile, Figure
Figure 7 provides
7 provides a residual normality
a residualmodelnormality plot
plot for
forthe
18 of 21
the
Next, as a comparison, we investigate a weighted cubic polynomial
weighted cubic spline model. The plot in Figure 7 shows no indication towards deviation with
weighted
weight 𝑤 =cubic
1⁄𝑡 , spline
𝑖 = 1,2,model. The
… ,200 for plotthe
fitting in Figure 7 shows
model (43). no indication
The fitting towards deviation
of the weighted
from the normal distribution.
frompolynomial
cubic the normal distribution.
model is shown in Figure 8. From the visualization in Figure 8, it can be
seen that this weighted
Next, cubic polynomial
as a comparison, we approach tendsatoweighted
investigate approach the function
cubic 𝑓(𝑡) very model with
polynomial
globally. This is in⁄contrast to the weighted cubic spline with two
weight 𝑤 = 1 𝑡 , 𝑖 = 1,2, … ,200 for fitting the model (43). The fitting knot points in (44), which
of the weighted
approaches the function 𝑓(𝑡) more locally. Thus, the weighted cubic spline model with two
cubic polynomial model is shown in Figure 8. From the visualization in Figure 8, it can be
knot points is adequate as an approximation model for the model presented in (43).
seen that this weighted cubic polynomial approach tends to approach the function 𝑓(𝑡) very
globally. This is in contrast to the weighted cubic spline with two knot points in (44), which
approaches the function 𝑓(𝑡) more locally. Thus, the weighted cubic spline model with two
knot points is adequate as an approximation model for the model presented in (43).
Figure7.7.Plot
Figure Plotofofweighted
weightedcubic
cubicspline
splineresidual
residualnormality.
normality.
Symmetry 2022, 14, 2227 19 of 22
Figure 8. Plots of (a) weighted cubic polynomial and (b) curve 𝑓(𝑡).
Figure8.8.Plots
Figure Plotsofof(a)
(a)weighted
weightedcubic
cubicpolynomial
polynomialand
and(b) curvef (𝑓(𝑡).
(b)curve t ).
Based on the illustration above and Figures 4–8, if 𝑓 (𝑡) is an estimator for model
Based
(43),Based
that is, ononifthe 𝑓illustration
the(𝑡)
illustration above
is the Penalized
above andand Figures
Weighted
Figures 4–8,4–8,
Least )𝑓is(𝑡)
if fˆ(iftSquares
anisestimator
an estimator
(PWLS) for model
solution,
for model then
(43),
∗ (𝑡) ˆ 𝑓 (𝑡) ∗ (𝑡)
𝑔(43),is,=
that if 𝐸(𝑦),
that f (tis, ifthe
) isfrom is the (44)
Equation
Penalized Penalized Weighted
is an estimator
Weighted Least forLeast
Squares model
(PWLS) Squares
(43) (PWLS)
as well,
solution, then ∗
suchsolution,
ĝthat 𝑔 Eithen
(th) = (y)=,
𝑔∗ (𝑡)
𝐸[𝑓
from (𝑡)],
= as𝐸(𝑦),
Equation from
provided
(44) isEquation
byanTheorem (44) 1.
estimator isfor
an model
estimator
(43)for
asmodel
well, such (43) as well,
that ĝ∗ (such E fˆ(𝑔
t) = that
∗
t) (𝑡)
, as=
𝐸[𝑓 (𝑡)],
providedThe as by provided
plots of theby
Theorem Theorem 1.curves of 𝑰𝑴𝑺𝑬(𝛌) , 𝑩𝒊𝒂𝒔𝟐 (𝛌) , and 𝑽𝒂𝒓(𝛌) , where
1.asymptotic
𝟐 𝟐
𝑰𝑴𝑺𝑬(𝛌) The =
The 𝑩𝒊𝒂𝒔ofof(𝛌)
plots
plots the
𝟐 (𝛌)
+ asymptotic
the 𝑽𝒂𝒓(𝛌), are curves
asymptotic shown in
curves 𝑰𝑴𝑺𝑬(𝛌)
of Figure
of IMSE9 [53]. (λ), (𝛌)
, 𝑩𝒊𝒂𝒔 Bias (λ), 𝑽𝒂𝒓(𝛌)
, 2and and Var , where
(λ),
𝑰𝑴𝑺𝑬(𝛌) = 𝑩𝒊𝒂𝒔 + 2 𝑽𝒂𝒓(𝛌), are shown
where IMSE(λ) = Bias (λ) + Var(λ), are shown in Figure 9 [53]. in Figure 9 [53].
Figure 9. Plots of asymptotic curves of (a) IMSE(λ), (b) Bias2 (λ ), and (c) Var(λ).
Figure 9. Plots of asymptotic curves of (a) 𝑰𝑴𝑺𝑬(𝛌), (b) 𝑩𝒊𝒂𝒔𝟐 (𝛌), and (c) 𝑽𝒂𝒓(𝛌).
𝟐
Figure 9. Plots
Figure of asymptotic
9 shows that thecurves 𝑰𝑴𝑺𝑬(𝛌),
of (a) Mean
Integrated Square Error(𝛌),
(b) 𝑩𝒊𝒂𝒔 𝑽𝒂𝒓(𝛌).represented by
and (c) curve
(IMSE(λ))
curve (a) is the sum of the quadratic bias (Bias2 (λ)) curve represented by curve (b) and
the variance curve (Var(λ)) represented by curve (c). It can be seen from Figure 9b that
Bias2 (λ)
Bias (λ) ≤ O(λ), as n → ∞ , that is, Lim Sup λ < ∞ [14,50,54]. This means that
2
n→∞
Bias2 (λ)
λ remains bounded, as n → ∞ , which is provided by Theorem 2. Furthermore,
−1
1
1
/λ < ∞ [14,50,54].
Figure 9c shows that Var (λ) ≤ O 1 (that is, Lim Sup nλ 2m
nλ 2m n→∞
−1
1
/λ remains bounded, as n → ∞ , which is provided by The-
This means that nλ 2m
Symmetry 2022, 14, 2227 20 of 22
orem 3. Furthermore, Figure 9 shows thatI MSE(λ) = Bias2 (λ) + Var (λ) ≤ O(nλ). Accord-
I MSE(λ)
ing to [14,50,54], this means that Lim Sup nλ < ∞. In other words, I MSE(λ) ≤ O(nλ)
n→∞
I MSE(λ)
if nλ remains bounded, as n → ∞ . According to [14,50,54], for any small positive
I MSE(λ)
number δ > 0 we have Lim P nλ > δ ≤ Lim P(| I MSE(λ)| > δ) = 0; hence,
n→∞ n→∞
Lim P(| I MSE(λ)| ≤ δ) = 1 is consistent.
n→∞
4. Conclusions
The smoothing spline estimator with the RKHS approach has good ability to estimate
the MNR model, which is a nonparametric regression model where the responses are
correlated with each other, because the goodness of fit and smoothness of the estimation
curve is controlled by the smoothing parameter, making the estimator very suitable for
prediction purposes. Therefore, selection of the optimal smoothing parameter value cannot
be omitted, and is crucial to good regression function fitting of the MNR model based on
smoothing spline estimator using the RKHS approach. The estimator of the smoothing
spline regression function of the MNR model that we obtained is linear with respect to the
observations in Equation (22), and is a consistent estimator based on the integrated mean
square error (IMSE) criterion. The main influence of this study is lies in the easier estimation
of a multiresponse nonparametric regression model where there is a correlation between
responses using the RKHS approach based on a smoothing spline estimator. This approach
is easier, and faster, and more optimal, as estimation is carried out simultaneously instead
of response by response for each observation. In addition, the theory generated in this study
can be used to estimate the nonparametric component of the multiresponse semiparametric
regression model used to model Indonesian toddlers’ standard growth charts. In the future,
the results of this study can be further developed within the scope of statistical inference,
especially for the purpose of testing hypotheses involving multiresponse nonparametric
regression models and multiresponse semiparametric regression models.
Author Contributions: All authors have contributed to this research article. Conceptualization, B.L.
and N.C.; methodology, B.L. and N.C.; software, B.L., N.C., D.A. and E.Y.; validation, B.L., N.C., D.A.
and E.Y.; formal analysis, B.L., N.C. and D.A.; investigation, resource and data curation, B.L., N.C.,
D.A. and E.Y.; writing—original draft preparation, B.L. and N.C.; writing—review and editing, B.L.,
N.C. and D.A.; visualization, B.L. and N.C.; supervision, B.L., N.C. and D.A.; project administration,
N.C. and B.L.; funding acquisition, N.C. All authors have read and agreed to the published version of
the manuscript.
Funding: This research was funded by the Directorate of Research, Technology, and Community
Service (Direktorat Riset, Teknologi, dan Pengabdian kepada Masyarakat–DRTPM), the Ministry
of Education, Culture, Research, and Technology, the Republic of Indonesia through the Featured
Basic Research of Higher Education Grant (Hibah Penelitian Dasar Unggulan Perguruan Tinggi–
PDUPT, Tahun Ketiga dari Tiga Tahun) with master contract number 010/E5/PG.02.00.PT/2022 and
derivative contract number 781/UN3.15/PT/2022.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: The authors confirm that there are no available data.
Acknowledgments: The authors thank Airlangga University for technical support and DRTPM
of the Ministry of Education, Culture, Research, and Technology, the Republic of Indonesia for
financial support. The authors are grateful to the editors and anonymous peer reviewers of the
Symmetry journal, who provided comments, corrections, criticisms, and suggestions that were useful
for improving the quality of this article.
Conflicts of Interest: The authors declare no conflict of interest. In addition, the funders had no role
in the design of the study, in the collection, analysis, or interpretation of data, in the writing of the
article, or in the decision to publish the results.
Symmetry 2022, 14, 2227 21 of 22
References
1. Wahba, G. Spline Models for Observational Data; SIAM: Philadelphia, PA, USA, 1990.
2. Kimeldorf, G.; Wahba, G. Some results on Tchebycheffian spline functions. J. Math. Anal. Appl. 1971, 33, 82–95. [CrossRef]
3. Cox, D.D. Asymptotics for M-type smoothing spline. Ann. Stat. 1983, 11, 530–551. [CrossRef]
4. Oehlert, G.W. Relaxed boundary smoothing spline. Ann. Stat. 1992, 20, 146–160. [CrossRef]
5. Ana, E.; Chamidah, N.; Andriani, P.; Lestari, B. Modeling of hypertension risk factors using local linear of additive nonparametric
logistic regression. J. Phys. Conf. Ser. 2019, 1397, 012067. [CrossRef]
6. Chamidah, N.; Yonani, Y.S.; Ana, E.; Lestari, B. Identification the number of mycobacterium tuberculosis based on sputum image
using local linear estimator. Bullet. Elect. Eng. Inform. 2020, 9, 2109–2116. [CrossRef]
7. Cheruiyot, L.R. Local linear regression estimator on the boundary correction in nonparametric regression estimation. J. Stat.
Theory Appl. 2020, 19, 460–471. [CrossRef]
8. Cheng, M.-Y.; Huang, T.; Liu, P.; Peng, H. Bias reduction for nonparametric and semiparametric regression models. Stat. Sin.
2018, 28, 2749–2770. [CrossRef]
9. Chamidah, N.; Zaman, B.; Muniroh, L.; Lestari, B. Designing local standard growth charts of children in East Java province using
a local linear estimator. Int. J. Innov. Creat. Chang. 2020, 13, 45–67.
10. Delaigle, A.; Fan, J.; Carroll, R.J. A design-adaptive local polynomial estimator for the errors-in-variables problem. J. Amer. Stat.
Assoc. 2009, 104, 348–359. [CrossRef]
11. Francisco-Fernandez, M.; Vilar-Fernandez, J.M. Local polynomial regression estimation with correlated errors. Comm. Statist.
Theory Methods 2001, 30, 1271–1293. [CrossRef]
12. Benhenni, K.; Degras, D. Local polynomial estimation of the mean function and its derivatives based on functional data and
regular designs. ESAIM Probab. Stat. 2014, 18, 881–899. [CrossRef]
13. Kikechi, C.B. On local polynomial regression estimators in finite populations. Int. J. Stat. Appl. Math. 2020, 5, 58–63.
14. Wand, M.P.; Jones, M.C. Kernel Smoothing; Chapman & Hall: London, UK, 1995.
15. Cui, W.; Wei, M. Strong consistency of kernel regression estimate. Open J. Stats. 2013, 3, 179–182. [CrossRef]
16. De Brabanter, K.; De Brabanter, J.; Suykens, J.A.K.; De Moor, B. Kernel regression in the presence of correlated errors. J. Mach.
Learn. Res. 2011, 12, 1955–1976.
17. Chamidah, N.; Lestari, B. Estimating of covariance matrix using multi-response local polynomial estimator for designing children
growth charts: A theoretically discussion. J. Phy. Conf. Ser. 2019, 1397, 012072. [CrossRef]
18. Aydin, D.; Güneri, Ö.I.; Fit, A. Choice of bandwidth for nonparametric regression models using kernel smoothing: A simulation
study. Int. J. Sci. Basic Appl. Res. 2016, 26, 47–61.
19. Eubank, R.L. Nonparametric Regression and Spline Smoothing, 2nd ed.; Marcel Dekker: New York, NY, USA, 1999.
20. Wang, Y. Smoothing Splines: Methods and Applications; Taylor & Francis Group: Boca Raton, FL, USA, 2011.
21. Liu, A.; Qin, L.; Staudenmayer, J. M-type smoothing spline ANOVA for correlated data. J. Multivar. Anal. 2010, 101, 2282–2296.
[CrossRef]
22. Gao, J.; Shi, P. M-Type smoothing splines in nonparametric and semiparametric regression models. Stat. Sin. 1997, 7, 1155–1169.
23. Chamidah, N.; Lestari, B.; Massaid, A.; Saifudin, T. Estimating mean arterial pressure affected by stress scores using spline
nonparametric regression model approach. Commun. Math. Biol. Neurosci. 2020, 2020, 1–12.
24. Fatmawati, I.N.; Budiantara, B.L. Comparison of smoothing and truncated spline estimators in estimating blood pressures models.
Int. J. Innov. Creat. Change 2019, 5, 1177–1199.
25. Chamidah, N.; Lestari, B.; Budiantara, I.N.; Saifudin, T.; Rulaningtyas, R.; Aryati, A.; Wardani, P.; Aydin, D. Consistency and
asymptotic normality of estimator for parameters in multiresponse multipredictor semiparametric regression model. Symmetry
2022, 14, 336. [CrossRef]
26. Eilers, P.H.C.; Marx, B.D. Flexible smoothing with B-splines and penalties. Statist. Sci. 1996, 11, 86–121. [CrossRef]
27. Lu, M.; Liu, Y.; Li, C.-S. Efficient estimation of a linear transformation model for current status data via penalized splines. Stat.
Meth. Medic. Res. 2020, 29, 3–14. [CrossRef]
28. Wang, Y.; Guo, W.; Brown, M.B. Spline smoothing for bivariate data with applications to association between hormones. Stat. Sin.
2000, 10, 377–397.
29. Yilmaz, E.; Ahmed, S.E.; Aydin, D. A-Spline regression for fitting a nonparametric regression function with censored data. Stats
2020, 3, 11. [CrossRef]
30. Aydin, D. A comparison of the nonparametric regression models using smoothing spline and kernel regression. World Acad. Sci.
Eng. Technol. 2007, 36, 253–257.
31. Lestari, B.; Budiantara, I.N.; Chamidah, N. Smoothing parameter selection method for multiresponse nonparametric regression
model using spline and kernel estimators approaches. J. Phy. Conf. Ser. 2019, 1397, 012064. [CrossRef]
32. Lestari, B.; Budiantara, I.N.; Chamidah, N. Estimation of regression function in multiresponse nonparametric regression model
using smoothing spline and kernel estimators. J. Phy. Conf. Ser. 2018, 1097, 012091. [CrossRef]
33. Osmani, F.; Hajizadeh, E.; Mansouri, P. Kernel and regression spline smoothing techniques to estimate coefficient in rates model
and its application in psoriasis. Med. J. Islam. Repub. Iran 2019, 33, 1–5. [CrossRef]
34. Lestari, B.; Budiantara, I.N. Spline estimator and its asymptotic properties in multiresponse nonparametric regression model.
Songklanakarin J. Sci. Technol. 2020, 42, 533–548.
Symmetry 2022, 14, 2227 22 of 22
35. Mariati, M.P.A.M.; Budiantara, I.N.; Ratnasari, V. The application of mixed smoothing spline and Fourier series model in
nonparametric regression. Symmetry 2021, 13, 2094. [CrossRef]
36. Wang, Y.; Ke, C. Smoothing spline semiparametric nonlinear regression models. J. Comp. Graphical Stats. 2009, 18, 165–183.
[CrossRef]
37. Lestari, B.; Chamidah, N. Estimating regression function of multiresponse semiparametric regression model using smoothing
spline. J. Southwest Jiaotong Univ. 2020, 55, 1–9.
38. Gu, C. Smoothing Spline ANOVA Models; Springer: New York, NY, USA, 2002.
39. Aronszajn, N. Theory of reproducing kernels. Transact. Amer. Math. Soc. 1950, 68, 337–404. [CrossRef]
40. Berlinet, A.; Thomas-Agnan, C. Reproducing Kernel Hilbert Spaces in Probability and Statistics; Kluwer Academic: Norwell, MA,
USA, 2004.
41. Wahba, G. Support vector machines, reproducing kernel Hilbert spaces and the randomized GACV. Adv. Kernel Methods-Support
Vector Learn. 1999, 6, 69–87.
42. Scholkopf, B.; Smola, A.J. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press:
Cambridge, MA, USA, 2002.
43. Zeng, X.; Xia, Y. Asymptotic distribution for regression in a symmetric periodic Gaussian kernel Hilbert space. Stat. Sin. 2019,
29, 1007–1024. [CrossRef]
44. Hofmann, T.; Scholkopf, B.; Smola, A.J. Kernel methods in machine learning. Ann. Stat. 2008, 36, 1171–1220. [CrossRef]
45. Johnson, R.A.; Wichern, D.W. Applied Multivariate Statistical Analysis; Prentice Hall: New York, NY, USA, 1982.
46. Li, J.; Zhang, R. Penalized spline varying-coefficient single-index model. Commun. Stat.-Simul. Comp. 2010, 39, 221–239. [CrossRef]
47. Ruppert, D.; Carroll, R. Penalized Regression Splines, Working Paper; School of Operation Research and Industrial Engineering,
Cornell University: New York, NY, USA, 1997.
48. Tunç, C.; Tunç, O. On the stability, integrability and boundedness analyses of systems of integro-differential equations with
time-delay retardation. RACSAM 2021, 115, 115. [CrossRef]
49. Aydin, A.; Korkmaz, E. Introduce Gâteaux and Frêchet derivatives in Riesz spaces. Appl. Appl. Math. Int. J. 2020, 15, 16.
50. Sen, P.K.; Singer, J.M. Large Sample in Statistics: An Introduction with Applications; Chapman & Hall: London, UK, 1993.
51. Eubank, R.L. Spline Smoothing and Nonparametric Regression; Marcel Dekker, Inc.: New York, NY, USA, 1988.
52. Speckman, P. Spline smoothing and optimal rate of convergence in nonparametric regression models. Ann. Stat. 1985, 13, 970–983.
[CrossRef]
53. Budiantara, I.N. Estimator Spline dalam Regresi Nonparametrik dan Semiparametrik. Ph.D. Dissertation, Universitas Gadjah
Mada, Yogyakarta, Indonesia, 2000.
54. Serfling, R.J. Approximation Theorems of Mathematical Statistics; John Wiley: New York, NY, USA, 1980.