Fuzzy Least Squares and Fuzzy Orthogonal Least Squares Linear
Regressions
Julien Rosset
a
and Laurent Donz
´
e
b
ASAM Group, University of Fribourg, Boulevard de P
´
erolles 90, 1700 Fribourg, Switzerland
Keywords:
Fuzzy Least Squares Linear Regression, Fuzzy Orthogonal Least Squares Linear Regression, Fuzzy Orthogo-
Nality, Fuzzy Logic, Fuzzy Methods in Data Analysis.
Abstract:
We examine the well known fuzzy least squares linear regression method. We discuss the constrained and
unconstrained solutions. Based on the concept of fuzzy orthogonality, we propose the fuzzy orthogonal least
squares method to solve fuzzy linear regression problems. We show that, in case of (fuzzy) orthogonal re-
gressors, an important property of the least squares method remains valid. We obtain the same estimates of
the parameters of the model if we regress on all regressors, or on each regressor considered separately. An
empirical application illustrates our methods.
1 INTRODUCTION
Fuzzy regression is no longer a new topic in fuzzy
analysis. Indeed, for decades, all kinds of propo-
sitions have been made to perform fuzzy regression
analyses. Focus has been put particularly on solving
least squares problems or least absolute deviations.
We intend to review some results and complete them
with the so-called orthogonal least squares method.
We especially investigate the fuzzy least squares li-
near regression and the orthogonal fuzzy least squares
linear regression. We provide essentially some strate-
gies to deal with fuzzy data in a regression context.
We proposed constrained and unconstrained solutions
and discussed them.
In some situations, the analyst could profit from
the orthogonality property of the independent vari-
ables. Indeed, by using a proper definition of fuzzy
orthogonal variables, we show an important feature
of the fuzzy orthogonal least squares method to solve
a linear regression problem. As in the classical case,
with crisp data, in a situation of orthogonal regressors,
the estimates of the model parameters are the same if
the estimation is done with all regressors of the model,
or if we regress the dependent variable on each re-
gressor alone. We verify empirically that, with our
fuzzy orthogonal least squares regression, this prop-
erty holds in a fuzzy context.
a
https://orcid.org/0000-0002-7883-1512
b
https://orcid.org/0000-0003-3522-4672
The main contribution of this paper is the proposi-
tion of a fuzzy orthogonal linear least squares regres-
sion method preserving the crisp orthogonal linear
least squares property in a fully fuzzy environment.
A nice feature of the method is to respect the fuzzy
arithmetic. Our work is organized as follow. We be-
gin our study with a small literature review in section
2. Notation is fixed in the next section, 3. Section 4 is
devoted to the problem of the fuzzy least squares re-
gression, proposing two methods to deal with crisp or
fuzzy input and fuzzy output. A discussion of the con-
strained and unconstrained solutions of the two meth-
ods is given in section 5. After briefly defining fuzzy
orthogonality, we present the method of the fuzzy or-
thogonal least squares dealing with crisp and fuzzy
input and fuzzy output in section 6. Section 7 gives us
the possibility to illustrate our methods by empirical
applications. Finally, section 8 concludes our study.
2 LITERATURE REVIEW
In 1982, Tanaka introduced a possibilistic approach
to fuzzy regression analysis. The method consists in
using possibilistic restrictions to minimize the fuzzi-
ness of the model’s fuzzy parameters. In this work,
the quadratic programming approach, which allows
both the minimisation of the estimated deviations of
the central tendency and the minimisation of the es-
timated deviations in the spreads of fuzzy observa-
Rosset, J. and Donzé, L.
Fuzzy Least Squares and Fuzzy Orthogonal Least Squares Linear Regressions.
DOI: 10.5220/0012182700003595
In Proceedings of the 15th International Joint Conference on Computational Intelligence (IJCCI 2023), pages 359-368
ISBN: 978-989-758-674-3; ISSN: 2184-3236
Copyright © 2023 by SCITEPRESS Science and Technology Publications, Lda. Under CC license (CC BY-NC-ND 4.0)
359
tions, will be of concern. (Tanaka and Lee, 1997)
studied the fuzzy linear regression model by means of
quadratic programming to minimise the distances be-
tween the estimated output centres and the observed
outputs while minimising the spreads of the estimated
outputs. (Tanaka and Lee, 1998) proposed an inter-
val regression analysis based on a quadratic program-
ming approach to deal with the problem of fuzzy co-
efficients becoming crisp when using linear program-
ming in possibilistic regression analysis. (Lee and
Tanaka, 1998) proposed a fuzzy regression analysis
based on a quadratic programming approach to again
integrate the central tendency of least squares and the
possibilistic properties of fuzzy regression. (Lee and
Tanaka, 1999) also explored a fuzzy linear regression
model with non-symmetric fuzzy coefficients using
quadratic programming and created a lower and up-
per approximation model. (Donoso et al., 2006) pro-
posed two new fuzzy regression models, the quadratic
possibilistic model and the quadratic non-possibilistic
model, which do not focus on the minimisation of the
uncertainty of the estimated results but on the min-
imisation of the quadratic deviations between the ob-
servations and estimated outputs. However, (Donoso
et al., 2006)’s regression models were only dealing
with fuzzy regressors.
To palliate this, (D’Urso and Massari, 2013) pro-
posed a general linear regression model for studying
the dependence of fuzzy response variable, on a set of
crisp or fuzzy explanatory variables. They also sug-
gested a robust fuzzy regression method, based on the
Least Median Squares estimation approach in an at-
tempt of neutralising the effects of crisp and fuzzy
outliers. In this direction, (Kashani et al., 2021) pro-
posed a penalized estimation method to estimate the
coefficients of a linear regression model with a fuzzy
response variable and crisp explanatory variables. (Li
et al., 2023) constructed a fuzzy multiple linear least
squares regression model based on two distance mea-
sures between LR-type fuzzy numbers. (Stanojevi
and Stanojevi, 2022) described a fuzzy quadratic least
squares regression for a fuzzy response variable and a
single crisp explanatory variable which gives regres-
sion coefficients with positive spreads.
Fuzzy inner product spaces and fuzzy orthog-
onality have been discussed by (Ithoh, 2017) and
(Mostofian et al., 2017). They proposed new defini-
tions of a fuzzy inner product space. They also de-
fined a suitable notion of fuzzy orthogonality in the
fuzzy world and investigated some properties. (Gior-
dani and Kiers, 2004) proposed two extensions of the
classical principal component analysis dealing with
fuzzy symmetric numbers. However, The lack of a
properly defined fuzzy orthogonality made the de-
rived results losing in significance. (Yabuuch and
Watada, 2017) performed a principal component anal-
ysis over crisp data belonging to fuzzy groups. In
their work (Yabuuch and Watada, 2017) introduced
new definitions of expectation, variance and covari-
ance to work with the concept of fuzzy groups. How-
ever, their results are limited to crisp input data.
We wanted to investigate more thoroughly the
fuzzy linear least square regression methods involving
both fuzzy response and explanatory variables since
these have not received much attention except from
(D’Urso and Massari, 2013). However, due to the
complexity of their recursive approach, constructing a
fuzzy orthogonal linear least squares method inspired
from it seemed to be a difficult task. Because we
wanted to preserve certain properties, our approach
was to build a fuzzy linear least squares regression
method that could be, in accordance with the concept
of fuzzy orthogonality, extended to a fuzzy orthogo-
nal linear least squares method allowing the indivi-
dual computation of the estimates.
3 NOTATION
Let us define by ˜x a fuzzy number. We write by
µ
˜x
(·), the membership function. We consider also
the α-cuts of ˜x denoted by ˜x
α
or by its equivalent
in interval form by [x
L,α
, x
R,α
]. In practice, triangu-
lar fuzzy numbers are often used. We denote them
by a triplet ˜x = (x
L
, x, x
R
) with x
L
x x
R
R. In-
dexed fuzzy triangular number will be denoted by
˜x
k
= (x
L
k
, x
k
, x
R
k
). If not specified otherwise, lower-
case bold letters with no index will be used for n-
components column vectors of fuzzy numbers, e.g.
˜y = (˜y
1
, . . . , ˜y
n
)
. A (n × m)-matrix is noted in capital
bold letters, e.g.
˜
X, with the i-th line and j-th column
given respectively by ˜x
i
= ( ˜x
i1
, . . . , ˜x
im
)
, i = 1, . . . , n,
and ˜x
j
= ( ˜x
1 j
, . . . , ˜x
n j
)
, j = 1, . . . , m.
The fuzzy multiplication operator between two
fuzzy numbers is denoted by
e
. The fuzzy multipli-
cation between two fuzzy triangular numbers ˜a and
˜
b
is defined as
˜a
e
˜
b =(min(a
L
b
L
, a
L
b
R
, b
R
a
L
, a
R
b
R
),
ab, max(a
L
b
L
, a
L
b
R
, b
R
a
L
, a
R
b
R
)). (1)
See (Viertl, 2018) for more details.
4 FUZZY LEAST SQUARES
REGRESSION
In the following, several types of fuzzy regression
models will be presented. We decided to estimate the
FCTA 2023 - 15th International Conference on Fuzzy Computation Theory and Applications
360
parameters of the models by the least squares tech-
nique. We will first deal with two different cases.
In the first one, we will analyse a regression model
with crisp independent variables and a fuzzy depen-
dent variable. The second case will involve both fuzzy
independent and dependent variables. Then we will
discuss their constrained and unconstrained solutions.
As we are only interested in the estimation of the pa-
rameters and not to build a statistical model, we won’t
incorporate random error terms in the specification of
the model. The fuzziness of the variables will be mod-
elled by triangular fuzzy numbers, and the estimated
parameters too.
4.1 Case 1: Fuzzy Dependent Variable
and Crisp Independent Variables
This case is solved with (Donoso et al., 2006, p.1305-
1306) approach. Let us consider the following regres-
sion model
˜
:
˜y
i
=
m
j=1
˜
β
j
x
i j
, i = 1, . . . , n, (2)
where
˜
β
j
is the j-th parameter of the model. Note that
˜
β
j
is assumed to be fuzzy. Our aim is to find fuzzy
triangular estimators
ˆ
˜
β
j
that fit (2) best by minimising
the following weighted sum of squares J given in (3):
J =
n
i=1
k
1
(y
i
m
j=1
β
j
x
i j
)
2
+ k
2
(y
L
i
m
j=1
β
L
j
x
i j
)
2
+ k
3
(y
R
i
m
j=1
β
R
j
x
i j
)
2
!
= k
1
n
i=1
(y
i
x
i
β)
2
+ k
2
n
i=1
(y
L
i
x
i
β
L
)
2
+ k
3
n
i=1
(y
R
i
x
i
β
R
)
2
, (3)
under the constraints
(β
j
β
L
j
), (β
R
j
β
j
) 0 for j = 1, . . . , m. (4)
The quantities k
1
, k
2
and k
3
are tuning weights associ-
ated with the three sums of squares corresponding to
the central, left and right values of the fuzzy numbers.
The constraints given in (4) make sure the estimators
ˆ
˜
β
j
are fuzzy triangular numbers, i.e.
ˆ
˜
β
j
= (
ˆ
β
L
j
,
ˆ
β,
ˆ
β
R
j
).
4.2 Case 2: Fuzzy Dependent Variable
and Fuzzy Independent Variables
Let us suppose now that the independent variables are
also fuzzy, i.e. ˜x
i j
= (x
L
i j
, x
i j
, x
R
i j
), i = 1, . . . , n, j =
1, . . . , m. The fuzzy linear regression model becomes:
˜y
i
=
m
j=1
˜
β
j
e
˜x
i j
, i = 1, . . . , n, (5)
where
e
denotes the fuzzy multiplication operator.
We find the fuzzy triangular estimators
ˆ
˜
β by min-
imising the sum of squares
n
i=1
˜y
i
m
j=1
˜
β
j
e
˜x
i j
2
.
Rewriting it using the fuzzy multiplication rule yields
the following objective function J given in (6):
J =
n
i=1
k
1
(y
i
m
j=1
β
j
x
i j
)
2
+ k
2
(y
L
i
m
j=1
(min(β
L
j
x
L
i j
, β
R
j
x
L
i j
, β
L
j
x
R
i j
, β
R
j
x
R
i j
))
2
+ k
3
(y
R
i
m
j=1
(max(β
L
j
x
L
i j
, β
R
j
x
L
i j
, β
L
j
x
R
i j
, β
R
j
x
R
i j
))
2
!
.
(6)
Minimising (6) is not so direct, and a proper strat-
egy needs to be developed. If we know the sign
of the independent triangular fuzzy variable ˜x
i j
and
the sign of the unknown fuzzy parameters
˜
β
j
for all
j = 1, ..., m, we can easily know what will be the mini-
mum and maximum values to be calculated in the ob-
jective function (6) and thus solve the regression prob-
lem under the constraints (4). In practice, the sign of
the observations is evidently known, but not the sign
of the parameters. However, on one hand, by using
the crisp value of the fuzzy triangular observations,
i.e. when α = 1, we can perform the classical least
squares to get the estimates
ˆ
β
j
, j = 1, . . . , m, and if
they are sufficiently far away from zero, the sign of
ˆ
˜
β
j
will be the one of
ˆ
β
j
. If, on the other hand,
ˆ
β
j
is close
to zero, one has to first estimate
˜
β
j
, j = 1, . . . , m, via
(3). We note this first estimates by
ˆ
˜
β
(1)
j
. We then pick
their signs to fix the ones of the unknown estimators
β
L
j
and β
R
j
and use them to solve problem (6).
Suppose we estimated the sign of the un-
known fuzzy parameters as written above and then
chose, according to the signs of the fuzzy ob-
servation ˜x
i j
and
˜
β
j
, i = 1, ..., n, j = 0, ..., m
the corresponding minimum x
min
i j
and maximum
x
max
i j
that solve min(β
L
j
x
L
i j
, β
R
j
x
L
i j
, β
L
j
x
R
i j
, β
R
j
x
R
i j
) and
Fuzzy Least Squares and Fuzzy Orthogonal Least Squares Linear Regressions
361
max(β
L
j
x
L
i j
, β
R
j
x
L
i j
, β
L
j
x
R
i j
, β
R
j
x
R
i j
) respectively. The ob-
jective function (6) can then be rewritten as:
J =
n
i=1
k
1
(y
i
m
j=1
β
j
x
i j
)
2
+ k
2
(y
L
i
m
j=1
(β
min
j
x
min
i j
)
2
+k
3
(y
R
i
m
j=1
(β
max
j
x
max
i j
)
2
!
, (7)
where β
min
j
, β
max
j
are the coefficients associated to
x
min
i j
and x
max
i j
respectively. The special case where
x
i j
> 0, i = 1, . . . , n, deserves our attention. In this
case, β
min
j
and β
max
j
are obviously equal to β
L
j
and β
R
j
respectively.
Notice that, in general, when the fuzzy observations
˜x
i j
are not necessarily all positive, the coefficients
β
min
j
or β
max
j
may appear simultaneously in both left
and right squared terms in (6). This is due to the fuzzy
arithmetic rule for the product that sometimes yields
β
min
j
or β
max
j
as the coefficients, which both maximise
and minimise the fuzzy product. This general case is
no longer a simple quadratic problem. This is why
one recommends adding a constant valued vector to
each covariate to ensure their positivity and thus al-
lowing to solve a proprer quadratic problem. Making
so, only the constant of the problem is affected letting
the slopes unchanged.
5 CONSTRAINED AND
UNCONSTRAINED SOLUTIONS
The constraint (4) assures that the solutions are fuzzy
triangular numbers. If we do not impose (4), we end
up with solutions
˜
β
j
which are sometimes of the form
(b, c, a) with b > c > a, a, b, c R. This clearly goes
against the definition of fuzzy triangular numbers. A
way to bypass this not plausible solution is to consider
the fuzzy parameters as fuzzy intervals.
In order to avoid confusion, it is very important
to make a clear distinction between a vector of fuzzy
numbers and a fuzzy vector. (Viertl, 2018, p.14) de-
fines a fuzzy vector as:
Definition 1 (Fuzzy vector and fuzzy interval).
A k-dimensional fuzzy vector ˜x
with membership
function µ
˜x
is such that:
1. µ
˜x
: R
k
[0, 1].
2. The support of µ
˜x
is a bounded set.
3. α (0, 1] the so-called α-cut C
α
( ˜x
) := {x
R
k
|µ
˜x
(x) α} is non-empty, bounded, and
a finite union of simply connected and closed
bounded sets.
The set of all k-dimensional fuzzy vectors is de-
noted by F (R
k
). A k-dimensional fuzzy vector is
called a k-dimensional fuzzy interval if all α-cuts are
connected compact sets.
Applying the concept of fuzzy intervals, we are
able to understand the main difference between the
constrained solution and the unconstrained one. In
the first case, when the constraints (4) are satisfied,
one ends up with a vector of fuzzy parameters
˜
β sat-
isfying β
L
k
β
k
β
R
k
meaning that, the upper regres-
sion coefficients β
R
k
must always give a steeper slope
to the regression line, and conversely, β
L
k
must al-
ways give a lower slope. When the constraints (4)
are dropped, one ends up with a fuzzy interval
˜
β
j
,
j = 1, . . . , m. In this case, the fitted solution is such
that
ˆ
˜y
lower
ˆ
˜y
upper
. This allows the slopes to violate
(4) so long the lower regression line is below the up-
per one on the interval formed by the observations X
or
˜
X.
As an example, let us consider a regression model
with two crisp independent variables, the first one be-
ing the constant of the model. The unconstrained so-
lution
ˆ
˜
β
can be written as:
ˆ
˜
β
=
ˆ
˜
β
0
ˆ
˜
β
1
!
=
ˆ
β
,lower
0
,
ˆ
β
0
,
ˆ
β
,upper
0
ˆ
β
,lower
1
,
ˆ
β
1
,
ˆ
β
,upper
1
!
, (8)
where
ˆ
β
,lower
0
is the smaller β
0
coefficient estimated
and
ˆ
β
,upper
0
is the greater β
0
coefficient estimated.
The coefficients
ˆ
β
,lower
1
and
ˆ
β
,upper
1
are defined anal-
ogously. It can be easily shown that the estimated
parameters (8) satisfy the definition of a fuzzy inter-
val. Indeed, let C
α
be the rectangle formed by the
vertices (β
,lower,α
0
, β
,lower,α
1
), (β
,lower,α
0
, β
,upper,α
1
),
(β
,upper,α
0
, β
,lower,α
1
), (β
,upper,α
0
, β
,upper,α
1
) where α
stands for the α-cut of the fuzzy number. The more
we reduce the fuzziness, that is the closer to 1 α is,
the closer
ˆ
˜
β
,α
will be to the crisp solution given when
α = 1. Thus, C
α
i
C
α
j
for α
i
α
j
. Assuming the ex-
istence of a solution, C
α
is non-empty, bounded and a
finite union of simply connected and bounded rectan-
gles. Furthermore, The vector-membership function
µ
ˆ
˜
β
: R
2
[0, 1] defined as :
µ
ˆ
˜
β
:= max{α · I
C
α
(x) : α (0, 1]}, x R
2
,
has its support bounded by the rectangle C
0
.
By following the same reasoning, one can easily
verify that in the case of m covariates, the solution is
also a fuzzy interval, where C
α
is the m-dimensional
rectangle which has the form of a hypercube of 2
m
vertices.
FCTA 2023 - 15th International Conference on Fuzzy Computation Theory and Applications
362
6 FUZZY ORTHOGONAL LEAST
SQUARES
Let us first remind the definition of the fuzzy inner
product and the concept of fuzzy orthogonality, de-
scribed by (Ithoh, 2017, p.13).
Definition 2 (Fuzzy inner product).
Let X be a nonzero vector space over the field R
and
˜
X = {x
λ
|x X, λ (0, 1]} be the set of all fuzzy
points in X. A function ⟨·, ·⟩ :
˜
X ×
˜
X R is said to be
a fuzzy inner product on
˜
X if
1. a) x
λ
, x
µ
0; b) x
λ
, x
µ
= 0 if and only if x = 0;
2. kx
λ
, y
µ
= kx
λ
, y
µ
;
3. x
λ
+ y
λ
, z
µ
= x
λ
, z
µ
+ y
λ
, z
µ
;
4. x
λ
, y
µ
= y
µ
, x
λ
;
5. x
λ
, y
µ
x
λ
, y
ν
if 0 < ν µ 1;
6. for every x
λ
, y
µ
˜
X and ε > 0, there exists 0 <
δ < µ such that x
λ
, y
µδ
< x
λ
, y
µ
+ ε;
7. for every x
λ
, y
µ
˜
X and ε > 0, there exists 0 <
δ < 1 µ such that x
λ
, y
µ+δ
> x
λ
, y
µ
ε.
The pair (X, ⟨·, ·⟩) is called a strong fuzzy inner
product space.
Remark 1.
1. Let (X, ⟨·, ·⟩) be a strong fuzzy inner product
space. If
x
λ
1
, y
µ
1
= x
λ
2
, y
µ
2
x, y X,
λ
i
, µ
i
(0, 1], i = 1, 2,
(9)
then (X, ⟨·, ·⟩) becomes a usual inner product
space;
2. If we use a stronger condition,
x
λ
+ y
µ
, z
ν
= x
λ
, y
µ
+ y
µ
, z
ν
, (10)
then (X, ⟨·, ·⟩) is just a usual inner product space.
The concept of fuzzy orthogonality is defined as:
Definition 3 (Fuzzy orthogonality).
Let ˜x and ˜y be vectors in a fuzzy inner product
space. One says that ˜x is fuzzy orthogonal to ˜y if
x
λ
, y
µ
= 0, for some λ, µ (0, 1].
The fuzzy orthogonality between two vectors of a
fuzzy inner product space ˜x and ˜y is denoted by ˜x
F
˜y.
In case x
λ
is orthogonal to y
µ
for these specific values
of λ, µ (0, 1], i.e x
λ
, y
µ
= 0, we note it x
λ
y
µ
.
We can now introduce the fuzzy orthogonal least
squares estimators of a fuzzy regression model. Our
aim is first to transform the independent variables in
such a way that at the end of the procedure, the trans-
formed variables are mutually (fuzzy) orthogonal. We
then can use these transformed variables as new re-
gressors in the regression equation and estimate the
model as described in section 4. The orthogonalisa-
tion procedure will be discussed for two cases: inde-
pendent crisp or fuzzy variables.
6.1 Case 3: Fuzzy Dependent Variable
and Crisp Independent Variables
We can use a classic orthonormalisation procedure
to project X onto an orthonormal basis. Let ¯x =
( ¯x
1
, ..., ¯x
m
)
be the vector of the means associated to
the m independent variables X
1
, ..., X
m
; S the esti-
mated variance-covariance matrix; G the matrix of all
the eigenvectors of S; ι
n
= (1, . . . , 1)
a (n × 1) vec-
tor. We project X onto an orthonormal basis using
the transformation
X
= (X ι
n
¯x
)G. (11)
Then we solve (3) with X
instead of X. Due to
the orthogonality of X
, the individual coefficients
˜
β
j
of the regression model can be estimated also by
regressing the dependent variable on the variable X
j
only, j = 1, . . . , m. As a consequence, the orthogonal-
isation of the regressors allows us to find uncorrelated
fuzzy estimators
˜
β
j
.
Note that, sometimes, by convention, G is used
instead of G in (11). One has to be careful with mul-
tiplying or not by a minus sign the orthogonal projec-
tion (11) since it can greatly affect the fuzzy triangu-
lar estimators
˜
β
j
. This effect will be discussed more
in depth in section 7.3.
6.2 Case 4: Fuzzy Dependent Variable
and Fuzzy Independent Variables
When both the dependent and independent variables
of the regression model are fuzzy, we first need to
solve the min() and max() functions in (6). We per-
form then the orthogonalisation procedure (11) to find
X
min
, X
and X
max
from X
min
, X and X
max
.
This allows us to write the objective function (7) with
orthogonal fuzzy data as
J =k
1
n
i=1
(y
i
m
j=1
β
j
x
i j
)
2
+ k
2
n
i=1
(y
L
i
m
j=1
β
L
j
x
min
i j
)
2
+ k
3
n
i=1
(y
R
i
m
j=1
β
R
j
x
max
i j
)
2
. (12)
Again, the properties of the orthogonality permit
us to compute the fuzzy least squares individually for
Fuzzy Least Squares and Fuzzy Orthogonal Least Squares Linear Regressions
363
each covariate X
j
while giving the same solution as
when computed with all covariates X
j
at once.
Observe that the restructuring of the initial data
into ˜x
i j
= (x
min
i j
, x
i j
, x
max
i j
) allowing us to compute the
fuzzy products in (6) is possible if we know the signs
of the fuzzy unknown parameters
˜
β. This ambiguity
can be solved by adequately permuting the observa-
tions to take into account the fuzzy arithmetic. Af-
ter the orthogonalisation procedure, in case of neg-
ative sign occuring, the signs of the observations
˜x
i j
= (x
min
i j
, x
i j
, x
max
i j
) may differ from the signs of
˜x
i j
= (x
min
i j
, x
i j
, x
max
i j
).
To correct this and to preserve both the fuzzy
arithmetic and the quadratic nature of the problem,
one should shift ˜x
j
, j = 1, . . . , m, by adding a con-
stant K
j
= |min( ˜x
j
)|. This operation ensures that
the signs of the observations ˜x
j
are positive. As
we already said, this transformation has the effect
of changing the intercept value however, it does not
change the value of the slope coefficients. In practical
applications, one may use this transformation to en-
sure that the signs of the observations
˜
X are positive
as well as ensure that the orthonormalised observa-
tions
˜
X
are positive too.
Lastly, note that the fuzzy orthogonal covariates ˜x
j
meet the definition of fuzzy orthogonality of (Ithoh,
2017). Indeed, by construction their left part, center
and right part are orthogonal at any given α-cut,
˜x
j L
, ˜x
k L
= 0,
x
j
, x
k
= 0,
˜x
j R
, ˜x
k R
= 0, j, k = 1, . . . , m, j ̸= k, (13)
˜x
j
satisfying Masuo Itoh’s definition of fuzzy or-
thogonaly.
7 APPLICATION
Let us apply the above fuzzy least squares regression
methods and discuss them. We considered the fuzzy
data set given in Table 1 inspired by (Tanaka and Lee,
1998).
7.1 Fuzzy Dependent Variable and
Crisp Independent Variables
Let us postulate the following fuzzy regression model
(14):
˜
Y
i
=
˜
β
0
X
0
+
˜
β
1
X
i
+
˜
β
2
X
2
i
, i = 1, . . . , 22. (14)
Using k
1
= k
2
= k
3
= 1 in (3), we find, following the
methodology described in section 4.1 the fuzzy tri-
angular estimates given in Table 2. In Table 2, The
studied model is displayed in the first column while
the second column features the fuzzy regression coef-
ficients obtained with the use of a given model under
constraints 4. These fuzzy estimates are fuzzy trian-
gular numbers since they are solutions of constrained
models. The third column of Table 2 depicts the fuzzy
estimates given by solving the models in the first col-
umn under no constraint, thus these fuzzy estimates
are of the form of a fuzzy interval, as explained in
definition 1.
Table 3 gives the sum of squared residuals. The
first column tells which model has been used and if
the fuzzy estimates have been obtained under con-
straints or not. Then, the four remaining columns give
the sum of squared residuals with respect to the left,
center and right values and their addition computed
using the objective function of the respective models.
7.2 Fuzzy Dependent Variable and
Fuzzy Independent Variables
We consider again the data in Table 1. We fuzzify
the covariates X and X
2
by triangular fuzzy numbers.
Let be
˜
X = (X
L
, X, X
R
),
˜
X
2
=
˜
X
e
˜
X = (X
L
2
, X
2
, X
R
2
)
with X
L
= X u
1
, X
R
= X +u
2
where u
1
U(0, 0.5)
and u
2
U(0, 0.7). Note that
˜
X
2
is easily computed
since the observations are positive. The problem is
now given by the fuzzy regression equation (15):
˜
Y
i
=
˜
β
0
X
0
+
˜
β
1
e
˜
X
i
+
˜
β
2
e
˜
X
2
i
i = 1, . . . , 22. (15)
Using the methodology of section 4.2 the constrained
and unconstrained solutions are given in Table 2 and
the sum of squared residuals in Table 3.
7.3 Fuzzy Dependent Variable and
Crisp Orthogonal Independent
Variables
By means of an SVD decomposition and using (11),
we orthonormalise X onto X
. The orthonormalised
covariates X
1
and X
2
of X and X
2
are given in Table
1. The model is expressed by the regression equation
(16):
˜
Y
i
=
˜
β
0
X
0,i
+
˜
β
1
X
1,i
+
˜
β
2,i
X
2
, i = 1, . . . , 22. (16)
Following the methodology discussed in section 6.1,
the estimated parameters, without and with a multi-
plication of a minus sign, are given in Table 2. Note
FCTA 2023 - 15th International Conference on Fuzzy Computation Theory and Applications
364
Table 1: Fuzzy Dataset.
˜
Y X
0
X X
2
X
1
X
2
(15,22.5,30) 1 1 1 -1.93184373 -0.35491140
(20,28.75,37.5) 1 2 4 -1.80915821 -0.25981071
(15,25,35) 1 3 9 -1.67727776 -0.17390496
(25,42.5,60) 1 4 16 -1.53620237 -0.09719415
(25,40,55) 1 5 25 -1.38593203 -0.02967828
(40,52.5,65) 1 6 36 -1.22646676 0.02864264
(55,75,95) 1 7 49 -1.05780654 0.07776863
(70,85,100) 1 8 64 -0.87995139 0.11769968
(80,105,130) 1 9 81 -0.69290129 0.14843578
(90,120,150) 1 10 100 -0.49665625 0.16997695
(115,145,175) 1 11 121 -0.29121627 0.18232317
(140,167.5,195) 1 12 144 -0.07658135 0.18547446
(155,187.5,220) 1 13 169 0.14724851 0.17943080
(175,212.5,250) 1 14 196 0.38027331 0.16419220
(200,240,280) 1 15 225 0.62249305 0.13975866
(240,275,310) 1 16 256 0.87390773 0.10613018
(270,305,340) 1 17 289 1.13451735 0.06330676
(300,342.5,385) 1 18 324 1.40432191 0.01128840
(340,380,420) 1 19 361 1.68332142 -0.04992490
(380,420,460) 1 20 400 1.97151586 -0.12033314
(420,460,500) 1 21 441 2.26890525 -0.19993632
(465,507.5,550) 1 22 484 2.57548957 -0.28873445
that for the unconstrained solution (Fuzzy intervals),
the solutions are the same.
As the regressors of model (16) are mutually or-
thogonal, we can find the estimates of the parameter
˜
β
1
and
˜
β
2
by estimating successively the models:
˜
Y
i
=
˜
β
0
X
0,i
+
˜
β
1
X
1,i
, i = 1, . . . , 22, (17)
and
˜
Y
i
=
˜
β
0
X
0,i
+
˜
β
2,i
X
2
, i = 1, . . . , 22. (18)
Finally, remark that one can come back to the es-
timates computed with observations X using a linear
transformation. Let G be the matrix of the eigenvec-
tors of the variance-covariance matrix of the regres-
sors,
L,
= β
L
,
= β,
R,
= β
R
.
(19)
This transformation only works for the unconstrained
solutions. In the constrained case, this is in general
not true.
At this point, an important observation has to be
made. When using orthogonal observations X
we
find a different fuzziness for the fuzzy estimators
ˆ
˜
β
2
and
ˆ
˜
β
2
:
ˆ
˜
β
2
= (1.010993, 1.010993, 1.010993)
is crisp while
ˆ
˜
β
2
= (130.0262, 111.6804, 93.33466)
is fuzzy. This phenomenon is due to the constraints
(4). They force the fuzzy ”slope” represented by
˜
β
2
to be steeper for the upper β
R
2
and lower for β
L
2
. De-
pending on the reference frame used, this constrains
more or less the solutions. Because of this, one has to
be careful with the convention of multiplying or not
a minus sign after having projected the observations
onto an orthonormal basis. The solution
ˆ
˜
β
of model
(16) would have become
ˆ
˜
β
=
(165.2273, 192.6705, 220.11364)
(108.2216, 108.2216, 108.2216)
(111.6804, 111.6804, 111.6804)
,
(20)
if we choose to multiply by a minus sign, and thus
ˆ
˜
β
2
would also be crisp. Notice, on the other hand, that
the unconstrained solution is unaffected by the choice
of sign.
Fuzzy Least Squares and Fuzzy Orthogonal Least Squares Linear Regressions
365
7.4 Fuzzy Dependent Variable and
Fuzzy Orthogonal Independent
Variables
The model considered is given by the regression equa-
tion (21):
˜
Y
i
=
˜
β
0
X
0,i
+
˜
β
1
˜
X
1,i
+
˜
β
2,i
˜
X
2,i
, i = 1, . . . , 22. (21)
The orthogonal covariates X
1
and X
2
are obtained by
orthogonalisation of
˜
X
1
and
˜
X
2
as explained in section
6.2. Then, by application of the method discussed
in section 6.2 the resulting estimated parameters are
given in Table 2 and the associated sum of squared
residuals in Table 3. Note that the estimated parame-
ters can be computed individually for each covariate
˜
X
j
and yielding the same results. Moreover, using the
transformation (19) with X
min
and X
max
instead of
X for the left, respectively right fuzzy estimates β
L,
,
β
R,
will return the original estimates computed with
X
min
, X and X
max
.
7.5 General Discussion
Notice that, in table 2, in case 1, 2 and 3 (with mi-
nus sign), the fuzzy interval estimates seem to bet-
ter preserve the fuzziness than the fuzzy triangular
ones. This show that, empirically constrained mod-
els tend to have crispier estimates than unconstrained
ones. Moreover, as depicted in case 3, we can see how
the freedom of multiplying or not by a minus sign the
orthogonal projection (11) affects the fuzziness of the
estimates.
7.6 Sums of Squared Residuals
Lastly, in Table 3, notice that models (14) and its
orthonormalised version (16) share the same sum of
squared residuals. Moroever, this is also true for mod-
els 15 and 21. This could in fact be resulting from the
transformation 19 which allows one to retrieve the es-
timates found with non orthogonal regressors with the
orthonormalised ones.
8 CONCLUSION
We reexamine the fuzzy least squares method to solve
the so-called fuzzy linear regression problems. We
deal with two cases in particular. First, we consid-
ered that the independent variables are crisp, and sec-
ond, we treat the case of fuzzy independent variables.
In both situation, the dependent variable is fuzzy.
We develop a proper strategy to efficiently deal with
the fuzziness appearing in the observations. More-
over, we present and discuss two different types of
solutions arising from constrained and unconstrained
fuzzy least squares regression problems which are re-
spectively fuzzy triangular valued and fuzzy interval
valued.
Then, the extension of the method to orthogonal
fuzzy least squares regression methods has been in-
vestigated. In case of (fuzzy) orthogonal indepen-
dent variables, an important property of the classical
least squares method has been preserved. Due to the
orthogonality of the regressors, the individual coef-
ficients
˜
β
j
of the regression model can be estimated
also by regressing the dependent variable on the jth
covariate only. As a consequence, the orthogonalisa-
tion of the regressors allows us to find uncorrelated
fuzzy estimators
˜
β
j
. Moreover, in the unconstrained
case, we showed that there exist a linear transforma-
tion to find the coefficients associated to the original
regression model, i.e. the model before the orthogo-
nalisation of the regressors. We also highlighted that
the resulting sum of squares residuals of a model and
its orthonormal counterpart are the same. These dis-
coveries seem to be very promising to make progress
in statistical inference with fuzzy data, in particular in
the study of the fuzzy distributions of the estimates.
REFERENCES
Donoso, S., Mar
´
ın, N., and Villa, M. A. (2006). Quadratic
programming models for fuzzy regression. Intelligent
Data Engineering and Automated Learning IDEAL
2006, 7th International Conference, Burgos, Spain,
September 20-23, 2006. Proceedings. Lecture Notes
in Computer Science, pages 1304–1311.
D’Urso, P. and Massari, R. (2013). Weighted Least Squares
and Least Median Squares estimation for the fuzzy
linear regression analysis. Metron, 71(3):279–306.
Giordani, P. and Kiers, H. A. (2004). Principal Compo-
nent Analysis of symmetric fuzzy data. Computa-
tional Statistics & Data Analysis, 45(3):519–548.
Ithoh, M. (2017). Fuzzy inner product spaces and fuzzy
orthogonality. Advances in Fuzzy Sets and Systems,
5(1-10).
Kashani, M., Arashi, M., Rabiei, M., D’Urso, P., and Gio-
vanni, L. D. (2021). A fuzzy penalized regression
model with variable selection. Expert Systems with
Applications, 175:114696.
Lee, H. and Tanaka, H. (1998). Fuzzy regression analysis
by quadratic programming reflecting central tendency.
Metrika, 25(1):65–80.
Lee, H. and Tanaka, H. (1999). Fuzzy approximations with
non-symmetric fuzzy parameters in fuzzy regression
analysis. Journal of the Operations Research Society
of Japan, 42(1):98–112.
FCTA 2023 - 15th International Conference on Fuzzy Computation Theory and Applications
366
Li, Y., He, X., and Liu, X. (2023). Fuzzy multiple linear
least squares regression analysis. Fuzzy Sets and Sys-
tems, 459:118–143.
Mostofian, E., Azhini, M., and Bodaghi, A. (2017). Fuzzy
inner product spaces and fuzzy orthogonality. Tbilisi
Mathematical Journal, 10(2):157–171.
Stanojevi, B. and Stanojevi, M. (2022). Quadratic least
square regression in fuzzy environment. Elsevier B.V.
Tanaka, H. and Lee, H. (1997). Fuzzy linear regression
combining central tendency and possibilistic proper-
ties. Proceedings of 6th International Fuzzy Systems
Conference, 1:63–68.
Tanaka, H. and Lee, H. (1998). Interval regression analysis
by quadratic programming approach. IEEE Transac-
tions on Fuzzy Systems, 6(1):473–481.
Viertl, R. (2018). Statistical Methods for Fuzzy Data. Wi-
ley.
Yabuuch, Y. and Watada, J. (2017). Fuzzy Principal Com-
ponent Analysis and Its Application. International
Journal of Biomedical Soft Computing and Human
Sciences: the official journal of the Biomedical Fuzzy
Systems Association, 3(1):83–92.
Fuzzy Least Squares and Fuzzy Orthogonal Least Squares Linear Regressions
367
APPENDIX
Table 2: Models estimations
1
.
Models Fuzzy triangular estimates
2
Fuzzy intervals
3
Case 1 (14)
(15.673, 22.605, 29.537)
(2.1602, 0.3766, 1.4069)
(1.0109, 1.0109, 1.0109)
(20.032, 22.605, 25.178)
(3.2499, 0.3766, 2.4966)
(1.0583, 1.0109, 0.9636)
Case 2 (15)
(14.473, 21.298, 30.283)
(1.3693, 0.0499, 0.7347)
(0.9967, 0.9967, 0.9967)
(21.861, 22.605, 23.275)
(3.4034, 0.3766, 2.4123)
(1.0849, 1.0109, 0.9259)
Case 3 (16) (without (-)
4
)
(165.22, 192.67, 220.11)
(100.18, 108.22, 116.26)
(130.02, 111.68, 93.334)
(165.22, 192.67, 220.11)
(100.18, 108.22, 116.26)
(130.02, 111.68, 93.334)
Case 3 (16) (with (-)
5
)
(165.22, 192.67, 220.11)
(108.22, 108.22, 108.22)
(111.68, 111.68, 111.68)
(165.22, 192.67, 220.11)
(100.18, 108.22, 116.26)
(130.02, 111.68, 93.334)
Case 4 (21) (without (-))
(165.28, 192.67, 220.11)
(100.24, 108.22, 116.26)
(124.05, 111.68, 91.63)
(165.28, 192.67, 220.11)
(100.24, 108.22, 116.26)
(124.05, 111.68, 91.62)
1
Table displaying the fuzzy regression coefficients obtained via a given fuzzy regression model shown in the first column.
2
Solutions of a given constrained fuzzy least squares regression model.
3
Solutions of a given unconstrained fuzzy least squares regression model.
4
The orthogonal projection has not been multiplied by a minus sign.
5
The orthogonal projection has been multiplied by a minus sign.
Table 3: Sum of squared residuals
1
.
Models Constraints Left value
2
Center value
3
Right value
4
Total value
5
(14) constrained 489.8294 256.9274 555.5718 1302.329
(14) unconstrained 426.2246 256.9274 491.9669 1175.119
(15) constrained 555.5714 303.3096 1189.367 2048.248
(15) unconstrained 599.9359 256.9274 630.1572 1487.021
(16) (with (-)
6
) constrained 489.8294 256.9274 555.5718 1302.329
(16)(w/o (-)
7
) constrained 3306.798 256.9274 3372.54 6936.265
(16) unconstrained 426.2246 256.9274 491.9669 1175.119
(21) (w/o (-)) constrained &
unconstrained
599.9359 256.9274 630.1572 1487.021
1
Table displaying the squared residuals of the different models studied throughout this work.
2
Squared residuals of the left fuzzy parts.
3
Squared residuals of the central fuzzy parts.
4
Squared residuals of the right fuzzy parts.
5
Sum of the left, center and right squared residuals.
6
The orthogonal projection has been multiplied by a minus sign.
7
The orthogonal projection has not been multiplied by a minus sign.
FCTA 2023 - 15th International Conference on Fuzzy Computation Theory and Applications
368