Loading [MathJax]/jax/element/mml/optable/CombDiactForSymbols.js
Research article

Fuzzy gradient descent for the linear fuzzy real number system

  • Received: 11 July 2019 Accepted: 31 July 2019 Published: 09 August 2019
  • MSC : 00A69

  • Many problems in education, finance, and engineering design require that decisions be made under uncertainty. In these fields, Machine Learning is often used to search for patterns and information from data. To find patterns in Fuzzy Data, Fuzzy Machine Learning techniques can be used. In this paper, we focus on solving and manipulating Fuzzy Nonlinear problems in the Linear Fuzzy Real (LFR) number system using the Gradient Descent. The Gradient Descent is the most often used learning algorithm in Machine Learning. Thus, we propose the LFR Gradient Descent method for solving nonlinear equations in the LFR number system.

    Citation: Frank Rogers. Fuzzy gradient descent for the linear fuzzy real number system[J]. AIMS Mathematics, 2019, 4(4): 1078-1086. doi: 10.3934/math.2019.4.1078

    Related Papers:

    [1] Shexiang Hai, Liang He . The steepest descent method for fuzzy optimization problems under granular differentiability. AIMS Mathematics, 2025, 10(4): 10163-10186. doi: 10.3934/math.2025463
    [2] Kashif Hussain, Ala Amourah, Jamal Salah, Ali Fareed Jameel, Nidal Anakira . Autonomous block method for uncertainty analysis in first-order real-world models using fuzzy initial value problem. AIMS Mathematics, 2025, 10(4): 9614-9636. doi: 10.3934/math.2025443
    [3] Abdulaziz M. Alanazi, R. Sriraman, R. Gurusamy, S. Athithan, P. Vignesh, Zaid Bassfar, Adel R. Alharbi, Amer Aljaedi . System decomposition method-based global stability criteria for T-S fuzzy Clifford-valued delayed neural networks with impulses and leakage term. AIMS Mathematics, 2023, 8(7): 15166-15188. doi: 10.3934/math.2023774
    [4] Zengtai Gong, Han Gao, Ting Xie . Preinvexity of n-dimensional fuzzy number-valued functions: characterization, variational inequality and optimization problems. AIMS Mathematics, 2022, 7(4): 6099-6127. doi: 10.3934/math.2022340
    [5] Jun Hu, Jie Wu, Mengzhe Wang . Research on VIKOR group decision making using WOWA operator based on interval Pythagorean triangular fuzzy numbers. AIMS Mathematics, 2023, 8(11): 26237-26259. doi: 10.3934/math.20231338
    [6] M Junaid Basha, S Nandhini . A fuzzy based solution to multiple objective LPP. AIMS Mathematics, 2023, 8(4): 7714-7730. doi: 10.3934/math.2023387
    [7] Noor Jamal, Muhammad Sarwar, Nabil Mlaiki, Ahmad Aloqaily . Solution of linear correlated fuzzy differential equations in the linear correlated fuzzy spaces. AIMS Mathematics, 2024, 9(2): 2695-2721. doi: 10.3934/math.2024134
    [8] Zengtai Gong, Jun Wu, Kun Liu . The dual fuzzy matrix equations: Extended solution, algebraic solution and solution. AIMS Mathematics, 2023, 8(3): 7310-7328. doi: 10.3934/math.2023368
    [9] Shovan Dogra, Madhumangal Pal, Qin Xin . Picture fuzzy sub-hyperspace of a hyper vector space and its application in decision making problem. AIMS Mathematics, 2022, 7(7): 13361-13382. doi: 10.3934/math.2022738
    [10] Ashraf Al-Quran . T-spherical linear Diophantine fuzzy aggregation operators for multiple attribute decision-making. AIMS Mathematics, 2023, 8(5): 12257-12286. doi: 10.3934/math.2023618
  • Many problems in education, finance, and engineering design require that decisions be made under uncertainty. In these fields, Machine Learning is often used to search for patterns and information from data. To find patterns in Fuzzy Data, Fuzzy Machine Learning techniques can be used. In this paper, we focus on solving and manipulating Fuzzy Nonlinear problems in the Linear Fuzzy Real (LFR) number system using the Gradient Descent. The Gradient Descent is the most often used learning algorithm in Machine Learning. Thus, we propose the LFR Gradient Descent method for solving nonlinear equations in the LFR number system.


    Problems in production planning and scheduling, location, transportation, finance, and engineering design sometimes require that decisions be made under uncertainty. In education, there exist Fuzzy Educational datasets, which are useful sources of information for educational software developers, students, parents, researchers and other educational stakeholders [2]. Fuzzy decision making was initially introduced by Bellman and Zadeh [1]. This concept was then adopted to mathematical programming [14], uncertain programming [12,14], fuzzy posets [6], linear programming [3,7,8,10,13,14] and many other concepts.

    Linear Fuzzy Real (LFR) numbers are a system of numbers that have properties comparable to the set of real numbers and the set of fuzzy numbers [9]. LFR numbers are used in the study of fuzzy random variables [4,9] and in the linear optimization problem [8]. The set of LFR numbers is a set that shows intermediate properties which are unique to the set and not to those of either the real numbers or the "general" fuzzy numbers. Because of the unique properties of LFR numbers, we can solve fuzzy linear and nonlinear problems [11].

    At the core of machine learning techniques, there are many types of cost functions. The process of increasing the accuracy of our model by iterating over a training data involves searching for the lowest error in the cost function. The cost function is used to check the error in predictions of a Machine Learning model. In many learning algorithms the Gradient Descent is used to minimize the cost function [5].

    In this paper, the Gradient Descent method is used to find both the extreme values and find the solutions of fuzzy equations.

    The Crisp Gradient Descent algorithm begins with an initial value x0 and generates the sequence {xn}n=0, by

    xn=x(n1)γF.

    The algorithm finds the minimum of a function f by taking small steps proportion to the negative of the gradient of the function f. The symbol γ is the measure of the small steps and F is the gradient of f.

    Solving fuzzy equations also implies that the Gradient Descent method can be used to minimize fuzzy cost functions since the values of x that satisfies

    f(x)=bf(x)b=0.

    Will also be the global minimizer of

    f(x)b22.

    The paper is outlined as follows: Linear Fuzzy Real numbers are reviewed in Section 2. In Section 3, we look at nonlinear equations in LFR. In Section 4, we illustrate LFR Gradient Descent method to solve systems of nonlinear equations in LFR. Finally, conclusions are made in Section 5.

    In this section, we describe the Linear Fuzzy Real numbers [7,10]. Considering the set of all real numbers R, one way to associate a fuzzy number with a fuzzy subset of real numbers is as a function μ: R → [0, 1], where the value μ(x) is to represent a degree of belonging to the subset of R.

    Definition 2.1. (Linear fuzzy real number) Let μ: R → [0, 1] be a function such that:

    1. μ(x) = 1 if x = b;

    2. μ(x) = 0 if xa or xc;

    3. μ(x) = (x - a)/(b - a) if a < x < b;

    4. μ(x) = (c - x)/(c - b) if b < x < c.

    Then μ(a, b, c) is called a linear fuzzy real number with associated triple of real numbers (a, b, c) where abc shown in Figure 1.

    Figure 1.  Linear fuzzy real number μ(a, b, c).

    We let LFR be the set of all linear fuzzy real numbers. We note that any real number b can be written as a linear fuzzy real number, r(b), where r(b) = μ(b, b, b) and so R LFR. As a linear fuzzy real number, we consider r(b) to represent the real number b itself. A linear fuzzy real number μ(a, b, c) is defined to be positive if a > 0, negative if c < 0, and zeroic if a ≤ 0 and c ≥ 0. The hybrid nature of LFR allows one to map the set to an isomorphic set where we can use inverse operations to solve fuzzy linear equations [10].

    Operations, functions, and linear equations are also defined in LFR as follows:

    Definition 2.2. (Operations in LFR) For given two linear fuzzy real numbers μ1 = μ(a1, b1, c1) and μ2 = μ(a2, b2, c2), we define addition, subtraction, multiplication, and division by

    1. μ1 + μ2 = μ(a1 + a2, b1 + b2, c1 + c2);

    2. μ1 - μ2 = μ(a1 - c2, b1 - b2, c1 - a2);

    3. μ1 · μ2 = μ(min{a1a2, a1c2, a2c1, c1c2}, b1b2, max{a1a2, a1c2, a2c1, c1c2}).;

    4. μ1μ2=μ11μ2 where 1μ2=μ(min{1a2,1b2,1c2},median{1a2,1b2,1c2},max{1a2,1b2,1c2}).

    Definition 2.3. (Function in LFR) Given real-valued function f: RR and an LFR μ(a, b, c), the LFR-valued function f: LFRLFR is defined as

    {\dddot f}(\mu (a, b, c)) = \mu \left( {{{\rm{a}}^*}, {{\rm{b}}^*}, {{\rm{c}}^*}} \right),

    where a* = min{f(a), f(b), f(c)}, b* = median{f(a), f(b), f(c)}, c* = max{f(a), f(b), f(c)}. We note that if a = b = c then a* = b* = c*, i.e., {\bar f} (r(b)) = r(f(b)). Hence {\dddot f} is an extension of the function f.

    Definition 2.4. (Linear equation in LFR) A linear equation over LFR is an equation of the form

    \mu_{1} \cdot \mu_{x}+\mu_{2} = \mu_{3},

    where the μi are LFR's for i = 1, 2, 3 and μx = μ(α, β, γ) is an unknown LFR with a triple of unknown real numbers (α, β, γ).

    Definition 2.5. (Ordering Properties of LFR) Given μ1, μ2LFR, μ1μ2 provided that

    a_{\mathit{1}} \leq a_{\mathit{2}}, b_{\mathit{1}} \leq b_{\mathit{2}}, c_{\mathit{1}} \leq c_{\mathit{2}}.

    If 0 \le \mu (a, b, c) then 0 ≤ a ≤ b ≤ c, hence μ is a non-negative linear fuzzy real number. Therefore, if μ is non-negative and zeroic, then a = 0 precisely.

    (LFR, ≤) is a complete ordered set. However, it is not linearly ordered. If we let μ1 = μ(3, 4, 5) and let μ2 = μ(a, 5, 6) and state that a < 3, then it is not true that μ1 ≤ μ2 nor is it true that μ1 ≥ μ2. Therefore, μ1 and μ2 are incomparable in this order [10].

    Nonlinear equations can be found in many applications, all the way from light diffraction to planetary orbits for example [10]. In this section, we discuss how to solve nonlinear equations in LFR so that we may locate the maximum or minimum of nonlinear functions. It is important that we discuss locating fuzzy zeros of nonlinear functions. Thus, we will find a μx in LFR such that a nonlinear equation f(μx) = 0, where f: LFRLFR is a nonlinear function.

    Solving and optimizing nonlinear equations over linear fuzzy real numbers is possible with a modification of the Gradient Descent method. This method is also known as the Steepest Descent method over real numbers.

    Definition 3.1. A function f satisfies Lipschitz if there is a real number N such that |f(x)-f(y)| N|x-y|.

    We can apply Definition 3.1 to the function g = \nabla f , such that there is a real number P such that |g(x)-g(y)| P|x-y|. Thus, the function g satisfies Lipschitz.

    Proposition 3.1. The function f satisfies Lipschitz condition in the LFR environment for a real number N.

    Proof: Given the existence of f and f'( μ) where f is continuous in a compact interval. Let us suppose that \left|f^{\prime}(\mu)\right| \leq k for all \mu_{x} \in\left[\mu_{a}, \mu_{b}\right] such that μxμy,

    Thus, we have \left|\frac{f\left(\mu_{x}\right)-f\left(\mu_{y}\right)}{\left(\mu_{x}-\mu_{y}\right)}\right| = \left|f^{\prime}\left(\mu_{c}\right)\right| \leq k .

    For some \mu_{c} \in\left[\mu_{x}, \mu_{y}\right] by the mean value theorem. Then \left|f\left(\mu_{x}\right)-f\left(\mu_{y}\right)\right| \leq k\left|\mu_{x}-\mu_{y}\right| .

    Therefore, the function f satisfies Lipschitz condition on any interval \left[\mu_{a}, \mu_{b}\right]

    Since we can apply Taylor's theorem to f in LFR [10], Proposition 3.1 coupled with the application of the mean value theorem, there is a constant M, such that f is Lipschitz continuous in LFR.

    Proposition 3.2. Suppose the function {\dddot f} : LF{R^n} \to LFR , where {\dddot f} is an extension of the real function f. {\dddot f} is convex and differentiable and the gradient {\dddot f} is an extension of the gradient and real function \nabla f . \nabla f is Lipschitz continuous with constant K > 0. Then if we run gradient descent for j iterations with a fixed step {\rm{t}} \le \frac{1}{K} , it will yield a solution that satisfies

    \dddot {f}\left(\mu_{x}^{(k)}\right)-\dddot {f}\left(\mu_{x}^{*}\right) \leq \frac{\left\|\mu_{x}^{(0)}-\mu_{x}^{*}\right\|^{2}}{2 t j} (1)

    Where {\dddot f}(\mu_{x}^{*}) is the optimal value. \mu_{x}^{*} is the value that minimizes the function, {\dddot f} .

    Proof

    Because \nabla{\dddot f} is Lipschitz continuous with constant K, \nabla^{2} \dddot{f}\left(\mu_{x}\right) -KI is a negative semidefinite fuzzy matrix. \nabla^{2} \dddot{f} is an extension of the hessian and real function {\nabla ^2}f . The identity matrix, I, is a matrix in LFR[10].

    Because of LFR's hybrid nature we can expand around {\dddot f}(\mu_{x}) apply the rules of differential calculus[10]. This yields the following inequality.

    \dddot {f}\left(\mu_{y}\right) \leq \dddot {f}\left(\mu_{x}\right)+\nabla \dddot {f}\left(\mu_{x}\right)^{T}\left(\mu_{y}-\mu_{x}\right)+\frac{1}{2} \nabla^{2} \dddot {f}\left(\mu_{x}\right)\left\|\mu_{y}-\mu_{x}\right\|^{2} (2)

    In general, we wish to minimize the increase in the expansion of the area of a fuzzy number. For this reason, in the case of the expansion, we use b, the crisp maximum value of \mu_{x} in the Jacobian and Hessian. Thus, we use a modified expansion

    \dddot {f}\left(\mu_{y}\right) \leq \dddot {f}\left(\mu_{x}\right)+\nabla \dddot {f}(b)^{T}\left(\mu_{y}-\mu_{x}\right)+\frac{1}{2} \nabla^{2} \dddot {f}(b)\left\|\mu_{y}-\mu_{x}\right\|^{2} (3)

    Where {\nabla ^2}{\dddot f}\left(b \right) -KI is a semidefinite matrix. Thus

    \dddot {f}\left(\mu_{y}\right) \leq \dddot {f}\left(\mu_{x}\right)+\nabla \dddot {f}(b)^{T}\left(\mu_{y}-\mu_{x}\right)+\frac{1}{2} K\left\|\mu_{y}-\mu_{x}\right\|^{2} (4)

    Taking the next step in the gradient descent update by letting {\mu _y} - {\mu _x}t{\nabla {\dddot f}}\left(b \right)

    Then

    \begin{aligned} \dddot {f}\left(\mu_{y}\right) & \leq \dddot {f}\left(\mu_{x}\right)+\nabla \dddot {f}(b)^{T}(-t \nabla \dddot {f}(b))+\frac{1}{2} K\|-t \nabla \dddot {f}(b)\|^{2} \\ & = \dddot {f}\left(\mu_{x}\right)-\nabla \dddot {f}(b)^{T}(t \nabla \dddot {f}(b))+\frac{1}{2} K\|t \nabla \dddot {f}(b)\|^{2} \\ & = \dddot {f}\left(\mu_{x}\right)-t\|\nabla \dddot {f}(b)\|^{2}+\frac{1}{2} K t^{2}\|\nabla \dddot {f}(b)\|^{2} \\ & = \dddot {f}\left(\mu_{x}\right)-\left(1-\frac{1}{2} K t\right) t\|\nabla \dddot {f}(b)\|^{2} \end{aligned} (5)

    If we set {{t}} \le \frac{1}{K} , we find that -\left(1-\frac{1}{2} K t\right) \leq-\frac{1}{2} , thus we have

    \dddot {f}\left(\mu_{y}\right) \leq \dddot {f}\left(\mu_{x}\right)-\frac{1}{2} t\|\nabla \dot{f}(b)\|^{2} (6)

    This inequality implies that the objective value decreases with each iteration of the gradient descent. Since {\dddot f} is convex.

    \dddot {f}\left(\mu_{x}^{*}\right) \geq \dddot {f}\left(\mu_{x}\right)+\nabla \dddot {f}(b)^{T}\left(\mu_{x}^{*}-\mu_{x}\right) (7)

    which implies that

    \dddot {f}\left(\mu_{x}\right) \leq \dddot {f}\left(\mu_{x}^{*}\right)+\nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right). (8)

    Thus

    \begin{array}{l}{\dddot {f}\left(\mu_{y}\right) \leq \dddot {f}\left(\mu_{x}^{*}\right)+\nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right)-\frac{1}{2} t\|\nabla \dot{f}(b)\|^{2}} \\ {\dddot {f}\left(\mu_{y}\right)-\dddot {f}\left(\mu_{x}^{*}\right) \leq \nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right)-\frac{1}{2} t\|\nabla \dddot {f}(b)\|^{2}}\end{array} (9)
    \leq \frac{1}{2 t}\left(2 t \nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right)-t^{2}\|\nabla \dddot {f}(b)\|^{2}\right)
    \begin{array}{l}{ \leq \frac{1}{2 t}\left(2 t \nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right)-t^{2}\|\nabla \dddot {f}(b)\|^{2}-\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}+\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}\right)} \\ { = \frac{1}{2 t}\left(2 t \nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right)+\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}-t^{2}\|\nabla \dddot {f}(b)\|^{2}-\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}\right)}\end{array}
    = \frac{1}{2 t}\left(2 t \nabla \dddot {f}(b)^{T}\left(\mu_{x}-\mu_{x}^{*}\right)+\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}-t^{2}\|\nabla \dddot {f}(b)\|^{2}-\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}\right)

    By factoring, we have

    = \frac{1}{2 t}\left(\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}-\left(\left\|\mu_{x}-\mu_{x}^{*}\right\|-t\|\nabla \dddot {f}(b)\|\right)^{2}\right). (10)

    And thus, the triangle inequality yields

    \dddot {f}\left(\mu_{y}\right)-\dddot {f}\left(\mu_{x}^{*}\right) \leq \frac{1}{2 t}\left(\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}-\left\|\mu_{x}-t \nabla \dddot {f}(b)-\mu_{x}^{*}\right\|\right)^{2}. (11)

    If we note that another step of the gradient descent, by definition yields {\mu _y} - {\mu _x}t\nabla {\dddot f}\left(b \right)

    Then {\mu _y} - {\mu _x}t\nabla {\dddot f}\left(b \right) and

    \dddot {f}\left(\mu_{y}\right)-\dddot {f}\left(\mu_{x}^{*}\right) \leq \frac{1}{2 t}\left(\left\|\mu_{x}-\mu_{x}^{*}\right\|^{2}-\left\|\mu_{y}-\mu_{x}^{*}\right\|^{2}\right) (12)

    This holds for {\mu _y} on every iteration of the modified gradient descent. If we sum over the iterations.

    \begin{aligned} \sum\nolimits_{i = 1}^j {} \dddot {\theta}\left(\mu_{x}^{(i)}\right)-\dddot {f}\left(\mu_{x}^{*}\right) & \leq \sum\nolimits_{i = 1}^j {} \left(\left\|\mu_{x}^{(i-1)}-\mu_{x}^{*}\right\|^{2}-\left\|\mu_{x}^{(i)}-\mu_{x}^{*}\right\|^{2}\right) \\ & = \frac{1}{2 t}\left(\left\|\mu_{x}^{(0)}-\mu_{x}^{*}\right\|^{2}-\left\|\mu_{x}^{(k)}-\mu_{x}^{*}\right\|^{2}\right.\\ & \leq \frac{1}{2 t}\left(\left\|\mu_{x}^{(0)}-\mu_{x}^{*}\right\|^{2}\right) \end{aligned} (13)

    Given the update (6) and using the fact that f is decreasing. We can conclude the following:

    \begin{aligned} \dddot {f}\left(\mu_{x}^{(j)}\right)-\dddot {f}\left(\mu_{x}^{*}\right) & \leq \frac{1}{j} \sum\nolimits_{i = 1}^j {} \dddot {f}\left(\mu_{x}^{(i)}\right)-\dddot {f}\left(\mu_{x}^{*}\right) \\ & \leq \frac{\left\|\mu_{x}^{(0)}-\mu_{x}^{*}\right\|^{2}}{2 t_{j}} \end{aligned} (14)

    Which is what we sought to prove. Thus, the minimal value {\dddot f}(\mu_{x}^{*}) is reached where both \mu_{x}^{*} and {\dddot f}(\mu_{x}^{*}) are Linear Fuzzy Real (LFR) number as defined by Definition 2.1. It follows that Proposition 3.2 applies to (1). As well as the modified form (3).

    The Modified Gradient Descent method begins with an initial approximation LFR, {\mu ^{(0)}}_x = \mu \left({{a^{(0)}}, {b^{(0)}}, {c^{(0)}}} \right) , and generates the sequence \left\{ {\mu _x^{(n)}} \right\}_{n = 0}^\infty with \mu \left({{a^{({\rm{n}})}}, {b^{({\rm{n}})}}} \right) , by

    \mu_{x}^{(n)} = \mu_{x}^{(n-1)}-\frac{1}{\alpha} \nabla f\left(b^{n-1}\right)^{T}, \text { for } n \geq 1.

    The stopping criterion of this method is \left| {{b^n} - {b^{n - 1}}} \right| < \epsilon , where \epsilon is a preset small value.

    Example 3.1. Find the local minimum of an LFR nonlinear equation {\mu ^4}_x - 3{\mu ^3}_x + 2.

    This is an equation with fuzzy coefficients. With typical calculation, it is expected that the local minimum occurs "around" 2.25.

    If we set f (μx) = μ4x - 3 μ3x +2, then f '(μx) = 4μ3x - 9μ2x . If we use an initial approximation μ(0)x = μ(4.5, 5, 5.5). If we set \gamma = \frac{1}{\alpha } . Then we can set \gamma = 0.001 and \epsilon = 0.00001 . Using Python, after 356 iterations, we find a fuzzy solution μ(*)x = μ(1.7505, 2.2505, 2.7505).

    Example 3.2. Find the extreme values of an LFR nonlinear equation

    \mathit{5}^{*}\left(\mu_{x 1}\right)^{\mathit{2}}+\left(\mu_{x 2}\right)^{2}+4 \mu_{x 1}^{*} \mu_{x 2}-14 \mu_{x 1}-6 \mu_{x 2}+20

    This is a multivariable equation with fuzzy coefficients. With typical calculation, it is expected that the optimal solution occurs "around" (1, 1).

    Let us set {\rm{F}}({\rm{x}}) = 5*{\left({{\mu _{x1}}} \right)^\mathit{2}} + {\left({{\mu _{x2}}} \right)^\mathit{2}} + 4{\mu _{x1}}*{\mu _{x2}} - 14{\mu _{x1}} - 6{\mu _{x2}} + 20

    then we have

    \nabla F = \left[\begin{array}{c}{10 * \mu_{x 1}+4 * \mu_{x 2}-14} \\ {4 * \mu_{x 1}+2 * \mu_{x 2}-6}\end{array}\right]

    If we use an initial approximation \mu _{x1}^{(0)} = \mu (- 0.5, 0, 0.5) , and \mu _{x2}^{(0)} = \mu (9.5, 10, 10.5) . If we set \gamma=0.02 and \epsilon = 0.000001 . Using Python, after 59583 iterations, we find fuzzy solutions μ(*)x1 = μ(0.44704, 0.944244, 1.44704) and μ(*)x2 = μ(0.63582, 1.13461, 1.63582).

    In this section, we discuss using Gradient Descent to solve system of nonlinear equations in LFR. To illustrate, suppose we solve the following system of two LFR nonlinear equations:

    \left\{\begin{array}{l}{f_{1}\left(\mu_{x 1}, \mu_{x 2}\right) = 0} \\ {f_{2}\left(\mu_{x 1}, \mu_{x 2}\right) = 0}\end{array}\right\} (15)

    Let the vector function F and the Jacobian matrix J = \nabla F be the followings:

    F\left(\mu_{x 1}, \mu_{x 2}\right) = \left[\begin{array}{l}{f_{1}\left(\mu_{x 1}, \mu_{x 2}\right)} \\ {f_{2}\left(\mu_{x 1}, \mu_{x 2}\right)}\end{array}\right] (16)

    And

    J = \left[\begin{array}{ll}{\frac{\partial f_{1}}{\partial \mu_{x 1}}} & {\frac{\partial f_{1}}{\partial \mu_{x 2}}} \\ {\frac{\partial f_{2}}{\partial \mu_{x 1}}} & {\frac{\partial f_{2}}{\partial \mu_{x 2}}}\end{array}\right] (17)

    Then LFR Gradient Descent method generates the solution sequences \left\{\mu_{x 1}^{(n)}\right\}_{n = 0}^{\infty} \text { and }\left\{\mu_{x 2}^{(n)}\right\}_{n = 0}^{\infty} such as:

    \left[\begin{array}{c}{\mu_{x 1}^{(n)}} \\ {\mu_{x 2}^{(n)}}\end{array}\right] = \left[\begin{array}{c}{\mu_{x 1}^{(n-1)}} \\ {\mu_{x 2}^{(n-1)}}\end{array}\right]-\gamma\left(J\left(\mu_{x 1}^{(n-1)}, \mu_{x 2}^{(n-1)}\right)\right) (18)

    Where \mu_{x 1}^{(n)} = \mu\left(a_{x 1}^{(n)}, b_{x 1}^{(n)}, c_{x 1}^{(n)}\right) \quad \text { and } \quad \mu_{x 2}^{(n)} = \mu\left(a_{x 2}^{(n)}, b_{x 2}^{(n)}, c_{x 2}^{(n)}\right) .

    The stopping criterion of this method is \min \left\{\left|b_{x 1}^{(n)}-b_{x 1}^{(n-1)}\right|, \left|b_{x 2}^{(n)}-b_{x 2}^{(n-1)}\right|\right\} < \epsilon , where \epsilon is a preset small value.

    Example 4.1. Solve an LFR system of nonlinear equations:

    \left\{\begin{array}{c}{\mu_{x 1}+\mu_{x 2}+\mu_{x 3} = 6} \\ {\mu_{x 1}^{2}+\mu_{x 2} = 5} \\ {\mu_{x 2}^{2}+\mu_{x 3} = 4}\end{array}\right.

    We easily find F and J as

    F\left(\mu_{x 1}, \mu_{x 2}, \mu_{x 3}\right) = \left[\begin{array}{c}{\mu_{x 1}+\mu_{x 2}+\mu_{x 3}-6} \\ {\mu_{x 1}^{2}+\mu_{x 2}-5} \\ {\mu_{x 2}^{2}+\mu_{x 3}-4}\end{array}\right]

    And

    J = \left[\begin{array}{ccc}{1} & {1} & {1} \\ {2 \mu_{x 1}} & {1} & {0} \\ {0} & {2 \mu_{x 2}} & {1}\end{array}\right]

    With typical calculation, it is expected that the optimal solution occurs "around" (2, 1, 3). If we use initial approximations \mu_{x 1}^{(0)} = \mu_{x 2}^{(0)} = \mu_{x 3}^{(0)} = \mu(-1, 2, 4) . And set \gamma = 0.02 and \epsilon = 0.000001 .

    Then in Python after 121193 iterations, LFR Gradient Descent Method yields

    \begin{array}{*{20}{l}} {\mu {{_{}^{(*)}}_{x1}} = \mu ( - 1.004, 1.996, 3.996)}\\ {\mu {{_{}^{(*)}}_{x2}} = \mu ( - 1.978, 1.022, 3.022)}\\ {\mu {{_{}^{(*)}}_{x3}} = \mu ( - 0.035, 2.965, 4.965)} \end{array}

    We often do not know the consequence of each possible decision with precision, sometimes we can only know this case with uncertainty. In many cases the LFR approach may be a good approach. The simplest case for example would be tolerance, where all possible values within the interval are equally likely. LFR presents an advantage where it may represent an interval, a crisp point, or a fuzzy number that can be written discretely or continuously. This is an advantage in the case of certain problems where it is already known that "crisp optima" in the purest sense do not exist, but projecting to the middle μ(a, b, c) → μ(b, b, b) = b produces a "crisp good choice" for an optimal value in optimization [12] and machine learning. It is also possible to approximate other fuzzy number types, using LFR. In this paper, we used LFR coupled with the Gradient Descent to solve and optimize Fuzzy Nonlinear problems and systems of fuzzy problems. As has been shown, LFR is a viable environment for machine learning techniques.

    This work has been financially supported by Project Engage: ROAR a grant funded by the United States Department of Education. Grant number: P120A160122.

    The author declares that there is no conflict of interests.



    [1] R. E. Bellman, L. A. Zadeh, Decision making in a fuzzy environment, Manage. Sci., 17 (1970), 141-164. doi: 10.1287/mnsc.17.4.B141
    [2] J. Cheng, Data-Mining Research in Education, Report, International School of Software, Wuhan University, 2017.
    [3] D. Dubois, H. Prade, System of linear fuzzy constraints, Fuzzy Set. Syst., 3 (1980), 37-48. doi: 10.1016/0165-0114(80)90004-4
    [4] B. Monk, A Proposed Theory of Fuzzy Random Variables, Dissertation, University of Alabama, 2001.
    [5] A. C. Muller, S. Guido, Introduction to Machine Learning with Python: A Guide for Data Scientists, O'Reilly Media, 2016.
    [6] J. Neggers, H. Kim, Fuzzy posets on sets, Fuzzy Set. Syst., 117 (2001), 391-402. doi: 10.1016/S0165-0114(98)00249-8
    [7] J. Neggers, H. Kim, On Linear Fuzzy Real Numbers, Manuscript for book under development, 2007.
    [8] C. V. Negoita, Fuzziness in management, OPSA/TIMS, Miami, 1970.
    [9] R. Prevo, Entropies of families of fuzzy random variables: an introduction to an in-depth exploration of several classes of important examples, Dissertation, University of Alabama, 2002.
    [10] F. Rogers, Optimal Choices in an LFR System, Dissertation, University of Alabama, 2005.
    [11] F. Rogers, Y. Jun, Fuzzy Nonlinear Optimization for the Linear Fuzzy Real Number System, International Mathematical Forum, 4 (2009), 587-596.
    [12] N. V. Sahindas, Optimization under uncertainty: state-of-the-art and opportunities, Comput. Chem. Eng., 28 (2004), 971-983. doi: 10.1016/j.compchemeng.2003.09.017
    [13] H. Tanaka, K. Asai, Fuzzy linear programming problems with fuzzy numbers, Fuzzy Set. Syst., 13 (1984), 1-10. doi: 10.1016/0165-0114(84)90022-8
    [14] H. J. Zimmermann, Fuzzy mathematical programming, Comput. Oper. Res., 10 (1983), 291-298. doi: 10.1016/0305-0548(83)90004-7
  • Reader Comments
  • © 2019 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(5005) PDF downloads(563) Cited by(0)

Figures and Tables

Figures(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog