Next Article in Journal
Spin-Dependent Scattering of Scalar and Vector Dark Matter on the Electron
Next Article in Special Issue
Some New Generalizations of Reverse Hilbert-Type Inequalities via Supermultiplicative Functions
Previous Article in Journal
A Hierarchical Universal Algorithm for Geometric Objects’ Reflection Symmetry Detection
Previous Article in Special Issue
Inequalities for Approximation of New Defined Fuzzy Post-Quantum Bernstein Polynomials via Interval-Valued Fuzzy Numbers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Accelerated Fixed-Point Algorithm for Classification and Convex Minimization Problems in Hilbert Spaces with Directed Graphs

by
Kobkoon Janngam
1 and
Rattanakorn Wattanataweekul
2,*
1
Graduate Ph.D. Degree Program in Mathematics, Department of Mathematics, Faculty of Science, Chiang Mai University, Chiang Mai 50200, Thailand
2
Department of Mathematics, Statistics and Computer, Faculty of Science, Ubon Ratchathani University, Ubon Ratchathani 34190, Thailand
*
Author to whom correspondence should be addressed.
Submission received: 29 April 2022 / Revised: 16 May 2022 / Accepted: 19 May 2022 / Published: 21 May 2022
(This article belongs to the Special Issue Functional Equations and Inequalities 2021)

Abstract

:
A new accelerated algorithm for approximating the common fixed points of a countable family of G-nonexpansive mappings is proposed, and the weak convergence theorem based on our main results is established in the setting of Hilbert spaces with a symmetric directed graph G. As an application, we apply our results for solving classification and convex minimization problems. We also apply our proposed algorithm to estimate the weight connecting the hidden layer and output layer in a regularized extreme learning machine. For numerical experiments, the proposed algorithm gives a higher performance of accuracy of the testing set than that of FISTA-S, FISTA, and nAGA.

1. Introduction

Let H be a real Hilbert space with the norm · and C be a nonempty closed convex subset of H . A mapping T : C C is said to be nonexpansive if it satisfies the following symmetric contractive-type condition:
T x T y x y ,
for all x , y C ; see [1].
The element x C is a fixed point of T if T x = x and F ( T ) : = { x C : x = T x } stands for the set of all fixed points of T.
Fixed point theory, i.e., the study of the conditions under which a map admits a fixed point, is an extensive area of research due to its numerous applications in many fields. It started with Banach’s work, which established the existence of a unique fixed point for a contraction using a classical theorem known as the Banach contraction principle; see [2]. The contraction principle of Banach has been expanded and generalized in various directions due to its applications in mathematics and other fields. One of the more recent generalizations is due to Jachymski.
Jachymski [3] introduced the structure of the graph on metric spaces using fixed point theory and obtained certain conditions for self-mapping to be a Picard operator. Several authors [4,5,6,7] proved fixed point theorems for a new type of contraction on a metric space endowed with graphs. Aleomraninejad et al. [8] used the idea of Reich et al. [9] and proved a strong convergence theorem for G-contractive and G-nonexpansive mappings. On hyperbolic metric spaces, Alfuraidan and Khamsi [10] gave a definition of G-monotone nonexpansive multivalued mappings and proved the existence of a fixed point for multivalued contraction and monotone single-valued mappings. Later on, Alfuraidan [11] presented and study the existence of fixed points for G-monotone nonexpansive and extended the results of Jachymski [3]. For approximating common fixed points of a finite family of G-nonexpansive mappings, Suantai et al. [12] used the shrinking projection with the parallel monotone hybrid method. They also used a graph to prove a strong convergence theorem in Hilbert spaces under specific conditions, and they then applied their iterative scheme to signal recovery.
In the past decade, algorithms for approximating fixed points of G-nonexpansive mappings without inertial techniques have been proposed by many researchers; see [3,8,10,11,12,13,14,15,16,17,18]. We require more efficient algorithms for solving such problems. As a result, some accelerated fixed-point algorithms using inertial techniques have been proposed to improve convergence behavior; see [19,20,21,22,23,24,25,26,27]. Recently, Janngam et al. [28] proved the weak convergence theorem for a countable family of G-nonexpansive mappings in a Hilbert space by using a coordinate affine structure with an inertial technique. They also applied their method to image recovery.
Inspired by previous research described above, we introduce a new accelerated algorithm based on the concept of the inertial technique for finding a common fixed point of a family of G-nonexpansive mappings in Hilbert spaces. We employ our result to solve data classification and convex minimization problems and also compare our algorithm efficiency to that of FISTA-S, FISTA, and nAGA.
This paper is classified as follows: in Section 2, we give certain terminology as well as some facts that will be useful in later sections. We investigate and prove our algorithm’s weak convergence in Section 3. For application, we apply our method for solving convex minimization and data classification problems in Section 4 and provide numerical experiments of classification problems in Section 5. The last section of our paper, Section 6, is a summary.

2. Preliminaries

Let C be a nonempty subset of a real Banach space X. Let Δ = { ( u , u ) : u C } , where Δ stands for the diagonal of the Cartesian product C × C . Consider a directed graph G in which the set V ( G ) of its vertices corresponds to C, and the set E ( G ) of its edges contains all loops. A directed graph G is said to have parallel edges if two or more edges with both the same tail vertex and the same head vertex.
Assume that G does not have parallel edges. Then, G = ( V ( G ) , E ( G ) ) . The conversion of a graph G is denoted by G 1 . Thus, we have
E ( G 1 ) = { ( u , v ) C × C : ( v , u ) E ( G ) } .
Let us give some definitions of basic graph properties which are used in this paper (see [29] for more details).
Definition 1. 
A graph G is said to be
(i)
symmetric if for all ( x , y ) E ( G ) ; we have ( y , x ) E ( G ) ;
(ii)
transitive if for any u , v , w V ( G ) with ( u , v ) , ( v , w ) E ( G ) ; then, ( u , w ) E ( G ) ;
(iii)
connected if there is a path between any two vertices of the graph G .
The definition of G-contraction [3] and G-nonexpansive mappings [13] are given as follows.
Definition 2. 
A mapping T : C C is said to be
(i)
G-contraction if
(a)
T is edge-preserving, i.e., ( T u , T v ) E ( G ) for all ( u , v ) E ( G ) .
(b)
There exists ρ [ 0 , 1 ) such that T u T v ρ u v for all ( u , v ) E ( G ) , where ρ is called a contraction factor.
(ii)
G-nonexpansive if
(a)
T is edge-preserving.
(b)
T u T v u v for all ( u , v ) E ( G ) .
Example 1. 
Let C = [ 0 , 2 ] R . Suppose that ( u , v ) E ( G ) if and only if 0.4 u , v 1.6 or u = v , where u , v R . Let S , T : C C be defined by
S u = c o s ( t a n ( u 1 ) ) s i n ( π 2 ) a n d T u = 1 + l n ( u ) 3 ,
for all u C . Then, both S and T are G-nonexpansive but not nonexpansive (see [30] for more details).
Definition 3. 
A mapping T : C C is called G-demiclosed at 0 if u n u and T u n 0 , then T u = 0 for all sequence { u n } C with ( u n , u n + 1 ) E ( G ) .
To prove our main result, we need the definition of the coordinate affine of the graph G as follows.
Definition 4 
([28]). Assume that Λ : = n = 1 F ( T n ) and Λ × Λ E ( G ) . Then, E ( G ) is said to be
(i)
left coordinate affine if α ( x , y ) + β ( u , y ) E ( G ) for all ( x , y ) , ( u , y ) E ( G ) and all α, β R with α + β = 1 .
(ii)
right coordinate affine if α ( x , y ) + β ( x , z ) E ( G ) for all ( x , y ) , ( x , z ) E ( G ) and all α, β R with α + β = 1 .
We say that E ( G ) is coordinate affine if E ( G ) is both left and right coordinate affine.
The results of the following lemmas can be used to prove our main theorem; see also [19,31,32].
Lemma 1 
([31]). Let { η n } , { ν n } and { ϑ n } be sequences of nonnegative real numbers such that n = 1 ϑ n < and n = 1 ν n < . Suppose that
η n + 1 ( 1 + ϑ n ) η n + ν n f o r   a l l n 1 .
Then, lim n η n exists.
Lemma 2 
([32]). For a real Hilbert space H, the following results hold:
(i) For any u , υ H and γ [ 0 , 1 ] ,
γ u + ( 1 γ ) υ 2 = γ u 2 + ( 1 γ ) υ 2 γ ( 1 γ ) u υ 2 .
(ii) For any u , υ H ,
u ± υ 2 = u 2 ± 2 u , υ + υ 2 .
Lemma 3 
([19]). Let { υ n } and { μ n } be sequences of nonnegative real numbers such that
υ n + 1 ( 1 + μ n ) υ n + μ n υ n 1 f o r   a l l n 1 .
Then,
υ n + 1 M · j = 1 n ( 1 + 2 μ j ) ,
where M = max { υ 1 , υ 2 } . If n = 1 μ n < , then { υ n } is bounded.
Let { u n } be a sequence in X . We write u n u to indicate that a sequence { u n } converges weakly to a point u H . Similarly, u n u will symbolize the strong convergence. For  v C , if there is a subsequence { u n k } of { u n } such that u n k v , then v is called a weak cluster point of { u n } . The set of all weak cluster points of { u n } is denoted by ω w ( u n ) .
The following lemma was proved by Moudafi and Al-Shemas; see [33].
Lemma 4 
([33]). Let { u n } be a sequence in a real Hilbert space H such that there exists Λ H satisfying:
(i) For any p Λ , lim n u n p exists.
(ii) Any weak cluster point of { u n } Λ .
Then, there exists x * Λ such that u n x * .
Let { T n } and ψ be families of nonexpansive mappings of C into itself such that F ( ψ ) n = 1 F ( T n ) , where F ( ψ ) is the set of all common fixed points of each T ψ . A sequence { T n } satisfies the NST-condition (I) with ψ if, for any bounded sequence { u n } in  C ,
lim n T n u n u n = 0 implies lim n T u n u n = 0 ,
for all T ψ ; see [34]. If ψ = { T } , then { T n } satisfies the NST-condition (I) with T .
Example 2 
([30]). Let T ψ . Define T n = β n I + ( 1 β n ) T , where 0 < s β n t < 1 for all n N . Therefore, { T n } is a family of G-nonexpansive mappings and satisfies the NST-condition.
Let A : H 2 H be a maximal monotone operator and c > 0 . The resolvent of A is defined by J c A = ( I + c A ) 1 , where I is an identity operator. If A = f for some f Γ 0 ( H ) , where Γ 0 ( H ) stands for the set of proper lower semicontinuous convex functions from H ( , + ] , then J c A = p r o x c f . The forward-backward operator of lower semi-continuous and convex functions of f , g : R n ( , + ] has the following definition:
A forward-backward operator T is defined by T : = p r o x λ g ( I λ f ) for λ > 0 , where f is the gradient operator of function f and p r o x λ g x : = a r g m i n y H g ( y ) + 1 2 λ y x 2 (see [35,36]). Moreau [37] defined the operator p r o x λ g as the proximity operator with respect to λ and g. If λ ( 0 , 2 / L ) , then T is nonexpansive and L is a Lipschitz constant of  f .
We have the following remark for the definition of the proximity operator; see [38].
Remark 1. 
Let g : R n R be given by g ( x ) = λ x 1 . The proximity operator of g is evaluated by the following formula
p r o x λ · 1 ( x ) = ( s i g n ( x i ) m a x ( | x i | λ , 0 ) ) i = 1 n ,
where x = ( x 1 , x 2 , , x n ) and x 1 = i = 1 n | x i | .
The following lemma was proved by Bussaban et al.; see [20].
Lemma 5. 
Let H be a real Hilbert space and T be the forward-backward operator of f and g, where g is a proper lower semi-continuous convex function from H into R { } , and f is a convex differentiable function from H into R with gradient f being L-Lipschitz constant for some L > 0 . If  { T n } is the forward-backward operator of f and g such that a n a with a, a n ( 0 , 2 / L ) , then { T n } satisfies the N S T -condition (I) with T.

3. Main Results

Let C be a nonempty closed and convex subset of a real Hilbert space H with a directed graph G = ( V ( G ) , E ( G ) ) such that V ( G ) = C . Let { T n } be a family of G-nonexpansive mappings of C into itself such that n = 1 F ( T n ) .
The following proposition is useful for our main theorem.
Proposition 1. 
Let x * n = 1 F ( T n ) and y 0 , x 1 C be such that ( x * , y 0 ) , ( x * , x 1 ) E ( G ) . Let { x n } be a sequence generated by Algorithm 1. Suppose E ( G ) is symmetric, transitive, and a right coordinate affine. Then, ( x * , x n ) , ( x * , y n ) , ( x * , z n ) and ( x n , x n + 1 ) E ( G ) for all n N .
Algorithm 1: (ASA) An Accelerated S-algorithm
1:
Initial. Take y 0 , x 1 C are arbitrary, n = 1 , β n [ a , b ] ( 0 , 1 ) , θ n 0 and n = 1 θ n < and α n 1 .
2:
Step 1. Compute y n , z n and x n + 1 by
z n = ( 1 β n ) x n + β n T n x n , y n = ( 1 α n ) T n x n + α n T n z n , x n + 1 = y n + θ n ( y n y n 1 ) .
Then, n : = n + 1 and go to Step 1.
Proof. 
We shall prove the results by using strong mathematical induction. From Algorithm 1, we obtain
( x * , z 1 ) = x * , ( 1 β 1 ) x 1 + β 1 T 1 x 1 = ( 1 β 1 ) ( x * , x 1 ) + β 1 ( x * , T 1 x 1 ) .
Since ( x * , x 1 ) E ( G ) and T n is edge preserving, we obtain ( x * , z 1 ) E ( G ) . Again, by Algorithm 1, we obtain
( x * , y 1 ) = ( x * , ( 1 α 1 ) T 1 x 1 + α 1 T 1 z 1 ) ) = ( 1 α 1 ) ( x * , T 1 x 1 ) + α 1 ( x * , T 1 z 1 ) .
Since ( x * , z 1 ) E ( G ) and T n is edge preserving, we obtain ( x * , y 1 ) E ( G ) . Next, we assume that ( x * , z k ) , ( x * , y k ) and ( x * , x k ) E ( G ) for all k < n . By Algorithm 1, we obtain
( x * , z k + 1 ) = ( x * , ( 1 β k + 1 ) x k + 1 + β k + 1 ( T k + 1 x k + 1 ) ) = ( 1 β k + 1 ) ( x k , x k + 1 ) + β k + 1 ( x * , T k + 1 x k + 1 ) ,
( x * , y k + 1 ) = ( x * , ( 1 α k + 1 ) T k + 1 y k + 1 + α k + 1 T k + 1 z k + 1 = ( 1 α k + 1 ) ( x * , T k + 1 y k + 1 + α k + 1 ( x * , T k + 1 z k + 1 )
and
( x * , x k + 1 ) = ( x * , y k + θ k ( y k y k + 1 ) ) = ( x * , ( 1 + θ k ) y k θ k y k 1 ) = ( 1 + θ k ) ( x * , y k ) θ k ( x * , y k 1 ) .
By (1)–(3) and since E ( G ) is right coordinate affine and T n is edge preserving, we obtain ( x * , x k + 1 ) , ( x * , y k + 1 ) and ( x * , z k + 1 ) are in E ( G ) . By strong mathematical induction, we conclude that ( x * , x n ) , ( x * , y n ) , ( x * , z n ) E ( G ) for all n N . Since E ( G ) is symmetric, we obtain ( x n , x * ) E ( G ) . Since ( x n , x * ) , ( x * , x n + 1 ) E ( G ) and E ( G ) is transitive, we obtain ( x n , x n + 1 ) E ( G ) as required. □
We now prove the weak convergence of G-nonexpansive mapping in a real Hilbert space by using Algorithm 1.
Theorem 1. 
Let C be a nonempty closed and convex subset of a real Hilbert space H with a directed graph G = ( V ( G ) , E ( G ) ) with V ( G ) = C and E ( G ) is symmetric, transitive, and right coordinate affine. Let y 0 , x 1 C and { x n } be a sequence in H defined by Algorithm 1. Suppose { T n } satisfies NST-condition (I) with T such that F ( T ) n = 1 F ( T n ) and ( x * , y 0 ) , ( x * , x 1 ) E ( G ) for all x * n = 1 F ( T n ) . Then, x n x * F ( T ) .
Proof. 
Let x * n = 1 F ( T n ) . By the definition of z n and y n , we obtain
x * z n ( 1 β n ) x * x n + β n x * T n x n = x * x n
and
x * y n ( 1 α n ) x * T n x n + α n x * T n z n ( 1 α n ) x * x n + α n x * z n
which implies that
x * y n x * x n .
By the definition of x n , we obtain
x * x n x * y n 1 + θ n 1 y n 2 y n 1 .
From (6) and (7), we obtain
x * y n x * y n 1 + θ n 1 y n 2 y n 1 .
It follows from (8) that
x * y n ( 1 + θ n 1 ) x * y n 1 + θ n 1 x * y n 2 .
Applying Lemma 3, we obtain x * y n M · j = 1 n ( 1 + 2 θ j ) , where M = max { x * y 1 , x * y 2 } . Since n = 1 θ n < , we obtain that { y n } is bounded and so are { z n } and { x n } . Thus,
n = 1 θ n y n y n 1 <
By Lemma 1 and (9), we obtain lim n x * y n exists. By Lemma 2(i) and the definition of z n , we obtain
x * z n 2 = ( 1 β n ) ( x * x n ) + β n ( x * T n x n ) 2 = ( 1 β n ) x * x n 2 + β n x * T n x n 2 ( 1 β n ) β n T n x n x n 2 x * x n 2 ( 1 β n ) β n T n x n x n 2 .
Let lim n x * y n = a . From the boundedness of { x n } and (6), we obtain
lim inf n x * x n a .
Since x * x n 1 x * y n + θ n y n y n 1 and (10), we obtain
lim sup n x * x n a .
It follows from (12) and (13) that
lim n x * x n = a .
From (4), one can easily see that lim sup n x * z n a . By (5) with α n 1 , we obtain that a lim inf n x * z n . Thus,
lim n x * z n = a .
Use the facts that (11), (14), and (15) yield
T n x n x n 0 .
According to { T n } satisfying the NST-condition (I) with T , we obtain that T x n x n 0 as n . Let ω w ( x n ) be the set of all weak cluster point of { x n } . Thus, ω w ( x n ) F ( T ) by demicloseness of I T at 0 . From Lemma 4, we conclude that x n x * with x * F ( T ) as required. □
Corollary 1. 
Let C be a nonempty closed and convex subset of a real Hilbert space H and let { T n } be a family of nonexpansive mappings of C into itself. Let y 0 , x 1 C , and { x n } be a sequence in H defined by Algorithm 1. Suppose that { T n } satisfies NST-condition (I) with T such that F ( T ) n = 1 F ( T n ) . Then, { x n } converges weakly to a point in F ( T ) .

4. Applications

In the past decade, extreme learning machine (ELM) [39], a new learning algorithm for single-hidden layer feedforward networks (SLFNs), has been extensively studied in various research topics for machine learning and artificial intelligence such as face classification, image segmentation, regression, and data classification problems. ELM was theoretically proven to have extremely fast learning speed and good performance better than the gradient-based learning such as backpropagation in most of the cases. The target of this model is to find the parameter β that solves the following minimization problem, called ordinary least square (OLS),
min β H β T ,
where · is the l 2 -norm defined by x 2 = i = 1 n | x i | 2 , T R N × m is the target of data, β R M × m is a weight which connects hidden layer and output layer, and H R N × M is the hidden layer output matrix. In general mathematical modeling, there are several methods to estimate the solution of (16); in this case, the solution β obtained by β = H T , where H is the Moore–Penrose generalized inverse of H . However, in a real situation, the number of unknown variable M is much more than the number of training data N, which causes the network to possibly lead to overfitting. On the other hand, the accuracy is low while the number of hidden nodes M is small. Thus, in order to improve (16), several regularization methods were introduced. The classical two standard techniques for improving (16) are subset selection and ridge regression (sometimes called Tikhonov regularization) [40].
In this paper, we focus on the following problem, called least absolute shrinkage and selection operator (LASSO) [41],
min β H β T 2 2 + λ β 1 ,
where λ is a regularization parameter. LASSO tries to retain the good features of both subset selection and ridge regression [41]. After the regularization methods and the original ELM were introduced for improving performance of OLS, five years later, the regularized extreme learning machine [42] was proposed and applied to solve regression problems. In general, the (17) can be rewritten as minimization of f + g , that is,
min x F ( x ) : = f ( x ) + g ( x ) ,
where f is a smooth convex function with gradient having Lipschitz constant L and g is a convex smooth (or possible non-smooth) function. The solution of (18) can be rewritten into x ˜ is a minimizer of ( f + g ) if and only if 0 f ( x ˜ ) + g ( x ˜ ) , where f ( x ˜ ) is the gradient of f and g ( x ˜ ) is a subdifferential of g by using Fermat’s rule (see [35] for more details). In fixed point theory, Parikh et al. [43] characterized (18) as follows: x ˜ is a minimizer of f + g if and only if
x ˜ = p r o x λ g ( I λ f ) ( x ˜ ) = J λ g ( I λ f ) ( x ˜ ) ,
where p r o x λ g is the proximity operator of λ g , λ > 0 and J g is defined by J g = ( I + g ) 1 , J g is the resolvent of g and I is an identity operator. The problem (18) can be rewritten into a general problem, called a zero of sum of two operators problem, by finding x ˜ such that
x ˜ z e r ( A + B ) ,
where A , B : H 2 H are two set-valued operators and z e r ( A + B ) : = { x : 0 A x + B x } . In this case, we assume that A : H 2 H is a maximal monotone operator and B : H H is an L-Lipschitz operator. For convenience, (19) also can be rewritten as:
x ˜ = T x ˜ ,
where T = p r o x λ g ( I λ f ) . It is also known that T is nonexpansive if λ ( 0 , 2 / L ) when L is a Lipschitz constant of f .
We are interested in applying our proposed method for solving a convex minimization problem and compared the convergence behavior of our proposed algorithm with the others and give some applications to solve classification problems. Our proposed method will be used to solve (18). Over the past two decades, several algorithms have been introduced for solving the problem (18). A simple and classical algorithm is the forward-backward algorithm (FBA), which was introduced by Lions and Mercier [21].
The forward-backward algorithm (FBA) is defined by
y n = x n γ f x n , x n + 1 = x n + ρ n ( J γ g y n x n ) ,
where n 1 , x 0 H and L is a Lipschitz constant of f , γ ( 0 , 2 / L ) , δ = 2 ( γ L / 2 ) and { ρ n } is a sequence in [ 0 , δ ] such that n N ρ n ( δ ρ n ) = + . A technique for improving speed and giving a better convergence behavior of the algorithms was introduced firstly by Polyak [44] by adding an inertial step called inertial techniques. Since then, many authors have employed the inertial technique to accelerate their algorithms for various kinds of problems; see [19,20,22,23,24,25,26]. The performance of FBA can be improved using an iterative method with the inertial steps described below.
A fast iterative shrinkage-thresholding algorithm (FISTA) [25] is defined by
y n = T x n , s n + 1 = 1 + 1 + 4 s n 2 2 , μ n = s n 1 s n + 1 , x n + 1 = y n + μ n ( y n y n 1 ) ,
where n 1 , s 1 = 1 , x 1 = y 0 R n , T : = p r o x 1 L g ( I 1 L f ) and μ n is the inertial step size. Beck and Teboulle [25] solved the image recovery and proved the convergence rate using FISTA. The inertial step size μ n of the FISTA was firstly introduced by Nesterov [45].
A fast iterative shrinkage-thresholding algorithm-Siteration (FISTA-S) [27] is defined by
s n + 1 = 1 + 1 + 4 s n 2 2 , μ n = s n 1 s n + 1 , y n = x n + μ n ( x n x n 1 ) , z n = ( 1 β n ) y n + β n T n y n , x n + 1 = ( 1 α n ) T n y n + α n T n z n ,
where x 0 , x 1 H , α n , β n [ a , b ] ( 0 , 1 ) T n = p r o x c n g ( I c n f ) and T = p r o x c g ( I c f ) . Bussaban et al. [27] solved the image recovery and proved the weak convergence theorem using FISTA-S.
A new accelerated proximal gradient algorithm (nAGA) [26] is defined by
y n = x n + μ n ( x n x n 1 ) , x n + 1 = T n [ ( 1 ρ n ) y n + ρ n T n y n ] ,
where n 1 , T n = p r o x a n g ( I a n f ) with a n ( 0 , 2 / L ) and { μ n } , { ρ n } are sequences in ( 0 , 1 ) and x n x n 1 2 μ n 0 . The nAGA was introduced for proving a convergence theorem by Verma and Shukla [26]. The nonsmooth convex minimization problem with sparsity, including regularizers, was solved using this method for the multitask learning framework.
Theorem 2. 
Let H be a Hilbert space, A : H 2 H be maximal monotone operator and B : H H be an L-Lipschitz operator. Let a ( 0 , 2 / L ) and { a n } ( 0 , 2 / L ) such that a n a . Define T n = J c n A ( I c n B ) and T = J c A ( I c B ) . Suppose that z e r ( A + B ) . Let { x n } be a sequence in H defined by Algorithm 1. Then, { x n } converges weakly to a point in z e r ( A + B ) .
Proof. 
Using Proposition 26.1(iv) (see [35]), we have { T n } and T are nonexpansive mappings such that F ( T ) = F ( T n ) = z e r ( A + B ) . Then, the proof is completed by Theorem 1 and Lemma 5. □
The convergence of Algorithm 2 is obtained by using our main result.
Algorithm 2: (FBASA) A forward-backward accelerated S-algorithm
1:
Initial. Take y 0 , x 1 C are arbitary, n = 1 , β n [ a , b ] ( 0 , 1 ) , θ n 0 , n = 1 θ n < and α n 1 .
2:
Step 1. Compute y n , z n and x n + 1 by using
z n = ( 1 β n ) x n + β n p r o x a n g ( I a n f ) y n , y n = ( 1 α n ) p r o x a n g ( I a n f ) + α n p r o x a n g ( I a n f ) z n , x n + 1 = y n + θ n ( y n y n 1 ) .
Then, n : = n + 1 and go to Step 1.
Theorem 3. 
For f , g : R n ( , ] , f is a smooth convex function with a gradient having a Lipschitz constant L and g is a convex function. Let a n ( 0 , 2 / L ) be such that { a n } converges to a and let T : = p r o x a g ( I a f ) and T n : = p r o x a n g ( I a n f ) and let { x n } be a sequence generated by Algorithm 2, where β n , α n and θ n are the same as in Algorithm 1. Then,
(i) x * x n + 1 M · j = 1 n ( 2 θ j + 1 ) , where M = m a x { x * x 1 , x * x 2 } and x * A r g m i n ( f + g ) ;
(ii) { x n } converges weakly to a point in Argmin ( f + g ) .
Proof. 
We know that T and { T n } are nonexpansive operators, and F ( T ) = n = 1 F ( T n ) = A r g m i n ( f + g ) for all n; see [35]. Then, { T n } satisfies the NST-condition (I) with T by using Lemma 5. We get the desired result immediately from Theorem 1 by putting G = R n × R n , the complete graph, on R n . □

5. Numerical Experiments

The classification problem is one of the most important problems in the convex minimization problem. We illustrate the process of reformulating the data classification problem in machine learning.
We first present a basic idea of extreme learning machines for data classification problem, and use our algorithm to find this problem through numerical experiments. Moreover, the performance of Algorithm 2, FISTA-S, FISTA, and nAGA are compared.
Extreme learning machine (ELM). Let R:= { ( x k , t k ) : x k R n , t k R m , k = 1 , 2 , , N } be a training set of different samples N, where x k is input data and t k is a target. A standard SLFNs with activation function Ψ ( x ) (for instance sigmoid), and M hidden nodes can be rewritten as
j = 1 M β j Ψ ( w j , x i + c j ) = o i , i = 1 , , N ,
where β j is the weight vector connecting the j-th the output node and hidden node, w j is the weight vector connecting the j-th the input node and hidden node, and c j is the threshold of the j-th hidden node. The objective of standard SLFNs is to estimate these N different samples with i = 1 N t i o i = 0 , that is, there exist β j , w j , c j such that
j = 1 M β j Ψ ( w j , x i + c j ) = t i , i = 1 , , N .
We can derive a simple equation from the above N equations as follows:
H β = T ,
H = Ψ ( w 1 , x 1 + c 1 ) Ψ ( w M , x 1 + c M ) Ψ ( w j , x N + c 1 ) Ψ ( w N , x N + c M ) ,
β = [ β 1 T , , β m × M T ] m × M T , T = [ t 1 T , , t N T ] m × N T .
A standard SLFN goal is to estimate β j , w j , and c j to solve (18), whereas an ELM goal is to find only β j with w j and c j chosen at random.
In an experiment on classification problems, we employ the model (17) to solve the convex minimization problem. We set f ( x ) = H β T 2 2 and g ( x ) = λ β 1 . Next, we use the Iris dataset to classify iris plant types, and the Heart Disease UCI dataset to identify heart patients which are detailed as follows:
Iris dataset [46]. This dataset has three classes of 50 examples, each of which represents a different variety of iris plant. The purpose is to identify each iris plant species based on the length of its sepals and petals.
Heart Disease UCI dataset [47]. Although there are 76 attributes in the original dataset, all published experiments only use 14 of them. Data on patients with heart disease are provided in this dataset. We divide the data into two classes based on the predicted attributes.
The dataset was graciously provided by https://archive.ics.uci.edu (accessed on April 2020).
All control parameters are set to the values shown in Table 1, L = 2 H 1 2 , where H 1 is a hidden layer output matrix of a training matrix, M = 100 , and Ψ ( x ) is sigmoid. Each dataset is given a training set, as indicated in Table 2. We evaluated the output data’s accuracy by
accuracy = correct   predicted   data all   data × 100
From the results in Table 3, we conclude that the proposed learning algorithm under selection with the identical number of hidden nodes M has high performance in terms of the accuracy. The weight computed by Algorithm 2 converges faster to the optimal weight and performs accuracy better than those computed by FISTA-S, FISTA, and nAGA.

6. Conclusions

We generate and study an algorithm for approximating the common fixed points of a countable family of G-nonexpansive mappings and prove the weak convergence of our algorithm. In addition, we give an application of our result for solving data classification and convex minimization problems. Finally, our numerical experiments assert that our proposed algorithm provides more accuracy than FISTA-S, FISTA, and nAGA.
In future work, we aim to find new models and methods for data prediction and classification of real datasets in medical science and create new innovations for health care service.

Author Contributions

Conceptualization, R.W.; Formal analysis, K.J. and R.W.; Investigation, K.J.; Methodology, R.W.; Supervision, R.W.; Validation, R.W.; Writing—original draft, K.J.; Writing—review and editing, R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The first author was supported by Fundamental Fund 2022, Chiang Mai University, Thailand, under the supervision of Suthep Suantai. The second author would like to thank Ubon Ratchathani University, Thailand.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Berinde, V. A Modified Krasnosel’skiǐ–Mann Iterative Algorithm for Approximating Fixed Points of Enriched Nonexpansive Mappings. Symmetry 2022, 14, 123. [Google Scholar] [CrossRef]
  2. Banach, S. Sur les oprations dans les ensembles abstraits et leur application aux quations intgrales. Fund. Math. 1922, 3, 133–181. [Google Scholar] [CrossRef]
  3. Jachymski, J. The contraction principle for mappings on a metric space with a graph. Proc. Am. Math. Soc. 2008, 136, 1359–1373. [Google Scholar] [CrossRef]
  4. Bojor, F. Fixed point of ψ-contraction in metric spaces endowed with a graph. Anna. Univ. Crai. Math. Comp. Sci. Ser. 2010, 37, 85–92. [Google Scholar]
  5. Chifu, C.; Petruşel, G. Generalized contractions in metric spaces endowed with a graph. Fixed Point Theory Appl. 2012, 2012, 161. [Google Scholar] [CrossRef] [Green Version]
  6. Acar, Ö.; Altun, I. Multivalued F-contractive mappings with a graph and some fixed point results. Publ. Math. Debr. 2016, 88, 305–317. [Google Scholar] [CrossRef]
  7. Acar, Ö.; Aydi, H.; De la Sen, M. New Fixed Point Results via a Graph Structure. Mathematics 2021, 9, 1013. [Google Scholar] [CrossRef]
  8. Aleomraninejad, S.M.A.; Rezapour, S.; Shahzad, N. Some fixed point result on a metric space with a graph. Topol. Appl. 2012, 159, 659–663. [Google Scholar] [CrossRef] [Green Version]
  9. Riech, S.; Zaslavski, A.J. Convergence of inexact iterative schemes for nonexpansive set-valued mappings. Fixed Point Theory Appl. 2010, 2010, 518243. [Google Scholar] [CrossRef] [Green Version]
  10. Alfuraidan, M.R.; Khamsi, M.A. Fixed points of monotone nonexpansive mappings on a hyperbolic metric space with a graph. Fixed Point Theory Appl. 2015, 2015, 44. [Google Scholar] [CrossRef]
  11. Alfuraidan, M.R. Fixed points of monotone nonexpansive mappings with a graph. Fixed Point Theory Appl. 2015, 2015, 49. [Google Scholar] [CrossRef] [Green Version]
  12. Suantai, S.; Kankam, K.; Cholamjiak, P.; Cholamjiak, W. A parallel monotone hybrid algorithm for a finite family of G-nonexpansive mappings in Hilbert spaces endowed with a graph applicable in signal recovery. Comp. Appl. Math. 2021, 40, 145. [Google Scholar] [CrossRef]
  13. Tiammee, J.; Kaewkhao, A.; Suantai, S. On Browder’s convergence theorem and Halpern iteration process for G-nonexpansive mappings in Hilbert spaces endowed with graphs. Fixed Point Theory Appl. 2015, 2015, 187. [Google Scholar] [CrossRef] [Green Version]
  14. Tripak, O. Common fixed points of G-nonexpansive mappings on Banach spaces with a graph. Fixed Point Theory Appl. 2016, 2016, 87. [Google Scholar] [CrossRef] [Green Version]
  15. Sridarat, P.; Suparaturatorn, R.; Suantai, S.; Cho, Y.J. Convergence analysis of SP-iteration for G-nonexpansive mappings with directed graphs. Bull. Malays. Math. Sci. Soc. 2019, 42, 2361–2380. [Google Scholar] [CrossRef]
  16. Glowinski, R.; Tallec, P.L. Augmented Lagrangian and Operator-Splitting Methods in Nonlinear Mechanic; SIAM: Philadelphia, PA, USA, 1989. [Google Scholar]
  17. Haubruge, S.; Nguyen, V.H.; Strodiot, J.J. Convergence analysis and applications of the Glowinski Le Tallec splitting method for finding a zero of the sum of two maximal monotone operators. J. Optim. Theory Appl. 1998, 97, 645–673. [Google Scholar] [CrossRef]
  18. Noor, M.A. New approximation schemes for general variational inequalities. J. Math. Anal. Appl. 2000, 251, 217–229. [Google Scholar] [CrossRef] [Green Version]
  19. Hanjing, A.; Suantai, S. A fast image restoration algorithm based on a fixed point and optimization method. Mathematics 2020, 8, 378. [Google Scholar] [CrossRef] [Green Version]
  20. Bussaban, L.; Suantai, S.; Kaewkhao, A. A parallel inertial S-iteration forward-backward algorithm for regression and classification problems. Carpathian J. Math. 2020, 36, 21–30. [Google Scholar] [CrossRef]
  21. Lions, P.L.; Mercier, B. Splitting Algorithms for the Sum of Two Nonlinear Operators. SIAM J. Numer. Anal. 1979, 16, 964–979. [Google Scholar] [CrossRef]
  22. Janngam, K.; Suantai, S. An accelerated forward-backward algorithm with applications to image restoration problems. Thai J. Math. 2021, 19, 325–339. [Google Scholar]
  23. Alakoya, T.O.; Jolaoso, L.O.; Mewomo, O.T. Two modifications of the inertial Tseng extragradient method with self-adaptive step size for solving monotone variational inequality problems. Demonstr. Math. 2020, 53, 208–224. [Google Scholar] [CrossRef]
  24. Gebrie, A.G.; Wangkeeree, R. Strong convergence of an inertial extrapolation method for a split system of minimization problems. Demonstr. Math. 2020, 53, 332–351. [Google Scholar] [CrossRef]
  25. Beck, A.; Teboulle, M. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef] [Green Version]
  26. Verma, M.; Shukla, K. A new accelerated proximal gradient technique for regularized multitask learning framework. Pattern Recogn. Lett. 2018, 95, 98–103. [Google Scholar] [CrossRef]
  27. Bussaban, L.; Kaewkhao, A.; Suantai, S. Inertial s-iteration forward-backward algorithm for a family of nonexpansive operators with applications to image restoration problems. Filomat 2021, 35, 771–882. [Google Scholar] [CrossRef]
  28. Janngam, K.; Wattanataweekul, R. An Accelerated Fixed-Point Algorithm with an Inertial Technique for a Countable Family of G-Nonexpansive Mappings Applied to Image Recovery. Symmetry 2022, 14, 662. [Google Scholar] [CrossRef]
  29. Johnsonbaugh, R. Discrete Mathematics; Pearson: Hoboken, NJ, USA, 1997. [Google Scholar]
  30. Suantai, S.; Donganont, M.; Cholamjiak, W. Hybrid Methods for a Countable Family of G-Nonexpansive Mappings in Hilbert Spaces Endowed with Graphs. Mathematics 2019, 7, 936. [Google Scholar] [CrossRef] [Green Version]
  31. Tan, K.; Xu, H.K. Approximating fixed points of nonexpansive mappings by the ishikawa iteration process. J. Math. Anal. Appl. 1993, 178, 301–308. [Google Scholar] [CrossRef] [Green Version]
  32. Takahashi, W. Introduction to Nonlinear and Convex Analysis; Yokohama Publishers: Yokohama, Japan, 2009. [Google Scholar]
  33. Moudafi, A.; Al-Shemas, E. Simultaneous iterative methods for split equality problem. Trans. Math. Program. Appl. 2013, 1, 1–11. [Google Scholar]
  34. Nakajo, K.; Shimoji, K.; Takahashi, W. Strong convergence to a common fixed point of families of nonexpansive mappings in Banach spaces. J. Nonlinear Convex Anal. 2007, 8, 11–34. [Google Scholar]
  35. Bauschke, H.H.; Combettes, P.L. Convex Analysis and Monotone Operator Theory in Hilbert Spaces, 2nd ed.; Springer: Berlin, Germany, 2017. [Google Scholar]
  36. Combettes, P.L.; Wajs, V.R. Signal recovery by proximal forward-backward splitting. Multiscale Model. Simul. 2005, 4, 168–1200. [Google Scholar] [CrossRef] [Green Version]
  37. Moreau, J.J. Fonctions convexes duales et points proximaux dans un espace hilbertien. C. R. Acad. Sci. Paris Ser. A Math. 1962, 255, 2897–2899. [Google Scholar]
  38. Beck, A. First-Order Methods in Optimization; Tel-Aviv University: Tel Aviv-Yafo, Israel, 2017; pp. 129–177. ISBN 978-1-61197-498-0. [Google Scholar]
  39. Huang, G.-B.; Zhu, Q.-Y.; Siew, C.-K. Extreme learning machine: Theory and applications. Neurocomputing 2006, 70, 489–501. [Google Scholar] [CrossRef]
  40. Tikhonov, A.N.; Arsenin, V.Y. Solutions of Ill-Posed Problems; V.H. Winston: Washington, DC, USA, 1977. [Google Scholar]
  41. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. B Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  42. Martnez-Martnez, J.M.; Escandell-Montero, P.; Soria-Olivas, E.; Martn-Guerrero, J.D.; MagdalenaBenedito, R.; Gmez-Sanchis, J. Regularized extreme learning machine for regression problems. Neurocomputing 2011, 74, 3716–3721. [Google Scholar] [CrossRef]
  43. Parikh, N.; Boyd, S. Proximal Algorithms. Found. Trends Optim. 2014, 1, 127–239. [Google Scholar] [CrossRef]
  44. Polyak, B. Some methods of speeding up the convergence of iteration methods. USSR Comput. Math. Math. Phys. 1964, 4, 1–17. [Google Scholar] [CrossRef]
  45. Nesterov, Y. A method for solving the convex programming problem with convergence rate O(1/k2). Dokl. Akad. Nauk SSSR 1983, 269, 543–547. [Google Scholar]
  46. Dua, D.; Karra, E. Taniskidou, UCI Machine Learning Repository; University of California Irvinea: Irvinea, CA, USA, 2017. [Google Scholar]
  47. Lichman, M. UCI Machine Learning Repository; University of California Irvinea: Irvinea, CA, USA, 2013. [Google Scholar]
Table 1. Parameters of each methods.
Table 1. Parameters of each methods.
MethodsSetting
Algorithm 2 α n = β n = 0.5 , c = 1 / L , θ n = 0.9 if
1 n 400 , and 1 / 2 n otherwise
FISTA-S α n = β n = 0.5 , c = 1 / L , s 1 = 1 ,
s n + 1 = ( 1 + 1 + 4 s n 2 ) / 2 , ϱ n = ( s n 1 ) / s n + 1
if 1 n 400 , and 1 / 2 n otherwise
FISTA s 1 = 1 , s n + 1 = ( 1 + 1 + 4 s n 2 ) / 2 ,
ϱ n = ( s n 1 ) / s n + 1
nAGA ρ n = 0.9 , γ = 1 / L , s 1 = 1 ,
s n + 1 = ( 1 + 1 + 4 s n 2 ) / 2 , ϱ n = ( s n 1 ) / s n + 1
Table 2. Training and testing sets of the Iris and Heart Disease UCI datasets.
Table 2. Training and testing sets of the Iris and Heart Disease UCI datasets.
DatasetAttributesSample
TrainTest
Heart Disease UCI1421390
Iris410545
Table 3. Performance comparison using different methods at the 400th iteration.
Table 3. Performance comparison using different methods at the 400th iteration.
DatasetAlgorithm 2FISTA-SFISTAnAGA
TrainTestTrainTestTrainTestTrainTest
Heart Disease UCI68.2662.3739.9637.6353.7650.5067.8360.22
Iris10098.1095.5694.2995.5694.2994.2993.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Janngam, K.; Wattanataweekul, R. A New Accelerated Fixed-Point Algorithm for Classification and Convex Minimization Problems in Hilbert Spaces with Directed Graphs. Symmetry 2022, 14, 1059. https://0-doi-org.brum.beds.ac.uk/10.3390/sym14051059

AMA Style

Janngam K, Wattanataweekul R. A New Accelerated Fixed-Point Algorithm for Classification and Convex Minimization Problems in Hilbert Spaces with Directed Graphs. Symmetry. 2022; 14(5):1059. https://0-doi-org.brum.beds.ac.uk/10.3390/sym14051059

Chicago/Turabian Style

Janngam, Kobkoon, and Rattanakorn Wattanataweekul. 2022. "A New Accelerated Fixed-Point Algorithm for Classification and Convex Minimization Problems in Hilbert Spaces with Directed Graphs" Symmetry 14, no. 5: 1059. https://0-doi-org.brum.beds.ac.uk/10.3390/sym14051059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop