• No results found

Mathematical programming for the sum of two convex functions with applications to lasso problem, split feasibility problems, and image deblurring problem

N/A
N/A
Protected

Academic year: 2020

Share "Mathematical programming for the sum of two convex functions with applications to lasso problem, split feasibility problems, and image deblurring problem"

Copied!
23
0
0

Loading.... (view fulltext now)

Full text

(1)

R E S E A R C H

Open Access

Mathematical programming for the sum

of two convex functions with applications to

lasso problem, split feasibility problems, and

image deblurring problem

Chih Sheng Chuang

1

, Zenn-Tsun Yu

2

and Lai-Jiu Lin

3*

*Correspondence: maljlin@cc.ncue.edu.tw 3Department of Mathematics, National Changhua University of Education, Changhua, 50058, Taiwan

Full list of author information is available at the end of the article

Abstract

In this paper, two iteration processes are used to find the solutions of the

mathematical programming for the sum of two convex functions. In infinite Hilbert space, we establish two strong convergence theorems as regards this problem. As applications of our results, we give strong convergence theorems as regards the split feasibility problem with modified CQ method, strong convergence theorem as regards the lasso problem, and strong convergence theorems for the mathematical programming with a modified proximal point algorithm and a modified

gradient-projection method in the infinite dimensional Hilbert space. We also apply our result on the lasso problem to the image deblurring problem. Some numerical examples are given to demonstrate our results. The main result of this paper entails a unified study of many types of optimization problems. Our algorithms to solve these problems are different from any results in the literature. Some results of this paper are original and some results of this paper improve, extend, and unify comparable results in existence in the literature.

MSC: 90C33; 90C34; 90C59

Keywords: lasso problem; mathematical programming for the sum of two functions; split feasibility problem; gradient-projection algorithm; proximal point algorithm

1 Introduction

LetRbe the set of real numbers,HandHbe (real) Hilbert spaces with inner product·,·

and norm·. LetCandQbe nonempty, closed, convex subsets ofHandH, respectively.

Let(H) be the space of all proper lower semicontinuous convex functions fromH to

(–∞,∞]. In this paper, we consider the following minimization problem:

arg min

xH

h(x) +g(x), (.)

whereh,g∈(H), andg:H→Ris a Fréchet differential function.

LetAbe am×nreal matrix,x∈Rn,bRm,γ  be a regularization parameter and

t≥. Tibshirani [] studied the following minimization problem:

min

x

Axb

 subject tox≤t, (.)

(2)

wherex= n

i=|xi|forx= (x,x, . . . ,xn)∈Rn, andy= m

i=(yi)fory= (y,y, . . . ,

ym)∈Rm. Problem (.) is called lasso, an abbreviation of ‘the least absolute shrinkage and

select operator’. This is a special case of problem (.). Besides, we know that problem (.) is equivalent to the following problem:

min

x

Axb

+γx. (.)

Therefore, problems (.) and (.) are special cases of problem (.).

Due to the involvement of thenorm, which promotes the sparsity phenomena of many

real world problems arising from image/signal processing, statistical regression, machin-ing learnmachin-ing and so on, the lasso receives much attention (see Combettes and Wajs [], Xu [], and Wang and Xu []).

Letf :H→(–∞,∞] be proper and letCbe a nonempty subset ofdom(f). Thenfis said to be uniformly convex onCif

fαx+ ( –α)y+α( –α)pxyαf(x) + ( –α)f(y)

for allx,yC, and for allα∈(, ), wherep: [,∞)→[,∞] is an increasing function,p vanishes only at .

Letg(H) andλ∈(,∞). The proximal operator ofgis defined by

proxgx=arg min

vH

g(v) +  vx

, xH.

The proximal operator ofg(H) of orderλ∈(,∞) is defined as the proximal operator

ofλg, that is,

proxλgx=arg min

vH

g(v) +λvx

, xH.

The following results are important results on the solution of the problem (.).

Theorem .(Douglas-Rachford-algorithm) [] Let f and g be functions in(H)such

that(∂f +∂g)–=.Let{λ

n}n∈Nbe a sequence in[, ]such thatn∈Nλn( –λn) = +∞.

Letγ ∈(,∞),and x∈H.Set

⎪ ⎨ ⎪ ⎩

yn=proxγgxn,

zn=proxγf(ynxn),

xn+=xn+λn(znxn), n∈N.

Then there exists xH such that the following hold:

(i) proxγgx∈arg minxH(f+g)(x); (ii) {ynzn}n∈Nconverges strongly to;

(iii) {xn}n∈Nconverges weakly tox;

(iv) {yn}n∈Nand{zn}n∈Nconverge weakly toproxγgx;

(v) suppose that one of the following holds:

(a) f is uniformly convex on every nonempty subset ofdom∂f;

(3)

Then{yn}n∈Nand{zn}n∈Nconverge strongly toproxγgx,which is the unique minimizer of

f +g.

Theorem . (Forward-backward algorithm) [] Let f(H), let g:H→Rbe

con-vex and differentiable with a

β-Lipschitz continuous gradient for some β ∈(,∞), let

γ ∈(, β),and setδ=min{,βγ}+ .Furthermore, let{λn}n∈N be a sequence in(, δ]

such thatnNλn(δ–λn) = +∞.Suppose thatarg min(f +g)=∅and let

yn=xnγg(xn),

xn+=xn+λn(proxγfynxn), n∈N.

Then the following hold:

(i) (xn)n∈Nconverges weakly to a point inarg minxH(f +g)(x);

(ii) suppose thatinfn∈Nλn∈(,∞)and one of the following hold:

(a) f is uniformly convex on every nonempty bounded subset ofdom∂f; (b) gis uniformly convex on every nonempty bounded subset ofH.

Then{xn}n∈Nconverges strongly to the unique minimizer of f +g

Theorem .(Tseng’s algorithm) [] Let D be a nonempty subset of H,f(H)be such

thatdom∂fD,and let g(H)be such that Gâteaux differentiable on D.Suppose that

C is a nonempty,closed,convex subset of D such that C∩arg minxH(f+g)(x)=∅,and that

g is β-Lipschitz continuous relative to C∪dom∂f for someβ∈(,∞).Let x∈C and

γ ∈(,β),and set

⎧ ⎪ ⎪ ⎪ ⎨ ⎪ ⎪ ⎪ ⎩

yn=xn–∇g(xn),

zn=proxγfyn,

rn=znγg(zn),

xn+=PC(xnyn+rn), n∈N.

Then

(a) {xn}n∈Nand{zn}n∈Nconverges weakly to a point inC∩arg minxH(f +g)(x);

(b) suppose thatf orgis uniformly convex on every nonempty subset ofdom∂f.

Then{xn}n∈Nand{zn}n∈Nconverges strongly in C∩arg minxH(f +g)(x).

Combettes and Wajs [] used the proximal gradient method to generalize a sequence

{xn}by the algorithm:x∈His chosen arbitrarily, and

xn+=proxλng(I–λnf)xn.

We observed that Combettes and Wajs [] showed that{xn}converges weakly to a

solu-tion of the minimizasolu-tion problem (.) under suitable condisolu-tions. In , Xu [] gave an iteration process and proved weak convergence theorems to the solution for the problem (.). Next, Wang and Xu [] studied problem (.) by the following two types of iteration processes:

(4)

For these two iteration processes, Wang and Xu [] proved that they converge strongly to a solution of the problem (.) under suitable conditions.

Let I be the identity function ofH, and letf:C×CH be a bifunction. Letg :

H→(–∞,∞] be a proper convex Fréchet differentiable function with Fréchet derivative

g onint(dom(g)),C⊂int(dom(g)), andh:H→(–∞,∞] be a proper convex lower

semicontinuous function. LetPC be the metric projection ofHintoC. Throughout this

paper, we use these notations unless specified otherwise.

Motivated by the results of the above problems, in this paper, we introduce the following iterations to study problem (.).

Iteration (I) Letx∈Cbe chosen arbitrarily, and

yn=J Af λ PCxn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn), n∈N,

wheref(x,y) =yx,g(x)+h(y) –h(x).

Iteration (II) Letx∈Cbe chosen arbitrarily, and

yn=Jr∂h(I–rg)xn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn), n∈N.

Then we establish two strong convergence theorems without the uniformly convex as-sumption on the functions we consider. Our results improve Combettes and Wajs [], Xu [], Douglas-Rachford [], Theorem ., and Tseng []. Our results is also different from Wang and Xu [].

We also apply our results to study the following problems.

(AP) Split feasibility problem:

Findx¯∈Hsuch thatx¯∈CandAx¯∈Q, (SFP)

whereA:HHis a linear and bounded operator.

In , the split feasibility problem (SFP) in finite dimensional Hilbert spaces was first introduced by Censor and Elfving [] for modeling inverse problems which arise from phase retrievals and in medical image reconstruction. Since then, the split feasibility problem (SFP) has received much attention due to its applications in signal processing, image reconstruction, with particular progress in intensity-modulated radiation therapy, approximation theory, control theory, biomedical engineering, communications, and geo-physics. For examples, one can refer to [–] and related literature.

In , Byrne [] first introduced the so-called CQ algorithm which generates a se-quence{xn}by the following recursive procedure:

xn+=PC

xnρnA∗(I–PQ)Axn

,

where the stepsizeρnis chosen in the interval (, /A), andPC andPQare the

(5)

method to study the split feasibility problem in finite dimensional spaces, but in the in-finite dimensional Hilbert space, a strong convergence theorem may not be true for the split feasibility problem by the CQ algorithm []. Hence, some modified CQ algorithms are introduced.

In this paper, we give a new iteration to study the (SFP), we also give our modified CQ method to study (SFP), we study two strong convergence theorem to the solution of this problem. We establish a strong convergence of this problem. Our results are different from Theorem . of Xu [] and improve other results of Xu [].

(AP) Lasso problem:

arg min

x∈Rn

Axb

+γx.

We give two iterations to study the lasso problem, and we establish two strong convergence theorems of the lasso problem.

(AP) Mathematical programming for convex function:

arg min

xH

h(x),whereh∈(H).

Rockafellar [] used a proximal point algorithm and proved a weak convergence theo-rem to the solution of this problem. We establish a modified proximal point algorithm and prove a strongly convergence theorem to study this problem. Our result improves the results given by Rockafellar [].

(AP) Mathematical programming for convex Fréchet differentiable function:

Findarg min

xH

g(x),whereg:H→Ris a Fréchet differentiable function.

Recently, Xu [] studied weak convergence and strong convergence theorem of this prob-lem with various types of relaxed gradient-projection iterations. He also used the viscosity nature of the gradient-projection method and the regularized method to study strong con-vergence theorems of this problem.

A special case of one of our iteration is modified gradient-projection algorithm. We use this modified gradient-projection algorithm to establish a strong convergence theorem of this problem (AP), and our results improve recent results given by Xu in [].

In this paper, we apply a recent result of Yu and Lin [] to find the solution of the math-ematical programming for two convex functions, then we apply our results on mathemat-ical programming for two convex functions to study the above problems. We establish a strongly convergent theorem for these problems and apply our result on the lasso problem to the image deblurring problem. Some numerical examples are given to demonstrate our results. The main result of this paper gives a unified study of many types of optimization problems. Our algorithms to solve these problems are different from any results in the literature. Some results of this paper are original and some results of this paper improve, extend, and unify comparable results existence in the literature.

2 Preliminaries

Throughout this paper, we denote the strong convergence of{xn}toxHbyxnx. Let

(6)

(i) Tis called nonexpansive ifTxTyxyfor allx,yC.

(ii) Tis strongly monotone if there existsγ¯> such thatxy,TxTy ≥ ¯γxy for allx,yC.

(iii) Tis Lipschitz continuous if there existsL> such thatTxTyLxyfor all

x,yC.

(iv) Letα> . ThenTisα-inverse-strongly monotone ifxy,TxTyαTxTy for allx,yC. We denoteT isα-ism if T isα-inverse-strongly monotone.

(v) Tis firmly nonexpansive if

TxTy≤ xy–(I–T)x– (I–T)y for everyx,yC,

that is,

TxTy≤ xy,TxTy for everyx,yC.

LetB:HHbe a multivalued mapping. The effective domain ofBis denoted byD(B), that is,D(B) ={xH:Bx=∅}. Thus:

(i) Bis a monotone operator onHifxy,uv ≥for allx,yD(B),uBxand

vBy.

(ii) Bis a maximal monotone operator onHifBis a monotone operator onHand its graph is not properly contained in the graph of any other monotone operator onH.

Lemma .[] Let G:HH be maximal monotone.Let JrGbe the resolvent of G

de-fined by JG

r = (I+rG)–for each r> .Then the following hold: (i) For eachr> ,JG

r is single-valued and firmly nonexpansive.

(ii) D(JG

r ) =HandFix(JrG) ={xD(G) : ∈Gx}.

LetCbe a nonempty, closed, convex subset of a real Hilbert spaceH. Letg:C×C→R be a function. The Ky Fan inequality problem [] is to findzCsuch that

g(z,y)≥ for eachyC. (EP)

The solution set of Ky Fan inequality problem (KF) is denoted byKF(C,g).

For solving the Ky Fan inequalities problem, let us assume that the bifunctiong:C× C→Rsatisfies the following conditions:

(A) g(x,x) = for eachxC;

(A) gis monotone,i.e.,g(x,y) +g(y,x)≤for anyx,yC; (A) for eachx,y,zC,lim suptg(tz+ ( –t)x,y)g(x,y);

(A) for eachxC, the scalar functionyg(x,y)is convex and lower semicontinuous.

We have the following result from Blum and Oettli [].

Lemma .[] Let g:C×C→Rbe a bifunction which satisfies conditions(A)-(A).

Then for each r> and each xH,there exists zC such that

g(z,y) +

ryz,zx ≥

(7)

In , Combettes and Hirstoaga [] established the following important properties of resolvent operator.

Lemma .[] Let g:C×C→Rbe a function satisfying conditions(A)-(A).For r> , define Trg:HC by

Trgx=

zC:g(z,y) +

ryz,zx ≥,∀yC

for all xH.Then the following hold:

(i) Trg is single-valued;

(ii) Trgis firmly nonexpansive,that is,TrgxTrgy≤ xy,TrgxTrgyfor allx,yH;

(iii) {xH:Trgx=x}={xC:g(x,y)≥,∀yC};

(iv) {xC:g(x,y)≥,∀yC}is a closed and convex subset ofC.

We call suchTrg the resolvent ofgforr> .

Takahashiet al.[] gave the following lemma.

Lemma .[] Let g:C×C→Rbe a bifunction satisfying the conditions(A)-(A).

Define Agas follows:

Agx=

{zH:g(x,y)yx,z,∀yC} if xC,

if x∈/C. (L.)

ThenKF(C,g) =A–

gand Agis a maximal monotone operator with the domain of AgC.

Furthermore,for any xH and r> ,the resolvent Trg of g coincides with the resolvent of

Ag,i.e.,Trgx= (I+rAg)–x.

Letf :H→(–∞,∞] be proper. The subdifferential∂f off is the set valued operator, defined∂f ={uH:f(y)≥f(x) +yx,ufor allyH}.

LetxH. Thenf is subdifferentiable atxif∂f(x)=∅. Then the elements of∂f(x) are called the subgradient off atx.

The directional derivative off atxin the directionyis

f(x,y) =lim

α↓

f(x+ay) –f(x)

α ,

provided that the limit exists in [–∞,∞].

Let x∈domf and suppose that f(x,y) is linear and continuous, thenf is said to be Gâteaux differentiable atx. By the Riesz representation theorem, there exists a unique vector∇f(x)∈Hsuch thatf(x,y) =y,f(x)for allyH.

LetxH, letμ(x) denote the family of all neighborhood ofx, letHbe a Hilbert space,

letCμ(x) and letf :C→(–∞,∞]. Thenf is said to be Fréchet differentiable atxif there exists an operator∇f(x)∈B(H,R), called the Fréchet derivative off atx, such that

lim

=y→

|f(x+y) –f(x) –y,f(x)|

y = .

(8)

LetCbe a nonempty, closed, convex subset ofH. The indicator functionιCdefined by

ιCx=

, xC,

∅, x∈/C,

is a proper lower semicontinuous convex function and its subdifferential∂ιCdefined by

∂ιCx=

zH:yx,zιC(y) –ιC(x),∀yH

is a maximal monotone operator (see Lemma .). Furthermore, we also define the normal coneNCuofCatuas follows:

NCu=

zH:z,vu ≤,∀vC.

We can define the resolventJ∂iC

λ of∂iCforλ> ,i.e.,

J∂iC

λ x= (I+λ∂iC)–x

for allxH. Since

∂iCx=

zH:iCx+z,yxiCy,yH

=zH:z,yx ≤,∀yC

=NCx

for allxC, we have

u=J∂iC

λ xxu+λ∂iCu

xuλNCu

xu,yu ≤, ∀yC

u=PCx.

For details see [].

Lemma .[] Let g(H)andλ∈(,∞).Thus:

(i) IfCis a nonempty,closed,convex subset ofHandg=iCis the indicator function of

C,then the proximal operatorproxλg=PCfor allλ∈(,∞),wherePCis the metric

projection operator fromHtoC. (ii) proxλgis firmly nonexpansive. (iii) proxλg= (I+λ∂g)–=J∂g

λ .

Lemma .[] Let f,g(H).Let x∗∈H andλ∈(,∞).Assume that f is finite valued

and Fréchet differentiable function on H with Fréchet derivativef.Then xis a solution to the problemarg minxHf(x) +g(x)if and only if x∗=proxλg(I–λf)x∗.

Lemma .[] Let CH be nonempty,closed,convex subset,let A:HH,and let

(9)

Lemma .[] Let f(H),then∂f is maximum monotone.

Lemma .[] Let f and g be functions in(H)such that one of the following holds: (i) domf ∩int(dom)g=∅;

(ii) domg=H.

Then∂(f +g) =∂f +∂g.

A mapping:HHis said to be averaged if= ( –α)I+αT, whereα∈(, ) and

T :HH is nonexpansive. In this case, we say thatisα-averaged. Clearly, a firmly

nonexpansive mapping is-averaged.

Lemma .[] Let T:HH be a mapping.Then the following hold:

(i) Tis nonexpansive if and only if the complement(I–T)is/-ism; (ii) ifSisυ-ism,then forγ > ,γSisυ/γ-ism;

(iii) Sis averaged if and only if the complementISisυ-ism for someυ> /; (iv) ifSandTare both averaged,then the product(composite)ST is averaged; (v) if the mappings {Ti}ni=are averaged and have a common fixed point,then

n

i=Fix(Ti) =Fix(T· · ·Tn).

Lemma .[] Let f :H→(–∞,∞]be proper and convex.Suppose that f is Gâteaux

differentiable at x.Then∂f(x) ={∇f(x)}.

3 Common solution of variational inequality problem, fixed point, and Ky Fan inequalities problem

For eachi= , , letFi:CHbe aκi-inverse-strongly monotone mapping ofCintoH

withκi> . For eachi= , , letGibe a maximal monotone mapping onHsuch that the

domain of Gi is included inCand define the setG–i  asGi– ={xH: ∈Gix}. Let

JG

λ = (I+λG)– andJrG = (I+rG)–for each n∈N, λ>  andr> . Let{θn} ⊂Hbe

a sequence. LetV be aγ¯-strongly monotone andL-Lipschitz continuous operator with

¯

γ >  andL> . LetT:CHbe a nonexpansive mapping. Throughout this paper, we use these notations and assumptions unless specified otherwise.

In this paper, we say conditions (D) hold if the following conditions are satisfied:

(i)  <lim infn→∞αn≤lim supn→∞αn< ;

(ii) limn→∞βn= , and

n=βn=∞;

(iii) limn→∞θn= .

The following strong convergence theorem is needed in this paper.

Theorem .[] Suppose that=Fix(T)∩Fix(JλG(I–λF))∩Fix(JrG(I–rF))=∅.Take

μ∈Rsuch that <μ< Lγ¯.A sequence{xn} ⊂H is defined as follows:x∈C is chosen

arbitrarily,

⎧ ⎪ ⎨ ⎪ ⎩

yn=JλG(I–λF)JrG(I–rF)xn,

sn=Tyn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)sn)

for each n∈N,{λ,r} ⊂(,∞),{αn} ⊂(, ),and{βn} ⊂(, ).Assume that conditions(D)

hold and <λ< κand <r< κ.Then

lim

(10)

where

¯

x=PFix

(T)∩Fix(JλG(IλF))∩Fix(JGr(IrF))(x¯–Vx).¯

This pointx is also the unique solution to the hierarchical variational inequality:¯

Vx,¯ qx¯ ≥, ∀q∈Fix(T)∩FixJG

λ (I–λF)

∩FixJG

r (I–rF)

.

For eachi= , , letfi:C×C→Rbe a bifunction satisfying conditions (A)-(A). An

iteration is used to find common solutions of a variational inequality problem, Ky Fan inequalities problems, and a fixed point set of a mapping:

Findx¯∈Hsuch thatx¯∈Fix(T)∩KF(C,f)∩KF(C,f) and Vx,¯ qx¯ ≥, ∀q∈Fix(T)∩KF(C,f)∩KF(C,f).

Theorem . For each i= , ,let fi:C×C→Rbe a bifunction satisfying conditions

(A)-(A),and let Afibe defined as(L.)in Lemma..Suppose that:=Fix(T)∩KF(C,f)∩

KF(C,f)=∅.Takeμ∈Rsuch that <μ<Lγ¯.A sequence{xn} ⊂H is defined as follows:

x∈C chosen arbitrarily,and

⎪ ⎨ ⎪ ⎩

yn=J Af

λ J

Af

r xn,

sn=Tyn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)sn)

for each n∈N,{λ,r} ⊂(,∞),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ This pointx¯∈is also the unique solution to the hierarchical variational inequality:

Vx,¯ qx¯ ≥, ∀q.

Proof For eachi= , , by Lemma ., we know thatAfiis a maximal monotone operator

with the domain ofAfiCandKF(C,fi) =Af–i. For eachi= , , letFi= , andGi=Afi

in Theorem .. By Lemma .(ii), we have, for eachi= , ,

FixJGi

λ (I–λFi)

=Gi– =A–fi  =KF(C,fi).

This implies that=. By Theorem .,limn→∞xn=x, where¯ x¯=P(x¯–Vx). This¯ pointx¯∈is also the unique solution to the hierarchical variational inequality:

Vx,¯ qx¯ ≥, ∀q.

Thus,

¯

x∈Fix(T)∩KF(C,f)∩KF(C,f)

and

Vx,¯ qx¯ ≥, ∀q∈Fix(T)∩KF(C,f)∩KF(C,f).

(11)

As a simple consequence of Theorem ., we study the common solution of the Ky Fan inequalities problems.

Theorem . Let f:C×C→Rbe a bifunction satisfying conditions(A)-(A)and let

Afbe defined as(L.)in Lemma..Suppose that:=KF(C,f)=∅.Takeμ∈Rsuch

that <μ<Lγ¯.A sequence{xn} ⊂H is defined as follows:x∈C is chosen arbitrarily,and

yn=J Af λ PCxn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,{λ,r} ⊂(,∞),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ This pointx¯∈KF(C,f)is also the unique solution to the hierarchical variational inequality:

Vx,¯ qx¯ ≥, ∀q.

Proof LetI|CandiCbe the restriction of the identity function onCand the indicate

func-tion onCrespectively and letT=I|C,f=iC in Theorem ., then Theorem . follows

from Theorem ..

Theorem . Let:=Fix(T)=∅.Takeμ∈Rsuch that <μ<Lγ¯.A sequence{xn} ⊂H

is defined as follows:x∈C is chosen arbitrarily,and

sn=TPCxn,

xn+=αnxn+ ( –αn)(βnθn+ ( –βn)sn)

for each n∈N,and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Thenlimn→∞xn=x,¯

wherex¯=P(x¯–Vx).¯ This pointx¯∈Fix(T)is also the unique solution to the hierarchical variational inequality:

Vx,¯ qx¯ ≥, ∀q.

Proof For eachi= , , letfi=iC,Afi=∂iCin Theorem ., whereiCis the indicator

func-tion ofC. ThenKF(C,fi) =CandJ Afi

r =PC. Therefore, Theorem . follows immediately

from Theorem ..

4 Mathematical programming for the sum of two convex functions

In the following theorem, an iteration is used to find the solution of the optimization prob-lem for the sum of two convex functions:

arg min

yC

(g+h)(y).

Theorem . Let g :H →(–∞,∞) be a convex Fréchet differentiable function with

Fréchet derivativegon H,and h:H→(–∞,∞]be a proper convex lower

(12)

defined as(L.)in Lemma.. Takeμ∈Rsuch that  <μ< Lγ¯.Suppose that :=

H–VI(C,∇g,h)=∅,where

H–VI(C,∇g,h) =

xC:yx,gx+h(y) –h(x)≥for all yC

.

A sequence{xn} ⊂H is defined as follows:x∈C is chosen arbitrarily,and

yn=J Af

λ PCxn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,λ∈(,∞),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ Further,x is the unique solution to the hierarchical¯ variational inequality:

Vx,¯ qx¯ ≥, ∀q.

Proof Since∇gis the Fréchet derivative of the convex functiong, it follows from

Corol-lary . of [] that∇gis continuous onC. By Proposition . of [],gis monotone

onC. Hence,gis bounded on any line segment ofC. By Proposition . of [],

yx,g(x)

g(y) –g(x) for allx,yC. (.)

Sinceh:C→Ris a proper convex lower semicontinuous function, it is easy to see that

for eachx,y,zC,

lim sup

t↓

f

tz+ ( –t)x,y

≤lim sup

t↓

ytz+ ( –t)x,∇g

tz+ ( –t)x+lim sup

t↓

h(y) –h

tz+ ( –t)x

=lim sup

t↓

ytz– ( –t)x– (y–x),g

tz+ ( –t)x+(y–x),g

tz+ ( –t)x

+lim sup

t↓

h(y) –h

tz+ ( –t)x

yx,gx+h(y) –h(x)

=f(x,y).

This shows that condition (A) is satisfied. It is easy to see thatfalso satisfies conditions

(A), (A), and (A). We seeKF(C,f) =H–VI(C,∇g,h) and= =∅. By

Theo-rem .,limn→∞xn=x, where¯ x¯=P(x¯–Vx),¯ x¯∈KF(C,f). This pointx¯∈is also the unique solution to the hierarchical variational inequality:

Vx,¯ qx¯ ≥, ∀q.

By=andx¯∈KF(C,f), we have

(13)

and

yx,¯ ∇g(x)¯

+h(y) –h(x)¯ ≥ for allyC. (.)

By (.) and (.), we have

g(y) +h(y) –g(x) –¯ h(x)¯ ≥

yx,g(x)

+h(y) –h(x)¯ ≥

for allyC. Thenx¯∈arg minyC(g+h)(y).

Example . Leth(x) =x,g(x) =x+ x+ ,C= [–, ],H=R,λ= ,V =I, αn= 

for alln∈N,βn=,n,C= [–, ],θn= ,f(x,y) =yx,g(x)+h(y) –h(x). Then

f(x,y) = (yx)(x+y+x+ ), this implies thatf(–,y) = (y+)≥ for ally∈[–, ], and

–

∈H–VI([–, ],∇g,h)=∅.

We also see Af(–) = (–∞, –],Af() = [,∞), andAf(x) = x+  ifx∈(–, ). Let yn=J

Af

PCxn,xn+=xn+( –,n)yn.

It is easy to see thatPCxn= yn+  andyn=(PCxn– ).

Hencexn+=xn+( – ,n)(PCxn– ). It is easy to see all the conditions of

Theo-rem . are satisfied.

Letx= , thenx= –.,x= –.,x= –.,x= –., . . . ,

we seelimn→∞xn=x¯= –∈arg minx∈[–,]gx+hx.

Next, an iteration is used to find the solution of the following optimization problem for the convex differentiable function:

arg min

yC

g(y).

Corollary . Let g:H→Rbe a convex Fréchet differentiable function with Fréchet

derivativeg.Let f(x,y) =yx,gxfor all x,yH and let Afbe defined as(L.) in Lemma..Suppose that,:=arg minyCg(y)=∅.Takeμ∈Rsuch that <μ< Lγ¯. A sequence{xn} ⊂His defined as follows:x∈C is chosen arbitrarily,and

yn=J Af

λ PCxn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,λ∈(,∞)and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P,(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx,¯ qx¯ ≥, ∀q,.

Proof By Lemma ., we know thatVI(C,∇g) =arg minyCg(y). Therefore, Corollary .

follows immediately from Theorem . by lettingh= .

Next, another iteration is used to find the solution of the following optimization problem for a convex function:

arg min

zC

(14)

Corollary . Let f(x,y) =h(y) –h(x),for all x,yC,and let Afbe defined as(L.) in Lemma..Suppose that,:=arg minyCh(y)=∅.Takeμ∈Rsuch that <μ< Lγ¯. A sequence{xn} ⊂His defined as follows:x∈C is chosen arbitrarily,and

yn=J Af λ PCxn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,λ∈(,∞),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P,(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx,¯ qx¯ ≥, ∀q,.

Proof Putg=  in Theorem .. Then Corollary . follows from Theorem ..

In the following theorem, an iteration is used to find the solution of the following opti-mization problem for the sum of two convex functions:

arg min

yH

g(y) +h(y).

Theorem . Let g: H→(–∞,∞) be a convex Fréchet differentiable function with

Fréchet derivativegon H, and h :H →(–∞,∞] be a proper convex lower

semi-continuous function.Suppose thatgis Lipschitz with Lipschitz constant L and:=

arg minxH(g+h)(x)=∅.Takeμ∈Rsuch that <μ< Lγ¯.A sequence{xn} ⊂His

de-fined as follows:x∈H is chosen arbitrarily,and

yn=proxrh(I–rg)xn=J

∂h

r (I–rg)xn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,r∈(,L

), and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx,¯ qx¯ ≥, ∀q.

We give two different proofs for this theorem.

Proof I PutC=H,G=∂h,F=∇g,T=I|C,F= ,G=∂iHin Theorem ., where

I|Cis the restriction ofIonCandiCis the indicate function ofC. Sincegis Lipschitz

continuous with Lipschitz constant L, it follows from Corollary  of [] that∇g is 

L-strongly-inverse-monotone. Sincehis a proper convex lower semicontinuous func-tion, it follows from Lemma . that∂h is a set valued maximum monotone mapping.

By Lemma .,∂g={∇g}. It follows fromdom(f)∩int(dom(g))=∅,dom(g) =H, and

Lemma . that

(15)

Hence,

=arg min

yH

(g+h)(y) =

xH:x∈(∂h+∇g)–

=Fix(I|C)∩Fix

J∂h

 (I–∇g)

∩Fix(PC)

=Fix(T)∩FixJ∂h

 (I–∇g)

∩FixJ∂iH

=.

Therefore, we get the conclusion of Theorem . from Theorem ..

Proof II Let C=H,T =proxrh(I–rg) in Theorem .. Since∇g is Lipschitz

con-stantL, it follows from Corollary  of [] that∇gis L-inverse- strongly-monotone.

By Lemma .,rgisrL-ism and (I–rg) is averaged. Since∂his maximum

mono-tone, it follows from Lemma .,proxrh =J

∂h

r is firmly nonexpansive. Henceproxrh is

-averaged. Then by Lemma .,T is averaged and nonexpansive. We have

Fix(T) =FixJ∂h

r (I–rg)PH

=Fixproxrh

(I–rg)

∩Fix(PH)

=arg min

xH

g(x) +h(x)∩H

=arg min

xC

g(x) +h(x).

Hence,=, and we get the conclusion of Theorem . from Theorem ..

Remark .

(a) The iterations in Theorems . and . are different.

(b) Theorem ., Theorem ., Theorem ., and the results given by Combettes and Wajs [], and Xu [] are weak convergence theorems for the problem:

arg min

xH

g(x) +h(x).

Further, Theorem ., Theorem ., and Theorem . gave strong convergence theorems of this problem under the uniform convex assumption onhorg. Therefore, Theorem .

is different from these results. Besides, Theorem . is also different from the result given by Wang and Xu [] and related algorithms in the literature.

Example . Leth(x) =x,g(x) =x+ x+ ,H=R,α= ,V=I,αn= for alln∈N,

βn=,n,C= [–, ],θn= . Then∂h(x) ={x},∇g(x) = x+ ,θn=  for alln∈N. We

see –∈arg minx∈Rg(x) +h(x)=∅, let

yn=J

∂h

 (I–∇hg)xn= (I+∂h)–(I–∇g)xn,

xn+=xn+( –,n)yn

for eachn∈N.

(16)

We obtain (I–∇g)xn= (I+∂h)yn=yn+ yn= yn=xn– xn– .

From this we obtainyn=–(xn+) and

xn+=

 xn+

   –  ,n yn = xn

   –  ,n

(xn+ )

= xn

 – 

,n

(xn+ )

 .

Letx= , thenx= .,x= –.,x= –.,x= –,x=

–., x = –., x = –., x= –., x= –.,

x= –.,x= –,x= –., . . . . From the relation

xn+=

 xn

 – 

,n

–(xn+ )

andx= , it is easy to see that the sequence{xn}n∈Nis nonincreasing for somenmand

bounded. Hencelimn→∞xnexists. Letlimn→∞xn=x. From the relation¯

xn+=

 xn

 – 

,n

(xn+ )

 ,

we see thatx¯= x¯–

¯

x+

 . Thereforex¯= – 

∈arg minxHg(x) +h(x).

In the following corollary, an iteration is used to find the solution of the following opti-mization problem:

Findx¯∈arg min

yH

h(y).

Corollary . Let h:H→(–∞,∞]be a proper convex lower semicontinuous function.

Takeμ∈Rsuch that <μ< Lγ¯.Suppose that ,:=arg minyHh(y)=∅. A sequence {xn} ⊂H is defined as follows:x∈H is chosen arbitrarily,and

yn=Jα∂hxn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,α∈(,∞)and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P,(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx,¯ qx¯ ≥, ∀q,.

Remark . In , Rockafellar [] proved the following in the Hilbert space setting: If his a proper convex lower semicontinuous function onH, the solution setarg minyHh(y)

is nonempty andlim infn→∞βn> . Let

xn+=arg min

yH

h(y) +

 βn

yxn

(17)

then{xn}converges weakly to a minimizer ofh. We see that Corollary . gives a

dif-ferent iteration which converges strongly to the solution of the following problem: Find

¯

x∈arg minyHh(y).

Next, a modified gradient-projection algorithm is used to find the solution of the fol-lowing mathematical program:

Findx¯∈arg min

yC

g(y).

Theorem . Let g: H→(–∞,∞) be a convex Fréchet differentiable function with

Fréchet derivativegon H.Takeμ∈Rsuch that <μ<Lγ¯.Suppose thatgis Lipschitz

continuous with Lipschitz constant Land:=arg minyCg(y)=∅.A sequence{xn} ⊂H

is defined as follows:x∈C is chosen arbitrarily,and

yn=PC(I–αg)xn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,α∈(,L

),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ Further,x is the unique solution to the hierarchical¯ variational inequality:

Vx,¯ qx¯ ≥, ∀q.

Proof Leth=iC, whereiCdenotes the indicator function ofC. From Lemma .,proxλh= PC, and

arg min

xC

h(x) +g(x)

=arg min

xC

g(x),

Theorem . follows immediately from Theorem ..

Remark . We know an iteration, defined by

xn+=PC(I–αng)xn, n∈N,

is called a gradient-projection algorithm, where∇gis Lipschitz continuous. In , Xu

[] used the gradient-projection algorithm and the relaxed gradient-projection algorithm and studied the problem

Findx¯∈arg min

yC

g(y),

(18)

In the end of this section, an iteration is used to find the solution of the following opti-mization problem:

Findx¯∈arg min

yHg(y).

Corollary . Let g :H→(–∞,∞) be a convex Fréchet differentiable function with

Fréchet derivativegon H. Take μ∈R such that  <μ< Lγ¯. Suppose thatgis

Lipschitz continuous with Lipschitz constant Land,:=arg minyHg(y)=∅.A sequence {xn} ⊂H is defined as follows:x∈H is chosen arbitrarily,and

yn= (I–αg)xn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,α∈(,L

),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then limn→∞xn=x,¯ wherex¯=P,(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx,¯ qx¯ ≥, ∀q,.

Proof Leth=iH. By Lemma ., we know thatproxλh=J

∂iH

λ =PH=I, and

arg min

xH

h(x) +g(x)

=arg min

xH

g(x).

Hence, Corollary . follows immediately from Theorem ..

5 Split feasibility problems and lasso problems

In the following theorem, a modified Byrne CQ iteration is used to find the solution of the following split feasibility problem: Findx¯∈Csuch thatAx¯∈Q.

Theorem . Let A:HHbe a bounded linear operator,Abe the adjoint of A.Take

μ∈Rsuch that <μ<Lγ¯.Suppose that:={xC:AxQ} =∅.A sequence{xn} ⊂H

is defined as follows:x∈C is chosen arbitrarily,and

yn=PC(I–αA∗(A–PQA))xn,

xn+=αnxn+ ( –αn)(βnθn+ (I–βnV)yn)

for each n∈N,α∈(,A),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx¯–θ,qx¯ ≥, ∀q.

Proof Letg(x) = AxPQAx

 

 . It is easy to see thatg(x) =minyQ

Axyis a convex

function. Then for anyvH, we have

g(x+v) – g(x) – 

Av, (IPQ)Ax

=(I–PQ)A(x+v), (IPQ)A(x+v)

–(I–PQ)Ax, (I–PQ)Ax

(19)

– Av, (IPQ)Ax

=Ax+AvPQAx+

PQAxPQA(x+v)

,Ax+AvPQAx

+PQAxPQA(x+v)

–(I–PQ)Ax, (I–PQ)Ax

– Av, (IPQ)Ax

=Av,Av+PQAxPQA(x+v),PQAxPQA(x+v)

+ (I–PQ)Ax,PQAxPQA(x+v)

+ Av,PQAxPQA(x+v)

. (.)

PQis a self-adjoint operator andPQ=PQ, therefore, we have

(I–PQ)Ax,PQAxPQA(x+v)

=(I–PQ)Ax,PQAx

–(I–PQ)Ax,PQA(x+v)

=PQ(I–PQ)Ax,Ax

PQ(I–PQ)Ax,A(x+v)

= . (.)

SincePQ:HQis a nonexpansive mapping, we have

PQAxPQA(x+v),PQAxPQA(x+v)

=PQAxPQA(x+v)

 

Ax– (Ax+Av)

=Av

A· v (.)

and

Av,PQAxPQA(x+v)

≤AvPQAxPQA(x+v)

≤AvAxA(x+v) ≤AvAv

≤A· v. (.)

We also see that

Av,AvA· v. (.)

By (.), (.), (.), (.), and (.), we have

lim

v→

|g(x+v) –g(x) –v,A∗(I–PQ)Ax|

v

=lim

v→

|g(x+v) – g(x) – v,A∗(I–PQ)Ax|

v ≤lim

v→A  v

(20)

This shows thatg is Fréchet differentiable with Fréchet derivative∇g=A∗(A–PQA).

SinceAis a bounded operator andPQis a firmly nonexpansive mapping,

g(x) –∇g(y)=A∗(A–PQA)xA∗(A–PQA)y

A(I–PQ)Ax– (I–PQ)Ay

AAxAy

A· xy. (.)

This shows that∇gis a Lipschitz function with Lipschitz constantA. By the

assump-tion=∅, we know that=. Then we get the conclusion of Theorem . from

Theorem ..

Remark . In , Xu [] used various algorithms to establish weak convergence

the-orems in infinite dimensional Hilbert spaces for the split feasibility problem (see Theo-rems ., ., ., . and . of []). Also, Xu [] established a strongly theorem for this problem in the infinite dimensional Hilbert space (see Theorem . of []). Now, Theo-rem . gives an algorithm to study the split feasibility problem which converges strongly to the split feasibility problem, and this result improves Byrne’s CQ algorithm [], and the results given by Xu [].

By Theorem ., we get the following result for split feasibility problem.

Corollary . Let A:HHbe a bounded linear operator,Abe the adjoint of A.Take

μ∈Rsuch that <μ< .Suppose that:={xC:AxQ} =∅.A sequence{xn} ⊂H is

defined as follows:x∈C is chosen arbitrarily,and

yn=PC(I–αA∗(A–PQA))xn,

xn+=αnxn+ ( –αn)(( –βn)yn)

for each n∈N,α∈(,A),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P.

Proof Letθn=θ=  for alln∈NandV =Iin Theorem ., then Corollary . follows

from Theorem ..

Next, an iteration is used to find the solution of the following lasso problem:

Findx¯∈arg min

x∈Rn

Axb

 +γx

.

Theorem . Let A be a m×n real matrix,x∈Rn,bRm,andγbe a regularization

parameter.Takeμ∈Rsuch that <μ<Lγ¯.Suppose that:=arg minx∈Rn

Axb  +

γx=∅.A sequence{xn} ⊂H is defined as follows:x∈Rnis chosen arbitrarily,and

yn=Jα∂γ·(xnαA∗(Axnb)),

(21)

for each n∈N,α∈(,A),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx¯–θ,qx¯ ≥, ∀q.

Proof Letg(x) =Axb, andh(x) =γx. For eachv∈Rn,

lim

v→

|g(x+v) –g(x) –v,A∗(Ax–b)| v

=lim

v→

|A(x+v) –b,A(x+v) –bAxb,Axb– v,A∗(Ax–b)| v

= . (.)

Then∇g(x) =A∗(Ax–b), andhis a proper convex and lower semicontinuous function

onH. Hence,

g(x) –g(y)=A(Axb) –A(Ayb)

≤ Axy

and∇g is a Lipschitz function with Lipschitz constantA. Therefore, Theorem .

follows from Theorem ..

The following is a special case of Theorem ..

Corollary . Let A be a m×n real matrix,x∈Rn,bRm,andγ ≥be a regularization parameter.Takeμ∈Rsuch that <μ< .Suppose that:=arg minx∈Rn

Axb+

γx=∅.A sequence{xn} ⊂Rnis defined as follows:x∈Rnis chosen arbitrarily,and

yn=J

∂γ·

α (xnαA∗(Axnb)),

xn+=αnxn+ ( –αn)(( –βn)yn)

for each n∈N,α∈(,A),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P().

Proof Letθn=θ =  for alln∈NandV=I in Theorem ., then Corollary . follows

from Theorem ..

Apply Corollary ., an iteration is used to find the solution to the split feasibility prob-lem: Findx¯∈C,Ax¯∈Q.

Theorem . Let A:HHbe a bounded linear operator,Abe the adjoint of A.Let

Afbe defined as(L.)in Lemma..Takeμ∈Rsuch that  <μ<

γ¯

L.Suppose that :={xC:AxQ} =∅.Let f(x,y) =yx,A∗(A–PQA)x. A sequence{xn} ⊂H is

defined as follows:x∈C is chosen arbitrarily,and

yn=J Af

λ PCxn,

(22)

for each n∈N,α∈(,A),and{αn,βn} ⊂(, ).Assume that conditions(D)hold.Then

limn→∞xn=x,¯ wherex¯=P(x¯–Vx).¯ Further,x is also the unique solution to the hierar-¯ chical variational inequality:

Vx¯–θ,qx¯ ≥, ∀q.

Proof Let g(x) =

AxPQAx

 , then shows that g is Fréchet differentiable with Fréchet

derivative∇g=A∗(A–PQA), andga Lipschitz function with Lipschitz constantA.

Applying Corollary . and following the same argument as Theorem ., we can prove

Theorem ..

6 Image deblurring problem

This section mainly focuses on the image deblurring problems, which has received a lot of attention in recent years. Until now, some researchers have proposed many novel al-gorithms for this problem based on different deblurring models; for examples, see []. Now, by Corollary ., we can consider the image deblurring problem.

All pixels of the original images described in the examples were first scaled into the range between  and .

The image went through a Gaussian blur of size × and standard deviation  (applied by the MATLAB functions imfilter and fspecial) followed by an additive zero-mean white Gaussian noise with standard deviation –. The original and observed images are given

in Figures -.

Remark . In the literature, we may observe that there are many fast algorithms for

the image deblurring problem. Here, we show that we can also consider this problem by Corollary ..

Figure 1 The original image.

Figure 2 The blurred image.

(23)

7 Conclusion and remarks

In this paper, we apply a recent fixed point theorem in [] to study mathematical pro-gramming for the sum of two convex functions, mathematical propro-gramming of convex function, the split feasibility problem, and the lasso problem. We establish strong conver-gence theorems as regards these problems. The study of such problems will give many other applications in science, nonlinear analysis, and statistics.

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors contributed equally to this work. All authors read and approved the final manuscript.

Author details

1Department of Applied Mathematics, National Sun Yat Sen University, Kaohsiung, 700, Taiwan.2Department of Electronic Engineering, Nan Kai University of Technology, Nantou, 542, Taiwan.3Department of Mathematics, National Changhua University of Education, Changhua, 50058, Taiwan.

Acknowledgements

Prof. CS Chuang was supported by the National Science Council of Republic of China.

Received: 14 May 2015 Accepted: 23 July 2015

References

1. Tibshirani, R: Regression shrinkage and selection for the lasso. J. R. Stat. Soc., Ser. B58, 267-288 (1996) 2. Combettes, PL, Wajs, R: Signal recovery by proximal forward-backward splitting. Multiscale Model. Simul.4,

1168-1200 (2005)

3. Xu, HK: Properties and iterative methods for the lasso and its variants. Chin. Ann. Math., Ser. B35, 501-518 (2014) 4. Wang, Y, Xu, HK: Strong convergence for the proximal gradient methods. J. Nonlinear Convex Anal.15, 581-593

(2014)

5. Douglas, J, Rachford, HH: On the numerical solution of heat conduction in two and three space variable. Trans. Am. Math. Soc.82, 421-439 (1956)

6. Bauschken, HH, Combettes, PL: Convex Analysis and Monotone Operator Theory in Hilbert Space. Springer, Berlin (2011)

7. Tseng, P: Further applications of a splitting algorithm to decomposition in variational inequalities and convex programming. Math. Program., Ser. B48, 249-263 (1990)

8. Censor, Y, Elfving, T: A multiprojection algorithm using Bregman projection in a product space. Numer. Algorithms8, 221-239 (1994)

9. Byrne, C: Iterative oblique projection onto convex sets and the split feasibility problem. Inverse Probl.18, 441-453 (2002)

10. Byrne, C: A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl.20, 103-120 (2004)

11. Censor, Y, Bortfeld, T, Martin, B, Trofimov, A: A unified approach for inversion problems in intensity-modulated radiation therapy. Phys. Med. Biol.51, 2353-2365 (2003)

12. López, G, Martín-Márquez, V, Xu, HK: Iterative algorithms for the multiple-sets split feasibility problem. In: Censor, Y, Jiang, M, Wang, G (eds.) Biomedical Mathematics: Promising Directions in Imaging, Therapy Planning and Inverse Problems, pp. 243-279. Medical Physics Publishing, Madison (2010)

13. Stark, H: Image Recovery: Theory and Applications. Academic Press, San Diego (1987)

14. Xu, HK: Iterative methods for the split feasibility problem in infinite-dimensional Hilbert spaces. Inverse Probl.26, 105018 (2010)

15. Rockafellar, TA: Monotone operators and the proximal point algorithm. SIAM J. Control Optim.14, 877-898 (1976) 16. Xu, HK: Averaged mappings and the gradient projection algorithm. J. Optim. Theory Appl.150, 360-378 (2011) 17. Yu, ZT, Lin, LJ: Hierarchical problems with applications to mathematical programming with multiple sets split

feasibility constraints. Fixed Point Theory Appl.2013, 283 (2013)

18. Takahashi, W: Nonlinear Functional Analysis: Fixed Point Theory and Its Applications. Yokohama Publishers, Yokohama (2000)

19. Fan, K: A minimax inequalities and its applications. In: Shisha, O (ed.) Inequalities III, pp. 103-113. Academic Press, San Diego (1972)

20. Blum, E, Oettli, W: From optimization and variational inequalities to equilibrium problems. Math. Stud.63, 123-146 (1994)

21. Combettes, PL, Hirstoaga, SA: Equilibrium programming in Hilbert spaces. J. Nonlinear Convex Anal.6, 117-136 (2005) 22. Takahashi, S, Takahashi, W, Toyoda, M: Strong convergence theorems for maximal monotone operators with

nonlinear mappings in Hilbert spaces. J. Optim. Theory Appl.147, 27-41 (2010)

23. Ekaland, I, Temam, R: Convex Analysis and Variational Problems. North-Holland, Amsterdam (1976)

24. Combettes, PL: Solving monotone inclusions via compositions of nonexpansive averaged operators. Optimization 53(5-6), 475-504 (2004)

25. Baillon, JB, Haddad, G: Quelques propriétés des opérateurs angle-bornés etn-cycliquement monotones. Isr. J. Math. 26, 137-150 (1977)

Figure

Figure 1 The original image.

References

Related documents

Interestingly, we also identified a significant reduction in the proportion of cells in G1/G0 (Fig. 6D), which may contribute towards a significant, yet moderate, increase in total

In spite of early diagnosis and better management of gallbladder disease, it is feared that this may not be a revisit by this, once not so uncommon entity, and a clinician should

The appropriate selection of suppliers, the optimal amounts of raw materials shipped from the selected suppliers to the plants in each period, the optimal amounts of products

2018-NIA-AA-RF: 2018 National Institute of Age-Alzheimer ’ s Association re- search framework; AD: Alzheimer ’ s disease; aMCI: Amnestic mild cognitive impairment;

Malmberg K, Ryden L, Efendic S, Herlitz J, Nicol P, Waldenstrom A, Wedel H, Welin L: Randomized trial of insulin-glucose infusion followed by subcutaneous insulin treatment in

designation of personal identity relies upon the continuity of mental features, memories included therein, with a non-branching qualifier (see Parfit, 1984, Nozick 1981, and