41
Received March 31, 2015, accepted April 18, 2015, date of publication May 6, 2015, date of current version May 20, 2015. Digital Object Identifier 10.1109/ACCESS.2015.2430359 A Survey of Sparse Representation: Algorithms and Applications ZHENG ZHANG 1,2 , (Student Member, IEEE), YONG XU 1,2 , (Senior Member, IEEE), JIAN YANG 3 , (Member, IEEE), XUELONG LI 4 , (Fellow, IEEE), AND DAVID ZHANG 5 , (Fellow, IEEE) 1 Bio-Computing Research Center, Shenzhen Graduate School, Harbin Institute of Technology, Shenzhen 518055, China 2 Key Laboratory of Network Oriented Intelligent Computation, Shenzhen 518055, China 3 College of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing 210094, China 4 State Key Laboratory of Transient Optics and Photonics, Center for Optical Imagery Analysis and Learning, Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, Shaanxi, China 5 Biometrics Research Center, The Hong Kong Polytechnic University, Hong Kong Corresponding author: Y. Xu ([email protected]) This work was supported in part by the National Natural Science Foundation of China under Grant 61370163, Grant 61233011, and Grant 61332011, in part by the Shenzhen Municipal Science and Technology Innovation Council under Grant JCYJ20130329151843309, Grant JCYJ20130329151843309, Grant JCYJ20140417172417174, and Grant CXZZ20140904154910774, in part by the China Post-Doctoral Science Foundation Funded Project under Grant 2014M560264, and in part by the Shaanxi Key Innovation Team of Science and Technology under Grant 2012KCT-04. ABSTRACT Sparse representation has attracted much attention from researchers in fields of signal processing, image processing, computer vision, and pattern recognition. Sparse representation also has a good reputation in both theoretical research and practical applications. Many different algorithms have been proposed for sparse representation. The main purpose of this paper is to provide a comprehensive study and an updated review on sparse representation and to supply guidance for researchers. The taxonomy of sparse representation methods can be studied from various viewpoints. For example, in terms of different norm minimizations used in sparsity constraints, the methods can be roughly categorized into five groups: 1) sparse representation with l 0 -norm minimization; 2) sparse representation with l p -norm (0 < p < 1) minimization; 3) sparse representation with l 1 -norm minimization; 4) sparse representation with l 2,1 -norm minimization; and 5) sparse representation with l 2 -norm minimization. In this paper, a comprehensive overview of sparse representation is provided. The available sparse representation algorithms can also be empirically categorized into four groups: 1) greedy strategy approximation; 2) constrained optimization; 3) proximity algorithm-based optimization; and 4) homotopy algorithm-based sparse representation. The rationales of different algorithms in each category are analyzed and a wide range of sparse representation applications are summarized, which could sufficiently reveal the potential nature of the sparse representation theory. In particular, an experimentally comparative study of these sparse representation algorithms was presented. INDEX TERMS Sparse representation, compressive sensing, greedy algorithm, constrained optimization, proximal algorithm, homotopy algorithm, dictionary learning. I. INTRODUCTION With advancements in mathematics, linear representation methods (LRBM) have been well studied and have recently received considerable attention [1], [2]. The sparse represen- tation method is the most representative methodology of the LRBM and has also been proven to be an extraordinary powerful solution to a wide range of application fields, especially in signal processing, image processing, machine learning, and computer vision, such as image denoising, debluring, inpainting, image restoration, super-resolution, visual tracking, image classification and image segmen- tation [3]–[10]. Sparse representation has shown huge potential capabilities in handling these problems. Sparse representation, from the viewpoint of its origin, is directly related to compressed sensing (CS) [11]–[13], which is one of the most popular topics in recent years. Donoho [11] first proposed the original concept of compressed sensing. CS theory suggests that if a signal is sparse or compressive, 490 2169-3536 2015 IEEE. Translations and content mining are permitted for academic research only. Personal use is also permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information. VOLUME 3, 2015 www.redpel.com1+917620593389 www.redpel.com1+917620593389

A survey of sparse representation algorithms and applications

Embed Size (px)

Citation preview

Page 1: A survey of sparse representation algorithms and applications

Received March 31, 2015, accepted April 18, 2015, date of publication May 6, 2015, date of current version May 20, 2015.

Digital Object Identifier 10.1109/ACCESS.2015.2430359

A Survey of Sparse Representation:Algorithms and ApplicationsZHENG ZHANG1,2, (Student Member, IEEE), YONG XU1,2, (Senior Member, IEEE),JIAN YANG3, (Member, IEEE), XUELONG LI4, (Fellow, IEEE), ANDDAVID ZHANG5, (Fellow, IEEE)1Bio-Computing Research Center, Shenzhen Graduate School, Harbin Institute of Technology, Shenzhen 518055, China2Key Laboratory of Network Oriented Intelligent Computation, Shenzhen 518055, China3College of Computer Science and Technology, Nanjing University of Science and Technology, Nanjing 210094, China4State Key Laboratory of Transient Optics and Photonics, Center for Optical Imagery Analysis and Learning, Xi’an Institute of Optics andPrecision Mechanics, Chinese Academy of Sciences, Xi’an 710119, Shaanxi, China5Biometrics Research Center, The Hong Kong Polytechnic University, Hong Kong

Corresponding author: Y. Xu ([email protected])

This work was supported in part by the National Natural Science Foundation of China under Grant 61370163, Grant 61233011, andGrant 61332011, in part by the Shenzhen Municipal Science and Technology Innovation Council under Grant JCYJ20130329151843309,Grant JCYJ20130329151843309, Grant JCYJ20140417172417174, and Grant CXZZ20140904154910774, in part by the ChinaPost-Doctoral Science Foundation Funded Project under Grant 2014M560264, and in part by the Shaanxi Key InnovationTeam of Science and Technology under Grant 2012KCT-04.

ABSTRACT Sparse representation has attracted much attention from researchers in fields of signalprocessing, image processing, computer vision, and pattern recognition. Sparse representation also has agood reputation in both theoretical research and practical applications. Many different algorithms have beenproposed for sparse representation. The main purpose of this paper is to provide a comprehensive studyand an updated review on sparse representation and to supply guidance for researchers. The taxonomy ofsparse representation methods can be studied from various viewpoints. For example, in terms of differentnorm minimizations used in sparsity constraints, the methods can be roughly categorized into five groups:1) sparse representation with l0-norm minimization; 2) sparse representation with lp-norm (0 < p < 1)minimization; 3) sparse representation with l1-norm minimization; 4) sparse representation with l2,1-normminimization; and 5) sparse representation with l2-norm minimization. In this paper, a comprehensiveoverview of sparse representation is provided. The available sparse representation algorithms can also beempirically categorized into four groups: 1) greedy strategy approximation; 2) constrained optimization;3) proximity algorithm-based optimization; and 4) homotopy algorithm-based sparse representation. Therationales of different algorithms in each category are analyzed and a wide range of sparse representationapplications are summarized, which could sufficiently reveal the potential nature of the sparse representationtheory. In particular, an experimentally comparative study of these sparse representation algorithms waspresented.

INDEX TERMS Sparse representation, compressive sensing, greedy algorithm, constrained optimization,proximal algorithm, homotopy algorithm, dictionary learning.

I. INTRODUCTIONWith advancements in mathematics, linear representationmethods (LRBM) have been well studied and have recentlyreceived considerable attention [1], [2]. The sparse represen-tation method is the most representative methodology of theLRBM and has also been proven to be an extraordinarypowerful solution to a wide range of application fields,especially in signal processing, image processing, machinelearning, and computer vision, such as image denoising,

debluring, inpainting, image restoration, super-resolution,visual tracking, image classification and image segmen-tation [3]–[10]. Sparse representation has shown hugepotential capabilities in handling these problems.

Sparse representation, from the viewpoint of its origin, isdirectly related to compressed sensing (CS) [11]–[13], whichis one of the most popular topics in recent years. Donoho [11]first proposed the original concept of compressed sensing.CS theory suggests that if a signal is sparse or compressive,

4902169-3536 2015 IEEE. Translations and content mining are permitted for academic research only.

Personal use is also permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

VOLUME 3, 2015

www.redpel.com1+917620593389

www.redpel.com1+917620593389

Page 2: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

the original signal can be reconstructed by exploiting a fewmeasured values, which are much less than the onessuggested by previously used theories such as Shannon’ssampling theorem (SST). Candès et al. [13], from the mathe-matical perspective, demonstrated the rationale of CS theory,i.e. the original signal could be precisely reconstructed byutilizing a small portion of Fourier transformationcoefficients. Baraniuk [12] provided a concrete analysis ofcompressed sensing and presented a specific interpretation onsome solutions of different signal reconstruction algorithms.All these literature [11]–[17] laid the foundation ofCS theory and provided the theoretical basis for futureresearch. Thus, a large number of algorithms based onCS theory have been proposed to address different problemsin various fields. Moreover, CS theory always includes thethree basic components: sparse representation, encodingmeasuring, and reconstructing algorithm. As an indispens-able prerequisite of CS theory, the sparse representationtheory [4], [7]–[10], [17] is the most outstanding techniqueused to conquer difficulties that appear in many fields.For example, the methodology of sparse representation isa novel signal sampling method for the sparse or com-pressible signal and has been successfully applied to signalprocessing [4]–[6].

Sparse representation has attracted much attention inrecent years and many examples in different fields can befound where sparse representation is definitely beneficial andfavorable [18], [19]. One example is image classification,where the basic goal is to classify the given test image intoseveral predefined categories. It has been demonstrated thatnatural images can be sparsely represented from the perspec-tive of the properties of visual neurons. The sparse represen-tation based classification (SRC) method [20] first assumesthat the test sample can be sufficiently represented by samplesfrom the same subject. Specifically, SRC exploits the linearcombination of training samples to represent the test sampleand computes sparse representation coefficients of the linearrepresentation system, and then calculates the reconstructionresiduals of each class employing the sparse representationcoefficients and training samples. The test sample will beclassified as a member of the class, which leads to theminimum reconstruction residual. The literature [20] has alsodemonstrated that the SRC method has great superioritieswhen addressing the image classification issue on corruptedor disguised images. In such cases, each natural image can besparsely represented and the sparse representation theory canbe utilized to fulfill the image classification task.

For signal processing, one important task is to extract keycomponents from a large number of clutter signals or groupsof complex signals in coordination with differentrequirements. Before the appearance of sparse representation,SST and Nyquist sampling law (NSL) were the traditionalmethods for signal acquisition and the general proceduresincluded sampling, coding compression, transmission, anddecoding. Under the frameworks of SST and NSL, thegreatest difficulty of signal processing lies in efficient

sampling from mass data with sufficient memory-saving.In such a case, sparse representation theory can simultane-ously break the bottleneck of conventional sampling rules,i.e. SST and NSL, so that it has a very wide applicationprospect. Sparse representation theory proposes to integratethe processes of signal sampling and coding compression.Especially, sparse representation theory employs a moreefficient sampling rate to measure the original sample byabandoning the pristine measurements of SST and NSL, andthen adopts an optimal reconstruction algorithm to recon-struct samples. In the context of compressed sensing, it isfirst assumed that all the signals are sparse or approximatelysparse enough [4], [6], [7]. Compared to the primary signalspace, the size of the set of possible signals can be largelydecreased under the constraint of sparsity. Thus, massivealgorithms based on the sparse representation theory havebeen proposed to effectively tackle signal processing issuessuch as signal reconstruction and recovery. To this end, thesparse representation technique can save a significant amountof sampling time and sample storage space and it is favorableand advantageous.

A. CATEGORIZATION OF SPARSEREPRESENTATION TECHNIQUESSparse representation theory can be categorized fromdifferent viewpoints. Because different methods have theirindividual motivations, ideas, and concerns, there arevarieties of strategies to separate the existing sparse represen-tation methods into different categories from the perspectiveof taxonomy. For example, from the viewpoint of ‘‘atoms’’,available sparse representation methods can be categorizedinto two general groups: naive sample based sparserepresentation and dictionary learning based sparse repre-sentation. However, on the basis of the availability of labelsof ‘‘atoms’’, sparse representation and learning methods canbe coarsely divided into three groups: supervised learning,semi-supervised learning, and unsupervised learningmethods. Because of the sparse constraint, sparse representa-tion methods can be divided into two communities: structureconstraint based sparse representation and sparse constraintbased sparse representation. Moreover, in the field of imageclassification, the representation based classificationmethods consist of two main categories in terms of the wayof exploiting the ‘‘atoms’’: the holistic representation basedmethod and local representation based method [21]. Morespecifically, holistic representation based methods exploittraining samples of all classes to represent the test sample,whereas local representation based methods only employtraining samples (or atoms) of each class or several classes torepresent the test sample. Most of the sparse representationmethods are holistic representation based methods. A typicaland representative local sparse representation methods is thetwo-phase test sample sparse representation (TPTSR)method [9]. In consideration of different methodologies,the sparse representation method can be grouped intotwo aspects: pure sparse representation and hybrid

VOLUME 3, 2015 491

www.redpel.com2+917620593389

www.redpel.com2+917620593389

Page 3: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

sparse representation, which improves the pre-existing sparserepresentation methods with the aid of other methods. Theliterature [23] suggests that sparse representation algorithmsroughly fall into three classes: convex relaxation,greedy algorithms, and combinational methods. In theliterature [24], [25], from the perspective of sparseproblem modeling and problem solving, sparse decompo-sition algorithms are generally divided into two sections:greedy algorithms and convex relaxation algorithms. On theother hand, if the viewpoint of optimization is taken intoconsideration, the problems of sparse representation can bedivided into four optimization problems: the smooth convexproblem, nonsmooth nonconvex problem, smooth noncon-vex problem, and nonsmooth convex problem. Furthermore,Schmidt et al. [26] reviewed some optimization techniquesfor solving l1-norm regularization problems and roughlydivided these approaches into three optimization strategies:sub-gradient methods, unconstrained approximationmethods, and constrained optimization methods. The supple-mentary file attached with the paper also offers more usefulinformation to make fully understandings of the ‘taxonomy’of current sparse representation techniques in this paper.

In this paper, the available sparse representation methodsare categorized into four groups, i.e. the greedy strategyapproximation, constrained optimization strategy, proximityalgorithm based optimization strategy, and homotopyalgorithm based sparse representation, with respect to theanalytical solution and optimization viewpoints.

(1) In the greedy strategy approximation for solving sparserepresentation problem, the target task is mainly to solvethe sparse representation method with l0-norm minimiza-tion. Because of the fact that this problem is an NP-hardproblem [27], the greedy strategy provides an approximatesolution to alleviate this difficulty. The greedy strategysearches for the best local optimal solution in each iterationwith the goal of achieving the optimal holistic solution [28].For the sparse representation method, the greedy strategyapproximation only chooses the most k appropriate samples,which are called k-sparsity, to approximate the measurementvector.

(2) In the constrained optimization strategy, the core ideais to explore a suitable way to transform a non-differentiableoptimization problem into a differentiable optimizationproblem by replacing the l1-norm minimization term, whichis convex but nonsmooth, with a differentiable optimiza-tion term, which is convex and smooth. More specifically,the constrained optimization strategy substitutes the l1-normminimization term with an equal constraint condition on theoriginal unconstraint problem. If the original unconstraintproblem is reformulated into a differentiable problem withconstraint conditions, it will become an uncomplicatedproblem in the consideration of the fact that l1-normminimization is global non-differentiable.

(3) Proximal algorithms can be treated as a powerfultool for solving nonsmooth, constrained, large-scale, ordistributed versions of the optimization problem [29]. In the

proximity algorithm based optimization strategy for sparserepresentation, the main task is to reformulate the originalproblem into the specific model of the corresponding prox-imal operator such as the soft thresholding operator, hardthresholding operator, and resolvent operator, and thenexploits the proximity algorithms to address the originalsparse optimization problem.

(4) The general framework of the homotopy algorithm isto iteratively trace the final desired solution starting fromthe initial point to the optimal point by successively adjust-ing the homotopy parameter [30]. In homotopy algorithmbased sparse representation, the homotopy algorithm is usedto solve the l1-norm minimization problem with k-sparseproperty.

B. MOTIVATION AND OBJECTIVESIn this paper, a survey on sparse representation and overviewavailable sparse representation algorithms from viewpointsof the mathematical and theoretical optimization is provided.This paper is designed to provide foundations of the studyon sparse representation and aims to give a good start tonewcomers in computer vision and pattern recognitioncommunities, who are interested in sparse representationmethodology and its related fields. Extensive state-of-artsparse representation methods are summarized and the ideas,algorithms, andwide applications of sparse representation arecomprehensively presented. Specifically, there is concentra-tion on introducing an up-to-date review of the existing litera-ture and presenting some insights into the studies of the latestsparse representation methods. Moreover, the existing sparserepresentation methods are divided into different categories.Subsequently, corresponding typical algorithms in differentcategories are presented and their distinctness is explicitlyshown. Finally, the wide applications of these sparse repre-sentation methods in different fields are introduced.

The remainder of this paper is mainly composed offour parts: basic concepts and frameworks are shownin Section II and Section III, representative algorithms arepresented in Section IV-VII and extensive applications areillustrated in Section VIII, massive experimental evaluationsare summarized in Section IX. More specifically, the funda-mentals and preliminary mathematic concepts are presentedin Section II, and then the general frameworks of the existingsparse representation with different norm regularizations aresummarized in Section III. In Section IV, the greedy strategyapproximation method is presented for obtaining a sparserepresentation solution, and in Section V, the constrainedoptimization strategy is introduced for solving the sparserepresentation issue. Furthermore, the proximity algorithmbased optimization strategy and Homotopy strategy foraddressing the sparse representation problem are outlinedin Section VI and Section VII, respectively. Section VIIIpresents extensive applications of sparse representation inwidespread and prevalent fields including dictionary learn-ing methods and real-world applications. Finally, Section IXoffers massive experimental evaluations and conclusions are

492 VOLUME 3, 2015

www.redpel.com3+917620593389

www.redpel.com3+917620593389

Page 4: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

FIGURE 1. The structure of this paper. The main body of this paper mainly consists of four parts: basic concepts and frameworks in Section II-III,representative algorithms in Section IV-VII and extensive applications in Section VIII, massive experimental evaluations in Section IX. Conclusion issummarized in Section X.

drawn and summarized in Section X. The structure of the thispaper has been summarized in Fig. 1.

II. FUNDAMENTALS AND PRELIMINARY CONCEPTSA. NOTATIONSIn this paper, vectors are denoted by lowercase letters withbold face, e.g. x. Matrices are denoted by uppercase letter,e.g. X and their elements are denoted with indexes such as Xi.In this paper, all the data are only real-valued.

Suppose that the sample is from space Rd and thus allthe samples are concatenated to form a matrix, denotedasD ∈ Rd×n. If any sample can be approximately representedby a linear combination of dictionaryD and the number of thesamples is larger than the dimension of samples in D,i.e. n > d , dictionary D is referred to as an over-completedictionary. A signal is said to be compressible if it is asparse signal in the original or transformed domain whenthere is no information or energy loss during the process oftransformation.

‘‘sparse’’ or ‘‘sparsity’’ of a vector means that someelements of the vector are zero. We use a linear combinationof a basis matrix A ∈ RN×N to represent a signal x ∈ RN×1,i.e. x = Aswhere s ∈ RN×1 is the column vector of weightingcoefficients. If only k (k � N ) elements of s are nonzero andthe rest elements in s are zero, we call the signal x is k-sparse.

B. BASIC BACKGROUNDThe standard inner product of two vectors, x and y from theset of real n dimensions, is defined as

〈x, y〉 = xT y = x1y1 + x2y2 + · · · + xnyn (II.1)

The standard inner product of two matrixes,X ∈ Rm×n and Y ∈ Rm×n from the set of realm×nmatrixes,is denoted as the following equation

〈X ,Y 〉 = tr(XTY ) =m∑i=1

n∑j=1

XijYij (II.2)

where the operator tr(A) denotes the trace of the matrix A,i.e. the sum of its diagonal entries.Suppose that v = [v1, v2, · · · , vn] is an n dimensional

vector in Euclidean space, thus

‖v‖p = (n∑i=1

|vi|p)1/p (II.3)

is denoted as the p-norm or the lp-norm (1 ≤ p ≤ ∞) ofvector v.

When p=1, it is called the l1-norm. It means the sum ofabsolute values of the elements in vector v, and its geometricinterpretation is shown in Fig. 2b, which is a square with aforty-five degree rotation.

When p=2, it is called the l2-norm or Euclidean norm. It isdefined as ‖v‖2 = (v21 + v

22 + · · · + v

2n)

1/2, and its geometricinterpretation in 2-D space is shown in Fig. 2c which is acircle.

In the literature, the sparsity of a vector v is always relatedto the so-called l0-norm, which means the number of thenonzero elements of vector v. Actually, the l0-norm is thelimit as p → 0 of the lp-norms [8] and the definition of

VOLUME 3, 2015 493

www.redpel.com4+917620593389

www.redpel.com4+917620593389

Page 5: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

FIGURE 2. Geometric interpretations of different norms in 2-D space [7]. (a), (b), (c), (d) are the unit ball of thel0-norm, l1-norm, l2-norm, lp-norm (0<p<1) in 2-D space, respectively. The two axes of the above coordinatesystems are x1 and x2.

the l0-norm is formulated as

‖v‖0 = limp→0‖v‖pp = lim

p→0

n∑i=1

|vi|p (II.4)

We can see that the notion of the l0-norm is very convenientand intuitive for defining the sparse representation problem.The property of the l0-norm can also be presented from theperspective of geometric interpretation in 2-D space, whichis shown in Fig. 2a, and it is a crisscross.

Furthermore, the geometric meaning of the lp-norm(0<p<1) is also presented, which is a form of similarrecessed pentacle shown in Fig. 2d.

On the other hand, it is assumed that f (x) is the functionof the lp-norm (p>0) on the parameter vector x, and then thefollowing function is obtained:

f (x) = ‖x‖pp = (n∑i=1

|xi|p) (II.5)

The relationships between different norms are summarizedin Fig. 3. From the illustration in Fig. 3, the conclusions areas follows. The l0-norm function is a nonconvex, nonsmooth,discontinuity, global nondifferentiable function. Thelp-norm (0<p<1) is a nonconvex, nonsmooth, global

FIGURE 3. Geometric interpretations of different norms in 1-D space [7].

nondifferentiable function. The l1-norm function is a convex,nonsmooth, global nondifferentiable function. The l2-normfunction is a convex, smooth, global differentiable function.

In order to more specifically elucidate the meaning andsolutions of different norm minimizations, the geometry in2-D space is used to explicitly illustrate the solutions of thel0-norm minimization in Fig. 4a, l1-norm minimizationin Fig. 4b, and l2-normminimization in Fig. 4c. Let S = {x∗ :Ax = y} denote the line in 2-D space and a hyperplane willbe formulated in higher dimensions. All possible solution x∗

must lie on the line of S. In order to visualize how to obtainthe solution of different norm-based minimization problems,we take the l1-norm minimization problem as an example toexplicitly interpret. Suppose that we inflate the l1-ball froman original status until it hits the hyperplane S at some point.Thus, the solution of the l1-norm minimization problem isthe aforementioned touched point. If the sparse solution ofthe linear system is localized on the coordinate axis, it willbe sparse enough. From the perspective of Fig. 4, it can beseen that the solutions of both the l0-norm and l1-normminimization are sparse, whereas for the l2-normminimization, it is very difficult to rigidly satisfy thecondition of sparsity. However, it has been demonstrated thatthe representation solution of the l2-normminimization is notstrictly sparse enough but ‘‘limitedly-sparse’’, which meansit possesses the capability of discriminability [31].

The Frobenius norm, L1-norm of matrix X ∈ Rm×n, andl2-norm or spectral norm are respectively defined as

‖X‖F = (n∑i=1

m∑j=1

X2j,i)

1/2, ‖X‖L1 =n∑i=1

m∑j=1

|xj,i|,

‖X‖2 = δmax(X ) = (λmax(XTX ))1/2 (II.6)

where δ is the singular value operator and the l2-norm of X isits maximum singular value [32].

The l2,1-norm or R1-norm is defined on matrix term, that is

‖X‖2,1 =n∑i=1

(m∑j=1

X2j,i)

1/2 (II.7)

As shown above, a norm can be viewed as a measure ofthe length of a vector v. The distance between two vectorsx and y, or matrices X and Y , can be measured by the length

494 VOLUME 3, 2015

www.redpel.com5+917620593389

www.redpel.com5+917620593389

Page 6: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

of their differences, i.e.

dist(x, y) = ‖x− y‖22, dist(X ,Y ) = ‖X − Y‖F (II.8)

which are denoted as the distance between x and y in thecontext of the l2-norm and the distance between X and Y inthe context of the Frobenius norm, respectively.

Assume that X ∈ Rm×n and the rank of X ,i.e. rank(X ) = r . The SVD of X is computed as

X = U3V T (II.9)

where U ∈ Rm×r with UTU = I and V ∈ Rn×r withV TV = I . The columns of U and V are called left andright singular vectors of X , respectively. Additionally, 3 is adiagonal matrix and its elements are composed of the singularvalues of X , i.e. 3 = diag(λ1, λ2, · · · , λr ) withλ1 ≥ λ2 ≥ · · · ≥ λr > 0. Furthermore, the singular valuedecomposition can be rewritten as

X =r∑i=1

λiuivi (II.10)

where λi, ui and vi are the i-th singular value, the i-th columnof U , and the i-th column of V , respectively [32].

III. SPARSE REPRESENTATION PROBLEM WITHDIFFERENT NORM REGULARIZATIONSIn this section, sparse representation is summarized andgrouped into different categories in terms of the normregularizations used. The general framework of sparse repre-sentation is to exploit the linear combination of some samplesor ‘‘atoms’’ to represent the probe sample, to calculate therepresentation solution, i.e. the representation coefficients ofthese samples or ‘‘atoms’’, and then to utilize therepresentation solution to reconstruct the desired results.The representation results in sparse representation, however,can be greatly dominated by the regularizer (or optimizer)imposed on the representation solution [33]–[36]. Thus,in terms of the different norms used in optimizers, thesparse representation methods can be roughly grouped intofive general categories: sparse representation with thel0-norm minimization [37], [38], sparse representationwith the lp-norm (0<p<1) minimization [39]–[41], sparserepresentation with the l1-norm minimization [42]–[45],sparse representation with the l2,1-norm minimiz-ation [46]–[50], sparse representation with the l2-normminimization [9], [22], [51].

A. SPARSE REPRESENTATION WITHl0-NORM MINIMIZATIONLet x1, x2, · · · , xn ∈ Rd be all the n known samples andmatrix X ∈ Rd×n (d<n), which is constructed by knownsamples, is the measurement matrix or the basis dictionaryand should also be an over-completed dictionary. Eachcolumn of X is one sample and the probe sample is y ∈ Rd ,which is a column vector. Thus, if all the known samples are

used to approximately represent the probe sample, it shouldbe expressed as:

y = x1α1 + x2α2 + · · · + xnαn (III.1)

where αi (i = 1, 2, · · · , n) is the coefficient of xi and Eq. III.1can be rewritten into the following equation for convenientdescription:

y = Xα (III.2)

where matrix X=[x1, x2, · · · , xn] and α=[α1,α2, · · · ,αn]T.However, problem III.2 is an underdetermined linear

system of equations and the main problem is how to solve it.From the viewpoint of linear algebra, if there is not any priorknowledge or any constraint imposed on the representationsolution α, problem III.2 is an ill-posed problem and willnever have a unique solution. That is, it is impossible to utilizeequation III.2 to uniquely represent the probe sample y usingthe measurement matrix X . To alleviate this difficulty, it isfeasible to impose an appropriate regularizer constraint orregularizer function on representation solution α. The sparserepresentation method demands that the obtained represen-tation solution should be sparse. Hereafter, the meaning of‘sparse’ or ‘sparsity’ refers to the condition that when thelinear combination of measurement matrix is exploited torepresent the probe sample, many of the coefficients shouldbe zero or very close to zero and few of the entries in therepresentation solution are differentially large.

The sparsest representation solution can be acquiredby solving the linear representation system III.2 with thel0-normminimization constraint [52]. Thus problem III.2 canbe converted to the following optimization problem:

α = argmin ‖α‖0 s.t. y = Xα (III.3)

where ‖ · ‖0 refers to the number of nonzero elements in thevector and is also viewed as the measure of sparsity. More-over, if just k (k < n) atoms from the measurement matrix Xare utilized to represent the probe sample, problem III.3 willbe equivalent to the following optimization problem:

y = Xα s.t. ‖α‖0 ≤ k (III.4)

Problem III.4 is called the k-sparse approximation problem.Because real data always contains noise, representation noiseis unavoidable in most cases. Thus the original model III.2can be revised to a modified model with respect to smallpossible noise by denoting

y = Xα + s (III.5)

where s ∈ Rd refers to representation noise and is boundedas ‖s‖2 ≤ ε. With the presence of noise, the sparse solutionsof problems III.3 and III.4 can be approximately obtained byresolving the following optimization problems:

α = argmin ‖α‖0 s.t. ‖y− Xα‖22 ≤ ε (III.6)

or

α = argmin ‖y− Xα‖22 s.t. ‖α‖0 ≤ ε (III.7)

VOLUME 3, 2015 495

www.redpel.com6+917620593389

www.redpel.com6+917620593389

Page 7: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

FIGURE 4. The geometry of the solutions of different norm regularization in 2-D space [7]. (a), (b) and (c) are the geometry of the solutionsof the l0-norm, l1-norm, l2-norm minimization, respectively.

Furthermore, according to the Lagrange multiplier theorem,a proper constant λ exists such that problems III.6 and III.7are equivalent to the following unconstrained minimizationproblem with a proper value of λ.

α = L(α, λ) = argmin ‖y− Xα‖22 + λ‖α‖0 (III.8)

where λ refers to the Lagrange multiplier associatedwith ‖α‖0.

B. SPARSE REPRESENTATION WITHl1-NORM MINIMIZATIONThe l1-norm originates from the Lasso problem [42], [43]and it has been extensively used to address issues in machinelearning, pattern recognition, and statistics [53]–[55].Although the sparse representation method with l0-normminimization can obtain the fundamental sparse solutionof α over thematrixX , the problem is still a non-deterministicpolynomial-time hard (NP-hard) problem and the solution isdifficult to approximate [27]. Recent literature [20], [56]–[58]has demonstrated that when the representation solutionobtained by using the l1-norm minimization constraint isalso content with the condition of sparsity and the solutionusing l1-norm minimization with sufficient sparsity can beequivalent to the solution obtained by l0-norm minimizationwith full probability. Moreover, the l1-norm optimizationproblem has an analytical solution and can be solved inpolynomial time. Thus, extensive sparse representationmethods with the l1-norm minimization have been proposedto enrich the sparse representation theory. The applicationsof sparse representation with the l1-norm minimization areextraordinarily and remarkably widespread. Correspond-ingly, the main popular structures of sparse representationwith the l1-norm minimization, similar to sparse representa-tion with l0-norm minimization, are generally used to solvethe following problems:

α = argminα‖α‖1 s.t. y = Xα (III.9)

α = argminα‖α‖1 s.t. ‖y− Xα‖22 ≤ ε (III.10)

or

α = argminα‖y− Xα‖22 s.t. ‖α‖1 ≤ τ (III.11)

α = L(α, λ) = argminα

12‖y− Xα‖22 + λ‖α‖1 (III.12)

where λ and τ are both small positive constants.

C. SPARSE REPRESENTATION WITHlp-NORM (0 < p < 1) MINIMIZATIONThe general sparse representation method is to solve a linearrepresentation system with the lp-norm minimizationproblem. In addition to the l0-norm minimization andl1-norm minimization, some researchers are trying to solvethe sparse representation problem with the lp-norm (0<p<1)minimization, especially p = 0.1, 1

2 ,13 , or 0.9 [59]–[61].

That is, the sparse representation problem with the lp-norm(0<p<1) minimization is to solve the following problem:

α = argminα‖α‖pp s.t. ‖y− Xα‖22 ≤ ε (III.13)

or

α = L(α, λ) = argminα‖y− Xα‖22 + λ‖α‖

pp (III.14)

In spite of the fact that sparse representation methods with thelp-norm (0<p<1) minimization are not the mainstreammethods to obtain the sparse representation solution,it tremendously influences the improvements of the sparserepresentation theory.

D. SPARSE REPRESENTATION WITH l2-NORMAND l2,1-NORM MINIMIZATIONThe representation solution obtained by the l2-normminimization is not rigorously sparse. It can only obtain a‘limitedly-sparse’ representation solution, i.e. the solutionhas the property that it is discriminative and distinguishablebut is not really sparse enough [31]. The objective functionof the sparse representation method with the l2-normminimization is to solve the following problem:

α = argminα‖α‖22 s.t. ‖y− Xα‖22 ≤ ε (III.15)

496 VOLUME 3, 2015

www.redpel.com7+917620593389

www.redpel.com7+917620593389

Page 8: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

or

α = L(α, λ) = argminα‖y− Xα‖22 + λ‖α‖

22 (III.16)

On the other hand, the l2,1-norm is also called the rotationinvariant l1-norm, which is proposed to overcome thedifficulty of robustness to outliers [62]. The objectivefunction of the sparse representation problem with thel2,1-norm minimization is to solve the following problem:

argminA‖Y − XA‖2,1 + µ‖A‖2,1 (III.17)

where Y = [y1, y2, · · · , yN ] refers to the matrix composedof samples, A = [a1, a2, · · · , aN ] is the correspondingcoefficient matrix of X , and µ is a small positive constant.Sparse representation with the l2,1-norm minimization canbe implemented by exploiting the proposed algorithms inliterature [46]–[48].

IV. GREEDY STRATEGY APPROXIMATIONGreedy algorithms date back to the 1950s. The core idea ofthe greedy strategy [7], [24] is to determine the position basedon the relationship between the atom and probe sample, andthen to use the least square to evaluate the amplitude value.Greedy algorithms can obtain the local optimized solutionin each step in order to address the problem. However, thegreedy algorithm can always produce the global optimalsolution or an approximate overall solution [7], [24].Addressing sparse representation with l0-norm regulariza-tion, i.e. problem III.3, is an NP hard problem [20], [56].The greedy strategy provides a special way to obtain anapproximate sparse representation solution. The greedy strat-egy actually can not directly solve the optimization problemand it only seeks an approximate solution for problem III.3.

A. MATCHING PURSUIT ALGORITHMThe matching pursuit (MP) algorithm [63] is the earliestand representative method of using the greedy strategy toapproximate problem III.3 or III.4. The main idea of theMP is to iteratively choose the best atom from the dictionarybased on a certain similarity measurement to approximatelyobtain the sparse solution. Taking as an example of the sparsedecomposition with a vector sample y over the over-completedictionary D, the detailed algorithm description is presentedas follows:

Suppose that the initialized representation residual isR0 = y, D = [d1, d2, · · · , dN ] ∈ Rd×N and each samplein dictionary D is an l2-norm unity vector, i.e. ‖d i‖ = 1.To approximate y, MP first chooses the best matching atomfrom D and the selected atom should satisfy the followingcondition:

|〈R0, d l0〉| = sup|〈R0, d i〉| (IV.1)

where l0 is a label index from dictionary D. Thus y can bedecomposed into the following equation:

y = 〈y, d l0〉d l0 + R1 (IV.2)

So y = 〈R0, d l0〉d l0 + R1 where 〈R0, d l0〉d l0 represents theorthogonal projection of y onto d l0 , and R1 is the representa-tion residual by using d l0 to represent y. Considering the factthat d l0 is orthogonal to R1, Eq. IV.2 can be rewritten as

‖y‖2 = |〈y, d l0〉|2+ ‖R1‖

2 (IV.3)

To obtain the minimum representation residual, theMP algorithm iteratively figures out the best matching atomfrom the over-completed dictionary, and then utilizes therepresentation residual as the next approximation target untilthe termination condition of iteration is satisfied. For the t-thiteration, the best matching atom is d lt and the approximationresult is found from the following equation:

Rt = 〈Rt , d lt 〉d lt + Rt+1 (IV.4)

where the d lt satisfies the equation:

|〈Rt , d lt 〉| = sup|〈Rt , d i〉| (IV.5)

Clearly, d lt is orthogonal to Rk+1, and then

‖Rk‖2 = |〈Rt , d lt 〉|2+ ‖Rt+1‖2 (IV.6)

For the n-th iteration, the representation residual‖Rn‖2 ≤ τ where τ is a very small constant and the probesample y can be formulated as:

y =n−1∑j=1

〈Rj, d lj〉d lj + Rn (IV.7)

If the representation residual is small enough, the probesample y can approximately satisfy the following equation:y ≈

∑n−1j=1 〈Rj, d lj〉d lj where n� N . Thus, the probe sample

can be represented by a small number of elements from a largedictionary. In the context of the specific representation error,the termination condition of sparse representation is that therepresentation residual is smaller than the presupposed value.More detailed analysis on matching pursuit algorithms can befound in the literature [63].

B. ORTHOGONAL MATCHING PURSUIT ALGORITHMThe orthogonal matching pursuit (OMP) algorithm [37], [64]is an improvement of the MP algorithm. The OMP employsthe process of orthogonalization to guarantee the orthogonaldirection of projection in each iteration. It has been verifiedthat the OMP algorithm can be converged in limitediterations [37]. The main steps of OMP algorithm have beensummarized in Algorithm 1.

C. SERIES OF MATCHING PURSUIT ALGORITHMSIt is an excellent choice to employ the greedy strategy toapproximate the solution of sparse representation with thel0-norm minimization. These algorithms are typical greedyiterative algorithms. The earliest algorithms were thematching pursuit (MP) and orthogonal matchingpursuit (OMP). The basic idea of theMP algorithm is to selectthe best matching atom from the overcomplete dictionaryto construct sparse approximation during each iteration,

VOLUME 3, 2015 497

www.redpel.com8+917620593389

www.redpel.com8+917620593389

Page 9: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Algorithm 1 Orthogonal Matching Pursuit AlgorithmTask: Approximate the constraint problem:

α = argminα ‖α‖0 s.t. y = XαInput: Probe sample y, measurement matrix X , sparsecoefficients vector αInitialization: t = 1, r0 = y, α = 0, D0 = φ, index set30 = φ where φ denotes empty set, τ is a small constant.While ‖rt‖ > τ doStep 1: Find the best matching sample, i.e. the biggest

inner product between rt−1 and xj (j 6∈ 3t−1)by exploiting λt = argmaxj 6∈3t−1 |〈rt−1, xj〉|.

Step 2: Update the index set 3t = 3t−1⋃λt and

reconstruct data set Dt = [Dt−1, xλt ].Step 3: Compute the sparse coefficient by using the least

square algorithm α = argmin ‖y− Dt α‖22.Step 4: Update the representation residual using

rt = y− Dt α.Step 5: t = t + 1.

EndOutput: D, α

to compute the signal representation residual, and then tochoose the best matching atom till the stopping criterion ofiteration is satisfied. Many more greedy algorithms based onthe MP and OMP algorithm such as the efficient orthogonalmatching pursuit algorithm [65] subsequently have beenproposed to improve the pursuit algorithm. Needell et al.proposed an regularized version of orthogonal matchingpursuit (ROMP) algorithm [38], which recovered all k sparsesignals based on the Restricted Isometry Property of randomfrequency measurements, and then proposed another variantof OMP algorithm called compressive sampling matchingpursuit (CoSaMP) algorithm [66], which incorporated severalexisting ideas such as restricted isometry property (RIP) andpruning technique into a greedy iterative structure of OMP.Some other algorithms also had an impressive influence onfuture research on CS. For example, Donoho et al. pro-posed an extension of OMP, called stage-wise orthogonalmatching pursuit (StOMP) algorithm [67], which depictedan iterative algorithm with three main steps, i.e. threholding,selecting and projecting. Dai and Milenkovic proposed anew method for sparse signal reconstruction named subspacepursuit (SP) algorithm [68], which sampled signals satisfy-ing the constraints of the RIP with a constant parameter.Do et al. presented a sparsity adaptive matching pur-suit (SAMP) algorithm [69], which borrowed the idea ofthe EM algorithm to alternatively estimate the sparsity andsupport set. Jost et al. proposed a tree-based matching pur-suit (TMP) algorithm [70], which constructed a tree structureand employed a structuring strategy to cluster similar signalatoms from a highly redundant dictionary as a new dictionary.Subsequently, La and Do proposed a new tree-based orthog-onal matching pursuit (TBOMP) algorithm [71], whichtreated the sparse tree representation as an additional prior

knowledge for linear inverse systems by using a smallnumber of samples. Recently, Karahanoglu and Erdoganconceived a forward-backward pursuit (FBP) method [72]with two greedy stages, in which the forward stage enlargedthe support estimation and the backward stage removedsome unsatisfied atoms. More detailed treatments of thegreedy pursuit for sparse representation can be found in theliterature [24].

V. CONSTRAINED OPTIMIZATION STRATEGYConstrained optimization strategy is always utilized to obtainthe solution of sparse representation with the l1-norm regu-larization. The methods that address the non-differentiableunconstrained problem will be presented by reformulating itas a smooth differentiable constrained optimization problem.These methods exploit the constrained optimization methodwith efficient convergence to obtain the sparse solution.Whatis more, the constrained optimization strategy emphasizesthe equivalent transformation of ‖α‖1 in problem III.12 andemploys the new reformulated constrained problem to obtaina sparse representation solution. Some typical methods thatemploy the constrained optimization strategy to solve theoriginal unconstrained non-smooth problem are introduced inthis section.

A. GRADIENT PROJECTION SPARSE RECONSTRUCTIONThe core idea of the gradient projection sparse representationmethod is to find the sparse representation solution alongwiththe gradient descent direction. The first key procedure ofgradient projection sparse reconstruction (GPSR) [73]provides a constrained formulation where each value of α canbe split into its positive and negative parts. Vectors α+ and α−are introduced to denote the positive and negative coefficientsof α, respectively. The sparse representation solution α can beformulated as:

α = α+ − α−, α+ ≥ 0, α− ≥ 0 (V.1)

where the operator (·)+ denotes the positive-part operator,which is defined as (x)+ = max{0, x}. Thus, ‖α‖1 = 1Td α++1Td α−, where 1d = [1, 1, · · · , 1︸ ︷︷ ︸

d

]T is a d–dimensional vector

with d ones. Accordingly, problem III.12 can be reformulatedas a constrained quadratic problem:

argminL(α) = argmin12‖y− X [α+ − α−]‖22

+ λ(1Td α+ + 1Td α−) s.t. α+ ≥ 0, α− ≥ 0(V.2)

or

argminL(α) = argmin12‖y− [X+,X−][α+ − α−]‖22

+ λ(1Td α+ + 1Td α−) s.t. α+ ≥ 0, α− ≥ 0(V.3)

Furthermore, problem V.3 can be rewritten as:

argmin G(z) = cT z+12zTAz s.t. z ≥ 0 (V.4)

498 VOLUME 3, 2015

www.redpel.com9+917620593389

www.redpel.com9+917620593389

Page 10: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

where z = [α+;α−], c = λ12d + [−XT y;XT y], 12d =

[1, · · · , 1︸ ︷︷ ︸2d

]T , A =(

XTX −XTX−XTX XTX

).

The GPSR algorithm employs the gradient descent andstandard line-search method [32] to address problemV.4. Thevalue of z can be iteratively obtained by utilizing

argmin zt+1 = zt − σ∇G(zt ) (V.5)

where the gradient of ∇G(zt ) = c+Azt and σ is the step sizeof the iteration. For step size σ , GPSR updates the step sizeby using

σ t = argminσ

G(zt − σgt ) (V.6)

where the function gt is pre-defined as

gti =

{(∇G(zt ))i, if zti > 0 or (∇G(zt ))i < 00, otherwise.

(V.7)

Problem V.6 can be addressed with the close-form solution

σ t =(gt )T (gt )(gt )TA(gt )

(V.8)

Furthermore, the basic GPSR algorithm employs thebacktracking linear search method [32] to ensure that the stepsize of gradient descent, in each iteration, is a more propervalue. The stop condition of the backtracking linear searchshould satisfy

G((zt − σ t∇G(zt ))+) > G(zt )− β∇G(zt )T

×(zt − (zt − σ t∇G(zt ))+) (V.9)

where β is a small constant. The main steps of GPSR aresummarized in Algorithm 2. For more detailed information,one can refer to the literature [73].

Algorithm 2 Gradient Projection Sparse Reconstruc-tion (GPSR)Task: To address the unconstraint problem:

α = argminα 12‖y− Xα‖

22 + λ‖α‖1

Input: Probe sample y, the measurement matrix X , smallconstant λInitialization: t = 0, β ∈ (0, 0.5), γ ∈ (0, 1), given α sothat z = [α+,α−].While not converged doStep 1: Compute σ t exploiting Eq. V.8 and σ t ← mid

(σmin, σ t , σmax), where mid(·, ·, ·) denotes themiddle value of the three parameters.

Step 2: While Eq. V.9 not satisfieddo σ t ← γ σ t end

Step 3: zt+1 = (zt − σ t∇G(zt ))+ and t = t + 1.EndOutput: zt+1, α

B. INTERIOR-POINT METHOD BASED SPARSEREPRESENTATION STRATEGYThe Interior-point method [32] is not an iterative algorithmbut a smooth mathematic model and it always incorpo-rates the Newton method to efficiently solve unconstrainedsmooth problems of modest size [29]. When the Newtonmethod is used to address the optimization issue, a complexNewton equation should be solved iteratively which is verytime-consuming. A method named the truncated Newtonmethod can effectively and efficiently obtain the solution ofthe Newton equation. A prominent algorithm called thetruncated Newton based interior-point method (TNIPM)exists, which can be utilized to solve the large-scalel1-regularized least squares (i.e. l1_ls) problem [74].The original problem of l1_ls is to solve problem III.12 and

the core procedures of l1_ls are shown below:(1) Transform the original unconstrained non-smoothproblem to a constrained smooth optimization problem.(2) Apply the interior-point method to reformulate theconstrained smooth optimization problem as a newunconstrained smooth optimization problem.(3) Employ the truncated Newton method to solve thisunconstrained smooth problem.

The main idea of the l1_ls will be briefly described. Forsimplicity of presentation, the following one-dimensionalproblem is used as an example.

|α| = arg min−σ≤α≤σ

σ (V.10)

where σ is a proper positive constant.Thus, problem III.12 can be rewritten as

α = argmin12‖y− Xα‖22 + λ‖α‖1

= argmin12‖y− Xα‖22 + λ

∑N

i=1min

−σi≤αi≤σiσi

= argmin12‖y− Xα‖22 + λ min

−σi≤αi≤σi

∑N

i=1σi

= arg min−σi≤αi≤σi

12‖y− Xα‖22 + λ

∑N

i=1σi (V.11)

Thus problem III.12 is also equivalent to solve thefollowing problem:

α = arg minα,σ∈RN

12‖y− Xα‖22 + λ

N∑i=1

σi

s.t. − σi≤ αi≤ σi (V.12)

or

α = arg minα,σ∈RN

12‖y− Xα‖22 + λ

N∑i=1

σi

s.t. σi + αi ≥ 0, σi − αi ≥ 0 (V.13)

The interior-point strategy can be used to transformproblem V.13 into an unconstrained smooth problem

α = arg minα,σ∈RN

G(α, σ )=v2‖y−Xα‖22 + λv

N∑i=1

σi−B(α, σ )

(V.14)

VOLUME 3, 2015 499

www.redpel.com10+917620593389

www.redpel.com10+917620593389

Page 11: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

where B(α, σ ) =∑N

i=1 log(σi + αi)+∑N

i=1 log(σi − αi) is abarrier function, which forces the algorithm to be performedwithin the feasible region in the context of unconstrainedcondition.

Subsequently, l1_ls utilizes the truncated Newton methodto solve problem V.14. The main procedures of addressingproblem V.14 are presented as follows:

First, the Newton system is constructed

H[4α

]= −∇G(α, σ ) ∈ R2N (V.15)

where H = −∇2G(α, σ ) ∈ R2N×2N is the Hessian

matrix, which is computed using the preconditioned conju-gate gradient algorithm, and then the direction of linear search[4α,4σ ] is obtained.

Second, the Lagrange dual of problem III.12 is used toconstruct the dual feasible point and duality gap:

a) The Lagrangian function and Lagrange dual ofproblem III.12 are constructed. The Lagrangian function isreformulated as

L(α, z,u) = zT z+ λ‖α‖1 + u(Xα − y− z) (V.16)

where its corresponding Lagrange dual function is

α = argmaxF(u) = −14uTu− uT y

s.t. |(XTu)i| ≤ λi (i = 1, 2, · · · ,N ) (V.17)

b) A dual feasible point is constructed

u = 2s(y− Xα), s = min{λ/|2yi − 2(XTXα)i|}∀i (V.18)

where u is a dual feasible point and s is the step size of thelinear search.

c) The duality gap is constructed, which is the gap betweenthe primary problem and the dual problem:

g = ‖y− Xα‖ + λ‖α‖1 − F(u) (V.19)

Third, the method of backtracking linear search is used todetermine an optimal step size of the Newton linear search.The stopping condition of the backtracking linear search is

G(α + ηt4α, σ + ηt4σ ) > G(α, σ )

+ ρηt∇G(α, σ )[4α,4σ ] (V.20)

where ρ ∈ (0, 0.5) and ηt ∈ (0, 1) is the step size of theNewton linear search.

Finally, the termination condition of the Newton linearsearch is set to

ζ = min{0.1, βg/‖h‖2} (V.21)

where the function h = ∇G(α, σ ), β is a small constant,and g is the duality gap. The main steps of algorithm l1_lsare summarized in Algorithm 3. For further description andanalyses, please refer to the literature [74].

The truncated Newton based interior-pointmethod (TNIPM) [75] is a very effective method to solvethe l1-norm regularization problems. Koh et al. [76] also

Algorithm 3 Truncated Newton Based Interior-PointMethod (TNIPM) for l1_lsTask: To address the unconstraint problem:

α = argminα 12‖y− Xα‖

22 + λ‖α‖1

Input: Probe sample y, the measurement matrix X , smallconstant λInitialization: t = 1, v = 1

λ, ρ ∈ (0, 0.5), σ = 1N

Step 1: Employ preconditioned conjugate gradient algo-rithm to obtain the approximation of H in Eq. V.15,and then obtain the descent direction of linear search[4αt ,4σ t ].Step 2: Exploit the algorithm of backtracking linear searchto find the optimal step size of Newton linear search ηt ,which satisfies the Eq. V.20.Step 3: Update the iteration point utilizing (αt+1, σ t+1) =(αt , σ t )+ (4αt +4σ t ).Step 4: Construct feasible point using eq. V.18 and dualitygap in Eq. V.19, and compute the termination tolerance ζin Eq. V.21.Step 5: If the condition g/F(u) > ζ is satisfied, stop;Otherwise, return to step 1, update v in Eq. V.14 andt = t + 1.Output: α

utilized the TNIPM to solve large scale logistic regressionproblems, which employed a preconditioned conjugategradient method to compute the search step size withwarm-start techniques. Mehrotra proposed to exploit theinterior-point method to address the primal-dual problem [77]and introduced the second-order derivation of Taylor polyno-mial to approximate a primal-dual trajectory. More analysesof interior-point method for sparse representation can befound in the literature [78].

C. ALTERNATING DIRECTION METHOD (ADM) BASEDSPARSE REPRESENTATION STRATEGYThis section shows how the ADM [44] is used to solve primaland dual problems in III.12. First, an auxiliary variable isintroduced to convert problem in III.12 into a constrainedproblem with the form of problem V.22. Subsequently, thealternative direction method is used to efficiently address thesub-problems of problem V.22. By introducing the auxiliaryterm s ∈ Rd , problem III.12 is equivalent to a constrainedproblem

argminα,s

12τ‖s‖2 + ‖α‖1 s.t. s = y− Xα (V.22)

The optimization problem of the augmented Lagrangianfunction of problem V.22 is considered

arg minα,s,λ

L(α, s,λ) =12τ‖s‖2 + ‖α‖1 − λT

×(s+ Xα − y)+µ

2‖s+ Xα − y‖22

(V.23)

500 VOLUME 3, 2015

www.redpel.com11+917620593389

www.redpel.com11+917620593389

Page 12: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

where λ ∈ Rd is a Lagrange multiplier vector and µ is apenalty parameter. The general framework of ADM is usedto solve problem V.23 as follows:

st+1 = argminL(s,αt ,λt ) (a)αt+1 = argminL(st+1,α,λt ) (b)λt+1 = λt − µ(st+1 + Xαt+1 − y) (c)

(V.24)

First, the first optimization problem V.24(a) is considered

argminL(s,αt ,λt ) =12τ‖s‖2 + ‖αt‖1 − (λt )T

×(s+ Xαt − y)+µ

2‖s+ Xαt − y‖22

=12τ‖s‖2−(λt )T s+

µ

2‖s+Xαt − y‖22

+‖αt‖1 − (λt )T (Xαt − y) (V.25)

Then, it is known that the solution of problem V.25 withrespect to s is given by

st+1 =τ

1+ µτ(λt − µ(y− Xαt )) (V.26)

Second, the optimization problem V.24(b) is considered

argminL(st+1,α, λt )

=12τ‖st+1‖2 + ‖α‖1 − (λ)T

×(st+1 + Xα − y)+µ

2‖st+1 + Xα − y‖22

which is equivalent to

argmin{‖α‖1 − (λt )T (st+1 + Xα − y)+µ

2‖st+1Xα − y‖22}

= ‖α‖1 +µ

2‖st+1 + Xα − y− λt/µ‖22

= ‖α‖1 + f (α) (V.27)

where f (α) = µ2 ‖s

t+1+Xα−y−λt/µ‖22. If the second order

Taylor expansion is used to approximate f (α), theproblem V.27 can be approximately reformulated as

argmin{‖α‖1 + (α − αt )TXT (st+1 + Xαt − y− λt/µ)

+12τ‖α − αt‖22} (V.28)

where τ is a proximal parameter. The solution ofproblem V.28 can be obtained by the soft thresholdingoperator

αt+1 = soft{αt − τXT (st+1 + Xαt − y− λt/µ),τ

µ}

(V.29)

where soft(σ, η) = sign(σ ) max{|σ | − η, 0}.Finally, the Lagrange multiplier vector λ is updated by

using Eq. V.24(c).The algorithm presented above utilizes the second order

Taylor expansion to approximately solve thesub-problem V.27 and thus the algorithm is denoted as aninexact ADM or approximate ADM. The main proceduresof the inexact ADM based sparse representation method are

summarized in Algorithm 4. More specifically, the inexactADM described above is to reformulate the unconstrainedproblem as a constrained problem, and then utilizes thealternative strategy to effectively address the correspond-ing sub-optimization problem. Moreover, ADM can alsoefficiently solve the dual problems of the primalproblems III.9-III.12. For more information, please refer tothe literature [44], [79].

Algorithm 4 Alternating Direction Method (ADM) BasedSparse Representation StrategyTask: To address the unconstraint problem:

α = argminα 12‖y− Xα‖

22 + τ‖α‖1

Input: Probe sample y, the measurement matrix X , smallconstant λInitialization: t = 0, s0 = 0, α0 = 0, λ0 = 0, τ = 1.01,µ is a small constant.Step 1: Construct the constraint optimization problem ofproblem III.12 by introducing the auxiliary parameterand its augmented Lagrangian function, i.e.problem (V.22) and (V.23).While not converged doStep 2: Update the value of the st+1 by using Eq. (V.25).Step 2: Update the value of the αt+1 by using Eq. (V.29).Step 3: Update the value of the λt+1 by using

Eq. (V.24(c)).Step 4: µt+1 = τµt and t = t + 1.

End WhileOutput: αt+1

VI. PROXIMITY ALGORITHM BASEDOPTIMIZATION STRATEGYIn this section, the methods that exploit the proximityalgorithm to solve constrained convex optimization problemsare discussed. The core idea of the proximity algorithm isto utilize the proximal operator to iteratively solve thesub-problem, which is much more computationally efficientthan the original problem. The proximity algorithm isfrequently employed to solve nonsmooth, constrainedconvex optimization problems [29]. Furthermore, the generalproblem of sparse representation with l1-norm regulariza-tion is a nonsmooth convex optimization problem, whichcan be effectively addressed by using the proximalalgorithm.

Suppose a simple constrained optimization problem is

min{h(x)|x ∈ χ} (VI.1)

where χ ⊂ Rn. The general framework of addressing theconstrained convex optimization problem VI.1 using theproximal algorithm can be reformulated as

xt = argmin{h(x)+τ

2‖x− xt‖2|x ∈ χ} (VI.2)

where τ and xt are given. For definiteness and without lossof generality, it is assumed that there is the following linear

VOLUME 3, 2015 501

www.redpel.com12+917620593389

www.redpel.com12+917620593389

Page 13: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

constrained convex optimization problem

argmin{F(x)+ G(x)|x ∈ χ} (VI.3)

The solution of problem VI.3 obtained by employing theproximity algorithm is:

xt+1 = argmin{F(x)+ 〈∇G(xt ), x− xt 〉 +12τ‖x− xt‖2}

= argmin{F(x)+12τ‖x− θ t‖2} (VI.4)

where θ = xt − τ∇G(xt ). More specifically, for the sparserepresentation problem with l1-norm regularization, the mainproblem can be reformulated as:

minP(α) = {λ‖α‖1 | Aα = y}

or minP(α) = {λ‖α‖1 + ‖Aα − y‖22 | α ∈ Rn} (VI.5)

which are considered as the constrained sparse representationof problem III.12.

A. SOFT THRESHOLDING OR SHRINKAGE OPERATORFirst, a simple form of problem III.12 is introduced, whichhas a closed-form solution, and it is formulated as:

α∗ = minαh(α) = λ‖α‖1 +

12‖α − s‖2

=

N∑j=1

λ|αj| +

N∑j=1

12(αj − sj)2 (VI.6)

where α∗ is the optimal solution of problem VI.6, and thenthere are the following conclusions:(1) if αj > 0, then h(α) = λα + 1

2‖α − s‖2 and its derivative

is h′(αj) = λ+ α∗j − sj.Let h′(αj) = 0⇒ α∗j = sj − λ, where it indicates sj > λ;(2) if αj < 0, then h(α) = −λα+ 1

2‖α−s‖2 and its derivative

is h′(αj) = −λ+ α∗j − sj.Let h′(αj) = 0⇒ α∗j = sj + λ, where it indicates sj < −λ;(3) if −λ ≤ sj ≤ λ, and then α∗j = 0.So the solution of problem VI.6 is summarized as

α∗j =

sj − λ, if sj > λ

sj + λ, if sj < −λ0, otherwise

(VI.7)

The equivalent expression of the solution isα∗ = shrink(s, λ), where the j-th component of shrink(s, λ)is shrink(s, λ)j = sign(sj) max{|sj| − λ, 0}. The operatorshrink(•) can be regarded as a proximal operator.

B. ITERATIVE SHRINKAGE THRESHOLDINGALGORITHM (ISTA)The objective function of ISTA [80] has the form of

argminF(α) =12‖Xα − y‖22 + λ‖α‖1 = f (α)+ λg(α)

(VI.8)

and is usually difficult to solve. Problem VI.8 can beconverted to the form of an easy problemVI.6 and the explicitprocedures are presented as follows.

First, Taylor expansion is used to approximatef (α) = 1

2‖Xα−y‖22 at a point of α

t . The second order Taylorexpansion is

f (α) = f (αt )+ (α − αt )T∇f (αt )

+12(α − αt )THf (αt )(α − αt )+ · · · (VI.9)

where Hf (αt ) is the Hessian matrix of f (α) at αt . For thefunction f (α), ∇f (α) = XT (Xα − y) and Hf (α) = XTXcan be obtained.

f (α) =12‖Xαt − y‖22 + (α − αt )TXT (Xαt − y)

+12(α − αt )TXTX (α − αt ) (VI.10)

If the Hessian matrix Hf (α) is replaced or approximated inthe third term above by using a scalar 1

τI , and then

f (α) ≈12‖Xαt − y‖22 + (α − αt )TXT (Xαt − y)

+12τ

(α − αt )T (α − αt ) = Qt (α,αt ) (VI.11)

Thus problem VI.8 using the proximal algorithm can besuccessively addressed by

αt+1 = argminQt (α,αt )+ λ‖α‖1 (VI.12)

Problem VI.12 is reformulated to a simple form ofproblem VI.6 by

Qt (α,αt ) =12‖Xαt − y‖22 + (α − αt )TXT (Xαt − y)

+12τ‖α − αt‖22

=12‖Xαt−y‖22 +

12τ‖α − αt+ τXT (Xαt− y)‖22

−τ

2‖XT (Xαt − y)‖22

=12τ‖α − (αt − τXT (Xαt − y))‖22 + B(α

t )

(VI.13)

where the term B(αt ) = 12‖Xα

t− y‖22 −

τ2‖X

T (Xαt − y)‖2

in problem VI.12 is a constant with respect to variable α, andit can be omitted. As a result, problem VI.12 is equivalent tothe following problem:

αt+1 = argmin12τ‖α − θ (αt )‖22 + λ‖α‖1 (VI.14)

where θ (αt ) = αt − τXT (Xαt − y).The solution of the simple problem VI.6 is applied to

solve problem VI.14 where the parameter t is replaced bythe equation θ (αt ), and the solution of problem VI.14 isαt+1 = shrink(θ (αt ), λτ ). Thus, the solution of ISTA isreached. The techniques used here are called linearization orpreconditioning and more detailed information can be foundin the literature [80], [81].

502 VOLUME 3, 2015

www.redpel.com13+917620593389

www.redpel.com13+917620593389

Page 14: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

C. FAST ITERATIVE SHRINKAGE THRESHOLDINGALGORITHM (FISTA)The fast iterative shrinkage thresholding algorithm (FISTA)is an improvement of ISTA. FISTA [82] not only preservesthe efficiency of the original ISTA but also promotes theeffectiveness of ISTA so that FISTA can obtain globalconvergence.

Considering that the HessianmatrixHf (α) is approximatedby using a scalar 1

τI for ISTA in Eq. VI.9, FISTA utilizes the

minimumLipschitz constant of the gradient∇f (α) to approx-imate the Hessian matrix of f (α), i.e. L(f ) = 2λmax(XTX ).Thus, the problem VI.8 can be converted to the problembelow:

f (α) ≈12‖Xαt − y‖22 + (α − αt )TXT (Xαt − y)

+L2(α − αt )T (α − αt ) = Pt (α,αt ) (VI.15)

where the solution can be reformulated as

αt+1 = argminL2‖α − θ (αt )‖22 + λ‖α‖1 (VI.16)

where θ (αt ) = αt − 1LX

T (Xαt − y).Moreover, to accelerate the convergence of the algorithm,

FISTA also improves the sequence of iteration points, insteadof employing the previous point it utilizes a specific linearcombinations of the previous two points {αt ,αt−1}, i.e.

αt = αt +µt − 1µt+1

(αt − αt−1) (VI.17)

where µt is a positive sequence, which satisfiesµt ≥ (t+1)/2, and the main steps of FISTA are summarizedin Algorithm 5. The backtracking linear research strategycan also be utilized to explore a more feasible value of Land more detailed analyses on FISTA can be found in theliterature [82], [83].

Algorithm 5 Fast Iterative Shrinkage ThresholdingAlgorithm (FISTA)Task: To address the problem α = argminF(α) =12‖Xα − y‖

22 + λ‖α‖1

Input: Probe sample y, the measurement matrix X , smallconstant λInitialization: t = 0, µ0

= 1, L = 23max(XTX ), i.e.Lipschitz constant of ∇f .While not converged doStep 1: Exploit the shrinkage operator in equation VI.7

to solve problem VI.16.

Step 2: Update the value ofµ usingµt+1 = 1+√

1+4(µt )2

2 .Step 3: Update iteration sequence αt using

equation VI.17.EndOutput: α

D. SPARSE RECONSTRUCTION BY SEPARABLEAPPROXIMATION (SpaRSA)Sparse reconstruction by separable approxi-mation (SpaRSA) [84] is another typical proximity algorithmbased on sparse representation, which can be viewed as anaccelerated version of ISTA. SpaRSA provides a generalalgorithmic framework for solving the sparse representationproblem and here a simple specific SpaRSA with adaptivecontinuation on ISTA is introduced. The main contributionsof SpaRSA are trying to optimize the parameter λ inproblem VI.8 by using the worm-starting technique, i.e.continuation, and choosing a more reliable approximationof Hf (α) in problem VI.9 using the Barzilai-Borwein (BB)spectral method [85]. The worm-starting technique andBB spectral approach are introduced as follows.

1) UTILIZING THE WORM-STARTING TECHNIQUETO OPTIMIZE λThe values of λ in the sparse representation methodsdiscussed above are always set to be a specific small con-stant. However, Hale et al. [86] concluded that the techniquethat exploits a decreasing value of λ from a warm-startingpoint can more efficiently solve the sub-problem VI.14 thanISTA that is a fixed point iteration scheme. SpaRSA usesan adaptive continuation technique to update the value of λso that it can lead to the fastest convergence. The procedureregenerates the value of λ using

λ = max{γ ‖XT y‖∞, λ} (VI.18)

where γ is a small constant.

2) UTILIZING THE BB SPECTRAL METHODTO APPROXIMATE Hf (α)ISTA employs 1

τI to approximate the matrix Hf (α), which

is the Hessian matrix of f (α) in problem VI.9 and FISTAexploits the Lipschitz constant of ∇f (α) to replace Hf (α).However, SpaRSA utilizes the BB spectral method to choosethe value of τ to mimic the Hessian matrix. The value of τ isrequired to satisfy the condition:

1τ t+1

(αt+1 − αt ) ≈ ∇f (αt+1)−∇f (αt ) (VI.19)

which satisfies the minimization problem

1τ t+1

= argmin ‖1τ(αt+1 − αt )− (∇f (αt+1)−∇f (αt ))‖22

=(αt+1 − αt )T (∇f (αt+1)−∇f (αt ))

(αt+1 − αt )T (αt+1 − αt )(VI.20)

For problem VI.14, SpaRSA requires that the valueof λ is a decreasing sequence using the Eq. VI.18 and thevalue of τ should meet the condition of Eq. VI.20. Thesparse reconstruction by separable approximation (SpaRSA)is summarized in Algorithm 6 and more information can befound in the literature [84].

VOLUME 3, 2015 503

www.redpel.com14+917620593389

www.redpel.com14+917620593389

Page 15: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Algorithm 6 Sparse Reconstruction by SeparableApproximation (SpaRSA)Task: To address the problem

α = argminF(α) = 12‖Xα − y‖

22 + λ‖α‖1

Input: Probe sample y, the measurement matrix X , smallconstant λInitialization: t = 0, i = 0, y0 = y, 1

τ 0I ≈ Hf (α) = XTX ,

tolerance ε = 10−5.Step 1: λt = max{γ ‖XT yt‖∞, λ}.Step 2: Exploit shrinkage operator to solve

problem VI.14, i.e. αi+1 = shrink(αi − τ iXT (XTαt − y), λtτ i).

Step 3: Update the value of 1τ i+1

using the Eq. VI.20.

Step 4: If ‖αi+1−αi‖

αi≤ ε, go to step 5; Otherwise,

return to step 2 and i = i+ 1.Step 5: yt+1 = y− Xαt+1.Step 6: If λt = λ, stop; Otherwise, return to step 1 and

t = t + 1.Output: αi

E. l1/2-NORM REGULARIZATION BASEDSPARSE REPRESENTATIONSparse representation with the lp-norm (0<p<1) regulariza-tion leads to a nonconvex, nonsmooth, and non-Lipschitzoptimization problem and its general forms are described asproblems III.13 and III.14. The lp-norm (0<p<1) regulariza-tion problem is always difficult to be efficiently addressedand it has also attracted wide interests from large numbers ofresearch groups. However, the research group led byZongbenXu summarizes the conclusion that themost impres-sive and representative algorithm of the lp-norm (0<p<1)regularization is sparse representation with the l1/2-normregularization [87]. Moreover, they have proposed someeffective methods to solve the l1/2-norm regularizationproblem [60], [88].

In this section, a half proximal algorithm is introducedto solve the l1/2-norm regularization problem [60], whichmatches the iterative shrinkage thresholding algorithm for thel1-norm regularization discussed above and the iterative hardthresholding algorithm for the l0-norm regularization. Sparserepresentation with the l1/2-norm regularization is explicitlyto solve the problem as follows:

α = argmin{F(α) = ‖Xα − y‖22 + λ‖α‖1/21/2} (VI.21)

where the first-order optimality condition of F(α) on α canbe formulated as

∇F(α) = XT (Xα − y)+λ

2∇(‖α‖1/21/2) = 0 (VI.22)

which admits the following equation:

XT (y− Xα) =λ

2∇(‖α‖1/21/2) (VI.23)

where ∇(‖α‖1/21/2) denotes the gradient of the regulariza-

tion term ‖α‖1/21/2. Subsequently, an equivalent transformation

of Eq. VI.23 is made by multiplying a positive constant τ andadding a parameter α to both sides. That is,

α + τXT (y− Xα) = α + τλ

2∇(‖α‖1/21/2) (VI.24)

To this end, the resolvent operator [60] is introduced tocompute the resolvent solution of the right part of Eq. VI.24,and the resolvent operator is defined as

Rλ, 12

(•) =(I +

λτ

2∇(‖ • ‖1/21/2)

)−1(VI.25)

which is very similar to the inverse function of the right partof Eq. VI.24. The resolvent operator is always satisfied nomatter whether the resolvent solution of ∇(‖ • ‖1/21/2) exists ornot [60]. Applying the resolvent operator to solveproblem VI.24

α = (I +λτ

2∇(‖ • ‖1/21/2))

−1(α + τX t (y− Xα))

= Rλ,1/2(α + τXT (y− Xα)) (VI.26)

can be obtained which is well-defined. θ (α) = α + τXT

(y − Xα) is denoted and the resolvent operator can beexplicitly expressed as:

Rλ, 12

(x) = (fλ, 12

(x1), fλ, 12(x2), · · · , fλ, 12

(xN ))T (VI.27)

where

fλ, 12

(xi) =23xi(1+ cos(

2π3−

23gλ(xi)),

gλ(xi) = arg cos(λ

8(|xi|3

)−32 ) (VI.28)

which have been demonstrated in the literature [60].Thus the half proximal thresholding function for the

l1/2-norm regularization is defined as below:

hλτ, 12

(xi) =

{fλτ, 12

(xi), if |xi| >3√544 (λτ )

23

0, otherwise(VI.29)

where the threshold3√544 (λτ )

23 has been conceived and

demonstrated in the literature [60].Therefore, if Eq. VI.29 is applied to Eq. VI.27, the half

proximal thresholding function, instead of the resolventoperator, for the l1/2-norm regularization problem VI.25 canbe explicitly reformulated as:

α = Hλτ, 12

(θ (α)) (VI.30)

where the half proximal thresholding operator H [60] isdeductively constituted by Eq. VI.29.

Up to now, the half proximal thresholding algorithm hasbeen completely structured by Eq. VI.30. However, theoptions of the regularization parameter λ in Eq. VI.24 canseriously dominate the quality of the representation solutionin problemVI.21, and the values of λ and τ can be specificallyfixed by

τ =1− ε‖X‖2

and λ =

√969τ|[θ (α)]k+1|

32 (VI.31)

504 VOLUME 3, 2015

www.redpel.com15+917620593389

www.redpel.com15+917620593389

Page 16: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

where ε is a very small constant, which is very close tozero, the k denotes the limit of sparsity (i.e. k-sparsity), and[•]k refers to the k-th largest component of [•]. The halfproximal thresholding algorithm for l1/2-norm regularizationbased sparse representation is summarized in Algorithm 7and more detailed inferences and analyses can be found inthe literature [60], [88].

Algorithm 7 The Half Proximal Thresholding Algorithm forl1/2-Norm RegularizationTask: To address the problem

α = argminF(α) = ‖Xα − y‖22 + λ‖α‖1/21/2

Input: Probe sample y, the measurement matrix XInitialization: t = 0, ε = 0.01, τ = 1−ε

‖X‖2.

While not converged doStep 1: Compute θ (αt ) = αt + τXT (y− Xαt ).Step 2: Compute λt =

√969τ |[θ (α

t )]k+1|32 in Eq. VI.31.

Step 3: Apply the half proximal thresholding operator toobtain the representation solution αt+1 =Hλtτ,

12(θ (αt )).

Step 4: t = t + 1.EndOutput: α

F. AUGMENTED LAGRANGE MULTIPLIER BASEDOPTIMIZATION STRATEGYThe Lagrange multiplier is a widely used tool to eliminatethe equality constrained problem and convert it to address theunconstrained problem with an appropriate penalty function.Specifically, the sparse representation problem III.9 can beviewed as an equality constrained problem and the equivalentproblem III.12 is an unconstrained problem, which augmentsthe objective function of problem III.9 with a weightedconstraint function. In this section, the augmented Lagrangianmethod (ALM) is introduced to solve the sparse representa-tion problem III.9.

First, the augmented Lagrangian function of problem III.9is conceived by introducing an additional equalityconstrained function, which is enforced on the Lagrangefunction in problem III.12. That is,

L(α, λ) = ‖α‖1 +λ

2‖y− Xα‖22 s.t. y− Xα = 0

(VI.32)

Then, a new optimization problem VI.32 with the form of theLagrangain function is reformulated as

argminLλ(α, z) = ‖α‖1 +λ

2‖y− Xα‖22 + z

T (y− Xα)

(VI.33)

where z ∈ Rd is called the Lagrange multiplier vector or dualvariable and Lλ(α, z) is denoted as the augmented Lagrangianfunction of problem III.9. The optimization problem VI.33is a joint optimization problem of the sparse representation

coefficient α and the Lagrange multiplier vector z.Problem VI.33 is solved by optimizing α and z alternativelyas follows:

αt+1 = argminLλ(α, zt )

= argmin(‖α‖1 +λ

2‖y− Xα‖22 + (zt )TXα) (VI.34)

zt+1 = zt + λ(y− Xαt+1) (VI.35)

where problem VI.34 can be solved by exploiting theFISTA algorithm. Problem VI.34 is iteratively solved and theparameter z is updated using Eq. VI.35 until the terminationcondition is satisfied. Furthermore, if the method of employ-ingALM to solve problemVI.33 is denoted as the primal aug-mented Lagrangian method (PALM) [89], the dual functionof problem III.9 can also be addressed by the ALM algorithm,which is denoted as the dual augmented Lagrangianmethod (DALM) [89]. Subsequently, the dual optimizationproblem III.9 is discussed and the ALM algorithm is utilizedto solve it.First, consider the following equation:

‖α‖1 = max‖θ‖∞≤1

〈θ ,α〉 (VI.36)

which can be rewritten as

‖α‖1 = max{〈θ ,α〉 − IB1∞}or ‖α‖1 = sup{〈θ ,α〉 − IB1∞} (VI.37)

where Bλp = {x ∈ RN | ‖x‖p ≤ λ} and I�(x) is a indicator

function, which is defined as I�(x) ={0, x ∈ �∞, x 6∈ �.

Hence,

‖α‖1 = max{〈θ ,α〉 : θ ∈ B1∞} (VI.38)

Second, consider the Lagrange dual problem ofproblem III.9 and its dual function is

g(λ) = infα{‖α‖1+λ

T (y−Xα)}=λT y− supα{λTXα−‖α‖1}

(VI.39)

where λ ∈ Rd is a Lagrangian multiplier. If the definition ofconjugate function is applied to Eq. VI.37, it can be verifiedthat the conjugate function of IB1∞ (θ ) is ‖α‖1. Thus Eq. VI.39can be equivalently reformulated as

g(λ) = λT y− IB1∞ (XTλ) (VI.40)

The Lagrange dual problem, which is associated with theprimal problem III.9, is an optimization problem:

maxλλT y s.t. (XTλ) ∈ B1∞ (VI.41)

Accordingly,

minλ,z−λT y s.t. z− XTλ = 0, z ∈ B1∞ (VI.42)

Then, the optimization problemVI.42 can be reconstructed as

arg minλ,z,µ

L(λ, z,µ) = −λT y− µT (z− XTλ)

2‖z−XTλ‖22 s.t. z ∈ B1∞ (VI.43)

VOLUME 3, 2015 505

www.redpel.com16+917620593389

www.redpel.com16+917620593389

Page 17: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

where µ ∈ Rd is a Lagrangian multiplier and τ is a penaltyparameter.

Finally, the dual optimization problem VI.43 is solved anda similar alternating minimization idea of PALM can also beapplied to problem VI.43, that is,

zt+1 = arg minz∈B1∞

Lτ (λt , z,µt )

= arg minz∈B1∞{−µT (z−XTλt )+

τ

2‖z−XTλt‖22}

= arg minz∈B1∞{τ

2‖z− (XTλt +

2τµT )‖22}

= PB1∞ (XTλt +

1τµT ) (VI.44)

where PB1∞ (u) is a projection, or called a proximal operator,onto B1∞ and it is also called group-wise soft-thresholding.For example, let x = PB1∞ (u), then the i-th component ofsolution x satisfies xi = sign(ui) min{|ui|, 1}

λt+1 = argminλLτ (λ, zt+1,µt )

= argminλ{−λT y+ (µt )TXTλ+

τ

2‖zt+1 − XTλ‖22}

= Q(λ) (VI.45)

Take the derivative of Q(λ) with respect to λ and obtain

λt+1 = (τXXT )−1(τXzt+1 + y− Xµt ) (VI.46)

µt+1 = µt − τ (zt+1 − XTλt+1) (VI.47)

The DALM for sparse representation with l1-norm regu-larization mainly exploits the augmented Lagrange methodto address the dual optimization problem of problem III.9and a proximal operator, the projection operator, is utilizedto efficiently solve the subproblem. The algorithm of DALMis summarized in Algorithm 8. For more detailed description,please refer to the literature [89].

Algorithm 8 Dual Augmented Lagrangian Method forl1-Norm RegularizationTask: To address the dual problem of α =

argminα ‖α‖1 s.t. y = XαInput: Probe sample y, the measurement matrix X , a smallconstant λ0.Initialization: t = 0, ε = 0.01, τ = 1−ε

‖X‖2, µ0= 0.

While not converged doStep 1: Apply the projection operator to compute

zt+1 = PB1∞ (XTλt + 1

τµT ).

Step 2: Update the value of λt+1 = (τXXT )−1

(τXzt+1 + y− Xµt ).Step 3: Update the value of µt+1 = µt−

τ (zt+1 − XTλt+1).Step 4: t = t + 1.

End WhileOutput: α = µ[1 : N ]

G. OTHER PROXIMITY ALGORITHM BASEDOPTIMIZATION METHODSThe theoretical basis of the proximity algorithm is to firstconstruct a proximal operator, and then utilize the proximaloperator to solve the convex optimization problem. Massiveproximity algorithms have followed up with improvedtechniques to improve the effectiveness and efficiency ofproximity algorithm based optimization methods. For exam-ple, Elad et al. proposed an iterative method named parallelcoordinate descent algorithm (PCDA) [90] by introducing theelement-wise optimization algorithm to solve the regularizedlinear least squares with non-quadratic regularizationproblem.

Inspired by belief propagation in graphical models,Donoho et al. developed a modified version of theiterative thresholding method, called approximate messagepassing (AMP) method [91], to satisfy the requirement thatthe sparsity undersampling tradeoff of the new algorithmis equivalent to the corresponding convex optimizationapproach. Based on the development of the first-ordermethodcalled Nesterov’s smoothing framework in convex opti-mization, Becker et al. proposed a generalized Nesterov’salgorithm (NESTA) [92] by employing the continuation-like scheme to accelerate the efficiency and flexibility.Subsequently, Becker et al. [93] further constructed a generalframework, i.e. templates for convex cone solvers (TFOCS),for solving massive certain types of compressed sensingreconstruction problems by employing the optimal first-ordermethod to solve the smoothed dual problem of the equiva-lent conic formulation of the original optimization problem.Further detailed analyses and inference information relatedto proximity algorithms can be found in theliterature [29], [83].

VII. HOMOTOPY ALGORITHM BASEDSPARSE REPRESENTATIONThe concept of homotopy derives from topology and thehomotopy technique is mainly applied to address a nonlinearsystem of equations problem. The homotopy method wasoriginally proposed to solve the least square problem withthe l1-penalty [94]. The main idea of homotopy is to solvethe original optimization problems by tracing a continuousparameterized path of solutions along with varyingparameters. Having a highly intimate relationship with theconventional sparse representationmethod such as least angleregression (LAR) [43], OMP [64] and polytope facespursuit (PFP) [95], the homotopy algorithm has been success-fully employed to solve the l1-norm minimization problems.In contrast to LAR and OMP, the homotopy method is morefavorable for sequentially updating the sparse solution byadding or removing elements from the active set. Somerepresentative methods that exploit the homotopy-basedstrategy to solve the sparse representation problem withthe l1-norm regularization are explicitly presented in thefollowing parts of this section.

506 VOLUME 3, 2015

www.redpel.com17+917620593389

www.redpel.com17+917620593389

Page 18: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

A. LASSO HOMOTOPYBecause of the significance of parameters in l1-normminimization, the well-known LASSO homotopy algorithmis proposed to solve the LASSO problem in III.9 by tracingthe whole homotopy solution path in a range of decreasingvalues of parameter λ. It is demonstrated that problem III.12with an appropriate parameter value is equivalent toproblem III.9 [30]. Moreover, it is apparent that as we changeλ from a very large value to zero, the solution ofproblem III.12 is converging to the solution ofproblem III.9 [30]. The set of varying value λ conceivesthe solution path and any point on the solution path is theoptimality condition of problem III.12. More specifically, theLASSO homotopy algorithm starts at an large initial value ofparameter λ and terminates at a point of λ, which approx-imates zero, along the homotopy solution path so that theoptimal solution converges to the solution of problem III.9.The fundamental of the homotopy algorithm is that thehomotopy solution path is a piecewise linear path with adiscrete number of operations while the value of the homo-topy parameter changes, and the direction of each segmentand the step size are absolutely determined by the signsequence and the support of the solution on the correspondingsegment, respectively [96].

Based on the basic ideas in a convex optimization problem,it is a necessary condition that the zero vector should bea solution of the subgradient of the objective function ofproblem III.12. Thus, we can obtain the subgradiential of theobjective function with respect to α for any given value of λ,that is,

∂L∂α= −XT (y− Xα)+ λ∂‖α‖1 (VII.1)

where the first term r = XT (y − Xα) is called the vectorof residual correlations, and ∂‖α‖1 is the subgradientobtained by

∂‖α‖1 =

{θ ∈ RN

∣∣∣∣ θi = sgn(αi), αi 6= 0θi ∈ [−1, 1], αi = 0

}Let3 and u denote the support of α and the sign sequence

of α on its support 3, respectively. X3 denotes that theindices of all the samples in X3 are all included in thesupport set3. If we analyze the KKT optimality condition forproblem III.12, we can obtain the following two equivalentconditions of problem VII.1, i.e.

X3(y− Xα) = λu; ‖XT3c (y− Xα)‖∞ ≤ λ (VII.2)

where 3c denotes the complementary set of the set 3.Thus, the optimality conditions in VII.2 can be divided intoN constraints and the homotopy algorithm maintains both ofthe conditions along the optimal homotopy solution path forany λ ≥ 0. As we decrease the value of λ to λ−τ , for a smallvalue of τ , the following conditions should be satisfied

XT3(y− Xα)+ τXT3Xδ = (λ− τ )u (a)

‖p+ τq‖∞ ≤ λ− τ (b) (VII.3)

where p = XT (y − Xα), q = XTXδ and δ is the updatedirection.

Generally, the homotopy algorithm is implementediteratively and it follows the homotopy solution path byupdating the support set by decreasing parameter λ froma large value to the desired value. The support set of thesolution will be updated and changed only at a critical pointof λ, where either an existing nonzero element shrinks tozero or a new nonzero element will be added into the supportset. The two most important parameters are the step size τand the update direction δ. At the l-th stage (if (XT3X3)

−1

exists), the homotopy algorithm first calculates the updatedirection, which can be obtained by solving

XT3X3δl = u (VII.4)

Thus, the solution of problem VII.4 can be written as

δl =

{(XT3X3)

−1u, on 30, otherwise

(VII.5)

Subsequently, the homotopy algorithm computes the stepsize τ to the next critical point by tracing the homotopysolution path. i.e. the homotopy algorithm moves along theupdate direction until one of constraints in VII.3 is notsatisfied. At this critical point, a new nonzero elementmust enter the support 3, or one of the nonzero elementsin α will be shrink to zero, i.e. this element must be removedfrom the support set 3. Two typical cases may lead to a newcritical point, where either condition of VII.3 is violated. Theminimum step size which leads to a critical point can be easilyobtained by computing τ ∗l = min(τ+l , τ

l ), and τ+l and τ−l arecomputed by

τ+l = mini∈3c

(λ− pi

1− xTi X3δl,

λ+ pi1+ xTi X3δl

)+

(VII.6)

τ−l = mini∈3

(−αil

δil

)+

(VII.7)

where pi = xTi (y − xiαil) and min(·)+ denotes that theminimum is operated over only positive arguments. τ+l isthe minimum step size that turns an inactive element at theindex i+ in to an active element, i.e. the index i+ should beadded into the support set. τ−l is the minimum step size thatshrinks the value of a nonzero active element to zero at theindex i− and the index i− should be removed from the supportset. The solution is updated by αl+1 = αl + τ

∗l δ, and its

support and sign sequence are renewed correspondingly.The homotopy algorithm iteratively computes the step size

and the update direction, and updates the homotopy solutionand its corresponding support and sign sequence till thecondition ‖p‖∞ = 0 is satisfied so that the solution ofproblem III.9 is reached. The principal steps of the LASSOhomotopy algorithm have been summarized in Algorithm 9.For further description and analyses, please refer to theliterature [30], [96].

VOLUME 3, 2015 507

www.redpel.com18+917620593389

www.redpel.com18+917620593389

Page 19: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Algorithm 9 Lasso Homotopy AlgorithmTask: To addrss the Lasso problem:

α = argminα ‖y− Xα‖22 s.t. ‖α‖1 ≤ ε

Input: Probe sample y, measurement matrix X .Initialization: l = 1, initial solution αl and its supportset 3l .Repeat:Step 1: Compute update direction δl by using

Eq. (VII.5).Step 2: Compute τ+l and τ−l by using Eq. (VII.6) and

Eq. (VII.7).Step 3: Compute the optimal minimum step size τ ∗l by

using τ ∗l = min{τ+l , τ−

l }.Step 4: Update the solution αl+1 by using αl+1 = αl

+τ ∗l δl .Step 5: Update the support set:

If τ+l == τ−

l thenRemove the i− from the support set, i.e.3l+1 = 3l\i−.

elseAdd the i+ into the support set, i.e.3l+1 = 3l

⋃i+

End ifStep 6: l = l + 1.

Until ‖XT (y− Xα)‖∞ = 0Output: αl+1

B. BPDN HOMOTOPYProblem III.11, which is called basis pursuitdenoising (BPDN) in signal processing, is the unconstrainedLagrangian function of the LASSO problem III.9, which isan unconstrained problem. The BPDN homotopy algorithmis very similar to the LASSO homotopy algorithm. If weconsider the KKT optimality condition for problem III.12,the following condition should be satisfied for the solution α

‖XT (y− Xα)‖∞ ≤ λ (VII.8)

As for any given value of λ and the support set 3, thefollowing two conditions also need to be satisfied

XT3(y− Xα) = λu; ‖XT3c (y− Xα)‖∞ ≤ λ (VII.9)

The BPDN homotopy algorithm directly computes thehomotopy solution by

α =

{(XT3X3)

−1(XT3y− λu), on 30, otherwise

(VII.10)

which is somewhat similar to the soft-thresholding operator.The value of the homotopy parameter λ is initialized with alarge value, which satisfies λ0 > ‖XT y‖∞. As the value ofthe homotopy parameter λ decreases, the BPDN homotopyalgorithm traces the solution in the direction of (XT3X3)

−1utill the critical point is obtained. Each critical point is reachedwhen either an inactive element is transferred into an activeelement, i.e. its corresponding index should be added into the

support set, or an nonzero active element value in α shrinksto zero, i.e. its corresponding index should be removed fromthe support set. Thus, at each critical point, only one elementis updated, i.e. one element being either removed from oradded into the active set, and each operation is very com-putationally efficient. The algorithm is terminated when thevalue of the homotopy parameter is lower than its desiredvalue. The BPDN homotopy algorithm has been summarizedin Algorithm 10. For further detail description and analyses,please refer to the literature [43].

Algorithm 10 BPDN Homotopy AlgorithmTask: To address the Lasso problem:

α = argminα ‖y− Xα‖22 + λ‖α‖1Input: Probe sample y, measurement matrix X .Initialization: l = 0, initial solution α0 and its supportset 30, a large value λ0, step size τ , tolerance ε.Repeat:Step 1: Compute update direction δl+1 by using

δl+1 = (XT3X3)−1ul .

Step 2: Update the solution αl+1 by using Eq. (VII.10).Step 3: Update the support set and the sign sequence set.Step 6: λl+1 = λl − τ , l = l + 1.

Until λ ≤ εOutput: αl+1

C. ITERATIVE REWEIGHTING l1-NORMMINIMIZATION VIA HOMOTOPYBased on the homotopy algorithm, Asif and Romberg [96]presented a enhanced sparse representation objective func-tion, a weighted l1-norm minimization, and then providedtwo fast and accurate solutions, i.e. the iterative reweight-ing algorithm, which updated the weights with a new ones,and the adaptive reweighting algorithm, which adaptivelyselected the weights in each iteration. Here the iterativereweighting algorithm via homotopy is introduced. Theobjective function of the weighted l1-norm minimization isformulated as

argmin12‖Xα − y‖22 + ‖Wα‖1 (VII.11)

where W = diag[w1,w2, · · · ,wN ] is the weight of thel1-norm and also is a diagonal matrix. For more explicitdescription, problem VII.11 can be rewritten as

argmin12‖Xα − y‖22 +

N∑i=1

wi|αi| (VII.12)

A common method [43], [73] to update the weight W isachieved by exploiting the solution of problem VII.12, i.e. α,at the previous iteration, and for the i-th element of the weightwi is updated by

wi =λ

|αi| + σ(VII.13)

where parameters λ and σ are both small constants. In orderto efficiently update the solution of problem (7-9), the homo-topy algorithm introduces a new weight of the l1-norm and

508 VOLUME 3, 2015

www.redpel.com19+917620593389

www.redpel.com19+917620593389

Page 20: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

a new homotopy based reweighting minimization problem isreformulated as

argmin12‖Xα − y‖22 +

N∑i=1

((1− σ )wi + σ wi)|αi|

(VII.14)

where wi denotes the new obtained weight by thehomotopy algorithm, parameter τ is denoted as the homotopyparameter varying from 0 to 1. Apparently, problem VII.14can be evolved to problem VII.12 with the increasing valueof the homotopy parameter by tracing the homotopy solutionpath. Similar to the LASSO homotopy algorithm,problem VII.14 is also piecewise linear along the homotopypath, and for any value of σ , the following conditions shouldbe satisfied

xTi (Xα − y) = −((1− σ )wi + σ wi)ui for i ∈ 3 (a)

|xTi (Xα − y)| < (1− σ )wi + σ wi for i ∈ 3c (b)

(VII.15)

where xi is the i-th column of the measurement X , wi and wiare the given weight and new obtained weight, respectively.Moreover, for the optimal step size σ , when the homotopyparameter changes from σ to σ + τ in the update direction δ,the following optimality conditions also should be satisfied

XT3(Xα − y)+ τXT3Xδ

= −((1− σ )W + σ W )u+ τ (W − W )u (a)

|p− τq| ≤ r + τ s (b)

(VII.16)

where u is the sign sequence of α on its support 3,pi = xTi (Xα − y), qi = xTi Xδ, ri = (1 − σ )wi + σ wiand si = wi − wi. Thus, at the l-th stage (if (XTi Xi)

−1

exists), the update direction of the homotopy algorithm canbe computed by

δl =

{(XT3X3)

−1(W − W )u, on 30, otherwise

(VII.17)

The step size which can lead to a critical point can be com-puted by τ ∗l = min(τ+l , τ

l ), and τ+l and τ−l are computed by

τ+l = mini∈3c

(ri − piqi − si

,−ri − piqi + si

)+

(VII.18)

τ−l = mini∈3

(−αil

δil

)+

(VII.19)

where τ+l is the minimum step size so that the index i+ shouldbe added into the support set and τ−l is the minimum stepsize that shrinks the value of a nonzero active element tozero at the index i−. The solution and homotopy parameterare updated by αl+1 = αl + τ

∗l δ, and σl+1 = σl + τ

∗l ,

respectively. The homotopy algorithm updates its support setand sign sequence accordingly until the new critical pointof the homotopy parameter σl+1 = 1. The main steps ofthis algorithm are summarized in Algorithm 11 and moreinformation can be found in literature [96].

Algorithm 11 Iterative Reweighting Homotopy Algorithmfor Weighted l1-Norm MinimizationTask: To addrss the weighted l1-norm minimization:

α = argmin 12‖Xα − y‖

22 +W‖α‖1

Input: Probe sample y, measurement matrix X .Initialization: l = 1, initial solution αl and its supportset 3l , σ1 = 0.Repeat:Step 1: Compute update direction δl by using

Eq. (VII.17).Step 2: Compute p, q, r and s by using Eq. (VII.16).Step 2: Compute τ+l and τ−l by using Eq. (VII.18) and

Eq. (VII.19).Step 3: Compute the step size τ ∗l by using

τ ∗l = min{τ+l , τ−

l }.Step 4: Update the solution αl+1 by using αl+1 =

αl + τ∗l δl .

Step 5: Update the support set:If τ+l == τ

l thenShrink the value to zero at the index i− andremove i−, i.e. 3l+1 = 3l\i−.

elseAdd the i+ into the support set, i.e. 3l+1

= 3l⋃i+

End ifStep 6: σl+1 = σl + τl and l = l + 1.

Until σl+1 = 1Output: αl+1

D. OTHER HOMOTOPY ALGORITHMSFOR SPARSE REPRESENTATIONThe general principle of the homotopy method is to reach theoptimal solution along with the homotopy solution path byevolving the homotopy parameter from a known initial valueto the final expected value. There are extensive hotomopyalgorithms, which are related to the sparse representationwiththe l1-norm regularization. Malioutov et al. first exploited thehomotopy method to choose a suitable parameter for l1-normregularization with a noisy term in an underdeterminedsystem and employed the homotopy continuation-basedmethod to solve BPDN for sparse signal processing [97].Garrigues and Ghaoui [98] proposed a modified homotopyalgorithm to solve the Lasso problem with online observa-tions by optimizing the homotopy parameter from the currentsolution to the solution after obtaining the next new datapoint. Efron et al. [43] proposed a basic pursuit denois-ing (BPDN) homotopy algorithm, which shrinked the param-eter to a final value with series of efficient optimization steps.Similar to BPDN homotopy, Asif [99] presented a homotopyalgorithm for the Dantzing selector (DS) under the consider-ation of primal and dual solution. Asif and Romberg [100]proposed a framework of dynamic updating solutions forsolving l1-norm minimization programs based on homotopy

VOLUME 3, 2015 509

www.redpel.com20+917620593389

www.redpel.com20+917620593389

Page 21: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

algorithm and demonstrated its effectiveness in addressingthe decoding issue. More recent literature related to homo-topy algorithms can be found in the streaming recoveryframework [101] and a summary [102].

VIII. THE APPLICATIONS OF THE SPARSEREPRESENTATION METHODSparse representation technique has been successfully imple-mented to numerous applications, especially in the fields ofcomputer vision, image processing, pattern recognition andmachine learning. More specifically, sparse representationhas also been successfully applied to extensive real-worldapplications, such as image denoising, deblurring, inpainting,super-resolution, restoration, quality assessment, classifica-tion, segmentation, signal processing, object tracking,texture classification, image retrieval, bioinformatics,biometrics and other artificial intelligence systems.Moreover, dictionary learning is one of the most typicalrepresentative examples of sparse representation for realizingthe sparse representation of a signal. In this paper, we onlyconcentrate on the three applications of sparse representa-tion, i.e. sparse representation in dictionary learning, imageprocessing, image classification and visual tracking.

A. SPARSE REPRESENTATION IN DICTIONARY LEARNINGThe history of modeling dictionary could be traced backto 1960s, such as the fast Fourier transform (FFT) [103].An over-complete dictionary that can lead to sparse represen-tation is usually achieved by exploiting pre-specified set oftransformation functions, i.e. transform domain method [5],or is devised based on learning, i.e. dictionary learningmethods [104]. Both of the transform domain and dictionarylearning based methods transform image samples into otherdomains and the similarity of transformation coefficients areexploited [105]. The difference between them is that the trans-form domain methods usually utilize a group of fixed trans-formation functions to represent the image samples, whereasthe dictionary learning methods apply sparse representationson a over-complete dictionary with redundant information.Moreover, exploiting the pre-specified transform matrix intransform domain methods is attractive because of its fastand simplicity. Specifically, the transform domain methodsusually represent the image patches by using the orthonor-mal basis such as over-complete wavelets transform [106],super-wavelet transform [107], bandelets [108], curveletstransform [109], contourlets transform [110] and steerablewavelet filters [111]. However, the dictionary learningmethods exploiting sparse representation have the potentialcapabilities of outperforming the pre-determined dictionar-ies based on transformation functions. Thus, in this subsec-tion we only focus on the modern over-complete dictionarylearning methods.

An effective dictionary can lead to excellent reconstruc-tion results and satisfactory applications, and the choice ofdictionary is also significant to the success of sparserepresentation technique. Different tasks have different

dictionary learning rules. For example, imageclassification requires that the dictionary contains discrim-inative information such that the solution of sparse repre-sentation possesses the capability of distinctiveness. Thepurpose of dictionary learning is motivated from sparserepresentation and aims to learn a faithful and effectivedictionary to largely approximate or simulate the specificdata. In this section, some parameters are defined as matrixY = [y1, y2, · · · , yN ], matrix X = [x1, x2, · · · , xN ]T , anddictionary D = [d1, d2, · · · , dM ].

From the notations of the literature [23], [112], theframework of dictionary learning can be generally formulatedas an optimization problem

arg minD∈�,xi

{1N

N∑i=1

(12‖yi − Dxi‖

22 + λP(xi))

}(VIII.1)

where � = {D = [d1, d2, · · · , dM ] : dTi d i = 1,i = 1, 2, · · · ,M} (M here may not be equal to N ), N denotesthe number of the known data set (eg. training samples inimage classification), yi is the i-th sample vector from aknown set, D is the learned dictionary and xi is the sparsityvector. P(xi) and λ are the penalty or regularization termand a tuning parameter, respectively. The regularization termof problem VIII.1 controls the degree of sparsity. That is,different kinds of the regularization terms can immenselydominate the dictionary learning results.

One spontaneous idea of defining the penalty term P(xi)is to introduce the l0-norm regularization, which leads tothe sparsest solution of problem VIII.1. As a result, thetheory of sparse representation can be applied to dictionarylearning. The most representative dictionary learning basedon the l0-norm penalty is the K-SVD algorithm [8], which iswidely used in image denoising. Because the solution ofl0-norm regularization is usually a NP-hard problem, utilizinga convex relaxation strategy to replace l0-norm regularizationis an advisable choice for dictionary learning. As a convexrelaxation method of l0-norm regularization, the l1-norm reg-ularization based dictionary learning has been proposed inlarge numbers of dictionary learning schemes. In the stage ofconvex relaxation methods, there are three optimal forms forupdating a dictionary: the one by one atom updating method,group atoms updating method, and all atoms updatingmethod [112]. Furthermore, because of over-penalization inl1-norm regularization, non-convex relaxation strategies alsohave been employed to address dictionary learning problems.For example, Fan and Li proposed a smoothly clipped abso-lution deviation (SCAD) penalty [113], which employed aniterative approximate Newton-Raphson method for penaliz-ing least sequences and exploited the penalized likelihoodapproaches for variable selection in linear regression models.Zhang introduced and studied the non-convex minimaxconcave (MC) family [114] of non-convex piecewisequadratic penalties to make unbiased variable selection forthe estimation of regression coefficients, which was demon-strated its effectiveness by employing an oracle inequality.

510 VOLUME 3, 2015

www.redpel.com21+917620593389

www.redpel.com21+917620593389

Page 22: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Friedman proposed to use the logarithmic penalty for a modelselection [115] and used it to solve theminimization problemswith non-convex regularization terms. From the viewpoint ofupdating strategy, most of the dictionary learning methodsalways iteratively update the sparse approximation or repre-sentation solution and the dictionary alternatively, and moredictionary learning theoretical results and analyses can befound in the literature [104], [116].

Recently, varieties of dictionary learning methods havebeen proposed and researchers have attempted to exploitdifferent strategies for implementing dictionary learningtasks based on sparse representation. There are several meansto categorize these dictionary learning algorithms into variousgroups. For example, dictionary learning methods can bedivided into three groups in the context of different normsutilized in the penalty term, that is, l0-norm regularizationbased methods, convex relaxation methods and non-convexrelaxation methods [117]. Moreover, dictionary learningalgorithms can also be divided into three other categoriesin the presence of different structures. The first category isdictionary learning under the probabilistic framework suchas maximum likelihood methods [118], the method of opti-mal directions (MOD) [119], and the maximum a posterioriprobability method [120]. The second category is clusteringbased dictionary learning approaches such as KSVD [121],which can be viewed as a generalization of K -means. Thethird category is dictionary learning with certain structures,which are grouped into two significative aspects, i.e. directlymodeling the relationship between each atom and structuringthe corrections between each atom with purposive sparsitypenalty functions. There are two typical models for thesekinds of dictionary learning algorithms, sparse andshift-invariant representation of dictionary learning and struc-ture sparse regularization based dictionary learning, suchas hierarchical sparse dictionary learning [122] and groupor block sparse dictionary learning [123]. Recently, someresearchers [23] categorized the latest methods of dictionarylearning into four groups, online dictionary learning [124],joint dictionary learning [125], discriminative dictionarylearning [126], and supervised dictionary learning [127].

Although there are extensive strategies to divide theavailable sparse representation based dictionary learningmethods into different categories, the strategy used hereis to categorize the current prevailing dictionary learningapproaches into two main classes: supervised dictionarylearning and unsupervised dictionary learning, and then spe-cific representative algorithms are explicitly introduced.

1) UNSUPERVISED DICTIONARY LEARNINGFrom the viewpoint of theoretical basis, the main differenceof unsupervised and supervised dictionary learning relies onwhether the class label is exploited in the process of learningfor obtaining the dictionary. Unsupervised dictionary learn-ing methods have been widely implemented to solve imageprocessing problems, such as image compression, and featurecoding of image representation [128], [129].

a: KSVD FOR UNSUPERVISED DICTIONARY LEARNINGOne of the most representative unsupervised dictionarylearning algorithms is the KSVD method [121], which is amodification or an extension of method of directions (MOD)algorithm. The objective function of KSVD is

argminD,X{‖Y − DX‖2F } s.t. ‖xi‖0 ≤ k, i = 1, 2, · · · ,N

(VIII.2)

where Y ∈ Rd×N is the matrix composed of all the knownexamples, D ∈ Rd×N is the learned dictionary, X ∈ RN×N

is the matrix of coefficients, k is the limit of sparsity andxi denotes the i-th row vector of the matrix X . Problem VIII.2is a joint optimization problem with respect to D and X , andthe natural method is to alternatively optimize the D and Xiteratively.

More specifically, when fixing dictionary D,problem VIII.2 is converted to

argminX‖Y − DX‖2F s.t. ‖xi‖0 ≤ k, i = 1, 2, · · · ,N

(VIII.3)

which is called sparse coding and k is the limit of sparsity.Then, its subproblem is considered as follows:

argminxi‖yi − Dxi‖

22 s.t. ‖xi‖0 ≤ k, i = 1, 2, · · · ,N

where we can iteratively resort to the classical sparse repre-sentation with l0-norm regularization such as MP and OMP,for estimating xi.When fixing X , problem VIII.3 becomes a simple

regression model for obtaining D, that is

D = argminD‖Y − DX‖2F (VIII.4)

where D = YX†= YXT (XXT )−1 and the method is called

MOD. Considering that the computational complexity of theinverse problem in solving problem VIII.4 is O(n3), it isfavorable, for further improvement, to update dictionary Dby fixing the other variables. The strategy of the KSVDalgorithm rewrites the problem VIII.4 into

D = argminD‖Y − DX‖2F = argmin

D‖Y −

N∑j=1

d jxTj ‖2F

= argminD‖(Y −

∑j6=l

d jxTj )− d lxTl ‖

2F (VIII.5)

where xj is the j-th row vector of thematrixX . First the overallrepresentation residual El = Y−

∑j 6=l d jx

Tj is computed, and

then d l and xl are updated. In order to maintain the sparsityof xTl in this step, only the nonzero elements of xTl shouldbe preserved and only the nonzero items of El should bereserved, i.e. EPl , from d lxTl . Then, SVD decomposes EPl intoEPl = U3V T , and then updates dictionary d l . The specificKSVD algorithm for dictionary learning is summarized toAlgorithm 12 and more information can be found in theliterature [121].

VOLUME 3, 2015 511

www.redpel.com22+917620593389

www.redpel.com22+917620593389

Page 23: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Algorithm 12 The K-SVD Algorithm for DictionaryLearningTask: Learning a dictionary D: argminD,X ‖Y − DX‖2Fs.t. ‖xi‖0 ≤ k, i = 1, 2, · · · ,NInput: The matrix composed of given samples Y =

[y1, y2, · · · , ym].Initialization: Set the initial dictionary D to the l2–normunit matrix, i = 1.While not converged doStep 1: For each given example yi, employing the classical

sparse representation with l0-norm regularizationsolve problem VIII.3 for further estimating X i,set l = 1.While l is not equal to k do

Step 2: Compute the overall representationresidual El = Y −

∑j 6=l d jx

Tj .

Step 3: Extract the column items of Elwhich corresponds to the nonzeroelements of xTl and obtain EPl .

Step 4: SVD decomposes EPl into EPl =U3V T .

Step 5: Update d l to the first column of Uand update correspondingcoefficients in xTl by 3(1, 1)times the first column of V .

Step 6: l = l + 1.End While

Step 7: i = i+ 1.End WhileOutput: dictionary D

b: LOCALITY CONSTRAINED LINEAR CODING FORUNSUPERVISED DICTIONARY LEARNINGThe locality constrained linear coding (LLC)algorithm [129] is an efficient local coordinate linear codingmethod, which projects each descriptor into a local constraintsystem to obtain an effective codebook or dictionary. It hasbeen demonstrated that the property of locality is more essen-tial than sparsity, because the locality must lead to sparsitybut not vice-versa, that is, a necessary condition of sparsity islocality, but not the reverse [129].

Assume that Y = [y1, y2, · · · , yN ] ∈ Rd×N is a matrixcomposed of local descriptors extracted from examples andthe objective dictionary D = [d1, d2, · · · , dN ] ∈ Rd×N . Theobjective function of LLC is formulated as

argminxi,D

N∑i=1

‖yi − Dxi‖22 + µ‖b� xi‖22

s.t. 1T xi = 1, i = 1, 2, · · · ,N (VIII.6)

where µ is a small constant as a regularization parameterfor adjusting the weighting decay speed, � is the operatorof the element-wise multiplication, xi is the code for yi,1 ∈ RN×1 is defined as a vector with all elements as 1 andvector b is the locality adaptor, which is, more specifically,

set as

b = exp(dist(yi,D)

σ

)(VIII.7)

where dist(yi,D) = [dist(yi, d1), · · · , dist(yi, dN )] anddist(yi, d j) denotes the distance between yi and d j withdifferent distance metrics, such as Euclidean distance andChebyshev distance. Specifically, the i-th value of vector bis defined as bi = exp

(dist(yi,di)

σ

).

The K -Means clustering algorithm is applied to gener-ate the codebook D, and then the solution of LLC can bededuced as:

xi = (Ci + µ diag2(b))\1 (VIII.8)xi = xi/1T xi (VIII.9)

where the operator a\b denotes a−1b, and Ci = (DT − 1yTi )(DT − 1yTi )

T is the covariance matrix with respect to yi. Thisis called the LLC algorithm. Furthermore, the incrementalcodebook optimization algorithm has also been proposedto obtain a more effective and optimal codebook, and theobjective function is reformulated as

argminxi,D

N∑i=1

‖yi − Dxi‖22 + µ‖b� xi‖22

s.t. 1T xi = 1, ∀i; ‖d j‖22 ≤ 1,∀j (VIII.10)

Actually, the problem VIII.10 is a process of featureextraction and the property of ‘locality’ is achieved byconstructing a local coordinate system by exploiting the localbases for each descriptor, and the local bases in the algo-rithm are simply obtained by using the K nearest neighborsof yi. The incremental codebook optimization algorithm inproblem VIII.10 is a joint optimization problem with respectto D and xi, and it can be solved by iteratively optimizingone when fixing the other alternatively. The main steps ofthe incremental codebook optimization algorithm aresummarized in Algorithm 13 and more information can befound in the literature [129].

c: OTHERUNSUPERVISEDDICTIONARYLEARNINGMETHODSA large number of different unsupervised dictionarylearning methods have been proposed. The KSVD algorithmand LLC algorithm are only two typical unsuperviseddictionary learning algorithms based on sparse representa-tion. Additionally, Jenatton et al. [122] proposed a tree-structured dictionary learning problem, which exploitedtree-structured sparse regularization to model the relation-ship between each atom and defined a proximal operator tosolve the primal-dual problem. Zhou et al. [130] developeda nonparametric Bayesian dictionary learning algorithm,which utilized hierarchical Bayesian to model parameters andemployed the truncated beta-Bernoulli process to learn thedictionary. Ramirez and Sapiro [131] employed minimumdescription length to model an effective framework of sparserepresentation and dictionary learning, and this frameworkcould conveniently incorporate prior information into theprocess of sparse representation and dictionary learning.

512 VOLUME 3, 2015

www.redpel.com23+917620593389

www.redpel.com23+917620593389

Page 24: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Algorithm 13 The Incremental Codebook OptimizationAlgorithmTask: Learning a dictionary D: argminxi,D

∑Ni=1 ‖yi−

Dxi‖22 + µ‖b� xi‖22 s.t. 1T xi = 1, ∀i; ‖d j‖22 ≤ 1,∀j

Input: The matrix composed of given samples Y =

[y1, y2, · · · , yN ].Initialization: i = 1, ε = 0.01, D initialized by K -Meansclustering algorithm.While i is not equal to N doStep 1: Initialize b with 1× N zero vector.Step 2: Update locality constraint parameter b with

bj = exp(−dist(yi, d j)

σ

)for ∀j.

Step 3: Normalize b using the equation b = b−bminbmax−bmin

.Step 4: Exploit the LLC coding algorithm to obtain xi.Step 5: Keep the set of Di, whose corresponding entries

of the code xi are greater than ε, and drop outother elements, i.e. index←{j | abs{xi(j)}>ε}∀jand Di← D(:, index).

Step 6: Update xi exploiting argmax ‖yi − Dixi‖22

s.t. 1T xi = 1.Step 7: Update dictionary D using a classical gradient

descent method with respect to problem VIII.6.Step 8: i = i+ 1.

End WhileOutput: dictionary D

Some other unsupervised dictionary learning algorithmsalso have been validated. Mairal et al. proposed an onlinedictionary learning [132] algorithm based on stochasticapproximations, which treated the dictionary learning prob-lem as the optimization of a smooth convex problem overa convex set and employed an iterative online algorithmat each step to solve the subproblems. Yang and Zhangproposed a sparse variation dictionary learning (SVDL)algorithm [133] for face recognition with a single trainingsample, in which a joint learning framework of adaptiveprojection and a sparse variation dictionary with sparsebases were simultaneously constructed from the galleryimage set to the generic image set. Shi et al. proposeda minimax concave penalty based sparse dictionarylearning (MCPSDL) [112] algorithm, which employed anon-convex relaxation online scheme, i.e. a minimax con-cave penalty, instead of using regular convex relaxationapproaches as approximation of l0-norm penalty in sparserepresentation problem, and designed a coordinate descendalgorithm to optimize it. Bao et al. proposed a dictionarylearning by proximal algorithm (DLPM) [117], which pro-vided an efficient alternating proximal algorithm for solvingthe l0-norm minimization based dictionary learning problemand demonstrated its global convergence property.

2) SUPERVISED DICTIONARY LEARNINGUnsupervised dictionary learning just considers that theexamples can be sparsely represented by the learned

dictionary and leaves out the label information of theexamples. Thus, unsupervised dictionary learning canperform very well in data reconstruction, such as imagedenoising and image compressing, but is not beneficial toperform classification. On the contrary, supervised dictionarylearning embeds the class label into the process of sparserepresentation and dictionary learning so that this leads tothe learned dictionary with discriminative information foreffective classification.

a: DISCRIMINATIVE KSVD FOR DICTIONARY LEARNINGDiscriminative KSVD (DKSVD) [126] was designed to solveimage classification problems. Considering the priorities ofsupervised learning theory in classification, DKSVD incor-porates the dictionary learning with discriminative informa-tion and classifier parameters into the objective function andemploys the KSVD algorithm to obtain the global optimalsolution for all parameters. The objective function of theDKSVD algorithm is formulated as

〈D,C,X〉 = arg minD,C,X

‖Y − DX‖2F + µ‖H − CX‖2F

+ η‖C‖2F s.t. ‖xi‖0 ≤ k (VIII.11)

where Y is the given input samples, D is the learned dictio-nary, X is the coefficient term, H is the matrix composedof label information corresponding to Y , C is the parameterterm for classifier, and η and µ are the weights. With aview to the framework of KSVD, problem VIII.11 can berewritten as

〈D,C,X〉 = arg minD,C,X

(YõH

)−

(DõC

)X‖2F

+ η‖C‖2F s.t. ‖xi‖0 ≤ k (VIII.12)

In consideration of the KSVD algorithm, each columnof the dictionary will be normalized to l2-norm unit vector

and(

DõC

)will also be normalized, and then the penalty

term ‖C‖2F will be dropped out and problem VIII.12 will bereformulated as

〈Z ,X〉 = argminZ ,X‖W − ZX‖2F s.t. ‖xi‖0 ≤ k (VIII.13)

where W =

(YõH

), Z =

(DõC

)and apparently

the formulation VIII.13 is the same as the framework ofKSVD [121] in Eq. VIII.2 and it can be efficiently solvedby the KSVD algorithm.

More specifically, the DKSVD algorithm containstwo main phases: the training phase and classification phase.For the training phase, Y is the matrix composed of thetraining samples and the objective is to learn a discrimina-tive dictionary D and the classifier parameter C . DKSVDupdates Z column by column and for each columnvector zi, DKSVD employs the KSVD algorithmto obtain zi and its corresponding weight. Then, the DKSVDalgorithm normalizes the dictionary D and classifier

VOLUME 3, 2015 513

www.redpel.com24+917620593389

www.redpel.com24+917620593389

Page 25: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

parameter C by

D′ = [d ′1, d′

2, · · · , d′M ] = [

d1‖d1‖

,d2‖d2‖

, · · · ,dM‖dM‖

]

C ′ = [c′1, c′

2, · · · , c′M ] = [

c1‖d1‖

,c2‖d2‖

, · · · ,cM‖dM‖

]

x′i = xi × ‖d i‖ (VIII.14)

For the classification phase, Y is the matrix composedof the test samples. Based on the obtained learning resultsD′ and C ′, the sparse coefficient matrix xi can be obtained foreach test sample yi by exploiting the OMP algorithm, whichis to solve

xi = argmin ‖yi − D′x′i‖

22 s.t. ‖x′i‖0 ≤ k (VIII.15)

On the basis of the corresponding sparse coefficient xi, thefinal classification, for each test sample yi, can be performedby judging the label result by multiplying xi by classifier C ′,that is,

label = C ′ × xi (VIII.16)

where the label is the final predicted label vector. The classlabel of yi is the determined class index of label.The main highlight of DKSVD is that it employs the

framework of KSVD to simultaneously learn a discrimina-tive dictionary and classifier parameter, and then utilizes theefficient OMP algorithm to obtain a sparse representationsolution and finally integrate the sparse solution and learnedclassifier for ultimate effective classification.

b: LABEL CONSISTENT KSVD FOR DISCRIMINATIVEDICTIONARY LEARNINGBecause of the classification term, a competent dictionarycan lead to effectively classification results. The originalsparse representation for face recognition [20] regards theraw data as the dictionary, and then reports its promisingclassification results. In this section, a label consistentKSVD (LC-KSVD) [134], [135] is introduced to learn aneffective discriminative dictionary for image classification.As an extension of D-KSVD, LC-KSVD exploits the super-vised information to learn the dictionary and integrates theprocess of constructing the dictionary and optimal linear clas-sifier into a mixed reconstructive and discriminative objectivefunction, and then jointly obtains the learned dictionary andan effective classifier. The objective function of LC-KSVD isformulated as

〈D,A,C,X〉 = arg minD,A,C,X

‖Y − DX‖2F + µ‖L − AX‖2F

+η‖H − CX‖2F s.t. ‖xi‖0 ≤ k (VIII.17)

where the first term denotes the reconstruction error, thesecond term denotes the discriminative sparse-code error, andthe final term denotes the classification error. Y is the matrixcomposed of all the input data, D is the learned dictionary,X is the sparse code term, µ and η are the weights of thecorresponding contribution items, A is a linear transforma-tion matrix, H is the matrix composed of label informationcorresponding to Y , C is the parameter term for classifier

and L is a joint label matrix for labels of Y and D. Forexample, providing that Y = [y1 . . . y4] and D = [d1 . . . d4]where y1, y2, d1 and d2 are from the first class, and y3, y4, d3and d4 are from the second class, and then the joint label

matrix L can be defined as L =

1 1 0 01 1 0 00 0 1 10 0 1 1

. Similar

to the DKSVD algorithm, the objective function VIII.17 canalso be reformulated as

〈Z ,X〉 = argminZ ,X‖T − ZX‖22 s.t. ‖xi‖0 ≤ k (VIII.18)

where T =

Y√µL√ηH

, Z =

D√µA√ηC

.

The learning process of the LC-KSVD algorithm, as isDKSVD, can be separated into two sections, the training termand the classification term. In the training section, since prob-lemVIII.18 completely satisfies the framework of KSVD, theKSVD algorithm is applied to update Z atom by atom andcompute X . Thus Z and X can be obtained. Then, theLC-KSVD algorithm normalizes dictionary D, transformmatrix A, and the classifier parameter C by

D′ = [d ′1, d′

2, · · · , d′M ] = [

d1‖d1‖

,d2‖d2‖

, · · · ,dM‖dM‖

]

A′ = [a′1, a′

2, · · · , a′M ] = [

a1‖d1‖

,a2‖d2‖

, · · · ,aM‖dM‖

]

C ′ = [c′1, c′

2, · · · , c′M ] = [

c1‖d1‖

,c2‖d2‖

, · · · ,cM‖dM‖

]

(VIII.19)

In the classification section, Y is the matrix composed ofthe test samples. On the basis of the obtained dictionary D′,the sparse coefficient xi can be obtained for each test sam-ple yi by exploiting the OMP algorithm, which is to solve

xi = argmin ‖yi − D′x′i‖

22 s.t. ‖x′i‖0 ≤ k (VIII.20)

The final classification is based on a simple linearpredictive function

l = argmaxf{ f = C ′ × xi} (VIII.21)

where f is the final predicting label vector and the testsample yi is classified as a member of the l-th class.The main contribution of LC-KSVD is to jointly incor-

porate the discriminative sparse coding term and classifierparameter term into the objective function for learning adiscriminative dictionary and classifier parameter. TheLC-KSVDdemonstrates that the obtained solution, comparedto other methods, can prevent learning a suboptimal or localoptimal solution in the process of learning a dictionary [134].

c: FISHER DISCRIMINATION DICTIONARY LEARNINGFOR SPARSE REPRESENTATIONFisher discrimination dictionary learning (FDDL) [136]incorporates the supervised information (class label informa-tion) and the Fisher discrimination message into the objective

514 VOLUME 3, 2015

www.redpel.com25+917620593389

www.redpel.com25+917620593389

Page 26: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

function for learning a structured discriminative dictionary,which is used for pattern classification. The general model ofFDDL is formulated as

J (D,X ) = argminD,X{f (Y ,D,X )+ µ‖X‖1 + ηg(X )}

(VIII.22)

where Y is the matrix composed of input data, D is thelearned dictionary, X is the sparse solution, and µ and η aretwo constants for tradeoff contributions. The first componentis the discriminative fidelity term, the second component isthe sparse regularization term, and the third component is thediscriminative coefficient term, such as Fisher discriminationcriterion in Eq. (VIII.23).

Considering the importance of the supervised information,i.e. label information, in classification, FDDL respectivelyupdates the dictionary and computes the sparse representationsolution class by class. Assume that Y i denotes the matrixof i-th class of input data, vector X i denotes the sparserepresentation coefficient of the learned dictionary D overY i and X ij denotes the matrix composed of the sparserepresentation solutions, which correspond to the j-th classcoefficients from X i. Di is denoted as the learned dictionarycorresponding to the i-th class. Thus, the objective functionof FDDL is

J (D,X ) = argminD,X

(c∑i=1

f (Y i,D,X i)+ µ‖X‖1

+η(tr(SW (X )− SB(X ))+ λ‖X‖2F )) (VIII.23)

where f (Y i,D,X i) = ‖Y i − DX i‖2F + ‖Yi− DiX ii ‖

2F +∑

j 6=i ‖DjX ij ‖

2F and SW (X ) and SB(X ) are the within-class

scatter of X and between-class scatter of X , respectively.c is the number of the classes. To solve problem VIII.23,a natural idea of optimization is to alternatively optimizeD and X class by class, and then the process of optimizationis briefly introduced.

When fixingD, problem VIII.23 can be solved by comput-ing X i class by class, and its sub-problem is formulated as

J (X i) = argminX i

(f (Y i,D,X i)+ µ‖X i‖1 + ηg(X i)

)(VIII.24)

where g(X i) = ‖X i −Mi‖2F −

∑ct=1 ‖Mt −M‖2F + λ‖X

i‖2F

and Mj and M denote the mean matrices corresponding tothe j-th class of X i and X i, respectively. Problem VIII.23 canbe solved by the iterative projection method in theliterature [137].

When fixing α, problem VIII.23 can be rewritten as

J (D) = argminD

(‖Y i − DiX i −∑j 6=i

DjX j‖2F

+ ‖Y i − DiX ii ‖2F +

∑j 6=i

‖DiX ij ‖2F )

(VIII.25)

where X i here denotes the sparse representationof Y over Di. In this section, each column of the learned

dictionary is normalized to a unit vector with l2-norm. Theoptimization of problem VIII.25 computes the dictionaryclass by class and it can be solved by exploiting the algorithmin the literature [138].

The main contribution of the FDDL algorithm lies incombining the Fisher discrimination criterion into theprocess of dictionary learning. The discriminative powercomes from the method of constructing the discriminativedictionary using the function f in problem VIII.22 and simul-taneously formulates the discriminative sparse representationcoefficients by exploiting the function g in problem VIII.22.

d: OTHER SUPERVISED DICTIONARY LEARNINGFOR SPARSE REPRESENTATIONUnlike unsupervised dictionary learning, superviseddictionary learning emphasizes the significance of the classlabel information and incorporates it into the learningprocess to enforce the discrimination of the learneddictionary. Recently, massive supervised dictionarylearning algorithms have been proposed. For example,Yang et al. [138] presented a metaface dictionary learn-ing method, which is motivated by ‘metagenes’ in geneexpression data analysis. Ramirez et al. [139] produced adiscriminative non-parametric dictionary learning (DNDL)framework based on the OMP algorithm for image clas-sification. Kong and Wang [140] introduced a learneddictionary with commonalty and particularity, calledDL-COPAR, which integrated an incoherence penalty terminto the objective function for obtaining the class-specificsub-dictionary. Gao et al. [141] learned a hybrid dictio-nary, i.e. category-specific dictionary and shared dictio-nary, which incorporated a cross-dictionary incoherencepenalty and self-dictionary incoherence penalty into theobjective function for learning a discriminative dictionary.Jafari and Plumbley [142] presented a greedy adaptive dic-tionary learning method, which updated the learned dictio-nary with a minimum sparsity index. Some other superviseddictionary learning methods are also competent in imageclassification, such as supervised dictionary learning in [143].Zhou et al. [144] developed a joint dictionary learningalgorithm for object categorization, which jointly learned acommonly shared dictionary and multiply category-specificdictionaries for correlated object classes and incorporatedthe Fisher discriminant fidelity term into the process ofdictionary learning. Ramirez et al. proposed a method of dic-tionary learning with structured incoherence (DLSI) [139],which unified the dictionary learning and sparse decom-position into a sparse dictionary learning framework forimage classification and data clustering. Ma et al. presented adiscriminative low-rank dictionary learning for sparserepresentation (DLRD_SR) [145], in which the sparsity andthe low-rank properties were integrated into one dictionarylearning scheme where sub-dictionary with discriminativepower was required to be low-rank. Lu et al. developed asimultaneous feature and dictionary learning [146] methodfor face recognition, which jointly learned the feature

VOLUME 3, 2015 515

www.redpel.com26+917620593389

www.redpel.com26+917620593389

Page 27: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

projection matrix for subspace learning and thediscriminative structured dictionary. Yang et al. introduced alatent dictionary learning (LDL) [147] method for sparse rep-resentation based image classification, which simultaneouslylearned a discriminative dictionary and a latent representationmodel based on the correlations between label informationand dictionary atoms. Jiang et al. presented a submodulardictionary learning (SDL) [148] method, which integrated theentropy rate of a randomwalk on a graph and a discriminativeterm into a unified objective function and devised a greedy-based approach to optimize it. Si et al. developed a supportvector guided dictionary learning (SVGDL) [149] method,which constructed a discriminative term by using adaptivelyweighted summation of the squared distances for all pairwiseof the sparse representation solutions.

B. SPARSE REPRESENTATION IN IMAGE PROCESSINGRecently, sparse representation methods have beenextensively applied to numerous real-world applicat-ions [150], [151]. The techniques of sparse representationhave been gradually extended and introduced to imageprocessing, such as super-resolution image processing, imagedenoising and image restoration.

First, the general framework of image processing usingsparse representation especially for image reconstructionshould be introduced:

Step 1: Partition the degraded image into overlappedpatches or blocks.

Step 2: Construct a dictionary, denoted as D, and assumethat the following sparse representation formulation shouldbe satisfied for each patch or block x of the image:

α = argmin ‖α‖p s.t. ‖x − HDα‖22 ≤ ε

where H is a degradation matrix and 0 ≤ p ≤ 1.Step 3: Reconstruct each patch or block by exploiting

x = Dα.Step 4: Put the reconstructed patch to the image at

the corresponding location and average each overlappedpatches to make the reconstructed image more consistent andnatural.

Step 5: Repeat step 1 to 4 several times till a terminationcondition is satisfied.

The following part of this subsection is to explicitly intro-duce some image processing techniques using sparse repre-sentation.

The main task of super-resolution image processing is toextract the high super-resolution image from its low resolu-tion counterpart and this challenging problem has attractedmuch attention. The most representative work was proposedto exploit the sparse representation theory to generate a super-resolution (SRSR) image from a single low-resolution imagein literature [152].

SRSR is mainly performed on two compact learneddictionaries Dl and Dh, which are denoted as dictionar-ies of low-resolution image patches and its correspondinghigh-resolution image patches, respectively. Dl is directly

employed to recover high-resolution images fromdictionary Dh. Let X and Y denote the high-resolution and itscorresponding low-resolution images, respectively. x and yare a high-resolution image patch and its correspondinglow-resolution image patch, respectively. Thus, x = Py andP is the projection matrix. Moreover, if the low resolutionimage Y is produced by down-sampling and blurring fromthe high resolution image X , the following reconstructionconstraint should be satisfied

Y = SBX (VIII.26)

where S and B are a downsampling operator and a blurringfilter, respectively. However, the solution of problem VIII.26is ill-posed because infinite solutions can be achieved fora given low-resolution input image Y . To this end,SRSR [152] provides a prior knowledge assumption, whichis formulated as

x = Dhα s.t. ‖α‖0 ≤ k (VIII.27)

where k is a small constant. This assumption gives a priorknowledge condition that any image patch x can beapproximately represented by a linear combination ofa few training samples from dictionary Dh. As presentedin Subsection III-B, problem VIII.27 is an NP-hard problemand sparse representation with l1-norm regularization is intro-duced. If the desired representation solution α is sufficientlysparse, problem VIII.27 can be converted into the followingproblem:

argmin ‖α‖1 s.t. ‖x− Dhα‖22 ≤ ε (VIII.28)

or

argmin ‖x− Dhα‖22 + λ‖α‖1 (VIII.29)

where ε is a small constant and λ is the Lagrange mul-tiplier. The solution of problem VIII.28 can be achievedby two main phases, i.e. local model based sparse rep-resentation (LMBSR) and enhanced global reconstructionconstraint. The first phase of SRSR, i.e. LMBSR, is operatedon each image patch, and for each low-resolution imagepatch y, the following equation is satisfied

argmin ‖Fy− FDlα‖22 + λ‖α‖1 (VIII.30)

where F is a feature extraction operator. One-pass algorithmsimilar to that of [153] is introduced to enhance thecompatibility between adjacent patches. Furthermore, a mod-ified optimization problem is proposed to guarantee that thesuper-resolution reconstruction coincides with the previouslyobtained adjacent high-resolution patches, and the problem isreformulated as

argmin ‖α‖1 s.t. ‖Fy− FDlα‖22≤ε1; ‖v− LDhα‖22≤ε2

(VIII.31)

where v is the previously obtained high-resolution image onthe overlap region, and L refers to the region of overlap

516 VOLUME 3, 2015

www.redpel.com27+917620593389

www.redpel.com27+917620593389

Page 28: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

between the current patch and previously obtained high-resolution image. Thus problem VIII.31 can be rewritten as

argmin ‖y− Dα‖22 + λ‖α‖1 (VIII.32)

where y =[Fyv

]and D =

[FDlLDh

]. Problem VIII.32 can be

simply solved by previously introduced solution of the sparserepresentation with l1-norm minimization. Assume that theoptimal solution of problem VIII.32, i.e. α∗, is achieved, thehigh-resolution patch can be easily reconstructed byx = Dhα∗.

The second phase of SRSR enforces the global recon-struction constraint to eliminate possible unconformity ornoise from the first phase and make the obtained image moreconsistent and compatible. Suppose that the high-resolutionimage obtained by the first phase is denoted as matrix X0,we project X0 onto the solution space of the reconstructionconstraint VIII.26 and the problem is formulatedas follows

X∗ = argmin ‖X − X0‖22 s.t. Y = SBX (VIII.33)

Problem VIII.33 can be solved by the back-projectionmethod in [154] and the obtained image X∗ is regardedas the final optimal high-resolution image. The entiresuper-resolution via sparse representation is summarized inAlgorithm 14 and more information can be found in theliterature [152].

Algorithm 14 Super-Resolution via Sparse RepresentationInput: Training image patches dictionaries Dl and Dh,a low-resolution image Y .For each overlapped 3 × 3 patches y of Y using one-passalgorithm, from left to right and top to bottomStep 1: Compute optimal sparse representation

coefficients α∗ in problem (VIII.32).Step 2: Compute the high-resolution patch by x = Dhα∗.Step 3: Put the patch x into a high-resolution image X0

in corresponding location.EndStep 4: Compute the final super-resolution image X∗ in

problem (VIII.33).Output: X∗

Furthermore, extensive other methods based on sparserepresentation have been proposed to solve the super-resolution image processing problem. For example,Yang et al. presented a modified version called joint dic-tionary learning via sparse representation (JDLSR) [155],which jointly learned two dictionaries that enforced thesimilarity of sparse representation for low-resolution andhigh-resolution images. Tang et al. [156] first explicitlyanalyzed the rationales of the sparse representation theoryin performing the super-resolution task, and proposed toexploit the L2-Boosting strategy to learn coupled dictionaries,

which were employed to construct sparse coding space.Zhang et al. [157] presented an image super-resolutionreconstruction scheme by employing the dual-dictionarylearning and sparse representation method for image super-resolution reconstruction and Gao et al. [158] proposeda sparse neighbor embedding method, which incorporatedthe sparse neighbor search and HoG clustering methodinto the process of image super-resolution reconstruction.Fernandez-Granda and Candès [159] designed a transform-invariant group sparse regularizer by implementing adata-driven non-parametric regularizers with learned domaintransform on group sparse representation for high imagesuper-resolution. Lu et al. [160] proposed a geometry con-strained sparse representation method for single image super-resolution by jointly obtaining an optimal sparse solutionand learning a discriminative and reconstructive dictionary.Dong et al. [161] proposed to harness an adaptive sparseoptimization with nonlocal regularization based on adaptiveprincipal component analysis enhanced by nonlocal similarpatch grouping and nonlocal self-similarity quadraticconstraint to solve the image high super-resolution problem.Dong et al. [162] proposed to integrate an adaptive sparsedomain selection and an adaptive regularization based onpiecewise autoregressive models into the sparse represen-tations framework for single image super-resolution recon-struction. Mallat and Yu [163] proposed a sparse mixingestimator for image super-resolution, which introduced anadaptive estimator models by combining a group of linearinverse estimators based on different prior knowledge forsparse representation.

Noise in an image is unavoidable in the process of imageacquisition. The need for sparse representation may arisewhen noise exists in image data. In such a case, the imagewith noise may lead to missing information or distortion suchthat this results in a decrease of the precision and accuracy ofimage processing. Eliminating such noise is greatly beneficialto many applications. The main goal of image denoising isto distinguish the actual signal and noise signal so that wecan remove the noise and reconstruct the genuine image.In the presence of image sparsity and redundancy representa-tion [4], [7], sparse representation for image denoising firstextracts the sparse image components, which are regardedas useful information, and then abandons the representa-tion residual, which is treated as the image noise term, andfinally reconstructs the image exploiting the pre-obtainedsparse components, i.e. noise-free image. Extensive researcharticles for image denoising based on sparse representationhave been published. For example, Bruckstein et al. [8],Donoho and Tsaig [30], and Donoho [165] first discoveredthe connection between the compressed sensing and imagedenoising. Subsequently, the most representative work ofusing sparse representation to make image denoising wasproposed in literature [165], in which a global sparse repre-sentationmodel over learned dictionaries (SRMLD)was usedfor image denoising. The following prior assumption shouldbe satisfied: every image block of image x, denoted as z, can

VOLUME 3, 2015 517

www.redpel.com28+917620593389

www.redpel.com28+917620593389

Page 29: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

be sparsely represented over a dictionary D, i.e. the solutionof the following problem is sufficiently sparse:

argminα‖α‖0 s.t. Dα = z (VIII.34)

And an equivalent problem can be reformulated for a propervalue of λ, i.e.

argminα‖Dα − z‖22 + λ‖α‖0 (VIII.35)

If we take the above prior knowledge into full consideration,the objective function of SRMLD based on Bayesiantreatment is formulated as

arg minD,αi,x

δ‖x− y‖22 +M∑i=1

‖Dαi − Pix‖22 +M∑i=1

λi‖αi‖0

(VIII.36)

where x is the finally denoised image, y the measuredimage with white and additive Gaussian white noise, Pi isa projection operator that extracts the i-th block from image x,M is the number of the overlapping blocks, D is the learneddictionary, αi is the coefficients vector, δ is the weight of thefirst term and λi is the Lagrange multiplier. The first termin VIII.36 is the log-likelihood global constraint such thatthe obtained noise-free image x is sufficiently similar to theoriginal image y. The second and third terms are the priorknowledge of the Bayesian treatment, which is presentedin problem VIII.35. The optimization of problem VIII.35 isa joint optimization problem with respect to D, αi and x.It can be solved by alternatively optimizing one variablewhen fixing the others. The process of optimization is brieflyintroduced below.

When dictionary D and the solution of sparserepresentation αi are fixed, problem VIII.36 can berewritten as

argminxδ‖x− y‖22 +

M∑i=1

‖Dαi − z‖22 (VIII.37)

where z = Pix. Apparently, problem VIII.37 is a simpleconvex optimization problem and has a closed-form solution,which is given by

x =

(M∑i=1

PTi Pi + δI

)−1 ( M∑i=1

PTi Dαi + δy

)−1(VIII.38)

When x is given, problem (VIII.36) can be written as

argminD,αi

M∑i=1

‖Dαi − Pix‖22 +M∑i=1

λi‖αi‖0 (VIII.39)

where the problem can be divided into M sub-problems andthe i-th sub-problem can be reformulated as the followingdictionary learning problem:

argminD,αi‖Dαi − z‖22 s.t. ‖αi‖0 ≤ τ (VIII.40)

where z = Pix and τ is small constant. One can seethat the sub-problem VIII.39 is the same as problem VIII.2

and it can be solved by the KSVD algorithm previouslypresented in Subsection VIII-A2. The algorithmof image denoising exploiting sparse and redundantrepresentation over learned dictionary is summarizedin Algorithm 15, and more information can be found inliterature [165].

Algorithm 15 Image Denoising via Sparse and RedundantRepresentation Over Learned DictionaryTask: To denoise a measured image y from white andadditional Gaussian white noise:argminD,αi,x δ‖x−y‖

22+∑M

i=1 ‖Dαi−Pix‖22+∑M

i=1 λi‖αi‖0

Input: Measured image sample y, the number of trainingiteration T .Initialization: t = 1, set x = y, D initialized by anovercomplete DCT dictionary.While t ≤ T doStep 1: For each image patch Pix, employ the KSVD

algorithm to update the values of sparse representationsolution αi and corresponding dictionary D.Step 2: t = t + 1

End WhileStep 3: Compute the value of x by using Eq. (VIII.38).

Output: denoised image x

Moreover, extensive modified sparse representation basedimage denoising algorithms have been proposed. Forexample, Dabov et al. [166] proposed an enhanced sparse rep-resentation with a block-matching 3-D (BM3D) transform-domain filter based on self-similarities and an enhancedsparse representation by clustering similar 2-D image patchesinto 3-D data spaces and an iterative collaborative filter-ing procedure for image denoising. Mariral et al. [167]proposed the use of extending the KSVD-based grayscalealgorithm and a generalized weighted average algorithm forcolor image denoising. Protter and Elad [168] extended thetechniques of sparse and redundant representations for imagesequence denoising by exploiting spatio-temporal atoms,dictionary propagation over time and dictionary learning.Dong et al. [169] designed a clustering based sparse represen-tation algorithm, which was formulated by a double-headersparse optimization problem built upon dictionary learningand structural clustering. Recently, Jiang et al. [170] proposeda variational encoding framework with a weighted sparsenonlocal constraint, which was constructed by integratingimage sparsity prior and nonlocal self-similarity prior intoa unified regularization term to overcome the mixed noiseremoval problem. Gu et al. [171] studied a weighted nuclearnorm minimization (WNNM) method with F-norm fidelityunder different weighting rules optimized by non-local self-similarity for image denoising. Ji et al. [172] proposed apatch-based video denoising algorithm by stacking similarpatches in both spatial and temporal domain to formulate alow-rank matrix problem with the nuclear norm.Cheng et al. [173] proposed an impressive image denoising

518 VOLUME 3, 2015

www.redpel.com29+917620593389

www.redpel.com29+917620593389

Page 30: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

method based on an extension of the KSVD algorithm viagroup sparse representation.

The primary purpose of image restoration is to recoverthe original image from the degraded or blurred image.The sparse representation theory has been extensivelyapplied to image restoration. For example,Bioucas-Dias and Figueirdo [174] introduced a two-stepiterative shrinkage/thresholding (TwIST) algorithm for imagerestoration, which is more efficient and can be viewed as anextension of the IST method. Mairal et al. [175] presented amultiscale sparse image representation framework based onthe KSVD dictionary learning algorithm and shift-invariantsparsity prior knowledge for restoration of color images andvideo image sequence. Recently, Mairal et al. [176] pro-posed a learned simultaneous sparse coding (LSSC) model,which integrated sparse dictionary learning and nonlocalself-similarities of natural images into a unified frameworkfor image restoration. Zoran and Weiss [177] proposed anexpected patch log likelihood (EPLL) optimization model,which restored the image from patch to the whole imagebased on the learned prior knowledge of any patch acquiredbyMaximumA-Posteriori estimation instead of using simplepatch averaging. Bao et al. [178] proposed a fast orthogonaldictionary learning algorithm, in which a sparse image repre-sentation based orthogonal dictionary was learned in imagerestoration. Zhang et al. [179] proposed a group-based sparserepresentation, which combined characteristics from localsparsity and nonlocal self-similarity of natural images to thedomain of the group. Dong et al. [180], [181] proposed a cen-tralized sparse representation (CSR) model, which combinedthe local and nonlocal sparsity and redundancy properties forvariational problem optimization by introducing a concept ofsparse coding noise term.

Here we mainly introduce a recently proposed simple buteffective image restoration algorithm CSR model [180]. Fora degraded image y, the problem of image restoration can beformulated as

y = Hx+ v (VIII.41)

where H is a degradation operator, x is the originalhigh-quality image and v is the Gaussian white noise.Suppose that the following two sparse optimization problemsare satisfied

αx = argmin ‖α‖1 s.t. ‖x− Dα‖22 ≤ ε (VIII.42)

αy = argmin ‖α‖1 s.t. ‖x− HDα‖22 ≤ ε (VIII.43)

where y and x respectively denote the degraded imageand original high-quality image, and ε is a small constant.A new concept called sparse coding noise (SCN) is defined

vα = αy − αx (VIII.44)

Given a dictionary D, minimizing SCN can make the imagebetter reconstructed and improve the quality of the imagerestoration because x∗ = x − x = Dαy − Dαx = Dvα .

Thus, the objective function is reformulated as

αy = argminα‖y− HDα‖22 + λ‖α‖1 + µ‖α − αx‖1

(VIII.45)

where λ and µ are both constants. However, the value ofαx is difficult to directly evaluate. Because many nonlocalsimilar patches are associated with the given image patch i,clustering these patches via block matching is advisable andthe sparse code of searching similar patch l to patch i incluster �i, denoted by αil , can be computed. Moreover, theunbiased estimation of αx , denoted by E[αx], empirically canbe approximate to αx under some prior knowledge [180], andthen SCN algorithm employs the nonlocal means estimationmethod [182] to evaluate the unbiased estimation of αx , thatis, using theweighted average of allαil to approachE[αx], i.e.

θi =∑l∈�i

wilαil (VIII.46)

where wil = exp(−‖xi − xil‖22/h

)/N , xi = Dαi, xil = Dαil ,

N is a normalization parameter and h is a constant. Thus, theobjective function VIII.45 can be rewritten as

αy = argminα‖y− HDα‖22 + λ‖α‖1 + µ

M∑i=1

‖αi − θi‖1

(VIII.47)

where M is the number of the separated patches. In thej-th iteration, the solution of problem VIII.47 is iterativelyperformed by

αj+1y = argminα‖y− HDα‖22 + λ‖α‖1 + µ

M∑i=1

‖αi − θji ‖1

(VIII.48)

It is obvious that problem VIII.47 can be optimized by theaugmented Lagrange multiplier method [183] or the iterativeshrinkage algorithm in [184]. According to the maximumaverage posterior principle and the distribution of the sparsecoefficients, the regularization parameter λ and constant µcan be adaptively determined by λ =

2√2ρ2σi

and

µ =2√2ρ2ηi

, where ρ, σi and ηi are the standard deviations ofthe additive Gaussian noise, αi and the SCN signal, respec-tively. Moreover, in the process of image patches clusteringfor each given image patch, a local PCA dictionary is learnedand employed to code each patch within its correspondingcluster. The main procedures of the CSR algorithm are sum-marized in Algorithm 16 and readers may refer to litera-ture [180] for more details.

C. SPARSE REPRESENTATION IN IMAGE CLASSIFICATIONAND VISUAL TRACKINGIn addition to these effective applications in image process-ing, several other fields for sparse representation have beenproposed and extensively studied in image classification andvisual tracking. Since Wright et al. [20] proposed to employ

VOLUME 3, 2015 519

www.redpel.com30+917620593389

www.redpel.com30+917620593389

Page 31: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

Algorithm 16 Centralized Sparse Representation for ImageRestoration

Initialization: Set x = y, initialize regularization param-eter λ and µ, the number of training iteration T , t = 0,θ0 = 0.Step 1: Partition the degraded image into M overlappedpatches.While t ≤ T doStep 2: For each image patch, update the correspondingdictionary for each cluster via k-means and PCA.Step 3: Update the regularization parameters λ and µ byusing

λ =2√2ρ2

σtand µ =

2√2ρ2

ηt.

Step 4: Compute the nonlocal means estimation of theunbiased estimation of αx , i.e. θ

t+1i , by using Eq. (VIII.46)

for each image patch.Step 5: For a given θ t+1i , compute the sparse representa-tion solution, i.e. αt+1y , in problem (VIII.48) by using theextended iterative shrinkage algorithm in literature [184].Step 6: t = t + 1End WhileOutput: Restored image x = Dαt+1y

sparse representation to perform robust face recognition,more and more researchers have been applying the sparserepresentation theory to the fields of computer vision andpattern recognition, especially in image classification andobject tracking. Experimental results have suggested thatthe sparse representation based classification method cansomewhat overcome the challenging issues from illuminationchanges, random pixel corruption, large block occlusion ordisguise.

As face recognition is a representative component ofpattern recognition and computer vision applications, theapplications of sparse representation in face recognition cansufficiently reveal the potential nature of sparse representa-tion. The most representative sparse representation for facerecognition has been presented in literature [18] and thegeneral scheme of sparse representation based classificationmethod is summarized in Algorithm 17. Suppose that thereare n training samples, X = [x1, x2, · · · , xn] from c classes.Let Xi denote the samples from the i-th class and the testingsample is y.Numerous sparse representation based classification

methods have been proposed to improve the robustness,effectiveness and efficiency of face recognition. For example,Xu et al. [9] proposed a two-phase sparse representationbased classification method, which exploited the l2-normregularization rather than the l1-norm regularization toperform a coarse to fine sparse representation based clas-sification, which was very efficient in comparison withthe conventional l1-norm regularization based sparse repre-sentation. Deng et al. [185] proposed an extended sparserepresentation method (ESRM) for improving the robustness

Algorithm 17 The Scheme of Sparse Representation BasedClassification MethodStep 1: Normalize all the samples to have unit l2-norm.Step 2: Exploit the linear combination of all the trainingsamples to represent the test sample and the followingl1-norm minimization problem is satisfied

α∗ = argmin ‖α‖1 s.t. ‖y− Xα‖22 ≤ ε.Step 3: Compute the representation residual for each class

ri = ‖y− Xiα∗i ‖22

where α∗i here denotes the representation coefficients vec-tor associated with the i-th class.Step 4: Output the identity of the test sample y by judging

label(y) = argmini(ri).

of SRC by eliminating the variations in face recognition,such as disguise, occlusion, expression and illumination.Deng et al. [186] also proposed a framework of superposedsparse representation based classification, which emphasizedthe prototype and variation components from uncontrolledimages. He et al. [187] proposed utilizing the maximumcorrentropy criterion named CESR embedding non-negativeconstraint and half-quadratic optimization to present a robustface recognition algorithm. Yang et al. [188] developeda new robust sparse coding (RSC) algorithm, which firstobtained a sparsity-constrained regression model based onmaximum likelihood estimation and exploited an iterativelyreweighted regularized robust coding algorithm to solve thepre-proposed model. Some other sparse representation basedimage classification methods also have been developed. Forexample, Yang et al. [189] introduced an extension of thespatial pyramid matching (SPM) algorithm called ScSPM,which incorporated SIFT sparse representation into thespatial pyramid matching algorithm. Subsequently,Gao et al. [190] developed a kernel sparse representationwith the SPM algorithm called KSRSPM, and then proposedanother version of an improvement of the SPM calledLScSPM [191], which integrated the Laplacian matrix withlocal features into the objective function of the sparserepresentation method. Kulkarni and Li [192] proposed adiscriminative affine sparse codes method (DASC) on alearned affine-invariant feature dictionary from input imagesand exploited theAdaBoost-based classifier to perform imageclassification. Zhang et al. [193] proposed integrating thenon-negative sparse coding, low-rank and sparse matrixdecomposition (LR-Sc+SPM) method, which exploited non-negative sparse coding and SPM for achieving local featuresrepresentation and employed low-rank and sparse matrixdecomposition for sparse representation, for image classifica-tion. Recently, Zhang et al. [194] presented a low-rank sparserepresentation (LRSR) learning method, which preservedthe sparsity and spatial consistency in each procedure offeature representation and jointly exploited local featuresfrom the same spatial proximal regions for imageclassification. Zhang et al. [195] developed a structuredlow-rank sparse representation (SLRSR) method for image

520 VOLUME 3, 2015

www.redpel.com31+917620593389

www.redpel.com31+917620593389

Page 32: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

classification, which constructed a discriminative dictionaryin training terms and exploited low-rank matrix recon-struction for obtaining discriminative representations.Tao et al. [196] proposed a novel dimension reductionmethodbased on the framework of rank preserving sparse learning,and then exploited the projected samples to make effectiveKinect-based scene classification. Zhang et al. [197]proposed a discriminative tensor sparse coding (RTSC)method for robust image classification. Recently, low-rankbased sparse representation became a popular topic such asnon-negative low-rank and sparse graph [198]. Some sparserepresentation methods in face recognition can be found in areview [83] and other more image classification methods canbe found in a more recent review [199].

Mei et al. employed the idea of sparse representation tovisual tracking [200] and vehicle classification [201], whichintroduced nonnegative sparse constraints and dynamic tem-plate updating strategy. It, in the context of the particle filterframework, exploited the sparse technique to guarantee thateach target candidate could be sparsely represented using thelinear combinations of fewest targets and particle templates.It also demonstrated that sparse representation can be prop-agated to address object tracking problems. Extensive sparserepresentation methods have been proposed to address thevisual tracking problem. In order to design an acceleratedalgorithm for l1 tracker, Li et al. [202] proposed two real-time compressive sensing visual tracking algorithms basedon sparse representation, which adopted dimension reductionand the OMP algorithm to improve the efficiency of recoveryprocedure in tracking, and also developed a modified versionof fusing background templates into the tracking procedurefor robust object tracking. Zhang et al. [203] directly treatedobject tracking as a pattern recognition problem by regardingall the targets as training samples, and then employed thesparse representation classification method to do effectiveobject tracking. Zhang et al. [204] employed the conceptof sparse representation based on a particle filter frameworkto construct a multi-task sparse learning method denoted asmulti-task tracking for robust visual tracking. Additionally,because of the discriminative sparse representation betweenthe target and the background, Jia et al. [205] conceived astructural local sparse appearance model for robust objecttracking by integrating the partial and spatial informationfrom the target based on an alignment-pooling algorithm.Liu et al. [206] proposed constructing a two-stage sparseoptimization based online visual tracking method, whichjointly minimized the objective reconstruction error andmaximized the discriminative capability by choosing dis-tinguishable features. Liu et al. [207] introduced a localsparse appearance model (SPT) with a static sparse dictio-nary learned from k-selection and dynamic updated basisdistribution to eliminate potential drifting problems in theprocess of visual tracking. Bao et al. [208] developed a fastreal time l1-tracker called theAPG-l1 tracker, which exploitedthe accelerated proximal gradient algorithm to improve thel1-tracker solver in [200]. Zhong et al. [209] addressed the

object tracking problem by developing a sparsity-basedcollaborative model, which combined a sparsity-based clas-sifier learned from holistic templates and a sparsity-based template model generated from local representations.Zhang et al. [210] proposed to formulate a sparse featuremeasurement matrix based on an appearance model byexploiting non-adaptive random projections, and employeda coarse-to-fine strategy to accelerate the computational effi-ciency of tracking task. Lu et al. [211] proposed to employboth non-local self-similarity and sparse representation todevelop a non-local self-similarity regularized sparse repre-sentation method based on geometrical structure informationof the target template data set. Wang et al. [212] proposeda sparse representation based online two-stage trackingalgorithm, which learned a linear classifier based on localsparse representation on favorable image patches. Moredetailed visual tracking algorithms can be found in the recentreviews [213], [214].

IX. EXPERIMENTAL EVALUATIONIn this section, we take the object categorization problem asan example to evaluate the performance of different sparserepresentation based classification methods. We analyze andcompare the performance of sparse representation with themost typical algorithms: OMP [37], l1_ls [76], PALM [89],FISTA [82], DALM [89], homotopy [99] and TPTSR [9].

Plenties of data sets have been collected for object cate-gorization, especially for image classification. Several imagedata sets are used in our experimental evaluations.ORL: The ORL database includes 400 face images taken

from 40 subjects each providing 10 face images [215]. Forsome subjects, the images were taken at different times,with varying lighting, facial expressions, and facial details.All the images were taken against a dark homogeneousbackground with the subjects in an upright, frontal position(with tolerance for some side movement). Each image wasresized to a 56×46 image matrix by using the down-samplingalgorithm.LFW Face Dataset: The Labeled Faces in the Wild (LFW)

face database is designed for the study of uncon-strained identity verification and face recognition [216].It contains more than 13,000 images of faces collected fromthe web under the unconstrained conditions. Each face hasbeen labeled with the name of the people pictured. 1680 ofthe people pictured have two or more distinct photos in thedatabase. In our experiments, we chose 1251 images from86 peoples and each subject has 10-20 images [217].Each image was manually cropped and was resized to32×32 pixels.Extended YaleB Face Dataset: The extended YaleB

database contains 2432 front face images of 38 individualsand each subject having around 64 near frontal images underdifferent illuminations [218]. The main challenge of thisdatabase is to overcome varying illumination conditions andexpressions. The facial portion of each original image wascropped to a 192×168 image. All images in this data set for

VOLUME 3, 2015 521

www.redpel.com32+917620593389

www.redpel.com32+917620593389

Page 33: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

FIGURE 5. Classification accuracies of using different sparse representation based classification methods versus varying values of theregularization parameter λ on the (a) ORL (b) LFW (c) Coil20 and (d) Fifteen scene datasets.

our experiments simply resized these face images to32×32 pixels.COIL20 Dataset: Columbia Object Image Library

(COIL-20) database consists of 1,440 size normalized gray-scale images of 20 objects [219]. Different object imagesare captured at every angle in a 360 rotation. Images of theobjects were taken from varying angles at pose intervals offive degrees and each object has 72 images.Fifteen Scene Dataset: This dataset contains

4485 images under 15 natural scene categories pre-sented in literature [220] and each category includes210 to 410 images. The 15 scenes categories are office,kitchen, living room, bedroom, store, industrial, tall building,inside cite, street, highway, coast, open country, mountain,forest and suburb. A wide range of outdoor and indoor scenesare included in this dataset. The average image size is around250 × 300 pixels and the spatial pyramid matching featuresare used in our experiments.

A. PARAMETER SELECTIONParameter selection, especially selection of the regularizationparameter λ in different minimization problems, plays animportant role in sparse representation. In order to makefair comparisons with different sparse representationalgorithms, performing the optimal parameter selectionfor different sparse representation algorithms on differentdatasets is advisable and indispensable. In this subsection,we perform extensive experiments for selecting the bestvalue of the regularization parameter λ with a wide range ofoptions. Specifically, we implement the l1_ls, FISTA,DALM,homotopy and TPTSR algorithms on different databases toanalyze the importance of the regularization parameter. Fig. 5summarizes the classification accuracies of exploiting differ-ent sparse representation based classification methods withvarying values of regularization parameter λ on the two facedatasets, i.e. ORL and LFW face datasets, and two objectdatasets, i.e. COIL20 and Fifteen scene datasets. On the ORL

522 VOLUME 3, 2015

www.redpel.com33+917620593389

www.redpel.com33+917620593389

Page 34: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

TABLE 1. Classification accuracies (mean classification error rates ± standard deviation %) of different sparse representation algorithms with differentnumbers of training samples. The bold numbers are the lowest error rates and the least time cost of different algorithms.

and LFW face datasets, we respectively selected the first fiveand eight face images of each subject as training samples andthe rest of image samples for testing. As for the experimentson the COIL20 and fifteen scene datasets, we respectivelytreated the first ten images of each subject in both datasetsas training samples and used all the remaining images as testsamples. Moreover, from Fig. 5, one can see that the valueof regularization parameter λ can significantly dominate theclassification results, and the values of λ for achieving thebest classification results on different datasets are distinctlydifferent. An interesting scenario is that the performanceof the TPTSR algorithm is almost not influenced by thevariation of regularization parameter λ in the experimentson fifteen scene dataset, as shown in Fig. 5(d). However, thebest classification accuracy can be always obtained withinthe range of 0.0001 to 1. Thus, the value of the regularizationparameter is set within the range from 0.0001 to 1.

B. EXPERIMENTAL RESULTSIn order to test the performance of different kinds ofsparse representation methods, an empirical study of exper-imental results is conducted in this subsection and seventypical sparse representation based classification methods areselected for performance evaluation followed with extensiveexperimental results. For all datasets, following most previ-ous published work, we randomly choose several samples ofevery class as training samples and used the rest as test

samples and the experiments are repeated 10 times withthe optimal parameter obtained using the cross validationapproach. The gray-level features of all images in these datasets are used to perform classification. For the sake of com-putational efficiency, principle component analysis algorithmis used as a preprocessing step to preserve 98% energy ofall the data sets. The classification results and computationaltime have been summarized in Table 1. From the experi-mental results on different databases, we can conclude thatthere still does not exist one extraordinary algorithm thatcan achieve the best classification accuracy on all databases.However, some algorithms are noteworthy to be paid muchmore attention. For example, the l1_ls algorithm in mostcases can achieve better classification results than the otheralgorithms on the ORL database, and when the number oftraining samples of each class is five, the l1_ls algorithm canobtain the highest classification result of 95.90%. The TPTSRalgorithm is very computationally efficient in comparisonwith other sparse representation with l1-norm minimizationalgorithms and the classification accuracies obtained by theTPTSR algorithm are very similar and sometimes evenbetter than the other sparse representation based classificationalgorithms.

The computational time is another indicator for measuringthe performance of one specific algorithm. As shownin Table 1, the average computational time of each algorithmis shown at the bottom of the table for one specific number of

VOLUME 3, 2015 523

www.redpel.com34+917620593389

www.redpel.com34+917620593389

Page 35: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

training samples. Note that the computational time of OMPand TPTSR algorithms are drastically lower than that of othersparse representation with l1-norm minimization algorithms.This is mainly because the sparse representation withl1-norm minimization algorithms always iteratively solvethe l1-norm minimization problem. However, the OMP andTPTSR algorithms both exploit the fast and efficient leastsquares technique, which guarantees that the computationaltime is significantly less than other l1-norm based sparserepresentation algorithms.

C. DISCUSSIONLots of sparse representation methods have been available inpast decades and this paper introduces various sparserepresentation methods from some viewpoints, includingtheir motivations, mathematical representations and the mainalgorithms. Based on the experimental results summarizedin Section IX, we have the following observations.

First, a challenging task of choosing a suitable regulariza-tion parameter for sparse representation should make furtherextensive studies. We can see that the value of the regulariza-tion parameter can remarkably influence the performance ofthe sparse representation algorithms and adjusting the param-eters in sparse representation algorithms requires expensivelabor. Moreover, adaptive parameter selection based sparserepresentation methods is preferable and very few methodshave been proposed to solve this critical issue.

Second, although sparse representation algorithms haveachieved distinctly promising performance on somereal-world databases, many efforts should be made inpromoting the accuracy of sparse representation based clas-sification, and the robustness of sparse representation shouldbe further enhanced. In terms of the recognition accuracy, thealgorithms of l1_ls, homotopy and TPTSR achieve the bestoverall performance. Considering the experimental results ofexploiting the seven algorithms on the five databases, thel1_ls algorithm has eight highest classification accuracies,followed by homotopy and TPTSR, in comparison with otheralgorithms. One can see that the sparse representation basedclassification methods still can not obtain satisfactory resultson some challenge databases. For example, all these repre-sentative algorithms can achieve relatively inferior experi-mental results on the LFW dataset shown in Subsection IX-B,because the LFWdataset is designed for studying the problemof unconstrained face recognition [216] and most of the faceimages are captured under complex environments. One cansee that the PALM algorithm has the worst classificationaccuracy on the LFW dataset and the classification accuracyeven decreases mostly with the increase of the number ofthe training samples. Thus, devising more robust sparserepresentation algorithm is an urgent issue.

Third, enough attention should be paid on thecomputational inefficiency of sparse representation withl1-norm minimization. One can see that high computationalcomplexity is one of the most major drawbacks of thecurrent sparse representation methods and also hampers its

applications in real-time processing scenarios. In terms ofspeed, PALM, FISTA and DALM take much longer time toconverge than the other methods. The average computationaltime of OMP and TPTSR is the two lowest algorithms. More-over, compared with the l1-regularized sparse representationbased classification methods, the TPTSR has very competi-tive classification accuracy but significantly low complexity.Efficient and effective sparse representation methods areurgently needed by real-time applications. Thus, developingmore efficient and effective methods is essential for futurestudy on sparse representation.

Finally, the extensive experimental results have demon-strated that there is no absolute winner that can achievethe best performance for all datasets in terms of classifica-tion accuracy and computational efficiency. However, l1_ls,TPTSR and homotopy algorithms as a whole outperform theother algorithms. As a compromising approach, the OMPalgorithm can achieve distinct efficiency without sacrificingmuch recognition rate in comparison with other algorithmsand it also has been extensively applied to some complexlearning algorithms as a function.

X. CONCLUSIONSparse representation has been extensively studied in recentyears. This paper summarizes and presents various avail-able sparse representation methods and discusses theirmotivations, mathematical representations and extensiveapplications. More specifically, we have analyzed their rela-tions in theory and empirically introduced the applicationsincluding dictionary learning based on sparse representationand real-world applications such as image processing, imageclassification, and visual tracking.

Sparse representation has become a fundamental tool,which has been embedded into various learning systemsand also has received dramatic improvements and unprece-dented achievements. Furthermore, dictionary learning isan extremely popular topic and is closely connected withsparse representation. Currently, efficient sparse represen-tation, robust sparse representation, and dictionary learningbased on sparse representation seem to be the main streamsof research on sparse representation methods. The low-rankrepresentation technique has also recently aroused intensiveresearch interests and sparse representation has beenintegrated into low-rank representation for constructing morereliable representation models. However, the mathematicaljustification of low-rank representation seems not to beelegant as sparse representation. Because employing the ideasof sparse representation as a prior can lead to state-of-the-art results, incorporating sparse representation with low-rankrepresentation is worth further research. Moreover, subspacelearning also has been becoming one of the most prevailingtechniques in pattern recognition and computer vision. It isnecessary to further study the relationship between sparserepresentation and subspace learning, and constructing morecompact models for sparse subspace learning becomes oneof the popular topics in various research fields. The transfer

524 VOLUME 3, 2015

www.redpel.com35+917620593389

www.redpel.com35+917620593389

Page 36: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

learning technique has emerged as a new learning frameworkfor classification, regression and clustering problems in datamining and machine learning. However, sparse representa-tion research still has been not fully applied to the transferlearning framework and it is significant to unify the sparserepresentation and low-rank representation techniques intothe transfer learning framework to solve domain adaption,multitask learning, sample selection bias and covariate shiftproblems. Furthermore, researches on deep learning seemsto become an overwhelming trend in the computer visionfield. However, dramatically expensive training effort is themain limitation of current deep learning technique and howto fully introduce current sparse representation methods intothe framework of deep learning is valuable and unsolved.

The application scope of sparse representation has emergedand has been widely extended to machine learning andcomputer vision fields. Nevertheless, the effectiveness andefficiency of sparse representation methods cannot perfectlymeet the need for real-world applications. Especially, thecomplexities of sparse representation have greatly affectedthe applicability, especially the applicability to large scaleproblems. Enhancing the robustness of sparse representa-tion is considered as another indispensable problem whenresearchers design algorithms. For image classification, therobustness should be seriously considered, such as the robust-ness to random corruptions, varying illuminations, outliers,occlusion and complex backgrounds. Thus, developing anefficient and robust sparse representation method for sparserepresentation is still the main challenge and to design a moreeffective dictionary is being expected and is beneficial to theperformance improvement.

Sparse representation still has wide potential for variouspossible applications, such as event detection, scenereconstruction, video tracking, object recognition, objectpose estimation, medical image processing, genetic expres-sion and natural language processing. For example, the studyof sparse representation in visual tracking is an importantdirection andmore depth studies are essential to future furtherimprovements of visual tracking research.

In addition, most sparse representation and dictionarylearning algorithms focus on employing the l0-norm orl1-norm regularization to obtain a sparse solution. However,there are still only a few studies on l2,1-norm regularizationbased sparse representation and dictionary learningalgorithms. Moreover, other extended studies of sparse repre-sentation may be fruitful. In summary, the recent prevalenceof sparse representation has extensively influenced differentfields. It is our hope that the review and analysis presented inthis paper can help and motivate more researchers to proposeperfect sparse representation methods.

ACKNOWLEDGMENTThe authors would like to thank Jian Wu for many inspiringdiscussions and he is ultimately responsible for many of theideas in the algorithm and analysis. They would also like tothank Dr. Zhihui Lai, Dr. Jinxing Liu and Xiaozhao Fang

for constructive suggestions. Moreover, they thank the editor,an associate editor, and referees for helpful comments andsuggestions which greatly improved this paper.

REFERENCES[1] B. K. Natarajan, ‘‘Sparse approximate solutions to linear systems,’’ SIAM

J. Comput., vol. 24, no. 2, pp. 227–234, 1995.[2] M. Huang et al., ‘‘Brain extraction based on locally linear representation-

based classification,’’ NeuroImage, vol. 92, pp. 322–339, May 2014.[3] X. Lu and X. Li, ‘‘Group sparse reconstruction for image segmentation,’’

Neurocomputing, vol. 136, pp. 41–48, Jul. 2014.[4] M. Elad, M. A. T. Figueiredo, and Y. Ma, ‘‘On the role of sparse and

redundant representations in image processing,’’ Proc. IEEE, vol. 98,no. 6, pp. 972–982, Jun. 2010.

[5] S. Mallat, A Wavelet Tour of Signal Processing: The Sparse Way.New York, NY, USA: Academic, 2008.

[6] J.-L. Starck, F. Murtagh, and J. M. Fadili, Sparse Image and SignalProcessing: Wavelets, Curvelets, Morphological Diversity. Cambridge,U.K.: Cambridge Univ. Press, 2010.

[7] M. Elad, Sparse and Redundant Representations: From Theory toApplications in Signal and Image Processing. New York, NY, USA:Springer-Verlag, 2010.

[8] A. M. Bruckstein, D. L. Donoho, and M. Elad, ‘‘From sparse solutions ofsystems of equations to sparse modeling of signals and images,’’ SIAMRev., vol. 51, no. 1, pp. 34–81, 2009.

[9] Y. Xu, D. Zhang, J. Yang, and J.-Y. Yang, ‘‘A two-phase test samplesparse representation method for use with face recognition,’’ IEEE Trans.Circuits Syst. Video Technol., vol. 21, no. 9, pp. 1255–1262, Sep. 2011.

[10] J. Wright, Y. Ma, J. Mairal, G. Sapiro, T. S. Huang, and S. Yan, ‘‘Sparserepresentation for computer vision and pattern recognition,’’ Proc. IEEE,vol. 98, no. 6, pp. 1031–1044, Jun. 2010.

[11] D. L. Donoho, ‘‘Compressed sensing,’’ IEEE Trans. Inf. Theory, vol. 52,no. 4, pp. 1289–1306, Apr. 2006.

[12] R. G. Baraniuk, ‘‘Compressive sensing [lecture notes],’’ IEEE SignalProcess. Mag., vol. 24, no. 4, pp. 118–121, Jul. 2007.

[13] E. J. Candès, J. Romberg, and T. Tao, ‘‘Robust uncertainty principles:Exact signal reconstruction from highly incomplete frequency informa-tion,’’ IEEE Trans. Inf. Theory, vol. 52, no. 2, pp. 489–509, Feb. 2006.

[14] E. J. Candès and M. B. Wakin, ‘‘An introduction to compressive sam-pling,’’ IEEE Signal Process. Mag., vol. 25, no. 2, pp. 21–30, Mar. 2008.

[15] Y. Tsaig and D. L. Donoho, ‘‘Extensions of compressed sensing,’’ SignalProcess., vol. 86, no. 3, pp. 549–571, 2006.

[16] E. J. Candès, ‘‘Compressive sampling,’’ in Proc. Int. Congr. Math.,Madrid, Spain, Aug. 2006, pp. 1433–1452.

[17] E. Candès and J. Romberg, ‘‘Sparsity and incoherence in compressivesampling,’’ Inverse Problems, vol. 23, no. 3, p. 969, 2007.

[18] X. Lu, H. Wu, Y. Yuan, P. Yan, and X. Li, ‘‘Manifold regularized sparseNMF for hyperspectral unmixing,’’ IEEE Trans. Geosci. Remote Sens.,vol. 51, no. 5, pp. 2815–2826, May 2013.

[19] Y. Yuan, X. Li, Y. Pang, X. Lu, and D. Tao, ‘‘Binary sparse nonnegativematrix factorization,’’ IEEE Trans. Circuits Syst. Video Technol., vol. 19,no. 5, pp. 772–777, May 2009.

[20] J. Wright, A. Y. Yang, A. Ganesh, S. S. Sastry, and Y. Ma, ‘‘Robust facerecognition via sparse representation,’’ IEEE Trans. Pattern Anal. Mach.Intell., vol. 31, no. 2, pp. 210–227, Feb. 2009.

[21] Z. Zhang, Z. Li, B. Xie, L. Wang, and Y. Chen, ‘‘Integrating globality andlocality for robust representation based classification,’’ Math. ProblemsEng., vol. 2014, Mar. 2014, Art. ID 415856.

[22] I. Naseem, R. Togneri, and M. Bennamoun, ‘‘Linear regression for facerecognition,’’ IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 11,pp. 2106–2112, Nov. 2010.

[23] H. Cheng, Z. Liu, L. Yang, and X. Chen, ‘‘Sparse representation andlearning in visual recognition: Theory and applications,’’ Signal Process.,vol. 93, no. 6, pp. 1408–1425, 2013.

[24] J. A. Tropp, A. C. Gilbert, and M. J. Strauss, ‘‘Algorithms for simul-taneous sparse approximation: Part I: Greedy pursuit,’’ Signal Process.,vol. 86, no. 3, pp. 572–588, 2006.

[25] J. A. Tropp, ‘‘Algorithms for simultaneous sparse approximation. Part II:Convex relaxation,’’ Signal Process., vol. 86, no. 3, pp. 589–602, 2006.

[26] M. Schmidt, G. Fung, and R. Rosaless, ‘‘Optimization methods for`1-regularization,’’ Univ. British Columbia, Vancouver, BC, Canada,Tech. Rep. TR-2009-19, 2009.

VOLUME 3, 2015 525

www.redpel.com36+917620593389

www.redpel.com36+917620593389

Page 37: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

[27] E. Amaldi and V. Kann, ‘‘On the approximability of minimizing nonzerovariables or unsatisfied relations in linear systems,’’ Theoretical Comput.Sci., vol. 209, nos. 1–2, pp. 237–260, 1998.

[28] J. A. Tropp, ‘‘Greed is good: Algorithmic results for sparse approxima-tion,’’ IEEE Trans. Inf. Theory, vol. 50, no. 10, pp. 2231–2242, Oct. 2004.

[29] N. Parikh and S. Boyd, ‘‘Proximal algorithms,’’ Found. Trends Optim.,vol. 1, no. 3, pp. 123–231, 2013.

[30] D. L. Donoho and Y. Tsaig, ‘‘Fast solution of `1-norm minimizationproblems when the solution may be sparse,’’ IEEE Trans. Inf. Theory,vol. 54, no. 11, pp. 4789–4812, Nov. 2008.

[31] Z. Zhang, L. Wang, Q. Zhu, Z. Liu, and Y. Chen, ‘‘Noise modelingand representation based classification methods for face recognition,’’Neurocomputing, vol. 148, pp. 420–429, Jan. 2015.

[32] S. Boyd and L. Vandenberghe, Convex Optimization. Cambridge, U.K.:Cambridge Univ. Press, 2009.

[33] J. Chen and X. Huo, ‘‘Theoretical results on sparse representations ofmultiple-measurement vectors,’’ IEEE Trans. Signal Process., vol. 54,no. 12, pp. 4634–4643, Dec. 2006.

[34] J. K. Pant, W.-S. Lu, and A. Antoniou, ‘‘Unconstrained regularized`p-norm based algorithm for the reconstruction of sparse signals,’’ inProc. IEEE Int. Symp. Circuits Syst., May 2011, pp. 1740–1743.

[35] R. Chartrand andW.Yin, ‘‘Iteratively reweighted algorithms for compres-sive sensing,’’ in Proc. IEEE Int. Conf. Acoust., Speech Signal Process.,Mar./Apr. 2008, pp. 3869–3872.

[36] J. Yang, L. Zhang, Y. Xu, and J.-Y. Yang, ‘‘Beyond sparsity: The role ofL1-optimizer in pattern classification,’’ Pattern Recognit., vol. 45, no. 3,pp. 1104–1118, 2012.

[37] J. A. Tropp and A. C. Gilbert, ‘‘Signal recovery from random measure-ments via orthogonal matching pursuit,’’ IEEE Trans. Inf. Theory, vol. 53,no. 12, pp. 4655–4666, Dec. 2007.

[38] D. Needell and R. Vershynin, ‘‘Uniform uncertainty principle and signalrecovery via regularized orthogonal matching pursuit,’’ Found. Comput.Math., vol. 9, no. 3, pp. 317–334, 2009.

[39] R. Saab, R. Chartrand, and O. Yilmaz, ‘‘Stable sparse approximations vianonconvex optimization,’’ inProc. IEEE Int. Conf. Acoust., Speech SignalProcess., Mar./Apr. 2008, pp. 3885–3888.

[40] R. Chartrand, ‘‘Exact reconstruction of sparse signals via nonconvexminimization,’’ IEEE Signal Process. Lett., vol. 14, no. 10, pp. 707–710,Oct. 2007.

[41] Z. Xu, ‘‘Data modeling: Visual psychology approach and L1/2 regulariza-tion theory,’’ in Proc. Int. Congr. Math., 2010, pp. 3151–3184.

[42] R. Tibshirani, ‘‘Regression shrinkage and selection via the lasso,’’ J. Roy.Statist. Soc. B, vol. 58, no. 1, pp. 267–288, 1996.

[43] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, ‘‘Least angle regres-sion,’’ Ann. Statist., vol. 32, no. 2, pp. 407–499, 2004.

[44] J. Yang and Y. Zhang, ‘‘Alternating direction algorithms for`1-problems in compressive sensing,’’ SIAM J. Sci. Comput., vol. 33,no. 1, pp. 250–278, 2011.

[45] M. Schmidt, G. Fung, and R. Rosales, ‘‘Fast optimization methodsfor L1 regularization: A comparative study and two new approaches,’’in Machine Learning. Berlin, Germany: Springer-Verlag, 2007,pp. 286–297.

[46] F. Nie, H. Huang, X. Cai, and C. Ding, ‘‘Efficient and robust featureselection via joint `2,1-norms minimization,’’ in Proc. Adv. Neural Inf.Process. Syst., 2010, pp. 1813–1821.

[47] Y. Yang, H. T. Shen, Z. Ma, Z. Huang, and X. Zhou, ‘‘`2,1-norm regular-ized discriminative feature selection for unsupervised learning,’’ in Proc.22nd Int. Joint Conf. Artif. Intell., 2011, vol. 22, no. 1, pp. 1589–1594.

[48] X. Shi, Y. Yang, Z. Guo, and Z. Lai, ‘‘Face recognition by sparse dis-criminant analysis via joint L2,1-norm minimization,’’ Pattern Recognit.,vol. 47, no. 7, pp. 2447–2453, 2014.

[49] J. Liu, S. Ji, and J. Ye, ‘‘Multi-task feature learning via efficient`2,1-norm minimization,’’ in Proc. 25th Conf. Uncertainty Artif. Intell.,2009, pp. 339–348.

[50] C. Hou, F. Nie, X. Li, D. Yi, and Y. Wu, ‘‘Joint embedding learning andsparse regression: A framework for unsupervised feature selection,’’ IEEETrans. Cybern., vol. 44, no. 6, pp. 793–804, Jun. 2014.

[51] D. Zhang, M. Yang, and X. Feng, ‘‘Sparse representation or collaborativerepresentation: Which helps face recognition?’’ in Proc. IEEE Int. Conf.Comput. Vis. (ICCV), Nov. 2011, pp. 471–478.

[52] D. L. Donoho and M. Elad, ‘‘Optimally sparse representation in general(nonorthogonal) dictionaries via `1 minimization,’’ Proc. Nat. Acad.Sci. USA, vol. 100, no. 5, pp. 2197–2202, 2003.

[53] L. Liu, L. Shao, F. Zheng, and X. Li, ‘‘Realistic action recognition viasparsely-constructed Gaussian processes,’’ Pattern Recognit., vol. 47,no. 12, pp. 3819–3827, 2014.

[54] V. M. Patel and R. Chellappa, ‘‘Sparse representations, compressivesensing and dictionaries for pattern recognition,’’ in Proc. IEEE 1st AsianConf. Pattern Recognit., Nov. 2011, pp. 325–329.

[55] Y. Yuan, X. Lu, and X. Li, ‘‘Learning hash functions using sparse recon-struction,’’ inProc. Int. Conf. Internet Multimedia Comput. Service, 2014,pp. 14–18.

[56] D. L. Donoho, ‘‘For most large underdetermined systems of linearequations the minimal `1-norm solution is also the sparsest solution,’’Commun. Pure Appl. Math., vol. 59, no. 6, pp. 797–829, 2006.

[57] E. J. Candès, J. K. Romberg, and T. Tao, ‘‘Stable signal recovery fromincomplete and inaccurate measurements,’’ Commun. Pure Appl. Math.,vol. 59, no. 8, pp. 1207–1223, 2006.

[58] E. J. Candès and T. Tao, ‘‘Near-optimal signal recovery from randomprojections: Universal encoding strategies?’’ IEEE Trans. Inf. Theory,vol. 52, no. 12, pp. 5406–5425, Dec. 2006.

[59] Q. Lyu, Z. Lin, Y. She, and C. Zhang, ‘‘A comparison of typical`p minimization algorithms,’’ Neurocomputing, vol. 119, pp. 413–424,Nov. 2013.

[60] Z. Xu, X. Chang, F. Xu, and H. Zhang, ‘‘L1/2 regularization:A thresholding representation theory and a fast solver,’’ IEEE Trans.Neural Netw. Learn. Syst., vol. 23, no. 7, pp. 1013–1027, Jul. 2012.

[61] S. Guo, Z. Wang, and Q. Ruan, ‘‘Enhancing sparsity via `p (0<p<1)minimization for robust face recognition,’’ Neurocomputing, vol. 99,pp. 592–602, Jan. 2013.

[62] C. Ding, D. Zhou, X. He, and H. Zha, ‘‘R1-PCA: Rotational invariantL1-norm principal component analysis for robust subspace factorization,’’in Proc. 23rd Int. Conf. Mach. Learn., 2006, pp. 281–288.

[63] S. G. Mallat and Z. Zhang, ‘‘Matching pursuits with time-frequency dic-tionaries,’’ IEEE Trans. Signal Process., vol. 41, no. 12, pp. 3397–3415,Dec. 1993.

[64] Y. C. Pati, R. Rezaiifar, and P. S. Krishnaprasad, ‘‘Orthogonal matchingpursuit: Recursive function approximation with applications to waveletdecomposition,’’ in Proc. 27th Asilomar Conf. Signals, Syst., Comput.,Nov. 1993, pp. 40–44.

[65] S. N. Vitaladevuni, P. Natarajan, and R. Prasad, ‘‘Efficient orthogonalmatching pursuit using sparse random projections for scene and videoclassification,’’ in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), Nov. 2011,pp. 2312–2319.

[66] D. Needell and J. A. Tropp, ‘‘CoSaMP: Iterative signal recovery fromincomplete and inaccurate samples,’’ Appl. Comput. Harmon. Anal.,vol. 26, no. 3, pp. 301–321, 2009.

[67] D. L. Donoho, Y. Tsaig, I. Drori, and J.-L. Starck, ‘‘Sparse solu-tion of underdetermined systems of linear equations by stagewiseorthogonal matching pursuit,’’ IEEE Trans. Inf. Theory, vol. 58, no. 2,pp. 1094–1121, Feb. 2012.

[68] W. Dai and O. Milenkovic, ‘‘Subspace pursuit for compressive sensingsignal reconstruction,’’ IEEE Trans. Inf. Theory, vol. 55, no. 5,pp. 2230–2249, May 2009.

[69] T. T. Do, L. Gan, N. Nguyen, and T. D. Tran, ‘‘Sparsity adaptivematching pursuit algorithm for practical compressed sensing,’’ in Proc.42nd Asilomar Conf. Signals, Syst., Comput., Oct. 2008, pp. 581–587.

[70] P. Jost, P. Vandergheynst, and P. Frossard, ‘‘Tree-based pursuit:Algorithm and properties,’’ IEEE Trans. Signal Process., vol. 54, no. 12,pp. 4685–4697, Dec. 2006.

[71] C. La and M. N. Do, ‘‘Tree-based orthogonal matching pursuit algorithmfor signal reconstruction,’’ in Proc. IEEE Int. Conf. Image Process.,Oct. 2006, pp. 1277–1280.

[72] N. B. Karahanoglu and H. Erdogan, ‘‘Compressed sensing signalrecovery via forward–backward pursuit,’’ Digit. Signal Process., vol. 23,no. 5, pp. 1539–1548, 2013.

[73] M. A. T. Figueiredo, R. D. Nowak, and S. J. Wright, ‘‘Gradient projectionfor sparse reconstruction: Application to compressed sensing and otherinverse problems,’’ IEEE J. Sel. Topics Signal Process., vol. 1, no. 4,pp. 586–597, Dec. 2007.

[74] S.-J. Kim, K. Koh, M. Lustig, S. Boyd, and D. Gorinevsky, ‘‘An interior-point method for large-scale `1-regularized least squares,’’ IEEE J. Sel.Topics Signal Process., vol. 1, no. 4, pp. 606–617, Dec. 2007.

[75] L. F. Portugal, M. G. C. Resende, G. Veiga, and J. J. Júdice,‘‘A truncated primal-infeasible dual-feasible network interior pointmethod,’’ Networks, vol. 35, no. 2, pp. 91–108, 2000.

526 VOLUME 3, 2015

www.redpel.com37+917620593389

www.redpel.com37+917620593389

Page 38: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

[76] K. Koh, S.-J. Kim, and S. Boyd, ‘‘An interior-point method for large-scale`1-regularized logistic regression,’’ J. Mach. Learn. Res., vol. 8, no. 8,pp. 1519–1555, 2007.

[77] S. Mehrotra, ‘‘On the implementation of a primal-dual interior pointmethod,’’ SIAM J. Optim., vol. 2, no. 4, pp. 575–601, 1992.

[78] S. J. Wright, Primal-Dual Interior-Point Methods, vol. 54. Philadelphia,PA, USA: SIAM, 1997.

[79] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, ‘‘Distributedoptimization and statistical learning via the alternating direction methodof multipliers,’’ Found. Trends Mach. Learn., vol. 3, no. 1, pp. 1–122,2011.

[80] M. A. T. Figueiredo and R. D. Nowak, ‘‘A bound optimization approachto wavelet-based image deconvolution,’’ in Proc. IEEE Int. Conf. ImageProcess., vol. 2. Sep. 2005, pp. II-782–II-785.

[81] P. L. Combettes and J. C. Pesquet, ‘‘Proximal splitting methods insignal processing,’’ in Fixed-Point Algorithms for Inverse Problems inScience and Engineering. New York, NY, USA: Springer-Verlag, 2011,pp. 185–212.

[82] A. Beck and M. Teboulle, ‘‘A fast iterative shrinkage-thresholdingalgorithm for linear inverse problems,’’ SIAM J. Imag. Sci., vol. 2, no. 1,pp. 183–202, 2009.

[83] A. Y. Yang, S. S. Sastry, A. Ganesh, and Y. Ma, ‘‘Fast `1-minimizationalgorithms and an application in robust face recognition: A review,’’in Proc. 17th IEEE Int. Conf. Image Process. (ICIP), Sep. 2010,pp. 1849–1852.

[84] S. J. Wright, R. D. Nowak, and M. A. T. Figueiredo, ‘‘Sparsereconstruction by separable approximation,’’ IEEE Trans. SignalProcess., vol. 57, no. 7, pp. 2479–2493, Jul. 2009.

[85] Y.-H. Dai, W. W. Hager, K. Schittkowski, and H. Zhang, ‘‘The cyclicBarzilai–Borwein method for unconstrained optimization,’’ IMA J.Numer. Anal., vol. 26, no. 3, pp. 604–627, 2006.

[86] E. T. Hale, W. Yin, and Y. Zhang, ‘‘A fixed-point continuationmethod for `1-regularized minimization with applications to compressedsensing,’’ Dept. Comput. Appl. Math., Rice Univ., Houston, TX, USA,CAAM Tech. Rep. TR07-07, 2007.

[87] J. Zeng, Z. Xu, B. Zhang, W. Hong, and Y. Wu, ‘‘Accelerated L1/2regularization based SAR imaging via BCR and reduced Newton skills,’’Signal Process., vol. 93, no. 7, pp. 1831–1844, 2013.

[88] J. Zeng, S. Lin, Y. Wang, and Z. Xu, ‘‘L1/2 regularization: Convergenceof iterative half thresholding algorithm,’’ IEEE Trans. Signal Process.,vol. 62, no. 9, pp. 2317–2329, May 2014.

[89] A. Y. Yang, Z. Zhou, A. G. Balasubramanian, S. S. Sastry, and Y. Ma,‘‘Fast `1-minimization algorithms for robust face recognition,’’ IEEETrans. Image Process., vol. 22, no. 8, pp. 3234–3246, Aug. 2013.

[90] M. Elad, B. Matalon, and M. Zibulevsky, ‘‘Coordinate and subspaceoptimization methods for linear least squares with non-quadraticregularization,’’ Appl. Comput. Harmon. Anal., vol. 23, no. 3,pp. 346–367, 2007.

[91] D. L. Donoho, A. Maleki, and A. Montanari, ‘‘Message-passingalgorithms for compressed sensing,’’ Proc. Nat. Acad. Sci. USA, vol. 106,no. 45, pp. 18914–18919, 2009.

[92] S. Becker, J. Bobin, and E. J. Candès, ‘‘NESTA: A fast and accuratefirst-order method for sparse recovery,’’ SIAM J. Imag. Sci., vol. 4, no. 1,pp. 1–39, 2011.

[93] S. R. Becker, E. J. Candès, and M. C. Grant, ‘‘Templates for convex coneproblems with applications to sparse signal recovery,’’ Math. Program.Comput., vol. 3, no. 3, pp. 165–218, 2011.

[94] M. R. Osborne, B. Presnell, and B. A. Turlach, ‘‘A new approach tovariable selection in least squares problems,’’ IMA J. Numer. Anal.,vol. 20, no. 3, pp. 389–403, 2000.

[95] M. D. Plumbley, ‘‘Recovery of sparse representations by polytopefaces pursuit,’’ in Independent Component Analysis and Blind SignalSeparation. Berlin, Germany: Springer-Verlag, 2006, pp. 206–213.

[96] M. S. Asif and J. Romberg, ‘‘Fast and accurate algorithms for re-weighted`1-norm minimization,’’ IEEE Trans. Signal Process., vol. 61, no. 23,pp. 5905–5916, Dec. 2012.

[97] D. M. Malioutov, M. Cetin, and A. S. Willsky, ‘‘Homotopy continua-tion for sparse signal representation,’’ in Proc. IEEE Int. Conf. Acoust.,Speech, Signal Process. (ICASSP), vol. 5. Mar. 2005, pp. v733–v736.

[98] P. Garrigues and L. E. Ghaoui, ‘‘An homotopy algorithm for the lassowith online observations,’’ in Advances in Neural Information ProcessingSystems. Red Hook, NY, USA: Curran Associates, 2009, pp. 489–496.

[99] M. S. Asif, ‘‘Primal dual pursuit: A homotopy based algorithmfor the dantzig selector,’’ M.S. thesis, School Elect. Comput. Eng.,Georgia Inst. Technol., Atlanta, GA, USA, 2008.

[100] M. S. Asif and J. Romberg, ‘‘Dynamic updating for `1 minimization,’’IEEE J. Sel. Topics Signal Process., vol. 4, no. 2, pp. 421–434, Apr. 2010.

[101] M. S. Asif and J. Romberg. (2013). ‘‘Sparse recovery of streaming signalsusing `1-homotopy.’’ [Online]. Available: http://arxiv.org/abs/1306.3331

[102] M. S. Asif, ‘‘Dynamic compressive sensing: Sparse recovery algorithmsfor streaming signals and video,’’ Ph.D. dissertation, Dept. Elect. Comput.Eng., Georgia Inst. Technol., Atlanta, GA, USA, 2013.

[103] J. W. Cooley and J. W. Tukey, ‘‘An algorithm for the machine calculationof complex Fourier series,’’Math. Comput., vol. 19, no. 90, pp. 297–301,1965.

[104] R. Rubinstein, A. M. Bruckstein, and M. Elad, ‘‘Dictionaries for sparserepresentation modeling,’’ Proc. IEEE, vol. 98, no. 6, pp. 1045–1057,Jun. 2010.

[105] L. Shao, R. Yan, X. Li, and Y. Liu, ‘‘From heuristic optimization to dictio-nary learning: A review and comprehensive comparison of image denois-ing algorithms,’’ IEEE Trans. Cybern., vol. 44, no. 7, pp. 1001–1013,Jul. 2014.

[106] E. P. Simoncelli and E. H. Adelson, ‘‘Noise removal via Bayesianwavelet coring,’’ in Proc. Int. Conf. Image Process., vol. 1. Sep. 1996,pp. 379–382.

[107] W. He, Y. Zi, B. Chen, F. Wu, and Z. He, ‘‘Automatic faultfeature extraction of mechanical anomaly on induction motor bearingusing ensemble super-wavelet transform,’’ Mech. Syst. Signal Process.,vols. 54–55, pp. 457–480, Mar. 2015.

[108] E. Le Pennec and S. Mallat, ‘‘Sparse geometric image representationswith bandelets,’’ IEEE Trans. Image Process., vol. 14, no. 4, pp. 423–438,Apr. 2005.

[109] J.-L. Starck, E. J. Candès, and D. L. Donoho, ‘‘The curvelet transformfor image denoising,’’ IEEE Trans. Image Process., vol. 11, no. 6,pp. 670–684, Jun. 2002.

[110] M. N. Do and M. Vetterli, ‘‘The contourlet transform: An efficientdirectional multiresolution image representation,’’ IEEE Trans. ImageProcess., vol. 14, no. 12, pp. 2091–2106, Dec. 2005.

[111] E. P. Simoncelli, W. T. Freeman, E. H. Adelson, and D. J. Heeger,‘‘Shiftablemultiscale transforms,’’ IEEE Trans. Inf. Theory, vol. 38, no. 2,pp. 587–607, Mar. 1992.

[112] J. Shi, X. Ren, G. Dai, J. Wang, and Z. Zhang, ‘‘A non-convex relaxationapproach to sparse dictionary learning,’’ inProc. IEEEConf. Comput. Vis.Pattern Recognit. (CVPR), Jun. 2011, pp. 1809–1816.

[113] J. Fan and R. Li, ‘‘Variable selection via nonconcave penalized likelihoodand its oracle properties,’’ J. Amer. Statist. Assoc., vol. 96, no. 456,pp. 1348–1360, 2001.

[114] C.-H. Zhang, ‘‘Nearly unbiased variable selection under minimax con-cave penalty,’’ Ann. Statist., vol. 38, no. 2, pp. 894–942, 2010.

[115] J. H. Friedman, ‘‘Fast sparse regression and classification,’’ Int.J. Forecasting, vol. 28, no. 3, pp. 722–738, 2012.

[116] I. Tošić and P. Frossard, ‘‘Dictionary learning,’’ IEEE Signal Process.Mag., vol. 28, no. 2, pp. 27–38, Mar. 2011.

[117] C. Bao, H. Ji, Y. Quan, and Z. Shen, ‘‘`0 norm based dictionary learningby proximal methods with global convergence,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2014, pp. 3858–3865.

[118] B. A. Olshausen and D. J. Field, ‘‘Sparse coding with an overcompletebasis set: A strategy employed by V1?’’ Vis. Res., vol. 37, no. 23,pp. 3311–3325, 1997.

[119] K. Engan, S. O. Aase, and J. H. Husøy, ‘‘Multi-frame compression:Theory and design,’’ Signal Process., vol. 80, no. 10, pp. 2121–2140,2000.

[120] K. Kreutz-Delgado, J. F. Murray, B. D. Rao, K. Engan, T.-W. Lee, andT. J. Sejnowski, ‘‘Dictionary learning algorithms for sparse representa-tion,’’ Neural Comput., vol. 15, no. 2, pp. 349–396, 2003.

[121] M. Aharon, M. Elad, and A. Bruckstein, ‘‘K-SVD: An algorithm fordesigning overcomplete dictionaries for sparse representation,’’ IEEETrans. Signal Process., vol. 54, no. 11, pp. 4311–4322, Nov. 2006.

[122] R. Jenatton, J. Mairal, G. Obozinski, and F. Bach, ‘‘Proximal methods forsparse hierarchical dictionary learning,’’ in Proc. 27th Int. Conf. Mach.Learn., 2010, pp. 487–494.

[123] S. Bengio, F. Pereira, Y. Singer, and D. Strelow, ‘‘Group sparse coding,’’in Proc. Adv. Neural Inf. Process. Syst., 2009, pp. 82–89.

[124] B. Zhao, F. F. Li, and E. P. Xing, ‘‘Online detection of unusual eventsin videos via dynamic sparse coding,’’ in Proc. IEEE Conf. Comput. Vis.Pattern Recognit., Jun. 2011, pp. 3313–3320.

VOLUME 3, 2015 527

www.redpel.com38+917620593389

www.redpel.com38+917620593389

Page 39: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

[125] T. Zhang, B. Ghanem, S. Liu, and N. Ahuja, ‘‘Robust visual tracking viamulti-task sparse learning,’’ in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 2042–2049.

[126] Q. Zhang and B. Li, ‘‘Discriminative K-SVD for dictionary learning inface recognition,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.,Jun. 2010, pp. 2691–2698.

[127] Y. Yang, Y. Yang, Z. Huang, H. T. Shen, and F. Nie, ‘‘Tag localizationwith spatial correlations and joint group sparsity,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2011, pp. 881–888.

[128] O. Bryt and M. Elad, ‘‘Compression of facial images using theK-SVD algorithm,’’ J. Vis. Commun. Image Represent., vol. 19, no. 4,pp. 270–282, 2008.

[129] J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong, ‘‘Locality-constrained linear coding for image classification,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2010, pp. 3360–3367.

[130] M. Zhou et al., ‘‘Nonparametric Bayesian dictionary learning for analysisof noisy and incomplete images,’’ IEEE Trans. Image Process., vol. 21,no. 1, pp. 130–144, Jan. 2012.

[131] I. Ramirez and G. Sapiro, ‘‘An MDL framework for sparse codingand dictionary learning,’’ IEEE Trans. Signal Process., vol. 60, no. 6,pp. 2913–2927, Jun. 2012.

[132] J. Mairal, F. Bach, J. Ponce, and G. Sapiro, ‘‘Online learning for matrixfactorization and sparse coding,’’ J. Mach. Learn. Res., vol. 11, pp. 19–60,Mar. 2010.

[133] M. Yang, L. Van Gool, and L. Zhang, ‘‘Sparse variation dictionarylearning for face recognition with a single training sample per person,’’in Proc. IEEE Int. Conf. Comput. Vis. (ICCV), Dec. 2013, pp. 689–696.

[134] Z. Jiang, Z. Lin, and L. S. Davis, ‘‘Learning a discriminative dictionary forsparse coding via label consistent K-SVD,’’ in Proc. IEEE Conf. Comput.Vis. Pattern Recognit. (CVPR), Jun. 2011, pp. 1697–1704.

[135] Z. Jiang, Z. Lin, and L. S. Davis, ‘‘Label consistent K-SVD: Learninga discriminative dictionary for recognition,’’ IEEE Trans. Pattern Anal.Mach. Intell., vol. 35, no. 11, pp. 2651–2664, Nov. 2013.

[136] M. Yang, L. Zhang, X. Feng, and D. Zhang, ‘‘Fisher discriminationdictionary learning for sparse representation,’’ in Proc. IEEE Int. Conf.Comput. Vis., Nov. 2011, pp. 543–550.

[137] L. Rosasco, A. Verri, M. Santoro, S. Mosci, and S. Villa, ‘‘Iterativeprojection methods for structured sparsity regularization,’’ MassachusettsInst. Technol., Cambridge, MA, USA, Tech. Rep. MIT-CSAIL-TR-2009-050, CBCL-282, 2009.

[138] M. Yang, L. Zhang, J. Yang, and D. Zhang, ‘‘Metaface learning for sparserepresentation based face recognition,’’ in Proc. IEEE Int. Conf. ImageProcess. (ICIP), Sep. 2010, pp. 1601–1604.

[139] I. Ramirez, P. Sprechmann, and G. Sapiro, ‘‘Classification and clus-tering via dictionary learning with structured incoherence and sharedfeatures,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2010,pp. 3501–3508.

[140] S. Kong and D. Wang, ‘‘A dictionary learning approach for classification:Separating the particularity and the commonality,’’ in Proc. 12th Eur.Conf. Comput. Vis. (ECCV), 2012, pp. 186–199.

[141] S. Gao, I. W.-H. Tsang, and Y. Ma, ‘‘Learning category-specific dictio-nary and shared dictionary for fine-grained image categorization,’’ IEEETrans. Image Process., vol. 23, no. 2, pp. 623–634, Feb. 2014.

[142] M. G. Jafari and M. D. Plumbley, ‘‘Fast dictionary learning for sparserepresentations of speech signals,’’ IEEE J. Sel. Topics Signal Process.,vol. 5, no. 5, pp. 1025–1031, Sep. 2011.

[143] J. Mairal, F. R. Bach, J. Ponce, G. Sapiro, and A. Zisserman, ‘‘Superviseddictionary learning,’’ in Proc. Adv. Neural Inf. Process. Syst., 2009,pp. 1033–1040.

[144] N. Zhou, Y. Shen, J. Peng, and J. Fan, ‘‘Learning inter-related visualdictionary for object recognition,’’ in Proc. IEEE Conf. Comput. Vis.Pattern Recognit., Jun. 2012, pp. 3490–3497.

[145] L. Ma, C. Wang, B. Xiao, and W. Zhou, ‘‘Sparse representation forface recognition based on discriminative low-rank dictionary learn-ing,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2012,pp. 2586–2593.

[146] J. Lu, G. Wang, W. Deng, and P. Moulin, ‘‘Simultaneous feature anddictionary learning for image set based face recognition,’’ in Proc. 13thEur. Conf. Comput. Vis. (ECCV), 2014, pp. 265–280.

[147] M. Yang, D. Dai, L. Shen, and L. Van Gool, ‘‘Latent dictionary learn-ing for sparse representation based classification,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2014, pp. 4138–4145.

[148] Z. Jiang, G. Zhang, and L. S. Davis, ‘‘Submodular dictionary learningfor sparse coding,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.,Jun. 2012, pp. 3418–3425.

[149] S. Cai, W. Zuo, L. Zhang, X. Feng, and P. Wang, ‘‘Support vector guideddictionary learning,’’ inProc. 13th Eur. Conf. Comput. Vis. (ECCV), 2014,pp. 624–639.

[150] X. Qin, J. Shen, X. Li, and Y. Jia, ‘‘A new sparse feature-basedpatch for dense correspondence,’’ in Proc. IEEE Int. Conf. MultimediaExpo (ICME), Jul. 2014, pp. 1–6.

[151] L. He, D. Tao, X. Li, and X. Gao, ‘‘Sparse representation for blindimage quality assessment,’’ in Proc. IEEE Conf. Comput. Vis. PatternRecognit. (CVPR), Jun. 2012, pp. 1146–1153.

[152] J. Yang, J. Wright, T. Huang, and Y. Ma, ‘‘Image super-resolutionas sparse representation of raw image patches,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2008, pp. 1–8.

[153] W. T. Freeman, T. R. Jones, and E. C. Pasztor, ‘‘Example-based super-resolution,’’ IEEE Comput. Graph. Appl., vol. 22, no. 2, pp. 56–65,Mar./Apr. 2002.

[154] M. Irani and S. Peleg, ‘‘Motion analysis for image enhancement: Reso-lution, occlusion, and transparency,’’ J. Vis. Commun. Image Represent.,vol. 4, no. 4, pp. 324–335, 1993.

[155] J. Yang, J. Wright, T. S. Huang, and Y. Ma, ‘‘Image super-resolutionvia sparse representation,’’ IEEE Trans. Image Process., vol. 19, no. 11,pp. 2861–2873, Nov. 2010.

[156] Y. Tang, Y. Yuan, P. Yan, and X. Li, ‘‘Greedy regression in sparsecoding space for single-image super-resolution,’’ J. Vis. Commun. ImageRepresent., vol. 24, no. 2, pp. 148–159, 2013.

[157] J. Zhang, C. Zhao, R. Xiong, S. Ma, and D. Zhao, ‘‘Image super-resolution via dual-dictionary learning and sparse representation,’’ inProc. IEEE Int. Symp. Circuits Syst. (ISCAS), May 2012, pp. 1688–1691.

[158] X. Gao, K. Zhang, D. Tao, and X. Li, ‘‘Image super-resolution withsparse neighbor embedding,’’ IEEE Trans. Image Process., vol. 21, no. 7,pp. 3194–3205, Jul. 2012.

[159] C. Fernandez-Granda and E. J. Candès, ‘‘Super-resolution via transform-invariant group-sparse regularization,’’ in Proc. IEEE Int. Conf. Comput.Vis. (ICCV), Dec. 2013, pp. 3336–3343.

[160] X. Lu, H. Yuan, P. Yan, Y. Yuan, and X. Li, ‘‘Geometry constrained sparsecoding for single image super-resolution,’’ in Proc. IEEE Conf. Comput.Vis. Pattern Recognit. (CVPR), Jun. 2012, pp. 1648–1655.

[161] W. Dong, G. Shi, L. Zhang, and X. Wu, ‘‘Super-resolution with nonlocalregularized sparse representation,’’ Proc. SPIE, vol. 7744, p. 77440H,Jul. 2010.

[162] W. Dong, D. Zhang, G. Shi, and X. Wu, ‘‘Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regular-ization,’’ IEEE Trans. Image Process., vol. 20, no. 7, pp. 1838–1857,Jul. 2011.

[163] S. Mallat and G. Yu, ‘‘Super-resolution with sparse mixing estimators,’’IEEE Trans. Image Process., vol. 19, no. 11, pp. 2889–2900, Nov. 2010.

[164] D. L. Donoho, ‘‘De-noising by soft-thresholding,’’ IEEE Trans. Inf.Theory, vol. 41, no. 3, pp. 613–627, May 1995.

[165] M. Elad and M. Aharon, ‘‘Image denoising via sparse and redundantrepresentations over learned dictionaries,’’ IEEE Trans. Image Process.,vol. 15, no. 12, pp. 3736–3745, Dec. 2006.

[166] K. Dabov, A. Foi, V. Katkovnik, and K. Egiazarian, ‘‘Image denoising bysparse 3-D transform-domain collaborative filtering,’’ IEEE Trans. ImageProcess., vol. 16, no. 8, pp. 2080–2095, Aug. 2007.

[167] J. Mairal, M. Elad, and G. Sapiro, ‘‘Sparse representation for color imagerestoration,’’ IEEE Trans. Image Process., vol. 17, no. 1, pp. 53–69,Jan. 2008.

[168] M. Protter and M. Elad, ‘‘Image sequence denoising via sparse andredundant representations,’’ IEEE Trans. Image Process., vol. 18, no. 1,pp. 27–35, Jan. 2009.

[169] W. Dong, X. Li, D. Zhang, and G. Shi, ‘‘Sparsity-based image denoisingvia dictionary learning and structural clustering,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2011, pp. 457–464.

[170] J. Jiang, L. Zhang, and J. Yang, ‘‘Mixed noise removal byweighted encod-ing with sparse nonlocal regularization,’’ IEEE Trans. Image Process.,vol. 23, no. 6, pp. 2651–2662, Jun. 2014.

[171] S. Gu, L. Zhang, W. Zuo, and X. Feng, ‘‘Weighted nuclear norm min-imization with application to image denoising,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2014, pp. 2862–2869.

[172] H. Ji, C. Liu, Z. Shen, and Y. Xu, ‘‘Robust video denoising using low rankmatrix completion,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.,Jun. 2010, pp. 1791–1798.

528 VOLUME 3, 2015

www.redpel.com39+917620593389

www.redpel.com39+917620593389

Page 40: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

[173] P. Cheng, C. Deng, S. Wang, and C. Zhang, ‘‘Image denoising via groupsparse representation over learned dictionary,’’ Proc. SPIE, vol. 8919,p. 891916, Oct. 2013.

[174] J. M. Bioucas-Dias and M. A. T. Figueiredo, ‘‘A new TwIST:Two-step iterative shrinkage/thresholding algorithms for image restora-tion,’’ IEEE Trans. Image Process., vol. 16, no. 12, pp. 2992–3004,Dec. 2007.

[175] J. Mairal, G. Sapiro, and M. Elad, ‘‘Learning multiscale sparse represen-tations for image and video restoration,’’Multiscale Model. Simul., vol. 7,no. 1, pp. 214–241, 2008.

[176] J. Mairal, F. Bach, J. Ponce, G. Sapiro, and A. Zisserman, ‘‘Non-localsparse models for image restoration,’’ in Proc. IEEE 12th Int. Conf.Comput. Vis., Sep./Oct. 2009, pp. 2272–2279.

[177] D. Zoran and Y. Weiss, ‘‘From learning models of natural image patchesto whole image restoration,’’ in Proc. IEEE Int. Conf. Comput. Vis.,Nov. 2011, pp. 479–486.

[178] C. Bao, J.-F. Cai, and H. Ji, ‘‘Fast sparsity-based orthogonal dictionarylearning for image restoration,’’ in Proc. IEEE Int. Conf. Comput. Vis.,Dec. 2013, pp. 3384–3391.

[179] J. Zhang, D. Zhao, and W. Gao, ‘‘Group-based sparse representationfor image restoration,’’ IEEE Trans. Image Process., vol. 23, no. 8,pp. 3336–3351, Aug. 2014.

[180] W. Dong, D. Zhang, and G. Shi, ‘‘Centralized sparse representation forimage restoration,’’ in Proc. IEEE Int. Conf. Comput. Vis., Nov. 2011,pp. 1259–1266.

[181] W.Dong, L. Zhang, G. Shi, andX. Li, ‘‘Nonlocally centralized sparse rep-resentation for image restoration,’’ IEEE Trans. Image Process., vol. 22,no. 4, pp. 1620–1630, Apr. 2013.

[182] A. Buades, B. Coll, and J.-M. Morel, ‘‘A non-local algorithm for imagedenoising,’’ in Proc. IEEE Comput. Soc. Conf. Comput. Vis. PatternRecognit. (CVPR), vol. 2. Jun. 2005, pp. 60–65.

[183] A. Nedic, D. P. Bertsekas, and A. E. Ozdaglar, Convex Analysisand Optimization. Belmont, MA, USA: Athena Scientific,2003.

[184] I. Daubechies, M. Defrise, and C. D. Mol, ‘‘An iterative thresh-olding algorithm for linear inverse problems with a sparsity con-straint,’’ Commun. Pure Appl. Math., vol. 57, no. 11, pp. 1413–1457,Nov. 2004.

[185] W. Deng, J. Hu, and J. Guo, ‘‘Extended SRC: Undersampled face recog-nition via intraclass variant dictionary,’’ IEEE Trans. Pattern Anal. Mach.Intell., vol. 34, no. 9, pp. 1864–1870, Sep. 2012.

[186] W. Deng, J. Hu, and J. Guo, ‘‘In defense of sparsity based face recog-nition,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2013,pp. 399–406.

[187] R. He, W.-S. Zheng, and B.-G. Hu, ‘‘Maximum correntropy criterion forrobust face recognition,’’ IEEE Trans. Pattern Anal. Mach. Intell., vol. 33,no. 8, pp. 1561–1576, Aug. 2011.

[188] M. Yang, L. Zhang, J. Yang, and D. Zhang, ‘‘Robust sparse coding forface recognition,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.,Jun. 2011, pp. 625–632.

[189] J. Yang, K. Yu, Y. Gong, and T. Huang, ‘‘Linear spatial pyramid match-ing using sparse coding for image classification,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2009, pp. 1794–1801.

[190] S. Gao, I. W.-H. Tsang, and L.-T. Chia, ‘‘Kernel sparse representation forimage classification and face recognition,’’ in Proc. Eur. Conf. Comput.Vis. (ECCV), 2010, pp. 1–14.

[191] S. Gao, I. W. Tsang, L.-T. Chia, and P. Zhao, ‘‘Local features are notlonely—Laplacian sparse coding for image classification,’’ in Proc. IEEEConf. Comput. Vis. Pattern Recognit., Jun. 2010, pp. 3555–3561.

[192] N. Kulkarni and B. Li, ‘‘Discriminative affine sparse codes for imageclassification,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit.,Jun. 2011, pp. 1609–1616.

[193] C. Zhang, J. Liu, Q. Tian, C. Xu, H. Lu, and S. Ma, ‘‘Image classifi-cation by non-negative sparse coding, low-rank and sparse decomposi-tion,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2011,pp. 1673–1680.

[194] T. Zhang, B. Ghanem, S. Liu, C. Xu, and N. Ahuja, ‘‘Low-rank sparsecoding for image classification,’’ in Proc. IEEE Int. Conf. Comput. Vis.,Dec. 2013, pp. 281–288.

[195] Y. Zhang, Z. Jiang, and L. S. Davis, ‘‘Learning structured low-rankrepresentations for image classification,’’ in Proc. IEEE Conf. Comput.Vis. Pattern Recognit., Jun. 2013, pp. 676–683.

[196] D. Tao, L. Jin, Y. Zhao, and X. Li, ‘‘Rank preserving sparse learning forKinect based scene classification,’’ IEEE Trans. Cybern., vol. 43, no. 5,pp. 1406–1417, Oct. 2013.

[197] Y. Zhang, Z. Jiang, and L. S. Davis, ‘‘Discriminative tensor sparsecoding for image classification,’’ in Proc. Brit. Mach. Vis. Conf., 2013,pp. 1–11.

[198] L. Zhuang, S. Gao, J. Tang, J. Wang, Z. Lin, and Y. Ma. (2014).‘‘Constructing a non-negative low rank and sparse graph with data-adaptive features.’’ [Online]. Available: http://arxiv.org/abs/1409.0964

[199] R. Rigamonti et al., ‘‘On the relevance of sparsity for imageclassification,’’ Comput. Vis. Image Understand., vol. 125, pp. 115–127,Aug. 2014.

[200] X. Mei and H. Ling, ‘‘Robust visual tracking using `1 minimiza-tion,’’ in Proc. IEEE 12th Int. Conf. Comput. Vis., Sep./Oct. 2009,pp. 1436–1443.

[201] X.Mei andH. Ling, ‘‘Robust visual tracking and vehicle classification viasparse representation,’’ IEEE Trans. Pattern Anal. Mach. Intell., vol. 33,no. 11, pp. 2259–2272, Nov. 2011.

[202] H. Li, C. Shen, and Q. Shi, ‘‘Real-time visual tracking using compressivesensing,’’ in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2011,pp. 1305–1312.

[203] S. Zhang, H. Yao, X. Sun, and S. Liu, ‘‘Robust object tracking based onsparse representation,’’ Proc. SPIE, vol. 7744, pp. 77441N-1–77441N-8,Aug. 2010.

[204] T. Zhang, B. Ghanem, S. Liu, and N. Ahuja, ‘‘Robust visual tracking viamulti-task sparse learning,’’ in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 2042–2049.

[205] X. Jia, H. Lu, and M.-H. Yang, ‘‘Visual tracking via adaptive structurallocal sparse appearancemodel,’’ inProc. IEEEConf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 1822–1829.

[206] B. Liu, L. Yang, J. Huang, P. Meer, L. Gong, and C. Kulikowski, ‘‘Robustand fast collaborative tracking with two stage sparse optimization,’’ inProc. Eur. Conf. Comput. Vis. (ECCV), 2010, pp. 624–637.

[207] B. Liu, J. Huang, C. Kulikowsk, and L. Yang, ‘‘Robust tracking usinglocal sparse appearance model and K-selection,’’ in Proc. IEEE Conf.Comput. Vis. Pattern Recognit., Jun. 2011, pp. 1313–1320.

[208] C. Bao, Y. Wu, H. Ling, and H. Ji, ‘‘Real time robust L1 tracker usingaccelerated proximal gradient approach,’’ in Proc. IEEE Conf. Comput.Vis. Pattern Recognit., Jun. 2012, pp. 1830–1837.

[209] W. Zhong, H. Lu, and M.-H. Yang, ‘‘Robust object tracking via sparsity-based collaborative model,’’ in Proc. IEEE Conf. Comput. Vis. PatternRecognit., Jun. 2012, pp. 1838–1845.

[210] K. Zhang, L. Zhang, and M.-H. Yang, ‘‘Fast compressive tracking,’’IEEE Trans. Pattern Anal. Mach. Intell., vol. 36, no. 10, pp. 2002–2015,Oct. 2014.

[211] X. Lu, Y. Yuan, and P. Yan, ‘‘Robust visual tracking with discrimina-tive sparse learning,’’ Pattern Recognit., vol. 46, no. 7, pp. 1762–1771,Jul. 2013.

[212] N. Wang, J. Wang, and D.-Y. Yeung, ‘‘Online robust non-negative dictio-nary learning for visual tracking,’’ in Proc. IEEE Int. Conf. Comput. Vis.,Dec. 2013, pp. 657–664.

[213] S. Zhang, H. Yao, X. Sun, and X. Lu, ‘‘Sparse coding based visual track-ing: Review and experimental comparison,’’ Pattern Recognit., vol. 46,no. 7, pp. 1772–1788, Jul. 2013.

[214] A.W.M. Smeulders, D.M. Chu, R. Cucchiara, S. Calderara, A. Dehghan,and M. Shah, ‘‘Visual tracking: An experimental survey,’’ IEEETrans. Pattern Anal. Mach. Intell., vol. 36, no. 7, pp. 1442–1468,Jul. 2013.

[215] F. S. Samaria and A. C. Harter, ‘‘Parameterisation of a stochastic modelfor human face identification,’’ in Proc. 2nd IEEE Workshop Appl.Comput. Vis., Dec. 1994, pp. 138–142.

[216] G. B. Huang, M. Ramesh, T. Berg, and E. Learned-Miller, ‘‘Labeledfaces in the wild: A database for studying face recognition in uncon-strained environments,’’ Univ. Massachusetts, Amherst, MA, USA,Tech. Rep. 07-49, 2007.

[217] S.-J. Wang, J. Yang, M.-F. Sun, X.-J. Peng, M.-M. Sun, andC.-G. Zhou, ‘‘Sparse tensor discriminant color space for faceverification,’’ IEEE Trans. Neural Netw. Learn. Syst., vol. 23, no. 6,pp. 876–888, Jun. 2012.

[218] A. S. Georghiades, P. N. Belhumeur, and D. Kriegman, ‘‘From fewto many: Illumination cone models for face recognition under variablelighting and pose,’’ IEEE Trans. Pattern Anal. Mach. Intell., vol. 23, no. 6,pp. 643–660, Jun. 2001.

VOLUME 3, 2015 529

www.redpel.com40+917620593389

www.redpel.com40+917620593389

Page 41: A survey of sparse representation algorithms and applications

Z. Zhang et al.: Survey of Sparse Representation

[219] S. A. Nene, S. K. Nayar, and H. Murase, ‘‘Columbia object image library(COIL-20),’’ Dept. Comput. Sci., Columbia Univ., New York, NY, USA,Tech. Rep. CUCS-005-96, 1996.

[220] S. Lazebnik, C. Schmid, and J. Ponce, ‘‘Beyond bags of features: Spatialpyramid matching for recognizing natural scene categories,’’ in Proc.IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., Jun. 2006,pp. 2169–2178.

ZHENG ZHANG received the B.S. degree fromthe Henan University of Science and Technology,in 2012, and the M.S. degree from the ShenzhenGraduate School, Harbin Institute of Technology,Shenzhen, China, in 2014, where he is currentlypursuing the Ph.D. degree in computer science andtechnology. His current research interests includepattern recognition, machine learning, andcomputer vision.

YONG XU was born in Sichuan, China, in 1972.He received the B.S. and M.S. degrees from theAir Force Institute of Meteorology, China, in 1994and 1997, respectively, and the Ph.D. degree inpattern recognition and intelligence systems fromthe Nanjing University of Science and Technol-ogy, in 2005. He is currently with the ShenzhenGraduate School, Harbin Institute of Technology.His current interests include pattern recognition,biometrics, machine learning, and video analysis.

JIAN YANG received the B.S. degree inmathemat-ics from Xuzhou Normal University, in 1995, theM.S. degree in applied mathematics fromChangsha Railway University, in 1998, and thePh.D. degree from the Nanjing University ofScience and Technology (NUST), in 2002, witha focus on pattern recognition and intelligencesystems. In 2003, he was a Post-DoctoralResearcher with the University of Zaragoza.He was a Post-Doctoral Fellow with the Biomet-

rics Centre, The Hong Kong Polytechnic University, from 2004 to 2006, andthe Department of Computer Science, New Jersey Institute of Technology,from 2006 to 2007. He is currently a Professor with the School of ComputerScience and Technology, NUST. His journal papers have been cited morethan 1600 times in the ISI Web of Science, and 2800 times in GoogleScholar. He has authored over 80 scientific papers in pattern recognition andcomputer vision. His research interests include pattern recognition, computervision, and machine learning. He is currently an Associate Editor of PatternRecognition Letters and the IEEE TRANSACTIONS ON NEURAL NETWORKS AND

LEARNING SYSTEMS, respectively.

XUELONG LI (M’02–SM’07–F’12) is currently a Full Professor with theCenter for OPTical IMagery Analysis and Learning, State Key Laboratoryof Transient Optics and Photonics, Xi’an Institute of Optics and PrecisionMechanics, Chinese Academy of Sciences, Shaanxi, P.R. China.

DAVID ZHANG (F’08) received the degree incomputer science from Peking University, theM.Sc. degree in computer science in 1982,the Ph.D. degree from the Harbin Institute ofTechnology, in 1985, and the Ph.D. degree inelectrical and computer engineering from theUniversity of Waterloo, ON, Canada, in 1994.From 1986 to 1988, he was a Post-Doctoral Fellowwith Tsinghua University, and then an AssociateProfessor with Academia Sinica, Beijing. He is

currently the Chair Professor with The Hong Kong Polytechnic Univer-sity, where he was the Founding Director of the Biometrics TechnologyCentre supported by the Hong Kong SAR Government in 1998. He servesas the Visiting Chair Professor with Tsinghua University, and an AdjunctProfessor with Shanghai Jiao Tong University, Peking University, the HarbinInstitute of Technology, and the University of Waterloo. He has authoredover 10 books and 200 journal papers. He is a Croucher Senior ResearchFellow, a Distinguished Speaker of the IEEE Computer Society, and a fellowof the International Association for Pattern Recognition. He is the Founderand Editor-in-Chief of the International Journal of Image and Graphics,a Book Editor of International Series on Biometrics (Springer), an Orga-nizer of the first International Conference on Biometrics Authentication, anAssociate Editor of more than ten international journals, including the IEEETRANSACTIONS and Pattern Recognition, and the Technical Committee Chairof the IEEE CIS.

530 VOLUME 3, 2015

www.redpel.com41+917620593389

www.redpel.com41+917620593389