首页 Total generalized variation

Total generalized variation

举报
开通vip

Total generalized variation Total Generalized Variation Kristian Bredies Karl Kunisch Thomas Pock May 14, 2010 Abstract The novel concept of total generalized variation of a function u is introduced and some of its essential properties are proved. Differ- ently from the bounded vari...

Total generalized variation
Total Generalized Variation Kristian Bredies Karl Kunisch Thomas Pock May 14, 2010 Abstract The novel concept of total generalized variation of a function u is introduced and some of its essential properties are proved. Differ- ently from the bounded variation semi-norm, the new concept involves higher order derivatives of u. Numerical examples illustrate the high quality of this functional as a regularization term for mathematical imaging problems. In particular this functional selectively regularizes on different regularity levels and, as a side effect, does not lead to a staircasing effect. Keywords: Bounded variation, total generalized variation, tensor fields, regularization, image denoising. AMS Subject Classification: 49J52, 49N45, 68U10. 1 Introduction Most mathematical formulations of inverse problems and in particular of mathematical imaging problems are cast in the form min u F(u) +R(u), (1.1) where F represents the data fidelity and R the regularization term. If G denotes the forward modeling operator then the most common fidelity term is of the form F(u) = 1 2 ‖G(u)− z‖2, (1.2) where z stands for the possibly error-prone data and ‖ · ‖ denotes an ap- propriately chosen Hilbertian norm. Similarly the most frequently chosen regularization term is given by R(u) = α 2 |u|2, (1.3) where α is the regularization parameter and | · | again denotes a Hilbertian norm or semi-norm. It is now becoming well-accepted that the mathematical 1 and computational simplicity of the norm-of-squares terms must be put into perspective with some serious shortcomings. If the errors in the data contain outliers or if the error is of impulsive type the fidelity terms suggested by methods from robust statistics should be preferred over (1.2). Similarly (1.3) is frequently not an appropriate choice. In fact, the regularization term penalizes a certain property of u, which is quantified in the choice of the norm | · |, and the natural proportionality by which this should enter into R would be 1-homogeneous rather than quadratic. In the present paper the focus will be on the choice of R. One of the early proposal for a refined choice was given in [ROF]. It uses the bounded variation semi-norm R(u) = α ∫ Ω |Du|, (1.4) where u is defined on the bounded domain Ω ⊂ IRd. This choice is highly effective when compared to e.g. R(u) = α2 ∫ Ω |∇u|2 dx if the data z to be reconstructed are piecewise constant, since it is more apt to preserve corners and edges. The bounded variation semi-norm, however, also has some shortcomings, most notably the staircasing phenomenon. To briefly explain this effect, we assume that G = I so that (1.1) describes the imaging denoising problem. If the true image contains not only flat, but also slanted regions, then the image reconstructed on the basis of bounded variation semi-norm tends to be piecewise constant (staircasing). This staircasing effect was rigorously established in [Ni, CCN, R], for example. For diverse other aspects on the topic of constructing appropriate regularization or filter functionals in image reconstruction we refer to [SGGHL] and the references given there. In this paper we propose and analyze the regularization term of the form TGVkα(u) = sup {∫ Ω udivk v dx ∣∣∣ v ∈ Ckc (Ω, Symk(IRd)), ‖divl v‖∞ ≤ αl, l = 0, . . . , k − 1 } , (1.5) where Symk(IRd) denotes the space of symmetric tensors of order k with arguments in IRd, and αl are fixed positive parameters. For the definition of the remaining quantities, we ask for the readers’ patience until Section 2. Suffice it to say at this moment that for k = 1, α0 = 1 the semi-norm TGVkα coincides with the bounded variation semi-norm. We refer to TGV k α as total generalized bounded variation of order k with weight α ∈ IRk. From the definition of TGVkα it is immediately clear that it involves (generalized) derivatives of u of order i = 1, . . . , k, and that the kernel of TGVkα is the set of polynomials of degree less than k. Intuitively the total generalized bounded variation further automatically balances the first to the k-th derivatives of u among themselves. It will be shown that TGVkα shares some properties 2 of TV: It is also rotationally invariant and for k = 2, the total generalized variation of the indicator function of a smooth set Ω′ ⊂⊂ Ω equals α1 PerΩ′ = α1 TV(χΩ′), where PerΩ′ denotes the perimeter of Ω′. If differs, however, for functions which are not piecewise constant. As a further preview we point out that in dimension 1, with Ω = ]0, L[, k = 2, α0, α1 > 0 we have for u(x) = 2∑ i=1 pi(x)χΩi , with pi(x) = aix+ bi, for each a1, a2, b1, b2 ∈ IR and Ω1 = ]0, c[, Ω2 = ]c, 1[ that TGV2α(u) = α1|p2(c)− p1(c)|+ α0|p′1(c)− p′2(c)| provided that the jump point c is not near the boundary, i.e., c ∈ [α0/α1, L− α0/α1]. In particular, TGV2α(u) does not penalize the derivative of order l = 0, 1 unless it jumps at c. As already mentioned our motivation for studying TGV2α(u) is based on the fact that it involves and balances higher-order derivatives of u. As a consequence it reduces the staircasing effect of the bounded variation func- tional. This will be demonstrated in our numerical experiments. The use of higher-order derivatives with the aim of reducing staircasing is not new. In [CL] the inf-convolution functional min u1+u2=u ∫ Ω |∇u1|+ α|∇(∇u2)| dx was proposed and proved to be practically efficient, eliminating the stair- casing effect, for denoising problems with images which contain various grey levels as well as edges and corners. This idea was followed upon in a modified form in [CEP] where the regularization term is of the form min u1+u2=u ∫ Ω |∇u1|+ α|∆u2| dx, i.e. the second derivative is replaced by the Laplacian and a dual method for its numerical realization is derived. A different functional was proposed and tested in [CMM]. It is given by∫ Ω |∇u|+ αΦ(|∇u|)(L(u))2 dx, where Φ is a real-valued function that reflects the presence of edges in the sense that its value approaches 0 when the gradient |∇(u)| is large, and L(u) is an elliptic operator. For this choice of regularization functional the ab- sence of the staircasing effect was verified in [DFLM]. In [PS] regularization terms of the form R(u) = ∫ Ω |D∇l−1(u)|, 3 where ∫ Ω |D∇l−1(u)| denotes the total variation of the (l − 1)-th derivative of u ∈ W l−1,1, are considered, and special structure of minimizers of the resulting problems (1.1) are investigated. Higher-order regularization func- tionals in the discrete setting which are related to our computations for the second-order case were further proposed and tested in [SS]. As we shall see in Section 3, even for the case k = 2 the proposed functional TGVkα does not agree with those regularization functionals which were considered earlier in the literature. In particular, although convex, it is structurally different from any infimal-convolution functional, especially the approaches of [CL, CEP]. Let us give a briefly outline of the following sections. Section 2 contains a compact treatise of tensor and, in particular, symmetric tensor analysis in a manner that is useful for the variational analysis context. The defi- nition of total generalized variation norms and some of its basic properties are given in Section 3. Based on Fenchel duality an equivalent description of TGVkα(u) is derived for u sufficiently regular. Moreover the relationship to an appropriately defined k-fold inf-convolution is obtained. A subsection is devoted to the special case k = 2. The description of the numerical pro- cedure that was employed as well as carefully selected numerical denoising experiments are contained in Section 4. The Appendix contains some basic results involving symmetric k-tensor fields. 2 Spaces of symmetric tensor fields This section is mainly devoted to the introduction of the notions we are going to utilize in the main parts of this article. For many of the considerations which are going to follow, the concept of symmetric tensor fields plays a central role. Therefore, we give a rather extensive introduction to make this paper more self-contained and also for the convenience for those readers, who are familiar with tensor analysis. Most of the results, although scattered in several chapters, can also be found in the books [BG, S]. We mainly restrict ourselves to a general introduction of symmetric ten- sor fields. Some more specific results can be found in the appendix. Throughout the paper, d ≥ 1 denotes the dimension which is typically 2 or 3, in applications. Let T k(IRd) = {ξ : IRd × · · · × IRd︸ ︷︷ ︸ k-times → IR ∣∣ ξ k-linear} Symk(IRd) = {ξ : IRd × · · · × IRd︸ ︷︷ ︸ k-times → IR ∣∣ ξ k-linear and symmetric} be the vector space of k-tensors and symmetric k-tensors, respectively (ac- tually, these are spaces of (0, k)-tensors, but since we only deal with co- variant vectors, we omit the 0). Here ξ ∈ T k(IRd) is called symmetric, 4 if ξ(a1, . . . , ak) = ξ(api(1), . . . , api(k)) for all pi ∈ Sk, where Sk denotes the permutation group of {1, . . . , k}. The case k = 0 corresponds to scalar values, for k = 1, Sym(IRd, IR) = IRd while for k = 2, Sym2(IRd) = Sd×d, i.e. the space corresponds to sym- metric matrices. Note three basic operations for k-tensors. For ξ ∈ T k(IRd) and η ∈ T l(IRd) the tensor product (ξ ⊗ η)(a1, . . . , ak+l) = ξ(a1, . . . , ak)η(ak+1, . . . , ak+l) yields an element of T k+l(IRd). We define the trace tr(ξ) ∈ T k−2(IRd) of ξ ∈ T k(IRd) with k ≥ 2 by tr(ξ)(a1, . . . , ak−2) = d∑ i=1 ξ(ei, a1, . . . , ak−2, ei) where ei denotes the i-th standard basis vector. This operation can be iterated, for example trl(ξ⊗η) for ξ ∈ Symk+l(IRd) and η ∈ Syml(IRd) yields a symmetric k-tensor. Every k-tensor ξ ∈ T k(IRd) can be symmetrized by (9ξ)(a1, . . . , ak) = 1 k! ∑ pi∈Sk ξ(api(1), . . . , api(k)). The symmetrization is a projection, i.e. 92ξ = 9ξ. The spaces T k(IRd) and, consequently, Symk(IRd) will be equipped with the scalar product ξ · η = trk(ξ¯ ⊗ η) = ∑ p∈{1,...,d}k ξ(ep1 , . . . , epk)η(ep1 , . . . , epk), ξ¯(a1, . . . , ak) = ξ(ak, . . . , a1), leading canonically to the norm |ξ| = √ξ · ξ. Again, this is the absolute value for k = 0, for k = 1, this corresponds to the Euclidean norm in IRd and in case k = 2, we can identify ξ ∈ Sym2(IRd) with ξ = ξ11 · · · ξ1d... . . . ... ξ1d · · · ξdd  , |ξ| = ( d∑ i=1 ξ2ii + 2 ∑ i 0. Then, the total generalized variation of order k with weight α for u ∈ L1loc(Ω) is defined as the value of the functional TGVkα(u) = sup {∫ Ω udivk v dx ∣∣∣ v ∈ Ckc (Ω, Symk(IRd)), ‖divl v‖∞ ≤ αl, l = 0, . . . , k − 1 } (3.1) with taking the value ∞ if the respective set is unbounded from above. The space BGVkα(Ω) = {u ∈ L1(Ω) ∣∣ TGVkα(u) <∞} , ‖u‖BGVkα = ‖u‖1 + TGVkα(u) is called the space of functions of bounded generalized variation of order k with weight α. Remark 3.2. For k = 1 and α > 0, we see that TGV1α(u) = α sup {∫ Ω udiv v dx ∣∣∣ v ∈ C1c (Ω, Sym1(IRd)), ‖v‖∞ ≤ 1} = αTV(u). Thus one can indeed speak of a generalization of the total variation. In the following, we will derive some basic properties of the total gener- alized variation. Proposition 3.3. The following statements hold: 1. TGVkα is a semi-norm on the normed space BGV k α(Ω), 10 2. for u ∈ L1loc(Ω), TGVkα(u) = 0 if and only if u is a polynomial of degree less than k, 3. for fixed k and positive weights α, α˜ ∈ IRk, the semi-norms TGVkα and TGVkα˜ are equivalent, 4. TGVkα is rotationally invariant, i.e. for each orthonormal matrix O ∈ IRd×d and u ∈ BGVkα(Ω) we have that u˜ ∈ BGVkα(OTΩ) with TGVkα(u˜) = TGVkα(u), where u˜(x) = u(Ox), 5. for r > 0 and u ∈ BGVkα(Ω), we have, defining u˜(x) = u(rx), that u˜ ∈ BGVkα(r−1Ω) with TGVkα(u˜) = r −d TGVkα˜(u) , α˜l = αlr k−l. Proof. Let us begin proving the first statement. Note that TGV can be interpreted as the dual semi-norm in which the set Kkα(Ω) = {divk v ∣∣ v ∈ Ckc (Ω, Symk(IRd)), ‖divl v‖∞ ≤ αl, l = 0, . . . , k − 1} (3.2) is taking the role of the “predual unit ball”: TGV(u) = sup w∈Kkα ∫ Ω uw dx. (3.3) It is easy to see that Kkα(Ω) is balanced and convex. The former implies that TGVkα is positively one-homogeneous
本文档为【Total generalized variation】,请使用软件OFFICE或WPS软件打开。作品中的文字与图均可以修改和编辑, 图片更改请在作品中右键图片并更换,文字修改请直接点击文字进行修改,也可以新增和删除文档中的内容。
该文档来自用户分享,如有侵权行为请发邮件ishare@vip.sina.com联系网站客服,我们会及时删除。
[版权声明] 本站所有资料为用户分享产生,若发现您的权利被侵害,请联系客服邮件isharekefu@iask.cn,我们尽快处理。
本作品所展示的图片、画像、字体、音乐的版权可能需版权方额外授权,请谨慎使用。
网站提供的党政主题相关内容(国旗、国徽、党徽..)目的在于配合国家政策宣传,仅限个人学习分享使用,禁止用于任何广告和商用目的。
下载需要: 免费 已有0 人下载
最新资料
资料动态
专题动态
is_530019
暂无简介~
格式:pdf
大小:3MB
软件:PDF阅读器
页数:42
分类:
上传时间:2013-11-02
浏览量:121