[精品论文]Shape Recoverya Generalized Topology.doc
精品论文Shape Recovery by a Generalized TopologyPreserving SOMDong Huang and Zhang YiComputational Intelligence Laboratory, School of Computer Science andEngineering, University of Electronic Science and Technology of China, Chengdu610054, P. R. China.E-mail: donnyhuang, zhangyi.AbstractThis paper proposes a new deformable model, i.e., gTPSOM, for object shape re- covery. Inspired by ViSOM and Region-Aided Active Contour, the proposed model is formulated as generalized chain SOM with an adaptive force field. The adaptive force field is adjusted during the evolvement of the neuron chain according to local consistency of the image edge map. With the topology preserving property inherited from the data mapping model, i.e. ViSOM, the proposed model is suitable for both the precise edge detection and the complex shape recovery with boundary strength variations. Detailed formulation and analysis of the proposed model are given. Ex- periments on both synthesis and real images are carried out to demonstrate the performances.Key words: Shape Recovery; Topology Preserving Mapping; ViSOM; Region-Aided Active Contour1 IntroductionAn important goal in computer vision is to recover the objects shape of interest from visual data. Deformable models originated in the work of 1 and the 3D case for 2, have been extensively used in shape recovery and medical imaging 34. Their applications also include geometric modeling 5, computer animation 6, texture segmentation 7 and object tracking.1 This work was supported by National Science Foundation of China under Grant60471055 and Specialized Research Fund for the Doctoral Program of Higher Edu- cation under Grant 20040614017.Preprint submitted to Elsevier Preprint31 October 20071The most popular deformable model Snake or Active Contour model 1, describes a closed parametric curve that deforms dynamically and moves to- wards the desired image features under the influence of internal and external forces. The internal forces keep the contour smooth, while the external forces attract the snake towards lines, edges, or other low-level image features. The geodesic active contour 8 significantly improve the parametric snake by nat- urally handling topological changes. However, it still suffers from drawbacks such as edge leakage and sensitivity to initialization. The Gradient Vector Flow (GVF) snake 9 uses a bi-directional external force field that provides long-range capture of object boundaries from either side. The main drawback of GVF however is that the contour does not propagate where the vector flows are the saddle points or divergence points within a neighborhood. Therefore, their contours can only avoid getting trapped at these points by proper ini- tialization.There are also many works 1213 that integrate the conventional SOM and Snake. In these models, shape recovery is regarded as data mapping from the edges to the chain SOMs. Unlike the classical SOMs that read the input data in random sequences and adjust the network structure over space, in these models, SOM processes the whole input in parallel and organizes itself over time 14. However, to ensure the proper evolvement to the object boundaries, a number of “Batch” updating rules and parameters needs to be determined manually. Thus their performances are limited.Another framework of deformable model is based on charged particle dy- namics. In the Charged Particle Model (CPM) 15, the charges are attracted towards the objects contours of interest by an electric field computed using the edge magnitude. The electric field plays the same role as the external force in the snake model, while internal interactions are modelled by repulsive elec- trostatic forces, referred to as Coulomb forces. CPM is extremely flexible and greatly relieve the initialization problem. However, it is not suitable for the cases where continuous and closed final contours are required.Recently, Yang et al. proposed the Charged Active Contour based on Elec- trostatics (CACE) 18 that incorporate both the snake and particle based models CPM. CACE adaptively change the external force field with the prop- agation of the active contour by introducing a competition part. This CACE successfully move across the saddle points and divergence points. But this ability depends on parameters chosen according to the local edge magnitude. For this reason, CACE has difficulties in dealing with images of variant edge strength and complex shapes.In order to overcome the drawbacks in the deformable model reviewed above, we propose a deformable model by incorporating the Topology Pre- serving Self-Organizing Mapping into the neuron competition. We call this model the generalized Topology Preserving SOM (gTPSOM). It is inspired by the Visual Induced Self-Organizing Map (ViSOM) 11 where the mapping pre- serves the inter-point distances of the input data on the neuron map as well5as the topology. Following the ideas in 121314, the gTPSOM is driven in parallel by an adaptive force field, which imposes constrains on the local boundary variation. Region aided active contour and Level sets are employed to implement the proposed model. The gTPSOM model is suitable for both the precise edge detection and the complex shape recovery with boundary strength variation. Detailed formulation and analysis of the proposed model are presented. Experiments on both synthesis and real images are carried out to illustrate the performances.The rest of this paper is organized as follows. Section 2 gives detailed formulation of the adaptive force field staring with the self-organizing of a neuron chain. Then the proposed model is formulated as Active Contour and Level Sets. Relations between our model and ViSOM and CACE are also given. Section 3 presents the experimental results and discussions. Finally, the paper is concluded in Section 4.2 FormulationDenote I the input image of N pixels. The edge map of the image can be computed using either gradient operator I or Gaussian-based edge detector. We use the gradient operator throughout the following presentation. Our de- formable model is first formulated as a closed neuron chain (Fig. 1 (a) driven by an adaptive vector field. The neurons are attracted to the nearest edges while compete for these edges. For the efficiency of numerical computation,our model is then translated in to a Region-Aided Active Contour and LevelSets Formulation.2.1 Self-Organizing of the Neuron ChainConsider the edge map of the input image as a data set X = xp R2, p =1, · · · , P with all data points located in the pixels grids. The edge magnitudef (ri ) in each pixel ri (i = 1, · · · , N ) can be regarded as the data density ofthe data set, where ri = rx, ry T R2 is the 2-D coordinates of the pixel.k kOur model is designed to map the data distribution of the edge map to adeformable neuron chain (Fig. 1 (a) with M uniformly distributed neurons. In the 2-D image coordinate, the neurons in the chain SOM correspond to the control points of a smooth curve. The weight vector of the kth neuron wk = wx, wy T are used to represent the coordinate of the kth control point.k kBy repeatedly updating the weight vectors, the chain SOM approximatesthe two-dimensional input distribution by a one-dimensional neural network. In this way, the control points move toward the pixels of strong object bound- aries. Similar to the classic SOM, the weight vector of the kth neuron can be updated by moving the neuron along the force vector formulated as the difference between the input data point xp and the weight vector wk :ry Fxx vpvpFx kFpvkwx,wyT krx(a)(b)Fig. 1. (a) The deformable neuron chain with weight vector wk = wx, wy T in thekk2-D image coordinate; (b) The vector graph of the neuron updating.4wk = · · (xp wk ),where and are the learning rate and the neighborhood function respec- tively. For each input data point, the updating scale for the closer neuron (the winner) is larger than the neurons faraway (the neighbors of the winner). The neuron grids of the classic SOMs spread uniformly to the data space using the neighborhood function based on the Euclidean distances. In the proposed chain SOM, similar mechanism is used. The updating scale for the neighbors are reduced along the tangent direction of the chain. Thus to maintain the spatial structure of the chain SOM, neurons are “repelled” by theirs neighbors.For each data point xp (p = 1, · · · , N ), there is a winning neuron wv on the neuron chain that is nearest to it. Then the updating force from the kth neuron to the input data Fxp k (k = 1, · · · , M ) can be decomposed into two parts (See Fig. 1 (b) ):Fxp k = Fxp v + Fvk .The first force Fxp v , represents the updating force from the winner to the input data. While the second force Fvk is a lateral force between the neuron k to the winner, i.e., a contraction force. This contraction force brings neuronsin the neighborhood toward the winner, and thus forms a contraction around the winner on the chain. This is an unfavorable effect for shape recovery. If the neurons clutch to the stronger parts of the boundary while left the weaker parts empty, the boundary topology of the objects in the image is distorted.The proposed generalized Topology Preserving SOM (gTPSOM) focuses on this problem. To preserve the uniform boundary topology, the lateral con- traction force Fvk is constrained. The scale of the competition is controlled by a function of the mapping from input data space to the two neurons. Denote this function by h(wk , wvxp ). The updating force for the kth neuron presented with the data point xp is then:vkFxp k = (xp wk ) (wv wk )h(wk , wxp v )= Favkxp k+ F r ,xp kwhere F ais the attraction force and F ris the repelling force. The functionh(wk, wvxp) needs to be formulated so that the neuron chain can spread to theregions of continuous data distribution (the closed boundaries in the image),while cling to the dense regions that has already been occupied.2.2 The Adaptive Force FieldThe Classical SOMs read the input data in random but sequential or- der and adjust the neural networks over multi-passes. In contrast, to achieve smooth propagation of the neuron chain, we implement “Batch” updating 12 13 by generating the joint force vector field F = Fx, Fy 0 at each pixel location of the input image I .Here, we treat the neurons on the deformable chain as positive charges 15. The attraction force generated by the edge pixels ri (i = 1, · · · , N ) can be formulated by treating the edge pixels as fixed negative charges of magnitude f (ri ). The joint force field F is composed of the attraction force F a to the edge data points and repelling force F r between neurons.Then according to Coulombs Law, the attracting force field at the kthneuron is computed by summing the force vectors from all edge data point ri :Na X f (ri )wk riF =i=1·40|wk, ri |3where 0 is the permittivity of free space. Note the force vector at wk generated by ri is direct proportionate to the data density f (ri ), and inverse proportion- ate to the square of distance between ri and wk . The attraction force field is pre-computed and fixed through time. Here the ”Coulomb-Law” is not the only option. Any isotropic measure can be used if the resulted force vectors de- crease smoothly with the increasing of distance. Other natural choices include inverse of the squared distance and the Gaussian function.Next, we compute the repelling part of the force field. Let f (ri ) (i =1, · · · , N ) be the data density in the domain of the edge map, and f (wk ) (k = 1, · · · , M ) for the neuron domain. If f (wj ) > 0, the jth neuron is on thepotential edge pixel ri (ri = wj ). The neuron j can be regarded as the winner for data at ri , and the neuron k (wk = ri , k = 1, · · · , M ) is the neighbor. The repelling regularization is realized by lettingf (wj ) f (wk )h(wk , wj ) =, (1)f (wj )where control the resolution of the regularization. Note that the distance or spacing among data points in the input space is inverse proportional to thedensity. Following from the ideas in ViSOM, the function h(wk , wj ) is related to the density distribution in both domains. Both ViSOM and gTPSOM con- strain the lateral contraction force according to the desired proportion of the distance in the input space and that represented by the neuron network.If wk and wj are in regions of approximately the same data density, i.e. f (wk ) = f (wj ), the contraction force between them is less constrained. The attraction force field F a will lead both k and j to denser regions, i.e., stronger edges in the vicinity. On the other hand, If the data density at location of thekth neuron is lower than that of the jth neuron, that is f (wk ) < f (wj ), the kth neuron is on the relatively weaker edges or in the homogeneous regions. The kth neuron moves less to the denser region that has already been occupied by the jth neuron. An opposite force is applied to repel the kth neuron away from the jth neuron. This actually weaken the attraction force of the data points at ri , where ri = wj .Then, In the “Batch” updating paradigm, the repelling force field at the kth neuron are computed by summing the force vectors from all other winner neurons (wj = ri ) on the deformable chain. The neighborhood concept is based on the square of distance between wj and wk .M "f (wk ) # f (wj )wk wjr XFk =j=11f (wj )·40|wk. wj |3as:Finally, the joint updating force Fk at the kth neuron can be computedFk = F a + F rk kNM " #= X f (ri )wk riXf (wk )+1 f (wj )wk wj. (2)i=140 · |wk ri |3j=1f (wj )40 · |wk wj |3Note the joint updating force is related to the locations of neurons on the deformable chain. When the neuron chain is propa