11.07.2015 Views

A Modified SPIHT Algorithm for Image Coding With a Joint MSE and ...

A Modified SPIHT Algorithm for Image Coding With a Joint MSE and ...

A Modified SPIHT Algorithm for Image Coding With a Joint MSE and ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

CHANG AND CARIN: MODIFIED <strong>SPIHT</strong> ALGORITHM FOR IMAGE CODING 715tree-structured <strong>and</strong> Markovian, tied to the wavelet quadtree, employinga state transition matrix to quantify the degree of persistencebetween scales. The iterative EM algorithm <strong>for</strong> HMTs[4] was proposed to train the parameters (the mixture densityparameters <strong>and</strong> the probabilistic graph transition probabilities)to match the data in the maximum likelihood (ML) sense. Thetrained HMT provides a good approximation to the joint probabilityof the wavelet coefficients, yielding good classificationper<strong>for</strong>mance.The conventional HMT training process requires the availabilityof labeled imagery. Specifically, data must be provided<strong>for</strong> each texture class, followed by HMT training. The trainedHMTs can then be applied to segment new imagery that mightbe observed, assuming that this new imagery is characterized bysimilar textures. In our problem we assume little or no knowledgeof the anticipated image textural properties, <strong>and</strong>, there<strong>for</strong>e,determination of the textural classes is determined jointly withHMT training.B. HMT Mixture ModelWe assume that the statistics of the wavelet coefficients fromthe overall image may be represented as a mixture of HMTs,analogous to the well-known Gaussian mixture model (GMM)[1]. The probabilistic model <strong>for</strong> mixture components is givenbywhere are the wavelet coefficients of a wavelet tree, ,are the mixing coefficients of the textures,which may also be interpreted as the prior probabilities, <strong>and</strong> represents the parameters of the th HMTcomponent. The vector represents the cumulative set of modelparameters, specifically <strong>and</strong> , .We employ an iterative training procedure, analogous to thatfound in GMM design [1]. Let <strong>and</strong> represent themodel parameters <strong>for</strong> mixture component after iteration .Weestimate the probability that the th wavelet tree is generated bytexture (corresponding to the th HMT, denoted )as(1)where, <strong>for</strong> notational simplicity, we define. Equation (3) represents an approximation tothe conditional expectationusing. The samples that are associated withtexture with higher likelihood make a greater contribution tothe parameters of that texture component. The cumulative setof HMT parameters , e.g., state-transition probabilities,state-dependent parameters, etc., define the overall set of parameters<strong>for</strong> the th HMT.The mixing coefficients are updated aswhere we have again assumed .For initialization, we use all the data with equal probabilityto train an initial (single) HMT with parameters. We then cluster the data into (scalar) Gaussian mixtures(denoting textures) based on , assumingthat the data from the same texture have similar probabilityvalues. In this manner, we assign the initial probability. The same ideahas been applied to effectively use unlabeled sequential datain learning hidden Markov models [10], with this termed theextended Baum–Welch (EBW) algorithm.We segment the image via a maximum a posteriori (MAP)estimator, that is(5)where represents the HMT model parameters <strong>for</strong> mixturecomponent , after convergence is achieved <strong>for</strong> the a<strong>for</strong>ementionedtraining algorithm. The parameter , representing thenumber of textures in the image, is selected autonomously via anin<strong>for</strong>mation-theoretic model-selection method called the MDLprinciple, derived by Rissanen [18]. The MDL principle statesthat the best model is the one that minimizes the summed descriptionlength of the model <strong>and</strong> the likelihood of the data withrespect to the model, making a trade-off between the model accuracy<strong>and</strong> model succinctness [26]. In our case, we calculatethe value as(4)(2)The parameters of each HMT are updated by an augmented<strong>for</strong>m of the EM algorithm in [4]. In [4], the wavelettreesare each used separately within an “upward-downward”algorithm to update the parameters of eachindividual HMT. On iteration , HMT model parametersare initiated using parameters from the previous step. Letrepresent an arbitrary parameter from ,soupdated using wavelet tree . Then, the associated cumulativeparameter, based on all wavelet trees, is expressed as(3)where the first term denotes the accuracy of the mixturemodel, the second term reflects the model complexity, <strong>and</strong>is the number of the free parameters estimated. We chooseto minimize (6).III. QUANTIZATION BINSThe purpose of rescaling the wavelet coefficients is to helpthe encoder order the output bit stream with consideration ofthe ultimate recognition task, with the balance between <strong>MSE</strong><strong>and</strong> segmentation per<strong>for</strong>mance driven by a Lagrangian metric.Wavelet coefficients that play an important role in defining thesegmented texture class labels (determined automatically, asdiscussed in Section II) should be represented by a relatively(6)

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!