Top Banner
Future Generation Computer Systems 15 (1999) 223–243 Wavelet-based medical image compression Eleftherios Kofidis 1, * , Nicholas Kolokotronis, Aliki Vassilarakou, Sergios Theodoridis, Dionisis Cavouras Department of Informatics, Division of Communications and Signal Processing, University of Athens, Panepistimioupolis, TYPA Buildings, GR-15784 Athens, Greece Abstract In view of the increasingly important role played by digital medical imaging in modern health care and the consequent blow up in the amount of image data that have to be economically stored and/or transmitted, the need for the development of image compression systems that combine high compression performance and preservation of critical information is ever growing. A powerful compression scheme that is based on the state-of-the-art in wavelet-based compression is presented in this paper. Compression is achieved via efficient encoding of wavelet zerotrees (with the embedded zerotree wavelet (EZW) algorithm) and subsequent entropy coding. The performance of the basic version of EZW is improved upon by a simple, yet effective, way of a more accurate estimation of the centroids of the quantization intervals, at a negligible cost in side information. Regarding the entropy coding stage, a novel RLE-based coder is proposed that proves to be much simpler and faster yet only slightly worse than context-dependent adaptive arithmetic coding. A useful and flexible compromise between the need for high compression and the requirement for preservation of selected regions of interest is provided through two intelligent, yet simple, ways of achieving the so-called selective compression. The use of the lifting scheme in achieving compression that is guaranteed to be lossless in the presence of numerical inaccuracies is being investigated with interesting preliminary results. Experimental results are presented that verify the superiority of our scheme over conventional block transform coding techniques (JPEG) with respect to both objective and subjective criteria. The high potential of our scheme for progressive transmission, where the regions of interest are given the highest priority, is also demonstrated. c 1999 Elsevier Science B.V. All rights reserved. Keywords: Medical imaging; Image compression; Wavelet transform; Lifting; Reversible transforms; Wavelet zerotrees; Entropy coding; Selective compression; Progressive transmission 1. Introduction Medical imaging (MI) [20,37] plays a major role in contemporary health care, both as a tool in primary diagnosis and as a guide for surgical and therapeutic procedures. The trend in MI is increasingly digital. * Corresponding author. 1 Present address: D´ epartement Signal et Image, Institut National des T´ el´ ecommunications, 9 rue Charles Fourier, F-91011 ´ Evry cedex, France. This is mainly motivated by the advantages of digital storage and communication technology. Digital data can be easily archived, stored and retrieved quickly and reliably, and used in more than one locations at a time. Furthermore digital data do not suffer from aging and moreover are suited to image postprocessing op- erations [11,57]. The rapid evolution in both research and clinical practice of picture archiving and commu- nication systems (PACS) [22,40] and digital teleradi- ology [18,27] applications, along with the existence 0167-739X/99/$ – see front matter c 1999 Elsevier Science B.V. All rights reserved. PII:S0167-739X(98)00066-1
21

Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

Aug 19, 2018

Download

Documents

vuonghanh
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

Future Generation Computer Systems 15 (1999) 223–243

Wavelet-based medical image compression

Eleftherios Kofidis1,∗, Nicholas Kolokotronis, Aliki Vassilarakou, Sergios Theodoridis,Dionisis Cavouras

Department of Informatics, Division of Communications and Signal Processing, University of Athens, Panepistimioupolis, TYPA Buildings,GR-15784 Athens, Greece

Abstract

In view of the increasingly important role played by digital medical imaging in modern health care and the consequent blowup in the amount of image data that have to be economically stored and/or transmitted, the need for the development of imagecompression systems that combine high compression performance and preservation of critical information is ever growing.A powerful compression scheme that is based on the state-of-the-art in wavelet-based compression is presented in this paper.Compression is achieved via efficient encoding of wavelet zerotrees (with the embedded zerotree wavelet (EZW) algorithm)and subsequent entropy coding. The performance of the basic version of EZW is improved upon by a simple, yet effective,way of a more accurate estimation of the centroids of the quantization intervals, at a negligible cost in side information.Regarding the entropy coding stage, a novel RLE-based coder is proposed that proves to be much simpler and faster yet onlyslightly worse than context-dependent adaptive arithmetic coding. A useful and flexible compromise between the need forhigh compression and the requirement for preservation of selected regions of interest is provided through two intelligent,yet simple, ways of achieving the so-called selective compression. The use of the lifting scheme in achieving compressionthat is guaranteed to be lossless in the presence of numerical inaccuracies is being investigated with interesting preliminaryresults. Experimental results are presented that verify the superiority of our scheme over conventional block transform codingtechniques (JPEG) with respect to both objective and subjective criteria. The high potential of our scheme for progressivetransmission, where the regions of interest are given the highest priority, is also demonstrated.c©1999 Elsevier Science B.V.All rights reserved.

Keywords:Medical imaging; Image compression; Wavelet transform; Lifting; Reversible transforms; Wavelet zerotrees; Entropy coding;Selective compression; Progressive transmission

1. Introduction

Medical imaging (MI) [20,37] plays a major role incontemporary health care, both as a tool in primarydiagnosis and as a guide for surgical and therapeuticprocedures. The trend in MI is increasingly digital.

∗ Corresponding author.1 Present address: Departement Signal et Image, Institut National

des Telecommunications, 9 rue Charles Fourier, F-91011Evrycedex, France.

This is mainly motivated by the advantages of digitalstorage and communication technology. Digital datacan be easily archived, stored and retrieved quicklyand reliably, and used in more than one locations at atime. Furthermore digital data do not suffer from agingand moreover are suited to image postprocessing op-erations [11,57]. The rapid evolution in both researchand clinical practice of picture archiving and commu-nication systems (PACS) [22,40] and digital teleradi-ology [18,27] applications, along with the existence

0167-739X/99/$ – see front matterc©1999 Elsevier Science B.V. All rights reserved.PII: S0167-739X(98)00066-1

Page 2: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

224 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

of standards for the exchange of digital information(e.g., ACR/NEMA [2,36], DICOM [17]) further pro-mote digital MI.

However, the volume of image data generated andmaintained in a computerized radiology department,by far exceeds the capacity of even the most recent de-velopments in digital storage and communication me-dia. We are talking about Terabytes of data producedevery year in a medium-sized hospital. Moreover, evenif the current disk and communication technologiessuffice to cover the needs for the time being, the costis too large to be afforded by every medical care unitand it cannot provide a guarantee for the near futurein view of the increasing use of digital MI.Imagecompressioncomes up as a particularly cost-effectivesolution to the data management problem by reduc-ing the size of the images for economic representationand transmission, keeping at the same time as muchof the diagnostic relevant information as possible. Theunique nature of medical imagery, namely, the highimportance of the information it contains along withthe need for its fast and reliable processing and trans-mission, imposes particularly stringent requirementsfor an image compression scheme to be acceptable forthis kind of application:– No loss of relevant information is allowed, though

the term ‘relevant’ can take different meanings de-pending on the specific task. At present, the use of‘lossy’ compression, i.e. with the reconstructed im-age being different from the original, is very limitedin clinical practice, especially in primary diagnosis,due to the skepticism of the physicians about eventhe slightest data loss, which might, even in the-ory, induce some critical information loss. However,efficient compression schemes have been derivedwhich have shown satisfying behavior in preserv-ing features (e.g., edges) that are critical in medicaldiagnosis and interpretation [1,11,13,33,46]. More-over, there is a constant increase in the productionof lossy imaging systems, though at a lower percentas compared to the lossless ones, as reported in [69].

– Progressive coding is a very desirable feature in amedical compression scheme, since it allows thetransmission of large images through low-capacitylinks with the image being gradually built in thereceiving display workstation [67]. The physiciancan, for example, browse over a set of images be-longing to a patient and decide on the basis of rough

(small-sized and/or low-contrast) approximations ofthe images on which of them are useful at that timewithout having to wait for the reception of the wholeimage set.

– Compression has to be fast enough to accommo-date the significant burden imposed on a radiologydepartment. Higher emphasis has to be given onthe speed of decompression; the time between theretrieval command and the display on the screenshould not exceed 2 s [69]. Although this time con-straint is too stringent to be met by an implementa-tion in a general-purpose computer system, there ishowever an imperative need for as fast decompres-sion as possible.

– There are cases where the diagnosis can be basedon one or more regions of the image. In such cases,there must be a way ofselective compression, thatis, allocating more bits to the regions of interest soas to maintain a maximum fidelity whereas the restof the image can be subjected to a higher compres-sion, representing only a rough approximation of thestructures surrounding the selected portions. Thiscan considerably reduce the image size for storageand/or transmission, without sacrificing critical in-formation.Compression techniques based on thewavelet de-

compositionof the image have received much atten-tion in the recent literature on (medical) image com-pression and can meet to a large extent the require-ments imposed by the application. This is mainly dueto the unique ability of the wavelet transform to rep-resent the image in such a way that high compres-sion is allowed preserving at the same time fine detailsof paramount importance. Fast algorithms, based onwell-known digital signal processing tools, have beenderived for the computation of such decompositions,leading to efficient software and hardware implemen-tations which add to the practical value of these meth-ods.

In this paper, we propose using a compressionscheme composed of a wavelet decomposition in con-juction with a modern quantization algorithm and alossless entropy encoder. The quantization module isbased on the embedded zerotree wavelet (EZW) algo-rithmic scheme, which constitutes the state-of-the-artin wavelet-based compression. One of its main fea-tures is the exploitation of the characteristics of thewavelet representation to provide a sequence of em-

Page 3: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 225

bedded compressed versions with increasing fidelityand resolution, thus leading to a particularly efficientsolution to the problem of progressive transmission(PT). A simple yet effective way of exploiting thegenerally nonuniform probability distribution of thesignificant wavelet coefficients is proposed and shownto lead to improved quality of reconstruction at a neg-ligible side-information cost. Our scheme allows bothlossy and lossless (i.e., exactly reversible) compres-sion,for the whole of the image or for selected regionsof interest.By exploiting the priority given to the largemagnitude coefficients by the EZW algorithm, weprovide the possibility of protecting selected regionsof interest from being distorted in the compressionprocess, via a perfectly reversible (lossless) or a moreaccurate than their context lossy representation.

A recently introduced wavelet implementation (lift-ing) that combines speed and numerical robustness isalso investigated in both the general compression con-text, as well as in the lossless region of interest repre-sentation, where it proves to be an efficient means ofdecorrelation that in combination with sophisticatedentropy coding yields high compression performance.

In the entropy coding stage, we have employedarithmetic coding (AC) in view of its optimal perfor-mance. Moreover, the small size of the alphabet rep-resenting EZW output makes adaptive AC well-suitedfor this application. However, the high complexity ofAC might not be affordable by a simple and fast imple-mentation. A much simpler and more efficient losslesscoder that exploits, via run-length encoding (RLE), theoccurrence of long sequences of insignificant waveletcoefficients in natural images, is proposed here andshown to be only slightly worse than context-basedadaptive AC.

The rest of this paper is organized as follows. Sec-tion 2 contains a short presentation of the generalframework for image compression including measuresof compression performance and criteria of evaluatingquality in reconstructed images. The technical detailsof the compression module we propose are given inSection 3. The wavelet transform is compared againstthe more commonly used DCT, underlying the JointPhotographic Experts Group (JPEG) standard, andis proposed as a viable alternative which meets thestringent needs of medical image management. Inaddition to the classical wavelet decomposition, thelifting realization of the wavelet transform is also

investigated and claimed to be of particular interestin the present context. The quantization algorithm ispresented in detail and its advantages for these ap-plications are brought out. The various choices forthe selection of the final stage (i.e. entropy encoder)are also discussed, with emphasis on the proposedRLE-based coder. Two different methods for selectivecompression are discussed and a novel application ofwavelets to lossless compression is presented. Section4 contains experimental results on a real medical im-age, that demonstrate the superiority of the proposedapproach over JPEG with respect to both objectiveand subjective criteria. Conclusions are drawn in Sec-tion 5.

2. Image compression: Framework and perfor-mance evaluation

Image compression [24,25] aims at removing or atleast reducing the redundancy present in the originalimage representation. In real world images, there isusually an amount of correlation among nearby pixelswhich can be taken advantage of to get a more eco-nomical representation. The degree of compression isusually measured with the so-calledcompression ra-tio (CR), i.e., the ratio of the size of the original im-age over the size of the compressed one in bytes. Forexample, if an image with (contrast) resolution of 8bits/pixel (8 bpp) is converted to a 1 bpp representa-tion, the compression ratio will be 8:1. The averagenumber of bits per pixel (bpp) is referred to as thebitrate of the image. One can categorize the image com-pression schemes into two types:

(i) Lossless.These methods (also calledreversible)reduce the inter-pixel correlation to the degreethat the original image can be exactly recon-structed from its compressed version. It is thisclass of techniques that enjoys wide acceptancein the radiology community, since it ensuresthat no data loss will accompany the compres-sion/expansion process. However, although theattainable compression ratio depends on themodality, lossless techniques cannot give com-pression ratios larger that 2:1 to 4:1 [48].

(ii) Lossy. The compression achieved via losslessschemes is often inadequate to cope with thevolume of image data involved. Thus, lossy

Page 4: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

226 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

schemes (also calledirreversible) have to be em-ployed, which aim at obtaining a more compactrepresentation of the image at the cost of somedata loss, which however might not correspondto an equal amount of information loss. In otherwords, although the original image cannot befully reconstructed, the degradation that it hasundergone is not visible by a human observerfor the purposes of the specific task. Compres-sion ratios achieved through lossy compressionrange from 4:1 to 100:1 or even higher.

In general terms, one can describe an image com-pression system as a cascade of one or more of thefollowing stages [12,70]:– Transformation.A suitable transformation is ap-

plied to the image with the aim of converting it intoa different domain where the compression will beeasier. Another way of viewing this is via a changein the basis images composing the original. In thetransform domain, correlation and entropy can belower, and the energy can be concentrated in a smallportion of the transformed image.

– Quantization.This is the stage that is mostly respon-sible for the ‘lossy’ character of the system. It en-tails a reduction in the number of bits used to repre-sent the pixels of the transformed image (also calledtransform coefficients). Coefficients of low contri-bution to the total energy or the visual appearance ofthe image are coarsely quantized (represented witha small number of bits) or even discarded, whereasmore significant coefficients are subjected to a finerquantization. Usually, the quantized values are rep-resented via some indices to a set of quantizer levels(codebook).

– Entropy coding (lossless).Further compressionis achieved with the aid of some entropy codingscheme where the nonuniform distribution of thesymbols in the quantization result is exploited soas to assign fewer bits to the most likely symbolsand more bits to unlikely ones. This results in asize reduction of the resulting bit-stream on the av-erage. The conversion that takes place at this stageis lossless, that is, it can be perfectly cancelled.The above process is followed in the encoding

(compression) part of the coder/decoder (codec) sys-tem. In the decoding (expansion/decompression) part,the same steps are taken in reverse. That is, the com-pressed bit-stream is entropy decoded yielding the

Fig. 1. Block representation of a general transform coding system.

quantized transform coefficients, then ‘de-quantized’(i.e., substituting the quantized values for the corre-sponding indices) and finally inverse transformed toarrive at an approximation of the original image. Thewhole process is shown schematically in Fig. 1.

To compare different algorithms of lossy compres-sion several approaches of measuring the loss of qual-ity have been devised. In the MI context, where theultimate use of an image is its visual assessment andinterpretation, subjective and diagnostic evaluation ap-proaches are the most appropriate [12]. However, theseare largely dependent on the specific task at hand andmoreover they entail costly and time-consuming pro-cedures. In spite of the fact that they are often inad-equate in predicting the visual (perceptual) quality ofthe decompressed image, objective measures are oftenused since they are easy to compute and are applicableto all kinds of images regardless of the application. Inthis study we have employed the following two dis-tortion measures:

(i) Peak signal-to-noise ratio (PSNR):

PSNR= 10 log10

(P 2

MSE

)dB

where

MSE = 1

MN‖X − X‖2

2 = 1

MN

∑i,j

|Xi,j − Xi,j |2

is themean squared errorbetween the original,X, and the reconstructed,X, M ×N images, andP is the maximum possible value of an elementof X (e.g., 255 in an 8 bpp image).

(ii) Maximum absolute difference (MAD)

MAD = ‖X − X‖∞ = maxi,j

|Xi,j − Xi,j |.

Page 5: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 227

At this point it is of interest to note that a degrada-tion in the image might also be present even in a loss-less transform coding scheme (i.e., where no quanti-zation is present). This is because the transformationprocess is perfectly invertible only in theory, and if di-rectly implemented, it can be a cause of distortion dueto finite precision effects. In this work, an implemen-tation of the wavelet transform that enjoys the perfectreconstruction property in the presence of arithmeticerrors is being investigated.

3. The proposed compression scheme

The compression scheme proposed here and shownin the block diagram of Fig. 2 is built around thegeneral process described in Section 2 (Fig. 1). In thesequel, we elaborate on the specific choices for thestages of the general transform-coding process madein our system.

3.1. Transformation

3.1.1. Block-transform coding – discrete cosinetransform

Although a large variety of methods have beenproposed for medical image compression, includingpredictive coding, vector quantization approaches,segmentation-based coding schemes, etc. (e.g.,[8,11,29,55]) the class of techniques that are based onlinear transformations dominates the field [4,7,9,32–34,42,47,51,60,72]. The main advantage of transformcoding techniques is the ability of allocating a differ-ent number of bits to each transform coefficient so asto emphasize those frequency (or scale) componentsthat contribute more to the way the image is perceptedand de-emphasize the less significant components,thus providing an effective way of quantization noiseshaping and masking [26].

As noted already, the goal of the transformation stepis to decorrelate the input samples and achieve com-

Fig. 2. Block diagram of the proposed coder. Details on selectivecompression are omitted.

paction of energy in as few coefficients as possible. Itis well known that the optimum transform with respectto both these criteria is the Karhunen–Loeve trans-form (KLT) [26], which however is of limited practicalimportance due to its dependence on the input signalstatistics and the lack of fast algorithms for computingits basis functions. Therefore, suboptimal yet compu-tationally attractive transforms are often used in prac-tice, with thediscrete cosine transform (DCT)beingthe most prominent member of this class. DCT owesits wide acceptance to its near-optimal performance(it approaches KLT for exponentially correlated sig-nals with a correlation coefficient approaching one, amodel often used to represent real world images) andthe existence of fast algorithms for its computation,stemming from its close relationship to the discreteFourier transform (DFT) [31,44]. Moreover, 2D-DCTis separable, that is, it can be separated into two 1DDCTs, one for the rows and one for the columns ofthe image (row–column scheme), a feature that furthercontributes to its simplicity.

The usual practice of DCT-based coding (that usedalso in the JPEG standard for still continuous-tone im-age compression baseline mode) is to divide the imageinto small blocks, usually of 8× 8 or 16× 16 pix-els, and transform each block separately. This block-transform approach is followed mainly for computa-tional complexity and memory reasons. Moreover, itallows adaptation of the spectral analysis action of thetransform to the local characteristics of a nonstationaryimage. However, in low bit-rate compression, whereseveral of the high-frequency DCT coefficients are dis-carded or coarsely quantized, this approach leads toannoying blocking artifacts in the reconstructed image[31,64]. This is due to the poor frequency localizationof DCT, that is, the dispersion of spatially short activeareas (e.g., edges) in a large number of coefficients oflow energy [69]. This interdependency among adja-cent blocks can be alleviated by simply applying theDCT to the whole of the image, considering it as asingle block. This so-called full-frame DCT (FFDCT)has been widely applied in the medical imaging area(e.g., [9,32]), where the slightest blocking effect isdeemed unacceptable. However, in view of the highspatial and contrast resolution of most of the medicalimages [28], the computational and storage require-ments of such a scheme prevent its implementation ingeneral-purpose software or hardware systems, requir-

Page 6: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

228 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

Fig. 3. Two-band multirate analysis/synthesis system.

ing specialized multiprocessor modules with externalframe buffers for fast and efficient operation [9].

3.1.2. The discrete wavelet transformA more realistic solution to the blocking problem

is offered bysubband codingschemes [58,61,64], thatis, generalized linear transforms where block over-lapping is intrinsic in the method. This property hasa smoothing effect that mitigates the sharp transitionbetween adjacent blocks. Viewed from the point ofview of a multirate maximally decimated filter bank,this is equivalent to the filters (basis functions) beinglonger than the number of subbands (basis size). Thisis in contrast to the DCT case, where the inadequatesize of the basis vectors results in poor frequency lo-calization of the transform. A multiresolution (multi-scale) representation of the image that trades off spa-tial resolution for frequency resolution is provided bythe discrete wavelet transform (DWT).The best wayto describe DWT is via a filter-bank tree. Fig. 3 de-picts the simple 2-band filter bank system for the 1Dcase. The input signal is filtered through the lowpassand highpass analysis filtersH0 andH1, respectively,and the outputs are subsampled by a factor of 2, thatis, we keep every other sample. This sampling rate al-teration is justified by the halving of the bandwith ofthe original signal. After being quantized and/or en-tropy coded, the subband signals are combined againto form a full-band signal, by increasing their sam-pling rate (upsampling by a factor of 2) and filteringwith the lowpass and highpass synthesis filtersG0 andG1 that interpolate the missing samples. It is possibleto design the analysis/synthesis filters2 in such a waythat in the absence of quantization of the subband sig-nals, the reconstructed signal coincides with the orig-

2 Here we deal only with FIR filters, that is, functionsHi(z) andGi(z) are (Laurent) polynomials inz.

Fig. 4. Separable 2D 4-band filter bank: (a) analysis; (b) synthesis.

inal, i.e., x = x. Such analysis/synthesis systems aresaid to have theperfect reconstruction (PR)property.

A direct extension to the 2D case, which is also themore commonly used, is to decompose separately theimage into low and high frequency bands, in the ver-tical and horizontal frequency [63]. This is achievedvia the separable 2D analysis/synthesis system shownin Fig. 4, which ideally results in the decompositionof the frequency spectrum shown in Fig. 5.

The analysis into subbands has a decorrelating ef-fect on the input image. Nevertheless, it might oftenbe necessary to further decorrelate the lowpass sub-band. This is done via iterating the above scheme onthe lowpass signal. Doing this several times leads toa filter bank tree such as that shown in Fig. 6. Theresult of such a 3-stage decomposition can be seenin Fig. 7. This decomposition of the input spectruminto subbands which are equal in a logarithmic scaleis nothing but the DWT representation of the image[45,58,61,64,72].

Page 7: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 229

Fig. 5. Ideal decomposition of the 2D spectrum resulting from thefilter bank of Fig. 4(a).

Fig. 6. 2D DWT: (a) forward; (b) inverse. (H -and G-blocks rep-resent the 4-band analysis and synthesis banks of Fig. 4(a) and(b), respectively.)

Fig. 7. Wavelet decomposition of the image spectrum (3 levels).

Provided that the 2-band filter bank satisfies the per-fect reconstruction condition, the structure of Fig. 6represents an expansion of the image with the basisfunctions being given by the impulse responses of theequivalent synthesis filter bank and the coefficients be-ing the outputs of the analysis bank, referred to also asDWT coefficients. In this expansion, the basis func-tions range from spatially limited ones of high fre-quency that capture the small details in the image, e.g.,edges, lines, to more expanded of low frequency thatare matched to larger areas of the image with higherinter-sample correlation. In contrast to what happensin the block-transform approach, a small edge, for ex-ample, in the image is reflected in a small set of DWTcoefficients, with its size depending on how well lo-calized the particular wavelet function basis is. It ismainly for these good spatial-frequency localizationproperties that the DWT has drawn the attention ofthe majority of the MI compression community, asa tool for achieving high-compression with no visi-ble degradation and preservation of small-scale details[34,47,51,60,72].

3.1.3. The lifting schemeAs already noted, direct implementations of theo-

retically invertible transforms usually suffer from in-version errors due to the finite register length of thecomputer system. Thus, one can verify that even in adouble precision implementation of the wavelet treeof Fig. 6, where the 2-band filter bank satisfies theconditions for PR, the input image will not be exactlyreconstructed. This shortcoming of the direct DWTscheme gets even worse in a practical image compres-sion application where the wavelet coefficients willfirst have to be rounded to the nearest integer beforeproceeding to the next (quantization or entropy encod-ing) stage. The presence of this kind of error wouldprove to be an obstacle in the application of the DWTin lossless or even near-lossless compression. In thissection we focus on the notion oflifting, which rep-resents a generic and efficient solution to the perfectinversion problem [6,10,16,59].

To describe the idea underlying the lifting scheme,we first have to reformulate the analysis/synthesisstructure of Fig. 3, translating it into an equivalent,more compact scheme. Decomposing the analysis andsynthesis filters into theirpolyphase components:

Page 8: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

230 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

Fig. 8. Polyphase implementation of Fig. 3.

Hi(z) = Hi,0(z2) + z−1Hi,1(z

2)

Gi(z) = Gi,0(z2) + zGi,1(z

2),

and making use of some well-known multirate identi-ties, one obtains the polyphase structure of Fig. 8 [61].Thepolyphase matrixHHHp(z) is defined as

HHHp(z) =[

H0,0(z) H0,1(z)

H1,0(z) H1,1(z)

]

with a similar definition forGGGp(z). It is apparent fromFig. 8 that PR is ensured if and only if the identity

GGGTp(z)HHHp(z) = III

holds. The idea behind lifting is torealize these twotransfer matrices in such a way that the above iden-tity is preserved regardless of the numerical precisionused.

The term ‘lifting’ refers to a stepwise enhancementof the band splitting characteristics of a trivial PR filterbank, usually the split–merge filter bank correspond-ing to the choiceHHHp(z) = GGGp(z) = III in Fig. 8. Thisis done via multiplication of the analysis polyphasematrix by polynomial matrices that are of a specialform: they are triangular with unit diagonal elements.The synthesis polyphase matrix is of course also mul-tiplied by the corresponding inverses which are sim-ply determined by inspection, namely by changing thesign in the nondiagonal elements. The important pointwith respect to the realization of an analysis/synthesissystem with the aid of a lifting scheme is that thepolyphase matrix of any PR FIR system, which with-out loss of generality may be considered as havingunity determinant, can be factorized in terms of suchelementary factor matrices. It can readily be verifiedthat the effect of rounding-off the multiplication re-sults in the analysis bank of such a structure can beexactly cancelled by an analogous rounding in the syn-thesis stage, resulting in a realization of the DWT thatmaps integers to integers with the PR property being

Fig. 9. Lifting realization of the Haar filter bank: (a) analysis; (b)synthesis.

structurally guaranteed. It should be noticed that syn-thesis involves exactly the same operations with theanalysis bank, except in the reverse order and with thesigns changed.

A simple but informative example of the liftingscheme is provided by Fig. 9 illustrating the cor-responding realization of the forward and inversesecond-order Haar transform:

H = 1√2

[1 11 −1

],

In addition to its potential for providing a loss-less multiresolution of an image, lifting can also yieldsignificant savings in both computational and storagecomplexity. As shown in [14], for sufficiently long fil-ters, the operations count in the direct wavelet filterbank can drop by about one half when lifting stepsare used. Moreover, the triangular form of the factormatrices allows the computations to be performed inplace, in contrast to the classical algorithm [59].

3.2. Quantization

The role of the DWT in the compression schemepresented here is not merely that of a decorrelatingand energy compacting tool. It also yields a means ofrevealing the self-similarity [53] and multiscale struc-ture present in a real world image, making efficient en-coding and successive approximation of the image in-formation possible and well suited to the needs of pro-gressive transmission. These properties of the DWTare exploited in a very natural manner by the quanti-zation stage employing an EZW scheme.

The EZW algorithm is a particularly effective ap-proach to the following twofold problem: (1) achiev-

Page 9: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 231

ing the best image quality for a given compressionratio (bit-rate) and (2) encode the image in such away that all lower bit rate encodings are embedded atthe beginning of the final bit-stream. This embeddingproperty provides an attractive solution to the problemof progressive transmission, especially when transmis-sion errors due to channel noise have also to be dealtwith. The basic phases of EZW are very similar tothose of the DCT-based coding used in JPEG, thatis, first information on which of the coefficients areof significant magnitude is generated and then thosecoefficients that are significant are encoded via somequantization. However, the multiresolution character-istics of the wavelet representation allow a much moreefficient implementation of these two phases. First,the significance map, that is, the set of decisions asto whether a DWT coefficient is to be quantized aszero or not, is encoded taking advantage of the self-similarity3 across scales (subbands) inherent in thewavelet decomposition of natural images and the fre-quently satisfied hypothesis of a decaying spectrum.The quantization of the coefficients is performed ina successive approximation manner, via a decreasingsequence of thresholds. The first threshold is taken tobe half the magnitude of the largest coefficient. A co-efficient x is said to besignificantwith respect to athresholdT if |x| ≥ T , otherwise it is calledinsignif-icant. Due to the hierarchical nature of the waveletdecomposition, each coefficient (except in the highestfrequency subbands) can be seen to be related to a setof coefficients at the next finer scale of similar orien-tation and corresponding spatial location. In this way,a tree structure can be defined as shown for examplein Fig. 7. If the hypothesis stated above is satisfied, itis likely that the descendants of an insignificant coeffi-cient in that tree will also be insignificant with respectto the same threshold. In such a case, this coefficientis termed azero-tree root (ZTR)and this subtree ofinsignificant coefficients is said to be azero-treewithrespect to this threshold. In contrast to related algo-rithms, e.g. [30], where the insignificance of a treeis decided upon a global statistical quantity, a ZTRhere implies thatall of its descendants are of negli-gible magnitude with respect to the current threshold,

3 Interestingly, this property has been recently shown to be closelyrelated to the notion of self-similarity underlying fractal imagerepresentation [15].

thus preventing statistical insignificance from obscur-ing isolated significant coefficients. A consequence ofthis point is that the descendants of a zero-tree rootdo not have to be encoded; they arepredictably in-significant.Moreover, the fact that the encoder fol-lows a specific order in scanning the coefficients, start-ing from low-frequency and proceeding to higher fre-quency subbands with a specific scanning order withineach subband, excludes the necessity of transmissionof side position information as it is done for examplein threshold DCT coding. In the case that not all de-scendants are insignificant, this coefficient is encodedasisolated zero (IZ).Insignificant coefficients belong-ing to the highest frequency subbands (i.e., with nochildren) are encoded with thezero (Z)symbol. ThesymbolsPOSandNEG, for positive and negative sig-nificant, respectively, are used to encode a significantcoefficient.

During the encoding (decoding) two separate listsof coefficients are maintained. Thedominantlist keepsthe coordinates of the coefficients that have not yetbeen found to be significant. The quantized magni-tudes of those coefficients that have been found to besignificant are kept in thesubordinatelist. For eachthresholdTi in the decreasing sequence of thresholds(hereTi = Ti−1/2) both lists are scanned. In the dom-inant pass, every time a coefficient in the dominant listis found to be significant, it is added at the end of thesubordinate list, after having recorded its sign. More-over, it is removed from the dominant list, so that it isnot examined in the next dominant pass. In the sub-ordinate pass a bit is added to the representations ofthe magnitudes of the significant coefficients depend-ing on whether they fall in the lower or upper half ofthe interval [Ti, Ti−1). The coefficients are added tothe subordinate list in order of appearance, and theirreconstruction values are refined in that order. Hencethe largest coefficients are subjected to finer quanti-zation, enabling embedding and progressive transmis-sion. To ensure this ordering with respect to the mag-nitudes, a reordering in the subordinate list might beneeded from time to time. This reordering is doneon the basis of the reconstruction magnitudes knownalso to the decoder, hence no side information needsto be included. The two lists are scanned alternatelyas the threshold values decrease until the desired filesize (average bit rate) or average distortion has beenreached. A notable feature of this algorithm is that the

Page 10: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

232 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

size or distortion limits set by the application can beexactly met, making it suitable for rate or distortionconstrained storage/transmission.4

In the decoding stage, the same operations takeplace, where the subordinate information is used to re-fine the uncertainty intervals of the coefficients. Thereconstruction level for a coefficient can be simplythe center of the corresponding quantization interval.However, this implies a uniform probability distribu-tion, assumption which in some cases might be quiteinaccurate. Therefore, in our implementation,we re-construct the magnitudes on the basis of the average‘trend’ of the real coefficients towards the lower orupper boundary of the interval, as estimated in theencoding stage.This information is conveyed to thedecoder in the form of a number in the range (0,1)representing the relative distance of the average coef-ficient from the lower boundary. This in fact providesa rudimentary estimate of the centroid in each intervalfor a nonuniform probability density function.

The compressed file contains a header with someinformation necessary for the decoder such as imagedimensions, number of levels in the hierarchical de-composition, starting threshold, and centroid informa-tion. We have seen that subtracting the mean from theimage before transforming it, and adding it back atthe end of the decoding phase yielded some improve-ment in the quality of the decompressed image.5 Themean subtraction has the effect of bias reduction (de-trend) in the image and moreover, as reported in [62],ensures zero transmittance of the high-pass filters atthe zero frequency, thus avoiding the occurrence of ar-tifacts due to insufficiently accurate reconstruction ofthe lowpass subband. The image mean is also storedin the header of the compressed file.

It must be emphasized that, although the embed-ding property of this codec may be a cause of sub-optimality [19], it nevertheless allows the truncationof the encoding or decoding procedure at any pointwith the same result that would have been obtained ifthis point was instead the target rate at the starting of

4 This characteristic should be contrasted to what happens inJPEG implementations, where the user specification for the amountof compression in terms of theQuality Factor cannot in generallead to a specified file size.

5 An analogous operation takes place in the JPEG standard aswell, where an offset is subtracted from the DCT coefficients [66].

the compression/expansion process [52,53]. Progres-sive coding withgraceful degradationas required inmedical applications can be benefited very much bythe above property. The embedding property of ourcompression algorithm closely matches the require-ments for progressive transmission in fidelity (PFT),i.e., progressive enhancement of the numerical accu-racy (contrast resolution) of the image. The sequenceof reconstructed images produced by the correspond-ing sequence of thresholds (bit-planes) with the inher-ent ordering of importance imposed by EZW on thebits of the wavelet coefficients – precision, magnitude,scale, spatial location, can be used to build an efficientand noise-protected PT system effectively combiningPFT and progressive (spatial) resolution transmission(PRT).

3.3. Entropy coding

The symbol stream generated by EZW can bestored/transmitted directly or alternatively can beinput to an entropy encoder to achieve further com-pression with no additional distortion. Compressionis achieved by replacing the symbol stream with asequence of binary codewords, such that the averagelength of the resulting bit-stream is reduced. Thelength of a codeword should decrease with the prob-ability of the corresponding symbol. It is well knownthat the smallest possible number of bits per symbolneeded to encode a symbol sequence is given by theentropyof the symbol source [35]:

H = −∑

i

pi log2 pi,

wherepi denotes the probability of theith symbol.In an optimal code, theith symbol would be repre-sented by−log2 pi bits. Huffman coding[23] is themost commonly used technique. This is due to its op-timality, that is, it achieves minimum average codelength. Moreover, it is rather simple in its design andapplication. Its main disadvantage is that it assigns aninteger number of bits to each symbol, hence it can-not attain the entropy bound unless the probabilitiesare powers of 2. Thus, even if a symbol has a proba-bility of occurrence 99.9%, it will get at least one bitin the Huffman representation. This can be remediedby using block Huffman coding, that is, grouping the

Page 11: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 233

symbols into blocks, at the cost of an increase in com-plexity and decoding delay.

A more efficient method, which can theoreticallyachieve the entropy lower bound even if it is frac-tional, isarithmetic coding (AC)[21,39,41,69]. In thisapproach, there is no one-to-one correspondence be-tween symbols and codewords. It is the message, i.e.the sequence of symbols, that is rather assigned a code-word, and not the individual symbols. In this way, asymbol may be represented with less that 1 bit. Al-though there is a variety of arithmetic coders that havebeen reported and used, the underlying idea is thesame in all of them. Each symbol is assigned a subin-terval of the real interval [0,1), equal to its probabilityin the statistical model of the source. Starting from[0,1), each symbol is coded by narrowing this inter-val according to its allotted subinterval. The result isa subinterval of [0,1) and any number within it canuniquely identify the message. In practice, the subin-terval is refined incrementally, with bits being outputas soon as they are known. In addition to this incre-mental transmission/reception, practical implementa-tions employ integer arithmetic to cope with the high-precision requirements of the coding process [39,69].

AC achievesH bits/symbol provided that the esti-mation of the symbol probabilities is accurate. Thestatistical model can be estimated before coding be-gins or can be adaptive, i.e., computed in the courseof the processing of the symbols. An advantage of theadaptive case is that, since the same process is fol-lowed in the decoder, there is no need for storing themodel description in the compressed stream. One canimprove on the simple scheme described above, byemploying higher-order models, that is, utilizing con-ditional probabilities as well [39]. We have used sucha coder in our experiments. The model was adaptivelyestimated for each dominant and subordinate pass, thatis, the model was initialized at the start of each phase.This choice was made for two reasons: first, the abil-ity of meeting exactly the target rate is thus preserved,and second, the statistics of the significance symbolsand the refinement bits would in general differ.

In spite of its optimality, this method has not re-ceived much attention in the data compression area(though there are several reported applications of itin medical image compression [28,49,51]) mainly be-cause of its high complexity which considerably slowsdown the operation of the overall system. Of course,

it should be noted that specialized hardware could beused to tackle this bottleneck problem [41]. Further-more, approximations to pure AC can be used to speedup the execution at the cost of a slight degradationin compression performance [21,41]. In this work, wehave tested a much faster and cost-effective entropycoding technique, based on a rudimentary binary en-tropy coding of the significance map symbols and anRLE of the ZTR and Z symbols. Based on extendedexperimentation we have confirmed the fact that theZTR and Z symbols are much more frequent than therest of the symbols. Thus, we use a single bit for rep-resenting each of these symbols, and a few more forthe rest of them. For the lower frequency subbandswe use the codes ZTR:0, IZ:10, POS:110, NEG:1110,while the symbols in the highest frequency subbandsare coded as Z:0, POS:10, NEG:110. Notice that sincethe decoder knows exactly in which subband it is foundat any time, no ambiguity results from the fact that thesame codes are assigned to different symbols in lowand high frequency subbands. We use an RLE codeto exploit the occurrence of long bursts of zero coeffi-cients. RLE merely represents a long series of consec-utive symbols by the length of the series (run-length)and the symbol [39]. The codeword 111. . . 10 wherethere arek 1’s in total represents a run of 2k−1 − 1zero symbols. To avoid confusion with single-symbolcodewords,k is restricted to exceed 2 and 3 for thehighest and low subbands, respectively. We have notused this technique on the subordinate list because ofthe larger fluctuation in the quantization bit sequences.As it will be seen in the evaluation of our experimen-tal results,the performance of this RLE coder is onlyslightly worse than that of AC, in spite of its simplicityand computational efficiency.

3.4. Selective compression

A compromise between the need for high compres-sion imposed by the vast amount of image data thathave to be managed and the requirement for preser-vation of the diagnostically critical information withlimited or no data loss, is provided in our compres-sion scheme by theselective compression (SC)option.This means that the doctor can interactively select oneor more (rectangular) regions of the image that he/shewould like to be subjected to less degradation than

Page 12: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

234 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

their context in the compression process. Dependingon his/her judgement on the importance of these re-gions of interest (RoI), he/she can specify that they belosslessly compressed or provide for each a quantita-tive specification of its usefulness relative to the restof the image so that it can be more finely quantized.

(i) Lossy SC.The relative importance of the RoI ascompared to its context is quantified via a weightfactor, e.g., 4 or 16. The spatial localization ofthe DWT is taken advantage of in this case to al-low the translation of the spatial coordinates ofthe RoI to those in the wavelet domain. In otherwords, we make use of the fact that each pixeldepends on only a few wavelet coefficients andvice versa. The RoI is determined at each levelby simply halving the coordinates of its repre-sentation at the previous (finer scale) level. Thecorresponding regions in the image subbands aremultiplied by the weight factor yielding an am-plification of the wavelet coefficients of interest.These same coefficients are divided by this fac-tor in the reconstruction phase to undo the em-phasis effect.6 This trick exploits the high pri-ority given to the coefficients of large magnitudeby the EZW algorithm [56]. These coefficientsare coded first, so the bit budget is mostly spenton the RoI whereas the rest of the image under-goes a coarser quantization. The larger the weightassigned to a RoI, the better its reconstructionwith respect to its context. The amount of degra-dation suffered from the background image de-pends on the target bit rate, the number and areasof the RoIs selected. Obviously, if a large num-ber of small RoIs are selected and assigned largeweights, the quality of the rest of the image mightget too low if a high compression ratio is to beachieved. The selection process must guaranteethat the appearance of this part of the image issufficiently informative to aid the observer appre-ciate the context of the RoI. The parameters ofa successful selection are largely task-dependentand the doctor might have to reevaluate his/herchoices in the course of an interactive trial-and-error process.

6 The weight factor along with the coordinates of the RoI areincluded in the compressed file header.

Our approach of amplifying the wavelet co-efficients corresponding to the selected RoI hasshown better results than the method proposedby Shapiro[54] where the weighting is applieddirectly to the spatial domain, that is, beforewavelet transforming the image. The latter ap-proach makes use of the linearity of the DWT totranslate the weighting to the subband domain.The same reasoning followed in the wavelet-domain method is also adopted here with respectto the prioritization of the large coefficients inthe EZW coding. However, the amplification ofa portion of the image generates artificial stepswhich result in annoying ringing effects when fil-tered by the wavelet filter bank. The introductionof a margin around the selected region to achievea more gradual amplification with the aid of asmooth window [54] was not sufficient to solvethe problem. The visually unpleasing and per-haps misleading borders around the RoI do notshow up in the wavelet approach even when nosmoothing window is used.

(ii) Lossless SC.Although the above method for SCcan in principle be used for lossless compressionas well, a more accurate and reliable approachfor reversibly representing the RoIs has also beendeveloped. It is based on the idea of treating theRoIs in a totally different way than the rest ofthe image. After the selection has been done, theRoIs are separately kept and input to a losslessencoder while the image is compressed with thenormal approach. Prior to its entropy encoding,the RoI can undergo a lowering in its entropywith the aid of a suitable transformation. This op-eration will normally permit a further compres-sion gain since it will reduce the lower bound forthe achievable bit-rate. Extended studies of theperformance of several transform-entropy codercombinations for reversible medical image com-pression have been reported [28,48].In this work,we have used the DWT in its lifting represen-tation as the decorrelation method. Apart fromthe advantages emerging from the hierarchicaldecomposition implied by this choice, the struc-tural enforcement of the exact reversibility of thelifting scheme naturally meets the requirementfor purely lossless compression.In the expan-sion stage, the RoIs are ‘pasted’ onto their orig-

Page 13: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 235

inal places. To save on the bit-rate for the restof the image, the corresponding portions are firstde-emphasized in the encoding stage by atten-uating the corresponding coefficients. This is ineffect the exact counterpart of the trick used inthe lossy SC approach, where now the selectedwavelet coefficients are given the lowest priorityin the allocation of the bit-budget. There is alsothe possibility of performing the de-emphasis inthe spatial domain by also introducing a suffi-ciently wide margin (whose width depends onthe length of the filters used) to avoid leakage ofringing effects outside the protected region.

As it will become apparent in the experimentalresults presented in the next section, this SC ap-proach, despite its conceptual simplicity, ispar-ticularly effective with respect to providing saferepresentation of critical information and visu-ally pleasing results, without sacrificing com-pression performance.

4. Results

This section presents typical results from the appli-cation of our compression scheme to real medical im-agery, aiming to demonstrate its applicability to the de-manding problems inherent in the medical image com-pression area and its superior performance over that ofmore conventional approaches in terms of both objec-tive and subjective criteria. The test image used in theexperiments presented here is the 512× 512× 8 bppchest X-ray image shown in Fig. 10. The filter banksused in our simulations include the Haar as well asthe (9,7) pair developed by Antonini et al. [3]. Bothfilter banks were ranked among the best ones for im-age compression in the study of [65]. Both systemsenjoy the property of linear phase which is partic-ularly desirable in image processing applications. Ithas been shown to contribute to a better preservationof the details in low bit-rate compression, apart fromovercoming the need for phase compensation in treesof nonlinear-phase filters. Orthogonality, a property ofthe transform that allows the direct translation of dis-tortion reductions from the transform to the spatial do-main in transform coding schemes, is satisfied by bothfilter pairs, albeit only approximately by the (9,7) sys-

Fig. 10. Test image used in our experiments (512× 512× 8 bpp).

tem.7 Despite the simplicity and computational effi-ciency of the Haar wavelet and its lower susceptibilityto ringing distortions as compared to the longer (9,7)filter bank, the latter is deemed more suitable for suchapplications in view of its better space-frequency lo-calization which battles against the blocking behavioroften experienced in Haar-based low-bit rate compres-sion.

We have tested the performance of our scheme ver-sus the JPEG codec in its Corel 7 implementation. TheDWT stage employed a 5-level tree using the (9,7) fil-ter bank. A practical matter comes up when wavelettransforming an image, namely that of coping with thefinite input size, i.e., performing the filtering with theappropriate initial conditions that do not harm the PRproperty. In the experiments discussed here, we usedsymmetric periodic extension to cope with this prob-lem, that is, the image is extended as if it were viewedthrough a mirror placed at its boundaries. This choicehas the advantage of alleviating distortions resultingfrom large differences between its opposing bound-aries. The direct DWT scheme was used. The resultswith the lifting scheme are not included since for lossycompression they are comparable to those of the di-

7 This is in fact abiorthogonalanalysis/synthesis system, meaningthat the analysis filters are orthogonal to the synthesis ones butnot to each other.

Page 14: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

236 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

Fig. 11. PSNR comparison of EZW-AC, EZW-RLE, and JPEG.

Fig. 12. MAD comparison of EZW-AC, EZW-RLE, and JPEG.

rect scheme, usually slightly worse especially for highcompression ratios. We have tested both the AC andour RLE-based lossless coding algorithms in the en-tropy coding stage. The resulting PSNR for the threecodecs and for a wide range of compression ratiosis plotted in Fig. 11. We observe that the curves forthe DWT-based schemes are well-above that of JPEGfor compression ratios larger that 10:1. Furthermore,the loss from employing the simpler and faster RLEcoder in comparison to the much more sophisticatedfirst-order model-based AC is seen to not exceed 1 dB.These results enable us to propose the RLE-based

Fig. 13. Rate-distortion curves for EZW-RLE: (a) PSNR; (b) MAD.

coder as an attractive alternative to more complex en-tropy coding schemes, since it is proven to combineconceptual and implementational simplicity with highcoding efficiency.

The same ranking of the three compression ap-proaches is preserved also in terms of a comparisonof the resulting MAD values, as shown in Fig. 12.Moreover, it can be seen that, with respect to this fig-ure of merit, the DWT-based compressor consistentlyoutperforms JPEG, for all compression ratios exam-ined. The results of Fig. 12, albeit reflecting an objec-tive performance evaluation, might also be of use in

Page 15: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 237

Fig. 14. Results of 32:1 compression with (a) JPEG and (b) EZW-RLE. (c), (d) Details of (a), (b). Corresponding errors are shown in (e), (f).

conjecturing on the subjective visual quality of the re-constructed images [60]. Besides, isolated local errorsmight prove to be more severe, for the purposes of di-agnosis, than distortions that have been spread out onlarger areas of the image.

We should note that, as it is evident in both of theabove comparisons, the reconstruction quality degra-dation, albeit increasing in all three algorithms withincreasing compression, is far more pronounced in theJPEG case.

Page 16: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

238 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

Fig. 15. Lossy SC: weight=16, CR=32:1. (a) RoI shown in border. Reconstructed image using wavelet (b) and spatial (c) domain weighting.(d), (e) Error images for (b), (c).

We have also tested the DWT compressor in adenser set of compression ratios, from 5:1 to 50:1.The results for the PSNR and MAD are plotted inFig. 13 and should be indicative of the rate-distortionbehavior of this scheme.

Fig. 14 shows the compressed/decompressed chestimage at a ratio of 32:1 using JPEG and the pro-posed scheme with 5-level (9,7) DWT and RLE cod-ing. The latter achieves a MAD of 13, and a PSNRof 41.888 dB, as compared to 18 and 39.769 dB for

Page 17: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 239

Fig. 16. Lossless SC: (a) reconstructed and (b) error images for the RoI of Fig. 15(a). A CR of 32:1 was specified for the rest of the image.

the MAD and the PSNR furnished by the JPEG coder.Moreover, no visible artifact is present in the DWT-compressed image, whereas the blocking effect is ev-ident in the result of JPEG, as can be observed in themagnified images shown in the same figure. Viewingthe corresponding (normalized to the range [0,255]) er-ror images further reveals the prevalence of the block-ing effect in the JPEG image.

The SC capability of our system has also been testedin both its lossy and lossless modes. The rectangularregion shown with the black border in Fig. 15(a) is theRoI selected for this experiment. For the test of thelossy SC, a weight factor of 16 was chosen and thetarget rate was specified to correspond to a compres-sion ratio of 32:1. Fig. 15(b,c) shows the reconstructedimages where amplification has been performed in thewavelet and spatial domains, respectively. In the lattercase, a raised-cosine window has been employed tosmooth-out the sharp transition generated. In spite ofthe smoothing used, the image in Fig. 15(c) exhibitsan annoying ringing effect around the RoI, in con-trast to the satisfying appearance of the region and itscontext in Fig. 15(b). The error images shown in Fig.15(d,e) are also informative on the comparison of thetwo approaches.

The result of the application of the lossless SC tech-nique to the same RoI is shown in Fig. 16. A compres-sion ratio of 32:1 was again used for the backgroundof the image. Notice that, contrary to what one wouldexpect from such a simple ‘cut-and-paste’ technique,

the result is a visually pleasant image showing no bor-dering effects around the selected region. In the exam-ple of Fig. 16 we used an AC to losslessly compressthe RoI. We have also evaluated the performance that a1-level Haar lifting scheme would have in decorrelat-ing this region. The choice of the Haar transform wasmade on the basis of its simplicity and furthermore, toexploit the fact that the matrixH coincides with theKLT of order-2, which is the optimum decorrelating2 × 2 transform [26]. In effect, the use of the liftingscheme in this example corresponds to a lossless ver-sion of the second-order KLT. We have found that theeffect of the Haar lifting transform on the entropy ofthe RoI was a decrease from 6.8123 bpp to 4.8744 bpp,or equivalently a compression ratio of 1.4:1 if an op-timal entropy coder were used. Better results could beobtained by using a deeper lifting tree. For example,with a 3-level tree a reduction in entropy of 3:1 isachieved. Moreover, the multiresolution nature of theDWT adds the possibility for progressive transmissionof the RoI. We feel that the applicability of the liftingscheme in reversible image compression deserves fur-ther investigation, in view of our preliminary results.

We have also investigated the potential of ourscheme for PT applications. A few frames of a pro-gressive decompression using the above techniqueare shown in Fig. 17, where it is also assumed thatan RoI has been selectively compressed. Notice howthe inherent prioritization of EZW shows up with theRoI being received and gradually refined first.

Page 18: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

240 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

Fig. 17. Progressive decompression corresponding to Fig. 15(b). (Not all frames are shown.)

5. Concluding remarks

The conflicting demands for high-quality and lowbit-rate compression imposed by the nature of themedical image compression problem are too difficult

to be satisfied by the traditional coding techniques, in-cluding the common JPEG standard. To come up witha scheme that would seem promising in meeting therequirements for image management in contemporaryPACS and teleradiology applications, we have tried

Page 19: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 241

to employ the state-of-the-art techniques in the cur-rent image compression palette of tools. Our system isbased on a combination of the DWT, EZW compres-sion with nonuniform coefficient distribution takeninto account, and effective entropy coding techniquesincluding a particularly efficient RLE-based coder. Anotable feature of the proposed codec is the selectivecompression capability, which considerably increasesits usefulness for MI applications, by allowing highcompression and critical information preservation atthe same time.

We have tested our codec in a variety of imagesand compression ratios and evaluated its performancein terms of objective measures as well as subjectivevisual quality. The unique features of multiresolution(or hierarchical) decomposition with enhanced spatial-frequency localization provided by the DWT, alongwith the clever exploitation of the DWT properties inthe EZW algorithm are responsible for most of theperformance gain exhibited by our scheme in com-parison to JPEG. The ordering of the bits in termsof their importance in conjuction with the embeddingcharacteristic increase the potential for fast and ro-bust progressive transmission, in a manner that effec-tively combines the spectral selection and successiveapproximation methods of the progressive mode aswell as the hierarchical mode of the JPEG algorithm[5,43]. Nevertheless, despite the good results obtainedin our experiments, it should be emphasized that theperformance of a compression algorithm is in generaltask-dependent and a period of thorough and extendedtesting and objective and subjective measurement hasto be experienced prior to applying this methodologyto clinical practice.

The compression scheme presented in this paperconstitutes a powerful palette of techniques includinghigh-performance approaches for traditional compres-sion tasks as well as innovative features (such as SC)and novel configurations (e.g., lifting decorrelation)that pave the way to further research and development.One could also substitute for the various modules inthe general scheme so as to adapt the system to theperformance and/or functional requirements of a spe-cific application area. For example, instead of the ba-sic version of the EZW algorithm used here, one couldtry other variations, such as the SPHT algorithm [50],versions using adaptive thresholds [38], or extensionsusing statistical rate-distortion optimizations [71]. The

use of the lifting realization of the DWT in losslesscompression, with its multiresolution, low complexityand structurally guaranteed perfect inversion charac-teristics, has given interesting results in our study andit is perhaps a worthwhile way of extending the basicsystem.

The existence of powerful and cost-effective solu-tions to the problem of MI compression will be of ma-jor importance in the envisaged telemedical informa-tion society. We feel that the proposed framework isin the right direction towards the achievement of thisgoal.

References

[1] D.R. Aberle, et al. The effect of irreversible imagecompression on diagnostic accuracy in thoracic imaging,Invest. Radiology 28 (5) (1993) 398–403.

[2] American College of Radiology (ACR)/National ElectricalManufacturers Association (NEMA) Standards Publicationfor Data Compression Standards, NEMA Publication PS-2,Washington, DC, 1989.

[3] M. Antonini, et al. Image coding using wavelet transform,IEEE Trans. Image Process. 1(2) (1992) 205–220.

[4] A. Baskurt, I. Magnin, Adaptive coding method of X-raymammograms, in: Y. Kim (Ed.), Medical Imaging V: ImageCapture, Formatting, and Display, Proc. SPIE 1444 (1991)240–249.

[5] V. Bhaskaran, K. Konstantinides, Image and VideoCompression Standards: Algorithms and Architectures,Kluwer, Dordrecht, 1995.

[6] A.R. Calderbank et al., Wavelet transforms that map integersto integers, Technical Report, Department of Mathematics,Princeton University, 1996.

[7] J. Chen, M.J. Flynn, B. Gross, D. Spizarny, Observer detectionof image degradation caused by irreversible data compressionprocesses, in: Y. Kim (Ed.), Medical Imaging V: ImageCapture, Formatting, and Display, Proc. SPIE 1444 (1991)256–264.

[8] K. Chen, T.V. Ramabadran, Near-lossless compression ofmedical images through entropy-coded DPCM, IEEE Trans.Med. Imag. 13 (3) (1994) 538–548.

[9] P.S. Cho, K.K. Chan, B.K.T. Ho, Data storage andcompression, in: Ref. [22], pp. 71–81.

[10] R. Claypoole et al., Nonlinear wavelet transforms forimage coding, Technical Report, Bell Laboratories, LucentTechnologies, 1997.

[11] P.C. Cosman, et al. Thoracic CT images: Effect of lossyimage compression on diagnostic accuracy, Radiology 190(1994) 517–524.

[12] P.C. Cosman, R.M. Gray, R.A. Olshen, Evaluating qualityof compressed medical images: SNR, subjective rating, anddiagnostic quality, Proc. IEEE 82 (6) (1994) 919–932.

Page 20: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

242 E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243

[13] G.G. Cox, et al. The effects of lossy compression on thedetection of subtle pulmonary nodules, Med. Phys. 23 (1)(1996) 127–132.

[14] I. Daubechies, W. Sweldens, Factoring wavelet transformsinto lifting steps, Technical Report, Bell Laboratories, LucentTechnologies, 1996.

[15] G.M. Davis, A wavelet-based analysis of fractal imagecompression, IEEE Trans. Image Process. 7(2) (1998) 141–154.

[16] S. Dewitte, J. Cornelis, Lossless integer wavelet transform,IEEE Signal Process. Lett. 4 (6) (1997) 158–160.

[17] Digital Imaging and Communication in Medicine (DICOM)version 3, American College of Radiology (ACR)/NationalElectrical Manufacturers Association (NEMA) StandardsDraft.

[18] J.H. Dripps, K. Boddy, G. Venters, TELEMEDICINE:Requirements, standards and applicability to remote casescenarios in Europe, in: J. Noothoven van Goor, J.P.Christensen (Eds.), Advances in Medical Informatics, IOSPress, 1992, pp. 340–348.

[19] W.H.R. Equitz, T.M. Cover, Successive refinement ofinformation, IEEE Trans. Inform. Theory 37(2) (1991) 269–275.

[20] R.A. Greenes, J.F. Brinkley, Radiology systems, in: E.H.Shortliffe, L.E. Perreault (Eds.), Medical Informatics –Computer Applications in Health Care, Addison-Wesley,Reading, MA, 1990.

[21] P.G. Howard, J.S. Vitter, Arithmetic coding for datacompression, Proc. IEEE 82 (6) (1994) 857–865.

[22] H.K. Huang et al., Picture Archiving and CommunicationSystems (PACS), NATO ASI Series, vol. F74, PACS inMedicine, Springer, Berlin, 1991.

[23] D.A. Huffman, A method for the construction of minimumredundancy codes, Proc. IRE 40 (1952) 1098–1101.

[24] A.K. Jain, P.M. Farrelle, V.R. Algazi, Image data compression,in: M.P. Ekstrom (Ed.), Digital Image Processing Techniques,Academic Press, New York, 1984, pp. 171–226.

[25] N. Jayant, Signal compression: Technology targets andresearch directions, IEEE J. Selected Areas in Commun. 10(5) (1992) 796–818.

[26] N.J. Jayant, P. Noll, Digital Coding of Waveforms, Prentice-Hall, Englewood Cliffs, NJ, 1984.

[27] R.G. Jost et al., High resolution teleradiology applicationswithin the hospital, in: R.G. Jost (Ed.), Medical Imaging V:PACS Design and Evaluation, Proc. SPIE 1446, (1991) 2–9.

[28] G.R. Kuduvalli, R.M. Rangayyan, Performance analysis ofreversible image compression techniques for high-resolutiondigital teleradiology, IEEE Trans. Med. Imag. 11(3) (1992)430–445.

[29] H. Lee, et al. A predictive classified vector quantizer andits subjective quality evaluation for X-ray CT images, IEEETrans. Med. Imag. 14 (2) (1995) 397–406.

[30] A.S. Lewis, G. Knowles, Image compression using the 2-Dwavelet transform, IEEE Trans. Image Process. 1(2) (1992)244–250.

[31] J.S. Lim, Two-Dimensional Signal and Image Processing,Prentice-Hall, Englewood Cliffs, NJ, 1990.

[32] S.-C.B. Lo et al., Full-frame entropy coding for radiologicalimage compression, in: Y. Kim (Ed.), Medical Imaging V:Image Capture, Formatting, and Display, Proc. SPIE 1444(1991) 265–271.

[33] H. MacMahon, et al. Data compression: Effect on diagnosticaccuracy in digital chest radiography, Radiology 178 (1991)175–179.

[34] A. Manduca, Compressing images with wavelet/subbandcoding, IEEE Eng. Med. Biol., Sept./Oct. 1995, pp. 639–646.

[35] M. Mansuripur, Introduction to Information Theory, Prentice-Hall, Englewood Cliffs, NJ, 1987.

[36] K.M. McNeill, et al. Evaluation of the ACR-NEMA standardfor communications in digital radiology, IEEE Trans. Med.Imag. 9 (3) (1990) 281–289.

[37] A. Macovski, Medical Imaging Systems, Prentice-Hall,Englewood Cliffs, NJ, 1983.

[38] A. Munteanu et al., Performance evaluation of the waveletbased techniques used in the lossless and lossy compressionof medical and preprint images, IRIS, TR-0046, 1997.

[39] M. Nelson, J.-L. Gailly, The Data Compression Book, M&TBooks, 1996.

[40] R. Passariello et al., PACS-IMACS: Operation evaluationand basic requirements for prospective evolution of PACStechnology, in: J. Noothoven van Goor, J. P. Christensen(Eds.), Advances in Medical Informatics, IOS Press, 1992,pp. 295–299.

[41] W.B. Pennebaker et al., Papers on arithmetic coding, IBM J.Res. Develop. 32 (6) (1988).

[42] A. Ramaswamy, W.B. Mikhael, A mixed transform approachfor efficient compression of medical images, IEEE Trans.Med. Imag. 15(3) (1996) 343–352.

[43] K.R. Rao, J.J. Hwang, Techniques and Standards for Image,Video and Audio Coding, Prentice-Hall, Englewood Cliffs,NJ, 1996.

[44] K.R. Rao, P. Yip, Discrete Cosine Transform-Algorithms,Advantages, Applications, Academic Press, New York, 1990.

[45] O. Rioul, M. Vetterli, Wavelets and signal processing, IEEESignal Process. Mag., Oct. 1991, pp. 14–38.

[46] E.A. Riskin, et al. Variable rate vector quantization formedical image compression, IEEE Trans. Med. Imag. 9(3)(1990) 290–298.

[47] O. Rompelman, Medical image compression: Possibleapplications of subband coding, in: Ref. [70], pp. 319–352.

[48] P. Roos, et al. Reversible intraframe compression of medicalimages, IEEE Trans. Med. Imag. 7(4) (1988) 328–336.

[49] P. Roos, M.A. Viergever, Reversible image data compressionbased on HINT decorrelation and arithmetic coding, in: Y.Kim (Ed.), Medical Imaging V: Image Capture, Formatting,and Display, Proc. SPIE 1444 (1991) 283–290.

[50] A. Said, W.A. Pearlman, A new, fast, and efficient imagecodec based on set partitioning in hierarchical trees, IEEETrans. Circuits and Systems for Video Techn. 6(3) (1996)243–250.

[51] P. Saipetch et al., Applying wavelet transforms with arithmeticcoding to radiological image compression, IEEE Eng. Med.Biol., Sept./Oct. 1995, pp. 587–593.

[52] J.M. Shapiro, An embedded wavelet hierarchical image coder,Proc. ICASSP’92, pp. IV-657–IV-660.

Page 21: Wavelet-based medical image compression - … · for an image compression scheme to be acceptable for ... Compression techniques based on the wavelet de-composition of the image have

E. Kofidis et al. / Future Generation Computer Systems 15 (1999) 223–243 243

[53] J.M. Shapiro, Embedded image coding using zerotrees ofwavelet coefficients, IEEE Trans. Signal Process. 41(12)(1993) 3445–3462.

[54] J.M. Shapiro, Smart compression using the embedded zerotreewavelet (EZW) algorithm, Proceedings of the 27th AnnualAsilomar Conference on Signals, Systems and Computers,1993, pp. 486–490.

[55] L. Shen, R.M. Rangayyan, A segmentation-based losslessimage coding method for high-resolution image compression,IEEE Trans. Med. Imag. 16 (3) (1997) 301–307.

[56] A. Signoroni, R. Leonardi, Progressive medical imagecompression using a diagnostic quality measure on regions-of-interest, Proc. EUSIPCO’98.

[57] R.L. Smathers, W.R. Brody, Digital radiography: current andfuture trends, British J. Radiology 58 (688) (1985) 285–307.

[58] G. Strang, T. Nguyen, Wavelets and Filter Banks, Wellesley–Cambridge Press, 1996.

[59] W. Sweldens, The lifting scheme: A new philosophy inbiorthogonal wavelet construction, in: A.F. Laine, M. Unser(Eds.), Wavelet Applications in Signal and Image ProcessingIII, Proc. SPIE 2569 (1995) 68–79.

[60] G.R. Thoma, L.R. Long, Compressing and transmitting visiblehuman images, IEEE MultiMedia, April–June 1997, pp. 36–45.

[61] P.P. Vaidyanathan, Multirate Systems and Filter Banks,Prentice-Hall, Englewood Cliffs, NJ, 1993.

[62] L. Vandendorpe, Optimized quantization for image subbandcoding, Signal Process. Image Commun. 4 (1991) 65–79.

[63] M. Vetterli, Multi-dimensional subband coding: Some theoryand algorithms, Signal Process. 6 (1984) 97–112.

[64] M. Vetterli, J. Kovacevic, Wavelets and Subband Coding,Prentice-Hall, Englewood Cliffs, NJ, 1995.

[65] J.D. Villasenor, B. Belzer, J. Liao, Wavelet filter evaluationfor image compression, IEEE Trans. Image Process. 4 (8)(1995) 1053–1060.

[66] G.K. Wallace, The JPEG still picture compression standard,IEEE Trans. Consumer Electr. 38 (1) (1992) 18–34.

[67] P.H. Westerink, J. Biemond, D.E. Boekee, Progressivetransmission of images using subband coding, Proc.ICASSP’89, pp. 1811–1814.

[68] I.H. Witten, R.M. Neal, J.G. Cleary, Arithmetic coding fordata compression, Commun. ACM, 30 (6) (1987) 520–540.

[69] S. Wong, et al. Radiological image compression – A review,Proc. IEEE 83(2) (1995) 194–219.

[70] J.W. Woods (Ed.), Subband Image Coding, Kluwer, Dodrecht,1991.

[71] Z. Xiong, N.P. Galatsanos, M.T. Orchard, Marginal analysisprioritization for image compression based on a hierarchicalwavelet decomposition, Proc. ICASSP’93, pp. V-546–V-549.

[72] Z. Yang et al., Effect of wavelet bases on compressing digitalmammograms, IEEE Eng. Med. Biol., Sept./Oct. 1995, pp.570–577.