Chapter 5

Compression of Digital Holographic Data

Abstract

Both lossless and lossy compression methods have been attempted in compressing digital holographic data. In the early phase, the performance of lossless coding and quantization methods were mainly investigated, while recent research works are more focused on lossy compression with wavelets transform, as summarized in this chapter.

In the second part, we discuss in more detail several wavelet-based holographic data compression methods based on Fresnelet (Liebling et al., 2003; Darakis and Soraghan, 2006a), separable and nonseparable vector lifting schemes (Xing et al., 2014a, 2015), arbitrary packet decomposition and directional wavelet transform (Blinder et al., 2013, 2014), and Morlet transform (Viswanathan et al., 2014).

Keywords

Digital holographic data compression

Quantization

Transform

Wavelets

Fresnelet

Vector lifting schemes

Packet decomposition

Directional wavelet transform

Morlet transform

Even though holography was invented more than 60 years ago, it has received more attention recently thanks to tremendous achievements in computer and digital technologies. Numerous research works have resulted in significant progresses to generate high-resolution digital holograms with cost-effective solutions. However, compression of digital holographic data is still a relatively new field of research.

Figure 5.1 shows the overall schematic of a complete digital holographic communication system. The different components include processing steps for digital hologram acquisition, compression, transmission, decompression, and display.

f05-01-9780128028544
Figure 5.1 Digital holographic communication system components.

In detail, a digital hologram is first captured from a still 3D object by using a charged coupled device. For moving 3D objects, a temporal sequence of holograms is generated. Alternatively, the hologram can also be obtained by computational methods from virtual 3D object models. Next, the interference pattern should be converted into a suitable representation. The resulting representation is then compressed and encoded with properly designed coding tools. The encoded data are then transmitted. At the receiver side, the transmitted is first decoded. This allows reconstruction of the object wave, which can then be rendered on the holographic display device.

5.1 Overview of State of the Art for Digital Holographic Data Compression

Both lossless and lossy compression methods have been attempted in compressing digital holographic data. In the early phase, the performance of lossless coding and quantization methods were mainly investigated, while recent research works are more focused on lossy compression with wavelets transform. As summarized in the following, most of the research works on compressing holographic data are categorized into two kinds of approaches: quantization-based and transform-based methods.

5.1.1 Quantization Based

Some lossless and lossy data compression methods are applied to real-imaginary information obtained by phase-shifting interferometry (PSI) in Naughton et al. (2002). Lossless techniques such as Lempel-Ziv, Lempel-Ziv-Welch, Huffman, and Burrows-Wheeler are used to separately code real and imaginary components of the holographic data. However, due to the low spatial redundancies resulting from the speckle nature of holograms, lossless compression methods are usually inefficient (Darakis et al., 2006). As a consequence, lossy compression of holographic data seems essential.

Lossy compression techniques such as subsampling and quantization are applied as well in Naughton et al. (2002). Hologram resampling resulted in a high degradation in reconstructed image quality, but for resizing to a side length of 0.5, a compression rate of 18.6 could be achieved with a high degree of median filtering. Quantization proved to be a very effective technique. Each real and imaginary component could be reduced to 4 bits with an acceptable reconstruction error, resulting in a compression rate of 16.

Mills and Yamaguchi (2005) also confirmed the effectiveness of quantization in both numerical simulation and optical experiments. The influence of bit-depth reduction has been investigated by quantizing the spot-array patterns (the intensity information). They also pointed out that the use of 4 bits appears to be adequate for visual recognition. However, as the number of bits decreases, the resulting quality falls rapidly when only applying quantization.

Furthermore, Naughton et al. (2003) improved the quantization of real-imaginary information with a bit packing operation for real-time networking application. A speedup metric is also defined, which combines space gains due to compression with temporal overheads due to the compression routine and the transmission serialization. It has been reported that quantization to 4 bits results in a compression rate of 16, with normalized root mean square (NRMS) errors in the reconstructed object intensity lower than 0.06.

A companding histogram approach is applied for nonuniform quantization in Shortt et al. (2006b). A companding quantizer is a nonuniform quantizer composed of a compressor, a uniform quantizer, and an expander. A pattern-determined grid is used for the compander, which is fixed-interval sampled. Its pattern is firstly determined by quantization experiments with some DHs in order to produce a hard-coded pattern of clusters. In this way, the computation burden of iterations can be significantly reduced by considering the cluster pattern as a codebook. The compressor and expander works depending on the density of input data. If the input data are dense, the grid is compressed; contrarily, if the input data are sparse, the grid is expanded. The companding approach combines the efficiency of uniform quantization with the improved performance of nonuniform quantization. It exploits a priori knowledge of the distribution of the values in the data and performs well when the data distribution can be described. Shortt et al. first applied the companding quantization method for two reasons: (1) the nonuniform distribution of holographic data; and (2) the computation cost and time delay of iterative techniques. They developed two kinds of companding grid: the diamond companding grid and the logarithmic spiral one. The first one is generated based on a logarithmic sampling distribution, which performs as well as some iterative techniques, for example, k-means algorithm, with higher efficiency when the number of clusters are in the range of [25, 100]. The second companding grid is developed to improve the performance for smaller numbers of clusters. It uses a logarithmic spiral function to produce the cluster pattern. Improvements have been obtained over the performance of the diamond one.

Similarly, histogram quantization is proposed in Shortt et al. (2007). The histogram quantization technique is a noniterative nonuniform quantization technique, specially designed for digital holographic data with better exploitation of data distribution. In this technique, highest peaks in a representation, for example, real and imaginary data, are extracted to define clusters. The most frequently occurring value in each section is selected as the highest peak. The range ri of the ith section is defined by

ri={[a+iδ,a+(i+1)δ),ifi<N1[a+iδ,a+(i+1)δ],ifi=N1,

si1_e  (5.1)

where the range of the set of values is [a,b], N is the number of sections, δ=(ba)Nsi2_e and i ∈{0,1,…,N − 1}. The peaks from real values and imaginary values are then paired to give complex-valued clusters. Each pixel is quantized to the value of its nearest cluster. The significant advantage of this quantization technique is its less time-complexity under the same requirement of reconstruction quality, compared to iterative techniques. This quantization technique can also be considered as joint quantization, similar to VQ, as the two sets of holographic data are jointly quantized. Shortt et al. also applied this technique on amplitude-phase representation; however, better reconstruction performance is reported for real-imaginary representation, in agreement with the conclusion obtained above.

Differently, based on SQ, a multiple description coding (MDC) method is applied on amplitude-phase information using maximum-a-posteriori in Arrifano et al. (2013). It takes advantage of MDC for optimally coding data between available channels and mitigate channel errors.

A comparative study of several quantization schemes, USQ, ASQ, and VQ, is presented in Xing et al. (2014b). It is shown that jointly encoding the components can bring more benefits for real-imaginary and shifted distance representations that present higher inter-component redundancies. Conversely, the amplitude-phase representation is better encoded separately due to the larger influence of phase information on the reconstruction quality.

5.1.2 Transform Based

Transform-based coding is widely used for image compression, thanks to its ability to compact the signal energy efficiently. Therefore, researchers have naturally investigated the use of transform-based coding, and more specifically wavelet transforms, for the compression of holograms.

Shortt et al. (2006a) introduced wavelet analysis for the compression of real-imaginary components. A 1D-DWT is applied to each component with different resolution levels. The wavelet coefficients are then quantized. It has been found that three levels performed the best on average. In addition, compression of optically obtained and computer-generated phase-shifting interference patterns by standard JPEG and JPEG 2000 compression techniques are, respectively, addressed in Darakis and Soraghan (2006b), with reported compression ratios in the range 20-27 at acceptable reconstruction levels, and Xing et al. (2013a) showing the higher performance of JPEG 2000.

However, standard wavelets are typically designed to process piecewise smooth signals, while holograms contain features which are spread out from the objects. Therefore, applying standard wavelets directly to the hologram is not efficient. For this purpose, Liebling et al. (2003) developed a family of wavelet bases—Fresnelets—obtained by applying the Fresnel transform operator to B-splines bases, which is specially tailored to the specificities of DH. The particular suitability of the Fresnel B-splines has been concluded.

Darakis and Soraghan (2006a) introduced the use of Fresnelets into phase-shifting digital holography (PSDH) for holographic data compression. The real-imaginary components are first separately decomposed to Fresnelet coefficients to a required scale depth. The real and imaginary Fresnelet coefficients are then fed to the SPIHT algorithm. Experimental results verified its extensive flexibility for the compression of PSI holographic data.

However, Viswanathan et al. (2013) analyzed that Fresnelets have limitations for showing localization in frequency regarding a viewpoint-based degraded reconstruction. Instead, they proposed to use Gabor wavelets, which are suitable for measuring the local spatial frequencies so that the coefficients can be pruned corresponding to a viewpoint selection. Also, the angular spectrum method is used to perform the reconstruction instead of the Fresnel transform. The experimental results prove that Gabor wavelets are able to suppress the unwanted orders created in the reconstruction for off-axis holograms and have better time-frequency localization for view-dependent compression techniques. Moreover, they proposed to use Morlet wavelets for transforming a hologram and partly reconstructing a scene by using a sparse set of Morlet transformed coefficients in Viswanathan et al. (2014). It has been shown that a view-dependent representation together with Morlet wavelets form a good starting step for compressing holographic data for next generation 3DTV applications.

On the other hand, Blinder et al. (2013) also investigated wavelet coding on off-axis holograms. Differently, the properties of the off-axis holograms are first examined by independent component analysis, which reveals the importance of orientation and high frequencies in off-axis holograms. For this reason, standard decomposition schemes are not suitable for compressing holograms. Based on the standard JPEG 2000 algorithm, some alternative decomposition schemes are proposed to decompose further the high-frequency subbands. They are combined with direction-adaptive wavelets (Chang and Girod, 2007). Significant improvements have been reported for lossy compression compared with the standard DWT using the Mallat decomposition. Furthermore, they proposed a wavelet packet decomposition scheme combined with directional wavelet transforms for compressing microscopic off-axis holograms in Blinder et al. (2014). Still, the JPEG 2000 standard is modified to cope with holograms by applying the new wavelet decomposition with directional wavelet transforms. This extended JPEG 2000 algorithm shows higher compression efficiency.

Wavelet-based methods are considered effective for compressing holographic data. However, all the coding schemes described above encode different holographic data independently. It could be more efficient to combine the wavelet transform and joint encoding methods to compress phase-shifting holographic data.Xing et al. (2014a) first introduced a joint coding method based on the concept of vector lifting scheme (VLS) to compress shifted distance information. Experimental results indicate the advantage of applying this joint coding scheme. A significant gain of about 2 dB and 0.15 in terms of PSNR and structural similarity (SSIM), respectively, has been achieved compared to independent JPEG 2000 coding scheme. Subsequently, the scheme has been further improved by changing the decomposition structure from a separable to a nonseparable scheme in Xing et al. (2015), resulting in significant improvements.

Several wavelet-based holographic data compression methods are further detailed in Section 5.2.

5.1.3 Extension to Digital Holographic Sequences

Some works on compressing hologram sequences have also been reported. However, most methods are simply based on 2D video compression techniques.

Seo et al. (2006) used a multiview prediction technique and a temporal motion prediction technique to remove the spatial and temporal data redundancies. The predicted and compensated data are then compressed by an MPEG-2 encoder. Enhanced performances have been reported.

The use of MPEG-4 advanced simple profile (ASP) (Pereira and Ebrahimi, 2002) for the compression of hologram sequences was investigated by Darakis and Naughton (2009). Although it has been designed for conventional video, the scheme is effective and achieves good reconstruction quality. Inter-frame coding is also shown to outperform intra-frame coding. In other words, the scheme successfully exploits temporal redundancies in the hologram sequences.

A 3D scanning method is introduced in Seo et al. (2007). More specifically, interference patterns are divided into blocks, and 2D discrete cosine transform (DCT) is performed. The resulting segments are then scanned in 3D, in order to form a video sequence. Finally, the resulting video sequence is encoded using the advanced video coding (AVC) standard—H.264/AVC (Wiegand et al., 2003). The authors show the effectiveness of the scheme and claim significantly improved compression performance when compared to earlier works.

Two video coding schemes, namely H.264/AVC (Wiegand et al., 2003) and the wavelet-based (Dirac, 2009), are compared in Darakis et al. (2010). The authors performed subjective experiments in order to identify the threshold of visually lossless quality. More specifically, the observers are successively shown two sequences in random order, the uncompressed reference and a compressed version, and are asked to identify the compressed one. The bitrate of the compressed sequence is varied until the just noticeable difference threshold is found. While performance is content-dependent, it is shown that compression ratios up 7.5 can be achieved with visually lossless quality.

The recent high efficiency video coding (HEVC) state-of-the-art video coding standard (Sullivan et al., 2012) has also been investigated to compress hologram sequences. In Xing et al. (2013b), HEVC is applied to holograms obtained from animated virtual objects using computer-generated phase-shifting holography. HEVC is shown to achieve high reconstruction quality with a bit rate of 15 Mbps. Moreover, HEVC is demonstrated to be significantly superior to H.264/AVC. Subjective quality assessment experiments using HEVC are also reported in Ahar et al. (2015).

5.2 Wavelet-Based Holographic Data Compression Methods

In this section, we discuss in more details several wavelet-based holographic data compression methods based on Fresnelet (Liebling et al., 2003; Darakis and Soraghan, 2006a), separable and nonseparable vector lifting schemes (Xing et al., 2014a, 2015), arbitrary packet decomposition and directional wavelet transform (Blinder et al., 2013, 2014), and Morlet transform (Viswanathan et al., 2014).

5.2.1 Fresnelet Scheme

Fresnelets are a family of B-spline wavelet basis functions, which are specially designed for processing digital holograms (Liebling et al., 2003). Wavelets are efficient for piecewise smooth signals, such as natural images. However, holograms have very different characteristics, and in this case, wavelets do not perform optimally. Fresnelets have been developed in order to address this drawback. They have been successfully applied to compress off-axis and phase-shifting holographic data (Darakis and Soraghan, 2006a). In this part, the basic principles of the design of Fresnelets will be represented.

As explained in Chapter 2, the Fresnel transform of the complex object wave U(x,y) in the wave traveling direction z = 0 to the distance z = d can be expressed by 2D convolution integral as

^Uτ(ξ,η)=exp(ikd)iλd++U(x,y)×exp[iπλd((ξx)2+(ηy)2)]dxdy=iexp(ikd)(U*Kτ)(ξ,η),

si3_e  (5.2)

with

Kτ(x,y)=1τ2exp[iπτ2(x2+y2)]=kτ(x)kτ(y),τ=λd,

si4_e  (5.3)

where * denotes convolution and Kτ(x,y) is the separable kernel.

The Fresnelet bases are the Fresnel transform of B-splines. B-splines are selected to generate a multiresolution analysis of L2 because they satisfy the requirements of a valid scaling function of L2. One-dimensional B-splines of degree n are defined as the (n + 1)-fold convolution of a rectangular pulse (Unser, 1999):

βn(x)=β0**β0n+1times(x),

si5_e  (5.4)

where

β0(x)={1,if12<x<1212,if|x|=120,otherwise.

si6_e  (5.5)

An alternative equivalent definition of the B-splines is:

βn(x)=Δn+1*(x)n+n!,

si7_e  (5.6)

where Δn+1 is the (n + 1)th centered finite-difference operator:

Δn+1=n+1k=0(1)kCkn+1δ(x+n+12k)

si8_e  (5.7)

and (x)n+=max(0,x)nsi9_e is the one-sided power function. So the explicit central B-splines can be represented by

βn(x)=n+1k=0(1)kCkn+1(x+n+12k)n+n!.

si10_e  (5.8)

More specially, a two-scale relation of the form can be shown as

βn(x2)=kh(k)βn(xk),

si11_e  (5.9)

where h(k) is the binomial filter:

h(k)=12nCkn+1.

si12_e  (5.10)

Moreover, in Unser (1999), a general family of semi-orthogonal spline wavelets in a two-scale relation has been shown of the form:

ψn(x2)=kg(k)βn(xk),

si13_e  (5.11)

so a Riesz basis of L2 is formed by the functions

{ψni,k=2i2ψ(2ixk)}.

si14_e  (5.12)

These wavelets are linear combinations of B-splines which are specified by g(k), while g(k) is the quadrature mirror filter of h(k), which forms Riesz bases of L2.

Fresnelets are then obtained by Fresnel transformed B-spline bases:

ˆβnτ(x)=(βn*kτ)(x).

si15_e  (5.13)

They can be used to decompose the signals into different scales. Particularly, for the two-scale relation, they become

ˆβnτ2(x2)=kh(k)ˆβn(xk).

si16_e  (5.14)

The semi-orthogonal ones corresponding to Eq. (5.11) can be derived in the same way.

In order to apply Fresnelets in 2D data, it can be simply derived by extending B-splines to 2D as

βn(x,y)=βn(x)βn(y),

si17_e  (5.15)

where ⋅ denotes the tensor product. The Fresnelets coefficients of the complex wave Ûτ0si18_e at scale j = 1 can be obtained by

LL(x2,y2)=Ûτ0*[ˆβnτ0(x2)ˆβnτ0(y2)]HL(x2,y2)=Ûτ0*[ˆψnτ0(x2)ˆβnτ0(y2)]LH(x2,y2)=Ûτ0*[ˆβnτ0(x2)ˆψnτ0(y2)]HH(x2,y2)=Ûτ0*[ˆψnτ0(x2)ˆψnτ0(y2)].

si19_e  (5.16)

For more scales decomposition, the subband LL needs to be decomposed further.

In Darakis and Soraghan (2006a), the obtained Fresnelets coefficients obtained from real-imaginary information are further quantized by uniform quantization and compressed by SPIHT coding. The comparison is conducted between lossless B-spline wavelets transform coded hologram and SPIHT coded Fresnelet coefficients. For the same compression rate, the Fresnelets-based scheme leads to smaller NRMS.

5.2.2 Separable Vector Lifting Scheme

As shown in Chapter 4, in the cases of shifted distance and real-imaginary representations, the two components of the hologram show similar visual content. Thus, it is expected that these kinds of data include redundancies, and from this point of view, efficient hologram compression schemes can be designed by exploiting the dependencies between these patterns. For this purpose, Xing et al. (2014a, 2015) first proposed to apply a joint coding method, based on the concept of a vector lifting scheme (VLS), to phase-shifting holographic data.

Most of the existing joint coding schemes, which have been developed in the literature for video and stereo/multiview data compression purpose, consist of two steps. Assuming that two input correlated images are available to be encoded, the first step consists of selecting one image as a reference and encoding it independently of the other one. Then, the second image, selected as a target image, is predicted from the first one, and the difference between the two images, called the residual, is encoded. Typically, DCT or DWT can be used for encoding both the reference and residual images (Moellenhoff and Maier, 1998; Boulgouris and Strintzis, 2002). Contrary to this standard scheme, the main feature of VLS is that it does not generate a residual image, but two compact representations of both images. VLS is an extended scheme based on lifting scheme (LS) (Sweldens, 1996; Hampson and Pesquet, 1998; Kaaniche et al., 2011a, 2012). LS is an alternative method for computing DWT, which is simpler and faster than the classical one. The original signal can also be easily obtained by the inverse transform. With these advantages, LS has been proven to be an efficient tool for still image coding. Many extensions have been applied to LS. One of them known as the quincunx lifting scheme has been found particularly useful for coding satellite images by using a quincunx grid (Gouze et al., 2004). Moreover, some directional transforms, such as oriented wavelet transform (Chappelier and Guillemot, 2006) and grouplet (Mallat, 2009), have also been developed with LS for image processing.

Figure 5.2 shows a generic separable LS structure for 1D signal with one level resolution. Biorthogonal wavelets can be constructed by a series of operators including split, predict, and update in the forward transform (analysis step).

f05-02-9780128028544
Figure 5.2 Generic lifting structure.

Split First, the input 1D signal sj is partitioned into two sample sets: the even samples sj(2n) and the odd samples sj(2n + 1).

Predict Next, predict each sample in one of the subsets by the neighboring samples in the other subset. A prediction error or detail signal, for the case of predicting odd sample (the same for the following contents), can be computed:

dj+1(n)=sj(2n+1)psj(n),

si20_e  (5.17)

where p is the prediction vector containing the prediction weights, sj(n)=(sj(2n2k))kPsi21_e is the reference vector containing the even samples for predicting the odd sample, and Psi22_e is the support of the predictor.

Update Then, a coarser approximation of the original signal is generated by a smoothing operation on the even samples using the detail coefficients:

sj+1(n)=sj(2n)+udj+1(n),

si23_e  (5.18)

where similarly, u is the update vector containing the weights of update operator, dj+1(n)=(dj+1(nk))kUsi24_e is the reference vector containing the detail coefficients, and Usi25_e is the support of the update operator.

It is obvious that the two sample subsets can be easily reconstructed from the forward transform by two inverse operators:

Undo update

sj(2n)=sj+1(n)udj+1(n),

si26_e  (5.19)

Undo predict

sj(2n+1)=dj+1(n)+psj(n).

si27_e  (5.20)

Then, they are merged to obtain the original signal. Above all, the LS is easier to implement than the filter bank-based method.

Similarly, the separable 2D-DWT can also be implemented by applying the 1D case to the lines and columns. Also, one can obtain a multiresolution representation by recursively repeating the analysis steps to the approximation coefficients.

Based on LS, VLS has been developed to encode jointly two sets of dependent signals which will be denoted by S(1) and S(2) (Benazza-Benyahia et al., 2002; Kaaniche et al., 2009). The block diagram of a separable vector lifting scheme is shown in Fig. 5.3.

f05-03-9780128028544
Figure 5.3 Principle of the VLS decomposition.

The principle of this multiscale decomposition is described for a given line x. In what follows, S(1)jsi28_e and S(2)jsi29_e designate the approximation coefficients of 2D signals, S(1) and S(2), at resolution level j. While j = 0 corresponds to the initial (full resolution) signal S(1) and S(2), note that the dimensions of S(1)jsi28_e and S(2)jsi29_e (j ≥ 1) are divided by 2j along the horizontal and vertical directions.

Decomposition of reference signal As can be seen in Fig. 5.3, the reference signal S(1) is firstly encoded using a classical lifting structure composed of a prediction and an update stage. To this end, for a given line x, the input signal S(1)j(x,y)si32_e is firstly partitioned into two datasets formed by the even S(1)j(x,2y)si33_e and odd samples S(1)j(x,2y+1)si34_e, respectively. Then, during the prediction step, each sample of one of the two subsets (say the odd ones) is predicted from the neighboring even samples, yielding the detail coefficients ˜d(1)j+1si35_e at the resolution (j + 1):

˜d(1)j+1(x,y)=S(1)j(x,2y+1)kP(1)jp(1)j,kS(1)j(x,2y2k),

si36_e  (5.21)

where the coefficients p(1)j,ksi37_e and the set P(1)jsi38_e represent, respectively, the weights and the support of the predictor of the odd samples S(1)j(x,2y+1)si34_e. After that, the update step aims at computing a coarser approximation ˜S(1)j+1si40_e of the original signal by smoothing the even sample S(1)j(x,2y)si33_e as follows:

˜S(1)j+1(x,y)=S(1)j(x,2y)+kU(1)1u(1)j,k˜d(1)j+1(x,yk),

si42_e  (5.22)

where the set U(1)jsi43_e denotes the spatial support of the update operator whose coefficients are u(1)j,ksi44_e.

Decomposition of target signal Once the reference signal S(1) is encoded in intra mode, the attention will be paid now to the target signal S(2). It is important to note that the main difference between a basic lifting scheme and the vector lifting scheme is that for the target signal S(2), the prediction step uses samples from itself and also their corresponding samples taken from the reference signal S(1).
As shown in Fig. 5.3, a P-U-P structure is used for the target signal S(2). More precisely, a first intra-prediction step is applied to generate an intermediate detail signal ˅d(2)j+1si45_e, which serves to compute the approximation signal ˜S(2)j+1si46_e through the update step. After that, an hybrid prediction is performed by exploiting simultaneously the intra- and inter-redundancies in order to compute the final detail signal ˜d(2)j+1si47_e. Thus, the resulting decomposition implies the following equations:

˅d(2)j+1(x,y)=S(2)j(x,2y+1)kP(2)jp(2)j,kS(2)j(x,2y2k),

si48_e  (5.23)

˜S(2)j+1(x,y)=S(2)j(x,2y)+kU(2)ju(2)j,k˅d(2)j+1(x,yk),

si49_e  (5.24)

˜d(2)j+1(x,y)=˅d(2)j+1(x,y)(kQjqj,k˜S(2)j+1(x,yk)+kP(1,2)jp(1,2)j,kS(1)j(x,2y+1k)),

si50_e  (5.25)

where P(2)jsi51_e (respectively, P(1,2)jsi52_e) is the spatial support of the intra-signal (respectively, inter-signals) whereas its weights are designated by p(2)j,ksi53_e (respectively, p(1,2)j,ksi54_e), and Qjsi55_e (respectively, qj,k) is the support (respectively, weights) of the second intra-signal predictor.
Since a separable decomposition has been considered, these steps are iterated on the columns y of the resulting subbands ˜S(1)j+1si40_e, ˜d(1)j+1si35_e, ˜S(2)j+1si46_e, ˜d(2)j+1si47_e in order to produce the approximation subbands S(1)j+1si60_e and S(2)j+1si61_e as well as three details subbands, for each signal, oriented horizontally, vertically, and diagonally. This decomposition is again repeated on the approximation subbands over J resolution levels, yielding the multiresolution representation of the two input signals.
Finally, at the last resolution level J, instead of encoding the approximation subband of the target signal S(2)jsi29_e, it is proposed to encode the residual subband given by:

e(2)J(x,y)=S(2)J(x,y)kP(1,2)Jp(1,2)J,kS(1)J(x,yk).

si63_e  (5.26)

Prediction and update filters Prediction and update filters with spatial supports P(1)j={1,0}si64_e and U(1)j={0,1}si65_e are considered for signal S(1). In addition, the weights of the update filter are set for all the resolution levels to: u(1)j,0=u(1)j,1=14si66_e. However, it is proposed to optimize the prediction weights p(1)j,1si67_e and p(1)j,0si68_e in order to design a coding scheme well adapted to the content of the hologram data. Since the detail coefficients can be viewed as prediction errors, the prediction filter coefficients can be optimized at each resolution level j by minimizing the variance of the detail signal ˜d(1)j+1si35_e. More specifically, the well-known Yule-Walker equations are applied to solve the optimization problem:

E[Sj(x,y)Sj(x,y)]pj=E[S(1)j(x,2y+1)Sj(x,y)],

si70_e  (5.27)

where

Sj(x,y)=(S(1)j(x,2y),S(1)j(x,2y+2)).

si71_e  (5.28)

pj=(p(1)j,0,p(1)j,1)si72_e is the prediction weighting vector and E[⋅] denotes the mathematical expectation.
Concerning the second signal S(2), the same intra-prediction and update filters used with S(1) will be employed to generate the signals ˅d(2)j+1si45_e and ˜S(2)j+1si46_e. Then, the second prediction stage is performed by setting Qj={1,0}si75_e and P(1,2)j={1,0,1}si76_e for j ∈{0,…,J − 1} and P(1,2)J={0}si77_e. The coefficients qj,k and p(1,2)j,ksi54_e are also optimized by minimizing the variance of the detail signal ˜d(2)j+1si47_e by Yule-Walker equations.

The subband coefficients resulting from the VLS decomposition are then quantized and encoded using the EBCOT algorithm similarly to the JPEG 2000 standard (see Chapter 4).

The effectiveness of the joint coding scheme based on separable VLS for holographic data compression purposes has been validated in Xing et al. (2014a). More specifically, the authors compared the performance of the VLS scheme with two reference methods using the shifted distance holographic data representation. The first method corresponds to the state-of-the-art hologram compression technique where the inputs D(1) and D(2) are separately encoded by using existing still image coders. To this end, the 9/7 wavelet transform retained in the lossy compression mode of JPEG 2000 has been used. In what follows, this scheme will be designated by “Independent.” The second one is the standard joint coding scheme, where the reference image D(1) and a residual one, given by D(2)D(1), are also encoded by applying the 9/7 transform. We recall that this technique has been considered in most of joint coding schemes developed in the context of stereo and video data compression. It will be designated by “Standard.” Note that JPEG 2000 has been used as an entropy encoder for these different hologram compression methods.Figure 5.4 shows the corresponding rate-distortion results in terms of PSNR, respectively, SSIM, versus the bitrate given in bit per pixel (bpp). Both PSNR and SSIM are computed between the original and reconstructed objects. For more details about the experimental results, readers can refer to Xing et al. (2014a).

f05-04-9780128028544
Figure 5.4 Rate-distortion performance of three different hologram compression schemes: independent, standard, and VLS, applied on D(1) and D(2), for the objects: (a) “Bunny-1,” (b) “Bunny-2,” and (c) “Girl.”

5.2.3 Nonseparable Vector Lifting Scheme

The VLS approach presented in Section 5.2.2 has been performed in a separable way by cascading the 1D decomposition along the horizontal direction, then along the vertical direction. However, according to the visual patterns of the difference data D(1) and D(2), it can be noticed that these signals present some structures that are neither horizontal nor vertical.

Therefore, to exploit better the characteristics of the hologram data, a nonseparable decomposition has been proposed in Xing et al. (2015), with the objective to build an efficient content-adaptive decomposition. This 2D nonseparable vector lifting scheme (NS-VLS) is described in more detail in this section. We use the same notations as in Section 5.2.2.

5.2.3.1 Basic Concept of Decomposition Structure

At each pixel location (x,y), the approximation coefficients of the first (respectively, second) image D(1)j(x,y)si80_e (respectively, D(2)j(x,y)si81_e) are divided into four polyphase components denoted by

{D(i)0,j(x,y)=D(i)j(2x,2y),D(i)1,j(x,y)=D(i)j(2x,2y+1),D(i)2,j(x,y)=D(i)j(2x+1,2y),D(i)3,j(x,y)=D(i)j(2x+1,2y+1),

si82_e  (5.29)

where i ∈{1,2}.

Figure 5.5 shows the NS-VLS analysis structure. One image (here D(1)) is selected as a reference image and encoded independently of the other one. To this end, a classical nonseparable lifting structure (Kaaniche et al., 2011b), composed of three prediction steps and an update one, is applied to D(1)jsi83_e in order to generate the detail signals oriented diagonally D(HH,1)j+1si84_e, vertically D(LH,1)j+1si85_e, and horizontally D(HL,1)j+1si86_e, as well as the approximation one D(1)j+1si87_e. Thus, the output wavelet coefficients, at the resolution level (j + 1), can be written as follows:

D(HH,1)j+1(x,y)=D(1)3,j(x,y)((P(HH,1)0,j)D(HH,1)0,j+(P(HH,1)1,j)D(HH,1)1,j+(P(HH,1)2,j)D(HH,1)2,j),

si88_e  (5.30)

D(LH,1)j+1(x,y)=D(1)2,j(x,y)((P(LH,1)0,j)D(LH,1)0,j+(P(LH,1)1,j)D-(HH,1)j+1),

si89_e  (5.31)

D(HL,1)j+1(x,y)=D(1)1,j(x,y)((P(HL,1)0,j)D(HL,1)0,j+(P(HL,1)1,j)¯D(HH,1)j+1),

si90_e  (5.32)

D(1)j+1(x,y)=D(1)0,j(x,y)+((U(HL,1)0,j)D(HL,1)j+1+(U(LH,1)1,j)D(LH,1)j+1+(U(HH,1)2,j)D(HH,1)j+1),

si91_e  (5.33)

where for every i ∈{0,1,2} and o ∈{HL, LH, HH},

 P(o,1)i,j=(p(o,1)i,j(s,t))(s,t)P(o,1)i,jsi92_e represents the prediction weighting vector whose support is denoted by P(o,1)i,jsi93_e;

 D(o,1)i,j=(D(1)i,j(x+s,y+t))(s,t)P(o,1)i,jsi94_e is a reference vector used to compute D(o,1)j+1(x,y)si95_e;

 D-(HH,1)j+1=(D(HH,1)j+1(x+s,y+t))(s,t)P(LH,1)1,jsi96_e and -D(HH,1)j+1=(D(HH,1)j+1(x+s,y+t))(s,t)P(HL,1)1,jsi97_e correspond to two reference vectors used in the second and the third prediction steps, respectively;

 U(o,1)i,j=(u(o,1)i,j(s,t))(s,t)U(o,1)i,jsi98_e is the update vector coefficients whose support is designated by U(o,1)i,jsi99_e; and

 D(o,1)j+1=(D(o,1)j+1(x+s,y+t))(s,t)U(o,1)i,jsi100_e is the reference vector containing the set of detail samples used in the update step.

f05-05-9780128028544
Figure 5.5 Proposed NS-VLS decomposition structure.

For the second image D(2)jsi101_e, selected as a target one, a joint wavelet decomposition is performed by taking into account its correlation with the reference one D(1)jsi83_e. More specifically, a lifting structure, similar to that used with D(1)jsi83_e, is firstly applied on D(2)jsi101_e in order to produce three intermediate detail signals ˜D(HH,2)j+1si105_e, ˜D(LH,2)j+1si106_e, and ˜D(HL,2)j+1si107_e, which will be used to compute the approximation coefficients D(2)j+1si108_e. Then, a second prediction stage is added.

Indeed, three hybrid prediction steps, which aim at exploiting simultaneously the intra- and inter-image redundancies, are applied in order to generate the final detail coefficients D(HH,2)j+1si109_e, D(LH,2)j+1si110_e, and D(HL,2)j+1si111_e. Thus, the final detail wavelet coefficients, at the resolution level (j + 1), are expressed as follows:

D(HH,2)j+1(x,y)=˜D(HH,2)j+1(x,y)((Q(HH,2)0,j)˜D(HH,2)0,j+1+(Q(HH,2)1,j)˜D(HH,2)1,j+1+(Q(HH,2)2,j)˜D(HH,2)2,j+1+(P(HH,1,2)0,j)D(HH,1)0,j+(P(HH,1,2)1,j)D(HH,1)1,j+(P(HH,1,2)2,j)D(HH,1)2,j+(P(HH,1,2)3,j)D(HH,1)3,j),

si112_e

D(LH,2)j+1(x,y)=˜D(LH,2)j+1(x,y)((Q(LH,2)0,j)˜D(LH,2)0,j+1+(Q(LH,2)1,j)D-(HH,2)j+1+(P(LH,1,2)0,j)D(LH,1)0,j+(P(LH,1,2)2,j)D(LH,1)2,j),

si113_e

D(HL,2)j+1(x,y)=˜D(HL,2)j+1(x,y)((Q(HL,2)0,j)˜D(HL,2)0,j+1+(Q(HL,2)1,j)¯D(HH,2)j+1+(P(HL,1,2)0,j)D(HL,1)0,j+(P(HL,1,2)1,j)D(HL,1)1,j),

si114_e

where for every i ∈{0,1,2,3} and o ∈{HL, LH, HH},

 Q(o,2)i,j=(q(o,2)i,j(s,t))(s,t)Q(o,2)i,jsi115_e is the intra-prediction weighting vector whose support is designated by Q(o,2)i,jsi116_e;

 P(o,1,2)i,j=(p(o,1,2)i,j(s,t))(s,t)P(o,1,2)i,jsi117_e is the hybrid prediction weighting vector whose support is denoted by P(o,1,2)i,jsi118_e;

 ˜D(o,2)0,j+1=(D(2)j+1(x+s,y+t))(s,t)Q(o,2)0,jsi119_e is a reference vector containing the approximation coefficients D(2)j+1si108_e used to compute the detail ones D(o,2)j+1(x,y)si121_e;

 ˜D(HH,2)1,j+1=(˜D(HL,2)j+1(x+s,y+t))(s,t)Q(HH,2)1,jsi122_e and ˜D(HH,2)2,j+1=(˜D(LH,2)j+1(x+s,y+t))(s,t)Q(HH,2)2,jsi123_e are two reference vectors, containing, respectively, the intermediate detail coefficients ˜D(HL,2)j+1si107_e and ˜D(LH,2)j+1si106_e, used to compute the final detail coefficients D(HH,2)j+1(x,y)si126_e;

 D-(HH,2)j+1=(D(HH,2)j+1(x+s,y+t))(s,t)Q(LH,2)1,jsi127_e and ¯D(HH,2)j+1=(D(HH,2)j+1(x+s,y+t))(s,t)Q(HL,2)1,jsi128_e are two intra-prediction vectors used to compute D(LH,2)j+1(x,y)si129_e and D(HL,2)j+1(x,y)si130_e; and

 D(o,1)i,j=(D(1)i,j(x+s,y+t))(s,t)P(o,1,2)i,jsi131_e is a vector containing some samples of the reference image D(1)jsi83_e used to exploit the inter-image redundancies during the computation of the final detail coefficients D(o,2)j+1(x,y)si121_e.

By repeating the same decomposition strategy on the approximation subbands over J resolution levels, two multiresolution representations of D(1) and D(2) are obtained. Finally, at the last resolution level J, instead of encoding the approximation subband of the target image D(2)jsi101_e, it would be interesting to exploit its correlation with D(1)jsi83_e and thus to encode the following residual subband e(2)Jsi136_e:

e(2)J(x,y)=D(2)J(x,y)p(1,2)JD(1)J(x,y),

si137_e  (5.34)

where p(1,2)Jsi138_e is a hybrid prediction coefficient that exploits the correlation between D(2)jsi101_e and D(1)jsi83_e.

5.2.3.2 Optimization Techniques

For the same reason as the VLS-based approach, the ultimate aim is to design an NS-VLS-based decomposition well adapted to the characteristics of the holographic data. Consequently, the optimization of all the operators is necessary. Due to the complexity of the NS-VLS structure, different strategies of optimization are applied:

Optimization of the predictors for D(1) For the first image D(1), the different prediction filters P(o,1)jsi141_e (with o ∈{HH, LH, HL}), used to generate the detail subbands D(o,1)j+1si142_e, are optimized at each resolution level by minimizing the variance of the detail coefficients. As mentioned before, the Yule-Walker equations should be satisfied.

Optimization of the update operators Different from the fixed assignment of the update operators in separable VLS, the update filter here at each level is optimized by minimizing the error between the approximation coefficients D(1)j+1si87_e and the decimated version of the signal resulting from an ideal low-pass filter. Due to the complexity of the equations for the application, the details will not be provided here. The readers can refer to Kaaniche et al. (2011b) for more details.

Optimization of the operators for D(2) For the second image D(2), all the prediction and update operators can also be optimized by adopting the same strategy used with D(1). However, since D(1) and D(2) present similar content, the optimization process of the first three prediction filters P(o,2)jsi144_e and the update one U(2)jsi145_e can be omitted by imposing these operators to be equal to those obtained with D(1). Thus, we have:

P(HH,2)j=P(HH,1)j,P(LH,2)j=P(LH,1)j,P(HL,2)j=P(HL,1)j,U(2)j=U(1)j.

si146_e  (5.35)

Note that this procedure presents the advantages of simplifying the optimization strategy and reducing the overhead cost corresponding to the number of filter coefficients that must be sent to the decoder. Finally, for the remaining hybrid prediction filters P(o,1,2)jsi147_e and Q(o,2)jsi148_e, they will be optimized by minimizing the variance of the detail coefficients D(o,2)j+1si149_e.

Figure 5.6 shows the improved performance of NS-VLS, compared to independent coding scheme, standard joint coding scheme, and separable VLS, when using the shifted distance holographic data representation.

f05-06-9780128028544
Figure 5.6 Rate-distortion performance of four different hologram compression schemes: independent, standard, separable VLS, and NS-VLS, applied on D(1) and D(2), for the objects: (a) “Luigi-1,” (b) “Luigi-2,” (c) “Bunny-1,” (d) “Bunny-2,” (e) “Girl,” and (f) “Teapot.”

In addition, different prediction filter lengths Lp have also been investigated. Indeed, increasing Lp may be interesting for two reasons. The first one is explained by the fact that the holographic data present repetitive circular structures similar to the propagation of waves. The second one is due to the objective of VLS, which consists of exploiting the inter-images redundancies through the prediction stage. In this respect, in addition to the case given by Lp = 2, other cases with Lp ∈{6,12,20,32} are also considered. The structures are designated by NS-VLS(2,2), NS-VLS(6,2), NS-VLS(12,2), NS-VLS(20,2), and NS-VLS(32,2). Some results are given in Fig. 5.7. Compared to the case Lp = 2, it can be observed that increasing this length up to 12 or 20 leads to greatly improved compression performance. However, very long filters (Lp = 32) lead to worse performance at low bitrate because of the expensive cost of encoding the filter coefficients compared to the case of small Lp values. Noted that the best filter length may differ according to the different content of holograms. For more details about the experimental results, readers can refer to Xing et al. (2015).

f05-07-9780128028544
Figure 5.7 Rate-distortion performance of NS-VLS with different prediction filter lengths for the objects: (a) “Luigi-1,” (b) “Luigi-2,” (c) “Bunny-1,” (d) “Bunny-2,” (e) “Girl,” and (f) “Teapot.”

5.2.4 Arbitrary Packet Decomposition and Direction-Adaptive Discrete Wavelet Transform

Digital holographic data exhibit different characteristics when compared to natural images. For instance, whereas images typically have a 1/f 2 power spectrum distribution, holographic data are characterized by significantly larger amount of high-frequency content. Moreover, the fringe patterns show a significant directionality. Straightforwardly, these characteristics need to be taken into account when designing a coding scheme.

In order to address the higher spatial frequency content, packet decompositions can be applied. More specifically, in contrast with the Mallat decomposition, packet decompositions further decompose the higher-frequency subbands, in addition to the decomposition of the low-frequency subbands, in order to achieve higher energy compaction. Such packet decompositions are supported in JPEG 2000 Part 2; however, the standard does not support a homogeneous decomposition style when the number of wavelet levels increases. For this purpose, a JPEG 2000 architecture compliant decomposition mechanism is proposed in Blinder et al. (2013, 2014) to fully enable arbitrary packet decompositions.

Exploiting directionality is also clearly an important property. For this purpose, in Blinder et al. (2013, 2014), it is proposed to replace the classical JPEG 2000 wavelet transform by a direction-adaptive discrete wavelet transform (DA-DWT). More precisely, the direction-adaptive wavelet filters introduced by Chang and Girod (2007) are used, in combination with the above wavelet packet decomposition.

The resulting coding architecture is shown in Fig. 5.8, where multiple DA-DWTs are applied on each evenly sized blocks of the hologram.

f05-08-9780128028544
Figure 5.8 JPEG 2000 architecture extended with an arbitrary packet decomposition and direction-adaptive discrete wavelet transform. Source: Adapted from Blinder et al. (2014)

The DA-DWT is also based on the idea of lifting scheme, to be precise, the directional lifting based. Moreover, it is able to represent more efficiently the sharp features in images. In order to explain DA-DWT, we follow the notations in Chang and Girod (2007). Let π={(lx,ly)Z2}si150_e be a 2D orthogonal sampling grid composed of four subgrids: πpq = {(lx,ly) ∈π|lx mod 2 = p,ly mod 2 = q}.S = {s[l]}, where s[l] = s[lx,ly] and l = (lx,ly), denotes a set of image samples on π.S0={s[l0]|l0π0=π00π01}si151_e and S1={s[l1]|l1π1=π10π11}si152_e denote the samples in even and odd rows, respectively.

Similar to the description of the lifting scheme in Section 5.2.2, the detail signal w1 = {w1[l1],l1 ∈π1} and approximation signal w0 = {w0[l0],l0 ∈π0} can be obtained by a prediction step and an update step in 1D vertical lifting of samples in S0, respectively, as follows:

w1[l1]=gH(s[l1]Ps,l1(S0)),l1π1w0[l0]=gL(s[l0]g1HUs,l0(w1)),l0π0,

si153_e  (5.36)

where gH, gL are scaling factors and P(⋅), U(⋅) are prediction and update operators, respectively. After further decomposition along the columns, four subbands in total are generated: w00 = LL defined on π00, w01 = LH defined on π01, w10 = HL defined on π10, and w11 = HH defined on π11.

In DA-DWT, the prediction and update operators with direction d = (dx,dy) are defined as

Pis,l1(s0)=KP1k=KPcP,k[l1(2k+1)d]Us,l0(w1)=KU1k=KUcU,k{l1|l1(2k+1)d*l1=l0}w1[l1],

si154_e  (5.37)

where d is defined such as to satisfy:

l1(2k+1)dπ0,l1π1,k=KP,,KP1,

si155_e  (5.38)

and i = 0,1,…,Nc − 1 is the direction index; Nc is the number of prediction directions; KP, KU, cP,k, and cU,k are defined by the adopted wavelet kernel; and dl1* denotes the direction selected at location l1 for Ps,l1(s0).

For each image block, the optimum direction d is selected by minimizing the prediction error w1[l1]. A speed-up process to search for the optimum direction has been also proposed in Stevens et al. (2008). For a multiscale decomposition, the directional lifting steps above are repeated for every LL subband.

Due to the high overhead cost for storing the direction indices of each subband, the DA-DWT are only applied for three main decompositions involved in the generation of the 1LL, 2LL, and 3LL subbands (Blinder et al., 2013).

Significant improvements in coding performance have been reported for off-axis holography and computer-generated holography compared to the conventional JPEG 2000 standard, with Bjøntegaard delta-peak signal-to-noise ratio improvements ranging up to 11 dB for lossy compression in the 0.125-2.00 bits per pixel (bpp) range, and bit-rate reductions of up to 1.6 bpp for lossless compression (Blinder et al., 2013, 2015).Table 5.1 demonstrates the significant improvements resulting from the wavelet packet decomposition and DA-DWT on five different test objects.

Table 5.1

BD-PSNR Improvements in (dB) Compared to the DWT Mallat, per Decomposition, With and Without DA-DWT

LossyPack1Pack2Full PacketDA-DWT+ MallatDA-DWT+ Pack1DA-DWT+ Pack2DA-DWT+ Full Packet
Neuron3.453.342.795.753.403.282.75
Erythocryte5.445.418.207.288.318.308.39
Microlenses0.060.302.102.892.652.612.37
Ball11.8111.9112.645.1412.7412.7412.57
Scratch−7.13−7.07−5.261.98−4.59−4.57−4.41
Average2.732.784.094.594.504.474.33

t0010

Source: Blinder et al. (2013).

5.2.5 Morlet Transform

The problem of viewpoint scalability is considered in Viswanathan et al. (2013). The authors address different aspects related to hologram display in the presence of a few viewers. In such a case, only a sparse set of holographic data is necessary to reconstruct a given viewpoint. In other words, the wavelet transform should allow coefficient pruning so that only relevant data are transmitted. To enable this view-based representation of the hologram, good space-frequency localization is needed. Since the wavelet families based on or derived from Gabor basis functions can provide the best space-frequency localization, in Viswanathan et al. (2014), Morlet wavelets are proposed to transform off-axis holograms.

Morlet wavelets are based on the concepts of Gabor functions. In order to design the Morlet wavelets, we start from the concepts of Gabor functions. A Gabor function is a Gaussian kernel multiplied by a sinusoid. Its 1D continuous form is defined as

gβ,f0(x)=Kexp(β2x2)exp(2iπf0x)β2=12σ2,

si156_e  (5.39)

where f0 is the frequency, K is the norm of the basis function, and σ is the standard deviation of the Gaussian function. In order to have equal number of oscillations for all frequencies, the setting

f0=12πβ

si157_e  (5.40)

is chosen. The 1D Morlet mother wavelet is further defined as

ψ(x)=Kexp(β2x2)(exp(2iπf0x)exp(π2β2f20)),

si158_e  (5.41)

where the term exp(π2β2f20)si159_e is introduced to eliminate the DC term of the off-axis hologram (Zhong 340et al., 2009). By extending Eq. (5.41) to 2D, we obtain

ψ(x,y)=Kexp(β2(x2+y2))(exp(2iπ(μ0x+ν0y))exp(π2β2(μ20+ν20))),

si160_e  (5.42)

where μ0 and ν0 are the spatial frequencies in x and y directions, respectively; K is the L2-norm of the basis. For the same purpose of having equal number of oscillations for all frequencies, the setting

μ20+ν20=π2β2

si161_e  (5.43)

is chosen. Consequently, we can obtain

exp(π2β2(μ20+ν20))=exp(π4)0.

si162_e  (5.44)

Representing the spatial frequencies μ0 and ν0 into polar form

μ0=F0cosθ,ν0=F0sinθ,

si163_e  (5.45)

where the variation of θ describes the variation of the spatial frequencies in x and y directions. Substituting Eqs. (5.40), (5.43), and (5.45) to Eq. (5.42), the 2D Morlet mother wavelet becomes

ψ(x,y)=Kexp(2f20(x2+y2)π2)exp(2iπ2β(xcosθ+ysinθ)),

si164_e  (5.46)

where the mother wavelet is centered at f0. In order to build the family of Morlet wavelets, the scaling parameter s for controlling the spatial and frequency resolution of decomposition is introduced to span the frequency plane, where the scaled frequencies are defined by

f=f0s.

si165_e  (5.47)

Therefore, the continuous 2D Morlet wavelet can be represented with the parameters s and θ by

ψs,θ(x,y)=Kexp(2f20s2π2(x2+y2))exp(2iπ2f0sπ(xcosθ+ysinθ)).

si166_e  (5.48)

For further application, the continuous transform has to be discretized to obtain the view-based representation of holograms. The discretization is mainly conducted from three aspects: the discretization of x and y, the discretization of s, and the discretization of θ.

The discretization of x and y is based on the consideration of the Gaussian function. Assume that the length of the Gaussian is L which is selected based on the tapering of the Gaussian function and there are N0 oscillations of the sinusoid with frequency f0, we have

L=N0f0.

si167_e  (5.49)

By Nyquist criterion, the sampling frequency fs should be twice of f0. We consequently have the actual size of the analog window function

Lω=Lfs=N0f02f0=2N0.

si168_e  (5.50)

The ration parameter θ is descretized by performing Nθ rotations of the sinusoid from 0 to π. The discrete spatial frequencies fx and fy in x and y directions are then represented by

fx=fccos(Θr)s,fy=fcsin(Θr)s,

si169_e  (5.51)

respectively, where Θ=πNθsi170_e, r is the discrete rotation parameter and fc=1Δsi171_e where Δ is the pixel period of the holographic display.

The discretization of the scale parameter s is dependent on the discrete viewer positions. Assume that θx and θy are the diffraction angles along x and y directions, respectively. Then the viewer plan can be divided into zones Rθx,θysi172_e, each of which corresponds to a unique Morlet transformed hologram zone Hθx,θysi173_e. The discrete diffraction angles are obtained by

θx=arcsin(λfccos(Θr)s),θy=arcsin(λfcsin(Θr)s).

si174_e  (5.52)

Above all, the discrete Morlet wavelet can be derived as

ψs,r(m,n)=Kexp(f2cs2π2(m2+n2))exp(2iπ2fcsπ(mcos(Θr)+nsin(Θr))),

si175_e  (5.53)

where N0mN0si176_e and N0nN0si177_e. The Morelet wavelet transform is then performed on the off-axis holograms.

References

Ahar A., Blinder D., Bruylants T., Schretter C., Munteanu A., Schelkens P. Subjective quality assessment of numerically reconstructed compressed holograms. In: Proc. SPIE, Applications of Digital Image Processing XXXVIII. 2015 San Diego, CA, USA.

Arrifano A., Antonini M., Pereira M. Multiple description coding of digital holograms using maximum-a-posteriori. In: European Workshop on Visual Information Processing. 2013:232–237.

Benazza-Benyahia A., Pesquet J.C., Hamdi M. Vector lifting schemes for lossless coding and progressive archival of multispectral images. IEEE Trans. Geosci. Remote Sens. 2002;40(9):2011–2024.

Blinder D., Bruylants T., Stijns E., Ottevaere H., Schelkens P. Wavelet coding of off-axis holographic images. Proc. SPIE. 2013;8856:88561L. doi:10.1117/12.2027114.

Blinder D., Bruylants T., Ottevaere H., Munteanu A., Schelkens P. JPEG 2000-based compression of fringe patterns for digital holographic microscopy. Opt. Eng. 2014;53(12):123102. doi:10.1117/1.OE.53.12.123102.

Boulgouris N., Strintzis M. A family of wavelet-based stereo image coders. IEEE Trans. Circuits Syst. Video Technol. 2002;12(10):898–903.

Chang C.L., Girod B. Direction-adaptive discrete wavelet transform for image compression. IEEE Trans. Image Process. 2007;16(5):1289–1302.

Chappelier V., Guillemot C. Oriented wavelet transform for image compression and denoising. IEEE Trans. Image Process. 2006;15(10):2892–2903. doi:10.1109/TIP.2006.877526.

Darakis E., Naughton T.J. Compression of digital hologram sequences using MPEG-4. In: Proc. SPIE, Holography: Advances and Modern Trends. 735811. doi:10.1117/12.820632. 2009;7358 URL.

Darakis E., Soraghan J. Use of Fresnelets for phase-shifting digital hologram compression. IEEE Trans. Image Process. 2006a;15(12):3804–3811. doi:10.1109/TIP.2006.884918.

Darakis E., Soraghan J.J. Compression of interference patterns with application to phase-shifting digital holography. Appl. Opt. 2006b. ;45(11):2437–2443. doi:10.1364/AO.45.002437. URL http://ao.osa.org/abstract.cfm?URI=ao-45-11-2437.

Darakis E., Naughton T.J., Soraghan J.J., Javidi B. Measurement of compression defects in phase-shifting digital holographic data. In: Proc. SPIE, Optical Information Systems IV. doi:10.1117/12.679445. 2006;6311 URL.

Darakis E., Kowiel M., Näsänen R., Naughton T.J. Visually lossless compression of digital hologram sequences. In: Proc. SPIE, Image Quality and System Performance VII. 2010.

Dirac. Dirac video compression. 2009. URL http://www.diracvideo.org/.

Gouze A., Antonini M., Barlaud M., Macq B. Design of signal-adapted multidimensional lifting scheme for lossy coding. IEEE Trans. Image Process. 2004;13(12):1589–1603. doi:10.1109/TIP.2004.837556.

Hampson F.J., Pesquet J.C. M-band nonlinear subband decompositions with perfect reconstruction. IEEE Trans. Image Process. 1998;7(11):1547–1560.

Kaaniche M., Benazza-Benyahia A., Pesquet-Popescu B., Pesquet J.C. Vector lifting schemes for stereo image coding. IEEE Trans. Image Process. 2009;18(11):2463–2475.

Kaaniche M., Benazza-Benyahia A., Pesquet-Popescu B., Pesquet J.C. Non separable lifting scheme with adaptive update step for still and stereo image coding. Signal Process. 2011a;91(12):2767–2782.

Kaaniche M., Benazza-Benyahia A., Pesquet-Popescu B., Pesquet J.C. Non-separable lifting scheme with adaptive update step for still and stereo image coding. Signal Process. 2011b. ;91(12):2767–2782. doi:10.1016/j.sigpro.2011.01.003. Advances in Multirate Filter Bank Structures and Multiscale Representations, URL http://www.sciencedirect.com/science/article/pii/S0165168411000041.

Kaaniche M., Pesquet-Popescu B., Benazza-Benyahia A., Pesquet J.C. Adaptive lifting scheme with sparse criteria for image coding. J. Adv. Signal Process. 2012;2012(1):10.

Liebling M., Blu T., Unser M. Fresnelets: new multiresolution wavelet bases for digital holography. IEEE Trans. Image Process. 2003;12(1):29–43. doi:10.1109/TIP.2002.806243.

Mallat S. Geometrical grouplets. Appl. Comput. Harmonic Anal. 2009. ;26(2):161–180. doi:10.1016/j.acha.2008.03.004. URL http://www.sciencedirect.com/science/article/pii/S1063520308000444.

Mills G., Yamaguchi I. Effects of quantization in phase-shifting digital holography. Appl. Opt. 2005. ;44(7):1216–1225. doi:10.1364/AO.44.001216. URL http://ao.osa.org/abstract.cfm?URI=ao-44-7-1216.

Moellenhoff M., Maier M. Transform coding of stereo image residuals. IEEE Trans. Image Process. 1998;7(6):804–812. doi:10.1109/83.679421.

Naughton T., Frauel Y., Javidi B., Tajahuerce E. Compression of digital holograms for three-dimensional object reconstruction and recognition. Appl. Opt. 2002;41(20):4124–4132.

Naughton T.J., McDonald J.B., Javidi B. Efficient compression of Fresnel fields for internet transmission of three-dimensional images. Appl. Opt. 2003. ;42(23):4758–4764. doi:10.1364/AO.42.004758. URL http://ao.osa.org/abstract.cfm?URI=ao-42-23-4758.

Pereira F., Ebrahimi T. The MPEG-4 Book. Upper Saddle River, NJ: Prentice Hall; 2002.

Seo Y.H., Choi H.J., Bae J.W., Yoo J.S., Kim D.W. Data compression technique for digital holograms using a temporally scalable coding method for 2-D images. In: 2006 IEEE International Symposium on Signal Processing and Information Technology. 2006:326–331.

Seo Y.H., Choi H.J., Kim D.W. 3D scanning-based compression technique for digital hologram video. Image Commun. 2007;22(2):144–156. doi:10.1016/j.image.2006.11.007.

Shortt A., Naughton T.J., Javidi B. Compression of digital holograms of three-dimensional objects using wavelets. Opt. Express. 2006a. ;14(7):2625–2630. doi:10.1364/OE.14.002625. URL http://www.opticsexpress.org/abstract.cfm?URI=oe-14-7-2625.

Shortt A.E., Naughton T.J., Javidi B. A companding approach for nonuniform quantization of digital holograms of three-dimensional objects. Opt. Express. 2006b. ;14(12):5129–5134. doi:10.1364/OE.14.005129. URL http://www.opticsexpress.org/abstract.cfm?URI=oe-14-12-5129.

Shortt A., Naughton T., Javidi B. Histogram approaches for lossy compression of digital holograms of three-dimensional objects. IEEE Trans. Image Process. 2007;16(6):1548–1556. doi:10.1109/TIP.2007.894269.

Stevens A.R., Munteanu J.C., Schelkens P. Optimized directional lifting with reduced complexity. In: EURASIP European Signal Processing Conference, EUSIPCO 2008, Lausanne, Switzerland; 2008:5.

Sullivan G., Ohm J., Han W.J., Wiegand T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012;22(12):1649–1668. doi:10.1109/TCSVT.2012.2221191.

Sweldens W. The lifting scheme: a custom-design construction of biorthogonal wavelets. Appl. Comput. Harmonic Anal. 1996;3(2):186–200.

Unser M. Splines: a perfect fit for signal and image processing. IEEE Signal Process. Mag. 1999;16(6):22–38. doi:10.1109/79.799930.

Viswanathan K., Gioia P., Morin L. Wavelet compression of digital holograms: towards a view-dependent framework. Proc. SPIE. 2013;8856:doi:10.1117/12.2027199.

Viswanathan K., Gioia P., Morin L. Morlet wavelet transformed holograms for numerical adaptive view-based reconstruction. Proc. SPIE. 2014;9216:92160G. doi:10.1117/12.2061588.

Wiegand T., Sullivan G., Bjontegaard G., Luthra A. Overview of the H.264/AVC video coding standard. IEEE Trans. Circuits Syst. Video Technol. 2003;13(7):560–576. doi:10.1109/TCSVT.2003.815165.

Xing Y., Pesquet-Popescu B., Dufaux F. Compression of computer generated hologram based on phase-shifting algorithm. In: European Workshop on Visual Information Processing. 2013:172–177.

Xing Y., Kaaniche M., Pesquet-Popescu B., Dufaux F. Vector lifting scheme for phase-shifting holographic data compression. Opt. Eng. 2014a;53(11):112312. doi:10.1117/1.OE.53.11.112312.

Xing Y., Pesquet-Popescu B., Dufaux F. Comparative study of scalar and vector quantization on different phase-shifting digital holographic data representations. In: 3DTV-Conference: The True Vision—Capture, Transmission and Display of 3D Video (3DTV-CON), 2014. 2014b:1–4.

Xing Y., Kaaniche M., Pesquet-Popescu B., Dufaux F. Adaptive nonseparable vector lifting scheme for digital holographic data compression. Appl. Opt. 2015. ;54(1):A98–A109. doi:10.1364/AO.54.000A98. URL http://ao.osa.org/abstract.cfm?URI=ao-54-1-A98.

Zhong J., Weng J., Hu C. Reconstruction of digital hologram by use of the wavelet transform. In: Advances in Imaging. Optical Society of America; 2009:DWB16. URL http://www.osapublishing.org/abstract.cfm?URI=DH-2009-DWB16.


2606 “To view the full reference list for the book, click here

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset