Institute for Systems Research
Permanent URI for this communityhttp://hdl.handle.net/1903/4375
Browse
Search Results
Item A Relationship between Quantization and Distribution Rates of Digitally Fingerprinted Data(2000) Karakos, Damianos; Papamarcou, Adrian; ISRThis paper considers a fingerprinting system where$2^{n R_W}$ distinct Gaussian fingerprints are embedded inrespective copies of an $n$-dimensional i.i.d. Gaussian image.Copies are distributed to customers in digital form, using$R_Q$ bits per image dimension.By means of a coding theorem, a rate regionfor the pair $(R_Q, R_W)$ is established such that (i) theaverage quadratic distortion between the original imageand each distributed copy does not exceed a specified level;and (ii) the error probability in decoding the embedded fingerprintin the distributed copy approaches zero asymptotically in $n$.Item A Practical Transmission System Based on the Human Visual Model for Satellite Channels(1999) Gu, Junfeng; Jiang, Yimin; Baras, John S.; Baras, John S.; ISR; CSHCNThis paper presents a practical architecture for joint source-channel coding of human visual model-based video transmission over a satellite channel. Perceptual distortion model just-noticeable-distortion (JND) is applied to improve the subjective quality of compressed videos. 3-D wavelet decomposition can remove spatial and temporal redundancy and provide the scalability of video quality.In order to conceal errors occurring under bad channel conditions, a novel slicing method and a joint source channel coding scenario that combines RCPC with CRC and utilizes the distortion information to allocate convolutional coding rates are proposed. A new performance index based on JND is proposed and used to evaluate the overall performance at different signal-to-noise ratios (SNR). Our system uses OQPSK modulation scheme.
The research and scientific content in this material has been submitted to Globecom'99. Item Combined Compression and Classification with Learning Vector Quantization(1998) Baras, John S.; Dey, Subhrakanti; ISRCombined compression and classification problems are becoming increasinglyimportant in many applications with large amounts of sensory data andlarge sets of classes. These applications range from aided target recognition(ATR), to medicaldiagnosis, to speech recognition, to fault detection and identificationin manufacturing systems. In this paper, we develop and analyze a learningvector quantization-based (LVQ) algorithm for the combined compressionand classification problem. We show convergence of the algorithm usingtechniques from stochastic approximation, namely, the ODE method. Weillustrate the performance of our algorithm with some examples.Item Accurate Segmentation and Estimation of Parametric Motion Fields for Object-based Video Coding using Mean Field Theory(1997) Haridasan, Radhakrishan; Baras, John S.; ISR; CSHCNWe formulate the problem of decomposing a scene into its constituent objects as one of partitioning the current frame into objects comprising it. The motion parameter is modeled as a nonrandom but unknown quantity and the problem is posed as one of Maximum Likelihood (ML) estimation. The MRF potentials which characterize the underlying segmentation field are defined in a way that the spatio-temporal segmentation is constrained by the static image segmentation of the current frame. To compute the motion parameter vector and the segmentation simultaneously we use the Expectation Maximization (EM) algorithm. The E-step of the EM algorithm, which computes the conditional expectation of the segmentation field, now reflects interdependencies more accurately because of neighborhood interactions. We take recourse to Mean Field theory to compute the expected value of the conditional MRF. Robust M-estimation methods are used in the M- step. To allow for motions of large magnitudes image frames are represented at various scales and the EM procedure is embedded in a hierarchical coarse-to-fine framework. Our formulation results in a highly parallel algorithm that computes robust and accurate segmentations as well as motion vectors for use in low bit rate video coding.This report has been submitted as a paper to the SPIE conference on Visual Communications and Image Processing - VCIP98 to be held in San Jose, California on Jan 24- 30, 1998. Item Wavelet Coding of Images: Adaptation, Scalability, and Transmission over Wireless Channels(1997) Jafarkhani, Hamid; Farvardin, N.; ISRIn this dissertation, we study the problem of image compression for storage and transmission applications separately. In addition to proposing new image coding systems, we consider different design constraints such as complexity and scalability.We propose a new classification scheme, dubbed spectral classification, which uses the spectral characteristics of the image blocks to classify them into one of a finite number of classes. The spectral classifier is used in adaptive image coding based on the discrete wavelet transform and shown to outperform gain-based classifiers while requiring a lower computational complexity. The resulting image coding system provides one of the best available rate-distortion performances in the literature. Also, we introduce a family of multiresolution image coding systems with different constraints on the complexity. For the class of rate-scalable image coding systems, we address the problem of progressive transmission and propose a method for fast reconstruction of a subband-decomposed progressively transmitted image.
Another important problem studied in this dissertation is the transmission of images over noisy channels, especially for the wireless channels in which the characteristics of the channel is time-varying. We propose an adaptive rate allocation scheme to optimally choose the rates of the source coder and channel coder pair in a tandem source-channel coding framework. Also, we suggest two adaptive coding systems for quantization and transmission over a finite-state channel using a combined source and channel coding scheme. Finally, we develop simple table- lookup encoders to reduce the complexity of channel-optimized quantizers while providing a slightly inferior performance. We propose the use of lookup tables for transcoding in heterogeneous networks
Item Low Complexity and High Throughput Fully DCT-Based Motion Compensated Video Coders(1996) Koc, Ut-Va; Liu, K.J.R.; ISRMany video coding standards such as H.261, MPEG1, MPEG2, HDTV and H.263, are based on the hybrid motion-compensated DCT approach. The common implementations adopt the conventional DCT-based motion-compensated video coder structure in which every raw video bit must go before being encoded through the performance- critical feedback loop consisting of a DCT, an Inverse DCT (IDCT) and a spatial-domain motion estimation/compensation. This heavily loaded feedback loop not only increases the overall complexity of the coder but also limits the throughput and becomes the bottleneck of a real-time high-end digital video system. In this dissertation, we propose a fully DCT-based motion-compensated video coder structure which eliminates IDCT and moves DCT out of the loop, resulting in a simple feedback loop with only one major component: Transform-Domain Motion Estimation/Compensation. Furthermore, different components can be jointly optimized if they operate in the same transform domain. Based on pseudo phases and sinusoidal orthogonal principles, we develop DCT Pseudo Phase Techniques to estimate displacement directly from the DCT coefficients of shifted signals/images. We develop the DCT-based motion estimation (DXT-ME) algorithm with the computational complexity, O(N2), compared to O(N4) for the full search block matching approach (BKM-ME). Simulation shows that the DXT-ME algorithm performs as well as, BKM-ME and other fast search approaches in terms of mean-square-error per pel (MSE) and bits per sample (BPS). Furthermore it has inherently highly parallel operations in computing the pseudo phases, suitable for VLSI implementation.We develop DCT-based subpixel motion estimation algorithms without the need of image interpolation as required by conventional methods, resulting in significant reduction in computations and data flow and flexible fully DCT-based coder design because the same hardware can support different levels of required accuracy. Simulation demonstrates comparable and even better performance of the DCT-based approach than BKM-ME in terms of MSE and BPS.
We devise the DCT-based motion compensation schemes via the bilinear and cubic interpolation without any increase in computations to achieve more accurate approximation and better visual quality for motion- compensated pictures. Less computations are required in DCT domain because of sparseness and block alignment of DCT blocks and use of fast DCT to reduce computations.
Item A Wavelet Approach to Wafer Temperature Measurement via Diffuse Reflectance Spectroscopy(1996) Krishnaprasad, Perinkulam S.; Kugarajah, Tharmarajah; Dayawansa, Wijesuriya P.; ISRA methodology for the determination of wafer temperature in Molecular Beam Epitaxy via diffuse reflectance measurements is developed. Approximate physical principles are not used, instead, patterns in the data (reflectance versus wavelength) are exploited via wavelet decomposition and Principal Component Analysis.Item Channel Codes That Exploit the Residual Redundancy in CELP- Encoded Speech(1996) Alajaji, Fady; Phamdo, N.; Fuja, Tom E.; ISRWe consider the problem of reliably transmitting CELP-encoded speech over noisy communication channels. Our objective is to design efficient coding/decoding schemes for the transmission of the CELP line spectral parameters (LSP's) over very noisy channels.We begin by quantifying the amount of ﲲesidual redundancy inherent in the LSP's of Federal Standard 1016 CELP. This is done by modeling the LSP's as first and second-order Markov chains. Two models for LSP generation are proposed; the first model characterizes the intra-frame correlation exhibited by the LSP's, while the second model captures both intra-frame and inter-frame correlation. By comparing the entropy rates of the models thus constructed with the CELP rates, it is shown that as many as one-third of the LSP bits in every frame of speech are redundant.
We next consider methods by which this residual redundancy can be exploited by an appropriately designed channel decoder. Before transmission, the LSP's are encoded with a forward error control (FEC) code; we consider both block (Reed- Solomon) codes and convolutional codes. Soft-decision decoders that exploit the residual redundancy in the LSP's are implemented assuming additive white Gaussian noise (AWGN) and independent Rayleigh fading environments. Simulation results employing binary phaseshift keying (BPSK) indicate coding gains of 2 to 5 dB over soft-decision decoders that do not exploit the residual redundancy.
Item Algorithm-Based Low-Power Digital Signal Processing System Designs(1995) Wu, A.Y.; Liu, K.J.R.; ISRIn most low-power VLSI designs, the supply voltage is usually reduced to lower the total power consumption. However, the device speed will be degraded as the supply voltage goes down. In order to meet the low-power/high-throughput constraint, the key issue is to ﲣompensate the increased delay so that the device can be operated at the slowest possible speed without affecting the system throughput rate.In this dissertation, new algorithmic- level techniques for compensating the increased delays based on the multirate approach are proposed.
Given the digital signal processing (DSP) problems, we apply the multirate approach to reformulate the algorithms so that the desired outputs can be obtained from the decimated input sequences. Since the data rate in the resulting multirate architectures is M- times slower than the original data rate while maintaining the same throughput rate, the speed penalty caused by the low supply voltage is compensated at the algorithmic/architectural level.
This new low-power design technique is applied to several important DSP applications. The first one is a design methodology for the low- power design of FIR/IIR systems. By following the proposed design procedures, users can convert a speed-demanding system function into its equivalent multirate transfer function. This methodology provides a systematic way for VLSI designers to design low- power/high-speed filtering architectures at the algorithmic/architectural level.
The multirate approach is also applied to the low-power transform coding architecture design. The resulting time-recursive multirate transform architectures inherit all advantages of the existing time-recursive transform architectures such as local communication, regularity, modularity, and linear hardware complexity, but the speed for updating the transform coefficients becomes M-times slower.
The last application is a programmable video co-processor system architecture that is capable of performing FIR/IIR filtering, subband filtering, discrete orthogonal transforms (DT) and adaptive filtering for the host processor in video applications. The system can be easily reconfigurated to perform multirate FIR/IIR/DT operations. Hence, we can either double the processing speed on-the-fly, based on the same processing elements, or apply this feature to the low- power implementation of this co- processor.
The methodology and the applications presented in this dissertation constitute a design framework for achieving low-power consumption at the algorithmic/architectural level for DSP applications.
Item Design of Structured Quantizers Based on Coset Codes(1995) Lee, Cheng-Chieh; Farvardin, N.; ISRFor memoryless sources, Entropy-Constrained Scalar Quantizers (ECSQs) can perform closely to the Gish-Pierce bound at high rates. There exist two fixed-rate variations of ECSQ -- Scalar- Vector Quantizer (SVQ) and Adaptive Entropy-Coded Quantizer (AECQ) -- that also perform closely to the Gish-Pierce bound. These quantization schemes have approximately cubic quantization cells while high-rate quantization theory suggests that quantization cells of the optimal quantizers should be approximately spherical. There are some coset codes whose Voronoi regions are very spherical. In this dissertation we present structured quantization schemes that combine these coset codes with the aforementioned quantizers (SVQ, ECSQ, and AECQ) so as to improve their performance beyond the Gish-Pierce bound.By combining trellis codes (that achieve a significant granular gain) with SVQ, ECSQ, and AECQ, we obtain Trellis-Based Scalar- Vector Quantizer (TB-SVQ), Entropy-Constrained Trellis- Coded Quantizer (ECTCQ), and Pathwise-Adaptive ECTCQ (PA-ECTCQ), respectively. With an 8-state underlying trellis code, these trellis-coded quantization schemes perform about 1.0 dB better than their naive counterparts. There are two approaches that can extend the quantizers (TB-SVQ, ECTCQ, and PA-ECTCQ) for quantizing sources with memory. The first is to combine the predictive coding operation of the Differential Pulse Code Modulation scheme with various quantizers, yielding Predictive TB-SVQ, Predictive ECTCQ, and Predictive PA-ECTCQ, respectively. There is a duality between quantizing sources with memory and transmitting data over channels with memory. Laroia, Tretter, and Farvardin have recently introduced a precoding idea that helps transmitting data efficiently over channels with memory. By exploiting this duality, the second approach combines the precoder with TB-SVQ and ECTCQ to arrive at Precoded TB-SVQ and Precoded ECTCQ, respectively. Simulation results indicate that the porformance of these quantizers are also close to the rate- distortion limit.
The PA-ECTCQ performance has been shown to be robust, in the presence of source scale and, to a lesser extent, shape mismatch conditions. We also considered adjusting the underlying entropy encoder based on the quantized output (which provide some approximate information on the source statistics). The performance of the resulting Shape-Adjusting PA-ECTCQ has been shown to be robust to a rather wide range of source shape mismatch conditions. are also close to the rate-distortion limit.