15 research outputs found
The design of finite-state machines for quantization using simulated annealing
Ankara : Department of Electrical and Electronics Engineering and Institute of Engineering and Sciences, Bilkent Univ., 1993.Thesis (Master's) -- Bilkent University, 1993.Includes bibliographical references leaves 121-125In this thesis, the combinatorial optimization algorithm known as simulated annealing
(SA) is applied to the solution of the next-state map design problem of
data compression systems based on finite-state machine decoders. These data
compression systems which include finite-state vector ciuantization (FSVQ),
trellis waveform coding (TWC), predictive trellis waveform coding (PTWC),
and trellis coded quantization (TCQ) are studied in depth. Incorporating generalized
Lloyd algorithm for the optimization of output map to SA, a finite-state
machine decoder design algorithm for the joint optimization of output map
and next-state map is constructed. Simulation results on several discrete-time
sources for FSVQ, TWC and PTWC show that decoders with higher performance
are obtained by the SA-I-CLA algorithm, when compared to other
related work in the literature. In TCQ, simulation results are obtained for
sources with memory and new observations are made.Kuruoğlu, Ercan EnginM.S
Optimal soft-decoding combined trellis-coded quantization/modulation.
Chei Kwok-hung.Thesis (M.Phil.)--Chinese University of Hong Kong, 2000.Includes bibliographical references (leaves 66-73).Abstracts in English and Chinese.Chapter Chapter 1 --- Introduction --- p.1Chapter 1.1 --- Typical Digital Communication Systems --- p.2Chapter 1.1.1 --- Source coding --- p.3Chapter 1.1.2 --- Channel coding --- p.5Chapter 1.2 --- Joint Source-Channel Coding System --- p.5Chapter 1.3 --- Thesis Organization --- p.7Chapter Chapter 2 --- Trellis Coding --- p.9Chapter 2.1 --- Convolutional Codes --- p.9Chapter 2.2 --- Trellis-Coded Modulation --- p.12Chapter 2.2.1 --- Set Partitioning --- p.13Chapter 2.3 --- Trellis-Coded Quantization --- p.14Chapter 2.4 --- Joint TCQ/TCM System --- p.17Chapter 2.4.1 --- The Combined Receiver --- p.17Chapter 2.4.2 --- Viterbi Decoding --- p.19Chapter 2.4.3 --- Sequence MAP Decoding --- p.20Chapter 2.4.4 --- Sliding Window Decoding --- p.21Chapter 2.4.5 --- Block-Based Decoding --- p.23Chapter Chapter 3 --- Soft Decoding Joint TCQ/TCM over AWGN Channel --- p.25Chapter 3.1 --- System Model --- p.26Chapter 3.2 --- TCQ with Optimal Soft-Decoder --- p.27Chapter 3.3 --- Gaussian Memoryless Source --- p.30Chapter 3.3.1 --- Theorem Limit --- p.31Chapter 3.3.2 --- Performance on PAM Constellations --- p.32Chapter 3.3.3 --- Performance on PSK Constellations --- p.36Chapter 3.4 --- Uniform Memoryless Source --- p.38Chapter 3.4.1 --- Theorem Limit --- p.38Chapter 3.4.2 --- Performance on PAM Constellations --- p.39Chapter 3.4.3 --- Performance on PSK Constellations --- p.40Chapter Chapter 4 --- Soft Decoding Joint TCQ/TCM System over Rayleigh Fading Channel --- p.42Chapter 4.1 --- Wireless Channel --- p.43Chapter 4.2 --- Rayleigh Fading Channel --- p.44Chapter 4.3 --- Idea Interleaving --- p.45Chapter 4.4 --- Receiver Structure --- p.46Chapter 4.5 --- Numerical Results --- p.47Chapter 4.5.1 --- Performance on 4-PAM Constellations --- p.48Chapter 4.5.2 --- Performance on 8-PAM Constellations --- p.50Chapter 4.5.3 --- Performance on 16-PAM Constellations --- p.52Chapter Chapter 5 --- Joint TCVQ/TCM System --- p.54Chapter 5.1 --- Trellis-Coded Vector Quantization --- p.55Chapter 5.1.1 --- Set Partitioning in TCVQ --- p.56Chapter 5.2 --- Joint TCVQ/TCM --- p.59Chapter 5.2.1 --- Set Partitioning and Index Assignments --- p.60Chapter 5.2.2 --- Gaussian-Markov Sources --- p.61Chapter 5.3 --- Simulation Results and Discussion --- p.62Chapter Chapter 6 --- Conclusion and Future Work --- p.64Chapter 6.1 --- Conclusion --- p.64Chapter 6.2 --- Future Works --- p.65Bibliography --- p.66Appendix-Publications --- p.7
High-performance compression of visual information - A tutorial review - Part I : Still Pictures
Digital images have become an important source of information in the modern world of communication systems. In their raw form, digital images require a tremendous amount of memory. Many research efforts have been devoted to the problem of image compression in the last two decades. Two different compression categories must be distinguished: lossless and lossy. Lossless compression is achieved if no distortion is introduced in the coded image. Applications requiring this type of compression include medical imaging and satellite photography. For applications such as video telephony or multimedia applications, some loss of information is usually tolerated in exchange for a high compression ratio. In this two-part paper, the major building blocks of image coding schemes are overviewed. Part I covers still image coding, and Part II covers motion picture sequences. In this first part, still image coding schemes have been classified into predictive, block transform, and multiresolution approaches. Predictive methods are suited to lossless and low-compression applications. Transform-based coding schemes achieve higher compression ratios for lossy compression but suffer from blocking artifacts at high-compression ratios. Multiresolution approaches are suited for lossy as well for lossless compression. At lossy high-compression ratios, the typical artifact visible in the reconstructed images is the ringing effect. New applications in a multimedia environment drove the need for new functionalities of the image coding schemes. For that purpose, second-generation coding techniques segment the image into semantically meaningful parts. Therefore, parts of these methods have been adapted to work for arbitrarily shaped regions. In order to add another functionality, such as progressive transmission of the information, specific quantization algorithms must be defined. A final step in the compression scheme is achieved by the codeword assignment. Finally, coding results are presented which compare stateof- the-art techniques for lossy and lossless compression. The different artifacts of each technique are highlighted and discussed. Also, the possibility of progressive transmission is illustrated
High ratio wavelet video compression through real-time rate-distortion estimation.
Thesis (M.Sc.Eng.)-University of Natal, Durban, 2003.The success of the wavelet transform in the compression of still images has prompted an
expanding effort to exercise this transform in the compression of video. Most existing video
compression methods incorporate techniques from still image compression, such techniques
being abundant, well defined and successful. This dissertation commences with a thorough
review and comparison of wavelet still image compression techniques. Thereafter an
examination of wavelet video compression techniques is presented. Currently, the most
effective video compression system is the DCT based framework, thus a comparison between
these and the wavelet techniques is also given.
Based on this review, this dissertation then presents a new, low-complexity, wavelet video
compression scheme. Noting from a complexity study that the generation of temporally
decorrelated, residual frames represents a significant computational burden, this scheme uses
the simplest such technique; difference frames. In the case of local motion, these difference
frames exhibit strong spatial clustering of significant coefficients. A simple spatial syntax is
created by splitting the difference frame into tiles. Advantage of the spatial clustering may then
be taken by adaptive bit allocation between the tiles. This is the central idea of the method.
In order to minimize the total distortion of the frame, the scheme uses the new p-domain rate-distortion
estimation scheme with global numerical optimization to predict the optimal
distribution of bits between tiles. Thereafter each tile is independently wavelet transformed and
compressed using the SPIHT technique.
Throughout the design process computational efficiency was the design imperative, thus leading
to a real-time, software only, video compression scheme. The scheme is finally compared to
both the current video compression standards and the leading wavelet schemes from the
literature in terms of computational complexity visual quality. It is found that for local motion
scenes the proposed algorithm executes approximately an order of magnitude faster than these
methods, and presents output of similar quality. This algorithm is found to be suitable for
implementation in mobile and embedded devices due to its moderate memory and
computational requirements
Distributed signal processing using nested lattice codes
Multi-Terminal Source Coding (MTSC) addresses the problem of compressing correlated sources
without communication links among them. In this thesis, the constructive approach of this problem
is considered in an algebraic framework and a system design is provided that can be applicable
in a variety of settings. Wyner-Ziv problem is first investigated: coding of an independent and
identically distributed (i.i.d.) Gaussian source with side information available only at the decoder
in the form of a noisy version of the source to be encoded. Theoretical models are first established
and derived for calculating distortion-rate functions. Then a few novel practical code implementations are proposed by using the strategy of multi-dimensional nested lattice/trellis coding. By
investigating various lattices in the dimensions considered, analysis is given on how lattice properties affect performance. Also proposed are methods on choosing good sublattices in multiple
dimensions. By introducing scaling factors, the relationship between distortion and scaling factor
is examined for various rates. The best high-dimensional lattice using our scale-rotate method can
achieve a performance less than 1 dB at low rates from the Wyner-Ziv limit; and random nested
ensembles can achieve a 1.87 dB gap with the limit. Moreover, the code design is extended to
incorporate with distributed compressive sensing (DCS). Theoretical framework is proposed and
practical design using nested lattice/trellis is presented for various scenarios. By using nested
trellis, the simulation shows a 3.42 dB gap from our derived bound for the DCS plus Wyner-Ziv
framework
Hidden Markov model based Finnish text-to-speech system utilizing glottal inverse filtering
Tässä työssä esitetään uusi Markovin piilomalleihin (hidden Markov model, HMM) perustuva äänilähteen käänteissuodatusta hyödyntävä suomenkielinen puhesynteesijärjestelmä. Uuden puhesynteesimenetelmän päätavoite on tuottaa luonnolliselta kuulostavaa synteettistä puhetta, jonka ominaisuuksia voidaan muuttaa eri puhujien, puhetyylien tai jopa äänen emootiosisällön mukaan. Näiden tavoitteiden mahdollistamiseksi uudessa puhesynteesimenetelmässä mallinnetaan ihmisen äänentuottojärjestelmää äänilähteen käänteissuodatuksen ja HMM-mallinnuksen avulla.
Uusi puhesynteesijärjestelmä hyödyntää äänilähteen käänteissuodatusmenetelmää, joka mahdollistaa äänilähteen ominaisuuksien parametrisoinnin erillään muista puheen parametreista, ja siten näiden parametrien mallintamisen erikseen HMM-järjestelmässä. Synteesivaiheessa luonnollisesta puheesta laskettuja glottispulsseja käytetään äänilähteen luomiseen, ja äänilähteen ominaisuuksia muokataan edelleen tilastollisen HMM-järjestelmän tuottaman parametrisen kuvauksen avulla, mikä imitoi oikeassa puheessa esiintyvää luonnollista äänilähteen ominaisuuksien vaihtelua.
Subjektiivisten kuuntelukokeiden tulokset osoittavat, että uuden puhesynteesimenetelmän laatu on huomattavasti parempi verrattuna perinteiseen HMM-pohjaiseen puhesynteesijärjestelmään. Lisäksi tulokset osoittavat, että uusi puhesynteesimenetelmä pystyy tuottamaan luonnolliselta kuulostavaa puhetta eri puhujien ominaisuuksilla.In this work, a new hidden Markov model (HMM) based text-to-speech (TTS) system utilizing glottal inverse filtering is described. The primary goal of the new TTS system is to enable producing natural sounding synthetic speech in different speaking styles with different speaker characteristics and emotions. In order to achieve these goals, the function of the real human voice production mechanism is modeled with the help of glottal inverse filtering embedded in a statistical framework of HMM.
The new TTS system uses a glottal inverse filtering based parametrization method that enables the extraction of voice source characteristics separate from other speech parameters, and thus the individual modeling of these characteristics in the HMM system. In the synthesis stage, natural glottal flow pulses are used for creating the voice source, and the voice source characteristics are further modified according to the adaptive all-pole model generated by the HMM system in order to imitate the natural variation in the real voice source.
Subjective listening tests show that the quality of the new TTS system is considerably better compared to a traditional HMM-based speech synthesizer. Moreover, the new system is clearly able to produce natural sounding synthetic speech with specific speaker characteristics
Ultra Low Power Digital Circuit Design for Wireless Sensor Network Applications
Ny forskning innenfor feltet trådløse sensornettverk åpner for nye og innovative produkter og løsninger. Biomedisinske anvendelser er blant områdene med størst potensial og det investeres i dag betydelige beløp for å bruke denne teknologien for å gjøre medisinsk diagnostikk mer effektiv samtidig som man åpner for fjerndiagnostikk basert på trådløse sensornoder integrert i et ”helsenett”. Målet er å forbedre tjenestekvalitet og redusere kostnader samtidig som brukerne skal oppleve forbedret livskvalitet som følge av økt trygghet og mulighet for å tilbringe mest mulig tid i eget hjem og unngå unødvendige sykehusbesøk og innleggelser. For å gjøre dette til en realitet er man avhengige av sensorelektronikk som bruker minst mulig energi slik at man oppnår tilstrekkelig batterilevetid selv med veldig små batterier. I sin avhandling ” Ultra Low power Digital Circuit Design for Wireless Sensor Network Applications” har PhD-kandidat Farshad Moradi fokusert på nye løsninger innenfor konstruksjon av energigjerrig digital kretselektronikk. Avhandlingen presenterer nye løsninger både innenfor aritmetiske og kombinatoriske kretser, samtidig som den studerer nye statiske minneelementer (SRAM) og alternative minnearkitekturer. Den ser også på utfordringene som oppstår når silisiumteknologien nedskaleres i takt med mikroprosessorutviklingen og foreslår løsninger som bidrar til å gjøre kretsløsninger mer robuste og skalerbare i forhold til denne utviklingen. De viktigste konklusjonene av arbeidet er at man ved å introdusere nye konstruksjonsteknikker både er i stand til å redusere energiforbruket samtidig som robusthet og teknologiskalerbarhet øker. Forskningen har vært utført i samarbeid med Purdue University og vært finansiert av Norges Forskningsråd gjennom FRINATprosjektet ”Micropower Sensor Interface in Nanometer CMOS Technology”