72 research outputs found
Exploiting parallelism within multidimensional multirate digital signal processing systems
The intense requirements for high processing rates of multidimensional Digital Signal Processing systems in practical applications justify the Application Specific Integrated Circuits designs and parallel processing implementations. In this dissertation, we propose novel theories, methodologies and architectures in designing high-performance VLSI implementations for general multidimensional multirate Digital Signal Processing systems by exploiting the parallelism within those applications. To systematically exploit the parallelism within the multidimensional multirate DSP algorithms, we develop novel transformations including (1) nonlinear I/O data space transforms, (2) intercalation transforms, and (3) multidimensional multirate unfolding transforms. These transformations are applied to the algorithms leading to systematic methodologies in high-performance architectural designs. With the novel design methodologies, we develop several architectures with parallel and distributed processing features for implementing multidimensional multirate applications. Experimental results have shown that those architectures are much more efficient in terms of execution time and/or hardware cost compared with existing hardware implementations
Design and Performance of Scalable High-Performance Programmable Routers - Doctoral Dissertation, August 2002
The flexibility to adapt to new services and protocols without changes in the underlying hardware is and will increasingly be a key requirement for advanced networks. Introducing a processing component into the data path of routers and implementing packet processing in software provides this ability. In such a programmable router, a powerful processing infrastructure is necessary to achieve to level of performance that is comparable to custom silicon-based routers and to demonstrate the feasibility of this approach. This work aims at the general design of such programmable routers and, specifically, at the design and performance analysis of the processing subsystem. The necessity of programmable routers is motivated, and a router design is proposed. Based on the design, a general performance model is developed and quantitatively evaluated using a new network processor benchmark. Operational challenges, like scheduling of packets to processing engines, are addressed, and novel algorithms are presented. The results of this work give qualitative and quantitative insights into this new domain that combines issues from networking, computer architecture, and system design
High ratio wavelet video compression through real-time rate-distortion estimation.
Thesis (M.Sc.Eng.)-University of Natal, Durban, 2003.The success of the wavelet transform in the compression of still images has prompted an
expanding effort to exercise this transform in the compression of video. Most existing video
compression methods incorporate techniques from still image compression, such techniques
being abundant, well defined and successful. This dissertation commences with a thorough
review and comparison of wavelet still image compression techniques. Thereafter an
examination of wavelet video compression techniques is presented. Currently, the most
effective video compression system is the DCT based framework, thus a comparison between
these and the wavelet techniques is also given.
Based on this review, this dissertation then presents a new, low-complexity, wavelet video
compression scheme. Noting from a complexity study that the generation of temporally
decorrelated, residual frames represents a significant computational burden, this scheme uses
the simplest such technique; difference frames. In the case of local motion, these difference
frames exhibit strong spatial clustering of significant coefficients. A simple spatial syntax is
created by splitting the difference frame into tiles. Advantage of the spatial clustering may then
be taken by adaptive bit allocation between the tiles. This is the central idea of the method.
In order to minimize the total distortion of the frame, the scheme uses the new p-domain rate-distortion
estimation scheme with global numerical optimization to predict the optimal
distribution of bits between tiles. Thereafter each tile is independently wavelet transformed and
compressed using the SPIHT technique.
Throughout the design process computational efficiency was the design imperative, thus leading
to a real-time, software only, video compression scheme. The scheme is finally compared to
both the current video compression standards and the leading wavelet schemes from the
literature in terms of computational complexity visual quality. It is found that for local motion
scenes the proposed algorithm executes approximately an order of magnitude faster than these
methods, and presents output of similar quality. This algorithm is found to be suitable for
implementation in mobile and embedded devices due to its moderate memory and
computational requirements
Exploiting parallelism within multidimensional multirate digital signal processing systems
The intense requirements for high processing rates of multidimensional Digital Signal Processing systems in practical applications justify the Application Specific Integrated Circuits designs and parallel processing implementations. In this dissertation, we propose novel theories, methodologies and architectures in designing high-performance VLSI implementations for general multidimensional multirate Digital Signal Processing systems by exploiting the parallelism within those applications. To systematically exploit the parallelism within the multidimensional multirate DSP algorithms, we develop novel transformations including (1) nonlinear I/O data space transforms, (2) intercalation transforms, and (3) multidimensional multirate unfolding transforms. These transformations are applied to the algorithms leading to systematic methodologies in high-performance architectural designs. With the novel design methodologies, we develop several architectures with parallel and distributed processing features for implementing multidimensional multirate applications. Experimental results have shown that those architectures are much more efficient in terms of execution time and/or hardware cost compared with existing hardware implementations
Downstream Bandwidth Management for Emerging DOCSIS-based Networks
In this dissertation, we consider the downstream bandwidth management in the context of emerging DOCSIS-based cable networks. The latest DOCSIS 3.1 standard for cable access networks represents a significant change to cable networks. For downstream, the current 6 MHz channel size is replaced by a much larger 192 MHz channel which potentially can provide data rates up to 10 Gbps. Further, the current standard requires equipment to support a relatively new form of active queue management (AQM) referred to as delay-based AQM. Given that more than 50 million households (and climbing) use cable for Internet access, a clear understanding of the impacts of bandwidth management strategies used in these emerging networks is crucial. Further, given the scope of the change provided by emerging cable systems, now is the time to develop and introduce innovative new methods for managing bandwidth. With this motivation, we address research questions pertaining to next generation of cable access networks. The cable industry has had to deal with the problem of a small number of subscribers who utilize the majority of network resources. This problem will grow as access rates increase to gigabits per second. Fundamentally this is a problem on how to manage data flows in a fair manner and provide protection. A well known performance issue in the Internet, referred to as bufferbloat, has received significant attention recently. High throughput network flows need sufficiently large buffer to keep the pipe full and absorb occasional burstiness. Standard practice however has led to equipment offering very large unmanaged buffers that can result in sustained queue levels increasing packet latency. One reason why these problems continue to plague cable access networks is the desire for low complexity and easily explainable (to access network subscribers and to the Federal Communications Commission) bandwidth management. This research begins by evaluating modern delay-based AQM algorithms in downstream DOCSIS 3.0 environments with a focus on fairness and application performance capabilities of single queue AQMs. We are especially interested in delay-based AQM schemes that have been proposed to combat the bufferbloat problem. Our evaluation involves a variety of scenarios that include tiered services and application workloads. Based on our results, we show that in scenarios involving realistic workloads, modern delay-based AQMs can effectively mitigate bufferbloat. However they do not address the other problem related to managing the fairness. To address the combined problem of fairness and bufferbloat, we propose a novel approach to bandwidth management that provides a compromise among the conflicting requirements. We introduce a flow quantization method referred to as adaptive bandwidth binning where flows that are observed to consume similar levels of bandwidth are grouped together with the system managed through a hierarchical scheduler designed to approximate weighted fairness while addressing bufferbloat. Based on a simulation study that considers many system experimental parameters including workloads and network configurations, we provide evidence of the efficacy of the idea. Our results suggest that the scheme is able to provide long term fairness and low delay with a performance close to that of a reference approach based on fair queueing. A further contribution is our idea for replacing `tiered\u27 levels of service based on service rates with tiering based on weights. The application of our bandwidth binning scheme offers a timely and innovative alternative to broadband service that leverages the potential offered by emerging DOCSIS-based cable systems
Recommended from our members
Case-based reasoning for product style construction and fuzzy analytic hierarchy process evaluation modeling using consumers linguistic variables
Key form features are relative to the style of a product and the expression style features depict the product description and are a measurement of attribute knowledge. The uncertainty definition leads to an improved and effective product style retrieval when combined with fuzzy sets. Firstly, a style knowledge and features database are constructed using fuzzy case based reasoning technology (FCBR). A similarity measurement method based on case-based reasoning and fuzzy model of the fuzzy proximity method may be defined by the Fuzzy Nearest-Neighbor (FNN) algorithm obtaining the style knowledge extraction. Secondly, the Linguistic Variables (LV) are used to assess the product characteristics to establish the product style evaluation database for simplifying the style presentation and decreasing the computational complexity. Thirdly, the model of product style feature set, extracted by FAHP and the final style related form features set, are acquired using LV. This research involves a case study for extracting the key form features of the style of high heel shoes. The proposed algorithms are generated by calculating the weights of each component of high heel shoes using FAHP with LV. The case study and results established that the proposed method is feasible and effective for extracting the style of the product
Recommended from our members
Diabetic plantar pressure analysis using image fusion
Plantar pressure images analysis is the key issue of designing comfortable shoe products through last customizing system, which has attracted the researchers’ curiosity toward image fusion as an application of medical and industrial imaging. In the current work, image fusion has been applied using wavelet transform and compared with Laplace Pyramid. Using image fusion rules of Mean-Max, we presented a plantar pressure image fusion method employing haar wavelet transform. It was compared in different composition layers with the Laplace pyramid transform. The experimental studies deployed the haar, db2, sym4, coif2, and bior5.5 wavelet basis functions for image fusion under decomposition layers of 3, 4, and 5. Evaluation metrics were measured in the case of the different layer number of wavelet decomposition to determine the best decomposition level and to evaluate the fused image quality using with different wavelet functions. The best wavelet basis function and decomposition layers were selected through the analysis and the evaluation measurements. This study established that haar wavelet transform with five decomposition levels on plantar pressure image achieved superior performance of 89.2817% mean, 89.4913% standard deviation, 5.4196 average gradient, 14.3364 spatial frequency, 5.9323 information entropy and 0.2206 cross entropy
On the design of fast and efficient wavelet image coders with reduced memory usage
Image compression is of great importance in multimedia systems and
applications because it drastically reduces bandwidth requirements for
transmission and memory requirements for storage. Although earlier
standards for image compression were based on the Discrete Cosine
Transform (DCT), a recently developed mathematical technique, called
Discrete Wavelet Transform (DWT), has been found to be more efficient
for image coding.
Despite improvements in compression efficiency, wavelet image coders
significantly increase memory usage and complexity when compared with
DCT-based coders. A major reason for the high memory requirements is
that the usual algorithm to compute the wavelet transform requires the
entire image to be in memory. Although some proposals reduce the memory
usage, they present problems that hinder their implementation. In
addition, some wavelet image coders, like SPIHT (which has become a
benchmark for wavelet coding), always need to hold the entire image in
memory.
Regarding the complexity of the coders, SPIHT can be considered quite
complex because it performs bit-plane coding with multiple image scans.
The wavelet-based JPEG 2000 standard is still more complex because it
improves coding efficiency through time-consuming methods, such as an
iterative optimization algorithm based on the Lagrange multiplier
method, and high-order context modeling.
In this thesis, we aim to reduce memory usage and complexity in
wavelet-based image coding, while preserving compression efficiency. To
this end, a run-length encoder and a tree-based wavelet encoder are
proposed. In addition, a new algorithm to efficiently compute the
wavelet transform is presented. This algorithm achieves low memory
consumption using line-by-line processing, and it employs recursion to
automatically place the order in which the wavelet transform is
computed, solving some synchronization problems that have not been
tackled by previous proposals. The proposed encodeOliver Gil, JS. (2006). On the design of fast and efficient wavelet image coders with reduced memory usage [Tesis doctoral no publicada]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/1826Palanci
Recommended from our members
An efficient local binary pattern based plantar pressure optical sensor image classification using convolutional neural networks
The objective of this study was to design and produce highly comfortable shoe products guided by a plantar pressure imaging data-set. Previous studies have focused on the geometric measurement on the size of the plantar, while in this research a plantar pressure optical imaging data-set based classification technology has been developed. In this paper, an improved local binary pattern (LBP) algorithm is used to extract texture-based features and recognize patterns from the data-set. A calculating model of plantar pressure imaging feature area is established subsequently. The data-set is classified by a neural network to guide the generation of various shoe-last surfaces. Firstly, the local binary mode is improved to adapt to the pressure imaging data-set, and the texture-based feature calculation is fully used to accurately generate the feature point set; hereafter, the plantar pressure imaging feature point set is then used to guide the design of last free surface forming. In the presented experiments of plantar imaging, multi-dimensional texture-based features and improved LBP features have been found by a convolution neural network (CNN), and compared with a 21-input-3-output two-layer perceptual neural network. Three feet types are investigated in the experiment, being flatfoot (F) referring to the lack of a normal arch, or arch collapse, Talipes Equinovarus (TE), being the front part of the foot is adduction, calcaneus varus, plantar flexion, or Achilles tendon contracture and Normal (N). This research has achieved an 82% accuracy rate with 10 hidden-layers CNN of rotation invariance LBP (RI-LBP) algorithm using 21 texture-based features by comparing other deep learning methods presented in the literature
Adaptive Resource Management for Uncertain Execution Platforms
Embedded systems are becoming increasingly complex. At the same time, the components that make up the system grow more uncertain in their properties. For example, current developments in CPU design focuses on optimizing for average performance rather than better worst case performance. This, combined with presence of 3rd party software components with unknown properties, makes resource management using prior knowledge less and less feasible. This thesis presents results on how to model software components so that resource allocation decisions can be made on-line. Both the single and multiple resource case is considered as well as extending the models to include resource constraints based on hardware dynam- ics. Techniques for estimating component parameters on-line are presented. Also presented is an algorithm for computing an optimal allocation based on a set of convex utility functions. The algorithm is designed to be computationally efficient and to use simple mathematical expres- sions that are suitable for fixed point arithmetics. An implementation of the algorithm and results from experiments is presented, showing that an adaptive strategy using both estimation and optimization can outperform a static approach in cases where uncertainty is high
- …