18 research outputs found
Regression and Singular Value Decomposition in Dynamic Graphs
Most of real-world graphs are {\em dynamic}, i.e., they change over time.
However, while problems such as regression and Singular Value Decomposition
(SVD) have been studied for {\em static} graphs, they have not been
investigated for {\em dynamic} graphs, yet. In this paper, we introduce,
motivate and study regression and SVD over dynamic graphs. First, we present
the notion of {\em update-efficient matrix embedding} that defines the
conditions sufficient for a matrix embedding to be used for the dynamic graph
regression problem (under norm). We prove that given an
update-efficient matrix embedding (e.g., adjacency matrix), after an update
operation in the graph, the optimal solution of the graph regression problem
for the revised graph can be computed in time. We also study dynamic
graph regression under least absolute deviation. Then, we characterize a class
of matrix embeddings that can be used to efficiently update SVD of a dynamic
graph. For adjacency matrix and Laplacian matrix, we study those graph update
operations for which SVD (and low rank approximation) can be updated
efficiently
Analysis of general weights in weighted ℓ1−2 minimization through applications
The weighted ℓ1−2 minimization has recently attracted some attention due to its capability to deal with highly coherent matrices. Notwithstanding the availability of its stable recovery guarantees, there appear to be some issues not addressed in the literature, which are (i). convergence of the solver for the weighted ℓ1−2 minimization analytically, and (ii). detailed analysis of relevance of general weights to applications. While establishing the convergence of the solver of the weighted ℓ1−2 minimization, we demonstrate the significance of general weights, w∈(0,1), empirically through some applications, including the reconstruction of magnetic resonance images. In particular, we show that the general weights attain significance when we do not have fully accurate or fully corrupt information about the support of the signal to be reconstructed from its linear measurements. We conclude the work by discussing a numerical scheme that chooses the partial support and the weights iteratively
Recommended from our members
Signal acquisition challenges in mobile systems
In recent decades, the advent of mobile computing has changed human lives by providing information that was not available in the past. The mobile computing platform opens a new door to the connected world in which various forms of hand-held and wearable systems are ubiquitous. A single mobile device plays multiple roles and shapes human lives towards a better future. In these systems, sensor-based data acquisition plays an essential role in generating and providing useful information.
The increased number of sensors is embedded in a single device in order to process various signal modalities. In practice, more than 30 data converters are required in designing a mobile system in which the data-converting blocks become among the most power-hungry components in battery-operated systems. Due to the increased variety of sensors, mobile systems are meant to face several obstacles. For example, the increased number of sensors increase system power consumption during the system operation. The increased power consumption directly affects operation time because mobile systems are powered by a limited energy source. Moreover, an increased amount of information also gives rise to bandwidth problems in communication due to the increased volume of data transmission. Also, this system design requires a larger area in a silicon die so that multiple signal paths can be placed without cross-channel interference. Therefore, the system design has presented a challenge in terms of trying to resolve the design constraints such as power consumption, bandwidth usage, storage space, and design complexity issues.
To overcome these obstacles, in this dissertation, efficient data acquisition and processing methods are investigated. Specifically, this thesis considers the problems of energy-efficient sampling and binary event detection.
This dissertation begins by presenting a new signal sampling scheme that enables higher precision signal conversion in compressed-sensing-based signal acquisition. The proposed scheme is based on the popular successive approximation register and employs a modified compressive sensing technique to increase the resolution of successive-approximation-register (SAR) analog-to-digital converter (ADC) architecture. Circuit-level architecture is discussed to implement the proposed scheme using the SAR ADC architecture. A non-uniform quantization scheme is proposed and it improves data quality after data acquisition. The proposed scheme is expected to be used for medium- or high- frequency data conversion.
Secondly, the possibility of using fewer ADCs than channels is studied by leveraging sparse-signal representation and blind-source-separation (BSS) techniques.
In particular, this dissertation examines the problem of using a single ADC or quantizer system for digitizing multi-channel inputs. Mixing and de-mixing strategies are extensively studied for sampling frequency-sparse signals and the proposed multi-channel architecture can be easily implemented using today's analog/mixed-signal circuits.
The third part of this dissertation investigates a binary hypothesis testing problem. In mobile devices such as smartphones and tablet PCs, a major portion of energy is consumed in user interfaces (LCD display and touch input processing). For accurate detection and better user interface, energy-efficient sensing and detection schemes are necessary to manage multiple sensor inputs. A highly efficient detection scheme is presented that can detect binary events reliably with a fraction of the energy consumption required in the conventional energy detection.Electrical and Computer Engineerin
Cooperative control of a network of multi-vehicle unmanned systems
Development of unmanned systems network is currently among one of the most important areas of activity and research with implications in variety of disciplines, such as communications, controls, and multi-vehicle systems. The main motivation for this interest can be traced back to practical applications wherein direct human involvement may not be possible due to environmental hazards or the extraordinary complexity of the tasks. This thesis seeks to develop, design, and analyze techniques and solutions that would ensure and guarantee the fundamental stringent requirements that are envisaged for these dynamical networks. In this thesis, the problem of team cooperation is solved by using synthesis-based approaches. The consensus problem is defined and solved for a team of agents having a general linear dynamical model. Stability of the team is guaranteed by using modified consensus algorithms that are achieved by minimizing a set of individual cost functions. An alternative approach for obtaining an optimal consensus algorithm is obtained by invoking a state decomposition methodology and by transforming the consensus seeking problem into a stabilization problem. In another methodology, the game theory approach is used to formulate the consensus seeking problem in a "more" cooperative framework. For this purpose, a team cost function is defined and a min-max problem is solved to obtain a cooperative optimal solution. It is shown that the results obtained yield lower cost values when compared to those obtained by using the optimal control technique. In game theory and optimal control approaches that are developed based on state decomposition, linear matrix inequalities are used to impose simultaneously the decentralized nature of the problem as well as the consensus constraints on the designed controllers. Moreover, performance and stability properties of the designed cooperative team is analyzed in presence of actuator anomalies corresponding to three types of faults. Steady state behavior of the team members are analyzed under faulty scenarios. Adaptability of the team members to the above unanticipated circumstances is demonstrated and verified. Finally, the assumption of having a fixed and undirected network topology is relaxed to address and solve a more realistic and practical situation. It is shown that the stability and consensus achievement of the network with a switching structure and leader assignment can still be achieved. Moreover, by introducing additional criteria, the desirable performance specifications of the team can still be ensured and guaranteed
Robust computational intelligence techniques for visual information processing
The third part is exclusively dedicated to the super-resolution of Magnetic Resonance Images. In one of these works, an algorithm based on the random shifting technique is developed. Besides, we studied noise removal and resolution enhancement simultaneously. To end, the cost function of deep networks has been modified by different combinations of norms in order to improve their training.
Finally, the general conclusions of the research are presented and discussed, as well as the possible future research lines that are able to make use of the results obtained in this Ph.D. thesis.This Ph.D. thesis is about image processing by computational intelligence techniques. Firstly, a general overview of this book is carried out, where the motivation, the hypothesis, the objectives, and the methodology employed are described. The use and analysis of different mathematical norms will be our goal. After that, state of the art focused on the applications of the image processing proposals is presented. In addition, the fundamentals of the image modalities, with particular attention to magnetic resonance, and the learning techniques used in this research, mainly based on neural networks, are summarized. To end up, the mathematical framework on which this work is based on, ₚ-norms, is defined.
Three different parts associated with image processing techniques follow. The first non-introductory part of this book collects the developments which are about image segmentation. Two of them are applications for video surveillance tasks and try to model the background of a scenario using a specific camera. The other work is centered on the medical field, where the goal of segmenting diabetic wounds of a very heterogeneous dataset is addressed.
The second part is focused on the optimization and implementation of new models for curve and surface fitting in two and three dimensions, respectively. The first work presents a parabola fitting algorithm based on the measurement of the distances of the interior and exterior points to the focus and the directrix. The second work changes to an ellipse shape, and it ensembles the information of multiple fitting methods. Last, the ellipsoid problem is addressed in a similar way to the parabola
Recommended from our members
Software reliability prediction
Two methods are proposed to find the maximum likelihood parameter estimates of a number of software reliability models. On the basis of the results from analysing 7 sets of real data, these methods are found to be both efficient and reliable.
The simple approach of adapting software reliability predictions by Keiller and Littlewood (1984) can produce improved predictions, but at the same time, introduces a lot of internal noise into the adapted predictions. This is due to the fact that the adaptor is a joined-up function.
An alternative adaptive procedure, which involves the parametric spline adaptor, can produce at least as good adapted predictions without the predictions being contaminated by internal noise as in the simple approach.
Miller and Sofer (1986a) proposed a method for estimating the failure rate of a program non-parametrically. Here, these non-parametric rates are used to produce reliability predictions and their quality is analysed and compared with the parametric predictions
Recommended from our members
Software reliability prediction
Two methods are proposed to find the maximum likelihood parameter estimates of a number of software reliability models. On the basis of the results from analysing 7 sets of real data, these methods are found to be both efficient and reliable.
The simple approach of adapting software reliability predictions by Keiller and Littlewood (1984) can produce improved predictions, but at the same time, introduces a lot of internal noise into the adapted predictions. This is due to the fact that the adaptor is a joined-up function.
An alternative adaptive procedure, which involves the parametric spline adaptor, can produce at least as good adapted predictions without the predictions being contaminated by internal noise as in the simple approach.
Miller and Sofer (1986a) proposed a method for estimating the failure rate of a program non-parametrically. Here, these non-parametric rates are used to produce reliability predictions and their quality is analysed and compared with the parametric predictions
A summary of the users perspective of LANDSAT-D and reference document of LANDSAT users
There are no author-identified significant results in this report