82,129 research outputs found
Recommended from our members
Comparison of Current Gravity Estimation and Determination Models
This paper will discuss the history of gravity estimation and determination models while analyzing methods that are in development. Some fundamental methods for calculating the gravity field include spherical harmonics solutions, local weighted interpolation, and global point mascon modeling (PMC). Recently, high accuracy measurements have become more accessible, and the requirements for high order geopotential modeling have become more stringent. Interest in irregular bodies, accurate models of the hydrological system, and on-board processing has demanded a comprehensive model that can quickly and accurately compute the geopotential with low memory costs. This trade study of current geopotential modeling techniques will reveal that each modeling technique has a unique use case. It is notable that the spherical harmonics model is relatively accurate but poses a cumbersome inversion problem. PMC and interpolation models, on the other hand, are computationally efficient, but require more research to become robust models with high levels of accuracy. Considerations of the trade study will suggest further research for the point mascon model. The PMC model should be improved through mascon refinement, direct solutions that stem from geodetic measurements, and further validation of the gravity gradient. Finally, the potential for each model to be implemented with parallel computation will be shown to lead to large improvements in computing time while reducing the memory cost for each technique.Aerospace Engineering and Engineering Mechanic
Energy rating of a water pumping station using multivariate analysis
Among water management policies, the preservation and the saving of energy demand in water supply and treatment systems play key roles. When focusing on energy, the customary metric to determine the performance of water supply systems is linked to the definition of component-based energy indicators. This approach is unfit to account for interactions occurring among system elements or between the system and its environment. On the other hand, the development of information technology has led to the availability of increasing large amount of data, typically gathered from distributed sensor networks in so-called smart grids. In this context, data intensive methodologies address the possibility of using complex network modeling approaches, and advocate the issues related to the interpretation and analysis of large amount of data produced by smart sensor networks.
In this perspective, the present work aims to use data intensive techniques in the energy analysis of a water management network.
The purpose is to provide new metrics for the energy rating of the system and to be able to provide insights into the dynamics of its operations. The study applies neural network as a tool to predict energy demand, when using flowrate and vibration data as predictor variables
Mathematical modeling of ultra wideband in vivo radio channel
This paper proposes a novel mathematical model for an in vivo radio channel at ultra-wideband frequencies (3.1–10.6 GHz), which can be used as a reference model for in vivo channel response without performing intensive experiments or simulations. The statistics of error prediction between experimental and proposed model is RMSE = 5.29, which show the high accuracy of the proposed model. Also, the proposed model was applied to the blind data, and the statistics of error prediction is RMSE = 7.76, which also shows a reasonable accuracy of the model. This model will save the time and cost on simulations and experiments, and will help in designing an accurate link budget calculation for a future enhanced system for ultra-wideband body-centric wireless systems
On the Resilience of RTL NN Accelerators: Fault Characterization and Mitigation
Machine Learning (ML) is making a strong resurgence in tune with the massive
generation of unstructured data which in turn requires massive computational
resources. Due to the inherently compute- and power-intensive structure of
Neural Networks (NNs), hardware accelerators emerge as a promising solution.
However, with technology node scaling below 10nm, hardware accelerators become
more susceptible to faults, which in turn can impact the NN accuracy. In this
paper, we study the resilience aspects of Register-Transfer Level (RTL) model
of NN accelerators, in particular, fault characterization and mitigation. By
following a High-Level Synthesis (HLS) approach, first, we characterize the
vulnerability of various components of RTL NN. We observed that the severity of
faults depends on both i) application-level specifications, i.e., NN data
(inputs, weights, or intermediate), NN layers, and NN activation functions, and
ii) architectural-level specifications, i.e., data representation model and the
parallelism degree of the underlying accelerator. Second, motivated by
characterization results, we present a low-overhead fault mitigation technique
that can efficiently correct bit flips, by 47.3% better than state-of-the-art
methods.Comment: 8 pages, 6 figure
Distributed computing methodology for training neural networks in an image-guided diagnostic application
Distributed computing is a process through which a set of computers connected by a network is used collectively to solve a single problem. In this paper, we propose a distributed computing methodology for training neural networks for the detection of lesions in colonoscopy. Our approach is based on partitioning the training set across multiple processors using a parallel virtual machine. In this way, interconnected computers of varied architectures can be used for the distributed evaluation of the error function and gradient values, and, thus, training neural networks utilizing various learning methods. The proposed methodology has large granularity and low synchronization, and has been implemented and tested. Our results indicate that the parallel virtual machine implementation of the training algorithms developed leads to considerable speedup, especially when large network architectures and training sets are used
W-NINE: a two-stage emulation platform for mobile and wireless systems
More and more applications and protocols are now running on wireless networks. Testing the implementation of such applications and protocols is a real challenge as the position of the mobile terminals and environmental effects strongly affect the overall performance. Network emulation is often perceived as a good trade-off between experiments on operational wireless networks and discrete-event simulations on Opnet or ns-2. However, ensuring repeatability and realism in network emulation while taking into account mobility in a wireless environment is very difficult. This paper proposes a network emulation platform, called W-NINE, based on off-line computations preceding online pattern-based traffic shaping. The underlying concepts of repeatability, dynamicity, accuracy and realism are defined in the emulation context. Two different simple case studies illustrate the validity of our approach with respect to these concepts
- …