153 research outputs found
A Case for Leveraging 802.11p for Direct Phone-to-Phone Communications
WiFi cannot effectively handle the demands of device-to-device communication between phones, due to insufficient range and poor reliability. We make the case for using IEEE 802.11p DSRC instead, which has been adopted for vehicle-to-vehicle communications, providing lower latency and longer range. We demonstrate a prototype motivated by a novel fabrication process that deposits both III-V and CMOS devices on the same die. In our system prototype, the designed RF front-end is interfaced with a baseband processor on an FPGA, connected to Android phones. It consumes 0.02uJ/bit across 100m assuming free space. Application-level power control dramatically reduces power consumption by 47-56%.Singapore-MIT Alliance for Research and TechnologyAmerican Society for Engineering Education. National Defense Science and Engineering Graduate Fellowshi
Assessment of the Regions Differentiation in the Context of the Economic Security Ensuring
In recent years, the problem of economic security has become significantly more acute. One of the threats to economic security is excessive differentiation of the country’s regions by their level of socio-economic development. This differentiation makes the country’s economic space heterogeneous. The corresponding threat to economic security requires neutralization. The key issue here is the choice of a method for assessing the level of differentiation. The article analyzes the methodological techniques used to assess regional differentiation. It is shown that they not only differ, but can also give contradictory results when applied. The need for unification and integration of approaches to assessing regional differentiation in the context of ensuring economic security is revealed. These recommendations can be considered when implementing economic policy at both the federal and regional levels
Efficient Core-selecting Incentive Mechanism for Data Sharing in Federated Learning
Federated learning is a distributed machine learning system that uses
participants' data to train an improved global model. In federated learning,
participants cooperatively train a global model, and they will receive the
global model and payments. Rational participants try to maximize their
individual utility, and they will not input their high-quality data truthfully
unless they are provided with satisfactory payments based on their data
quality. Furthermore, federated learning benefits from the cooperative
contributions of participants. Accordingly, how to establish an incentive
mechanism that both incentivizes inputting data truthfully and promotes stable
cooperation has become an important issue to consider. In this paper, we
introduce a data sharing game model for federated learning and employ
game-theoretic approaches to design a core-selecting incentive mechanism by
utilizing a popular concept in cooperative games, the core. In federated
learning, the core can be empty, resulting in the core-selecting mechanism
becoming infeasible. To address this, our core-selecting mechanism employs a
relaxation method and simultaneously minimizes the benefits of inputting false
data for all participants. However, this mechanism is computationally expensive
because it requires aggregating exponential models for all possible coalitions,
which is infeasible in federated learning. To address this, we propose an
efficient core-selecting mechanism based on sampling approximation that only
aggregates models on sampled coalitions to approximate the exact result.
Extensive experiments verify that the efficient core-selecting mechanism can
incentivize inputting high-quality data and stable cooperation, while it
reduces computational overhead compared to the core-selecting mechanism
Enchanting Program Specification Synthesis by Large Language Models using Static Analysis and Program Verification
Formal verification provides a rigorous and systematic approach to ensure the
correctness and reliability of software systems. Yet, constructing
specifications for the full proof relies on domain expertise and non-trivial
manpower. In view of such needs, an automated approach for specification
synthesis is desired. While existing automated approaches are limited in their
versatility, i.e., they either focus only on synthesizing loop invariants for
numerical programs, or are tailored for specific types of programs or
invariants. Programs involving multiple complicated data types (e.g., arrays,
pointers) and code structures (e.g., nested loops, function calls) are often
beyond their capabilities. To help bridge this gap, we present AutoSpec, an
automated approach to synthesize specifications for automated program
verification. It overcomes the shortcomings of existing work in specification
versatility, synthesizing satisfiable and adequate specifications for full
proof. It is driven by static analysis and program verification, and is
empowered by large language models (LLMs). AutoSpec addresses the practical
challenges in three ways: (1) driving \name by static analysis and program
verification, LLMs serve as generators to generate candidate specifications,
(2) programs are decomposed to direct the attention of LLMs, and (3) candidate
specifications are validated in each round to avoid error accumulation during
the interaction with LLMs. In this way, AutoSpec can incrementally and
iteratively generate satisfiable and adequate specifications. The evaluation
shows its effectiveness and usefulness, as it outperforms existing works by
successfully verifying 79% of programs through automatic specification
synthesis, a significant improvement of 1.592x. It can also be successfully
applied to verify the programs in a real-world X509-parser project
Revisiting QRS detection methodologies for portable, wearable, battery-operated, and wireless ECG systems
Cardiovascular diseases are the number one cause of death worldwide. Currently, portable battery-operated systems such as mobile phones with wireless ECG sensors have the potential to be used in continuous cardiac function assessment that can be easily integrated into daily life. These portable point-of-care diagnostic systems can therefore help unveil and treat cardiovascular diseases. The basis for ECG analysis is a robust detection of the prominent QRS complex, as well as other ECG signal characteristics. However, it is not clear from the literature which ECG analysis algorithms are suited for an implementation on a mobile device. We investigate current QRS detection algorithms based on three assessment criteria: 1) robustness to noise, 2) parameter choice, and 3) numerical efficiency, in order to target a universal fast-robust detector. Furthermore, existing QRS detection algorithms may provide an acceptable solution only on small segments of ECG signals, within a certain amplitude range, or amid particular types of arrhythmia and/or noise. These issues are discussed in the context of a comparison with the most conventional algorithms, followed by future recommendations for developing reliable QRS detection schemes suitable for implementation on battery-operated mobile devices.Mohamed Elgendi, Björn Eskofier, Socrates Dokos, Derek Abbot
Development and evaluation of a deep learning model for automatic segmentation of non-perfusion area in fundus fluorescein angiography
Diabetic retinopathy (DR) is the most prevalent cause of preventable vision loss worldwide, imposing a significant economic and medical burden on society today, of which early identification is the cornerstones of the management. The diagnosis and severity grading of DR rely on scales based on clinical visualized features, but lack detailed quantitative parameters. Retinal non-perfusion area (NPA) is a pathogenic characteristic of DR that symbolizes retinal hypoxia conditions, and was found to be intimately associated with disease progression, prognosis, and management. However, the practical value of NPA is constrained since it appears on fundus fluorescein angiography (FFA) as distributed, irregularly shaped, darker plaques that are challenging to measure manually. In this study, we propose a deep learning-based method, NPA-Net, for accurate and automatic segmentation of NPAs from FFA images acquired in clinical practice. NPA-Net uses the U-net structure as the basic backbone, which has an encoder-decoder model structure. To enhance the recognition performance of the model for NPA, we adaptively incorporate multi-scale features and contextual information in feature learning and design three modules: Adaptive Encoder Feature Fusion (AEFF) module, Multilayer Deep Supervised Loss, and Atrous Spatial Pyramid Pooling (ASPP) module, which enhance the recognition ability of the model for NPAs of different sizes from different perspectives. We conducted extensive experiments on a clinical dataset with 163 eyes with NPAs manually annotated by ophthalmologists, and NPA-Net achieved better segmentation performance compared to other existing methods with an area under the receiver operating characteristic curve (AUC) of 0.9752, accuracy of 0.9431, sensitivity of 0.8794, specificity of 0.9459, IOU of 0.3876 and Dice of 0.5686. This new automatic segmentation model is useful for identifying NPA in clinical practice, generating quantitative parameters that can be useful for further research as well as guiding DR detection, grading severity, treatment planning, and prognosis.</p
- …
