240 research outputs found

    Social Impacts of the Asian Crisis: Policy Challenges and Lessons

    Get PDF
    human development, economic growth, globalization, inequality, poverty

    Monitoring The Evolutionary Patterns of Technological Advances Based On the Dynamic Patent Lattice: A Modified Formal Concept Analysis Approach

    Get PDF
    The strategic importance of monitoring changes in technology has been highlighted for achieving and maintaining firms’ competitive positions. In this respect, among others, patent citation analysis has been the most frequently adopted tool. However, it is subject to some drawbacks that stem from only consideration of citing-cited information and time lags between citing and cited patents. In response, we propose a modified formal concept analysis (FCA) approach to developing dynamic patent lattice that can analyze the complex relations among patents and evolutionary patterns of technological advances. The FCA is a mathematical tool for grouping objects with shared properties based on the lattice theory. The distinct strength of FCA, vis-á-vis other methods, lies in structuring and displaying the relations among objects in the amount of data. The FCA is modified to take time periods into account for the purpose of technology monitoring. Specifically, patents are first collected and transformed into structured data. Next, the dynamic patent lattice is developed by executing a modified FCA algorithm based on patent context. Finally, quantitative indexes are defined and gauged to conduct a more detailed analysis and obtain richer information. The proposed dynamic patent lattice can be effectively employed to aid decision making in technology monitoring

    Interference Alignment Through User Cooperation for Two-cell MIMO Interfering Broadcast Channels

    Full text link
    This paper focuses on two-cell multiple-input multiple-output (MIMO) Gaussian interfering broadcast channels (MIMO-IFBC) with KK cooperating users on the cell-boundary of each BS. It corresponds to a downlink scenario for cellular networks with two base stations (BSs), and KK users equipped with Wi-Fi interfaces enabling to cooperate among users on a peer-to-peer basis. In this scenario, we propose a novel interference alignment (IA) technique exploiting user cooperation. Our proposed algorithm obtains the achievable degrees of freedom (DoF) of 2K when each BS and user have M=K+1M=K+1 transmit antennas and N=KN=K receive antennas, respectively. Furthermore, the algorithm requires only a small amount of channel feedback information with the aid of the user cooperation channels. The simulations demonstrate that not only are the analytical results valid, but the achievable DoF of our proposed algorithm also outperforms those of conventional techniques.Comment: This paper will appear in IEEE GLOBECOM 201

    The Minimum Scheduling Time for Convergecast in Wireless Sensor Networks

    Get PDF
    We study the scheduling problem for data collection from sensor nodes to the sink node in wireless sensor networks, also referred to as the convergecast problem. The convergecast problem in general network topology has been proven to be NP-hard. In this paper, we propose our heuristic algorithm (finding the minimum scheduling time for convergecast (FMSTC)) for general network topology and evaluate the performance by simulation. The results of the simulation showed that the number of time slots to reach the sink node decreased with an increase in the power. We compared the performance of the proposed algorithm to the optimal time slots in a linear network topology. The proposed algorithm for convergecast in a general network topology has 2.27 times more time slots than that of a linear network topology. To the best of our knowledge, the proposed method is the first attempt to apply the optimal algorithm in a linear network topology to a general network topology

    Lessons from the 1997 and 2008 Crises in Korea

    Get PDF

    Abnormality Diagnosis Model for Nuclear Power Plants Using Two-Stage Gated Recurrent Units

    Get PDF
    A nuclear power plant is a large complex system with tens of thousands of components. To ensure plant safety, the early and accurate diagnosis of abnormal situations is an important factor. To prevent misdiagnosis, operating procedures provide the anticipated symptoms of abnormal situations. While the more severe emergency situations total less than ten cases and can be diagnosed by dozens of key plant parameters, abnormal situations on the other hand include hundreds of cases and a multitude of parameters that should be considered for diagnosis. The tasks required of operators to select the appropriate operating procedure by monitoring large amounts of information within a limited amount of time can burden operators. This paper aims to develop a system that can, in a short time and with high accuracy, select the appropriate operating procedure and sub-procedure in an abnormal situation. Correspondingly, the proposed model has two levels of prediction to determine the procedure level and the detailed cause of an event. Simulations were conducted to evaluate the developed model, with results demonstrating high levels of performance. The model is expected to reduce the workload of operators in abnormal situations by providing the appropriate procedure to ultimately improve plant safety. (c) 2020 Korean Nuclear Society, Published by Elsevier Korea LLC

    Learning to Quantize Deep Networks by Optimizing Quantization Intervals with Task Loss

    Full text link
    Reducing bit-widths of activations and weights of deep networks makes it efficient to compute and store them in memory, which is crucial in their deployments to resource-limited devices, such as mobile phones. However, decreasing bit-widths with quantization generally yields drastically degraded accuracy. To tackle this problem, we propose to learn to quantize activations and weights via a trainable quantizer that transforms and discretizes them. Specifically, we parameterize the quantization intervals and obtain their optimal values by directly minimizing the task loss of the network. This quantization-interval-learning (QIL) allows the quantized networks to maintain the accuracy of the full-precision (32-bit) networks with bit-width as low as 4-bit and minimize the accuracy degeneration with further bit-width reduction (i.e., 3 and 2-bit). Moreover, our quantizer can be trained on a heterogeneous dataset, and thus can be used to quantize pretrained networks without access to their training data. We demonstrate the effectiveness of our trainable quantizer on ImageNet dataset with various network architectures such as ResNet-18, -34 and AlexNet, on which it outperforms existing methods to achieve the state-of-the-art accuracy
    corecore