113 research outputs found

    Stepwise Lithospheric Delamination Leads to Pulsed Cenozoic Uplifts of Central Tien Shan

    Get PDF
    AbstractThe Tien Shan provides an ideal site to study mechanism of intracontinental orogeny due to distant effect of Indo-Asian collision. We investigate lithospheric structures, in particular the lithosphere-asthenosphere boundary (LAB), of Central Tien Shan (CTS) using S wave receiver functions. The results show distinct structures across the orogen. Under the southern CTS, the LAB is shallower than that of the Tarim Basin; a 50 km vertical offset implies that part of the lithosphere has been delaminated. Under the middle CTS, two phases of negative velocity gradient are obtained, which may indicate a new LAB and an ongoing delamination underneath. Under the northern CTS and Kazakh Shield northward, the lithosphere is stable although the LAB inclines southward slightly. The two periods of lithospheric delamination under the southern and middle CTS account well for pulsed uplifts of the Tien Shan at ~11-8 Ma and ~5-0 Ma, respectively

    Index tracking model, downside risk and non-parametric kernel estimation

    Get PDF
    In this paper, we propose an index tracking model with the conditional value-at-risk (CVaR) constraint based on a non-parametric kernel (NPK) estimation framework. In theory, we demonstrate that the index tracking model with the CVaR constraint is a convex optimization problem. We then derive NPK estimators for tracking errors and CVaR, and thereby construct the NPK index tracking model. Monte Carlo simulations show that the NPK method outperforms the linear programming (LP) method in terms of estimation accuracy. In addition, the NPK method can enhance computational efficiency when the sample size is large. Empirical tests show that the NPK method can effectively control downside risk and obtain higher excess returns, in both bearish and bullish market environments

    Efficient Black-box Checking of Snapshot Isolation in Databases

    Full text link
    Snapshot isolation (SI) is a prevalent weak isolation level that avoids the performance penalty imposed by serializability and simultaneously prevents various undesired data anomalies. Nevertheless, SI anomalies have recently been found in production cloud databases that claim to provide the SI guarantee. Given the complex and often unavailable internals of such databases, a black-box SI checker is highly desirable. In this paper we present PolySI, a novel black-box checker that efficiently checks SI and provides understandable counterexamples upon detecting violations. PolySI builds on a novel characterization of SI using generalized polygraphs (GPs), for which we establish its soundness and completeness. PolySI employs an SMT solver and also accelerates SMT solving by utilizing the compact constraint encoding of GPs and domain-specific optimizations for pruning constraints. As demonstrated by our extensive assessment, PolySI successfully reproduces all of 2477 known SI anomalies, detects novel SI violations in three production cloud databases, identifies their causes, outperforms the state-of-the-art black-box checkers under a wide range of workloads, and can scale up to large-sized workloads.Comment: 20 pages, 15 figures, accepted by PVLD

    Predicting no-show medical appointments using machine learning

    Get PDF
    Health care centers face many issues due to the limited availability of resources, such as funds, equipment, beds, physicians, and nurses. Appointment absences lead to a waste of hospital resources as well as endangering patient health. This fact makes unattended medi- cal appointments both socially expensive and economically costly. This research aimed to build a predictive model to identify whether an appointment would be a no-show or not in order to reduce its consequences. This paper proposes a multi-stage framework to build an accurate predictor that also tackles the imbalanced property that the data exhibits. The first stage includes dimensionality reduction to compress the data into its most important components. The second stage deals with the imbalanced nature of the data. Different machine learning algorithms were used to build the classifiers in the third stage. Various evaluation metrics are also discussed and an evaluation scheme that fits the problem at hand is described. The work presented in this paper will help decision makers at health care centers to implement effective strategies to reduce the number of no-shows

    An insight into imbalanced Big Data classification: outcomes and challenges

    Get PDF
    Big Data applications are emerging during the last years, and researchers from many disciplines are aware of the high advantages related to the knowledge extraction from this type of problem. However, traditional learning approaches cannot be directly applied due to scalability issues. To overcome this issue, the MapReduce framework has arisen as a “de facto” solution. Basically, it carries out a “divide-and-conquer” distributed procedure in a fault-tolerant way to adapt for commodity hardware. Being still a recent discipline, few research has been conducted on imbalanced classification for Big Data. The reasons behind this are mainly the difficulties in adapting standard techniques to the MapReduce programming style. Additionally, inner problems of imbalanced data, namely lack of data and small disjuncts, are accentuated during the data partitioning to fit the MapReduce programming style. This paper is designed under three main pillars. First, to present the first outcomes for imbalanced classification in Big Data problems, introducing the current research state of this area. Second, to analyze the behavior of standard pre-processing techniques in this particular framework. Finally, taking into account the experimental results obtained throughout this work, we will carry out a discussion on the challenges and future directions for the topic.This work has been partially supported by the Spanish Ministry of Science and Technology under Projects TIN2014-57251-P and TIN2015-68454-R, the Andalusian Research Plan P11-TIC-7765, the Foundation BBVA Project 75/2016 BigDaPTOOLS, and the National Science Foundation (NSF) Grant IIS-1447795

    Rotor end factors for 2-D FEA of induction motors with smooth or slitted solid rotor

    No full text
    Modifying the equivalent rotor resistivity with rotor end factor in 2-dimension (2-D) finite element analysis (FEA) is an effective way to analyze the 3-dimension (3-D) solid rotor problems. For the smooth solid rotor, five different rotor end factors are discussed and compared with each other. It is theoretically clarified that the resistivity of rotor in 2-D FEA should be multiplied by the square of rotor end factors to take the 3-D end effect of solid rotor into account. For the slitted solid rotor, an improved rotor end factor is proposed based on the equivalent area algorithm of eddy currents in rotor, since the end factors of smooth solid rotor are not applicable. Finally, the time-harmonic finite element method (FEM) combined with the rotor end factor is applied to analyze the performance of solid rotor induction motor. The tested and computed results are in good agreement, which proves the effectiveness of rotor end factor for the simplication of 3-D solid rotor problems

    Safety risk evaluations of deep foundation construction schemes based on imbalanced data sets

    Get PDF
    Safety risk evaluations of deep foundation construction schemes are important to ensure safety. However, the amount of knowledge on these evaluations is large, and the historical data of deep foundation engineering is imbalanced. Some adverse factors influence the quality and efficiency of evaluations using traditional manual evaluation tools. Machine learning guarantees the quality of imbalanced data classifications. In this study, three strategies are proposed to improve the classification accuracy of imbalanced data sets. First, data set information redundancy is reduced using a binary particle swarm optimization algorithm. Then, a classification algorithm is modified using an Adaboost-enhanced support vector machine classifier. Finally, a new classification evaluation standard, namely, the area under the ROC curve, is adopted to ensure the classifier to be impartial to the minority. A transverse comparison experiment using multiple classification algorithms shows that the proposed integrated classification algorithm can overcome difficulties associated with correctly classifying minority samples in imbalanced data sets. The algorithm can also improve construction safety management evaluations, relieve the pressure from the lack of experienced experts accompanying rapid infrastructure construction, and facilitate knowledge reuse in the field of architecture, engineering, and construction

    A Hybrid Method for Generation of Typical Meteorological Years for Different Climates of China

    No full text
    Since a representative dataset of the climatological features of a location is important for calculations relating to many fields, such as solar energy system, agriculture, meteorology and architecture, there is a need to investigate the methodology for generating a typical meteorological year (TMY). In this paper, a hybrid method with mixed treatment of selected results from the Danish method, the Festa-Ratto method, and the modified typical meteorological year method is proposed to determine typical meteorological years for 35 locations in six different climatic zones of China (Tropical Zone, Subtropical Zone, Warm Temperate Zone, Mid Temperate Zone, Cold Temperate Zone and Tibetan Plateau Zone). Measured weather data (air dry-bulb temperature, air relative humidity, wind speed, pressure, sunshine duration and global solar radiation), which cover the period of 1994–2015, are obtained and applied in the process of forming TMY. The TMY data and typical solar radiation data are investigated and analyzed in this study. It is found that the results of the hybrid method have better performance in terms of the long-term average measured data during the year than the other investigated methods. Moreover, the Gaussian process regression (GPR) model is recommended to forecast the monthly mean solar radiation using the last 22 years (1994–2015) of measured data
    corecore