87 research outputs found
CSNL: A cost-sensitive non-linear decision tree algorithm
This article presents a new decision tree learning algorithm called CSNL that induces Cost-Sensitive Non-Linear decision trees. The algorithm is based on the hypothesis that nonlinear decision nodes provide a better basis than axis-parallel decision nodes and utilizes discriminant analysis to construct nonlinear decision trees that take account of costs of misclassification.
The performance of the algorithm is evaluated by applying it to seventeen datasets and the results are compared with those obtained by two well known cost-sensitive algorithms, ICET and MetaCost, which generate multiple trees to obtain some of the best results to date. The results show that CSNL performs at least as well, if not better than these algorithms, in more than twelve of the datasets and is considerably faster. The use of bagging with CSNL further enhances its performance showing the significant benefits of using nonlinear decision nodes.
The performance of the algorithm is evaluated by applying it to seventeen data sets and the results are
compared with those obtained by two well known cost-sensitive algorithms, ICET and MetaCost, which generate multiple trees to obtain some of the best results to date.
The results show that CSNL performs at least as well, if not better than these algorithms, in more than twelve of the data sets and is considerably faster.
The use of bagging with CSNL further enhances its performance showing the significant benefits of using non-linear decision nodes
Inducing safer oblique trees without costs
Decision tree induction has been widely studied and applied. In safety applications, such as determining whether a chemical process is safe or whether a person has a medical condition, the cost of misclassification in one of the classes is significantly higher than in the other class. Several authors have tackled this problem by developing cost-sensitive decision tree learning algorithms or have suggested ways of changing the
distribution of training examples to bias the decision tree learning process so as to take account of costs. A prerequisite for applying such algorithms is the availability of costs of misclassification.
Although this may be possible for some applications, obtaining reasonable estimates of costs of misclassification is not easy in the area of safety.
This paper presents a new algorithm for applications where the cost of misclassifications cannot be quantified, although the cost of misclassification in one class is known to be significantly higher than in another class. The algorithm utilizes linear discriminant analysis to identify oblique relationships between continuous attributes and then carries out an appropriate modification to ensure that the resulting tree errs on the side of safety. The algorithm is evaluated with respect to one of the best known cost-sensitive algorithms (ICET), a well-known oblique decision tree algorithm (OC1) and an algorithm that utilizes robust linear programming
Nanotechnology Applications for Chemical and Biological Sensors
Recent discoveries indicate that when the materials are brought down to sizes in the range 1–100 nm, theseexhibit unique electrical, optical, magnetic, chemical, and mechanical properties. Methods have now beenestablished to obtain the monodisperse nanocrystals of various metallic and semiconducting materials, single-walled and multi-walled nanotubes of carbon and other metallic and non-metallic materials together withorganic nanomaterials such as supra-molecular nanostructures, dendrimers, hybrid composites with tailoredfunctionalities. The high surface-to-volume ratio with an added element of porosity makes these highly potentialcandidates for chemical and biological sensor applications with higher degree of sensitivity and selectivity ascompared to their bulk counterparts. The paper reviews the recent developments and applications of chemicaland biological sensors based on nanomaterials of various structural forms.Defence Science Journal, 2008, 58(5), pp.636-649, DOI:http://dx.doi.org/10.14429/dsj.58.168
Use of Xpert MTB/RIF in Decentralized Public Health Settings and Its Effect on Pulmonary TB and DR-TB Case Finding in India
Background Xpert MTB/RIF, the first automated molecular test for tuberculosis, is transforming the diagnostic landscape in high-burden settings. This study assessed the impact of up-front Xpert MTB/RIF testing on detection of pulmonary tuberculosis (PTB) and rifampicin-resistant PTB (DR-TB) cases in India. Methods This demonstration study was implemented in 18 sub-district level TB programme units (TUs) in India in diverse geographic and demographic settings covering a population of 8.8 million. A baseline phase in 14 TUs captured programmatic baseline data, and an intervention phase in 18 TUs had Xpert MTB/RIF offered to all presumptive TB patients. We estimated changes in detection of TB and DR-TB, the former using binomial regression models to adjust for clustering and covariates. Results In the 14 study TUs, which participated in both phases, 10,675 and 70,556 presumptive TB patients were enrolled in the baseline and intervention phase, respectively, and 1,532 (14.4%) and 14,299 (20.3%) bacteriologically confirmed PTB cases were detected. The implementation of Xpert MTB/RIF was associated with increases in both notification rates of bacteriologically confirmed TB cases (adjusted incidence rate ratio [aIRR] 1.39; CI 1.18-1.64), and proportion of bacteriological confirmed TB cases among presumptive TB cases (adjusted risk ratio (aRR) 1.33; CI 1.6-1.52). Compared with the baseline strategy of selective drug-susceptibility testing only for PTB cases at high risk of drug-resistant TB, Xpert MTB/RIF implementation increased rifampicin resistant TB case detection by over fivefold. Among, 2765 rifampicin resistance cases detected, 1055 were retested with conventional drug susceptibility testing (DST). Positive predictive value (PPV) of rifampicin resistance detected by Xpert MTB/RIF was 94.7% (CI 91.3-98.1), in comparison to conventional DST. Conclusion Introduction of Xpert MTB/RIF as initial diagnostic test for TB in public health facilities significantly increased case-notification rates of all bacteriologically confirmed TB by 39% and rifampicin-resistant TB case notification by fivefold
A survey of cost-sensitive decision tree induction algorithms
The past decade has seen a significant interest on the problem of inducing decision trees that take account of costs of misclassification and costs of acquiring the features used for decision making. This survey identifies over 50 algorithms including approaches that are direct adaptations of accuracy based methods, use genetic algorithms, use anytime methods and utilize boosting and bagging. The survey brings together these different studies and novel approaches to cost-sensitive decision tree learning, provides a useful taxonomy, a historical timeline of how the field has developed and should provide a useful reference point for future research in this field
CHERI: A hybrid capability-system architecture for scalable software compartmentalization
CHERI extends a conventional RISC Instruction-
Set Architecture, compiler, and operating system to support
fine-grained, capability-based memory protection to mitigate
memory-related vulnerabilities in C-language TCBs. We describe
how CHERI capabilities can also underpin a hardware-software
object-capability model for application compartmentalization
that can mitigate broader classes of attack. Prototyped as an
extension to the open-source 64-bit BERI RISC FPGA softcore
processor, FreeBSD operating system, and LLVM compiler,
we demonstrate multiple orders-of-magnitude improvement in
scalability, simplified programmability, and resulting tangible
security benefits as compared to compartmentalization based on
pure Memory-Management Unit (MMU) designs. We evaluate
incrementally deployable CHERI-based compartmentalization
using several real-world UNIX libraries and applications.We thank our colleagues Ross Anderson, Ruslan Bukin,
Gregory Chadwick, Steve Hand, Alexandre Joannou, Chris
Kitching, Wojciech Koszek, Bob Laddaga, Patrick Lincoln,
Ilias Marinos, A Theodore Markettos, Ed Maste, Andrew W.
Moore, Alan Mujumdar, Prashanth Mundkur, Colin Rothwell,
Philip Paeps, Jeunese Payne, Hassen Saidi, Howie Shrobe, and
Bjoern Zeeb, our anonymous reviewers, and shepherd Frank
Piessens, for their feedback and assistance. This work is part of
the CTSRD and MRC2 projects sponsored by the Defense Advanced
Research Projects Agency (DARPA) and the Air Force
Research Laboratory (AFRL), under contracts FA8750-10-C-
0237 and FA8750-11-C-0249. The views, opinions, and/or
findings contained in this paper are those of the authors and
should not be interpreted as representing the official views
or policies, either expressed or implied, of the Department
of Defense or the U.S. Government. We acknowledge the EPSRC
REMS Programme Grant [EP/K008528/1], Isaac Newton
Trust, UK Higher Education Innovation Fund (HEIF), Thales
E-Security, and Google, Inc.This is the author accepted manuscript. The final version is available at http://dx.doi.org/10.1109/SP.2015.
Beyond the PDP-11: Architectural support for a memory-safe C abstract machine
We propose a new memory-safe interpretation of the C abstract machine that provides stronger protection to benefit security and debugging. Despite ambiguities in the specification intended to provide implementation flexibility, contemporary implementations of C have converged on a memory model similar to the PDP-11, the original target for C. This model lacks support for memory safety despite well documented impacts on security and reliability. Attempts to change this model are often hampered by assumptions embedded in a large body of existing C code, dating back to the memory model exposed by the original C compiler for the PDP-11. Our experience with attempting to implement a memory-safe variant of C on the CHERI experimental microprocessor led us to identify a number of problematic idioms. We describe these as well as their interaction with existing memory safety schemes and the assumptions that they make beyond the requirements of the C specification. Finally, we refine the CHERI ISA and abstract model for C, by combining elements of the CHERI capability model and fat pointers, and present a softcore CPU that implements a C abstract machine that can run legacy C code with strong memory protection guarantees.This work is part of the CTSRD and MRC2 projects that are sponsored by the Defense Advanced Research Projects Agency (DARPA) and the Air Force Research Laboratory (AFRL), under contracts FA8750-10-C-0237 and FA8750- 11-C-0249. The views, opinions, and/or findings contained in this paper are those of the authors and should not be interpreted as representing the official views or policies, either expressed or implied, of the Department of Defense or the U.S. Government. We gratefully acknowledge Google, Inc. for its sponsorship
- …