1,513 research outputs found
Regularizing soft decision trees
Recently, we have proposed a new decision tree family called soft decision trees where a node chooses both its left and right children with different probabilities as given by a gating function, different from a hard decision node which chooses one of the two. In this paper, we extend the original algorithm by introducing local dimension reduction via L-1 and L-2 regularization for feature selection and smoother fitting. We compare our novel approach with the standard decision tree algorithms over 27 classification data sets. We see that both regularized versions have similar generalization ability with less complexity in terms of number of nodes, where L-2 seems to work slightly better than L-1.Publisher's VersionAuthor Post Prin
Recommended from our members
Building more accurate decision trees with the additive tree.
The expansion of machine learning to high-stakes application domains such as medicine, finance, and criminal justice, where making informed decisions requires clear understanding of the model, has increased the interest in interpretable machine learning. The widely used Classification and Regression Trees (CART) have played a major role in health sciences, due to their simple and intuitive explanation of predictions. Ensemble methods like gradient boosting can improve the accuracy of decision trees, but at the expense of the interpretability of the generated model. Additive models, such as those produced by gradient boosting, and full interaction models, such as CART, have been investigated largely in isolation. We show that these models exist along a spectrum, revealing previously unseen connections between these approaches. This paper introduces a rigorous formalization for the additive tree, an empirically validated learning technique for creating a single decision tree, and shows that this method can produce models equivalent to CART or gradient boosted stumps at the extremes by varying a single parameter. Although the additive tree is designed primarily to provide both the model interpretability and predictive performance needed for high-stakes applications like medicine, it also can produce decision trees represented by hybrid models between CART and boosted stumps that can outperform either of these approaches
Deciding How to Decide: Dynamic Routing in Artificial Neural Networks
We propose and systematically evaluate three strategies for training
dynamically-routed artificial neural networks: graphs of learned
transformations through which different input signals may take different paths.
Though some approaches have advantages over others, the resulting networks are
often qualitatively similar. We find that, in dynamically-routed networks
trained to classify images, layers and branches become specialized to process
distinct categories of images. Additionally, given a fixed computational
budget, dynamically-routed networks tend to perform better than comparable
statically-routed networks.Comment: ICML 2017. Code at https://github.com/MasonMcGill/multipath-nn Video
abstract at https://youtu.be/NHQsDaycwy
Automated Segmentation of Pulmonary Lobes using Coordination-Guided Deep Neural Networks
The identification of pulmonary lobes is of great importance in disease
diagnosis and treatment. A few lung diseases have regional disorders at lobar
level. Thus, an accurate segmentation of pulmonary lobes is necessary. In this
work, we propose an automated segmentation of pulmonary lobes using
coordination-guided deep neural networks from chest CT images. We first employ
an automated lung segmentation to extract the lung area from CT image, then
exploit volumetric convolutional neural network (V-net) for segmenting the
pulmonary lobes. To reduce the misclassification of different lobes, we
therefore adopt coordination-guided convolutional layers (CoordConvs) that
generate additional feature maps of the positional information of pulmonary
lobes. The proposed model is trained and evaluated on a few publicly available
datasets and has achieved the state-of-the-art accuracy with a mean Dice
coefficient index of 0.947 0.044.Comment: ISBI 2019 (Oral
- …