39 research outputs found
Large Margin Distribution Machine Recursive Feature Elimination
We gratefully thank Dr Teng Zhang and Prof Zhi-Hua Zhou for providing the source code of “LDM” source code and their kind technical assistance. This work is supported by the National Natural Science Foundation of China (Nos. 61472159, 61572227) and Development Project of Jilin Province of China (Nos. 20160204022GX, 2017C033). This work is also partially supported by the 2015 Scottish Crucible Award funded by the Royal Society of Edinburgh and the 2016 PECE bursary provided by the Scottish Informatics & Computer Science Alliance (SICSA).Postprin
e-Distance Weighted Support Vector Regression
We propose a novel support vector regression approach called e-Distance
Weighted Support Vector Regression (e-DWSVR).e-DWSVR specifically addresses two
challenging issues in support vector regression: first, the process of noisy
data; second, how to deal with the situation when the distribution of boundary
data is different from that of the overall data. The proposed e-DWSVR optimizes
the minimum margin and the mean of functional margin simultaneously to tackle
these two issues. In addition, we use both dual coordinate descent (CD) and
averaged stochastic gradient descent (ASGD) strategies to make e-DWSVR scalable
to large scale problems. We report promising results obtained by e-DWSVR in
comparison with existing methods on several benchmark datasets
Less but Better: Generalization Enhancement of Ordinal Embedding via Distributional Margin
In the absence of prior knowledge, ordinal embedding methods obtain new
representation for items in a low-dimensional Euclidean space via a set of
quadruple-wise comparisons. These ordinal comparisons often come from human
annotators, and sufficient comparisons induce the success of classical
approaches. However, collecting a large number of labeled data is known as a
hard task, and most of the existing work pay little attention to the
generalization ability with insufficient samples. Meanwhile, recent progress in
large margin theory discloses that rather than just maximizing the minimum
margin, both the margin mean and variance, which characterize the margin
distribution, are more crucial to the overall generalization performance. To
address the issue of insufficient training samples, we propose a margin
distribution learning paradigm for ordinal embedding, entitled Distributional
Margin based Ordinal Embedding (\textit{DMOE}). Precisely, we first define the
margin for ordinal embedding problem. Secondly, we formulate a concise
objective function which avoids maximizing margin mean and minimizing margin
variance directly but exhibits the similar effect. Moreover, an Augmented
Lagrange Multiplier based algorithm is customized to seek the optimal solution
of \textit{DMOE} effectively. Experimental studies on both simulated and
real-world datasets are provided to show the effectiveness of the proposed
algorithm.Comment: Accepted by AAAI 201