1,597 research outputs found

    Online classifier adaptation for cost-sensitive learning

    Full text link
    In this paper, we propose the problem of online cost-sensitive clas- sifier adaptation and the first algorithm to solve it. We assume we have a base classifier for a cost-sensitive classification problem, but it is trained with respect to a cost setting different to the desired one. Moreover, we also have some training data samples streaming to the algorithm one by one. The prob- lem is to adapt the given base classifier to the desired cost setting using the steaming training samples online. To solve this problem, we propose to learn a new classifier by adding an adaptation function to the base classifier, and update the adaptation function parameter according to the streaming data samples. Given a input data sample and the cost of misclassifying it, we up- date the adaptation function parameter by minimizing cost weighted hinge loss and respecting previous learned parameter simultaneously. The proposed algorithm is compared to both online and off-line cost-sensitive algorithms on two cost-sensitive classification problems, and the experiments show that it not only outperforms them one classification performances, but also requires significantly less running time

    FiBiNET: Combining Feature Importance and Bilinear feature Interaction for Click-Through Rate Prediction

    Full text link
    Advertising and feed ranking are essential to many Internet companies such as Facebook and Sina Weibo. Among many real-world advertising and feed ranking systems, click through rate (CTR) prediction plays a central role. There are many proposed models in this field such as logistic regression, tree based models, factorization machine based models and deep learning based CTR models. However, many current works calculate the feature interactions in a simple way such as Hadamard product and inner product and they care less about the importance of features. In this paper, a new model named FiBiNET as an abbreviation for Feature Importance and Bilinear feature Interaction NETwork is proposed to dynamically learn the feature importance and fine-grained feature interactions. On the one hand, the FiBiNET can dynamically learn the importance of features via the Squeeze-Excitation network (SENET) mechanism; on the other hand, it is able to effectively learn the feature interactions via bilinear function. We conduct extensive experiments on two real-world datasets and show that our shallow model outperforms other shallow models such as factorization machine(FM) and field-aware factorization machine(FFM). In order to improve performance further, we combine a classical deep neural network(DNN) component with the shallow model to be a deep model. The deep FiBiNET consistently outperforms the other state-of-the-art deep models such as DeepFM and extreme deep factorization machine(XdeepFM).Comment: 8 pages,5 figure

    Modified Cramer-Rao bound for M-FSK signal parameter estimation in Cauchy and Gaussian noise

    Get PDF
    The Cramer-Rao bound (CRB) provides an efficient standard for evaluating the quality of standard parameter estimators. In this paper, a modified Cramer-Rao bounds (MCRB) for modulation parameter estimations of frequency-shift-keying (FSK) signals is proposed under the condition of the Gaussian and non-Gaussian additive interference. We extend the MCRB to the estimation of a vector of non-random parameters in the presence of nuisance parameters. Moreover, the MCRB is applied to the joint estimation of phase offset, frequency offsets, frequency deviation, and symbol period of FSK signal with two important special cases of alpha stable distributions, namely, the Cauchy and the Gaussian. The extensive simulation studies are conducted to contrast the MCRB for the modulation parameter vector in different noise environments

    Magnetic domain wall engineering in a nanoscale permalloy junction

    Get PDF
    Nanoscale magnetic junctions provide a useful approach to act as building blocks for magnetoresistive random access memories (MRAM), where one of the key issues is to control the magnetic domain configuration. Here, we study the domain structure and the magnetic switching in the Permalloy (Fe20Ni80) nanoscale magnetic junctions with different thicknesses by using micromagnetic simulations. It is found that both the 90-° and 45-° domain walls can be formed between the junctions and the wire arms depending on the thickness of the device. The magnetic switching fields show distinct thickness dependencies with a broad peak varying from 7 nm to 22 nm depending on the junction sizes, and the large magnetic switching fields favor the stability of the MRAM operation

    MemoNet:Memorizing Representations of All Cross Features Efficiently via Multi-Hash Codebook Network for CTR Prediction

    Full text link
    New findings in natural language processing(NLP) demonstrate that the strong memorization capability contributes a lot to the success of large language models.This inspires us to explicitly bring an independent memory mechanism into CTR ranking model to learn and memorize all cross features'representations. In this paper,we propose multi-Hash Codebook NETwork(HCNet) as the memory mechanism for efficiently learning and memorizing representations of all cross features in CTR tasks.HCNet uses multi-hash codebook as the main memory place and the whole memory procedure consists of three phases: multi-hash addressing,memory restoring and feature shrinking.HCNet can be regarded as a general module and can be incorporated into any current deep CTR model.We also propose a new CTR model named MemoNet which combines HCNet with a DNN backbone.Extensive experimental results on three public datasets show that MemoNet reaches superior performance over state-of-the-art approaches and validate the effectiveness of HCNet as a strong memory module.Besides, MemoNet shows the prominent feature of big models in NLP,which means we can enlarge the size of codebook in HCNet to sustainably obtain performance gains.Our work demonstrates the importance and feasibility of learning and memorizing representations of all cross features ,which sheds light on a new promising research direction

    Jacobi pseudo-spectral Galerkin method for second kind Volterra integro-differential equations with a weakly singular kernel

    Get PDF
    The Jacobi pseudo-spectral Galerkin method for the Volterra integro-differential equations of the second kind with a weakly singular kernel is proposed in this paper. We provide a rigorous error analysis for the proposed method, which indicates that the numerical errors (in the Lωα,β2-norm and the L∞-norm) will decay exponentially provided that the source function is sufficiently smooth. Numerical examples are given to illustrate the theoretical results

    FiBiNet++: Reducing Model Size by Low Rank Feature Interaction Layer for CTR Prediction

    Full text link
    Click-Through Rate (CTR) estimation has become one of the most fundamental tasks in many real-world applications and various deep models have been proposed. Some research has proved that FiBiNet is one of the best performance models and outperforms all other models on Avazu dataset. However, the large model size of FiBiNet hinders its wider application. In this paper, we propose a novel FiBiNet++ model to redesign FiBiNet's model structure, which greatly reduces model size while further improves its performance. One of the primary techniques involves our proposed "Low Rank Layer" focused on feature interaction, which serves as a crucial driver of achieving a superior compression ratio for models. Extensive experiments on three public datasets show that FiBiNet++ effectively reduces non-embedding model parameters of FiBiNet by 12x to 16x on three datasets. On the other hand, FiBiNet++ leads to significant performance improvements compared to state-of-the-art CTR methods, including FiBiNet
    corecore