30 research outputs found

    VAEFL: Integrating variational autoencoders for privacy preservation and performance retention in federated learning

    Get PDF
    Federated Learning (FL) heralds a paradigm shift in the training of artificial intelligence (AI) models by fostering collaborative model training while safeguarding client data privacy. In sectors where data sensitivity and AI model security are of paramount importance, such as fintech and biomedicine, maintaining the utility of models without compromising privacy is crucial with the growing application of AI technologies. Therefore, the adoption of FL is attracting significant attention. However, traditional FL methods are susceptible to Deep Leakage from Gradients (DLG) attacks, and typical defensive strategies in current research, such as secure multi-party computation and differential privacy, often lead to excessive computational costs or significant decreases in model accuracy. To address DLG attacks in FL, this study introduces VAEFL, an innovative FL framework that incorporates Variational Autoencoders (VAEs) to enhance privacy protection without undermining the predictive prowess of the models. VAEFL strategically partitions the model into a private encoder and a public decoder. The private encoder, remaining local, transmutes sensitive data into a latent space fortified for privacy, while the public decoder and classifier, through collaborative training across clients, learn to derive precise predictions from the encoded data. This bifurcation ensures that sensitive data attributes are not disclosed, circumventing gradient leakage attacks and simultaneously allowing the global model to benefit from the diverse knowledge of client datasets. Comprehensive experiments demonstrate that VAEFL not only surpasses standard FL benchmarks in privacy preservation but also maintains competitive performance in predictive tasks. VAEFL thus establishes a novel equilibrium between data privacy and model utility, offering a secure and efficient FL approach for the sensitive application of FL in the financial domain

    The FOXK1-CCDC43 Axis Promotes the Invasion and Metastasis of Colorectal Cancer Cells

    Get PDF
    Background/Aims: The CCDC43 gene is conserved in human, rhesus monkey, mouse and zebrafish. Bioinformatics studies have demonstrated the abnormal expression of CCDC43 gene in colorectal cancer (CRC). However, the role and molecular mechanism of CCDC43 in CRC remain unknown. Methods: The functional role of CCDC43 and FOXK1 in epithelial-mesenchymal transition (EMT) was determined using immunohistochemistry, flow cytometry, western blot, EdU incorporation, luciferase, chromatin Immunoprecipitation (ChIP) and cell invasion assays. Results: The CCDC43 gene was overexpressed in human CRC. High expression of CCDC43 protein was associated with tumor progression and poor prognosis in patients with CRC. Moreover, the induction of EMT by CCDC43 occurred through TGF-β signaling. Furthermore, a positive correlation between the expression patterns of CCDC43 and FOXK1 was observed in CRC cells. Promoter assays demonstrated that FOXK1 directly bound and activated the human CCDC43 gene promoter. In addition, CCDC43 was necessary for FOXK1- mediated EMT and metastasis in vitro and vivo. Taken together, this work identified that CCDC43 promoted EMT and was a direct transcriptional target of FOXK1 in CRC cells. Conclusion: FOXK1-CCDC43 axis might be helpful to develop the drugs for the treatment of CRC

    Large-Scale Video Hashing via Structure Learning

    No full text
    Recently, learning based hashing methods have become popular for indexing large-scale media data. Hashing methods map high-dimensional features to compact binary codes that are efficient to match and robust in preserving original similarity. However, most of the existing hashing methods treat videos as a simple aggregation of independent frames and index each video through combining the indexes of frames. The structure information of videos, e.g., discriminative local visual commonality and temporal consistency, is often neglected in the design of hash functions. In this paper, we propose a supervised method that explores the structure learning techniques to design efficient hash functions. The proposed video hashing method formulates a minimization problem over a structure-regularized empirical loss. In particular, the structure regularization exploits the common local visual patterns occurring in video frames that are associated with the same semantic class, and simultaneously preserves the temporal consistency over successive frames from the same video. We show that the minimization objective can be efficiently solved by an Accelerated Proximal Gradient (APG) method. Extensive experiments on two large video benchmark datasets (up to around 150K video clips with over 12 million frames) show that the proposed method significantly outperforms the state-ofthe-art hashing methods. 1

    Columbia-Ucf Trecvid2010 Multimedia Event Detection: Combining Multiple Modalities, Contextual Concepts, And Temporal Matching

    No full text
    TRECVID Multimedia Event Detection offers an interesting but very challenging task in detecting highlevel complex events (Figure 1) in user-generated videos. In this paper, we will present an overview and comparative analysis of our results, which achieved top performance among all 45 submissions in TRECVID 2010. Our aim is to answer the following questions. What kind of feature is more effective for multimedia event detection? Are features from different feature modalities (e.g., audio and visual) complementary for event detection? Can we benefit from generic concept detection of background scenes, human actions, and audio concepts? Are sequence matching and event-specific object detectors critical? Our findings indicate that spatial-temporal feature is very effective for event detection, and it\u27s also very complementary to other features such as static SIFT and audio features. As a result, our baseline run combining these three features already achieves very impressive results, with a mean minimal normalized cost (MNC) of 0.586. Incorporating the generic concept detectors using a graph diffusion algorithm provides marginal gains (mean MNC 0.579). Sequence matching with Earth Mover\u27s Distance (EMD) further improves the results (mean MNC 0.565). The event-specific detector ( batter ), however, didn\u27t prove useful from our current re-ranking tests. We conclude that it is important to combine strong complementary features from multiple modalities for multimedia event detection, and cross-frame matching is helpful in coping with temporal order variation. Leveraging contextual concept detectors and foreground activities remains a very attractive direction requiring further research
    corecore