50 research outputs found

    Towards Optimal Discrete Online Hashing with Balanced Similarity

    Full text link
    When facing large-scale image datasets, online hashing serves as a promising solution for online retrieval and prediction tasks. It encodes the online streaming data into compact binary codes, and simultaneously updates the hash functions to renew codes of the existing dataset. To this end, the existing methods update hash functions solely based on the new data batch, without investigating the correlation between such new data and the existing dataset. In addition, existing works update the hash functions using a relaxation process in its corresponding approximated continuous space. And it remains as an open problem to directly apply discrete optimizations in online hashing. In this paper, we propose a novel supervised online hashing method, termed Balanced Similarity for Online Discrete Hashing (BSODH), to solve the above problems in a unified framework. BSODH employs a well-designed hashing algorithm to preserve the similarity between the streaming data and the existing dataset via an asymmetric graph regularization. We further identify the "data-imbalance" problem brought by the constructed asymmetric graph, which restricts the application of discrete optimization in our problem. Therefore, a novel balanced similarity is further proposed, which uses two equilibrium factors to balance the similar and dissimilar weights and eventually enables the usage of discrete optimizations. Extensive experiments conducted on three widely-used benchmarks demonstrate the advantages of the proposed method over the state-of-the-art methods.Comment: 8 pages, 11 figures, conferenc

    MemoChat: Tuning LLMs to Use Memos for Consistent Long-Range Open-Domain Conversation

    Full text link
    We propose MemoChat, a pipeline for refining instructions that enables large language models (LLMs) to effectively employ self-composed memos for maintaining consistent long-range open-domain conversations. We demonstrate a long-range open-domain conversation through iterative "memorization-retrieval-response" cycles. This requires us to carefully design tailored tuning instructions for each distinct stage. The instructions are reconstructed from a collection of public datasets to teach the LLMs to memorize and retrieve past dialogues with structured memos, leading to enhanced consistency when participating in future conversations. We invite experts to manually annotate a test set designed to evaluate the consistency of long-range conversations questions. Experiments on three testing scenarios involving both open-source and API-accessible chatbots at scale verify the efficacy of MemoChat, which outperforms strong baselines.Comment: Codes, data and models will be available soo

    Dynamic Sparse No Training: Training-Free Fine-tuning for Sparse LLMs

    Full text link
    The ever-increasing large language models (LLMs), though opening a potential path for the upcoming artificial general intelligence, sadly drops a daunting obstacle on the way towards their on-device deployment. As one of the most well-established pre-LLMs approaches in reducing model complexity, network pruning appears to lag behind in the era of LLMs, due mostly to its costly fine-tuning (or re-training) necessity under the massive volumes of model parameter and training data. To close this industry-academia gap, we introduce Dynamic Sparse No Training (DSnoT), a training-free fine-tuning approach that slightly updates sparse LLMs without the expensive backpropagation and any weight updates. Inspired by the Dynamic Sparse Training, DSnoT minimizes the reconstruction error between the dense and sparse LLMs, in the fashion of performing iterative weight pruning-and-growing on top of sparse LLMs. To accomplish this purpose, DSnoT particularly takes into account the anticipated reduction in reconstruction error for pruning and growing, as well as the variance w.r.t. different input data for growing each weight. This practice can be executed efficiently in linear time since its obviates the need of backpropagation for fine-tuning LLMs. Extensive experiments on LLaMA-V1/V2, Vicuna, and OPT across various benchmarks demonstrate the effectiveness of DSnoT in enhancing the performance of sparse LLMs, especially at high sparsity levels. For instance, DSnoT is able to outperform the state-of-the-art Wanda by 26.79 perplexity at 70% sparsity with LLaMA-7B. Our paper offers fresh insights into how to fine-tune sparse LLMs in an efficient training-free manner and open new venues to scale the great potential of sparsity to LLMs. Codes are available at https://github.com/zyxxmu/DSnoT.Comment: Published as a conference paper at ICLR 202

    Unified and Dynamic Graph for Temporal Character Grouping in Long Videos

    Full text link
    Video temporal character grouping locates appearing moments of major characters within a video according to their identities. To this end, recent works have evolved from unsupervised clustering to graph-based supervised clustering. However, graph methods are built upon the premise of fixed affinity graphs, bringing many inexact connections. Besides, they extract multi-modal features with kinds of models, which are unfriendly to deployment. In this paper, we present a unified and dynamic graph (UniDG) framework for temporal character grouping. This is accomplished firstly by a unified representation network that learns representations of multiple modalities within the same space and still preserves the modality's uniqueness simultaneously. Secondly, we present a dynamic graph clustering where the neighbors of different quantities are dynamically constructed for each node via a cyclic matching strategy, leading to a more reliable affinity graph. Thirdly, a progressive association method is introduced to exploit spatial and temporal contexts among different modalities, allowing multi-modal clustering results to be well fused. As current datasets only provide pre-extracted features, we evaluate our UniDG method on a collected dataset named MTCG, which contains each character's appearing clips of face and body and speaking voice tracks. We also evaluate our key components on existing clustering and retrieval datasets to verify the generalization ability. Experimental results manifest that our method can achieve promising results and outperform several state-of-the-art approaches

    Inequalities involving two simplexes

    No full text

    Metabolomic Profile and Antibacterial Bioactivity of Akebia trifoliata (Thunb.) Koidz Pericarp Extract

    No full text
    Akebia trifoliata (A. trifoliata) is a significant medicinal and edible fruit crop and has some important bioactivities. However, there are few studies on the bacteriostatic activity of A. trifoliata, and the underlying mechanism of A. trifoliata for antibacterial activity is still unknown. Therefore, the bacteriostatic activity and antibacterial mechanism of A. trifoliata were investigated by a combination of chemical assays, using the UHPLC-TOF-MS/MS technique. The results indicated that alkaloids, triterpenoids, and flavonoids are the major secondary bioactive compounds in A. trifoliata that play a crucial role in antibacterial activity. We found that EEPA exhibited both bacteriostatic and bactericidal effects against all Gram-positive and Gram-negative bacteria tested, with IZDs ranging from 13.80 ± 0.79 to 17.00 ± 0.58 mm. Significant differences in terms of sensitivity between Gram-positive and Gram-negative bacteria were not observed. In contrast, both antibiotics (kanamycin sulfate and ampicillin sodium salt) exhibited much better antimicrobial activity against Gram-positive bacteria than Gram-negative bacteria. In addition, the primary antimicrobial mechanism was that EEPA increased cellular content leakage, altered the cell morphology, and destroyed the internal cell structure. Meanwhile, MA, UA, and OA, as the common triterpenoid components existing in plants, were used to analyze the relationships between the structures and the antimicrobial activities among homologous compounds, to determine the key functional group that plays an antibacterial role in MA, UA, and OA. As result, it was found that both the hydroxide and methyl groups present are important for their antibacterial activity. These findings suggested that EEPA exerted significant antimicrobial activity against S. aureus, E. coli, B. subtilis, and P. aeruginosa and might be a potential natural antibacterial
    corecore