7,509 research outputs found
Centralized Coded Caching with User Cooperation
In this paper, we consider the coded-caching broadcast network with user
cooperation, where a server connects with multiple users and the users can
cooperate with each other through a cooperation network. We propose a
centralized coded caching scheme based on a new deterministic placement
strategy and a parallel delivery strategy. It is shown that the new scheme
optimally allocate the communication loads on the server and users, obtaining
cooperation gain and parallel gain that greatly reduces the transmission delay.
Furthermore, we show that the number of users who parallelly send information
should decrease when the users' caching size increases. In other words, letting
more users parallelly send information could be harmful. Finally, we derive a
constant multiplicative gap between the lower bound and upper bound on the
transmission delay, which proves that our scheme is order optimal.Comment: 9 pages, submitted to ITW201
Rethinking the Construction of Effective Metrics for Understanding the Mechanisms of Pretrained Language Models
Pretrained language models are expected to effectively map input text to a
set of vectors while preserving the inherent relationships within the text.
Consequently, designing a white-box model to compute metrics that reflect the
presence of specific internal relations in these vectors has become a common
approach for post-hoc interpretability analysis of pretrained language models.
However, achieving interpretability in white-box models and ensuring the rigor
of metric computation becomes challenging when the source model lacks inherent
interpretability. Therefore, in this paper, we discuss striking a balance in
this trade-off and propose a novel line to constructing metrics for
understanding the mechanisms of pretrained language models. We have
specifically designed a family of metrics along this line of investigation, and
the model used to compute these metrics is referred to as the tree topological
probe. We conducted measurements on BERT-large by using these metrics. Based on
the experimental results, we propose a speculation regarding the working
mechanism of BERT-like pretrained language models, as well as a strategy for
enhancing fine-tuning performance by leveraging the topological probe to
improve specific submodules.Comment: Accepted by Findings of EMNLP202
- …