4 research outputs found
Segmented Recurrent Transformer: An Efficient Sequence-to-Sequence Model
Transformers have shown dominant performance across a range of domains
including language and vision. However, their computational cost grows
quadratically with the sequence length, making their usage prohibitive for
resource-constrained applications. To counter this, our approach is to divide
the whole sequence into segments and apply attention to the individual
segments. We propose a segmented recurrent transformer (SRformer) that combines
segmented (local) attention with recurrent attention. The loss caused by
reducing the attention window length is compensated by aggregating information
across segments with recurrent attention. SRformer leverages Recurrent
Accumulate-and-Fire (RAF) neurons' inherent memory to update the cumulative
product of keys and values. The segmented attention and lightweight RAF neurons
ensure the efficiency of the proposed transformer. Such an approach leads to
models with sequential processing capability at a lower computation/memory
cost. We apply the proposed method to T5 and BART transformers. The modified
models are tested on summarization datasets including CNN-dailymail, XSUM,
ArXiv, and MediaSUM. Notably, using segmented inputs of varied sizes, the
proposed model achieves higher ROUGE1 scores than a segmented
transformer and outperforms other recurrent transformer approaches.
Furthermore, compared to full attention, the proposed model reduces the
computational complexity of cross attention by around .Comment: EMNLP 2023 Finding
Lost 'n' found : a journey and beyond.
Lost ‘n’ Found is a 3-d animation about how a poor but optimistic guy, Ronald, while on his way to look for a home, got lost in a surrealistic space full of road signs. The humour of the story lies in the paradox of how these road signs confuse the protagonist instead of giving him the correct direction. The climax and twist of the story lies in the end, when he decided not to continue his endless journey, but instead decided to convert all his obstacles into hope- his new home.Bachelor of Fine Art