3,251 research outputs found
Text Processing Like Humans Do: Visually Attacking and Shielding NLP Systems
Visual modifications to text are often used to obfuscate offensive comments
in social media (e.g., "!d10t") or as a writing style ("1337" in "leet speak"),
among other scenarios. We consider this as a new type of adversarial attack in
NLP, a setting to which humans are very robust, as our experiments with both
simple and more difficult visual input perturbations demonstrate. We then
investigate the impact of visual adversarial attacks on current NLP systems on
character-, word-, and sentence-level tasks, showing that both neural and
non-neural models are, in contrast to humans, extremely sensitive to such
attacks, suffering performance decreases of up to 82\%. We then explore three
shielding methods---visual character embeddings, adversarial training, and
rule-based recovery---which substantially improve the robustness of the models.
However, the shielding methods still fall behind performances achieved in
non-attack scenarios, which demonstrates the difficulty of dealing with visual
attacks.Comment: Accepted as long paper at NAACL-2019; fixed one ungrammatical
sentenc
Chunks hierarchies and retrieval structures: Comments on Saariluoma and Laine
The empirical results of Saariluoma and Laine (in press) are discussed and their computer simulations are compared with CHREST, a computational model of perception, memory and learning in chess. Mathematical functions such as power functions and logarithmic functions account for Saariluoma and Laine's (in press) correlation heuristic and for CHREST very well. However, these functions fit human data well only with game positions, not with random positions. As CHREST, which learns using spatial proximity, accounts for the human data as well as Saariluoma and Laine's (in press) correlation heuristic, their conclusion that frequency-based heuristics match the data better than proximity-based heuristics is questioned. The idea of flat chunk organisation and its relation to retrieval structures is discussed. In the conclusion, emphasis is given to the need for detailed empirical data, including information about chunk structure and types of errors, for discriminating between various learning algorithms
Recommended from our members
Chunks in expert memory: Evidence for the magical number four⊠or is it two?
This study aims to test the divergent predictions of the chunking theory (Chase & Simon, 1973) and template theory (Gobet & Simon, 1996a; 2000) with respect to the number of chunks held in visual short-term memory and the size of chunks used by experts. We presented game and random chessboards in both a copy and a recall task. In a within-subject design, the stimuli were displayed using two presentation media: (a) physical board and pieces, as in Chase and Simonâs (1973) study; and (b) a computer display, as in Gobet and Simonâs (1998) study. Results show that, in most cases, no more than three chunks were replaced in the recall task, as predicted by template theory. In addition, with game positions in the computer condition, Masters replaced very large chunks (up to 15 pieces), again in line with template theory. Overall, the results suggest that the original chunking theory overestimated short-term memory capacity and underestimated the size of chunks used, in particular with Masters. They also suggest that Cowanâs (2001) proposal that STM holds four chunks may be an overestimate
Expertise effects in memory recall: A reply to Vicente and Wang
This article may not exactly replicate the final version published in the APA journal. It is not the copy of record.In the January 1998 Psychological Review, Vicente and Wang propose a "constraint attunement hypothesis" to explain the large effects of domain expertise upon memory recall observed in a number of task domains. They claim to find serious defects in alternative explanations of these effects which their theory overcomes. Re-examination of the evidence shows that their theory is not novel, but has been anticipated by those they criticize, and that other current published theories of the phenomena do not have the defects Vicente and Wang attribute to them. Vicente and Wang's views reflect underlying differences (a) about emphasis upon performance versus process in psychology, and (b) about how theories and empirical knowledge interact and progress with the development of a science
Recommended from our members
Some shortcomings of long-term working memory
Within the framework of their long-term working memory theory, Ericsson and Kintsch (1995) propose that experts rapidly store information in long-term memory through two mechanisms: elaboration of long-term memory patterns and schemas and use of retrieval structures. They use chess playersâ memory as one of their most compelling sources of empirical evidence. In this paper, I show that evidence from chess memory, far from supporting their theory, limits its generality. Evidence from other domains reviewed by Ericsson and Kintsch, such as medical expertise, is not as strong as claimed, and sometimes contradicts the theory outright. I argue that Ericsson and Kintschâs concept of retrieval structure conflates three different types of memory structures that possess quite different properties. One of these types of structuresâgeneric, general-purpose retrieval structuresâhas a narrower use than proposed by Ericsson and Kintsch: it applies only in domains where there is a conscious, deliberate intent by individuals to improve their memory. Other mechanisms, including specific retrieval structures, exist that permit a rapid encoding into long-term memory under other circumstances
MATREX: DCU machine translation system for IWSLT 2006
In this paper, we give a description of the machine translation system developed at DCU that was used for our first participation in the evaluation campaign of the International Workshop on Spoken Language Translation (2006). This system combines two types of approaches. First, we use an EBMT approach to collect aligned chunks based on two steps: deterministic chunking of both sides and chunk alignment. We use several chunking and alignment strategies. We also extract SMT-style aligned phrases, and the two types of resources are combined.
We participated in the Open Data Track for the following
translation directions: Arabic-English and Italian-English,
for which we translated both the single-best ASR hypotheses
and the text input. We report the results of the system for
the provided evaluation sets
Name Strategy: Its Existence and Implications
It is argued that colour name strategy, object name strategy, and chunking strategy in memory are all aspects of the same general phenomena, called stereotyping, and this in turn is an example of a know-how representation. Such representations are argued to have their origin in a principle called the minimum duplication of resources. For most the subsequent discussions existence of colour name strategy suffices. It is pointed out that the Berlinâ - Kayâ universal partial ordering of colours and the frequency of traffic accidents classified by colour are surprisingly similar; a detailed analysis is not carried out as the specific colours recorded are not identical. Some consequences of the existence of a name strategy for the philosophy of language and mathematics are discussed: specifically it is argued that in accounts of truth and meaning it is necessary throughout to use real numbers as opposed to bi-valent quantities; and also that the concomitant label associated with sentences should not be of unconditional truth, but rather several real-valued quantities associated with visual communication. The implication of real-valued truth quantities is that the Continuum Hypothesis of pure mathematics is side-stepped, because real valued quantities occur ab initio. The existence of name strategy shows that thought/sememes and talk/phonemes can be separate, and this vindicates the assumption of thought occurring before talk used in psycho-linguistic speech production models.
- âŠ