1,166 research outputs found
A sufficient condition for a number to be the order of a nonsingular derivation of a Lie algebra
A study of the set N_p of positive integers which occur as orders of
nonsingular derivations of finite-dimensional non-nilpotent Lie algebras of
characteristic p>0 was initiated by Shalev and continued by the present author.
The main goal of this paper is to show the abundance of elements of N_p. Our
main result shows that any divisor n of q-1, where q is a power of p, such that
, belongs to N_p. This extends its special
case for p=2 which was proved in a previous paper by a different method.Comment: 10 pages. This version has been revised according to a referee's
suggestions. The additions include a discussion of the (lower) density of the
set N_p, and the results of more extensive machine computations. Note that
the title has also changed. To appear in Israel J. Mat
Contextual Object Detection with a Few Relevant Neighbors
A natural way to improve the detection of objects is to consider the
contextual constraints imposed by the detection of additional objects in a
given scene. In this work, we exploit the spatial relations between objects in
order to improve detection capacity, as well as analyze various properties of
the contextual object detection problem. To precisely calculate context-based
probabilities of objects, we developed a model that examines the interactions
between objects in an exact probabilistic setting, in contrast to previous
methods that typically utilize approximations based on pairwise interactions.
Such a scheme is facilitated by the realistic assumption that the existence of
an object in any given location is influenced by only few informative locations
in space. Based on this assumption, we suggest a method for identifying these
relevant locations and integrating them into a mostly exact calculation of
probability based on their raw detector responses. This scheme is shown to
improve detection results and provides unique insights about the process of
contextual inference for object detection. We show that it is generally
difficult to learn that a particular object reduces the probability of another,
and that in cases when the context and detector strongly disagree this learning
becomes virtually impossible for the purposes of improving the results of an
object detector. Finally, we demonstrate improved detection results through use
of our approach as applied to the PASCAL VOC and COCO datasets
Subgraphs and network motifs in geometric networks
Many real-world networks describe systems in which interactions decay with
the distance between nodes. Examples include systems constrained in real space
such as transportation and communication networks, as well as systems
constrained in abstract spaces such as multivariate biological or economic
datasets and models of social networks. These networks often display network
motifs: subgraphs that recur in the network much more often than in randomized
networks. To understand the origin of the network motifs in these networks, it
is important to study the subgraphs and network motifs that arise solely from
geometric constraints. To address this, we analyze geometric network models, in
which nodes are arranged on a lattice and edges are formed with a probability
that decays with the distance between nodes. We present analytical solutions
for the numbers of all 3 and 4-node subgraphs, in both directed and
non-directed geometric networks. We also analyze geometric networks with
arbitrary degree sequences, and models with a field that biases for directed
edges in one direction. Scaling rules for scaling of subgraph numbers with
system size, lattice dimension and interaction range are given. Several
invariant measures are found, such as the ratio of feedback and feed-forward
loops, which do not depend on system size, dimension or connectivity function.
We find that network motifs in many real-world networks, including social
networks and neuronal networks, are not captured solely by these geometric
models. This is in line with recent evidence that biological network motifs
were selected as basic circuit elements with defined information-processing
functions.Comment: 9 pages, 6 figure
Generalization Error in Deep Learning
Deep learning models have lately shown great performance in various fields
such as computer vision, speech recognition, speech translation, and natural
language processing. However, alongside their state-of-the-art performance, it
is still generally unclear what is the source of their generalization ability.
Thus, an important question is what makes deep neural networks able to
generalize well from the training set to new data. In this article, we provide
an overview of the existing theory and bounds for the characterization of the
generalization error of deep neural networks, combining both classical and more
recent theoretical and empirical results
Quantity makes quality: learning with partial views
In many real world applications, the number of examples to learn from is plentiful, but we can only obtain limited information on each individual example. We study the possibilities of efficient, provably correct, large-scale learning in such settings. The main theme we would like to establish is that large amounts of examples can compensate for the lack of full information on each individual example. The type of partial information we consider can be due to inherent noise or from constraints on the type of interaction with the data source. In particular, we describe and analyze algorithms for budgeted learning, in which the learner can only view a few attributes of each training example (Cesa-Bianchi, Shalev-Shwartz, and Shamir 2010a; 2010c), and algorithms for learning kernel-based predictors, when individual examples are corrupted by random noise (Cesa-Bianchi, Shalev-Shwartz, and Shamir 2010b)
Private Incremental Regression
Data is continuously generated by modern data sources, and a recent challenge
in machine learning has been to develop techniques that perform well in an
incremental (streaming) setting. In this paper, we investigate the problem of
private machine learning, where as common in practice, the data is not given at
once, but rather arrives incrementally over time.
We introduce the problems of private incremental ERM and private incremental
regression where the general goal is to always maintain a good empirical risk
minimizer for the history observed under differential privacy. Our first
contribution is a generic transformation of private batch ERM mechanisms into
private incremental ERM mechanisms, based on a simple idea of invoking the
private batch ERM procedure at some regular time intervals. We take this
construction as a baseline for comparison. We then provide two mechanisms for
the private incremental regression problem. Our first mechanism is based on
privately constructing a noisy incremental gradient function, which is then
used in a modified projected gradient procedure at every timestep. This
mechanism has an excess empirical risk of , where is the
dimensionality of the data. While from the results of [Bassily et al. 2014]
this bound is tight in the worst-case, we show that certain geometric
properties of the input and constraint set can be used to derive significantly
better results for certain interesting regression problems.Comment: To appear in PODS 201
Primitive Words, Free Factors and Measure Preservation
Let F_k be the free group on k generators. A word w \in F_k is called
primitive if it belongs to some basis of F_k. We investigate two criteria for
primitivity, and consider more generally, subgroups of F_k which are free
factors.
The first criterion is graph-theoretic and uses Stallings core graphs: given
subgroups of finite rank H \le J \le F_k we present a simple procedure to
determine whether H is a free factor of J. This yields, in particular, a
procedure to determine whether a given element in F_k is primitive.
Again let w \in F_k and consider the word map w:G x G x ... x G \to G (from
the direct product of k copies of G to G), where G is an arbitrary finite
group. We call w measure preserving if given uniform measure on G x G x ... x
G, w induces uniform measure on G (for every finite G). This is the second
criterion we investigate: it is not hard to see that primitivity implies
measure preservation and it was conjectured that the two properties are
equivalent. Our combinatorial approach to primitivity allows us to make
progress on this problem and in particular prove the conjecture for k=2.
It was asked whether the primitive elements of F_k form a closed set in the
profinite topology of free groups. Our results provide a positive answer for
F_2.Comment: This is a unified version of two manuscripts: "On Primitive words I:
A New Algorithm", and "On Primitive Words II: Measure Preservation". 42
pages, 14 figures. Some parts of the paper reorganized towards publication in
the Israel J. of Mat
- …