158 research outputs found
Asymptotically Good Additive Cyclic Codes Exist
Long quasi-cyclic codes of any fixed index have been shown to be
asymptotically good, depending on Artin primitive root conjecture in (A.
Alahmadi, C. G\"uneri, H. Shoaib, P. Sol\'e, 2017). We use this recent result
to construct good long additive cyclic codes on any extension of fixed degree
of the base field. Similarly self-dual double circulant codes, and self-dual
four circulant codes, have been shown to be good, also depending on Artin
primitive root conjecture in (A. Alahmadi, F. \"Ozdemir, P. Sol\'e, 2017) and (
M. Shi, H. Zhu, P. Sol\'e, 2017) respectively. Building on these recent
results, we can show that long cyclic codes are good over \F_q, for many
classes of 's. This is a partial solution to a fifty year old open problem
Stable and verifiable state estimation methods and systems with spacecraft applications
The stability of a recursive estimator process (e.g., a Kalman filter is assured for long time periods by periodically resetting an error covariance P(t.sub.n) of the system to a predetermined reset value P.sub.r. The recursive process is thus repetitively forced to start from a selected covariance and continue for a time period that is short compared to the system's total operational time period. The time period in which the process must maintain its numerical stability is significantly reduced as is the demand on the system's numerical stability. The process stability for an extended operational time period T.sub.o is verified by performing the resetting step at the end of at least one reset time period T.sub.r whose duration is less than the operational time period T.sub.o and then confirming stability of the process over the reset time period T.sub.r. Because the recursive process starts from a selected covariance at the beginning of each reset time period T.sub.r, confirming stability of the process over at least one reset time period substantially confirms stability over the longer operational time period T.sub.o
On -ary shortened--perfect-like codes
We study codes with parameters of -ary shortened Hamming codes, i.e.,
. At first, we prove the fact mentioned in
[A.E.Brouwer et al. Bounds on mixed binary/ternary codes. IEEE Trans. Inf.
Theory 44 (1998) 140-161] that such codes are optimal, generalizing it to a
bound for multifold packings of radius- balls, with a corollary for multiple
coverings. In particular, we show that the punctured Hamming code is an optimal
-fold packing with minimum distance . At second, we show the existence of
-ary codes with parameters of shortened -perfect codes that cannot be
obtained by shortening a -perfect code. Keywords: Hamming graph; multifold
packings; multiple coverings; perfect codes
System for star catalog equalization to enhance attitude determination
An apparatus for star catalog equalization to enhance attitude determination includes a star tracker, a star catalog and a controller. The star tracker is used to sense the positions of stars and generate signals corresponding to the positions of the stars as seen in its field of view. The star catalog contains star location data that is stored using a primary and multiple secondary arrays sorted by both declination (DEC) and right ascension (RA), respectively. The star location data stored in the star catalog is predetermined by calculating a plurality of desired star locations, associating one of a plurality of stars with each of the plurality of desired star locations based upon a neighborhood association angle to generate an associated plurality of star locations: If an artificial star gap occurs during association, then the neighborhood association angle for reassociation is increased. The controller uses the star catalog to determine which stars to select to provide star measurement residuals for correcting gyroscope bias and spacecraft attitude
Spacecraft attitude control systems with dynamic methods and structures for processing star tracker signals
Methods are provided for dynamically processing successively-generated star tracker data frames and associated valid flags to generate processed star tracker signals that have reduced noise and a probability greater than a selected probability P.sub.slctd of being valid. These methods maintain accurate spacecraft attitude control in the presence of spurious inputs (e.g., impinging protons) that corrupt collected charges in spacecraft star trackers. The methods of the invention enhance the probability of generating valid star tracker signals because they respond to a current frame probability P.sub.frm by dynamically selecting the largest valid frame combination whose combination probability P.sub.cmb satisfies a selected probability P.sub.slctd. Noise is thus reduced while the probability of finding a valid frame combination is enhanced. Spacecraft structures are also provided for practicing the methods of the invention
System and method for calibrating inter-star-tracker misalignments in a stellar inertial attitude determination system
A method and apparatus for determining star tracker misalignments is disclosed. The method comprises the steps of defining a defining a reference frame for the star tracker assembly according to a boresight of the primary star tracker and a boresight of a second star tracker wherein the boresight of the primary star tracker and a plane spanned by the boresight of the primary star tracker and the boresight of the second star tracker at least partially define a datum for the reference frame for the star tracker assembly; and determining the misalignment of the at least one star tracker as a rotation of the defined reference frame
Neighborhood-based Hard Negative Mining for Sequential Recommendation
Negative sampling plays a crucial role in training successful sequential
recommendation models. Instead of merely employing random negative sample
selection, numerous strategies have been proposed to mine informative negative
samples to enhance training and performance. However, few of these approaches
utilize structural information. In this work, we observe that as training
progresses, the distributions of node-pair similarities in different groups
with varying degrees of neighborhood overlap change significantly, suggesting
that item pairs in distinct groups may possess different negative
relationships. Motivated by this observation, we propose a Graph-based Negative
sampling approach based on Neighborhood Overlap (GNNO) to exploit structural
information hidden in user behaviors for negative mining. GNNO first constructs
a global weighted item transition graph using training sequences. Subsequently,
it mines hard negative samples based on the degree of overlap with the target
item on the graph. Furthermore, GNNO employs curriculum learning to control the
hardness of negative samples, progressing from easy to difficult. Extensive
experiments on three Amazon benchmarks demonstrate GNNO's effectiveness in
consistently enhancing the performance of various state-of-the-art models and
surpassing existing negative sampling strategies. The code will be released at
\url{https://github.com/floatSDSDS/GNNO}
Examining the Effect of Pre-training on Time Series Classification
Although the pre-training followed by fine-tuning paradigm is used
extensively in many fields, there is still some controversy surrounding the
impact of pre-training on the fine-tuning process. Currently, experimental
findings based on text and image data lack consensus. To delve deeper into the
unsupervised pre-training followed by fine-tuning paradigm, we have extended
previous research to a new modality: time series. In this study, we conducted a
thorough examination of 150 classification datasets derived from the Univariate
Time Series (UTS) and Multivariate Time Series (MTS) benchmarks. Our analysis
reveals several key conclusions. (i) Pre-training can only help improve the
optimization process for models that fit the data poorly, rather than those
that fit the data well. (ii) Pre-training does not exhibit the effect of
regularization when given sufficient training time. (iii) Pre-training can only
speed up convergence if the model has sufficient ability to fit the data. (iv)
Adding more pre-training data does not improve generalization, but it can
strengthen the advantage of pre-training on the original data volume, such as
faster convergence. (v) While both the pre-training task and the model
structure determine the effectiveness of the paradigm on a given dataset, the
model structure plays a more significant role
- …