2,069 research outputs found

    Identification and characterization of the human ORC6 homolog

    Get PDF
    A new protein was cloned and identified as the sixth member of the human Origin Recognition Complex (ORC). The newly identified 30-kDa protein hsORC6 is 28% identical and 49% similar to ORC6p from Drosophila melanogaster, which is consistent with the identities and similarities found among the other ORC members reported in the two species. The human ORC6 gene is located on chromosome 16q12. ORC6 protein level did not change through the cell cycle. Like ORC1, ORC6 did not co-immunoprecipitate with other ORC subunits but was localized in the nucleus along with the other ORC subunits. Several cellular proteins co-immunoprecipitated with ORC6, including a 65-kDa protein that was hyperphosphorylated in G1 and dephosphorylated in mitosis. Therefore, unlike the tight stoichiometric association of six yeast ORC subunits in one holo-complex, only a small fraction of human ORC1 and ORC6 is likely to be associated with a subcomplex of ORC2, 3, 4 and 5 suggesting differences in the architecture and regulation of human ORC

    Inherent Weight Normalization in Stochastic Neural Networks

    Get PDF
    Multiplicative stochasticity such as Dropout improves the robustness and generalizability of deep neural networks. Here, we further demonstrate that always-on multiplicative stochasticity combined with simple threshold neurons are sufficient operations for deep neural networks. We call such models Neural Sampling Machines (NSM). We find that the probability of activation of the NSM exhibits a self-normalizing property that mirrors Weight Normalization, a previously studied mechanism that fulfills many of the features of Batch Normalization in an online fashion. The normalization of activities during training speeds up convergence by preventing internal covariate shift caused by changes in the input distribution. The always-on stochasticity of the NSM confers the following advantages: the network is identical in the inference and learning phases, making the NSM suitable for online learning, it can exploit stochasticity inherent to a physical substrate such as analog non-volatile memories for in-memory computing, and it is suitable for Monte Carlo sampling, while requiring almost exclusively addition and comparison operations. We demonstrate NSMs on standard classification benchmarks (MNIST and CIFAR) and event-based classification benchmarks (N-MNIST and DVS Gestures). Our results show that NSMs perform comparably or better than conventional artificial neural networks with the same architecture

    Non Gaussian information of heterogeneity in Soft Matter

    Full text link
    Heterogeneity in dynamics in the form of non-Gaussian molecular displacement distributions appears ubiquitously in soft matter. We address the quantification of such heterogeneity using an information-theoretic measure of the distance between the actual displacement distribution and its nearest Gaussian estimation. We explore the usefulness of this measure in two generic scenarios of random walkers in heterogeneous media. We show that our proposed measure leads to a better quantification of non-Gaussianity than the conventional ones based on moment ratios
    • …
    corecore