57,690 research outputs found
Learning-by-Doing, Organizational Forgetting, and Industry Dynamics
Learning-by-doing and organizational forgetting have been shown to be important in a variety of industrial settings. This paper provides a general model of dynamic competition that accounts for these economic fundamentals and shows how they shape industry structure and dynamics. Previously obtained results regarding the dominance properties of firms' pricing behavior no longer hold in this more general setting. We show that forgetting does not simply negate learning. Rather, learning and forgetting are distinct economic forces. In particular, a model with learning and forgetting can give rise to aggressive pricing behavior, market dominance, and multiple equilibria, whereas a model with learning alone cannot.
Learning-by-Doing, Organizational Forgetting, and Industry Dynamics
Learning-by-doing and organizational forgetting are empirically important in a variety of industrial settings. This paper provides a general model of dynamic competition that accounts for these fundamentals and shows how they shape industry structure and dynamics. We show that forgetting does not simply negate learning. Rather, they are distinct economic forces that interact in subtle ways to produce a great variety of pricing behaviors and industry dynamics. In particular, a model with learning and forgetting can give rise to aggressive pricing behavior, varying degrees of long-run industry concentration ranging from moderate leadership to absolute dominance, and multiple equilibria
The Importance of Forgetting: Limiting Memory Improves Recovery of Topological Characteristics from Neural Data
We develop of a line of work initiated by Curto and Itskov towards
understanding the amount of information contained in the spike trains of
hippocampal place cells via topology considerations. Previously, it was
established that simply knowing which groups of place cells fire together in an
animal's hippocampus is sufficient to extract the global topology of the
animal's physical environment. We model a system where collections of place
cells group and ungroup according to short-term plasticity rules. In
particular, we obtain the surprising result that in experiments with spurious
firing, the accuracy of the extracted topological information decreases with
the persistence (beyond a certain regime) of the cell groups. This suggests
that synaptic transience, or forgetting, is a mechanism by which the brain
counteracts the effects of spurious place cell activity
On the automated interpretation and indexing of American football
This work combines natural language understanding and image processing with incremental learning to develop a system that can automatically interpret and index American Football. We have developed a model for representing spatio-temporal characteristics of multiple objects in dynamic scenes in this domain. Our representation combines expert knowledge, domain knowledge, spatial knowledge and temporal knowledge. We also present an incremental learning algorithm to improve the knowledge base as well as to keep previously developed concepts consistent with new data. The advantages of the incremental learning algorithm are that is that it does not split concepts and it generates a compact conceptual hierarchy which does not store instances
DeepCare: A Deep Dynamic Memory Model for Predictive Medicine
Personalized predictive medicine necessitates the modeling of patient illness
and care processes, which inherently have long-term temporal dependencies.
Healthcare observations, recorded in electronic medical records, are episodic
and irregular in time. We introduce DeepCare, an end-to-end deep dynamic neural
network that reads medical records, stores previous illness history, infers
current illness states and predicts future medical outcomes. At the data level,
DeepCare represents care episodes as vectors in space, models patient health
state trajectories through explicit memory of historical records. Built on Long
Short-Term Memory (LSTM), DeepCare introduces time parameterizations to handle
irregular timed events by moderating the forgetting and consolidation of memory
cells. DeepCare also incorporates medical interventions that change the course
of illness and shape future medical risk. Moving up to the health state level,
historical and present health states are then aggregated through multiscale
temporal pooling, before passing through a neural network that estimates future
outcomes. We demonstrate the efficacy of DeepCare for disease progression
modeling, intervention recommendation, and future risk prediction. On two
important cohorts with heavy social and economic burden -- diabetes and mental
health -- the results show improved modeling and risk prediction accuracy.Comment: Accepted at JBI under the new name: "Predicting healthcare
trajectories from medical records: A deep learning approach
- …