11,354 research outputs found

    Thirty Years of Machine Learning: The Road to Pareto-Optimal Wireless Networks

    Full text link
    Future wireless networks have a substantial potential in terms of supporting a broad range of complex compelling applications both in military and civilian fields, where the users are able to enjoy high-rate, low-latency, low-cost and reliable information services. Achieving this ambitious goal requires new radio techniques for adaptive learning and intelligent decision making because of the complex heterogeneous nature of the network structures and wireless services. Machine learning (ML) algorithms have great success in supporting big data analytics, efficient parameter estimation and interactive decision making. Hence, in this article, we review the thirty-year history of ML by elaborating on supervised learning, unsupervised learning, reinforcement learning and deep learning. Furthermore, we investigate their employment in the compelling applications of wireless networks, including heterogeneous networks (HetNets), cognitive radios (CR), Internet of things (IoT), machine to machine networks (M2M), and so on. This article aims for assisting the readers in clarifying the motivation and methodology of the various ML algorithms, so as to invoke them for hitherto unexplored services as well as scenarios of future wireless networks.Comment: 46 pages, 22 fig

    Aquaculture Asia, Vol.13, No.2, pp.1-56, April-June 2008

    Get PDF
    Peter Edwards writes on rural aquaculture: From integrated carp polyculture to intensive monoculture in the Pearl River Delta, South China. Better management practices for Vietnamese catfish. Ipomoea aquatica ā€“ an aquaculture friendly macrophyte. A status overview of fisheries and aquaculture development in Pakistan with context to other Asian countries. The changing face of post-grad education in aquaculture: contributing to soaring production and sustainable practices. Hatchery management in Bangladesh. Production of Cirrhinus molitorella and Labeo chrysophekadion for culture based fisheries development in Lao PDR Part I: Captive spawning. Application of ipil-ipil leaf meal as feed Ingredient for monosex tilapia fry (Oreochromis niloticus) in terms of growth and economics. Fermented feed ingredients as fish meal replacer in aquafeed production Aquaculture and fishing management in coastal zone demarcation: the case of Thailand. Reservoir fisheries of freshwater prawn ā€“ success story of an emerging culture-based giant freshwater prawn fishery at Malampuzha Dam in Kerala, India. Determining and locating sea cage production area for sustainable tropical aquaculture. SPC Pacific-Asia marine fish mariculture technical workshop: ā€œFarming Marine Fishes for our Futureā€. Developing Better Management Practices for Marine Finfish Aquaculture. Breeding and seed production of silver pompano (Trachinotus blochii, Lacepede) at the Mariculture Development Center of Batam. Potential of silver pomfret (Pampus argenteus) as a new candidate species for aquaculture. NACA Newsletter

    Flexibility to contingency changes distinguishes habitual and goal-directed strategies in humans

    Get PDF
    Decision-making in the real world presents the challenge of requiring flexible yet prompt behavior, a balance that has been characterized in terms of a trade-off between a slower, prospective goal-directed model-based (MB) strategy and a fast, retrospective habitual model-free (MF) strategy. Theory predicts that flexibility to changes in both reward values and transition contingencies can determine the relative influence of the two systems in reinforcement learning, but few studies have manipulated the latter. Therefore, we developed a novel two-level contingency change task in which transition contingencies between states change every few trials; MB and MF control predict different responses following these contingency changes, allowing their relative influence to be inferred. Additionally, we manipulated the rate of contingency changes in order to determine whether contingency change volatility would play a role in shifting subjects between a MB and MF strategy. We found that human subjects employed a hybrid MB/MF strategy on the task, corroborating the parallel contribution of MB and MF systems in reinforcement learning. Further, subjects did not remain at one level of MB/MF behaviour but rather displayed a shift towards more MB behavior over the first two blocks that was not attributable to the rate of contingency changes but rather to the extent of training. We demonstrate that flexibility to contingency changes can distinguish MB and MF strategies, with human subjects utilizing a hybrid strategy that shifts towards more MB behavior over blocks, consequently corresponding to a higher payoff

    Reinforcement Learning for Racecar Control

    Get PDF
    This thesis investigates the use of reinforcement learning to learn to drive a racecar in the simulated environment of the Robot Automobile Racing Simulator. Real-life race driving is known to be difficult for humans, and expert human drivers use complex sequences of actions. There are a large number of variables, some of which change stochastically and all of which may affect the outcome. This makes driving a promising domain for testing and developing Machine Learning techniques that have the potential to be robust enough to work in the real world. Therefore the principles of the algorithms from this work may be applicable to a range of problems. The investigation starts by finding a suitable data structure to represent the information learnt. This is tested using supervised learning. Reinforcement learning is added and roughly tuned, and the supervised learning is then removed. A simple tabular representation is found satisfactory, and this avoids difficulties with more complex methods and allows the investigation to concentrate on the essentials of learning. Various reward sources are tested and a combination of three are found to produce the best performance. Exploration of the problem space is investigated. Results show exploration is essential but controlling how much is done is also important. It turns out the learning episodes need to be very long and because of this the task needs to be treated as continuous by using discounting to limit the size of the variables stored. Eligibility traces are used with success to make the learning more efficient. The tabular representation is made more compact by hashing and more accurate by using smaller buckets. This slows the learning but produces better driving. The improvement given by a rough form of generalisation indicates the replacement of the tabular method by a function approximator is warranted. These results show reinforcement learning can work within the Robot Automobile Racing Simulator, and lay the foundations for building a more efficient and competitive agent

    Aerospace medicine and biology: A continuing bibliography with indexes (supplement 323)

    Get PDF
    This bibliography lists 125 reports, articles and other documents introduced into the NASA Scientific and Technical Information System during April, 1989. Subject coverage includes; aerospace medicine and psychology, life support systems and controlled environments, safety equipment exobiology and extraterrestrial life, and flight crew behavior and performance

    Normative Evidence Accumulation in Unpredictable Environments

    Get PDF
    In our dynamic world, decisions about noisy stimuli can require temporal accumulation of evidence to identify steady signals, differentiation to detect unpredictable changes in those signals, or both. Normative models can account for learning in these environments but have not yet been applied to faster decision processes. We present a novel, normative formulation of adaptive learning models that forms decisions by acting as a leaky accumulator with non-absorbing bounds. These dynamics, derived for both discrete and continuous cases, depend on the expected rate of change of the statistics of the evidence and balance signal identification and change detection. We found that, for two different tasks, human subjects learned these expectations, albeit imperfectly, then used them to make decisions in accordance with the normative model. The results represent a unified, empirically supported account of decision-making in unpredictable environments that provides new insights into the expectation-driven dynamics of the underlying neural signals
    • ā€¦
    corecore