154,143 research outputs found
Recommended from our members
Language acquisition and machine learning
In this paper, we review recent progress in the field of machine learning and examine its implications for computational models of language acquisition. As a framework for understanding this research, we propose four component tasks involved in learning from experience - aggregation, clustering, characterization, and storage. We then consider four common problems studied by machine learning researchers - learning from examples, heuristics learning, conceptual clustering, and learning macro-operators - describing each in terms of our framework. After this, we turn to the problem of grammar acquisition, relating this problem to other learning tasks and reviewing four AI systems that have addressed the problem. Finally, we note some limitations of the earlier work and propose an alternative approach to modeling the mechanisms underlying language acquisition
Reinforcement learning for efficient network penetration testing
Penetration testing (also known as pentesting or PT) is a common practice for actively assessing the defenses of a computer network by planning and executing all possible attacks to discover and exploit existing vulnerabilities. Current penetration testing methods are increasingly becoming non-standard, composite and resource-consuming despite the use of evolving tools. In this paper, we propose and evaluate an AI-based pentesting system which makes use of machine learning techniques, namely reinforcement learning (RL) to learn and reproduce average and complex pentesting activities. The proposed system is named Intelligent Automated Penetration Testing System (IAPTS) consisting of a module that integrates with industrial PT frameworks to enable them to capture information, learn from experience, and reproduce tests in future similar testing cases. IAPTS aims to save human resources while producing much-enhanced results in terms of time consumption, reliability and frequency of testing. IAPTS takes the approach of modeling PT environments and tasks as a partially observed Markov decision process (POMDP) problem which is solved by POMDP-solver. Although the scope of this paper is limited to network infrastructures PT planning and not the entire practice, the obtained results support the hypothesis that RL can enhance PT beyond the capabilities of any human PT expert in terms of time consumed, covered attacking vectors, accuracy and reliability of the outputs. In addition, this work tackles the complex problem of expertise capturing and re-use by allowing the IAPTS learning module to store and re-use PT policies in the same way that a human PT expert would learn but in a more efficient way
Intermittent Connectivity for Exploration in Communication-Constrained Multi-Agent Systems
Motivated by exploration of communication-constrained underground environments using robot teams, we study the problem of planning for intermittent connectivity in multi-agent systems. We propose a novel concept of information-consistency to handle situations where the plan is not initially known by all agents, and suggest an integer linear program for synthesizing information-consistent plans that also achieve auxiliary goals. Furthermore, inspired by network flow problems we propose a novel way to pose connectivity constraints that scales much better than previous methods. In the second part of the paper we apply these results in an exploration setting, and propose a clustering method that separates a large exploration problem into smaller problems that can be solved independently. We demonstrate how the resulting exploration algorithm is able to coordinate a team of ten agents to explore a large environment
Optimizing Taxi Carpool Policies via Reinforcement Learning and Spatio-Temporal Mining
In this paper, we develop a reinforcement learning (RL) based system to learn
an effective policy for carpooling that maximizes transportation efficiency so
that fewer cars are required to fulfill the given amount of trip demand. For
this purpose, first, we develop a deep neural network model, called ST-NN
(Spatio-Temporal Neural Network), to predict taxi trip time from the raw GPS
trip data. Secondly, we develop a carpooling simulation environment for RL
training, with the output of ST-NN and using the NYC taxi trip dataset. In
order to maximize transportation efficiency and minimize traffic congestion, we
choose the effective distance covered by the driver on a carpool trip as the
reward. Therefore, the more effective distance a driver achieves over a trip
(i.e. to satisfy more trip demand) the higher the efficiency and the less will
be the traffic congestion. We compared the performance of RL learned policy to
a fixed policy (which always accepts carpool) as a baseline and obtained
promising results that are interpretable and demonstrate the advantage of our
RL approach. We also compare the performance of ST-NN to that of
state-of-the-art travel time estimation methods and observe that ST-NN
significantly improves the prediction performance and is more robust to
outliers.Comment: Accepted at IEEE International Conference on Big Data 2018. arXiv
admin note: text overlap with arXiv:1710.0435
- β¦