948 research outputs found
Modern meat: the next generation of meat from cells
Modern Meat is the first textbook on cultivated meat, with contributions from over 100 experts within the cultivated meat community.
The Sections of Modern Meat comprise 5 broad categories of cultivated meat: Context, Impact, Science, Society, and World.
The 19 chapters of Modern Meat, spread across these 5 sections, provide detailed entries on cultivated meat. They extensively tour a range of topics including the impact of cultivated meat on humans and animals, the bioprocess of cultivated meat production, how cultivated meat may become a food option in Space and on Mars, and how cultivated meat may impact the economy, culture, and tradition of Asia
Deep Generative Modelling of Human Behaviour
Human action is naturally intelligible as a time-varying graph of connected joints constrained by locomotor anatomy and physiology. Its prediction allows the anticipation of actions with applications across healthcare, physical rehabilitation and training, robotics, navigation, manufacture, entertainment, and security. In this thesis we investigate deep generative approaches to the problem of understanding human action. We show that the learning of generative qualities of the distribution may render discriminative tasks more robust to distributional shift and real-world variations in data quality. We further build, from the bottom-up, a novel stochastically deep generative modelling model taylored to the problem of human motion and demonstrate many of it’s state-of-the-art properties such as anomaly detection, imputation in the face of incomplete examples, as well as synthesis—and conditional synthesis—of new samples on massive open source human motion datasets compared to multiple baselines derived from the most relevant pieces of literature
Cooperative AI via Decentralized Commitment Devices
Credible commitment devices have been a popular approach for robust
multi-agent coordination. However, existing commitment mechanisms face
limitations like privacy, integrity, and susceptibility to mediator or user
strategic behavior. It is unclear if the cooperative AI techniques we study are
robust to real-world incentives and attack vectors. However, decentralized
commitment devices that utilize cryptography have been deployed in the wild,
and numerous studies have shown their ability to coordinate algorithmic agents
facing adversarial opponents with significant economic incentives, currently in
the order of several million to billions of dollars. In this paper, we use
examples in the decentralization and, in particular, Maximal Extractable Value
(MEV) (arXiv:1904.05234) literature to illustrate the potential security issues
in cooperative AI. We call for expanded research into decentralized commitments
to advance cooperative AI capabilities for secure coordination in open
environments and empirical testing frameworks to evaluate multi-agent
coordination ability given real-world commitment constraints.Comment: NeurIPS 2023- Multi-Agent Security Worksho
Recommended from our members
On the Creation and Use of Forward Models in Robot Motor Control
Advancements in robotics have the potential to aid humans in many realms of exploration as well as daily life: from search and rescue work, to space and deep sea exploration, to in-home assistance to improve the quality of life for those with limited mobility. One of the main milestones that needs to be met for robotics to achieve these ends is a robust ability to manipulate objects and locomote in cluttered and changing environments. A prerequisite to these skills is the ability to understand the current state of the world as well as how actions result in changes to the environment; in short, a robot needs a way to model itself and the world around it. With recent advances in machine learning and access to cheap and fast computation, one of the most promising avenues for creating robust models is to learn a neural network to approximate the dynamics of the system.
Learning a data-driven model that accurately replicates the dynamics of a robot and its environment is an active area of robotics research. This model needs to be accurate, it needs to operate using sensors that are often high dimensional, and it needs to be robust to changes within the system and the surrounding environment. In this thesis, we investigate ways to improve the processof learning data-driven dynamics models as well as ways to reduce the dimensionality of a robot’s state space.
We start by trying to improve the long-term accuracy of neural network based forward models. Learning forward models is more complicated than it appears on the surface. While it is easy to learn a model to predict the change of a system over a short horizon, it is challenging to assure this performance over a long horizon. We investigate the concept of adding temporal information into the loss function of the forward model during training; we demonstrate that this improves the accuracy of a model when it is used to predict over long horizons.
While we are currently working with low dimensional systems, we eventually want to apply our learned models to robots with high dimensional state spaces. To make learning feasible, we need to find ways to learn a lower dimensional representation of the state space (also known as a latent space) to make learning models in the real world computationally feasible. We present a method to improve the usefulness of a learned latent space using a method we call context training: we learn a latent space alongside a forward model to encourage the learned latent space to retain the variables critical to learning the dynamics of the system.
In all of our experiments, we spend significant time in analysis and evaluation. A large portion of literature demonstrating the effectiveness of data-driven forward models in robot control settings often only presents the final controller performance. We were often left curious about what the model was learning independent of the control scenario. We set out to do our own deep dive into exactly what data-driven forward models are predicting. We evaluate all of our models over long horizons. We also look deeper than just the mean and median loss values. We plot the full distribution of loss values over the entire horizon. The literature on data-driven models that do evaluate model prediction accuracy often focuses on the mean and median prediction errors; while these are important metrics, we found that looking at these metrics alone can sometimes obscure subtle but important effects. A high mean loss is often a result of poor performance on only a subset of the test dataset; one model can outperform other models with lower mean error values on a majority of the test set, but it can be skewed to look like the worst performer by having a few highly inaccurate outliers.
We observe that models often have a subset of a test dataset on which they perform best; we seek to limit the use of a model to regions of the test dataset where it has high accuracy by using an ensemble of models. We find that if we train an ensemble of forward models, the accuracy of the models is higher when they all agree on a prediction. Conversely, when the ensemble of models disagrees, the prediction is often poor. We explore this relationship and propose future ways to apply it.
Finally, we look into the application of improved model accuracy and context trained latent spaces. We start by testing the performance of our context training architecture as a method to reduce the state space dimensionality in a model-free reinforcement learning (MFRL) reaching task. We hypothesize that a policy trained with a latent space observation derived using our context trained encoder will outperform a policy trained with a latent space observation derived from a standard autoencoder. Unfortunately, we found no difference in task performance between the policies learned using either method. We end on a bright note by looking at the power of model-based control when we have access to an accurate model. We successfully use model predictive control (MPC) to generate robust locomotion for a simulated snake robot. With access to an accurate model, we are able to generate realistic snake gaits in a variety of environments with very little parameter tuning that are robust to changes in the environment
Reinforcement Learning Curricula as Interpolations between Task Distributions
In the last decade, the increased availability of powerful computing machinery has led to an increasingly widespread application of machine learning methods. Machine learning has been particularly successful when large models, typically neural networks with an ever-increasing number of parameters, can leverage vast data to make predictions.
While reinforcement learning (RL) has been no exception from this development, a distinguishing feature of RL is its well-known exploration-exploitation trade-off, whose optimal solution – while possible to model as a partially observable Markov decision process – evades computation in all but the simplest problems. Consequently, it seems unsurprising that notable demonstrations of reinforcement learning, such as an RL-based Go agent (AlphaGo) by Deepmind beating the professional Go player Lee Sedol, relied both on the availability of massive computing capabilities and specific forms of regularization that facilitate learning. In the case of AlphaGo, this regularization came in the form of self-play, enabling learning by interacting with gradually more proficient opponents.
In this thesis, we develop techniques that, similarly to the concept of self-play of AlphaGo, improve the learning performance of RL agents by training on sequences of increasingly complex tasks. These task sequences are typically called curricula and are known to side-step problems such as slow learning or convergence to poor behavior that may occur when directly learning in complicated tasks. The algorithms we develop in this thesis create curricula by minimizing distances or divergences between probability distributions of learning tasks, generating interpolations between an initial distribution of easy learning tasks and a target task distribution. Apart from improving the learning performance of RL agents in experiments, developing methods that realize curricula as interpolations between task distributions results in a nuanced picture of key aspects of successful reinforcement learning curricula.
In Chapter 1, we start this thesis by introducing required reinforcement learning notation and then motivating curriculum reinforcement learning from the perspective of continuation methods for non-linear optimization. Similar to curricula for reinforcement learning agents, continuation methods have been used in non-linear optimization to solve challenging optimization problems. This similarity provides an intuition about the effect of the curricula we aim to generate and their limits.
In Chapter 2, we transfer the concept of self-paced learning, initially proposed in the supervised learning community, to the problem of RL, showing that an automated curriculum generation for RL agents can be motivated by a regularized RL objective. This regularized RL objective implies generating a curriculum as a sequence of task distributions that trade off the expected agent performance against similarity to a specified distribution of target tasks. This view on curriculum RL contrasts existing approaches, as it motivates curricula via a regularized RL objective instead of generating them from a set of assumptions about an optimal curriculum. In experiments, we show that an approximate implementation of the aforementioned curriculum – that restricts the interpolating task distribution to a Gaussian – results in improved learning performance compared to regular reinforcement learning, matching or surpassing the performance of existing curriculum-based methods.
Subsequently, Chapter 3 builds up on the intuition of curricula as sequences of interpolating task distributions established in Chapter 2. Motivated by using more flexible task distribution representations, we show how parametric assumptions play a crucial role in the empirical success of the previous approach and subsequently uncover key ingredients that enable the generation of meaningful curricula without assuming a parametric model of the task distributions. One major ingredient is an explicit notion of task similarity via a distance function of two Markov Decision Processes. We turn towards optimal transport theory, allowing for flexible particle-based representations of the task distributions while properly considering the newly introduced metric structure of the task space. Combined with other improvements to our first method, such as a more aggressive restriction of the curriculum to tasks that are not too hard for the agent, the resulting approach delivers consistently high learning performance in multiple experiments.
In the final Chapter 4, we apply the refined method of Chapter 3 to a trajectory-tracking task, in which we task an RL agent to follow a three-dimensional reference trajectory with the tip of an inverted pendulum mounted on a Barrett Whole Arm Manipulator. The access to only positional information results in a partially observable system that, paired with its inherent instability, underactuation, and non-trivial kinematic structure, presents a challenge for modern reinforcement learning algorithms, which we tackle via curricula. The technically infinite-dimensional task space of target trajectories allows us to probe the developed curriculum learning method for flaws that have not surfaced in the rather low-dimensional experiments of the previous chapters. Through an improved optimization scheme that better respects the non-Euclidean structure of target trajectories, we reliably generate curricula of trajectories to be tracked, resulting in faster and more robust learning compared to an RL baseline that does not exploit this form of structured learning. The learned policy matches the performance of an optimal control baseline on the real system, demonstrating the potential of curriculum RL to learn state estimation and control for non-linear tracking tasks jointly.
In summary, this thesis introduces a perspective on reinforcement learning curricula as interpolations between task distributions. The methods developed under this perspective enjoy a precise formulation as optimization problems and deliver empirical benefits throughout experiments. Building upon this precise formulation may allow future work to advance the formal understanding of reinforcement learning curricula and, with that, enable the solution of challenging decision-making and control problems with reinforcement learning
Limited Information Shared Control and its Applications to Large Vehicle Manipulators
Diese Dissertation beschäftigt sich mit der kooperativen Regelung einer mobilen Arbeitsmaschine, welche aus einem Nutzfahrzeug und einem oder mehreren hydraulischen Manipulatoren besteht. Solche Maschinen werden für Aufgaben in der Straßenunterhaltungsaufgaben eingesetzt. Die Arbeitsumgebung des Manipulators ist unstrukturiert, was die Bestimmung einer Referenztrajektorie erschwert oder unmöglich macht. Deshalb wird in dieser Arbeit ein Ansatz vorgeschlagen, welcher nur das Fahrzeug automatisiert, während der menschliche Bediener ein Teil des Systems bleibt und den Manipulator steuert. Eine solche Teilautomatisierung des Gesamtsystems führt zu einer speziellen Klasse von Mensch-Maschine-Interaktionen, welche in der Literatur noch nicht untersucht wurde: Eine kooperative Regelung zwischen zwei Teilsystemen, bei der die Automatisierung keine Informationen von dem vom Menschen gesteuerten Teilsystem hat. Deswegen wird in dieser Arbeit ein systematischer Ansatz der kooperativen Regelung mit begrenzter Information vorgestellt, der den menschlichen Bediener unterstützen kann, ohne die Referenzen oder die Systemzustände des Manipulators zu messen. Außerdem wird ein systematisches Entwurfskonzept für die kooperative Regelung mit begrenzter Information vorgestellt. Für diese Entwurfsmethode werden zwei neue Unterklassen der sogenannten Potenzialspiele eingeführt, die eine systematische Berechnung der Parameter der entwickelten kooperativen Regelung ohne manuelle Abstimmung ermöglichen. Schließlich wird das entwickelte Konzept der kooperativen Regelung am Beispiel einer großen mobilen Arbeitsmaschine angewandt, um seine Vorteile zu ermitteln und zu bewerten. Nach der Analyse in Simulationen wird die praktische Anwendbarkeit der Methode in drei Experimenten mit menschlichen Probanden an einem Simulator untersucht. Die Ergebnisse zeigen die Überlegenheit des entwickelten kooperativen Regelungskonzepts gegenüber der manuellen Steuerung und der nicht-kooperativen Steuerung hinsichtlich sowohl der objektiven Performanz als auch der subjektiven Bewertung der Probanden. Somit zeigt diese Dissertation, dass die kooperative Regelung mobiler Arbeitsmaschinen mit den entwickelten theoretischen Konzepten sowohl hilfreich als auch praktisch anwendbar ist
Undergraduate and Graduate Course Descriptions, 2023 Spring
Wright State University undergraduate and graduate course descriptions from Spring 2023
Fictional Practices of Spirituality I: Interactive Media
"Fictional Practices of Spirituality" provides critical insight into the implementation of belief, mysticism, religion, and spirituality into worlds of fiction, be it interactive or non-interactive. This first volume focuses on interactive, virtual worlds - may that be the digital realms of video games and VR applications or the imaginary spaces of life action role-playing and soul-searching practices. It features analyses of spirituality as gameplay facilitator, sacred spaces and architecture in video game geography, religion in video games and spiritual acts and their dramaturgic function in video games, tabletop, or LARP, among other topics. The contributors offer a first-time ever comprehensive overview of play-rites as spiritual incentives and playful spirituality in various medial incarnations
Fundamentals of Business
Fundamentals of Business, fourth edition (2023) is an open education resource intended to serve as a no-cost, faculty-customizable primary text for one-semester undergraduate introductory business courses. It covers the following topics in business: Teamwork; economics; ethics; entrepreneurship; business ownership, management, and leadership; organizational structures and operations management; human resources and motivating employees; managing in labor union contexts; marketing and pricing strategy; hospitality and tourism, accounting and finance, personal finances, and technology in business
Analytics and Intuition in the Process of Selecting Talent
In management, decisions are expected to be based on rational analytics rather than intuition. But intuition, as a human evolutionary achievement, offers wisdom that, despite all the advances in rational analytics and AI, should be used constructively when recruiting and winning personnel. Integrating these inner experiential competencies with rational-analytical procedures leads to smart recruiting decisions
- …