757,441 research outputs found
The application of artificial intelligence techniques to large distributed networks
Data accessibility and transfer of information, including the land resources information system pilot, are structured as large computer information networks. These pilot efforts include the reduction of the difficulty to find and use data, reducing processing costs, and minimize incompatibility between data sources. Artificial Intelligence (AI) techniques were suggested to achieve these goals. The applicability of certain AI techniques are explored in the context of distributed problem solving systems and the pilot land data system (PLDS). The topics discussed include: PLDS and its data processing requirements, expert systems and PLDS, distributed problem solving systems, AI problem solving paradigms, query processing, and distributed data bases
Distributed human computation framework for linked data co-reference resolution
Distributed Human Computation (DHC) is a technique used to solve computational problems by incorporating the collaborative effort of a large number of humans. It is also a solution to AI-complete problems such as natural language processing. The Semantic Web with its root in AI is envisioned to be a decentralised world-wide information space for sharing machine-readable data with minimal integration costs. There are many research problems in the Semantic Web that are considered as AI-complete problems. An example is co-reference resolution, which involves determining whether different URIs refer to the same entity. This is considered to be a significant hurdle to overcome in the realisation of large-scale Semantic Web applications. In this paper, we propose a framework for building a DHC system on top of the Linked Data Cloud to solve various computational problems. To demonstrate the concept, we are focusing on handling the co-reference resolution in the Semantic Web when integrating distributed datasets. The traditional way to solve this problem is to design machine-learning algorithms. However, they are often computationally expensive, error-prone and do not scale. We designed a DHC system named iamResearcher, which solves the scientific publication author identity co-reference problem when integrating distributed bibliographic datasets. In our system, we aggregated 6 million bibliographic data from various publication repositories. Users can sign up to the system to audit and align their own publications, thus solving the co-reference problem in a distributed manner. The aggregated results are published to the Linked Data Cloud
Organisational Intelligence and Distributed AI
The analysis of this chapter starts from organisational theory, and from this it draws conclusions for the design, and possible organisational applications, of Distributed AI systems. We first review how the concept of organisations has emerged from non-organised "blackbox" entities to so-called "computerised" organisations. Within this context organisational researchers have started to redesign their models of intelligent organisations with respect to the availability of advanced computing technology. The recently emerged concept of Organisational Intelligence integrates these efforts in that it suggests five components of intelligent organisational skills (communication, memory, learning, cognition, problem solving). The approach integrates human and computer-based information processing and problem solving capabilities.<br/
Interplay between Distributed AI Workflow and URLLC
Distributed artificial intelligence (AI) has recently accomplished tremendous
breakthroughs in various communication services, ranging from fault-tolerant
factory automation to smart cities. When distributed learning is run over a set
of wireless connected devices, random channel fluctuations, and the incumbent
services simultaneously running on the same network affect the performance of
distributed learning. In this paper, we investigate the interplay between
distributed AI workflow and ultra-reliable low latency communication (URLLC)
services running concurrently over a network. Using 3GPP compliant simulations
in a factory automation use case, we show the impact of various distributed AI
settings (e.g., model size and the number of participating devices) on the
convergence time of distributed AI and the application layer performance of
URLLC. Unless we leverage the existing 5G-NR quality of service handling
mechanisms to separate the traffic from the two services, our simulation
results show that the impact of distributed AI on the availability of the URLLC
devices is significant. Moreover, with proper setting of distributed AI (e.g.,
proper user selection), we can substantially reduce network resource
utilization, leading to lower latency for distributed AI and higher
availability for the URLLC users. Our results provide important insights for
future 6G and AI standardization.Comment: Accepted in 2022 IEEE Global Communications Conference (GLOBECOM
The AI Bus architecture for distributed knowledge-based systems
The AI Bus architecture is layered, distributed object oriented framework developed to support the requirements of advanced technology programs for an order of magnitude improvement in software costs. The consequent need for highly autonomous computer systems, adaptable to new technology advances over a long lifespan, led to the design of an open architecture and toolbox for building large scale, robust, production quality systems. The AI Bus accommodates a mix of knowledge based and conventional components, running on heterogeneous, distributed real world and testbed environment. The concepts and design is described of the AI Bus architecture and its current implementation status as a Unix C++ library or reusable objects. Each high level semiautonomous agent process consists of a number of knowledge sources together with interagent communication mechanisms based on shared blackboards and message passing acquaintances. Standard interfaces and protocols are followed for combining and validating subsystems. Dynamic probes or demons provide an event driven means for providing active objects with shared access to resources, and each other, while not violating their security
Distributed Governance of Medical AI
Artificial intelligence (AI) promises to bring substantial benefits to medicine. In addition to pushing the frontiers of what is humanly possible, like predicting kidney failure or sepsis before any human can notice, it can democratize expertise beyond the circle of highly specialized practitioners, like letting generalists diagnose diabetic degeneration of the retina. But AI doesn’t always work, and it doesn’t always work for everyone, and it doesn’t always work in every context. AI is likely to behave differently in well-resourced hospitals where it is developed than in poorly resourced frontline health environments where it might well make the biggest difference for patient care. To make the situation even more complicated, AI is unlikely to go through the centralized review and validation process that other medical technologies undergo, like drugs and most medical devices. Even if it did go through those centralized processes, ensuring high-quality performance across a wide variety of settings, including poorly resourced settings, is especially challenging for such centralized mechanisms. What are policymakers to do? This short Essay argues that the diffusion of medical AI, with its many potential benefits, will require policy support for a process of distributed governance, where quality evaluation and oversight take place in the settings of application—but with policy assistance in developing capacities and making that oversight more straightforward to undertake. Getting governance right will not be easy (it never is), but ignoring the issue is likely to leave benefits on the table and patients at risk
- …