676 research outputs found
Towards a Scalable Dynamic Spatial Database System
With the rise of GPS-enabled smartphones and other similar mobile devices,
massive amounts of location data are available. However, no scalable solutions
for soft real-time spatial queries on large sets of moving objects have yet
emerged. In this paper we explore and measure the limits of actual algorithms
and implementations regarding different application scenarios. And finally we
propose a novel distributed architecture to solve the scalability issues.Comment: (2012
Swarm Based Implementation of a Virtual Distributed Database System in a Sensor Network
The deployment of unmanned aerial vehicles (UAVs) in recent military operations has had success in carrying out surveillance and combat missions in sensitive areas. An area of intense research on UAVs has been on controlling a group of small-sized UAVs to carry out reconnaissance missions normally undertaken by large UAVs such as Predator or Global Hawk. A control strategy for coordinating the UAV movements of such a group of UAVs adopts the bio-inspired swarm model to produce autonomous group behavior. This research proposes establishing a distributed database system on a group of swarming UAVs, providing for data storage during a reconnaissance mission. A distributed database system model is simulated treating each UAV as a distributed database site connected by a wireless network. In this model, each UAV carries a sensor and communicates to a command center when queried. Drawing equivalence to a sensor network, the network of UAVs poses as a dynamic ad-hoc sensor network. The distributed database system based on a swarm of UAVs is tested against a set of reconnaissance test suites with respect to evaluating system performance. The design of experiments focuses on the effects of varying the query input and types of swarming UAVs on overall system performance. The results show that the topology of the UAVs has a distinct impact on the output of the sensor database. The experiments measuring system delays also confirm the expectation that in a distributed system, inter-node communication costs outweigh processing costs
Implementing an Information Retrieval and Visualization Framework for Heterogeneous Data Types
In today\u27s information focused world, there is no lack of entities focused on information gathering. However, there is still a widespread epidemic of information starvation in the Department of Defense (DoD). This starvation is attributed to the lack of interoperability between information gatherers and information consumers. To alleviate this problem, the DoD has put forth a vision of a Joint Battlespace Infosphere (JBI). This research proposes a framework for sharing and finding resources in a JBI. The framework uses an extensible metadata specification, agent technology, and the Control of Agent Based Systems (CoABS). It provides several tools for publication and subscription of resources, including a visual query wizard and a visualization of the results. This framework and tools provide visual query capability for the heterogeneous resources within the JBI
Naval Integration into Joint Data Strategies and Architectures in JADC2
NPS NRP Technical ReportAs Joint capabilities mature and shape into the Joint All Domain C2 Concept, Services, COCOMs and Coalition Partners will need to invest into efforts that would seamlessly integrate into Joint capabilities. The objective for the Navy is to study the options for Navy, including Naval Special Warfare Command under SOCOM, on how to integrate Navy's data strategy and architecture under the unifying JADC2 umbrella. The other objectives are to explore alternatives considered by the SOCOM and the Air Force, which are responsible for JADC2 Information Advantage and Digital Mission Command & Control. A major purpose of Joint, Services/COCOMs, agencies and Coalition Partners capabilities is to provide shared core of integrated canonical services for data, information, and knowledge with representations for vertical interoperability across all command levels and JADC2, lateral interoperability between Naval Service/COCOMs, and any combination of JADC2 constituents, agencies, and coalition partners. Our research plan is to explore available data strategy options by leveraging previous NRP work (NPS-20-N313-A). We will participate in emerging data strategy by Navy JADC2 project Overmatch. By working with MITRE our team will explore Air Force JADC2 data strategy implemented in ABMS DataOne component. Our goal is to find a seamless integration between Naval Data Strategy and data strategies behind JADC2 Information Advantage and Digital Mission Command & Control capabilities. Our plan includes studying Service-to-Service and Service-to-COCOM interoperability options required for Joint operations with a goal to minimize OODA's loop latency across sensing, situation discovery & monitoring, and knowledge understanding-for-planning, deciding, and acting. Our team realizes JADC2 requires virtual model allowing interoperability between subordinate C2 for services, agencies, and partner. Without such flexible 'joint' intersection organizational principal hierarchical structure it would be impossible to define necessary temporal and spatial fidelities for each level of organizational command required for implanting JADC2. Research deliverables will document the results of the exploration of Joint, COCOM, Agency and Partner Data Strategies approaches as JADC2 interoperability options to the emerging JADC2. We strive for standard JADC2 interface. Keywords: JADC2, ABMS, DataOne, Information Advantage, Digital Mission Command, IntegrationN2/N6 - Information WarfareThis research is supported by funding from the Naval Postgraduate School, Naval Research Program (PE 0605853N/2098). https://nps.edu/nrpChief of Naval Operations (CNO)Approved for public release. Distribution is unlimited.
Semantics-driven Abstractive Document Summarization
The evolution of the Web over the last three decades has led to a deluge of scientific and news articles on the Internet. Harnessing these publications in different fields of study is critical to effective end user information consumption. Similarly, in the domain of healthcare, one of the key challenges with the adoption of Electronic Health Records (EHRs) for clinical practice has been the tremendous amount of clinical notes generated that can be summarized without which clinical decision making and communication will be inefficient and costly. In spite of the rapid advances in information retrieval and deep learning techniques towards abstractive document summarization, the results of these efforts continue to resemble extractive summaries, achieving promising results predominantly on lexical metrics but performing poorly on semantic metrics. Thus, abstractive summarization that is driven by intrinsic and extrinsic semantics of documents is not adequately explored. Resources that can be used for generating semantics-driven abstractive summaries include: • Abstracts of multiple scientific articles published in a given technical field of study to generate an abstractive summary for topically-related abstracts within the field, thus reducing the load of having to read semantically duplicate abstracts on a given topic. • Citation contexts from different authoritative papers citing a reference paper can be used to generate utility-oriented abstractive summary for a scientific article. • Biomedical articles and the named entities characterizing the biomedical articles along with background knowledge bases to generate entity and fact-aware abstractive summaries. • Clinical notes of patients and clinical knowledge bases for abstractive clinical text summarization using knowledge-driven multi-objective optimization. In this dissertation, we develop semantics-driven abstractive models based on intra- document and inter-document semantic analyses along with facts of named entities retrieved from domain-specific knowledge bases to produce summaries. Concretely, we propose a sequence of frameworks leveraging semantics at various granularity (e.g., word, sentence, document, topic, citations, and named entities) levels, by utilizing external resources. The proposed frameworks have been applied to a range of tasks including 1. Abstractive summarization of topic-centric multi-document scientific articles and news articles. 2. Abstractive summarization of scientific articles using crowd-sourced citation contexts. 3. Abstractive summarization of biomedical articles clustered based on entity-relatedness. 4. Abstractive summarization of clinical notes of patients with heart failure and Chest X-Rays recordings. The proposed approaches achieve impressive performance in terms of preserving semantics in abstractive summarization while paraphrasing. For summarization of topic-centric multiple scientific/news articles, we propose a three-stage approach where abstracts of scientific articles or news articles are clustered based on their topical similarity determined from topics generated using Latent Dirichlet Allocation (LDA), followed by extractive phase and abstractive phase. Then, in the next stage, we focus on abstractive summarization of biomedical literature where we leverage named entities in biomedical articles to 1) cluster related articles; and 2) leverage the named entities towards guiding abstractive summarization. Finally, in the last stage, we turn to external resources such as citation contexts pointing to a scientific article to generate a comprehensive and utility-centric abstractive summary of a scientific article, domain-specific knowledge bases to fill gaps in information about entities in a biomedical article to summarize and clinical notes to guide abstractive summarization of clinical text. Thus, the bottom-up progression of exploring semantics towards abstractive summarization in this dissertation starts with (i) Semantic Analysis of Latent Topics; builds on (ii) Internal and External Knowledge-I (gleaned from abstracts and Citation Contexts); and extends it to make it comprehensive using (iii) Internal and External Knowledge-II (Named Entities and Knowledge Bases)
Proceedings of the 2004 ONR Decision-Support Workshop Series: Interoperability
In August of 1998 the Collaborative Agent Design Research Center (CADRC) of the California Polytechnic State University in San Luis Obispo (Cal Poly), approached Dr. Phillip Abraham of the Office of Naval Research (ONR) with the proposal for an annual workshop focusing on emerging concepts in decision-support systems for military applications. The proposal was considered timely by the ONR Logistics Program Office for at least two reasons. First, rapid advances in information systems technology over the past decade had produced distributed collaborative computer-assistance capabilities with profound potential for providing meaningful support to military decision makers. Indeed, some systems based on these new capabilities such as the Integrated Marine Multi-Agent Command and Control System (IMMACCS) and the Integrated Computerized Deployment System (ICODES) had already reached the field-testing and final product stages, respectively.
Second, over the past two decades the US Navy and Marine Corps had been increasingly challenged by missions demanding the rapid deployment of forces into hostile or devastate dterritories with minimum or non-existent indigenous support capabilities. Under these conditions Marine Corps forces had to rely mostly, if not entirely, on sea-based support and sustainment operations. Particularly today, operational strategies such as Operational Maneuver From The Sea (OMFTS) and Sea To Objective Maneuver (STOM) are very much in need of intelligent, near real-time and adaptive decision-support tools to assist military commanders and their staff under conditions of rapid change and overwhelming data loads.
In the light of these developments the Logistics Program Office of ONR considered it timely to provide an annual forum for the interchange of ideas, needs and concepts that would address the decision-support requirements and opportunities in combined Navy and Marine Corps sea-based warfare and humanitarian relief operations. The first ONR Workshop was held April 20-22, 1999 at the Embassy Suites Hotel in San Luis Obispo, California. It focused on advances in technology with particular emphasis on an emerging family of powerful computer-based tools, and concluded that the most able members of this family of tools appear to be computer-based agents that are capable of communicating within a virtual environment of the real world. From 2001 onward the venue of the Workshop moved from the West Coast to Washington, and in 2003 the sponsorship was taken over by ONR’s Littoral Combat/Power Projection (FNC) Program Office (Program Manager: Mr. Barry Blumenthal). Themes and keynote speakers of past Workshops have included:
1999: ‘Collaborative Decision Making Tools’ Vadm Jerry Tuttle (USN Ret.); LtGen Paul Van Riper (USMC Ret.);Radm Leland Kollmorgen (USN Ret.); and, Dr. Gary Klein (KleinAssociates)
2000: ‘The Human-Computer Partnership in Decision-Support’ Dr. Ronald DeMarco (Associate Technical Director, ONR); Radm CharlesMunns; Col Robert Schmidle; and, Col Ray Cole (USMC Ret.)
2001: ‘Continuing the Revolution in Military Affairs’ Mr. Andrew Marshall (Director, Office of Net Assessment, OSD); and,Radm Jay M. Cohen (Chief of Naval Research, ONR)
2002: ‘Transformation ... ’ Vadm Jerry Tuttle (USN Ret.); and, Steve Cooper (CIO, Office ofHomeland Security)
2003: ‘Developing the New Infostructure’ Richard P. Lee (Assistant Deputy Under Secretary, OSD); and, MichaelO’Neil (Boeing)
2004: ‘Interoperability’ MajGen Bradley M. Lott (USMC), Deputy Commanding General, Marine Corps Combat Development Command; Donald Diggs, Director, C2 Policy, OASD (NII
Context Aware Computing for The Internet of Things: A Survey
As we are moving towards the Internet of Things (IoT), the number of sensors
deployed around the world is growing at a rapid pace. Market research has shown
a significant growth of sensor deployments over the past decade and has
predicted a significant increment of the growth rate in the future. These
sensors continuously generate enormous amounts of data. However, in order to
add value to raw sensor data we need to understand it. Collection, modelling,
reasoning, and distribution of context in relation to sensor data plays
critical role in this challenge. Context-aware computing has proven to be
successful in understanding sensor data. In this paper, we survey context
awareness from an IoT perspective. We present the necessary background by
introducing the IoT paradigm and context-aware fundamentals at the beginning.
Then we provide an in-depth analysis of context life cycle. We evaluate a
subset of projects (50) which represent the majority of research and commercial
solutions proposed in the field of context-aware computing conducted over the
last decade (2001-2011) based on our own taxonomy. Finally, based on our
evaluation, we highlight the lessons to be learnt from the past and some
possible directions for future research. The survey addresses a broad range of
techniques, methods, models, functionalities, systems, applications, and
middleware solutions related to context awareness and IoT. Our goal is not only
to analyse, compare and consolidate past research work but also to appreciate
their findings and discuss their applicability towards the IoT.Comment: IEEE Communications Surveys & Tutorials Journal, 201
Applying semantic web concepts to support Net-Centric Warfare using the Tactical Assessment Markup Language (TAML)
The ability to analyze data quickly and transform it into important information is vital for information superiority. However, the amount of available data is increasing and the time to make decisions is decreasing. There is too much data for humans to sift through and filter for decision making, so computer automation is necessary. The current approach to automating data processing is to hard-code programs to parse particular data formats, but this approach is not flexible enough to handle the constantly changing data world. The Extensible Markup Language (XML) offers a partial solution by providing a syntactic standard for data exchange. The Tactical Assessment Markup Language (TAML) is an XML vocabulary for exchanging undersea warfare tactical data that provides a standard syntax for message exchange. However, the meaning or semantics of the data is unknown to the machine processing the data. The Semantic Web is a set of technologies designed to add semantic information to data for machine processing. The technologies consist of several components including a common syntax for data exchange, common semantic representation, and a common ontology language. The Resource Description Framework (RDF) is used to explicitly state the relationships between resources or entities. The Web Ontology Language (OWL) is used to build models that explicitly define the concepts and properties in a domain. Since concept definitions are written in standard languages, a variety of reasoning engines might be used to process any ontology and its corresponding data instances. Reasoning engines can also apply algorithms to the data to infer useful information and present it to decision makers. Thus there is far less need for specialty hard-coded programs or proprietary data-representation schemes to hold semantic information, since the information needed to process data is captured in an OWL ontology, itself stored in XML format for exchange between systems. Building ontologies for specific domains such as undersea warfare allows programs to understand, process, and infer new information from coherent data. Applying Semantic Web technologies to XML languages such as TAML brings the armed forces closer to a knowledge-aware Global Information Grid (GIG).http://archive.org/details/applyingsemantic109452770US Navy (USN) author.Approved for public release; distribution is unlimited.Approved for public release; distribution is unlimited
The Digital Classicist 2013
This edited volume collects together peer-reviewed papers that initially emanated from presentations at Digital Classicist seminars and conference panels. This wide-ranging volume showcases exemplary applications of digital scholarship to the ancient world and critically examines the many challenges and opportunities afforded by such research. The chapters included here demonstrate innovative approaches that drive forward the research interests of both humanists and technologists while showing that rigorous scholarship is as central to digital research as it is to mainstream classical studies. As with the earlier Digital Classicist publications, our aim is not to give a broad overview of the field of digital classics; rather, we present here a snapshot of some of the varied research of our members in order to engage with and contribute to the development of scholarship both in the fields of classical antiquity and Digital Humanities more broadly
- …