4,082 research outputs found

    Generative AI in the Construction Industry: A State-of-the-art Analysis

    Full text link
    The construction industry is a vital sector of the global economy, but it faces many productivity challenges in various processes, such as design, planning, procurement, inspection, and maintenance. Generative artificial intelligence (AI), which can create novel and realistic data or content, such as text, image, video, or code, based on some input or prior knowledge, offers innovative and disruptive solutions to address these challenges. However, there is a gap in the literature on the current state, opportunities, and challenges of generative AI in the construction industry. This study aims to fill this gap by providing a state-of-the-art analysis of generative AI in construction, with three objectives: (1) to review and categorize the existing and emerging generative AI opportunities and challenges in the construction industry; (2) to propose a framework for construction firms to build customized generative AI solutions using their own data, comprising steps such as data collection, dataset curation, training custom large language model (LLM), model evaluation, and deployment; and (3) to demonstrate the framework via a case study of developing a generative model for querying contract documents. The results show that retrieval augmented generation (RAG) improves the baseline LLM by 5.2, 9.4, and 4.8% in terms of quality, relevance, and reproducibility. This study provides academics and construction professionals with a comprehensive analysis and practical framework to guide the adoption of generative AI techniques to enhance productivity, quality, safety, and sustainability across the construction industry.Comment: 74 pages, 11 figures, 20 table

    Clustering Methods for Network Data Analysis in Programming

    Get PDF
    In the modern world, data volumes are constantly increasing, and clustering has become an essential tool for identifying patterns and regularities in large datasets. The relevance of this study is associated with the growing need for effective data analysis methods in programming. The objective is to evaluate different clustering techniques within the programming domain and explore their suitability for analysing a wide range of datasets. Inductive and deductive methodologies, concrete illustrations, and visual techniques were employed. The clustering techniques were implemented using RStudio and Matlab tools. The study's findings facilitated the identification of crucial attributes of clustering techniques, including hierarchical structure, cluster quantity, and similarity metrics. The application of several data analysis and visualisation approaches, including k-means, c-means, hierarchical, least spanning tree, and linked component extraction, was illustrated. This study elucidated the clustering approaches that may be optimally employed in various contexts, resulting in enhanced precision in analyses and data-informed decision-making. The study's practical significance is in enhancing programmers' methodological toolset with tools for data analysis and processing

    Consistent Density Scanning and Information Extraction From Point Clouds of Building Interiors

    Get PDF
    Over the last decade, 3D range scanning systems have improved considerably enabling the designers to capture large and complex domains such as building interiors. The captured point cloud is processed to extract specific Building Information Models, where the main research challenge is to simultaneously handle huge and cohesive point clouds representing multiple objects, occluded features and vast geometric diversity. These domain characteristics increase the data complexities and thus make it difficult to extract accurate information models from the captured point clouds. The research work presented in this thesis improves the information extraction pipeline with the development of novel algorithms for consistent density scanning and information extraction automation for building interiors. A restricted density-based, scan planning methodology computes the number of scans to cover large linear domains while ensuring desired data density and reducing rigorous post-processing of data sets. The research work further develops effective algorithms to transform the captured data into information models in terms of domain features (layouts), meaningful data clusters (segmented data) and specific shape attributes (occluded boundaries) having better practical utility. Initially, a direct point-based simplification and layout extraction algorithm is presented that can handle the cohesive point clouds by adaptive simplification and an accurate layout extraction approach without generating an intermediate model. Further, three information extraction algorithms are presented that transforms point clouds into meaningful clusters. The novelty of these algorithms lies in the fact that they work directly on point clouds by exploiting their inherent characteristic. First a rapid data clustering algorithm is presented to quickly identify objects in the scanned scene using a robust hue, saturation and value (H S V) color model for better scene understanding. A hierarchical clustering algorithm is developed to handle the vast geometric diversity ranging from planar walls to complex freeform objects. The shape adaptive parameters help to segment planar as well as complex interiors whereas combining color and geometry based segmentation criterion improves clustering reliability and identifies unique clusters from geometrically similar regions. Finally, a progressive scan line based, side-ratio constraint algorithm is presented to identify occluded boundary data points by investigating their spatial discontinuity

    Exploration of Building Information Modeling and Integrated Project Cloud Service in early architectural design stages

    Full text link
    [EN] In the evolving Architecture, Engineering, and Construction (AEC) industry, the use of Building Information Modeling (BIM) and Integrated Project Cloud Service (IPCS) has become crucial. These tools are particularly essential during the early design stages, as they enable comprehensive management and integration of project information, thus promoting effective decision-making throughout project lifecycles. This combined approach enhances inter-organizational collaborations, improves design and construction practices, and creates a communal data platform for stakeholders. This research explores the effectiveness of the BIM-IPCS system in streamlining data exchange and information flow during early design, suggesting ways to minimize errors, speed up processes, and reduce construction costs through dependable networks. Conclusively, this study underscores the significant impact of the BIM-IPCS system on project management, ensuring well-coordinated and informed construction while advocating for its role in driving innovative and efficient project delivery in the AEC industry.Grateful acknowledgment is extended to the National Taiwan University of Science and Technology, the Public Works Information Institute of the Republic of China (CPWEIA), and Luxor Digital Co., Ltd. (LUXOR) for their substantial support and contributions to this research.Wagiri, F.; Shih, S.; Harsono, K.; Cheng, T.; Lu, M. (2023). Exploration of Building Information Modeling and Integrated Project Cloud Service in early architectural design stages. VITRUVIO - International Journal of Architectural Technology and Sustainability. 8(2):26-37. https://doi.org/10.4995/vitruvio-ijats.2023.2045326378

    Faculty of Computer Science

    Get PDF
    Information about the Faculty of Computer Science of the Technische Universität Dresden, data and facts and a selection of current research projects, 2009Informationen über die Fakultät Informatik der TU Dresden, Daten und Fakten sowie eine Auswahl aktueller Forschungsprojekte, 200

    Data-Driven Models, Techniques, and Design Principles for Combatting Healthcare Fraud

    Get PDF
    In the U.S., approximately 700billionofthe700 billion of the 2.7 trillion spent on healthcare is linked to fraud, waste, and abuse. This presents a significant challenge for healthcare payers as they navigate fraudulent activities from dishonest practitioners, sophisticated criminal networks, and even well-intentioned providers who inadvertently submit incorrect billing for legitimate services. This thesis adopts Hevner’s research methodology to guide the creation, assessment, and refinement of a healthcare fraud detection framework and recommended design principles for fraud detection. The thesis provides the following significant contributions to the field:1. A formal literature review of the field of fraud detection in Medicaid. Chapters 3 and 4 provide formal reviews of the available literature on healthcare fraud. Chapter 3 focuses on defining the types of fraud found in healthcare. Chapter 4 reviews fraud detection techniques in literature across healthcare and other industries. Chapter 5 focuses on literature covering fraud detection methodologies utilized explicitly in healthcare.2. A multidimensional data model and analysis techniques for fraud detection in healthcare. Chapter 5 applies Hevner et al. to help develop a framework for fraud detection in Medicaid that provides specific data models and techniques to identify the most prevalent fraud schemes. A multidimensional schema based on Medicaid data and a set of multidimensional models and techniques to detect fraud are presented. These artifacts are evaluated through functional testing against known fraud schemes. This chapter contributes a set of multidimensional data models and analysis techniques that can be used to detect the most prevalent known fraud types.3. A framework for deploying outlier-based fraud detection methods in healthcare. Chapter 6 proposes and evaluates methods for applying outlier detection to healthcare fraud based on literature review, comparative research, direct application on healthcare claims data, and known fraudulent cases. A method for outlier-based fraud detection is presented and evaluated using Medicaid dental claims, providers, and patients.4. Design principles for fraud detection in complex systems. Based on literature and applied research in Medicaid healthcare fraud detection, Chapter 7 offers generalized design principles for fraud detection in similar complex, multi-stakeholder systems.<br/

    Consistency Algorithms and Protocols for Distributed Interactive Applications

    Full text link
    The Internet has a major impact not only on how people retrieve information but also on how they communicate. Distributed interactive applications support the communication and collaboration of people through the sharing and manipulation of rich multimedia content via the Internet. Aside from shared text editors, meeting support systems, and distributed virtual environments, shared whiteboards are a prominent example of distributed interactive applications. They allow the presentation and joint editing of documents in video conferencing scenarios. The design of such a shared whiteboard application, the multimedia lecture board (mlb), is a main contribution of this thesis. Like many other distributed interactive applications, the mlb has a replicated architecture where each user runs an instance of the application. This has the distinct advantage that the application can be deployed in a lightweight fashion, without relying on a supporting server infrastructure. But at the same time, this peer-to-peer architecture raises a number of challenging problems: First, application data needs to be distributed among all instances. For this purpose, we present the network protocol RTP/I for the standardized communication of distributed interactive applications, and a novel application-level multicast protocol that realizes efficient group communication while taking application-level knowledge into account. Second, consistency control mechanisms are required to keep the replicated application data synchronized. We present the consistency control algorithms “local lag”, “Timewarp”, and “state request”, show how they can be combined, and discuss how to provide visual feedback so that the session members are able to handle conflicting actions. Finally, late-joining participants need to be initialized with the current application state before they are able to participate in a collaborative session. We propose a novel late-join algorithm, which is both flexible and scalable. All algorithms and protocols presented in this dissertation solve the aforementioned problems in a generic way. We demonstrate how they can be employed for the mlb as well as for other distributed interactive applications

    Visualizing Large Business Process Models: Challenges, Techniques, Applications

    Get PDF
    Large process models may comprise hundreds or thousands of process elements, like activities, gateways, and data objects. Presenting such process models to users and enabling them to interact with these models constitute crucial tasks of any process-aware information systems (PAISs). Existing PAISs, however, neither provide adequate techniques for visualizing and abstracting process models nor for interacting with them. In particular, PAISs do not provide tailored process visualizations as needed in complex application environments. This paper presents examples of large process models and discusses some of the challenges to be tackled when visualizing and abstracting respective models. Further, it presents a comprehensive framework that allows for personalized process model visualizations, which can be tailored to the specific needs of the different user groups. First, process model complexity can be reduced by abstracting the models, i.e., by eliminating or aggregating process elements not relevant in the given visualization context. Second, the appearance of process elements can be customized independent of the process modeling language used. Third, different visualization formats (e.g., process diagrams, process forms, and process trees) are supported. Finally, it will be discussed how tailored visualizations of process models may serve as basis for changing and evolving process models at a high level of abstraction

    Opportunities and Challenges of Applying Large Language Models in Building Energy Efficiency and Decarbonization Studies: An Exploratory Overview

    Full text link
    In recent years, the rapid advancement and impressive capabilities of Large Language Models (LLMs) have been evident across various domains. This paper explores the application, implications, and potential of LLMs in building energy efficiency and decarbonization studies. The wide-ranging capabilities of LLMs are examined in the context of the building energy field, including intelligent control systems, code generation, data infrastructure, knowledge extraction, and education. Despite the promising potential of LLMs, challenges including complex and expensive computation, data privacy, security and copyright, complexity in fine-tuned LLMs, and self-consistency are discussed. The paper concludes with a call for future research focused on the enhancement of LLMs for domain-specific tasks, multi-modal LLMs, and collaborative research between AI and energy experts
    • …
    corecore