4 research outputs found

    Risk assessment of subway station fire by using a Bayesian network-based scenario evolution model

    Get PDF
    Subway station fires frequently result in massive casualties, economic losses and even social panic due to the massive passenger flow, semiconfined space and limited conditions for escape and smoke emissions. The combination of different states of fire hazard factors increases the uncertainty and complexity of the evolution path of subway station fires and causes difficulty in assessing fire risk. Traditional methods cannot describe the development process of subway station fires, and thus, cannot assess fire risk under different fire scenarios. To realise scenario-based fire risk assessment, the elements that correspond to each scenario state during fire development in subway stations are identified in this study to explore the intrinsic driving force of fire evolution. Accordingly, a fire scenario evolution model of subway stations is constructed. Then, a Bayesian network is adopted to construct a scenario evolution probability calculation model for calculating the occurrence probability of each scenario state during subway station fire development and identifying critical scenario elements that promote fire evolution. Xi’an subway station system is used as a case to illustrate the application of Bayesian network-based scenario evolution model, providing a practical management tool for fire safety managers. The method adopted in this study enables managers to predict fire risk in each scenario and understand the evolution path of subway station fire, supporting the establishment of fire response strategies based on “scenario–response” planning.</p

    UV/Vis+ Photochemistry Database : Structure, Content and Applications

    Get PDF
    Acknowledgments This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors. However, the authors are indebted to those colleagues who support us in maintaining the database through the provision of spectral and other photochemical data and information. The National Center for Atmospheric Research is operated by the University Coporation for Atmopsheric Research, under the sponsorship of the National Science Foundation. Disclaimer: The views expressed in this paper are those of the authors and do not necessarily represent the views or policies of the U.S.EPA. Mention of trade names or products does not convey and should not be interpreted as conveying official U.S. EPA approval, endorsement, or recommendation.Peer reviewedPublisher PD

    Secrets of RLHF in Large Language Models Part I: PPO

    Full text link
    Large language models (LLMs) have formulated a blueprint for the advancement of artificial general intelligence. Its primary objective is to function as a human-centric (helpful, honest, and harmless) assistant. Alignment with humans assumes paramount significance, and reinforcement learning with human feedback (RLHF) emerges as the pivotal technological paradigm underpinning this pursuit. Current technical routes usually include \textbf{reward models} to measure human preferences, \textbf{Proximal Policy Optimization} (PPO) to optimize policy model outputs, and \textbf{process supervision} to improve step-by-step reasoning capabilities. However, due to the challenges of reward design, environment interaction, and agent training, coupled with huge trial and error cost of large language models, there is a significant barrier for AI researchers to motivate the development of technical alignment and safe landing of LLMs. The stable training of RLHF has still been a puzzle. In the first report, we dissect the framework of RLHF, re-evaluate the inner workings of PPO, and explore how the parts comprising PPO algorithms impact policy agent training. We identify policy constraints being the key factor for the effective implementation of the PPO algorithm. Therefore, we explore the PPO-max, an advanced version of PPO algorithm, to efficiently improve the training stability of the policy model. Based on our main results, we perform a comprehensive analysis of RLHF abilities compared with SFT models and ChatGPT. The absence of open-source implementations has posed significant challenges to the investigation of LLMs alignment. Therefore, we are eager to release technical reports, reward models and PPO code
    corecore