15,371 research outputs found

    Towards Safer Generative Language Models: A Survey on Safety Risks, Evaluations, and Improvements

    Full text link
    As generative large model capabilities advance, safety concerns become more pronounced in their outputs. To ensure the sustainable growth of the AI ecosystem, it's imperative to undertake a holistic evaluation and refinement of associated safety risks. This survey presents a framework for safety research pertaining to large models, delineating the landscape of safety risks as well as safety evaluation and improvement methods. We begin by introducing safety issues of wide concern, then delve into safety evaluation methods for large models, encompassing preference-based testing, adversarial attack approaches, issues detection, and other advanced evaluation methods. Additionally, we explore the strategies for enhancing large model safety from training to deployment, highlighting cutting-edge safety approaches for each stage in building large models. Finally, we discuss the core challenges in advancing towards more responsible AI, including the interpretability of safety mechanisms, ongoing safety issues, and robustness against malicious attacks. Through this survey, we aim to provide clear technical guidance for safety researchers and encourage further study on the safety of large models

    Spartan Daily, September 4, 1998

    Get PDF
    Volume 111, Issue 6https://scholarworks.sjsu.edu/spartandaily/9294/thumbnail.jp

    Spartan Daily, April 22, 1997

    Get PDF
    Volume 108, Issue 56https://scholarworks.sjsu.edu/spartandaily/9131/thumbnail.jp

    Spartan Daily, April 22, 1997

    Get PDF
    Volume 108, Issue 56https://scholarworks.sjsu.edu/spartandaily/9131/thumbnail.jp

    Spartan Daily, April 22, 1997

    Get PDF
    Volume 108, Issue 56https://scholarworks.sjsu.edu/spartandaily/9131/thumbnail.jp

    Bringing order into the realm of Transformer-based language models for artificial intelligence and law

    Full text link
    Transformer-based language models (TLMs) have widely been recognized to be a cutting-edge technology for the successful development of deep-learning-based solutions to problems and applications that require natural language processing and understanding. Like for other textual domains, TLMs have indeed pushed the state-of-the-art of AI approaches for many tasks of interest in the legal domain. Despite the first Transformer model being proposed about six years ago, there has been a rapid progress of this technology at an unprecedented rate, whereby BERT and related models represent a major reference, also in the legal domain. This article provides the first systematic overview of TLM-based methods for AI-driven problems and tasks in the legal sphere. A major goal is to highlight research advances in this field so as to understand, on the one hand, how the Transformers have contributed to the success of AI in supporting legal processes, and on the other hand, what are the current limitations and opportunities for further research development.Comment: Please refer to the published version: Greco, C.M., Tagarelli, A. (2023) Bringing order into the realm of Transformer-based language models for artificial intelligence and law. Artif Intell Law, Springer Nature. November 2023. https://doi.org/10.1007/s10506-023-09374-

    Multi-Granularity Archaeological Dating of Chinese Bronze Dings Based on a Knowledge-Guided Relation Graph

    Full text link
    The archaeological dating of bronze dings has played a critical role in the study of ancient Chinese history. Current archaeology depends on trained experts to carry out bronze dating, which is time-consuming and labor-intensive. For such dating, in this study, we propose a learning-based approach to integrate advanced deep learning techniques and archaeological knowledge. To achieve this, we first collect a large-scale image dataset of bronze dings, which contains richer attribute information than other existing fine-grained datasets. Second, we introduce a multihead classifier and a knowledge-guided relation graph to mine the relationship between attributes and the ding era. Third, we conduct comparison experiments with various existing methods, the results of which show that our dating method achieves a state-of-the-art performance. We hope that our data and applied networks will enrich fine-grained classification research relevant to other interdisciplinary areas of expertise. The dataset and source code used are included in our supplementary materials, and will be open after submission owing to the anonymity policy. Source codes and data are available at: https://github.com/zhourixin/bronze-Ding.Comment: CVPR2023 accepte

    Italian Schools and New Linguistic Minorities: Nationality Vs. Plurilingualism. Which Ways and Methodologies for Mapping these Contexts?

    Get PDF
    According to the latest findings of the MIUR (Ministry of Education, University and Research), Alunni con cittadinanza non italiana 2004-2005 (MIUR, October 2005), 4.2% of the school population in Italy is made up of non-Italian citizens, with no reference to students who have one Italian parent or adopted children. These findings show that schools have become multilingual, not so much or solely because of the proposed linguistic offerings, nor for the linguistic heritage of Italian-speakers, which alternates among dialect, regional Italian and standard Italian, but mostly because of the dimension created by the contacts developed between different linguistic and cultural heritages. The paper aims at emphasizing and showing different ways for mapping the role played and the weight exercised by these “new linguistic minorities” – (defined as such) so-called because they are related to immigrant settlements in the territory and, hence, “immigrant languages” – in redefining the linguistic landscape of a school and of a territory.Language Contact, Immigrant Languages, School System, Linguistic-Cultural Identity

    Principle-Driven Self-Alignment of Language Models from Scratch with Minimal Human Supervision

    Full text link
    Recent AI-assistant agents, such as ChatGPT, predominantly rely on supervised fine-tuning (SFT) with human annotations and reinforcement learning from human feedback (RLHF) to align the output of large language models (LLMs) with human intentions, ensuring they are helpful, ethical, and reliable. However, this dependence can significantly constrain the true potential of AI-assistant agents due to the high cost of obtaining human supervision and the related issues on quality, reliability, diversity, self-consistency, and undesirable biases. To address these challenges, we propose a novel approach called SELF-ALIGN, which combines principle-driven reasoning and the generative power of LLMs for the self-alignment of AI agents with minimal human supervision. Our approach encompasses four stages: first, we use an LLM to generate synthetic prompts, and a topic-guided method to augment the prompt diversity; second, we use a small set of human-written principles for AI models to follow, and guide the LLM through in-context learning from demonstrations (of principles application) to produce helpful, ethical, and reliable responses to user's queries; third, we fine-tune the original LLM with the high-quality self-aligned responses so that the resulting model can generate desirable responses for each query directly without the principle set and the demonstrations anymore; and finally, we offer a refinement step to address the issues of overly-brief or indirect responses. Applying SELF-ALIGN to the LLaMA-65b base language model, we develop an AI assistant named Dromedary. With fewer than 300 lines of human annotations (including < 200 seed prompts, 16 generic principles, and 5 exemplars for in-context learning). Dromedary significantly surpasses the performance of several state-of-the-art AI systems, including Text-Davinci-003 and Alpaca, on benchmark datasets with various settings.Comment: Accepted at NeurIPS 2023 (Spotlight). Project page: https://github.com/IBM/Dromedar

    Spartan Daily, January 26, 1996

    Get PDF
    Volume 106, Issue 2https://scholarworks.sjsu.edu/spartandaily/8787/thumbnail.jp
    • …
    corecore