146 research outputs found

    GPT-4 Is Too Smart To Be Safe: Stealthy Chat with LLMs via Cipher

    Full text link
    Safety lies at the core of the development of Large Language Models (LLMs). There is ample work on aligning LLMs with human ethics and preferences, including data filtering in pretraining, supervised fine-tuning, reinforcement learning from human feedback, and red teaming, etc. In this study, we discover that chat in cipher can bypass the safety alignment techniques of LLMs, which are mainly conducted in natural languages. We propose a novel framework CipherChat to systematically examine the generalizability of safety alignment to non-natural languages -- ciphers. CipherChat enables humans to chat with LLMs through cipher prompts topped with system role descriptions and few-shot enciphered demonstrations. We use CipherChat to assess state-of-the-art LLMs, including ChatGPT and GPT-4 for different representative human ciphers across 11 safety domains in both English and Chinese. Experimental results show that certain ciphers succeed almost 100% of the time to bypass the safety alignment of GPT-4 in several safety domains, demonstrating the necessity of developing safety alignment for non-natural languages. Notably, we identify that LLMs seem to have a ''secret cipher'', and propose a novel SelfCipher that uses only role play and several demonstrations in natural language to evoke this capability. SelfCipher surprisingly outperforms existing human ciphers in almost all cases. Our code and data will be released at https://github.com/RobustNLP/CipherChat.Comment: 13 pages, 4 figures, 9 table

    Encouraging Divergent Thinking in Large Language Models through Multi-Agent Debate

    Full text link
    Modern large language models (LLMs) like ChatGPT have shown remarkable performance on general language tasks but still struggle on complex reasoning tasks, which drives the research on cognitive behaviors of LLMs to explore human-like problem-solving strategies. Along this direction, one representative strategy is self-reflection, which asks an LLM to refine the solution with the feedback generated by itself iteratively. However, our study shows that such reflection-style methods suffer from the Degeneration-of-Thought (DoT) problem: once the LLM has established confidence in its solutions, it is unable to generate novel thoughts later through reflection even if its initial stance is incorrect. To address the DoT problem, we propose a Multi-Agent Debate (MAD) framework, in which multiple agents express their arguments in the state of "tit for tat" and a judge manages the debate process to obtain a final solution. Clearly, our MAD framework encourages divergent thinking in LLMs which would be helpful for tasks that require deep levels of contemplation. Experiment results on two challenging datasets, commonsense machine translation and counter-intuitive arithmetic reasoning, demonstrate the effectiveness of our MAD framework. Extensive analyses suggest that the adaptive break of debate and the modest level of "tit for tat" state are required for MAD to obtain good performance. Moreover, we find that LLMs might not be a fair judge if different LLMs are used for agents. Codes: https://github.com/Skytliang/Multi-Agents-DebateComment: Work in progres

    2′-Eth­oxy-1,3,3-trimethyl­spiro­[indoline-2,3′-3H-naphtho­[2,1-b][1,4]oxazine]

    Get PDF
    In the title compound, C24H24N2O2, the five-membered ring of the indoline ring system adopts an envelope conformation with the spiro C atom at the flap. The dihedral angle between the benzene ring of the indoline ring system and the naphthalene ring system is 71.70 (7)°. In the crystal structure, pair of weak C—H⋯O hydrogen bonds link the mol­ecules into centrosymmetric dimers

    Exploring Human-Like Translation Strategy with Large Language Models

    Full text link
    Large language models (LLMs) have demonstrated impressive capabilities in general scenarios, exhibiting a level of aptitude that approaches, in some aspects even surpasses, human-level intelligence. Among their numerous skills, the translation abilities of LLMs have received considerable attention. In contrast to traditional machine translation that focuses solely on source-target mapping, LLM-based translation can potentially mimic the human translation process that takes many preparatory steps to ensure high-quality translation. This work aims to explore this possibility by proposing the MAPS framework, which stands for Multi-Aspect Prompting and Selection. Specifically, we enable LLMs to first analyze the given source text and extract three aspects of translation-related knowledge: keywords, topics and relevant demonstrations to guide the translation process. To filter out the noisy and unhelpful knowledge, we employ a selection mechanism based on quality estimation. Experiments suggest that MAPS brings significant and consistent improvements over text-davinci-003 and Alpaca on eight translation directions from the latest WMT22 test sets. Our further analysis shows that the extracted knowledge is critical in resolving up to 59% of hallucination mistakes in translation. Code is available at https://github.com/zwhe99/MAPS-mt.Comment: V2: add more experiments and case studies; polish writin

    County-Scale Spatial Distribution of Soil Enzyme Activities and Enzyme Activity Indices in Agricultural Land: Implications for Soil Quality Assessment

    Get PDF
    Here the spatial distribution of soil enzymatic properties in agricultural land was evaluated on a county-wide (567 km2) scale in Changwu, Shaanxi Province, China. The spatial variations in activities of five hydrolytic enzymes were examined using geostatistical methods. The relationships between soil enzyme activities and other soil properties were evaluated using both an integrated total enzyme activity index (TEI) and the geometric mean of enzyme activities (GME). At the county scale, soil invertase, phosphatase, and catalase activities were moderately spatially correlated, whereas urease and dehydrogenase activities were weakly spatially correlated. Correlation analysis showed that both TEI and GME were better correlated with selected soil physicochemical properties than single enzyme activities. Multivariate regression analysis showed that soil OM content had the strongest positive effect while soil pH had a negative effect on the two enzyme activity indices. In addition, total phosphorous content had a positive effect on TEI and GME in orchard soils, whereas alkali-hydrolyzable nitrogen and available potassium contents, respectively, had negative and positive effects on these two enzyme indices in cropland soils. The results indicate that land use changes strongly affect soil enzyme activities in agricultural land, where TEI provides a sensitive biological indicator for soil quality

    LdsConv : learned depthwise separable convolutions by group pruning

    Get PDF
    Standard convolutional filters usually capture unnecessary overlap of features resulting in a waste of computational cost. In this paper, we aim to solve this problem by proposing a novel Learned Depthwise Separable Convolution (LdsConv) operation that is smart but has a strong capacity for learning. It integrates the pruning technique into the design of convolutional filters, formulated as a generic convolutional unit that can be used as a direct replacement of convolutions without any adjustments of the architecture. To show the effectiveness of the proposed method, experiments are carried out using the state-of-the-art convolutional neural networks (CNNs), including ResNet, DenseNet, SE-ResNet and MobileNet, respectively. The results show that by simply replacing the original convolution with LdsConv in these CNNs, it can achieve a significantly improved accuracy while reducing computational cost. For the case of ResNet50, the FLOPs can be reduced by 40.9%, meanwhile the accuracy on the associated ImageNet increases
    corecore