44 research outputs found

    Meta-Reasoning: Semantics-Symbol Deconstruction For Large Language Models

    Full text link
    Neural-symbolic methods have shown their effectiveness in enhancing the reasoning abilities of large language models (LLMs). However, existing methods primarily rely on mapping natural languages to more syntactically complete formal languages (e.g., Python and SQL). Those approaches necessitate that reasoning tasks be convertible into programs, which cater more to the computer execution mindset and deviate from human reasoning habits. To expand the real-world applicability and flexibility of symbolic methods, we propose Meta-Reasoning from the scope of linguistics itself. This method empowers LLMs to deconstruct questions and effectively capture more generalized knowledge autonomously. We find that Meta-Reasoning achieves improved in-context learning efficiency, reasoning accuracy, and output stability in six arithmetic and symbolic reasoning tasks. In particular, when applied to symbolic reasoning tasks such as Tracking Shuffled Objects, GPT-3 (text-davinci-002) surpasses the few-shot Chain-of-Thought prompting approach (+37.7%), with 99% accuracy after a single demonstration of Meta-Reasoning.Comment: Work in progres
    corecore