391 research outputs found

    Waiting time analysis of foreign currency exchange rates: Beyond the renewal-reward theorem

    Get PDF
    We evaluate the average waiting time between observing the price of financial markets and the next price change, especially in an on-line foreign exchange trading service for individual customers via the internet. Basic technical idea of our present work is dependent on the so-called renewal-reward theorem. Assuming that stochastic processes of the market price changes could be regarded as a renewal process, we use the theorem to calculate the average waiting time of the process. In the conventional derivation of the theorem, it is apparently hard to evaluate the higher order moments of the waiting time. To overcome this type of difficulties, we attempt to derive the waiting time distribution Omega(s) directly for arbitrary time interval distribution (first passage time distribution) of the stochastic process P_{W}(tau) and observation time distribution P_{O}(t) of customers. Our analysis enables us to evaluate not only the first moment (the average waiting time) but also any order of the higher moments of the waiting time. Moreover, in our formalism, it is possible to model the observation of the price on the internet by the customers in terms of the observation time distribution P_{O}(t). We apply our analysis to the stochastic process of the on-line foreign exchange rate for individual customers from the Sony bank and compare the moments with the empirical data analysis.Comment: 8pages, 11figures, using IEEEtran.cl

    NoisyICL: A Little Noise in Model Parameters Calibrates In-context Learning

    Full text link
    In-Context Learning (ICL) is suffering from unsatisfactory performance and under-calibration due to high prior bias and unfaithful confidence. Some previous works fine-tuned language models for better ICL performance with enormous datasets and computing costs. In this paper, we propose NoisyICL, simply perturbing the model parameters by random noises to strive for better performance and calibration. Our experiments on two models and 12 downstream datasets show that NoisyICL can help ICL produce more accurate predictions. Our further analysis indicates that NoisyICL enables the model to provide more fair predictions, and also with more faithful confidence. Therefore, we believe that NoisyICL is an effective calibration of ICL. Our experimental code is uploaded to Github.Comment: 20 pages, 28 figures, 7 tables (5 pages, 4 figures, 1 table in main body). ACL 2024 under revie

    R4C: A Benchmark for Evaluating RC Systems to Get the Right Answer for the Right Reason

    Get PDF
    Recent studies have revealed that reading comprehension (RC) systems learn to exploit annotation artifacts and other biases in current datasets. This prevents the community from reliably measuring the progress of RC systems. To address this issue, we introduce R4C, a new task for evaluating RC systems' internal reasoning. R4C requires giving not only answers but also derivations: explanations that justify predicted answers. We present a reliable, crowdsourced framework for scalably annotating RC datasets with derivations. We create and publicly release the R4C dataset, the first, quality-assured dataset consisting of 4.6k questions, each of which is annotated with 3 reference derivations (i.e. 13.8k derivations). Experiments show that our automatic evaluation metrics using multiple reference derivations are reliable, and that R4C assesses different skills from an existing benchmark.Comment: Accepted by ACL2020. See https://naoya-i.github.io/r4c/ for more informatio

    Discovering Highly Influential Shortcut Reasoning: An Automated Template-Free Approach

    Full text link
    Shortcut reasoning is an irrational process of inference, which degrades the robustness of an NLP model. While a number of previous work has tackled the identification of shortcut reasoning, there are still two major limitations: (i) a method for quantifying the severity of the discovered shortcut reasoning is not provided; (ii) certain types of shortcut reasoning may be missed. To address these issues, we propose a novel method for identifying shortcut reasoning. The proposed method quantifies the severity of the shortcut reasoning by leveraging out-of-distribution data and does not make any assumptions about the type of tokens triggering the shortcut reasoning. Our experiments on Natural Language Inference and Sentiment Analysis demonstrate that our framework successfully discovers known and unknown shortcut reasoning in the previous work

    Leveraging Unannotated Texts for Scientific Relation Extraction

    Get PDF
    corecore