7,328 research outputs found

    Breaking boundaries:Charge density waves, quantum measurement, and black holes in theoretical physics

    Get PDF
    This thesis, titled “Breaking Boundaries” is a journey through three topics united by the theme of boundaries in physics. First, the journey begins with an investigation into charge density waves (CDWs) and their nearly commensurate phase, focusing on the materials 2H-TaSe2 and 1T-TaS2. An extensive treatment of Ginzburg-Landau theory is covered with an extension into truly two-dimensional systems. This extension is used to study spiral patches of commensurate charge density waves observed in experiment. The research leads to a novel perspective on CDW behaviour with the existence of a spiral CDW phase in a range of materials. Secondly, transitioning to the quantum realm, the thesis addresses the quantum measurement problem, emphasizing the constraints any valid theory must possess. It critiques existing models, demonstrates the non-linearity of objective collapse theories, and proposes a minimal model that bridges quantum mechanics and classical physics. Thirdly, the thesis delves into black holes and specifically the phenomena of thermal radiation due to a horizon. First, we explore analogue models that mimic the thermal spectrum near a black hole horizon, to pave the way to experimental realization. Then we focus on the region far away from a black hole horizon and challenge the notion of remnant radiation at this position. With a theoretical toy model, we study the regime and find a non-evaporating black hole. This questions the validity of standard Hawking radiation calculations.In conclusion, the thesis navigates through the boundaries of material behaviours, the quantum-classical divide, and the enigmatic nature of black holes. It highlights the blurring and breaking of boundaries in physics, offering new perspectives and promising avenues for future discoveries

    On the Generation of Realistic and Robust Counterfactual Explanations for Algorithmic Recourse

    Get PDF
    This recent widespread deployment of machine learning algorithms presents many new challenges. Machine learning algorithms are usually opaque and can be particularly difficult to interpret. When humans are involved, algorithmic and automated decisions can negatively impact people’s lives. Therefore, end users would like to be insured against potential harm. One popular way to achieve this is to provide end users access to algorithmic recourse, which gives end users negatively affected by algorithmic decisions the opportunity to reverse unfavorable decisions, e.g., from a loan denial to a loan acceptance. In this thesis, we design recourse algorithms to meet various end user needs. First, we propose methods for the generation of realistic recourses. We use generative models to suggest recourses likely to occur under the data distribution. To this end, we shift the recourse action from the input space to the generative model’s latent space, allowing to generate counterfactuals that lie in regions with data support. Second, we observe that small changes applied to the recourses prescribed to end users likely invalidate the suggested recourse after being nosily implemented in practice. Motivated by this observation, we design methods for the generation of robust recourses and for assessing the robustness of recourse algorithms to data deletion requests. Third, the lack of a commonly used code-base for counterfactual explanation and algorithmic recourse algorithms and the vast array of evaluation measures in literature make it difficult to compare the per formance of different algorithms. To solve this problem, we provide an open source benchmarking library that streamlines the evaluation process and can be used for benchmarking, rapidly developing new methods, and setting up new experiments. In summary, our work contributes to a more reliable interaction of end users and machine learned models by covering fundamental aspects of the recourse process and suggests new solutions towards generating realistic and robust counterfactual explanations for algorithmic recourse

    Assessing generalisability of deep learning-based polyp detection and segmentation methods through a computer vision challenge

    Get PDF
    Polyps are well-known cancer precursors identified by colonoscopy. However, variability in their size, appearance, and location makes the detection of polyps challenging. Moreover, colonoscopy surveillance and removal of polyps are highly operator-dependent procedures and occur in a highly complex organ topology. There exists a high missed detection rate and incomplete removal of colonic polyps. To assist in clinical procedures and reduce missed rates, automated methods for detecting and segmenting polyps using machine learning have been achieved in past years. However, the major drawback in most of these methods is their ability to generalise to out-of-sample unseen datasets from different centres, populations, modalities, and acquisition systems. To test this hypothesis rigorously, we, together with expert gastroenterologists, curated a multi-centre and multi-population dataset acquired from six different colonoscopy systems and challenged the computational expert teams to develop robust automated detection and segmentation methods in a crowd-sourcing Endoscopic computer vision challenge. This work put forward rigorous generalisability tests and assesses the usability of devised deep learning methods in dynamic and actual clinical colonoscopy procedures. We analyse the results of four top performing teams for the detection task and five top performing teams for the segmentation task. Our analyses demonstrate that the top-ranking teams concentrated mainly on accuracy over the real-time performance required for clinical applicability. We further dissect the devised methods and provide an experiment-based hypothesis that reveals the need for improved generalisability to tackle diversity present in multi-centre datasets and routine clinical procedures

    Enhancing glomeruli segmentation through cross-species pre-training

    Get PDF
    The importance of kidney biopsy, a medical procedure in which a small tissue sample is extracted from the kidney for examination, is increasing due to the rising incidence of kidney disorders. This procedure helps diagnosing several kidney diseases which are cause of kidney function changes, as well as guiding treatment decisions, and evaluating the suitability of potential donor kidneys for transplantation. In this work, a deep learning system for the automatic segmentation of glomeruli in biopsy kidney images is presented. A novel cross-species transfer learning approach, in which a semantic segmentation network is trained on mouse kidney tissue images and then fine-tuned on human data, is proposed to boost the segmentation performance. The experiments conducted using two deep semantic segmentation networks, MobileNet and SegNeXt, demonstrated the effectiveness of the cross-species pre-training approach leading to an increased generalization ability of both models

    On the structure of repeated-root polycyclic codes over local rings

    Get PDF
    Producción CientíficaThis paper provides the Generalized Mattson Solomon polynomial for repeated-root polycyclic codes over local rings that gives an explicit decomposition of them in terms of idempotents. It also states some structural properties of repeated-root polycyclic codes over finite fields in terms of matrix product codes. Both approaches provide a description of the -dual code for a given polycyclic code.MCIN/AEI /10.13039/501100011033 - EU NextGenerationEU/ PRTR (Grant TED2021-130358B-I00)Bulgarian Ministry of Education and Science, Scientific Programme “Enhancing the Research Capacity in Mathematical Sciences (PIKOM)”, No. DO1-67/05.05.2022.TÜB˙ITAK within the scope of 2219 International Post Doctoral Research Fellowship Program with application number 1059B19210116

    Quantum-Classical hybrid systems and their quasifree transformations

    Get PDF
    The focus of this work is the description of a framework for quantum-classical hybrid systems. The main emphasis lies on continuous variable systems described by canonical commutation relations and, more precisely, the quasifree case. Here, we are going to solve two main tasks: The first is to rigorously define spaces of states and observables, which are naturally connected within the general structure. Secondly, we want to describe quasifree channels for which both the Schrödinger picture and the Heisenberg picture are well defined. We start with a general introduction to operator algebras and algebraic quantum theory. Thereby, we highlight some of the mathematical details that are often taken for granted while working with purely quantum systems. Consequently, we discuss several possibilities and their advantages respectively disadvantages in describing classical systems analogously to the quantum formalism. The key takeaway is that there is no candidate for a classical state space or observable algebra that can be put easily alongside a quantum system to form a hybrid and simultaneously fulfills all of our requirements for such a partially quantum and partially classical system. Although these straightforward hybrid systems are not sufficient enough to represent a general approach, we use one of the candidates to prove an intermediate result, which showcases the advantages of a consequent hybrid ansatz: We provide a hybrid generalization of classical diffusion generators where the exchange of information between the classical and the quantum side is controlled by the induced noise on the quantum system. Then, we present solutions for our initial tasks. We start with a CCR-algebra where some variables may commute with all others and hence generate a classical subsystem. After clarifying the necessary representations, our hybrid states are given by continuous characteristic functions, and the according state space is equal to the state space of a non-unital C*-algebra. While this C*-algebra is not a suitable candidate for an observable algebra itself, we describe several possible subsets in its bidual which can serve this purpose. They can be more easily characterized and will also allow for a straightforward definition of a proper Heisenberg picture. The subsets are given by operator-valued functions on the classical phase space with varying degrees of regularity, such as universal measurability or strong*-continuity. We describe quasifree channels and their properties, including a state-channel correspondence, a factorization theorem, and some basic physical operations. All this works solely on the assumption of a quasifree system, but we also show that the more famous subclass of Gaussian systems fits well within this formulation and behaves as expected

    The infrared structure of perturbative gauge theories

    Get PDF
    Infrared divergences in the perturbative expansion of gauge theory amplitudes and cross sections have been a focus of theoretical investigations for almost a century. New insights still continue to emerge, as higher perturbative orders are explored, and high-precision phenomenological applications demand an ever more refined understanding. This review aims to provide a pedagogical overview of the subject. We briefly cover some of the early historical results, we provide some simple examples of low-order applications in the context of perturbative QCD, and discuss the necessary tools to extend these results to all perturbative orders. Finally, we describe recent developments concerning the calculation of soft anomalous dimensions in multi-particle scattering amplitudes at high orders, and we provide a brief introduction to the very active field of infrared subtraction for the calculation of differential distributions at colliders. © 2022 Elsevier B.V

    LIPIcs, Volume 251, ITCS 2023, Complete Volume

    Get PDF
    LIPIcs, Volume 251, ITCS 2023, Complete Volum

    Clinically Labeled Contrastive Learning for OCT Biomarker Classification

    Full text link
    This paper presents a novel positive and negative set selection strategy for contrastive learning of medical images based on labels that can be extracted from clinical data. In the medical field, there exists a variety of labels for data that serve different purposes at different stages of a diagnostic and treatment process. Clinical labels and biomarker labels are two examples. In general, clinical labels are easier to obtain in larger quantities because they are regularly collected during routine clinical care, while biomarker labels require expert analysis and interpretation to obtain. Within the field of ophthalmology, previous work has shown that clinical values exhibit correlations with biomarker structures that manifest within optical coherence tomography (OCT) scans. We exploit this relationship by using the clinical data as pseudo-labels for our data without biomarker labels in order to choose positive and negative instances for training a backbone network with a supervised contrastive loss. In this way, a backbone network learns a representation space that aligns with the clinical data distribution available. Afterwards, we fine-tune the network trained in this manner with the smaller amount of biomarker labeled data with a cross-entropy loss in order to classify these key indicators of disease directly from OCT scans. We also expand on this concept by proposing a method that uses a linear combination of clinical contrastive losses. We benchmark our methods against state of the art self-supervised methods in a novel setting with biomarkers of varying granularity. We show performance improvements by as much as 5\% in total biomarker detection AUROC.Comment: Accepted in IEEE Journal of Biomedical and Health Informatics. arXiv admin note: text overlap with arXiv:2211.0509
    corecore