142 research outputs found

    Multi-functional Building

    Get PDF
    Diplomová práce je vypracovaná ve formě projektové dokumentace polyfunkčního domu. Objekt má jeden podzemní a čtyři nadzemní podlaží. Navržená budova má plochou střechu a dosahuje výšku 16,650 m nad podlahou 1.NP. Budova má nosnou systém ze betonového monolitického skeletu. Stropní deska je řešená jako deska lokálně podepřená sloupy. Budova se skládá ze dvou částí. V prvním části jsou obchodní prostory v 1.NP a ve podsklepené podlaží je prostor pro Fitness. Druhá část slouží pro bydlení skládající se ze 12 obytných jednotekMaster thesis is composed from project documentation of Multi-functional building. The object has one basement floor and four floor. The building has flat roof and come up to height 16,65m. The structural system is designed from reinforced concrete skeleton. The building has two main parts. The first is commercial space, which is located on first floor and basement. The second part is used for living. The building has 12 apartments.

    On the proof of some theorem on locally nilpotent subgroups in division rings

    Full text link
    In Hai-Thin (2009), there is a theorem, stating that every locally nilpotent subnormal subgroup in a division ring DD is central (see Hai-Thin (2009, Th. 2.2)). Unfortunately, there is some mistake in the proof of this theorem. In this note we give the another proof of this theorem.Comment: 3 page

    Diffeomorphic Information Neural Estimation

    Full text link
    Mutual Information (MI) and Conditional Mutual Information (CMI) are multi-purpose tools from information theory that are able to naturally measure the statistical dependencies between random variables, thus they are usually of central interest in several statistical and machine learning tasks, such as conditional independence testing and representation learning. However, estimating CMI, or even MI, is infamously challenging due the intractable formulation. In this study, we introduce DINE (Diffeomorphic Information Neural Estimator)-a novel approach for estimating CMI of continuous random variables, inspired by the invariance of CMI over diffeomorphic maps. We show that the variables of interest can be replaced with appropriate surrogates that follow simpler distributions, allowing the CMI to be efficiently evaluated via analytical solutions. Additionally, we demonstrate the quality of the proposed estimator in comparison with state-of-the-arts in three important tasks, including estimating MI, CMI, as well as its application in conditional independence testing. The empirical evaluations show that DINE consistently outperforms competitors in all tasks and is able to adapt very well to complex and high-dimensional relationships.Comment: Accepted at the Thirty-Seventh AAAI Conference on Artificial Intelligence (AAAI 2023

    Domain Generalisation via Risk Distribution Matching

    Full text link
    We propose a novel approach for domain generalisation (DG) leveraging risk distributions to characterise domains, thereby achieving domain invariance. In our findings, risk distributions effectively highlight differences between training domains and reveal their inherent complexities. In testing, we may observe similar, or potentially intensifying in magnitude, divergences between risk distributions. Hence, we propose a compelling proposition: Minimising the divergences between risk distributions across training domains leads to robust invariance for DG. The key rationale behind this concept is that a model, trained on domain-invariant or stable features, may consistently produce similar risk distributions across various domains. Building upon this idea, we propose Risk Distribution Matching (RDM). Using the maximum mean discrepancy (MMD) distance, RDM aims to minimise the variance of risk distributions across training domains. However, when the number of domains increases, the direct optimisation of variance leads to linear growth in MMD computations, resulting in inefficiency. Instead, we propose an approximation that requires only one MMD computation, by aligning just two distributions: that of the worst-case domain and the aggregated distribution from all domains. Notably, this method empirically outperforms optimising distributional variance while being computationally more efficient. Unlike conventional DG matching algorithms, RDM stands out for its enhanced efficacy by concentrating on scalar risk distributions, sidestepping the pitfalls of high-dimensional challenges seen in feature or gradient matching. Our extensive experiments on standard benchmark datasets demonstrate that RDM shows superior generalisation capability over state-of-the-art DG methods.Comment: Accepted at 2024 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV 2024
    corecore