13,828 research outputs found

    Combined 3D thinning and greedy algorithm to approximate realistic particles with corrected mechanical properties

    Full text link
    The shape of irregular particles has significant influence on micro- and macro-scopic behavior of granular systems. This paper presents a combined 3D thinning and greedy set-covering algorithm to approximate realistic particles with a clump of overlapping spheres for discrete element method (DEM) simulations. First, the particle medial surface (or surface skeleton), from which all candidate (maximal inscribed) spheres can be generated, is computed by the topological 3D thinning. Then, the clump generation procedure is converted into a greedy set-covering (SCP) problem. To correct the mass distribution due to highly overlapped spheres inside the clump, linear programming (LP) is used to adjust the density of each component sphere, such that the aggregate properties mass, center of mass and inertia tensor are identical or close enough to the prototypical particle. In order to find the optimal approximation accuracy (volume coverage: ratio of clump's volume to the original particle's volume), particle flow of 3 different shapes in a rotating drum are conducted. It was observed that the dynamic angle of repose starts to converge for all particle shapes at 85% volume coverage (spheres per clump < 30), which implies the possible optimal resolution to capture the mechanical behavior of the system.Comment: 34 pages, 13 figure

    Challenges and complexities in application of LCA approaches in the case of ICT for a sustainable future

    Get PDF
    In this work, three of many ICT-specific challenges of LCA are discussed. First, the inconsistency versus uncertainty is reviewed with regard to the meta-technological nature of ICT. As an example, the semiconductor technologies are used to highlight the complexities especially with respect to energy and water consumption. The need for specific representations and metric to separately assess products and technologies is discussed. It is highlighted that applying product-oriented approaches would result in abandoning or disfavoring of new technologies that could otherwise help toward a better world. Second, several believed-untouchable hot spots are highlighted to emphasize on their importance and footprint. The list includes, but not limited to, i) User Computer-Interfaces (UCIs), especially screens and displays, ii) Network-Computer Interlaces (NCIs), such as electronic and optical ports, and iii) electricity power interfaces. In addition, considering cross-regional social and economic impacts, and also taking into account the marketing nature of the need for many ICT's product and services in both forms of hardware and software, the complexity of End of Life (EoL) stage of ICT products, technologies, and services is explored. Finally, the impact of smart management and intelligence, and in general software, in ICT solutions and products is highlighted. In particular, it is observed that, even using the same technology, the significance of software could be highly variable depending on the level of intelligence and awareness deployed. With examples from an interconnected network of data centers managed using Dynamic Voltage and Frequency Scaling (DVFS) technology and smart cooling systems, it is shown that the unadjusted assessments could be highly uncertain, and even inconsistent, in calculating the management component's significance on the ICT impacts.Comment: 10 pages. Preprint/Accepted of a paper submitted to the ICT4S Conferenc

    A finder and representation system for knowledge carriers based on granular computing

    Get PDF
    In one of his publications Aristotle states ”All human beings by their nature desire to know” [Kraut 1991]. This desire is initiated the day we are born and accompanies us for the rest of our life. While at a young age our parents serve as one of the principle sources for knowledge, this changes over the course of time. Technological advances and particularly the introduction of the Internet, have given us new possibilities to share and access knowledge from almost anywhere at any given time. Being able to access and share large collections of written down knowledge is only one part of the equation. Just as important is the internalization of it, which in many cases can prove to be difficult to accomplish. Hence, being able to request assistance from someone who holds the necessary knowledge is of great importance, as it can positively stimulate the internalization procedure. However, digitalization does not only provide a larger pool of knowledge sources to choose from but also more people that can be potentially activated, in a bid to receive personalized assistance with a given problem statement or question. While this is beneficial, it imposes the issue that it is hard to keep track of who knows what. For this task so-called Expert Finder Systems have been introduced, which are designed to identify and suggest the most suited candidates to provide assistance. Throughout this Ph.D. thesis a novel type of Expert Finder System will be introduced that is capable of capturing the knowledge users within a community hold, from explicit and implicit data sources. This is accomplished with the use of granular computing, natural language processing and a set of metrics that have been introduced to measure and compare the suitability of candidates. Furthermore, are the knowledge requirements of a problem statement or question being assessed, in order to ensure that only the most suited candidates are being recommended to provide assistance

    Can banks circumvent minimum capital requirements? The case of mortgage portfolios under Basel II

    Get PDF
    The recent mortgage crisis has resulted in several bank failures as the number of mortgage defaults increased. The current Basel I capital framework does not require banks to hold sufficient amounts of capital to support their mortgage lending activities. The new Basel II capital rules are intended to correct this problem. However, Basel II models could become too complex and too costly to implement, often resulting in a trade-off between complexity and model accuracy. In addition, the variation of the model, particularly how mortgage portfolios are segmented, could have a significant impact on the default and loss estimated and, thus, could affect the amount of capital that banks are required to hold. This paper finds that the calculated Basel II capital varies considerably across the default prediction model and segmentation schemes, thus providing banks with an incentive to choose an approach that results in the least required capital for them. The authors also find that a more granular segmentation model produces smaller required capital, regardless of the economic environment. In addition, while borrowers' credit risk factors are consistently superior, economic factors have also played a role in mortgage default during the financial crisis.Capital ; Banks and banking ; Basel capital accord

    An adaptive stigmergy-based system for evaluating technological indicator dynamics in the context of smart specialization

    Full text link
    Regional innovation is more and more considered an important enabler of welfare. It is no coincidence that the European Commission has started looking at regional peculiarities and dynamics, in order to focus Research and Innovation Strategies for Smart Specialization towards effective investment policies. In this context, this work aims to support policy makers in the analysis of innovation-relevant trends. We exploit a European database of the regional patent application to determine the dynamics of a set of technological innovation indicators. For this purpose, we design and develop a software system for assessing unfolding trends in such indicators. In contrast with conventional knowledge-based design, our approach is biologically-inspired and based on self-organization of information. This means that a functional structure, called track, appears and stays spontaneous at runtime when local dynamism in data occurs. A further prototyping of tracks allows a better distinction of the critical phenomena during unfolding events, with a better assessment of the progressing levels. The proposed mechanism works if structural parameters are correctly tuned for the given historical context. Determining such correct parameters is not a simple task since different indicators may have different dynamics. For this purpose, we adopt an adaptation mechanism based on differential evolution. The study includes the problem statement and its characterization in the literature, as well as the proposed solving approach, experimental setting and results.Comment: mail: [email protected]
    corecore