636 research outputs found
Temperature dependence of the charge carrier mobility in gated quasi-one-dimensional systems
The many-body Monte Carlo method is used to evaluate the frequency dependent
conductivity and the average mobility of a system of hopping charges,
electronic or ionic on a one-dimensional chain or channel of finite length. Two
cases are considered: the chain is connected to electrodes and in the other
case the chain is confined giving zero dc conduction. The concentration of
charge is varied using a gate electrode. At low temperatures and with the
presence of an injection barrier, the mobility is an oscillatory function of
density. This is due to the phenomenon of charge density pinning. Mobility
changes occur due to the co-operative pinning and unpinning of the
distribution. At high temperatures, we find that the electron-electron
interaction reduces the mobility monotonically with density, but perhaps not as
much as one might intuitively expect because the path summation favour the
in-phase contributions to the mobility, i.e. the sequential paths in which the
carriers have to wait for the one in front to exit and so on. The carrier
interactions produce a frequency dependent mobility which is of the same order
as the change in the dc mobility with density, i.e. it is a comparably weak
effect. However, when combined with an injection barrier or intrinsic disorder,
the interactions reduce the free volume and amplify disorder by making it
non-local and this can explain the too early onset of frequency dependence in
the conductivity of some high mobility quasi-one-dimensional organic materials.Comment: 9 pages, 8 figures, to be published in Physical Review
Explainable product backorder prediction exploiting CNN: Introducing explainable models in businesses
Due to expected positive impacts on business, the application of artificial intelligence has been widely increased. The decision-making procedures of those models are often complex and not easily understandable to the company’s stakeholders, i.e. the people having to follow up on recommendations or try to understand automated decisions of a system. This opaqueness and black-box nature might hinder adoption, as users struggle to make sense and trust the predictions of AI models. Recent research on eXplainable Artificial Intelligence (XAI) focused mainly on explaining the models to AI experts with the purpose of debugging and improving the performance of the models. In this article, we explore how such systems could be made explainable to the stakeholders. For doing so, we propose a new convolutional neural network (CNN)-based explainable predictive model for product backorder prediction in inventory management. Backorders are orders that customers place for products that are currently not in stock. The company now takes the risk to produce or acquire the backordered products while in the meantime, customers can cancel their orders if that takes too long, leaving the company with unsold items in their inventory. Hence, for their strategic inventory management, companies need to make decisions based on assumptions. Our argument is that these tasks can be improved by offering explanations for AI recommendations. Hence, our research investigates how such explanations could be provided, employing Shapley additive explanations to explain the overall models’ priority in decision-making. Besides that, we introduce locally interpretable surrogate models that can explain any individual prediction of a model. The experimental results demonstrate effectiveness in predicting backorders in terms of standard evaluation metrics and outperform known related works with AUC 0.9489. Our approach demonstrates how current limitations of predictive technologies can be addressed in the business domain
The Automation of the Taxi Industry – Taxi Drivers’ Expectations and Attitudes Towards the Future of their Work
Advocates of autonomous driving predict that the occupation of taxi driver could be made obsolete by shared autonomous vehicles (SAV) in the long term. Conducting interviews with German taxi drivers, we investigate how they perceive the changes caused by advancing automation for the future of their business. Our study contributes insights into how the work of taxi drivers could change given the advent of autonomous driving: While the task of driving could be taken over by SAVs for standard trips, taxi drivers are certain that other areas of their work such as providing supplementary services and assistance to passengers would constitute a limit to such forms of automation, but probably involving a shifting role for the taxi drivers, one which focuses on the sociality of the work. Our findings illustrate how taxi drivers see the future of their work, suggesting design implications for tools that take various forms of assistance into account, and demonstrating how important it is to consider taxi drivers in the co-design of future taxis and SAV services
Arabic Sentiment Analysis with Noisy Deep Explainable Model
Sentiment Analysis (SA) is an indispensable task for many real-world
applications. Compared to limited resourced languages (i.e., Arabic, Bengali),
most of the research on SA are conducted for high resourced languages (i.e.,
English, Chinese). Moreover, the reasons behind any prediction of the Arabic
sentiment analysis methods exploiting advanced artificial intelligence
(AI)-based approaches are like black-box - quite difficult to understand. This
paper proposes an explainable sentiment classification framework for the Arabic
language by introducing a noise layer on Bi-Directional Long Short-Term Memory
(BiLSTM) and Convolutional Neural Networks (CNN)-BiLSTM models that overcome
over-fitting problem. The proposed framework can explain specific predictions
by training a local surrogate explainable model to understand why a particular
sentiment (positive or negative) is being predicted. We carried out experiments
on public benchmark Arabic SA datasets. The results concluded that adding noise
layers improves the performance in sentiment analysis for the Arabic language
by reducing overfitting and our method outperformed some known state-of-the-art
methods. In addition, the introduced explainability with noise layer could make
the model more transparent and accountable and hence help adopting AI-enabled
system in practice.Comment: This is the pre-print version of our accepted paper at the 7th
International Conference on Natural Language Processing and Information
Retrieval~(ACM NLPIR'2023
Trust your guts: fostering embodied knowledge and sustainable practices through voice interaction
Despite various attempts to prevent food waste and motivate conscious food handling, household members find it difficult to correctly assess the edibility of food. With the rise of ambient voice assistants, we did a design case study to support households’ in situ decision-making process in collaboration with our voice agent prototype, Fischer Fritz. Therefore, we conducted 15 contextual inquiries to understand food practices at home. Furthermore, we interviewed six fish experts to inform the design of our voice agent on how to guide consumers and teach food literacy. Finally, we created a prototype and discussed with 15 consumers its impact and capability to convey embodied knowledge to the human that is engaged as sensor. Our design research goes beyond current Human-Food Interaction automation approaches by emphasizing the human-food relationship in technology design and demonstrating future complementary human-agent collaboration with the aim to increase humans’ competence to sense, think, and act
Unveiling Black-boxes: Explainable Deep Learning Models for Patent Classification
Recent technological advancements have led to a large number of patents in a
diverse range of domains, making it challenging for human experts to analyze
and manage. State-of-the-art methods for multi-label patent classification rely
on deep neural networks (DNNs), which are complex and often considered
black-boxes due to their opaque decision-making processes. In this paper, we
propose a novel deep explainable patent classification framework by introducing
layer-wise relevance propagation (LRP) to provide human-understandable
explanations for predictions. We train several DNN models, including Bi-LSTM,
CNN, and CNN-BiLSTM, and propagate the predictions backward from the output
layer up to the input layer of the model to identify the relevance of words for
individual predictions. Considering the relevance score, we then generate
explanations by visualizing relevant words for the predicted patent class.
Experimental results on two datasets comprising two-million patent texts
demonstrate high performance in terms of various evaluation measures. The
explanations generated for each prediction highlight important relevant words
that align with the predicted class, making the prediction more understandable.
Explainable systems have the potential to facilitate the adoption of complex
AI-enabled methods for patent classification in real-world applications.Comment: This is the pre-print of the submitted manuscript on the World
Conference on eXplainable Artificial Intelligence (xAI2023), Lisbon,
Portugal. The published manuscript can be found here
https://doi.org/10.1007/978-3-031-44067-0_2
Designing for ethical innovation:a case study on ELSI co-design in emergency
The ever more pervasive ‘informationalization’ of crisis management and response brings both unprecedented opportunities and challenges. Recent years have seen the emergence of attention to ethical, legal and social issues (ELSI) in the field of Information and Communication Technology. However, disclosing (and addressing) ELSI issues in design is still a challenge because they are inherently relational, arising from interactions between people, the material and design of the artifact, and the context. In this article, we discuss approaches for addressing such ‘deeper’ and ‘wider’ political implications, values and ethical, legal and social implications that arise between practices, people and technology. Based on a case study from the BRIDGE project, which has provided the opportunity for deep engagement with these issues through the concrete exploration and experimentation with technologically augmented practices of emergency response, we present insights from our interdisciplinary work aiming to make design and innovation projects ELSI-aware. Crucially, we have seen in our study a need for a shift from privacy by design towards designing for privacy, collaboration, trust, accessibility, ownership, transparency etc., acknowledging that these are emergent practices that we cannot control by design, but rather that we can help to design for—calling for approaches that allow to make ELSI issues explicit and addressable in design-time
- …