462 research outputs found

    An Empirical Study of Bots in Software Development -- Characteristics and Challenges from a Practitioner's Perspective

    Full text link
    Software engineering bots - automated tools that handle tedious tasks - are increasingly used by industrial and open source projects to improve developer productivity. Current research in this area is held back by a lack of consensus of what software engineering bots (DevBots) actually are, what characteristics distinguish them from other tools, and what benefits and challenges are associated with DevBot usage. In this paper we report on a mixed-method empirical study of DevBot usage in industrial practice. We report on findings from interviewing 21 and surveying a total of 111 developers. We identify three different personas among DevBot users (focusing on autonomy, chat interfaces, and "smartness"), each with different definitions of what a DevBot is, why developers use them, and what they struggle with. We conclude that future DevBot research should situate their work within our framework, to clearly identify what type of bot the work targets, and what advantages practitioners can expect. Further, we find that there currently is a lack of general purpose "smart" bots that go beyond simple automation tools or chat interfaces. This is problematic, as we have seen that such bots, if available, can have a transformative effect on the projects that use them.Comment: To be published at the ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering (ESEC/FSE

    Deception against Deception: Toward A Deception Framework for Detection and Characterization of Covert Micro-targeting Campaigns on Online Social Networks

    Get PDF
    Micro-targeting campaigns on online social networks are an emerging class of social engineering attacks that prime individuals via personalized content for malicious purposes. Detecting micro-targeting campaigns is challenging due to their clandestine nature and the lack of visibility around users’ private communications. Our work aims to devise theories, methods, and tools to detect suspected micro-targeting campaigns. To this end, we propose to design and generate a network of decoy personas with characteristics similar to those of targeted groups in order to trap, engage, and identify micro-targeting campaigns. In this paper, we discuss our motivation to conduct this interdisciplinary research effort and introduce our focal research questions and preliminary design for a network of decoy personas

    Design Perspective on the Role of Advanced Bots for Self-Guided Learning

    Get PDF
    Virtual worlds are rapidly gaining acceptance in educational settings; with bots play an important role in these environments to help learners. Authentic learning can be significantly supported by bots to help self-guided learning in authentic tasks. in this paper, we investigate what is stopping educators from making more use of bots as a valuable resource and how these barriers can be overcome. This exploratory research uses interviews with six educators, who use educational bots. We show that while the experts have 'big plans' for bot use, the current educational implementations are 'low-level' and restrictive in their application. There is further confusion about appropriate pedagogical models and how to use them effectively as more than 'prompters' or 'extras'. While creation- and control-technologies are advancing, allowing use of bots as a 'hard technology' to guide learners through routine procedures; there is a lack of resources for automation as intelligence technologies are slower to develop and may required future partnerships with external parties before they are available useable by general educators

    Improving conversations with digital assistants through extracting, recommending, and verifying user inputs

    Get PDF
    Digital assistants, including chat bots and voice assistants, suffer from discrepancies and uncertainty in human text and speech inputs. Human dialogue is often varied, ambiguous, and inconsistent, making data entry prone to error and difficult for digital assistants to process. Finding and extracting pertinent information from unstructured user inputs improves and expands the use of digital assistants on any platform. By confirming data entries and providing relevant recommendations when invalid information is provided, the digital assistant enables the use of natural language and introduces a higher degree of flow into the conversation.This paper describes a series of input logic codifiers that form a corrective method to overcome errors and ambiguity typical of voice and text inputs. When users make a common mistake or forget data, the digital assistant can bridge the gap by recommending the most similar data that is available. The assistant measures the delta between the user’s utterance and valid entries using fuzzy logic to identify the closest and next closest data that relates to the unstructured text.Furthermore, there are endless ways to denote dates, locations, etc., making it difficult for digital assistants to extract accurate and relevant data from the user’s natural language. However, the assistant may infer the desired data format or reference from the dialogue provided and validate this with the user as a follow-on question. The desired data format or type is inferred using fuzzy extraction methods, such as fuzzy date extraction, to isolate the desired data format from the unstructured text. This extracted information is then verified or confirmed by the user to maintain data accuracy and avoid downstream data quality issues

    UniMS-RAG: A Unified Multi-source Retrieval-Augmented Generation for Personalized Dialogue Systems

    Full text link
    Large Language Models (LLMs) has shown exceptional capabilities in many natual language understanding and generation tasks. However, the personalization issue still remains a much-coveted property, especially when it comes to the multiple sources involved in the dialogue system. To better plan and incorporate the use of multiple sources in generating personalized response, we firstly decompose it into three sub-tasks: Knowledge Source Selection, Knowledge Retrieval, and Response Generation. We then propose a novel Unified Multi-Source Retrieval-Augmented Generation system (UniMS-RAG) Specifically, we unify these three sub-tasks with different formulations into the same sequence-to-sequence paradigm during the training, to adaptively retrieve evidences and evaluate the relevance on-demand using special tokens, called acting tokens and evaluation tokens. Enabling language models to generate acting tokens facilitates interaction with various knowledge sources, allowing them to adapt their behavior to diverse task requirements. Meanwhile, evaluation tokens gauge the relevance score between the dialogue context and the retrieved evidence. In addition, we carefully design a self-refinement mechanism to iteratively refine the generated response considering 1) the consistency scores between the generated response and retrieved evidence; and 2) the relevance scores. Experiments on two personalized datasets (DuLeMon and KBP) show that UniMS-RAG achieves state-of-the-art performance on the knowledge source selection and response generation task with itself as a retriever in a unified manner. Extensive analyses and discussions are provided for shedding some new perspectives for personalized dialogue systems
    • …
    corecore