330 research outputs found

    Where's the Liability in Harmful AI Speech?

    Full text link
    Generative AI, in particular text-based "foundation models" (large models trained on a huge variety of information including the internet), can generate speech that could be problematic under a wide range of liability regimes. Machine learning practitioners regularly "red team" models to identify and mitigate such problematic speech: from "hallucinations" falsely accusing people of serious misconduct to recipes for constructing an atomic bomb. A key question is whether these red-teamed behaviors actually present any liability risk for model creators and deployers under U.S. law, incentivizing investments in safety mechanisms. We examine three liability regimes, tying them to common examples of red-teamed model behaviors: defamation, speech integral to criminal conduct, and wrongful death. We find that any Section 230 immunity analysis or downstream liability analysis is intimately wrapped up in the technical details of algorithm design. And there are many roadblocks to truly finding models (and their associated parties) liable for generated speech. We argue that AI should not be categorically immune from liability in these scenarios and that as courts grapple with the already fine-grained complexities of platform algorithms, the technical details of generative AI loom above with thornier questions. Courts and policymakers should think carefully about what technical design incentives they create as they evaluate these issues.Comment: Published in the Journal of Free Speech Law (2023

    Supporting Source Code Search with Context-Aware and Semantics-Driven Query Reformulation

    Get PDF
    Software bugs and failures cost trillions of dollars every year, and could even lead to deadly accidents (e.g., Therac-25 accident). During maintenance, software developers fix numerous bugs and implement hundreds of new features by making necessary changes to the existing software code. Once an issue report (e.g., bug report, change request) is assigned to a developer, she chooses a few important keywords from the report as a search query, and then attempts to find out the exact locations in the software code that need to be either repaired or enhanced. As a part of this maintenance, developers also often select ad hoc queries on the fly, and attempt to locate the reusable code from the Internet that could assist them either in bug fixing or in feature implementation. Unfortunately, even the experienced developers often fail to construct the right search queries. Even if the developers come up with a few ad hoc queries, most of them require frequent modifications which cost significant development time and efforts. Thus, construction of an appropriate query for localizing the software bugs, programming concepts or even the reusable code is a major challenge. In this thesis, we overcome this query construction challenge with six studies, and develop a novel, effective code search solution (BugDoctor) that assists the developers in localizing the software code of interest (e.g., bugs, concepts and reusable code) during software maintenance. In particular, we reformulate a given search query (1) by designing novel keyword selection algorithms (e.g., CodeRank) that outperform the traditional alternatives (e.g., TF-IDF), (2) by leveraging the bug report quality paradigm and source document structures which were previously overlooked and (3) by exploiting the crowd knowledge and word semantics derived from Stack Overflow Q&A site, which were previously untapped. Our experiment using 5000+ search queries (bug reports, change requests, and ad hoc queries) suggests that our proposed approach can improve the given queries significantly through automated query reformulations. Comparison with 10+ existing studies on bug localization, concept location and Internet-scale code search suggests that our approach can outperform the state-of-the-art approaches with a significant margin

    Three-dimensional conversation : the shift to a public, asynchronous and persistent exchange in Malta

    Get PDF
    An observation of the evolution of the marketing messages of Telecommunications Company Vodafone between 2007 and 2013 sheds light on the significant changes that occurred in the communications arena throughout this period. The shift is not a hypothetical one; it is real and reflected in the shifting usage profiles of millions of mobile users. Moreover the shift is not limited to the changes in the technology which enables mediated conversation. Reference is made to existing literature to define the activity under study, understand the historical context of conversation, both in the mobile and online space, measure the present shifts and explore how findings can contribute to a better understanding of the future. In the context of the existing body of work and the significant changes that occurred over the past years, the research aims to propose a new model of conversation in response to the chosen research question, which asks, “how is conversation evolving as a result of take up of new media in Malta?“ A two‐step approach is adopted. The first research stream makes use of a data set of usage logs of a sample of smartphone adopters on the Vodafone network. A comparison of the usage logs before and after adoption is used to shed light on the influence of the device on the users’ conversations. The analysis is supported with two secondary experiments, one relating to the usage of mobile Internet on specific days during the year and the other extending the experiment to everyday conversation on Facebook. The second research stream consists of a review of the new media landscape with a specific focus on key themes. The findings are used to corroborate a model of shifting conversation. The model proposes that conversation is captured in three dimensions - a shift from synchronous to asynchronous conversation, from private to public and from transient to persistent exchanges

    Topic driven testing

    Get PDF
    Modern interactive applications offer so many interaction opportunities that automated exploration and testing becomes practically impossible without some domain specific guidance towards relevant functionality. In this dissertation, we present a novel fundamental graphical user interface testing method called topic-driven testing. We mine the semantic meaning of interactive elements, guide testing, and identify core functionality of applications. The semantic interpretation is close to human understanding and allows us to learn specifications and transfer knowledge across multiple applications independent of the underlying device, platform, programming language, or technology stack—to the best of our knowledge a unique feature of our technique. Our tool ATTABOY is able to take an existing Web application test suite say from Amazon, execute it on ebay, and thus guide testing to relevant core functionality. Tested on different application domains such as eCommerce, news pages, mail clients, it can trans- fer on average sixty percent of the tested application behavior to new apps—without any human intervention. On top of that, topic-driven testing can go with even more vague instructions of how-to descriptions or use-case descriptions. Given an instruction, say “add item to shopping cart”, it tests the specified behavior in an application–both in a browser as well as in mobile apps. It thus improves state-of-the-art UI testing frame- works, creates change resilient UI tests, and lays the foundation for learning, transfer- ring, and enforcing common application behavior. The prototype is up to five times faster than existing random testing frameworks and tests functions that are hard to cover by non-trained approaches.Moderne interaktive Anwendungen bieten so viele Interaktionsmöglichkeiten, dass eine vollstĂ€ndige automatische Exploration und das Testen aller Szenarien praktisch unmöglich ist. Stattdessen muss die Testprozedur auf relevante KernfunktionalitĂ€t ausgerichtet werden. Diese Arbeit stellt ein neues fundamentales Testprinzip genannt thematisches Testen vor, das beliebige Anwendungen u ̈ber die graphische OberflĂ€che testet. Wir untersuchen die semantische Bedeutung von interagierbaren Elementen um die Kernfunktionenen von Anwendungen zu identifizieren und entsprechende Tests zu erzeugen. Statt typischen starren Testinstruktionen orientiert sich diese Art von Tests an menschlichen AnwendungsfĂ€llen in natĂŒrlicher Sprache. Dies erlaubt es, Software Spezifikationen zu erlernen und Wissen von einer Anwendung auf andere zu ĂŒbertragen unabhĂ€ngig von der Anwendungsart, der Programmiersprache, dem TestgerĂ€t oder der -Plattform. Nach unserem Kenntnisstand ist unser Ansatz der Erste dieser Art. Wir prĂ€sentieren ATTABOY, ein Programm, das eine existierende Testsammlung fĂŒr eine Webanwendung (z.B. fĂŒr Amazon) nimmt und in einer beliebigen anderen Anwendung (sagen wir ebay) ausfĂŒhrt. Dadurch werden Tests fĂŒr Kernfunktionen generiert. Bei der ersten AusfĂŒhrung auf Anwendungen aus den DomĂ€nen Online Shopping, Nachrichtenseiten und eMail, erzeugt der Prototyp sechzig Prozent der Tests automatisch. Ohne zusĂ€tzlichen manuellen Aufwand. DarĂŒber hinaus interpretiert themen- getriebenes Testen auch vage Anweisungen beispielsweise von How-to Anleitungen oder Anwendungsbeschreibungen. Eine Anweisung wie "FĂŒgen Sie das Produkt in den Warenkorb hinzu" testet das entsprechende Verhalten in der Anwendung. Sowohl im Browser, als auch in einer mobilen Anwendung. Die erzeugten Tests sind robuster und effektiver als vergleichbar erzeugte Tests. Der Prototyp testet die ZielfunktionalitĂ€t fĂŒnf mal schneller und testet dabei Funktionen die durch nicht spezialisierte AnsĂ€tze kaum zu erreichen sind

    Extending the Reach of Fault Localization to Assist in Automated Debugging

    Get PDF
    Software debugging is one of the most time-consuming tasks in modern software maintenance. To assist developers with debugging, researchers have proposed fault localization techniques. These techniques aim to automate the process of locating faults in software, which can greatly reduce debugging time and assist developers in understanding the faults. Effective fault localization is also crucial for automated program repair techniques, as it helps identify potential faulty locations for patching. Despite recent efforts to advance fault localization techniques, their effectiveness is still limited. With the increasing complexity of modern software, fault localization may not always provide direct identification of the root causes of faults. Further, there is a lack of studies on their application in modern software development. Most prior studies have evaluated these techniques in traditional software development settings, where only a single snapshot of the system is considered. However, modern software development often involves continuous and fine-grained changes to the system. This dissertation proposes a series of approaches to explore new automated debugging solutions that can enhance software quality assurance and reliability practices, with a specific focus on extending the reach of fault localization in modern software development. The dissertation begins with an empirical study on user-reported logs in bug reports, revealing that re-constructed execution paths from these logs provide valuable debugging hints. To further assist developers in debugging, we propose using static analysis techniques for information-retrieval and path-guided fault localization. By leveraging execution paths from logs in bug reports, we can improve the effectiveness of fault localization techniques. Second, we investigate the characteristics of operational data in continuous integration that can help capture faults early in the testing phase. As there is currently no available continuous integration benchmark that incorporates continuous test execution and failure, we present T-Evos, a dataset that comprises various operational data in continuous integration settings. We propose automated fault localization techniques that integrate change information from continuous integration settings, and demonstrate that leveraging such fine-grained change information can significantly improve their effectiveness. Finally, the dissertation investigates the data cleanness in fault localization by examining developers' knowledge in fault-triggering tests. The study reveals a significant degradation in the performance of fault localization techniques when evaluated on faults without developer knowledge. Through case studies and experiments, the proposed techniques in this dissertation significantly improve the effectiveness of fault localization and facilitate their adoption in modern software development. Additionally, this dissertation provides valuable insights into new debugging solutions for future research

    Evolução da telefonia na web

    Get PDF
    Mestrado em Sistemas de InformaçãoCom a ameaça imposta Ă s operadoras por aplicaçÔes OTT como WhatsApp ou Skype, diversas iniciativas coordenadas pela GSMA foram criadas para tentar responder a este fenĂłmeno. Paralelamente, com a evolução de tecnologias como HTML5 e WebRTC, novos serviços como o Twilio tĂȘm surgido, oferecendo APIs para o desenvolvimento de novas aplicaçÔes Web. No entanto, a integração destas tecnologias em tradicionais redes de telecomunicaçÔes nĂŁo faz parte das actuais especificaçÔes. Sendo assim, o objectivo desta dissertação consiste na especificação e implementação de um protĂłtipo baseado nestas tecnologias emergentes, integrado com uma rede IMS. Primeiramente, foi feito um estudo do estado de arte, definindo requisitos e casos de uso a serem explorados. De seguida, o desenho da solução foi feito e implementado, tendo sido criado uma plataforma que alia WebRTC e a OneAPI da GSMA (que define funcionalidades bĂĄsicas para operadores), oferecendo interoperabilidade entre ambos os mundos. A solução Ă© composta por um servidor aplicacional que expĂ”e a API e gateway WebRTC, tendo sido testada e considerada adaptada Ă s necessidades estabelecidas.With the threat to operators by OTT applications such as Skype or WhatsApp, several initiatives coordinated by GSMA were created in an effort to respond to this phenomenon. In parallel, with the evolution of technologies such as HTML5 and WebRTC, new services such as Twilio are now available, offering APIs for web application development. However, the integration of these technologies and traditional telecommunication networks is not a part of the current standards. As such, the objective of this dissertation is the specification and implementation of a prototype based on these emerging technologies, integrated in an IMS network. First, a state-of-the-art analysis was made, defining requirements and use-cases to be explored. Secondly, the design and implementation of the solution was done, creating a platform that unites WebRTC and GSMA’s OneAPI (which exposes basic operator features), offering interoperability between both worlds. The solution is composed by an application server that exposes the API and a WebRTC gateway, having been successfully tested and adapted to the established needs

    Timeline design for visualising cultural heritage data

    Get PDF
    This thesis is concerned with the design of data visualisations of digitised museum, archive and library collections, in timelines. As cultural institutions digitise their collections—converting texts, objects, and artworks to electronic records—the volume of cultural data available grows. There is a growing perception, though, that we need to get more out of this data. Merely digitising does not automatically make collections accessible, discoverable and comprehensible, and standard interfaces do not necessarily support the types of interactions users wish to make. Data visualisations—this thesis focuses on interactive visual representations of data created with software—allow us to see an overview of, observe patterns in, and showcase the richness of, digitised collections. Visualisation can support analysis, exploration and presentation of collections for different audiences: research, collection administration, and the general public. The focus here is on visualising cultural data by time: a fundamental dimension for making sense of historical data, but also one with unique strangeness. Through cataloguing, cultural institutions define the meaning and value of items in their collections and the structure within which to make sense of them. By visualising threads in cataloguing data through time, can historical narratives be made visible? And is the data alone enough to tell the stories that people wish to tell? The intended audience for this research is cultural heritage institutions. This work sits at the crossroads between design, cultural heritage (particularly museology), and computing—drawing on the fields of digital humanities, information visualisation and human computer-interaction which also live in these overlapping spaces. This PhD adds clarity around the question of what cultural visualisation is (and can be) for, and highlights issues in the visualisation of qualitative or nominal data. The first chapter lays out the background, characterising cultural data and its visualisation. Chapter two walks through examples of existing cultural timeline visualisations, from the most handcrafted displays to automated approaches. At this point, the research agenda and methodology are set out. The next five chapters document a portfolio of visualisation projects, designing and building novel prototype timeline visualisations with data from the Wellcome Library and Victoria & Albert Museum, London, Cooper Hewitt Smithsonian Design Museum, New York City, and the Nordic Museum, Stockholm. In the process, a range of issues are identified for further discussion. The final chapters reflect on these projects, arguing that automated timeline visualisation can be a productive way to explore and present historical narratives in collection data, but a range of factors govern what is possible and useful. Trust in cultural data visualisation is also discussed. This research argues that visualising cultural data can add value to the data both for users and for data-holding institutions. However, that value is likely to be best achieved by customising a visualisation design to the dataset, audience and use case. Keywords: cultural heritage data; historical data; cultural analytics; cultural informatics; humanities visualisation; generous interfaces; digital humanities; design; information design; interface design; data visualisation; information visualisation; time; timeline; history; historiography; museums; museology; archives; chronographics

    Design and Evaluation of User Interfaces for Mobile Web Search

    Get PDF
    Mobiili tiedonhaku on jatkuvasti kasvava ja monimuotoistuva osa jokapÀivÀistÀ tiedonhankintaa. Aikaisemman tutkimuksen mukaan tarvitaan kuitenkin parempia kÀyttöliittymÀratkaisuja tukemaan mobiililaitteilla tapahtuvaa verkkotiedonhakua. VÀitöskirjatutkimuksessa suunniteltiin ja toteutettiin kaksi uutta hakukÀyttöliittymÀÀ, joita arvioitiin kÀyttÀjÀtutkimuksissa. EnsimmÀinen kÀyttöliittymÀ perustuu siihen, ettÀ hakutulokset luokitellaan ryhmiin niissÀ esiintyvien avainsanojen perusteella. KÀyttÀjÀtutkimusten tulokset osoittavat, ettÀ luokittelulla voidaan tukea mobiilikÀyttÀjien tutkivaa tiedonhakua. Toinen kÀyttöliittymÀ antaa hakutulosten yhteydessÀ yleiskuvan hakulauseen sijaintikohdista tulosdokumenteissa. Vaikkakin menetelmÀn kÀyttö vaatii opettelua, kÀyttÀjÀarviot osoittavat ettÀ se voi auttaa sivuuttamaan huonot hakutulokset, etenkin silloin kun muut hakutulosta kuvaavat tiedot ovat epÀselviÀ. LisÀksi vÀitöskirjassa tutkittiin aktiivisten mobiili-Internetin kÀyttÀjien tiedontarpeita verkkotiedonhaun kÀytön ymmÀrtÀmiseksi. Tutkimustulosten mukaan hakujen tekeminen ja verkon selaaminen ovat nÀiden kÀyttÀjien tÀrkeimpiÀ tiedonhankintatapoja. NiillÀ pyritÀÀn vastaamaan tiedontarpeisiin heti niiden ilmaantuessa, olipa kÀyttÀjÀ sitten kotona, liikkeessÀ tai sosiaalisessa vuorovaikutustilanteessa. Mobiili tiedonhankinta on vahvasti sidoksissa kÀyttötilanteeseen, mikÀ tulee huomioida hakukÀyttöliittymien suunnittelussa. Tulevaisuuden hakukÀyttöliittymÀt voivat esimerkiksi tukea tiedonhankintaa hyödyntÀmÀllÀ tietoa kÀyttÀjÀn sijainnista ja aktiviteeteista. Myös epÀmuodollisten ja tutkivien tiedontarpeiden kasvava rooli asettaa uusia haasteita vuorovaikutuksen suunnittelulle.Mobile Web search is a rapidly growing information seeking activity employed across different locations, situations, and activities. Current mobile search interfaces are based on the ranked result list, dominant in desktop interfaces. Research suggests that new paradigms are needed for better support of mobile searchers. For this dissertation, two such novel search interface techniques were designed, implemented, and evaluated. The first method, a clustering search interface that presents a category- based overview of the results, was studied both in a task-based experiment in a laboratory setting and in a longitudinal field study wherein it was used to address real information needs. The results indicate that clustering can support exploratory search needs when the searcher has trouble defining the information need, requires an overview of the search topic, or is interested in multiple results related to the same topic. The findings informed design guidelines for category-based search interfaces. How and when categorization is presented in the search interface needs to be carefully considered. Categorization methods should be improved, for better response to diverse information needs. Hybrid approaches employing contextually informed clustering, classification, and faceted browsing may offer the best match for user needs. The second presentation method, a visualization of the occurrences of the user s query phrase in a result document, can be incorporated into the ranked result list as an additional, unobtrusive result descriptor. It allows the searcher to see how often the query phrase appears in the result document, enabling the use of various evaluation strategies to assess the relevance of the results. Several iterations of the visualization were studied with users to form an understanding of the potential of this approach. The results suggest that a novel visualization can be useful in ruling out non-relevant results and can assist when the other result descriptors do not provide for a conclusive relevance assessment. However, users familiarity with well-established result descriptors means that users have to learn how to integrate the visualization into their search strategies and reconcile situations in which the visualization is in conflict with other metadata. In addition, the contextual triggers and information behaviors of mobile Internet users were studied, for understanding of the role of Web search as a mobile information seeking activity. The results from this study show that mobile Web search and browsing are important information seeking activities. They are engaged in to resolve emerging information needs as they appear, whether at home, on the go, or in social situations

    An ant-inspired, deniable routing approach in ad hoc question & answer networks

    Get PDF
    The ubiquity of the Internet facilitates electronic question and answering (Q&A) between real people with ease via community portals and social networking websites. It is a useful service which allows users to appeal to a broad range of answerers. In most cases however, Q&A services produce answers by presenting questions to the general public or associated digital community with little regard for the amount of time users spend examining and answering them. Ultimately, a question may receive large amounts of attention but still not be answered adequately. Several existing pieces of research investigate the reasons why questions do not receive answers on Q&A services and suggest that it may be associated with users being afraid of expressing themselves. Q&A works well for solving information needs, however, it rarely takes into account the privacy requirements of the users who form the service. This thesis was motivated by the need for a more targeted approach towards Q&A by distributing the service across ad hoc networks. The main contribution of this thesis is a novel routing technique and networking environment (distributed Q&A) which balances answer quality and user attention while protecting privacy through plausible deniability. Routing approaches are evaluated experimentally by statistics gained from peer-to-peer network simulations, composed of Q&A users modelled via features extracted from the analysis of a large Yahoo! Answers dataset. Suggestions for future directions to this work are presented from the knowledge gained from our results and conclusion
    • 

    corecore