39 research outputs found

    Efficient Prior Publication Identification for Open Source Code

    Full text link
    Free/Open Source Software (FOSS) enables large-scale reuse of preexisting software components. The main drawback is increased complexity in software supply chain management. A common approach to tame such complexity is automated open source compliance, which consists in automating the verication of adherence to various open source management best practices about license obligation fulllment, vulnerability tracking, software composition analysis, and nearby concerns.We consider the problem of auditing a source code base to determine which of its parts have been published before, which is an important building block of automated open source compliance toolchains. Indeed, if source code allegedly developed in house is recognized as having been previously published elsewhere, alerts should be raised to investigate where it comes from and whether this entails that additional obligations shall be fullled before product shipment.We propose an ecient approach for prior publication identication that relies on a knowledge base of known source code artifacts linked together in a global Merkle direct acyclic graph and a dedicated discovery protocol. We introduce swh-scanner, a source code scanner that realizes the proposed approach in practice using as knowledge base Software Heritage, the largest public archive of source code artifacts. We validate experimentally the proposed approach, showing its eciency in both abstract (number of queries) and concrete terms (wall-clock time), performing benchmarks on 16 845 real-world public code bases of various sizes, from small to very large

    Data Infrastructures in Ecology: An Infrastructure Studies Perspective

    Get PDF
    The development of information infrastructures that make ecological research data avail¬able has increased in recent years, contributing to fundamental changes in ecological re¬search. Science and Technology Studies (STS) and the subfield of Infrastructure Studies, which aims at informing infrastructures’ design, use, and maintenance from a social science point of view, provide conceptual tools for understanding data infrastructures in ecology. This perspective moves away from the language of engineering, with its discourse on physical structures and systems, to use a lexicon more “social” than “technical” to understand data infrastructures in their informational, sociological, and historical dimensions. It takes a holistic approach that addresses not only the needs of ecological research but also the diversity and dynamics of data, data work, and data management. STS research, having focused for some time on studying scientific practices, digital devices, and information systems, is expanding to investigate new kinds of data infrastructures and their interdependencies across the data landscape. In ecology, data sharing and data infrastructures create new responsibilities that require scientists to engage in opportunities to plan, experiment, learn, and reshape data arrangements. STS and Infrastructure Studies scholars are suggesting that ecologists as well as data specialists and social scientists would benefit from active partnerships to ensure the growth of data infrastructures that effectively support scientific investigative processes in the digital era.Ope

    Personal Data Stores (PDS): A Review

    Get PDF
    Internet services have collected our personal data since their inception. In the beginning, the personal data collection was uncoordinated and was limited to a few selected data types such as names, ages, birthdays, etc. Due to the widespread use of social media, more and more personal data has been collected by different online services. We increasingly see that Internet of Things (IoT) devices are also being adopted by consumers, making it possible for companies to capture personal data (including very sensitive data) with much less effort and autonomously at a very low cost. Current systems architectures aim to collect, store, and process our personal data in the cloud with very limited control when it comes to giving back to citizens. However, Personal Data Stores (PDS) have been proposed as an alternative architecture where personal data will be stored within households, giving us complete control (self-sovereignty) over our data. This paper surveys the current literature on Personal Data Stores (PDS) that enable individuals to collect, control, store, and manage their data. In particular, we provide a comprehensive review of related concepts and the expected benefits of PDS platforms. Further, we compare and analyse existing PDS platforms in terms of their capabilities and core components. Subsequently, we summarise the major challenges and issues facing PDS platforms’ development and widespread adoption

    A Graduatte Level Immersive-Simulattion Program for Teaching and Assessing Fundamental Skills in Entry Level Clinical Perfusionists.

    Get PDF
    Background: The clinical perfusionist is a member of the open-heart-surgery team and responsible for operating the life support equipment that replaces the function of the patient\u27s heart and lungs and arrests and restarts the patient\u27s heart in the course of a Cardiopulmonary Bypass (CPB) procedure. In the perfusionists scope of practice, the consequence of unskilled actions, inaccurate understanding or delayed decision making may result in significant patient morbidity or even death. Historically, perfusion students have learned and practiced their skills within a clinical preceptorship program in which an experienced clinician allows the novice student to operate the life support equipment under their direct supervision and consultation. While there is clinical evidence from numerous surgical specialties which establishes that learning curve associated errors have a negative effect on patient outcomes, this has not been researched for clinical perfusionists. Despite this evidence gap, the professions leaders have been instrumental in driving educational innovation and the development of medical simulation models that may reduce the patient\u27s exposure to learning curve associated morbidity by developing competence with high-risk clinical skills prior to patient contact. The purpose of this research is to develop, validate and apply novel medical simulation techniques and technologies to the preparation of entry level clinical perfusionists and demonstrate pre-clinical competence with the fundamental perfusion skills.Methods and Results: To inform the development of a skills curriculum we conducted two national surveys using online survey tools. Through these surveys we validated a list of fundamental skills, and the deconstructed sub-elements involved in the conduct of these skills. Additionally, we identified the typical ranges of physiologic and technical parameters that clinicians maintain during clinical procedures. With this foundational benchmark data we validated the performance of a simulated patient to establish that the patient surrogate generates data that is substantially similar to the physiologic and technical data that a perfusionist would manage during a live clinical procedure. This validated simulation technology was then incorporated into a high-fidelity simulation suite and applied to an innovative immersive curriculum which included hands on repetitive practice, live and video supported self, peer and expert observation and feedback as well as a battery of high-stakes assessments. The validity and fidelity of the simulated experience was established through analysis of over 800 opinions generated over 10 years by novice and expert perfusionists after performing simulated cases. Finally, the efficacy of the simulation curriculum was assessed by comparing our simulation trained students to a national pool of their peers from other schools and expert clinicians. Through this process we generated the first measurements of the typical learning curve for the fundamental skills of CPB, the first estimates of error rates for students navigating the learning curve and the first benchmark measures of competent performance in a simulated environment. This data establishes that students learning in traditional clinical training programs conduct three-fold more errors than experts and will have approximately 99 high-risk patient encounters prior to developing competence with fundamental skills. By comparison, simulation trained students demonstrated competence with fundamental skills that was similar to the experts with almost no high-risk patient encounters. Discussion: The implications to patient safety are clearly implied. These studies establish that there is a high level of agreement among clinicians regarding the skills that are necessary to operate perfusion equipment and that realistic simulation environments can be designed and applied to the development of student\u27s fundamental perfusion skills without exposing patients to the threat of students learning curve associated errors. This data may catalyze a larger national dialog regarding Entrustable Professional Activities for perfusionists and influence national accreditation standards for educational programs

    A Review of Age Friendly Virtual Assistive Technologies and their Effect on Daily Living for Carers and Dependent Adults

    Get PDF
    Many barriers exist in the lives of older adult’s, including health, transport, housing, isolation, disability and access to technology. The appropriate integration of technology within age-friendly communities continues to offer possible solutions to these barriers and challenges. Older adults and disabled people continue to be affected and marginalized due to lack of access to the digital world. Working collaboratively with planners, policy makers and developers, social and living spaces in the future will ensure that residents are equipped to live in an era that continues to be led by, and is dependent upon, access to technology. This review paper uniquely draws together the small volume of literature from the fields of gerontology, gerontechnology, human computer interaction (HCI), and disability. This paper examines the national and international age-friendly frameworks regarding older adults who are carers of dependent people with disabilities

    Self-branding strategies of online freelancers on Upwork

    Get PDF
    Self-branding is crucial for online freelancers as they must constantly differentiate themselves from competitors on online labor platforms to ensure a viable stream of income. By analyzing 39 interviews with freelancers and clients on the online labor platform Upwork, we identify five key self-branding strategies: boosting a profile, showcasing skills, expanding presence, maintaining relationships with clients, and individualizing brand. These self-branding strategies are contextualized within Goffman?s dramaturgical theory and through an affordances lens, showing immanent tensions. While online freelancers successfully leverage self-branding to improve their visibility on Upwork and beyond, the client perspective reveals a fine line between too little and too much self-branding. Online freelancers must brand themselves in visibility games when the game rules are largely opaque, riddled with uncertainty, and constantly evolving. We connect the findings to adjacent platform economy research and derive a self-branding as a performance framework

    The Software Heritage License Dataset (2022 Edition)

    Full text link
    Context: When software is released publicly, it is common to include with it either the full text of the license or licenses under which it is published, or a detailed reference to them. Therefore public licenses, including FOSS (free, open source software) licenses, are usually publicly available in source code repositories.Objective: To compile a dataset containing as many documents as possible that contain the text of software licenses, or references to the license terms. Once compiled, characterize the dataset so that it can be used for further research, or practical purposes related to license analysis.Method: Retrieve from Software Heritage-the largest publicly available archive of FOSS source code-all versions of all files whose names are commonly used to convey licensing terms. All retrieved documents will be characterized in various ways, using automated and manual analyses.Results: The dataset consists of 6.9 million unique license files. Additional metadata about shipped license files is also provided, making the dataset ready to use in various contexts, including: file length measures, MIME type, SPDX license (detected using ScanCode), and oldest appearance. The results of a manual analysis of 8102 documents is also included, providing a ground truth for further analysis. The dataset is released as open data as an archive file containing all deduplicated license files, plus several portable CSV files with metadata, referencing files via cryptographic checksums.Conclusions: Thanks to the extensive coverage of Software Heritage, the dataset presented in this paper covers a very large fraction of all software licenses for public code. We have assembled a large body of software licenses, characterized it quantitatively and qualitatively, and validated that it is mostly composed of licensing information and includes almost all known license texts. The dataset can be used to conduct empirical studies on open source licensing, training of automated license classifiers, natural language processing (NLP) analyses of legal texts, as well as historical and phylogenetic studies on FOSS licensing. It can also be used in practice to improve tools detecting licenses in source code

    Applying Wikipedia to Interactive Information Retrieval

    Get PDF
    There are many opportunities to improve the interactivity of information retrieval systems beyond the ubiquitous search box. One idea is to use knowledge bases—e.g. controlled vocabularies, classification schemes, thesauri and ontologies—to organize, describe and navigate the information space. These resources are popular in libraries and specialist collections, but have proven too expensive and narrow to be applied to everyday webscale search. Wikipedia has the potential to bring structured knowledge into more widespread use. This online, collaboratively generated encyclopaedia is one of the largest and most consulted reference works in existence. It is broader, deeper and more agile than the knowledge bases put forward to assist retrieval in the past. Rendering this resource machine-readable is a challenging task that has captured the interest of many researchers. Many see it as a key step required to break the knowledge acquisition bottleneck that crippled previous efforts. This thesis claims that the roadblock can be sidestepped: Wikipedia can be applied effectively to open-domain information retrieval with minimal natural language processing or information extraction. The key is to focus on gathering and applying human-readable rather than machine-readable knowledge. To demonstrate this claim, the thesis tackles three separate problems: extracting knowledge from Wikipedia; connecting it to textual documents; and applying it to the retrieval process. First, we demonstrate that a large thesaurus-like structure can be obtained directly from Wikipedia, and that accurate measures of semantic relatedness can be efficiently mined from it. Second, we show that Wikipedia provides the necessary features and training data for existing data mining techniques to accurately detect and disambiguate topics when they are mentioned in plain text. Third, we provide two systems and user studies that demonstrate the utility of the Wikipedia-derived knowledge base for interactive information retrieval

    Postmemorial Exhibitions: A Design Approach to Negotiate Cultural Trauma Among Children and Grandchildren of Former War Refugees and Guerrilla Supporters

    Full text link
    The fields of social memory and museum studies share a similar concern for intergenerational dynamics of memory, although the motivations and emphasis of study in each field is different. In museum studies, memorial museums have been noted to promote emotional engagement and critical reflection about mass atrocities. Early evidence from participatory exhibitions in other types of museums suggests that participatory approaches to exhibition design may support the desired outcomes of reflection and emotional impact in memorial museums. However, no studies to date have explored the role of participatory design in shaping communities’ memories about the violent past. In addition, the field of social memory studies has focused on understanding the nature of social memories that result from mass atrocities. Some scholars refer to this type of memory as cultural trauma. Cultural trauma is a socially relevant phenomenon because it alters group identity and may either lead to a betterment of living conditions for marginalized populations or their increased oppression depending on how societies negotiate the trauma. Importantly, cultural trauma has been noted to be a long-term phenomenon that affects different generations differently. Unlike psychological, individual trauma, cultural trauma is never solved, but socially negotiated. Moreover, the strategies through which a given generation negotiates cultural trauma may not prove ineffective or irrelevant for succeeding generations. Considering the complex and pervasive nature of this phenomenon, more work is needed to intervene in cultural trauma negotiations that does not attempt to ‘solve’ such traumas but supports its negotiation. My dissertation seizes on the unique opportunity of both academic traditions. Merging the insights from studies about studies about cultural trauma and memorial museums, I investigate how participation in the design process of a memorial exhibition shaped the memories of children who were born a decade after the end of the civil war of El Salvador (1980-1992), a historical episode characterized by atrocities perpetrated primarily against civilians. My research combined ethnographic methods such as participant observation and semi-structured interviews with the implementation of a participatory design project in El Salvador. During the eight months of my fieldwork, I worked with seven children and grandchildren of former guerrilla combatants and war refugees to produce an exhibition concept about their deeply held beliefs and feelings about the war, which they only have experienced through the stories shared with them by friends and family. I refer to my participants’ exhibition concept as a postmemorial exhibition. At the theoretical level, the concept of a postmemorial exhibition represents an opportunity to understand the impact of memorial exhibitions in social rather than individual frames of memory. Conversely, at the methodological level, the combined approach of ethnographic methods and participatory design in museums extends the toolkit of scholars of social memory to not only observe but intervene in the field. At stake in these interventions is the opportunity to promote more sympathetic and critical subjects through the process of postmemorial exhibition design, and thus, to contribute to the prevention of future episodes of mass atrocities.PHDInformationUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/162997/1/allanmar_1.pd

    Lawrence, Spring 2021

    Get PDF
    https://lux.lawrence.edu/alumni_magazines/1116/thumbnail.jp
    corecore