21,651 research outputs found

    On the role of pre and post-processing in environmental data mining

    Get PDF
    The quality of discovered knowledge is highly depending on data quality. Unfortunately real data use to contain noise, uncertainty, errors, redundancies or even irrelevant information. The more complex is the reality to be analyzed, the higher the risk of getting low quality data. Knowledge Discovery from Databases (KDD) offers a global framework to prepare data in the right form to perform correct analyses. On the other hand, the quality of decisions taken upon KDD results, depend not only on the quality of the results themselves, but on the capacity of the system to communicate those results in an understandable form. Environmental systems are particularly complex and environmental users particularly require clarity in their results. In this paper some details about how this can be achieved are provided. The role of the pre and post processing in the whole process of Knowledge Discovery in environmental systems is discussed

    Researching grammar learning strategies: Combining the macro- and micro-perspective

    Get PDF
    Udostępnienie publikacji Wydawnictwa Uniwersytetu Łódzkiego finansowane w ramach projektu „Doskonałość naukowa kluczem do doskonałości kształcenia”. Projekt realizowany jest ze środków Europejskiego Funduszu Społecznego w ramach Programu Operacyjnego Wiedza Edukacja Rozwój; nr umowy: POWER.03.05.00-00-Z092/17-00

    A National Dialogue on Health Information Technology and Privacy

    Get PDF
    Increasingly, government leaders recognize that solving the complex problems facing America today will require more than simply keeping citizens informed. Meeting challenges like rising health care costs, climate change and energy independence requires increased level of collaboration. Traditionally, government agencies have operated in silos -- separated not only from citizens, but from each other, as well. Nevertheless, some have begun to reach across and outside of government to access the collective brainpower of organizations, stakeholders and individuals.The National Dialogue on Health Information Technology and Privacy was one such initiative. It was conceived by leaders in government who sought to demonstrate that it is not only possible, but beneficial and economical, to engage openly and broadly on an issue that is both national in scope and deeply relevant to the everyday lives of citizens. The results of this first-of-its-kind online event are captured in this report, together with important lessons learned along the way.This report served as a call to action. On his first full day in office, President Obama put government on notice that this new, more collaborative model can no longer be confined to the efforts of early adopters. He called upon every executive department and agency to "harness new technology" and make government "transparent, participatory, and collaborative." Government is quickly transitioning to a new generation of managers and leaders, for whom online collaboration is not a new frontier but a fact of everyday life. We owe it to them -- and the citizens we serve -- to recognize and embrace the myriad tools available to fulfill the promise of good government in the 21st Century.Key FindingsThe Panel recommended that the Administration give stakeholders the opportunity to further participate in the discussion of heath IT and privacy through broader outreach and by helping the public to understand the value of a person-centered view of healthcare information technology

    Saggitarius: A DSL for Specifying Grammatical Domains

    Full text link
    Common data types like dates, addresses, phone numbers and tables can have multiple textual representations, and many heavily-used languages, such as SQL, come in several dialects. These variations can cause data to be misinterpreted, leading to silent data corruption, failure of data processing systems, or even security vulnerabilities. Saggitarius is a new language and system designed to help programmers reason about the format of data, by describing grammatical domains -- that is, sets of context-free grammars that describe the many possible representations of a datatype. We describe the design of Saggitarius via example and provide a relational semantics. We show how Saggitarius may be used to analyze a data set: given example data, it uses an algorithm based on semi-ring parsing and MaxSAT to infer which grammar in a given domain best matches that data. We evaluate the effectiveness of the algorithm on a benchmark suite of 110 example problems, and we demonstrate that our system typically returns a satisfying grammar within a few seconds with only a small number of examples. We also delve deeper into a more extensive case study on using Saggitarius for CSV dialect detection. Despite being general-purpose, we find that Saggitarius offers comparable results to hand-tuned, specialized tools; in the case of CSV, it infers grammars for 84% of benchmarks within 60 seconds, and has comparable accuracy to custom-built dialect detection tools.Comment: OOPSLA 202

    Semiotic Machine

    Get PDF

    Towards a Critical Understanding of Deepfakes: Developing a Teaching Module and More

    Get PDF
    Recently, computer-generated and computer-altered videos known as deepfakes have raised widespread concerns about the harms they may cause to democratic elections, national security, people’s reputation, and people’s autonomy over their words and actions as represented in videos and other media. How can we build towards a critical understanding of not only deepfakes, but also photos, videos, and the role of many other media objects surrounding us that inform us about the world? In this thesis, wanting to take a historical approach and noting the newness of deepfakes, I first investigate a historical case study regarding a manipulated photo from a 1950 U.S. Senate election campaign. Examining hearings conducted by the Senate into the use of misleading media in the election, I investigate how the incident sparked a debate between different groups of people over the trustworthiness of photographs and their proper role in elections. Next, I move forward in time and discuss the nature of deepfakes, presenting a brief history focusing on the different communities—academic, hobbyist, and commercial—that have played a role in the development of different, but related, technologies that all fall under the umbrella term of deepfakes. Some of this history is incorporated into the third part of this thesis, in which I present a teaching module I developed with the goals of guiding students to think critically about photos and videos and of raising awareness about deepfakes
    corecore