147,733 research outputs found

    An overview of process model quality literature - The Comprehensive Process Model Quality Framework

    Get PDF
    The rising interest in the construction and the quality of (business) process models resulted in an abundancy of emerged research studies and different findings about process model quality. The lack of overview and the lack of consensus hinder the development of the research field. The research objective is to collect, analyse, structure, and integrate the existing knowledge in a comprehensive framework that strives to find a balance between completeness and relevance without hindering the overview. The Systematic Literature Review methodology was applied to collect the relevant studies. Because several studies exist that each partially addresses this research objective, the review was performed at a tertiary level. Based on a critical analysis of the collected papers, a comprehensive, but structured overview of the state of the art in the field was composed. The existing academic knowledge about process model quality was carefully integrated and structured into the Comprehensive Process Model Quality Framework (CPMQF). The framework summarizes 39 quality dimensions, 21 quality metrics, 28 quality (sub)drivers, 44 (sub)driver metrics, 64 realization initiatives and 15 concrete process model purposes related to 4 types of organizational benefits, as well as the relations between all of these. This overview is thus considered to form a valuable instrument for both researchers and practitioners that are concerned about process model quality. The framework is the first to address the concept of process model quality in such a comprehensive way

    On the use of scaling relations for the Tolman test

    Get PDF
    The use of relations between structural parameters of early type galaxies to perform the Tolman test is reconsidered. Scaling relations such as the FP or the Kormendy relation, require the transformation from angular to metric sizes, to compare the relation at different z values. This transformation depends on the assumed world model: galaxies of a given angular size, at a given z, are larger (in kpc) in a non-expanding universe than in an expanding one. Furthermore, the luminosities of galaxies are expected to evolve with z in an expanding model. These effects are shown to conspire to reduce the difference between the predicted SB change with redshift in the expanding and non expanding cases. We find that the predictions for the visible photometric bands of the expanding models with passive luminosity evolution are very similar to those of the static model till z about 1, and therefore, the test cannot distinguish between the two world models. Recent good quality data are consistent with the predictions from both models. In the K-band, where the expected (model) luminosity evolutionary corrections are smaller, the differences between the xpanding and static models amount to about 0.4 (0.8) magnitudes at z = 0.4 (1). It is shown that, due to that small difference between the predictions in the covered z-range, and to the paucity and uncertainties of the relevant SB photometry, the existing K-band data is not adequate to distinguish between the different world metrics, and cannot be yet used to discard the static case. It is pointed out that the scaling relations could still be used to rule out the non-evolving case if it could be shown that the coefficients change with the redshift.Comment: Latex, 15 pages with 2 figures. To be published in ApJ Letter

    Estimating Accuracy of Personal Identifiable Information in Integrated Data Systems

    Get PDF
    Without a valid assessment of accuracy there is a risk of data users coming to incorrect conclusions or making bad decision based on inaccurate data. This dissertation proposes a theoretical method for developing data-accuracy metrics specific for any given person-centric integrated system and how a data analyst can use these metrics to estimate the overall accuracy of person-centric data. Estimating the accuracy of Personal Identifiable Information (PII) creates a corresponding need to model and formalize PII for both the real-world and electronic data, in a way that supports rigorous reasoning relative to real-world facts, expert opinions, and aggregate knowledge. This research provides such a foundation by introducing a temporal first-order logic language (FOL), called Person Data First-order Logic (PDFOL). With its syntax and semantics formalized, PDFOL provides a mechanism for expressing data- accuracy metrics, computing measurements using these metrics on person-centric databases, and comparing those measurements with expected values from real-world populations. Specifically, it enables data analysts to model person attributes and inter-person relations from real-world population or database representations of such, as well as real-world facts, expert opinions, and aggregate knowledge. PDFOL builds on existing first-order logics with the addition of temporal predicated based on time intervals, aggregate functions, and tuple-set comparison operators. It adapts and extends the traditional aggregate functions in three ways: a) allowing any arbitrary number free variables in function statement, b) adding groupings, and c) defining new aggregate function. These features allow PDFOL to model person-centric databases, enabling formal and efficient reason about their accuracy. This dissertation also explains how data analysts can use PDFOL statements to formalize and develop formal accuracy metrics specific to a person-centric database, especially if it is an integrated person- centric database, which in turn can then be used to assess the accuracy of a database. Data analysts apply these metrics to person-centric data to compute the quality-assessment measurements, YD. After that, they use statistical methods to compare these measurements with the real-world measurements, YR. Compare YD and YR with the hypothesis that they should be very similar, if the person-centric data is an accurate and complete representations of the real-world population. Finally, I show that estimated accuracy using metrics based on PDFOL can be good predictors of database accuracy. Specifically, I evaluated the performance of selected accuracy metrics by applying them to a person-centric database, mutating the database in various ways to degrade its accuracy, and the re-apply the metrics to see if they reflect the expected degradation. This research will help data analyst to develop an accuracy metrics specific to their person-centric data. In addition, PDFOL can provide a foundation for future methods for reasoning about other quality dimensions of PII

    Metrics for Measuring Data Quality - Foundations for an Economic Oriented Management of Data Quality

    Get PDF
    The article develops metrics for an economic oriented management of data quality. Two data quality dimensions are focussed: consistency and timeliness. For deriving adequate metrics several requirements are stated (e. g. normalisation, cardinality, adaptivity, interpretability). Then the authors discuss existing approaches for measuring data quality and illustrate their weaknesses. Based upon these considerations, new metrics are developed for the data quality dimensions consistency and timeliness. These metrics are applied in practice and the results are illustrated in the case of a major German mobile services provider

    Discourse Structure in Machine Translation Evaluation

    Full text link
    In this article, we explore the potential of using sentence-level discourse structure for machine translation evaluation. We first design discourse-aware similarity measures, which use all-subtree kernels to compare discourse parse trees in accordance with the Rhetorical Structure Theory (RST). Then, we show that a simple linear combination with these measures can help improve various existing machine translation evaluation metrics regarding correlation with human judgments both at the segment- and at the system-level. This suggests that discourse information is complementary to the information used by many of the existing evaluation metrics, and thus it could be taken into account when developing richer evaluation metrics, such as the WMT-14 winning combined metric DiscoTKparty. We also provide a detailed analysis of the relevance of various discourse elements and relations from the RST parse trees for machine translation evaluation. In particular we show that: (i) all aspects of the RST tree are relevant, (ii) nuclearity is more useful than relation type, and (iii) the similarity of the translation RST tree to the reference tree is positively correlated with translation quality.Comment: machine translation, machine translation evaluation, discourse analysis. Computational Linguistics, 201

    Drag it together with Groupie: making RDF data authoring easy and fun for anyone

    No full text
    One of the foremost challenges towards realizing a “Read-write Web of Data” [3] is making it possible for everyday computer users to easily find, manipulate, create, and publish data back to the Web so that it can be made available for others to use. However, many aspects of Linked Data make authoring and manipulation difficult for “normal” (ie non-coder) end-users. First, data can be high-dimensional, having arbitrary many properties per “instance”, and interlinked to arbitrary many other instances in a many different ways. Second, collections of Linked Data tend to be vastly more heterogeneous than in typical structured databases, where instances are kept in uniform collections (e.g., database tables). Third, while highly flexible, the problem of having all structures reduced as a graph is verbosity: even simple structures can appear complex. Finally, many of the concepts involved in linked data authoring - for example, terms used to define ontologies are highly abstract and foreign to regular citizen-users.To counter this complexity we have devised a drag-and-drop direct manipulation interface that makes authoring Linked Data easy, fun, and accessible to a wide audience. Groupie allows users to author data simply by dragging blobs representing entities into other entities to compose relationships, establishing one relational link at a time. Since the underlying representation is RDF, Groupie facilitates the inclusion of references to entities and properties defined elsewhere on the Web through integration with popular Linked Data indexing services. Finally, to make it easy for new users to build upon others’ work, Groupie provides a communal space where all data sets created by users can be shared, cloned and modified, allowing individual users to help each other model complex domains thereby leveraging collective intelligence

    Broad-Scale Relations between Conservation Reserve Program and Grassland Birds: Do Cover Type, Configuration and Contract Age Matter?

    Get PDF
    The Conservation Reserve Program (CRP) is a voluntary cropland set-aside program where environmentally-sensitive cropland is retired to a conservation practice. Grassland birds should benefit because most CRP is grass habitat and because amount of land in CRP is highest in agriculture-dominated areas of the United States where grassland habitat has been most impacted. We used the Breeding Bird Survey and Common Land Unit (CLU) data (spatially-explicit data of farm field boundaries and land cover) to identify relations between types and configurations of CRP and grassland bird abundance in 3 Midwestern states. All 13 species we studied were related to at least one aspect of CRP habitat - specific conservation practices (e.g., native vs. exotic grass), CRP habitat configuration, or habitat age. Treating all types of CRP as a single habitat type would have obscured bird-CRP relations. Based on our results, creating a mosaic of large and small set-aside patches could benefit both area-sensitive and edge-associated grassland birds. Additionally, northern bobwhite and other birds that use early successional grasslands would benefit from periodic disturbances. CRP, agrienvironment schemes, and other government-sponsored set-aside programs may be most successful when administered as part of a targeted, regional conservation plan

    Assessing architectural evolution: A case study

    Get PDF
    This is the post-print version of the Article. The official published can be accessed from the link below - Copyright @ 2011 SpringerThis paper proposes to use a historical perspective on generic laws, principles, and guidelines, like Lehman’s software evolution laws and Martin’s design principles, in order to achieve a multi-faceted process and structural assessment of a system’s architectural evolution. We present a simple structural model with associated historical metrics and visualizations that could form part of an architect’s dashboard. We perform such an assessment for the Eclipse SDK, as a case study of a large, complex, and long-lived system for which sustained effective architectural evolution is paramount. The twofold aim of checking generic principles on a well-know system is, on the one hand, to see whether there are certain lessons that could be learned for best practice of architectural evolution, and on the other hand to get more insights about the applicability of such principles. We find that while the Eclipse SDK does follow several of the laws and principles, there are some deviations, and we discuss areas of architectural improvement and limitations of the assessment approach
    • 

    corecore