11 research outputs found

    The applicability of a use value-based file retention method

    Get PDF
    The determination of the relative value of files is important for an organization while determining a retrieval service level for its files and a corresponding file retention policy. This paper discusses via a literature review methods for developing file retention policies based on the use values of files. On basis of these results we propose an enhanced version of one of them. In a case study, we demonstrate how one can develop a customized file retention policy by testing causal relations between file parameters and the use value of files. This case shows that, contrary to suggestions of previous research, the file type has no significant relation with the value of a file and thus should be excluded from a retention policy in this case. The case study also shows a strong relation between the position of a file user and the value of this file. Furthermore, we have improved the Information Value Questionnaire (IVQ) for subjective valuation of files. However, the resulting method needs software to be efficient in its application. Therefore, we developed a prototype for the automatic execution of a file retention policy. We conclude with a discussio

    StackInsights: Cognitive Learning for Hybrid Cloud Readiness

    Full text link
    Hybrid cloud is an integrated cloud computing environment utilizing a mix of public cloud, private cloud, and on-premise traditional IT infrastructures. Workload awareness, defined as a detailed full range understanding of each individual workload, is essential in implementing the hybrid cloud. While it is critical to perform an accurate analysis to determine which workloads are appropriate for on-premise deployment versus which workloads can be migrated to a cloud off-premise, the assessment is mainly performed by rule or policy based approaches. In this paper, we introduce StackInsights, a novel cognitive system to automatically analyze and predict the cloud readiness of workloads for an enterprise. Our system harnesses the critical metrics across the entire stack: 1) infrastructure metrics, 2) data relevance metrics, and 3) application taxonomy, to identify workloads that have characteristics of a) low sensitivity with respect to business security, criticality and compliance, and b) low response time requirements and access patterns. Since the capture of the data relevance metrics involves an intrusive and in-depth scanning of the content of storage objects, a machine learning model is applied to perform the business relevance classification by learning from the meta level metrics harnessed across stack. In contrast to traditional methods, StackInsights significantly reduces the total time for hybrid cloud readiness assessment by orders of magnitude

    Cognition-Based Networks: A New Perspective on Network Optimization Using Learning and Distributed Intelligence

    Get PDF
    IEEE Access Volume 3, 2015, Article number 7217798, Pages 1512-1530 Open Access Cognition-based networks: A new perspective on network optimization using learning and distributed intelligence (Article) Zorzi, M.a , Zanella, A.a, Testolin, A.b, De Filippo De Grazia, M.b, Zorzi, M.bc a Department of Information Engineering, University of Padua, Padua, Italy b Department of General Psychology, University of Padua, Padua, Italy c IRCCS San Camillo Foundation, Venice-Lido, Italy View additional affiliations View references (107) Abstract In response to the new challenges in the design and operation of communication networks, and taking inspiration from how living beings deal with complexity and scalability, in this paper we introduce an innovative system concept called COgnition-BAsed NETworkS (COBANETS). The proposed approach develops around the systematic application of advanced machine learning techniques and, in particular, unsupervised deep learning and probabilistic generative models for system-wide learning, modeling, optimization, and data representation. Moreover, in COBANETS, we propose to combine this learning architecture with the emerging network virtualization paradigms, which make it possible to actuate automatic optimization and reconfiguration strategies at the system level, thus fully unleashing the potential of the learning approach. Compared with the past and current research efforts in this area, the technical approach outlined in this paper is deeply interdisciplinary and more comprehensive, calling for the synergic combination of expertise of computer scientists, communications and networking engineers, and cognitive scientists, with the ultimate aim of breaking new ground through a profound rethinking of how the modern understanding of cognition can be used in the management and optimization of telecommunication network

    Evaluating the Applicability of a Use Value-Based File Retention Method

    Get PDF
    A well constructed file retention policy can help a company determine the relative value and the corresponding retrieval service level of the different files it owns. Though such a retention policy is useful, the method one can use to arrive at such a policy is under-researched. This paper discusses how one can arrive at a method (based on a systematic literature review) for developing file retention policies based on use values of files. In the case study, we demonstrate how one can develop a file retention policy by testing of causal relations between file retention policy parameters and the use value of files. This case study shows that, contrary to suggestions of previous research, the file type has no significant causal relation with the value of a file and thus should be excluded from a retention policy in this case. The case study also shows that there is a strong causal relation between the position of a user of a file and the value of this file. Furthermore, we have amended an existing subjective file valuation method, namely, the Information Value Questionnaire (IVQ). However, to make file retention methods effective and reliable a substantially more case experiences need to be collected

    Wirtschaftlichkeitsanalyse der automatisierten Verwaltung unstrukturierter Daten im Information Lifecycle Management

    Get PDF
    Ziel von Information Lifecycle Management (ILM) ist die Klassifizierung und kostengünstige Verwaltung von Informationen. Informationen, auf die in einem Unternehmen oft zugegriffen wird, können von Informationen mit geringerer Zugriffshäufigkeit getrennt und den Unternehmensanforderungen entsprechend auf dem jeweils sinnvollsten Speichermedium bereitgestellt und verwaltet werden. Das rasche Datenwachstum, hohe Speicher-, Administrations- und Betriebskosten sowie zahlreiche rechtliche Anforderungen sind die wesentlichen Gründe für die Entstehung von ILM. Zur Analyse des Kosten- und Nutzenverhältnisses, das bei der Implementierung und Anwendung von ILM entsteht, wird in dieser Arbeit ein Verfahren für die Wirtschaftlichkeitsanalyse zur Verwaltung unstrukturierter Daten im ILM konstruiert, implementiert, demonstriert und evaluiert. Es werden Klassifizierungs-, Verlagerungs- und Kostenfunktionen implementiert, um die Entstehung von Kosten und Kostensenkungspotentialen für verschiedene Nutzungsgrade von Informationen zu simulieren und auszuwerten. Es wird untersucht, welchen Einfluss die automatisierte Klassifizierung und Verwaltung unterschiedlich großer Datenmengen mit variierenden Zugriffshäufigkeiten und Nutzungsgraden auf die Wirtschaftlichkeit des ILM hat. Neben unternehmensinternen Speichermedien wird die Verwaltung von Daten im Cloud Computing in die Betrachtungen einbezogen. Die vorliegende Arbeit liefert eine Literaturanalyse zu den Konzepten ILM und Cloud Computing sowie einen wissenschaftlichen Beitrag dazu, sowohl rechtliche als auch funktionale und technische Anforderungen an ILM und Cloud Computing zu erörtern. Beendet wird die Arbeit durch eine Zusammenfassung sowie eine kritische Würdigung der Ergebnisse. Ein Ausblick gibt Hinweise und Verbesserungsvorschläge hinsichtlich der Wirtschaftlichkeitsanalyse unstrukturierter Daten im Information Lifecycle Management.The aim of Information Lifecycle Management (ILM) is the classification and cost-effective management of information. Highly accessed information should be separated from information with low access rates with the goal to use enterprise storage according to usage and business needs. The rapid growth of information, increasing storage-, administration- and operating costs as well as legal requirements are the main reasons for the emergence of ILM. To analyze the cost-benefit ratio resulting from implementing and running ILM, a method for the economic analysis for managing unstructured data in ILM is designed, implemented, demonstrated and evaluated. The author is implementing classification, displacement and cost functions to simulate the formation of costs and cost reduction potentials for different access rates and lifecycles of information. Automated classification is used for different volumes of data with varying access frequencies and degrees of utilization. Cloud Computing is included in the considerations as a way to store and manage data in addition to internal enterprise storage. The present study provides a literature review on ILM concepts and cloud computing. Legal, functional and technical requirements for ILM and Cloud Computing are discussed. A summary and a critical assessment of the results is finishing the work. A lookout gives hints and suggestions for improving the economic analysis of unstructured data in Information Lifecycle Management

    File Classification in Self-* Storage Systems

    No full text
    To tune and manage themselves, file and storage systems must understand key properties (e.g., access pattern, lifetime, size) of their various files. This paper describes how systems can automatically learn to classify the properties of files (e.g., read-only access pattern, short-lived, small in size) and predict the properties of new files, as they are created, by exploiting the strong associations between a file's properties and the names and attributes assigned to it. These associations exist, strongly but differently, in each of four real NFS environments studied. Decision tree classifiers can automatically identify and model such associations, providing prediction accuracies that often exceed 90%. Such predictions can be used to select storage policies (e.g., disk allocation schemes and replication factors) for individual files. Further, changes in associations can expose information about applications, helping autonomic system components distinguish growth from fundamental change

    File Classification in Self-* Storage Systems (CMU-PDL-04-101)

    No full text
    To tune and manage themselves, file and storage systems must understand key properties (e.g., access pattern, lifetime, size) of their various files. This paper describes how systems can automatically learn to classify the properties of files (e.g., read-only access pattern, short-lived, small in size) and predict the properties of new files, as they are created, by exploiting the strong associations between a file’s properties and the names and attributes assigned to it. These associations exist, strongly but differently, in each of four real NFS environments studied. Decision tree classifiers can automatically identify and model such associations, providing prediction accuracies that often exceed 90%. Such predictions can be used to select storage policies (e.g., disk allocation schemes and replication factors) for individual files. Further, changes in associations can expose information about applications, helping autonomic system components distinguish growth from fundamental change
    corecore