511,338 research outputs found

    Implications for enterprise software vendors (ESV)

    Get PDF
    Thesis (S.M.)--Massachusetts Institute of Technology, System Design and Management Program, 2009.Includes bibliographical references (leaves 67-69).'Cloud computing', is a broad concept and in general is a term used for internet-based computing resources that are in an unspecified remote location or locations and that are flexible and fungible. Clouds provide a wide range of computing capability available as a service where users are separated from the underlying technology by a set of APIs. These computing capabilities are made available by abstracting at different levels; at the hardware level, development platform or the applications level. Cloud computing is particularly helpful to application developers and IT operations because it allows them to focus on the service/application provided rather than worrying about scaling, failure, maintenance or reliability of these computing resources. By consolidating and sharing computing resources among multiple tenants thus improving utilization, cloud computing brings cost savings to end users. The higher the abstraction level, greater are the benefits resulting from better resource utilization and thus more cost savings, both for providers and end users. As computing resources become cheaper, network connectivity and bandwidth improve both in terms of availability and pricing and human resources becomes expensive, cloud computing is increasingly seen as viable replacement of enterprise owned local IT infrastructure.(cont.) With the adoption of cloud computing comes a major shift in the underlying architecture of how we develop, deploy, deliver and run applications compared to existing behavior where we run applications on local computing resources and thus increasing pressure on enterprise software vendors to adopt these new business model for software development and new alternate software delivery models that are supported by and derive the benefits of cloud computing. While legacy enterprise software can simply be installed and run on instances on the cloud using cloud based infrastructure services, maximum benefits are realized by end users when these applications itself are provided as a service in the form of a platform or software. To do so, in most cases, legacy enterprise software would have to go through an architecture overhaul to be able to deliver existing functionalities as a platform or software as a service. Enterprise software vendors would also have to change their current business models where large license revenues, high maintenance cost of antiquated versions and heavily invested customers are the standard and move to pay-per-use cloud computing model. This thesis aims to study the implications of 'Cloud Computing' trends on the development, distribution, business models and the business of enterprise software vendors. This thesis tries to chart and predict the progress of trends in computing towards 'cloud computing', connect those trends to enterprise software usage changes and determine the impact on enterprise software vendors.(cont.) This will help enterprise software vendors to determine what if any strategic options available will help adopt this technological innovation and conform to future enterprise software requirements based on this trend.by Leonard Francis.S.M

    AI and OR in management of operations: history and trends

    Get PDF
    The last decade has seen a considerable growth in the use of Artificial Intelligence (AI) for operations management with the aim of finding solutions to problems that are increasing in complexity and scale. This paper begins by setting the context for the survey through a historical perspective of OR and AI. An extensive survey of applications of AI techniques for operations management, covering a total of over 1200 papers published from 1995 to 2004 is then presented. The survey utilizes Elsevier's ScienceDirect database as a source. Hence, the survey may not cover all the relevant journals but includes a sufficiently wide range of publications to make it representative of the research in the field. The papers are categorized into four areas of operations management: (a) design, (b) scheduling, (c) process planning and control and (d) quality, maintenance and fault diagnosis. Each of the four areas is categorized in terms of the AI techniques used: genetic algorithms, case-based reasoning, knowledge-based systems, fuzzy logic and hybrid techniques. The trends over the last decade are identified, discussed with respect to expected trends and directions for future work suggested

    CHORUS Deliverable 3.4: Vision Document

    Get PDF
    The goal of the CHORUS Vision Document is to create a high level vision on audio-visual search engines in order to give guidance to the future R&D work in this area and to highlight trends and challenges in this domain. The vision of CHORUS is strongly connected to the CHORUS Roadmap Document (D2.3). A concise document integrating the outcomes of the two deliverables will be prepared for the end of the project (NEM Summit)

    Building Programmable Wireless Networks: An Architectural Survey

    Full text link
    In recent times, there have been a lot of efforts for improving the ossified Internet architecture in a bid to sustain unstinted growth and innovation. A major reason for the perceived architectural ossification is the lack of ability to program the network as a system. This situation has resulted partly from historical decisions in the original Internet design which emphasized decentralized network operations through co-located data and control planes on each network device. The situation for wireless networks is no different resulting in a lot of complexity and a plethora of largely incompatible wireless technologies. The emergence of "programmable wireless networks", that allow greater flexibility, ease of management and configurability, is a step in the right direction to overcome the aforementioned shortcomings of the wireless networks. In this paper, we provide a broad overview of the architectures proposed in literature for building programmable wireless networks focusing primarily on three popular techniques, i.e., software defined networks, cognitive radio networks, and virtualized networks. This survey is a self-contained tutorial on these techniques and its applications. We also discuss the opportunities and challenges in building next-generation programmable wireless networks and identify open research issues and future research directions.Comment: 19 page

    Quality of Web Mashups: A Systematic Mapping Study

    Full text link
    The final publication is available at Springer via http://dx.doi.org/10.1007/978-3-319-04244-2_8Web mashups are a new generation of applications based on the composition of ready-to-use, heterogeneous components. They are gaining momentum thanks to their lightweight composition approach, which represents a new opportunity for companies to leverage on past investments in SOA, Web services, and public APIs. Although several studies are emerging in order to address mashup development, no systematic mapping studies have been reported on how quality issues are being addressed. This paper reports a systematic mapping study on which and how the quality of Web mashups has been addressed and how the product quality-aware approaches have been defined and validated. The aim of this study is to provide a background in which to appropriately develop future research activities. A total of 38 research papers have been included from an initial set of 187 papers. Our results provided some findings regarding how the most relevant product quality characteristics have been addressed in different artifacts and stages of the development process. They have also been useful to detect some research gaps, such as the need of more controlled experiments and more quality-aware mashup development proposals for other characteristics which being important for the Web domain have been neglected such as Usability and ReliabilityThis work is funded by the MULTIPLE project (TIN2009-13838), the Senescyt program (scholarships 2011), and the Erasmus Mundus Programme of the European Commission under the Transatlantic Partnership for Excellence in Engineering - TEE Project.Cedillo Orellana, IP.; Fernández Martínez, A.; Insfrán Pelozo, CE.; Abrahao Gonzales, SM. (2013). Quality of Web Mashups: A Systematic Mapping Study. En Current Trends in Web Engineering. Springer. 66-78. https://doi.org/10.1007/978-3-319-04244-2_8S6678Alkhalifa, E.: The Future of Enterprise Mashups. Business Insights. E-Strategies for Resource Management Systems (2009)Beemer, B., Gregg, D.: Mashups: A Literature Review and Classification Framework. Future Internet 1, 59–87 (2009)Cappiello, C., Daniel, F., Matera, M.: A Quality Model for Mashup Components. In: Gaedke, M., Grossniklaus, M., Díaz, O. (eds.) ICWE 2009. LNCS, vol. 5648, pp. 236–250. Springer, Heidelberg (2009)Cappiello, C., Daniel, F., Matera, M., Pautasso, C.: Information Quality in Mashups. IEEE Internet Computing 14(4), 32–40 (2010)Cappiello, C., Matera, M., Picozzi, M., Daniel, F., Fernandez, A.: Quality-Aware Mashup Composition: Issues, Techniques and Tools. In: 8th International Conference on the Quality of Information and Communications Technology (QUATIC 2012), pp. 10–19 (2012)Fenton, N.E., Pfleeger, S.L.: Software Metrics: A Rigorous and Practical Approach, 2nd edn. International Thompson 1996, pp. I–XII, 1–638 (1996) ISBN 978-1-85032-275-7Fernandez, A., Insfran, E., Abrahão, S.: Usability evaluation methods for the web: A systematic mapping study. Information and Software Technology 53(8), 789–817 (2011)Garousi, V., Mesbah, A., Betin-Can, A., Mirshokraie, S.: A systematic mapping study of web application testing. Information and Software Technology 55(8), 1374–1396 (2013)Grammel, L., Storey, M.-A.: A survey of mashup development environments. In: Chignell, M., Cordy, J., Ng, J., Yesha, Y. (eds.) The Smart Internet. LNCS, vol. 6400, pp. 137–151. Springer, Heidelberg (2010)Hoyer, V., Fischer, M.: Market Overview of Enterprise Mashup Tools. In: Bouguettaya, A., Krueger, I., Margaria, T. (eds.) ICSOC 2008. LNCS, vol. 5364, pp. 708–721. Springer, Heidelberg (2008)ISO/IEC: ISO/IEC 25010 Systems and software engineering. Systems and software Quality Requirements and Evaluation (SQuaRE). System and software quality models (2011)Kitchenham, B., Charters, S.: Guidelines for performing Systematic Literature Reviews in Software Engineering. Version 2.3, ESBE Technical Report, Keele University, UK (2007)Mendes, E.: A systematic review on the Web engineering research. In: International Symposium on Empirical Software Engineering (ISESE 2005), pp. 498–507 (2005)OrangeLabs: State of the Art in Mashup tools, SocEDA project, pp. 1–59 (2011)Petersen, K., Feldt, R., Mujtaba, S., Mattsson, M.: Systematic mapping studies in software engineering. In: 12th International Conference on Evaluation and Assessment in Software Engineering (EASE), pp. 68–77 (2008)Raza, M., Hussain, F.K., Chang, E.: A methodology for quality-based mashup of data sources. In: 10th International Conference on Information Integration and Web-based Applications & Services (iiWAS 2008), pp. 528–533 (2008)Saeed, A.: A Quality-based Framework for Leveraging the Process of Mashup Component Selection (2009), https://gupea.ub.gu.se/handle/2077/21953Sharma, A., Hellmann, T.D., Maurer, F.: Testing of Web Services - A Systematic Mapping. In: 8th World Congress on Services (SERVICES 2012), pp. 346–352 (2012

    Discrete event simulation and virtual reality use in industry: new opportunities and future trends

    Get PDF
    This paper reviews the area of combined discrete event simulation (DES) and virtual reality (VR) use within industry. While establishing a state of the art for progress in this area, this paper makes the case for VR DES as the vehicle of choice for complex data analysis through interactive simulation models, highlighting both its advantages and current limitations. This paper reviews active research topics such as VR and DES real-time integration, communication protocols, system design considerations, model validation, and applications of VR and DES. While summarizing future research directions for this technology combination, the case is made for smart factory adoption of VR DES as a new platform for scenario testing and decision making. It is put that in order for VR DES to fully meet the visualization requirements of both Industry 4.0 and Industrial Internet visions of digital manufacturing, further research is required in the areas of lower latency image processing, DES delivery as a service, gesture recognition for VR DES interaction, and linkage of DES to real-time data streams and Big Data sets

    The future of computing beyond Moore's Law.

    Get PDF
    Moore's Law is a techno-economic model that has enabled the information technology industry to double the performance and functionality of digital electronics roughly every 2 years within a fixed cost, power and area. Advances in silicon lithography have enabled this exponential miniaturization of electronics, but, as transistors reach atomic scale and fabrication costs continue to rise, the classical technological driver that has underpinned Moore's Law for 50 years is failing and is anticipated to flatten by 2025. This article provides an updated view of what a post-exascale system will look like and the challenges ahead, based on our most recent understanding of technology roadmaps. It also discusses the tapering of historical improvements, and how it affects options available to continue scaling of successors to the first exascale machine. Lastly, this article covers the many different opportunities and strategies available to continue computing performance improvements in the absence of historical technology drivers. This article is part of a discussion meeting issue 'Numerical algorithms for high-performance computational science'
    corecore