58,281 research outputs found
Economic Impacts of GO TO 2040
The economy of the Chicago metropolitan region has reached a critical juncture. On the one hand, Chicagoland is currently a highly successful global region with extraordinary assets and outputs. The region successfully made the transition in the 1980s and 1990s from a primarily industrial to a knowledge and service-based economy. It has high levels of human capital, with strong concentrations in information-sector industries and knowledge-based functional clusters -- a headquarters region with thriving finance, business services, law, IT and emerging bioscience, advanced manufacturing and similar high-growth sectors. It combines multiple deep areas of specialization, providing the resilience that comes from economic diversity. It is home to the abundant quality-of-life amenities that flow from business and household prosperity.On the other hand, beneath this static portrait of our strengths lie disturbing signs of a potential loss of momentum. Trends in the last decade reveal slowing rates, compared to other regions, of growth in productivity and gross metropolitan product. Trends in innovation, new firm creation and employment are comparably lagging. The region also faces emerging challenges with respect to both spatial efficiency and governance.In this context, the Chicago Metropolitan Agency for Planning (CMAP) has just released GO TO 2040, its comprehensive, long-term plan for the Chicago metropolitan area. The plan contains recommendations aimed at shaping a wide range of regional characteristics over the next 30 years, during which time more than 2 million new residents are anticipated. Among the chief goals of GO TO 2040 are increasing the region's long-term economic prosperity, sustaining a high quality of life for the region's current and future residents and making the most effective use of public investments. To this end, the plan addresses a broad scope of interrelated issues which, in aggregate, will shape the long-term physical, economic, institutional and social character of the region.This report by RW Ventures, LLC is an independent assessment of the plan from a purely economic perspective, addressing the impacts that GO TO 2040's recommendations can be expected to have on the future of the regional economy. The assessment begins by describing how implementation of GO TO 2040's recommendations would affect the economic landscape of the region; reviews economic research and practice about the factors that influence regional economic growth; and, given both of these, articulates and illustrates the likely economic impacts that will flow from implementation of the plan. In the course of reviewing the economic implications of the plan, the assessment also provides recommendations of further steps, as the plan is implemented, for increasing its positive impact on economic growth
Evaluating Enterprize Delivery Using the TYPUS Metrics and the KILT Mode
The goal of this work is the technical, ecological, environmental and social examination of the life-cycle (LC) of any product (consumable, service, production) using the TYPUS metrics and the KILT model. The life-cycle starts when the idea of a product is born and lasts until complete dismissal through design, implementation and operation, etc. In the first phases requirements’ specification, analysis, several design steps (global plan, detailed design, assembly design, etc.) are followed by part manufacturing, assembly, testing, diagnostics and operation, advertisement, service, maintenance, etc. Then finally disassembly and dismissal are coming, but dismissal can be substituted by re-cycling (e.g. melting the metals) or re-use (used parts applications). Qualitative and quantitative evaluations of enterprise results are supported by the new models and metrics
CASP-DM: Context Aware Standard Process for Data Mining
We propose an extension of the Cross Industry Standard Process for Data
Mining (CRISPDM) which addresses specific challenges of machine learning and
data mining for context and model reuse handling. This new general
context-aware process model is mapped with CRISP-DM reference model proposing
some new or enhanced outputs
Recommended from our members
A review of microgrid development in the United States – A decade of progress on policies, demonstrations, controls, and software tools
Microgrids have become increasingly popular in the United States. Supported by favorable federal and local policies, microgrid projects can provide greater energy stability and resilience within a project site or community. This paper reviews major federal, state, and utility-level policies driving microgrid development in the United States. Representative U.S. demonstration projects are selected and their technical characteristics and non-technical features are introduced. The paper discusses trends in the technology development of microgrid systems as well as microgrid control methods and interactions within the electricity market. Software tools for microgrid design, planning, and performance analysis are illustrated with each tool's core capability. Finally, the paper summarizes the successes and lessons learned during the recent expansion of the U.S. microgrid industry that may serve as a reference for other countries developing their own microgrid industries
Rethinking Security Incident Response: The Integration of Agile Principles
In today's globally networked environment, information security incidents can
inflict staggering financial losses on organizations. Industry reports indicate
that fundamental problems exist with the application of current linear
plan-driven security incident response approaches being applied in many
organizations. Researchers argue that traditional approaches value containment
and eradication over incident learning. While previous security incident
response research focused on best practice development, linear plan-driven
approaches and the technical aspects of security incident response, very little
research investigates the integration of agile principles and practices into
the security incident response process. This paper proposes that the
integration of disciplined agile principles and practices into the security
incident response process is a practical solution to strengthening an
organization's security incident response posture.Comment: Paper presented at the 20th Americas Conference on Information
Systems (AMCIS 2014), Savannah, Georgi
High-Level Object Oriented Genetic Programming in Logistic Warehouse Optimization
DisertaÄŤnĂ práce je zaměřena na optimalizaci prĹŻbÄ›hu pracovnĂch operacĂ v logistickĂ˝ch skladech a distribuÄŤnĂch centrech. HlavnĂm cĂlem je optimalizovat procesy plánovánĂ, rozvrhovánĂ a odbavovánĂ. JelikoĹľ jde o problĂ©m patĹ™ĂcĂ do tĹ™Ădy sloĹľitosti NP-teĹľkĂ˝, je vĂ˝poÄŤetnÄ› velmi nároÄŤnĂ© nalĂ©zt optimálnĂ Ĺ™ešenĂ. MotivacĂ pro Ĺ™ešenĂ tĂ©to práce je vyplnÄ›nĂ pomyslnĂ© mezery mezi metodami zkoumanĂ˝mi na vÄ›deckĂ© a akademickĂ© pĹŻdÄ› a metodami pouĹľĂvanĂ˝mi v produkÄŤnĂch komerÄŤnĂch prostĹ™edĂch. Jádro optimalizaÄŤnĂho algoritmu je zaloĹľeno na základÄ› genetickĂ©ho programovánĂ Ĺ™ĂzenĂ©ho bezkontextovou gramatikou. HlavnĂm pĹ™Ănosem tĂ©to práce je a) navrhnout novĂ˝ optimalizaÄŤnĂ algoritmus, kterĂ˝ respektuje následujĂcĂ optimalizaÄŤnĂ podmĂnky: celkovĂ˝ ÄŤas zpracovánĂ, vyuĹľitĂ zdrojĹŻ, a zahlcenĂ skladovĂ˝ch uliÄŤek, kterĂ© mĹŻĹľe nastat bÄ›hem zpracovánĂ ĂşkolĹŻ, b) analyzovat historická data z provozu skladu a vyvinout sadu testovacĂch pĹ™ĂkladĹŻ, kterĂ© mohou slouĹľit jako referenÄŤnĂ vĂ˝sledky pro dalšà vĂ˝zkum, a dále c) pokusit se pĹ™edÄŤit stanovenĂ© referenÄŤnĂ vĂ˝sledky dosaĹľenĂ© kvalifikovanĂ˝m a trĂ©novanĂ˝m operaÄŤnĂm manaĹľerem jednoho z nejvÄ›tšĂch skladĹŻ ve stĹ™ednĂ EvropÄ›.This work is focused on the work-flow optimization in logistic warehouses and distribution centers. The main aim is to optimize process planning, scheduling, and dispatching. The problem is quite accented in recent years. The problem is of NP hard class of problems and where is very computationally demanding to find an optimal solution. The main motivation for solving this problem is to fill the gap between the new optimization methods developed by researchers in academic world and the methods used in business world. The core of the optimization algorithm is built on the genetic programming driven by the context-free grammar. The main contribution of the thesis is a) to propose a new optimization algorithm which respects the makespan, the utilization, and the congestions of aisles which may occur, b) to analyze historical operational data from warehouse and to develop the set of benchmarks which could serve as the reference baseline results for further research, and c) to try outperform the baseline results set by the skilled and trained operational manager of the one of the biggest warehouses in the middle Europe.
Physical Representation-based Predicate Optimization for a Visual Analytics Database
Querying the content of images, video, and other non-textual data sources
requires expensive content extraction methods. Modern extraction techniques are
based on deep convolutional neural networks (CNNs) and can classify objects
within images with astounding accuracy. Unfortunately, these methods are slow:
processing a single image can take about 10 milliseconds on modern GPU-based
hardware. As massive video libraries become ubiquitous, running a content-based
query over millions of video frames is prohibitive.
One promising approach to reduce the runtime cost of queries of visual
content is to use a hierarchical model, such as a cascade, where simple cases
are handled by an inexpensive classifier. Prior work has sought to design
cascades that optimize the computational cost of inference by, for example,
using smaller CNNs. However, we observe that there are critical factors besides
the inference time that dramatically impact the overall query time. Notably, by
treating the physical representation of the input image as part of our query
optimization---that is, by including image transforms, such as resolution
scaling or color-depth reduction, within the cascade---we can optimize data
handling costs and enable drastically more efficient classifier cascades.
In this paper, we propose Tahoma, which generates and evaluates many
potential classifier cascades that jointly optimize the CNN architecture and
input data representation. Our experiments on a subset of ImageNet show that
Tahoma's input transformations speed up cascades by up to 35 times. We also
find up to a 98x speedup over the ResNet50 classifier with no loss in accuracy,
and a 280x speedup if some accuracy is sacrificed.Comment: Camera-ready version of the paper submitted to ICDE 2019, In
Proceedings of the 35th IEEE International Conference on Data Engineering
(ICDE 2019
- …