1,961,952 research outputs found
A Service-based Model for Customer Intelligence in the Age of Big Data
The dominance of the service sector in today’s economy gives prominence to customer intelligence as a means for enterprises to provide optimal service. In fact, the revolution of big data has generated a vast amount of customer data and reshaped the dimensions of science, management, and engineering within enterprises. The big data era also acknowledges the role of customers as value co-creators. Therefore, the objective of this paper is to propose a service-based customer intelligence model, hereafter called SBCI (Service-based Customer Intelligence) model, to guide the development and application of customer intelligence. Laid the groundwork upon the service science, the model is proposed with three levels: i) the network of service systems level for customer value co-creation, ii) the service system level for the science, management, and engineering dimensions, and iii) the service level for customer intelligence services
Simulating Congestion Dynamics of Train Rapid Transit using Smart Card Data
Investigating congestion in train rapid transit systems (RTS) in today's
urban cities is a challenge compounded by limited data availability and
difficulties in model validation. Here, we integrate information from travel
smart card data, a mathematical model of route choice, and a full-scale
agent-based model of the Singapore RTS to provide a more comprehensive
understanding of the congestion dynamics than can be obtained through
analytical modelling alone. Our model is empirically validated, and allows for
close inspection of the dynamics including station crowdedness, average travel
duration, and frequency of missed trains---all highly pertinent factors in
service quality. Using current data, the crowdedness in all 121 stations
appears to be distributed log-normally. In our preliminary scenarios, we
investigate the effect of population growth on service quality. We find that
the current population (2 million) lies below a critical point; and increasing
it beyond a factor of leads to an exponential deterioration in
service quality. We also predict that incentivizing commuters to avoid the most
congested hours can bring modest improvements to the service quality provided
the population remains under the critical point. Finally, our model can be used
to generate simulated data for analytical modelling when such data are not
empirically available, as is often the case.Comment: 10 pages, 5 figures, submitted to International Conference on
Computational Science 201
Experimental Design at the Intersection of Mathematics, Science, and Technology in Grades K-6
Interdisciplinary courses, highlighting as they do the area(s) the disciplines have in common, often give the misperception of a single body of knowledge and/or way of knowing. However, discipline based courses often leave the equally mistaken notion that the disciplines have nothing in common. The task of the methods courses described in this paper is to reach an appropriate balance so that our pre-service elementary (K-6) teachers have a realistic perception of the independence and interdependence of mathematics and science. At the College of William and Mary each cohort of pre-service elementary teachers enrolls in mathematics and science methods courses taught in consecutive hours. Both instructors emphasize the importance of the content pedagogy unique to their disciplines such as strategies for teaching problem solving, computation, algebraic thinking, and proportional reasoning in mathematics and strategies for teaching students how to investigate and understand the concepts of science. The instructors model interdisciplinary instruction by collaboratively teaching common content pedagogy such as the use of technology, data analysis, and interpretation. Students also identify real-life application of the mathematical principles they are learning that can be applied to science. The concept of simultaneously teaching appropriately selected math and science skills are stressed. Given this approach students are not left with the notion that mathematics is the handmaid of science nor the notion that it is the queen of the sciences. Rather, they view mathematics as a co-equal partner
Market Model and Optimal Pricing Scheme of Big Data and Internet of Things (IoT)
Big data has been emerging as a new approach in utilizing large datasets to
optimize complex system operations. Big data is fueled with Internet-of-Things
(IoT) services that generate immense sensory data from numerous sensors and
devices. While most current research focus of big data is on machine learning
and resource management design, the economic modeling and analysis have been
largely overlooked. This paper thus investigates the big data market model and
optimal pricing scheme. We first study the utility of data from the data
science perspective, i.e., using the machine learning methods. We then
introduce the market model and develop an optimal pricing scheme afterward. The
case study shows clearly the suitability of the proposed data utility
functions. The numerical examples demonstrate that big data and IoT service
provider can achieve the maximum profit through the proposed market model
Multi-layer Architecture For Storing Visual Data Based on WCF and Microsoft SQL Server Database
In this paper we present a novel architecture for storing visual data.
Effective storing, browsing and searching collections of images is one of the
most important challenges of computer science. The design of architecture for
storing such data requires a set of tools and frameworks such as SQL database
management systems and service-oriented frameworks. The proposed solution is
based on a multi-layer architecture, which allows to replace any component
without recompilation of other components. The approach contains five
components, i.e. Model, Base Engine, Concrete Engine, CBIR service and
Presentation. They were based on two well-known design patterns: Dependency
Injection and Inverse of Control. For experimental purposes we implemented the
SURF local interest point detector as a feature extractor and -means
clustering as indexer. The presented architecture is intended for content-based
retrieval systems simulation purposes as well as for real-world CBIR tasks.Comment: Accepted for the 14th International Conference on Artificial
Intelligence and Soft Computing, ICAISC, June 14-18, 2015, Zakopane, Polan
A construção imaginativa de cuidados: a experiência de profissionais de enfermagem em um serviço de assistência remota
Indexación: Web of Science; ScieloThe direction of care delivery goes from the action to the being; a process built from
professional experience, which gains special characteristics when the service is delivered by
telephone. The goal of this research was to understand the interaction between professionals and users in a
remote care service; to do so, a research is presented, using Grounded Theory and Symbolic
Interactionism as theoretical references. Data were collected through eight interviews with professionals who
deliver care by telephone. The theoretical understanding permitted the creation of the theoretical
model of the Imaginative Construction of Care, which shows the interaction processes the professional
experiences when delivering care by telephone. In this model, individual and social facts are
added, showing the link between the concepts, with special emphasis on uncertainty, sensitivity and
professional responsibility, as essential components of this experience.http://www.scielo.br/scielo.php?script=sci_arttext&pid=S0104-11692012000400009&nrm=isohttp://ref.scielo.org/44chq
A Semantic Grid Service for Experimentation with an Agent-Based Model of Land-Use Change
Agent-based models, perhaps more than other models, feature large numbers of parameters and potentially generate vast quantities of results data. This paper shows through the FEARLUS-G project (an ESRC e-Social Science Initiative Pilot Demonstrator Project) how deploying an agent-based model on the Semantic Grid facilitates international collaboration on investigations using such a model, and contributes to establishing rigorous working practices with agent-based models as part of good science in social simulation. The experimental workflow is described explicitly using an ontology, and a Semantic Grid service with a web interface implements the workflow. Users are able to compare their parameter settings and results, and relate their work with the model to wider scientific debate.Agent-Based Social Simulation, Experiments, Ontologies, Replication, Semantic Grid
High-Performance Cloud Computing: A View of Scientific Applications
Scientific computing often requires the availability of a massive number of
computers for performing large scale experiments. Traditionally, these needs
have been addressed by using high-performance computing solutions and installed
facilities such as clusters and super computers, which are difficult to setup,
maintain, and operate. Cloud computing provides scientists with a completely
new model of utilizing the computing infrastructure. Compute resources, storage
resources, as well as applications, can be dynamically provisioned (and
integrated within the existing infrastructure) on a pay per use basis. These
resources can be released when they are no more needed. Such services are often
offered within the context of a Service Level Agreement (SLA), which ensure the
desired Quality of Service (QoS). Aneka, an enterprise Cloud computing
solution, harnesses the power of compute resources by relying on private and
public Clouds and delivers to users the desired QoS. Its flexible and service
based infrastructure supports multiple programming paradigms that make Aneka
address a variety of different scenarios: from finance applications to
computational science. As examples of scientific computing in the Cloud, we
present a preliminary case study on using Aneka for the classification of gene
expression data and the execution of fMRI brain imaging workflow.Comment: 13 pages, 9 figures, conference pape
- …