30,933 research outputs found
Improving the quality of the industrial enterprise management based on the network-centric approach
The article examines the network-centric approach to the industrial enterprise management to improve the ef ciency and effectiveness in the implementation of production plans and maximize responsiveness to customers. A network-centric management means the decentralized enterprise group management. A group means a set of enterprise divisions, which should solve by joint efforts a certain case that occurs in the production process. The network-centric management involves more delegation of authority to the lower elements of the enterprise’s organizational structure. The industrial enterprise is considered as a large complex system (production system) functioning and controlled amidst various types of uncertainty: information support uncertainty and goal uncertainty or multicriteria uncertainty. The information support uncertainty occurs because the complex system functioning always takes place in the context of incomplete and fuzzy information. Goal uncertainty or multicriteria uncertainty caused by a great number of goalsestablished for the production system. The network-centric management task de nition by the production system is formulated. The authors offer a mathematical model for optimal planning of consumers’ orders production with the participation of the main enterprise divisions. The methods of formalization of various types of uncertainty in production planning tasks are considered on the basis of the application of the fuzzy sets theory. An enterprise command center is offered as an effective tool for making management decisions by divisions. The article demonstrates that decentralized group management methods can improve the ef ciency and effectiveness of the implementation of production plans through the self-organization mechanisms of enterprise divisions.The work has been prepared with the financial support from the Russian Ministry of Education and Science (Contract No. 02.G25.31.0068 of 23.05.2013 as part of the measure to implement Decision of the Russian Government No. 218)
Data granulation by the principles of uncertainty
Researches in granular modeling produced a variety of mathematical models,
such as intervals, (higher-order) fuzzy sets, rough sets, and shadowed sets,
which are all suitable to characterize the so-called information granules.
Modeling of the input data uncertainty is recognized as a crucial aspect in
information granulation. Moreover, the uncertainty is a well-studied concept in
many mathematical settings, such as those of probability theory, fuzzy set
theory, and possibility theory. This fact suggests that an appropriate
quantification of the uncertainty expressed by the information granule model
could be used to define an invariant property, to be exploited in practical
situations of information granulation. In this perspective, a procedure of
information granulation is effective if the uncertainty conveyed by the
synthesized information granule is in a monotonically increasing relation with
the uncertainty of the input data. In this paper, we present a data granulation
framework that elaborates over the principles of uncertainty introduced by
Klir. Being the uncertainty a mesoscopic descriptor of systems and data, it is
possible to apply such principles regardless of the input data type and the
specific mathematical setting adopted for the information granules. The
proposed framework is conceived (i) to offer a guideline for the synthesis of
information granules and (ii) to build a groundwork to compare and
quantitatively judge over different data granulation procedures. To provide a
suitable case study, we introduce a new data granulation technique based on the
minimum sum of distances, which is designed to generate type-2 fuzzy sets. We
analyze the procedure by performing different experiments on two distinct data
types: feature vectors and labeled graphs. Results show that the uncertainty of
the input data is suitably conveyed by the generated type-2 fuzzy set models.Comment: 16 pages, 9 figures, 52 reference
Neuro-fuzzy knowledge processing in intelligent learning environments for improved student diagnosis
In this paper, a neural network implementation for a fuzzy logic-based model of the diagnostic process is proposed as a means to achieve accurate student diagnosis and updates of the student model in Intelligent Learning Environments. The neuro-fuzzy synergy allows the diagnostic model to some extent "imitate" teachers in diagnosing students' characteristics, and equips the intelligent learning environment with reasoning capabilities that can be further used to drive pedagogical decisions depending on the student learning style. The neuro-fuzzy implementation helps to encode both structured and non-structured teachers' knowledge: when teachers' reasoning is available and well defined, it can be encoded in the form of fuzzy rules; when teachers' reasoning is not well defined but is available through practical examples illustrating their experience, then the networks can be trained to represent this experience. The proposed approach has been tested in diagnosing aspects of student's learning style in a discovery-learning environment that aims to help students to construct the concepts of vectors in physics and mathematics. The diagnosis outcomes of the model have been compared against the recommendations of a group of five experienced teachers, and the results produced by two alternative soft computing methods. The results of our pilot study show that the neuro-fuzzy model successfully manages the inherent uncertainty of the diagnostic process; especially for marginal cases, i.e. where it is very difficult, even for human tutors, to diagnose and accurately evaluate students by directly synthesizing subjective and, some times, conflicting judgments
Big data analytics:Computational intelligence techniques and application areas
Big Data has significant impact in developing functional smart cities and supporting modern societies. In this paper, we investigate the importance of Big Data in modern life and economy, and discuss challenges arising from Big Data utilization. Different computational intelligence techniques have been considered as tools for Big Data analytics. We also explore the powerful combination of Big Data and Computational Intelligence (CI) and identify a number of areas, where novel applications in real world smart city problems can be developed by utilizing these powerful tools and techniques. We present a case study for intelligent transportation in the context of a smart city, and a novel data modelling methodology based on a biologically inspired universal generative modelling approach called Hierarchical Spatial-Temporal State Machine (HSTSM). We further discuss various implications of policy, protection, valuation and commercialization related to Big Data, its applications and deployment
- …