266,272 research outputs found
Attributes of Big Data Analytics for Data-Driven Decision Making in Cyber-Physical Power Systems
Big data analytics is a virtually new term in power system terminology. This concept delves into the way a massive volume of data is acquired, processed, analyzed to extract insight from available data. In particular, big data analytics alludes to applications of artificial intelligence, machine learning techniques, data mining techniques, time-series forecasting methods. Decision-makers in power systems have been long plagued by incapability and weakness of classical methods in dealing with large-scale real practical cases due to the existence of thousands or millions of variables, being time-consuming, the requirement of a high computation burden, divergence of results, unjustifiable errors, and poor accuracy of the model. Big data analytics is an ongoing topic, which pinpoints how to extract insights from these large data sets. The extant article has enumerated the applications of big data analytics in future power systems through several layers from grid-scale to local-scale. Big data analytics has many applications in the areas of smart grid implementation, electricity markets, execution of collaborative operation schemes, enhancement of microgrid operation autonomy, management of electric vehicle operations in smart grids, active distribution network control, district hub system management, multi-agent energy systems, electricity theft detection, stability and security assessment by PMUs, and better exploitation of renewable energy sources. The employment of big data analytics entails some prerequisites, such as the proliferation of IoT-enabled devices, easily-accessible cloud space, blockchain, etc. This paper has comprehensively conducted an extensive review of the applications of big data analytics along with the prevailing challenges and solutions
Information Splitting for Big Data Analytics
Many statistical models require an estimation of unknown (co)-variance
parameter(s) in a model. The estimation usually obtained by maximizing a
log-likelihood which involves log determinant terms. In principle, one requires
the \emph{observed information}--the negative Hessian matrix or the second
derivative of the log-likelihood---to obtain an accurate maximum likelihood
estimator according to the Newton method. When one uses the \emph{Fisher
information}, the expect value of the observed information, a simpler algorithm
than the Newton method is obtained as the Fisher scoring algorithm. With the
advance in high-throughput technologies in the biological sciences,
recommendation systems and social networks, the sizes of data sets---and the
corresponding statistical models---have suddenly increased by several orders of
magnitude. Neither the observed information nor the Fisher information is easy
to obtained for these big data sets. This paper introduces an information
splitting technique to simplify the computation. After splitting the mean of
the observed information and the Fisher information, an simpler approximate
Hessian matrix for the log-likelihood can be obtained. This approximated
Hessian matrix can significantly reduce computations, and makes the linear
mixed model applicable for big data sets. Such a spitting and simpler formulas
heavily depends on matrix algebra transforms, and applicable to large scale
breeding model, genetics wide association analysis.Comment: arXiv admin note: text overlap with arXiv:1605.0764
How can SMEs benefit from big data? Challenges and a path forward
Big data is big news, and large companies in all sectors are making significant advances in their customer relations, product selection and development and consequent profitability through using this valuable commodity. Small and medium enterprises (SMEs) have proved themselves to be slow adopters of the new technology of big data analytics and are in danger of being left behind. In Europe, SMEs are a vital part of the economy, and the challenges they encounter need to be addressed as a matter of urgency. This paper identifies barriers to SME uptake of big data analytics and recognises their complex challenge to all stakeholders, including national and international policy makers, IT, business management and data science communities.
The paper proposes a big data maturity model for SMEs as a first step towards an SME roadmap to data analytics. It considers the ‘state-of-the-art’ of IT with respect to usability and usefulness for SMEs and discusses how SMEs can overcome the barriers preventing them from adopting existing solutions. The paper then considers management perspectives and the role of maturity models in enhancing and structuring the adoption of data analytics in an organisation. The history of total quality management is reviewed to inform the core aspects of implanting a new paradigm. The paper concludes with recommendations to help SMEs develop their big data capability and enable them to continue as the engines of European industrial and business success. Copyright © 2016 John Wiley & Sons, Ltd.Peer ReviewedPostprint (author's final draft
Recommended from our members
The Emergence of Big Data Policing
The past decade has seen both the proliferation of surveillance in everyday life and the rise of “big data.” Through extensive qualitative research focusing on the Los Angeles Police Department (LAPD), PRC faculty research associate Sarah Brayne explores whether and how adopting big data analytics transforms police surveillance practices. This brief demonstrates that, in some cases, the adoption of big data analytics is associated with mere amplifications in prior practices, but in others, it is associated with fundamental transformations in surveillance activities.Population Research Cente
Crisis Analytics: Big Data Driven Crisis Response
Disasters have long been a scourge for humanity. With the advances in
technology (in terms of computing, communications, and the ability to process
and analyze big data), our ability to respond to disasters is at an inflection
point. There is great optimism that big data tools can be leveraged to process
the large amounts of crisis-related data (in the form of user generated data in
addition to the traditional humanitarian data) to provide an insight into the
fast-changing situation and help drive an effective disaster response. This
article introduces the history and the future of big crisis data analytics,
along with a discussion on its promise, challenges, and pitfalls
- …
