950 research outputs found
Smart Bike Sharing System to make the City even Smarter
These last years with the growing population in the smart city demands an
efficient transportation sharing (bike sharing) system for developing the smart
city. The Bike sharing as we know is affordable, easily accessible and reliable
mode of transportation. But an efficient bike sharing capable of not only
sharing bike also provides information regarding the availability of bike per
station, route business, time/day-wise bike schedule. The embedded sensors are
able to opportunistically communicate through wireless communication with
stations when available, providing real-time data about tours/minutes, speed,
effort, rhythm, etc. We have been based on our study analysis data to predict
regarding the bike's available at stations, bike schedule, a location of the
nearest hub where a bike is available etc., reduce the user time and effort
Ontology-based Classification and Analysis of non- emergency Smart-city Events
Several challenges are faced by citizens of urban centers while dealing with
day-to-day events, and the absence of a centralised reporting mechanism makes
event-reporting and redressal a daunting task. With the push on information
technology to adapt to the needs of smart-cities and integrate urban civic
services, the use of Open311 architecture presents an interesting solution. In
this paper, we present a novel approach that uses an existing Open311 ontology
to classify and report non-emergency city-events, as well as to guide the
citizen to the points of redressal. The use of linked open data and the
semantic model serves to provide contextual meaning and make vast amounts of
content hyper-connected and easily-searchable. Such a one-size-fits-all model
also ensures reusability and effective visualisation and analysis of data
across several cities. By integrating urban services across various civic
bodies, the proposed approach provides a single endpoint to the citizen, which
is imperative for smooth functioning of smart cities
LEVERAGING BIBLIOGRAPHIC RDF DATA FOR KEYWORD PREDICTION WITH ASSOCIATION RULE MINING (ARM)
The Semantic Web ( Web 3.03.0) has been proposed as an efficient way to access the increasingly large amounts of data on the internet. The Linked Open Data Cloud project at present is the major effort to implement the concepts of the Seamtic Web, addressing the problems of in homogeneity and large data volumes. RKBExplorer is one of many repositories implementing Open Data and contains considerable bibliographic information. Th is paper discusses bibliographic data data, an important part of cloud data. Effective searching of bibliographic datasets can be a challenge as many of the papers residing in these databases do not have sufficient or comprehensive keyword information. In these cases however, a search engine based on RKBExplorer is only able to use information to retrieve papers based on author names and title of papers without keywords keywords. In this paper we attempt to address this problem by using the data mining algorithm Association Rule Mining (ARM ) to develop keywords based on features retrieved from Resource Description Framework (RDF) data within a bibliographic citation. We have demonstrate the applicability of this method for predicting missing keywords for bibliographic entries in several typical databases
Terrain Classification using Transfer Learning on Hyperspectral Images: A Comparative study
A Hyperspectral image contains much more number of channels as compared to a
RGB image, hence containing more information about entities within the image.
The convolutional neural network (CNN) and the Multi-Layer Perceptron (MLP)
have been proven to be an effective method of image classification. However,
they suffer from the issues of long training time and requirement of large
amounts of the labeled data, to achieve the expected outcome. These issues
become more complex while dealing with hyperspectral images. To decrease the
training time and reduce the dependence on large labeled dataset, we propose
using the method of transfer learning. The hyperspectral dataset is
preprocessed to a lower dimension using PCA, then deep learning models are
applied to it for the purpose of classification. The features learned by this
model are then used by the transfer learning model to solve a new
classification problem on an unseen dataset. A detailed comparison of CNN and
multiple MLP architectural models is performed, to determine an optimum
architecture that suits best the objective. The results show that the scaling
of layers not always leads to increase in accuracy but often leads to
overfitting, and also an increase in the training time.The training time is
reduced to greater extent by applying the transfer learning approach rather
than just approaching the problem by directly training a new model on large
datasets, without much affecting the accuracy
- …