24,622 research outputs found
iTeleScope: Intelligent Video Telemetry and Classification in Real-Time using Software Defined Networking
Video continues to dominate network traffic, yet operators today have poor
visibility into the number, duration, and resolutions of the video streams
traversing their domain. Current approaches are inaccurate, expensive, or
unscalable, as they rely on statistical sampling, middle-box hardware, or
packet inspection software. We present {\em iTelescope}, the first intelligent,
inexpensive, and scalable SDN-based solution for identifying and classifying
video flows in real-time. Our solution is novel in combining dynamic flow rules
with telemetry and machine learning, and is built on commodity OpenFlow
switches and open-source software. We develop a fully functional system, train
it in the lab using multiple machine learning algorithms, and validate its
performance to show over 95\% accuracy in identifying and classifying video
streams from many providers including Youtube and Netflix. Lastly, we conduct
tests to demonstrate its scalability to tens of thousands of concurrent
streams, and deploy it live on a campus network serving several hundred real
users. Our system gives unprecedented fine-grained real-time visibility of
video streaming performance to operators of enterprise and carrier networks at
very low cost.Comment: 12 pages, 16 figure
NeuRoute: Predictive Dynamic Routing for Software-Defined Networks
This paper introduces NeuRoute, a dynamic routing framework for Software
Defined Networks (SDN) entirely based on machine learning, specifically, Neural
Networks. Current SDN/OpenFlow controllers use a default routing based on
Dijkstra algorithm for shortest paths, and provide APIs to develop custom
routing applications. NeuRoute is a controller-agnostic dynamic routing
framework that (i) predicts traffic matrix in real time, (ii) uses a neural
network to learn traffic characteristics and (iii) generates forwarding rules
accordingly to optimize the network throughput. NeuRoute achieves the same
results as the most efficient dynamic routing heuristic but in much less
execution time.Comment: Accepted for CNSM 201
WebPicker: Knowledge Extraction from Web Resources
We show how information distributed in several web resources and represented in different restricted languages can be extracted from its original sources and transformed into a common knowledge model represented in XML using WebPicker. This information, which has been built to cover different needs and functionalities, can be later imported into WebODE, integrated, enriched and exported into different representation formats using WebODE specific modules. We show a case study in the e-commerce domain, using products and services standards from several organizations and/or joint initiatives of industrial and services companies, and a product catalogue from an e-commerce platform
Large-scale Continuous Gesture Recognition Using Convolutional Neural Networks
This paper addresses the problem of continuous gesture recognition from
sequences of depth maps using convolutional neutral networks (ConvNets). The
proposed method first segments individual gestures from a depth sequence based
on quantity of movement (QOM). For each segmented gesture, an Improved Depth
Motion Map (IDMM), which converts the depth sequence into one image, is
constructed and fed to a ConvNet for recognition. The IDMM effectively encodes
both spatial and temporal information and allows the fine-tuning with existing
ConvNet models for classification without introducing millions of parameters to
learn. The proposed method is evaluated on the Large-scale Continuous Gesture
Recognition of the ChaLearn Looking at People (LAP) challenge 2016. It achieved
the performance of 0.2655 (Mean Jaccard Index) and ranked place in
this challenge
LCrowdV: Generating Labeled Videos for Simulation-based Crowd Behavior Learning
We present a novel procedural framework to generate an arbitrary number of
labeled crowd videos (LCrowdV). The resulting crowd video datasets are used to
design accurate algorithms or training models for crowded scene understanding.
Our overall approach is composed of two components: a procedural simulation
framework for generating crowd movements and behaviors, and a procedural
rendering framework to generate different videos or images. Each video or image
is automatically labeled based on the environment, number of pedestrians,
density, behavior, flow, lighting conditions, viewpoint, noise, etc.
Furthermore, we can increase the realism by combining synthetically-generated
behaviors with real-world background videos. We demonstrate the benefits of
LCrowdV over prior lableled crowd datasets by improving the accuracy of
pedestrian detection and crowd behavior classification algorithms. LCrowdV
would be released on the WWW
- …