49,157 research outputs found
Designing digital technologies and learning activities for different geometries
This chapter focuses on digital technologies and geometry education, a combination of topics that provides a suitable avenue for analysing closely the issues and challenges involved in designing and utilizing digital technologies for learning mathematics. In revealing these issues and challenges, the chapter examines the design of digital technologies and related forms of learning activities for a range of geometries, including Euclidean and co-ordinate geometries in two and three dimensions, and non-Euclidean geometries such as spherical, hyperbolic and fractal geometry. This analysis reveals the decisions that designers take when designing for different geometries on the flat computer screen. Such decisions are not only about the geometry but also about the learner in terms of supporting their perceptions of what are the key features of geometry
TextureNet: Consistent Local Parametrizations for Learning from High-Resolution Signals on Meshes
We introduce, TextureNet, a neural network architecture designed to extract
features from high-resolution signals associated with 3D surface meshes (e.g.,
color texture maps). The key idea is to utilize a 4-rotational symmetric
(4-RoSy) field to define a domain for convolution on a surface. Though 4-RoSy
fields have several properties favorable for convolution on surfaces (low
distortion, few singularities, consistent parameterization, etc.), orientations
are ambiguous up to 4-fold rotation at any sample point. So, we introduce a new
convolutional operator invariant to the 4-RoSy ambiguity and use it in a
network to extract features from high-resolution signals on geodesic
neighborhoods of a surface. In comparison to alternatives, such as PointNet
based methods which lack a notion of orientation, the coherent structure given
by these neighborhoods results in significantly stronger features. As an
example application, we demonstrate the benefits of our architecture for 3D
semantic segmentation of textured 3D meshes. The results show that our method
outperforms all existing methods on the basis of mean IoU by a significant
margin in both geometry-only (6.4%) and RGB+Geometry (6.9-8.2%) settings
Data-Driven Shape Analysis and Processing
Data-driven methods play an increasingly important role in discovering
geometric, structural, and semantic relationships between 3D shapes in
collections, and applying this analysis to support intelligent modeling,
editing, and visualization of geometric data. In contrast to traditional
approaches, a key feature of data-driven approaches is that they aggregate
information from a collection of shapes to improve the analysis and processing
of individual shapes. In addition, they are able to learn models that reason
about properties and relationships of shapes without relying on hard-coded
rules or explicitly programmed instructions. We provide an overview of the main
concepts and components of these techniques, and discuss their application to
shape classification, segmentation, matching, reconstruction, modeling and
exploration, as well as scene analysis and synthesis, through reviewing the
literature and relating the existing works with both qualitative and numerical
comparisons. We conclude our report with ideas that can inspire future research
in data-driven shape analysis and processing.Comment: 10 pages, 19 figure
GroundNet: Monocular Ground Plane Normal Estimation with Geometric Consistency
We focus on estimating the 3D orientation of the ground plane from a single
image. We formulate the problem as an inter-mingled multi-task prediction
problem by jointly optimizing for pixel-wise surface normal direction, ground
plane segmentation, and depth estimates. Specifically, our proposed model,
GroundNet, first estimates the depth and surface normal in two separate
streams, from which two ground plane normals are then computed
deterministically. To leverage the geometric correlation between depth and
normal, we propose to add a consistency loss on top of the computed ground
plane normals. In addition, a ground segmentation stream is used to isolate the
ground regions so that we can selectively back-propagate parameter updates
through only the ground regions in the image. Our method achieves the
top-ranked performance on ground plane normal estimation and horizon line
detection on the real-world outdoor datasets of ApolloScape and KITTI,
improving the performance of previous art by up to 17.7% relatively.Comment: Camera Ready for ACM MM 201
Learning to Calibrate Straight Lines for Fisheye Image Rectification
This paper presents a new deep-learning based method to simultaneously
calibrate the intrinsic parameters of fisheye lens and rectify the distorted
images. Assuming that the distorted lines generated by fisheye projection
should be straight after rectification, we propose a novel deep neural network
to impose explicit geometry constraints onto processes of the fisheye lens
calibration and the distorted image rectification. In addition, considering the
nonlinearity of distortion distribution in fisheye images, the proposed network
fully exploits multi-scale perception to equalize the rectification effects on
the whole image. To train and evaluate the proposed model, we also create a new
largescale dataset labeled with corresponding distortion parameters and
well-annotated distorted lines. Compared with the state-of-the-art methods, our
model achieves the best published rectification quality and the most accurate
estimation of distortion parameters on a large set of synthetic and real
fisheye images
Multi-view X-ray R-CNN
Motivated by the detection of prohibited objects in carry-on luggage as a
part of avionic security screening, we develop a CNN-based object detection
approach for multi-view X-ray image data. Our contributions are two-fold.
First, we introduce a novel multi-view pooling layer to perform a 3D
aggregation of 2D CNN-features extracted from each view. To that end, our
pooling layer exploits the known geometry of the imaging system to ensure
geometric consistency of the feature aggregation. Second, we introduce an
end-to-end trainable multi-view detection pipeline based on Faster R-CNN, which
derives the region proposals and performs the final classification in 3D using
these aggregated multi-view features. Our approach shows significant accuracy
gains compared to single-view detection while even being more efficient than
performing single-view detection in each view.Comment: To appear at the 40th German Conference on Pattern Recognition (GCPR)
201
Predicting Lung Nodule Malignancies by Combining Deep Convolutional Neural Network and Handcrafted Features
To predict lung nodule malignancy with a high sensitivity and specificity, we
propose a fusion algorithm that combines handcrafted features (HF) into the
features learned at the output layer of a 3D deep convolutional neural network
(CNN). First, we extracted twenty-nine handcrafted features, including nine
intensity features, eight geometric features, and twelve texture features based
on grey-level co-occurrence matrix (GLCM) averaged from thirteen directions. We
then trained 3D CNNs modified from three state-of-the-art 2D CNN architectures
(AlexNet, VGG-16 Net and Multi-crop Net) to extract the CNN features learned at
the output layer. For each 3D CNN, the CNN features combined with the 29
handcrafted features were used as the input for the support vector machine
(SVM) coupled with the sequential forward feature selection (SFS) method to
select the optimal feature subset and construct the classifiers. The fusion
algorithm takes full advantage of the handcrafted features and the highest
level CNN features learned at the output layer. It can overcome the
disadvantage of the handcrafted features that may not fully reflect the unique
characteristics of a particular lesion by combining the intrinsic CNN features.
Meanwhile, it also alleviates the requirement of a large scale annotated
dataset for the CNNs based on the complementary of handcrafted features. The
patient cohort includes 431 malignant nodules and 795 benign nodules extracted
from the LIDC/IDRI database. For each investigated CNN architecture, the
proposed fusion algorithm achieved the highest AUC, accuracy, sensitivity, and
specificity scores among all competitive classification models.Comment: 11 pages, 5 figures, 5 tables. This work has been submitted to the
IEEE for possible publicatio
EC-Net: an Edge-aware Point set Consolidation Network
Point clouds obtained from 3D scans are typically sparse, irregular, and
noisy, and required to be consolidated. In this paper, we present the first
deep learning based edge-aware technique to facilitate the consolidation of
point clouds. We design our network to process points grouped in local patches,
and train it to learn and help consolidate points, deliberately for edges. To
achieve this, we formulate a regression component to simultaneously recover 3D
point coordinates and point-to-edge distances from upsampled features, and an
edge-aware joint loss function to directly minimize distances from output
points to 3D meshes and to edges. Compared with previous neural network based
works, our consolidation is edge-aware. During the synthesis, our network can
attend to the detected sharp edges and enable more accurate 3D reconstructions.
Also, we trained our network on virtual scanned point clouds, demonstrated the
performance of our method on both synthetic and real point clouds, presented
various surface reconstruction results, and showed how our method outperforms
the state-of-the-arts.Comment: accepted by ECCV2018; project in https://yulequan.github.io/ec-net
Pixel2Mesh: Generating 3D Mesh Models from Single RGB Images
We propose an end-to-end deep learning architecture that produces a 3D shape
in triangular mesh from a single color image. Limited by the nature of deep
neural network, previous methods usually represent a 3D shape in volume or
point cloud, and it is non-trivial to convert them to the more ready-to-use
mesh model. Unlike the existing methods, our network represents 3D mesh in a
graph-based convolutional neural network and produces correct geometry by
progressively deforming an ellipsoid, leveraging perceptual features extracted
from the input image. We adopt a coarse-to-fine strategy to make the whole
deformation procedure stable, and define various of mesh related losses to
capture properties of different levels to guarantee visually appealing and
physically accurate 3D geometry. Extensive experiments show that our method not
only qualitatively produces mesh model with better details, but also achieves
higher 3D shape estimation accuracy compared to the state-of-the-art
Mini-Unmanned Aerial Vehicle-Based Remote Sensing: Techniques, Applications, and Prospects
The past few decades have witnessed the great progress of unmanned aircraft
vehicles (UAVs) in civilian fields, especially in photogrammetry and remote
sensing. In contrast with the platforms of manned aircraft and satellite, the
UAV platform holds many promising characteristics: flexibility, efficiency,
high-spatial/temporal resolution, low cost, easy operation, etc., which make it
an effective complement to other remote-sensing platforms and a cost-effective
means for remote sensing. Considering the popularity and expansion of UAV-based
remote sensing in recent years, this paper provides a systematic survey on the
recent advances and future prospectives of UAVs in the remote-sensing
community. Specifically, the main challenges and key technologies of
remote-sensing data processing based on UAVs are discussed and summarized
firstly. Then, we provide an overview of the widespread applications of UAVs in
remote sensing. Finally, some prospects for future work are discussed. We hope
this paper will provide remote-sensing researchers an overall picture of recent
UAV-based remote sensing developments and help guide the further research on
this topic
- …