472 research outputs found
Species misidentification in ecological studies : incidence and importance from the ecologists’ point of view
Natural scientists study a wide variety of species, but whether they have identified all studied samples correctly to species is rarely evaluated. Species misidentification in empirical research can cause significant losses of money, information, and time, and contribute to false results. Thus, I study the abundance of species misidentification and ecologists’ perceptions of such mistakes through a web survey targeting researchers from scientific institutes around the globe (including universities, research societies and museums) who completed their doctoral degree in any ecology-related field of science. I received 117 responses with either work or educational background from 30 countries. I found that species misidentification widely existed in respondents’ research: almost 70% of the respondents noticed species misidentification in their own research, while the estimated proportion of existing studies with species misidentification was 34% (95% CI: 28% - 40%). Although misidentification was mainly found during specimen collection, specimen handling and data analysis, misidentifications in reporting stages (writing, revision and after publishing) could persist until publication. Moreover, according to respondents, reviewers seldom comment about species identification methods or their accuracy, which may affect respondents’ (both leading and not leading a research team) low reporting frequency about the possibility of misidentification. Expert checking, training students, and DNA barcoding are the most prevalent approaches to ensure identification accuracy among respondents. My results imply that species misidentification might be widespread in existing ecological research. Although the problem of species misidentification is widely recognized, such an issue seldom be appropriately handled by respondents. To increase the accuracy of species identification and maintain academic integrity, I suggest that researchers need to focus more on the study species (e.g., sampling process, identification method, and accuracy) when writing and reviewing papers. Furthermore, I appeal for guidelines about reporting species identification methods and their accuracy in papers, as well as research on education about identification skills in universities, as these two topics may constrain the precision of species identification
Molecular dynamics simulation study of grain boundary migration in nanocrystalline Pd
We present a new methodology for measuring the grain boundary mobility for curved boundaries using molecular-dynamics simulation of grain growth in a small, specifically tailored Pd nanocrystalline structure. In the model system, the boundaries move under the forces provided by their curvature and in the presence of the triple junctions. As a consequence of grain boundary migration the boundary area per unit volume is reduced and the mean grain size of grains increases with time. Our investigation shows that at elevated temperatures the activation energy for grain growth in this specifically tailored microstructure is very close to that of grain boundary diffusion. These findings suggest that the migration mechanism of curved grain boundaries might be mediated by short distance diffusion of atoms in the grain boundaries
Supply interruption supply chain network model with uncertain demand: an application of chance-constrained programming with fuzzy parameters
The downstream supply interruption of manufacturers is a disaster for the company when the demand is uncertain in the market; a fuzzy programming with fuzzy parameters model of supply interruption supply chain network is established by simulating market operation rules. The aim of the current study is to build a fuzzy chance-constrained programming method which is developed for supporting the uncertainty of demand. This method ensured that the fuzzy constraints can be satisfied at specified confidence levels, leading to cost-effective solutions under acceptable risk magnitudes. Finally, through the case of the electronic product manufacturing enterprise, the feasibility and effectiveness of the proposed model are verified by adopting a sensitivity analysis of capacity loss level and minimizing objective function. Numerical simulation shows that selecting two manufacturing centers can effectively reduce the supply chain cost and maintain business continuity
Uso de materiales compuestos en el diseño de un árbol de transmisión
El presente proyecto tiene por objetivo el diseño de un árbol de
transmisión tubular utilizando los materiales compuestos, dada su importancia
para la aplicación en la industria automovilística como sustituto de los árboles de
transmisión convencional.
Al sustituir el material convencional como el acero por los materiales
compuestos, se consigue hasta un 50% de la reducción del peso del árbol, sin
modificar el comportamiento estructural del mismo. Asimismo, la posibilidad que
ofrecen los materiales compuestos de seleccionar el número de láminas, la
orientación del refuerzo de las mismas, su secuencia de apilamiento, etc.,
permiten un mejor aprovechamiento estructural del mismo.
Para conseguir un diseño factible con máxima reducción del peso, el
primer paso es la selección de los posibles materiales compuestos para el diseño
del árbol. Una vez seleccionados los materiales, se debe estudiar las cargas que
sufre dicho árbol. Para determinar el comportamiento del árbol frente al estado
de cargas, se construye un modelo matemático basado en la teoría clásica de
laminados.
Para obtener un diseño óptimo, que aproveche mejor las propiedades
estructurales del material, se ha desarrollado el programa OPTIEJE, mediante el
software Matlab.
Una vez obtenido el diseño analítico optimizado, se realiza un modelo de
elementos finitos del mismo mediante el programa ANSYS. Dicho análisis
numérico incluye un análisis estático para determinar el estado tensional y de
deformación en dicho elemento mecánico, un análisis modal del mismo para
determinar la frecuencia natural y un análisis de pandeo para determinar la
resistencia a pandeo del árbol de transmisión. Este análisis por el MEF permite,
además, determinar la validez del diseño de la unión adhesiva.
El paso final es la comparación de los resultados obtenidos por el diseño
analítico y los obtenidos por la simulación por el MEF, así como la determinación
del error cometido y la viabilidad del diseño.This document is mandated to design a tubular drive shaft with composite
materials, given its importance for the application in the automotive industry as a
substitute for traditional metallic drive shaft.
Replacing conventional materials like steel by composite materials, a 50%
weight reduction is archived without modifying performance of the drive shaft.
Furthermore, composite materials offer the possibility of the selection of the
number of layers, fiber orientation of each layer and stacking sequence, which
allow a better exploitation of structural proprieties of these materials.
In order to obtain an optimum design of the drive shaft with maximum
weight saving this document begins with a composite material selection. Then the
loads should to be defined once the materials are selected. To study the drive
shaft’s behavior against the loads a mathematic model based on the classical
laminate theory is built.
To find out an optimum design that takes maximum advantages of the
structural proprieties of the material, a Matlab based algorithm will be used.
Once obtained an analytical optimized design, a finite elements analysis
will be realized using the software ANSYS. The mentioned analysis contains a
static analysis to study the stress and deformation results of the mechanical
element; a modal analysis of this element to find out the natural frequency, a
buckling analysis to find out the buckling capacity of the drive shaft. Furthermore,
the finite elements analysis allows to determine the validity of the adhesive
bonding.
Finally a comparison of the analytic results and the ANSYS results is
necessary to determine the error between both method and the feasibility of the
designIngeniería Industria
Performance optimization of convolution calculation by blocking and sparsity on GPU
Convolution neural network (CNN) plays a paramount role in machine learning,
which has made significant contributions in medical image classification,
natural language processing, recommender system and so on. A successful
convolution neural network can achieve excellent performance with fast
execution time. The convolution operation dominates the total operation time of
convolution neural network. Therefore, in this paper, we propose a novel
convolution method on Graphic Processing Units (GPUs), which reduces the
convolution operation time and improves the execution speed by approximately 2X
than the state of the art convolution algorithm. Our work is based on the
observation that the sparsity of the input feature map of convolution operation
is relatively large, and the zero value of the feature map is redundancy for
convolution result. Therefore, we skip the zero value calculation and improve
the speed by compressing the feature map. Besides, the shape of the feature map
for the deep network is small, and the number of threads is limited. Therefore,
for a limited number of threads, it is necessary to reduce the amount of
calculation to increase the calculation speed. Our algorithm has a good effect
on the convolution operation for the feature map of the deep network with large
sparsity and small size
Learning Accurate Performance Predictors for Ultrafast Automated Model Compression
In this paper, we propose an ultrafast automated model compression framework
called SeerNet for flexible network deployment. Conventional
non-differen-tiable methods discretely search the desirable compression policy
based on the accuracy from exhaustively trained lightweight models, and
existing differentiable methods optimize an extremely large supernet to obtain
the required compressed model for deployment. They both cause heavy
computational cost due to the complex compression policy search and evaluation
process. On the contrary, we obtain the optimal efficient networks by directly
optimizing the compression policy with an accurate performance predictor, where
the ultrafast automated model compression for various computational cost
constraint is achieved without complex compression policy search and
evaluation. Specifically, we first train the performance predictor based on the
accuracy from uncertain compression policies actively selected by efficient
evolutionary search, so that informative supervision is provided to learn the
accurate performance predictor with acceptable cost. Then we leverage the
gradient that maximizes the predicted performance under the barrier complexity
constraint for ultrafast acquisition of the desirable compression policy, where
adaptive update stepsizes with momentum are employed to enhance optimality of
the acquired pruning and quantization strategy. Compared with the
state-of-the-art automated model compression methods, experimental results on
image classification and object detection show that our method achieves
competitive accuracy-complexity trade-offs with significant reduction of the
search cost.Comment: Accepted to IJC
Ground-VIO: Monocular Visual-Inertial Odometry with Online Calibration of Camera-Ground Geometric Parameters
Monocular visual-inertial odometry (VIO) is a low-cost solution to provide
high-accuracy, low-drifting pose estimation. However, it has been meeting
challenges in vehicular scenarios due to limited dynamics and lack of stable
features. In this paper, we propose Ground-VIO, which utilizes ground features
and the specific camera-ground geometry to enhance monocular VIO performance in
realistic road environments. In the method, the camera-ground geometry is
modeled with vehicle-centered parameters and integrated into an
optimization-based VIO framework. These parameters could be calibrated online
and simultaneously improve the odometry accuracy by providing stable
scale-awareness. Besides, a specially designed visual front-end is developed to
stably extract and track ground features via the inverse perspective mapping
(IPM) technique. Both simulation tests and real-world experiments are conducted
to verify the effectiveness of the proposed method. The results show that our
implementation could dramatically improve monocular VIO accuracy in vehicular
scenarios, achieving comparable or even better performance than state-of-art
stereo VIO solutions. The system could also be used for the auto-calibration of
IPM which is widely used in vehicle perception. A toolkit for ground feature
processing, together with the experimental datasets, would be made open-source
(https://github.com/GREAT-WHU/gv_tools)
OpenDigger: Data Mining and Information Service System for Open Collaboration Digital Ecosystem
The widespread development and adoption of open-source software have built an
ecosystem for open development and collaboration. In this ecosystem,
individuals and organizations collaborate to create high-quality software that
can be used by everyone. Social collaboration platforms like GitHub have
further facilitated large-scale, distributed, and fine-grained code
collaboration and technical interactions. Countless developers contribute code,
review code, report bugs, and propose new features on these platforms every
day, generating a massive amount of valuable behavioral data from the open
collaboration process. This paper presents the design and implementation of
OpenDigger, a comprehensive data mining and information service system for open
collaboration in the digital ecosystem. The goal is to build a data
infrastructure for the open-source domain and promote the continuous
development of the open-source ecosystem. The metrics and analysis models in
the OpenDigger system can mine various knowledge from the macro to micro levels
in the open-source digital ecosystem. Through a unified information service
interface, OpenDigger provides various open-source information services to
different user groups, including governments, enterprises, foundations, and
individuals. As a novel information service system in the open-source
ecosystem, this paper demonstrates the effectiveness of the metrics and models
in OpenDigger through several real-world scenarios, including products, tools,
applications, and courses. It showcases the significant and diverse practical
applications of the metrics and models in both algorithmic and business
aspects.Comment: in Chinese languag
- …