110 research outputs found
A NEW METHOD TO CONTROL THE REGIONAL STRATA MOVEMENT OF SUPER-THICK WEAK CEMENTATION OVERBURDEN IN DEEP MINING
In the western of china, the deep mining area with super-thick and weak cementation overburden is vast, sparsely populated and the ecological environment is extremely fragile. With the large-scale exploitation of deep coal resources, it is inevitable to face green mining problem, whose essence is the surface subsidence control. Therefore, it is necessary to study the control technology for the regional mining based on the evolution law of subsidence movement and energy-polling of super-thick and weak cementation overburden, and put forward the economically design scheme that can control strata movement and surface subsidence in a certain degree. Based on the key strata control theory, this paper puts forward the subsidence control scheme of partial filling -partial caving in multi-working face coordinated mining, and further studies its control mechanism through the numerical simulation and then analyzes the control effect of the strata movement and energy-polling in the fully caving mining, backfill mining, wide strip skip-mining and mixed filling mining method etc., the following conclusions are detailed as follows: (1) The maximum value of energy-polling occurs on the coal pillars or on both sides of goaf. With the width of goaf, the maximum value of energy-polling increases in a parabola. (2) In the partial filling-partial caving multiple working faces coordinated mining based on the main key stratum, the stress distribution of the composite backfill in the filling working face is parabolic, and it is high on both sides and low in the middle. Moreover, in the composite backfill, the stress concentration degree of a outside coal pillar is greater than that of the inside coal pillar. (3)The control mechanism of partial filling-partial caving harmonious mining based on main key layer structure is the double-control cooperative deformation system, formed by the composite backfill and the main and sub-key layers structure. They jointly control the movement and energy accumulation of overlying strata by greatly reducing the effective space to transmit upward, and absorb the wave subsidence trend of the overburden until it develops into a single flat subsidence basin. (4) Considering the recovery rate, pillar rate, area filling rate, technical difficulty and subsidence coefficient etc., the partial filling-partial caving multiple working faces coordinated mining based on the main key stratum is the most cost-effective mining method to control surface subsidence. This paper takes a guiding role in controlling the regional strata movement and surface subsidence of deep mining with super-thick and weak cementation overburden
TransVCL: Attention-enhanced Video Copy Localization Network with Flexible Supervision
Video copy localization aims to precisely localize all the copied segments
within a pair of untrimmed videos in video retrieval applications. Previous
methods typically start from frame-to-frame similarity matrix generated by
cosine similarity between frame-level features of the input video pair, and
then detect and refine the boundaries of copied segments on similarity matrix
under temporal constraints. In this paper, we propose TransVCL: an
attention-enhanced video copy localization network, which is optimized directly
from initial frame-level features and trained end-to-end with three main
components: a customized Transformer for feature enhancement, a correlation and
softmax layer for similarity matrix generation, and a temporal alignment module
for copied segments localization. In contrast to previous methods demanding the
handcrafted similarity matrix, TransVCL incorporates long-range temporal
information between feature sequence pair using self- and cross- attention
layers. With the joint design and optimization of three components, the
similarity matrix can be learned to present more discriminative copied
patterns, leading to significant improvements over previous methods on
segment-level labeled datasets (VCSL and VCDB). Besides the state-of-the-art
performance in fully supervised setting, the attention architecture facilitates
TransVCL to further exploit unlabeled or simply video-level labeled data.
Additional experiments of supplementing video-level labeled datasets including
SVD and FIVR reveal the high flexibility of TransVCL from full supervision to
semi-supervision (with or without video-level annotation). Code is publicly
available at https://github.com/transvcl/TransVCL.Comment: Accepted by the Thirty-Seventh AAAI Conference on Artificial
Intelligence(AAAI2023
Rare failure event analysis of structures under mixed uncertainties
Two challenges may exist in the reliability analysis of highly reliable structures in, e.g., aerospace engineering. The first one is that, the failure probability may be extremely small (typically, smaller than 1e-6), which commonly prevents us from generating accurate estimation with acceptable computational costs by using the available methods. The second one is that, the available information for the input variables may be subject to incompleteness (e.g., sparse data) and/or imprecision (e.g., measuring error), which, makes it impossible to generate precise probability models for the input variables. To address the above two challenges, this work proposes two effective algorithms based on combining the sampling techniques (i.e., extended Monte Carlo simulation and subset simulation), active learning techniques and high-dimensional model representation decomposition. The proposed methods can effectively estimate the failure probability function w.r.t. the uncertain distribution parameters of the input variables with small number of training samples even when the failure event is extremely rare. A numerical test example is introduced to illustrate the proposed methods
Optimization or Bayesian strategy? Performance of the Bhattacharyya distance in different algorithms of stochastic model updating
The Bhattacharyya distance has been developed as a comprehensive uncertainty quantification metric by capturing multiple uncertainty sources from both numerical predictions and experimental measurements. This work pursues a further investigation of the performance of the Bhattacharyya distance in different methodologies for stochastic model updating, and thus to prove the universality of the Bhattacharyya distance in various currently popular updating procedures. The first procedure is the Bayesian model updating where the Bhattacharyya distance is utilized to define an approximate likelihood function and the transitional Markov chain Monte Carlo algorithm is employed to obtain the posterior distribution of the parameters. In the second updating procedure, the Bhattacharyya distance is utilized to construct the objective function of an optimization problem. The objective function is defined as the Bhattacharyya distance between the samples of numerical prediction and the samples of the target data. The comparison study is performed on a four degrees-of-freedom mass-spring system. A challenging task is raised in this example by assigning different distributions to the parameters with imprecise distribution coefficients. This requires the stochastic updating procedure to calibrate not the parameters themselves, but their distribution properties. The second example employs the GARTEUR SM-AG19 benchmark structure to demonstrate the feasibility of the Bhattacharyya distance in the presence of practical experiment uncertainty raising from measuring techniques, equipment, and subjective randomness. The results demonstrate the Bhattacharyya distance as a comprehensive and universal uncertainty quantification metric in stochastic model updating
Learning Segment Similarity and Alignment in Large-Scale Content Based Video Retrieval
With the explosive growth of web videos in recent years, large-scale
Content-Based Video Retrieval (CBVR) becomes increasingly essential in video
filtering, recommendation, and copyright protection. Segment-level CBVR
(S-CBVR) locates the start and end time of similar segments in finer
granularity, which is beneficial for user browsing efficiency and infringement
detection especially in long video scenarios. The challenge of S-CBVR task is
how to achieve high temporal alignment accuracy with efficient computation and
low storage consumption. In this paper, we propose a Segment Similarity and
Alignment Network (SSAN) in dealing with the challenge which is firstly trained
end-to-end in S-CBVR. SSAN is based on two newly proposed modules in video
retrieval: (1) An efficient Self-supervised Keyframe Extraction (SKE) module to
reduce redundant frame features, (2) A robust Similarity Pattern Detection
(SPD) module for temporal alignment. In comparison with uniform frame
extraction, SKE not only saves feature storage and search time, but also
introduces comparable accuracy and limited extra computation time. In terms of
temporal alignment, SPD localizes similar segments with higher accuracy and
efficiency than existing deep learning methods. Furthermore, we jointly train
SSAN with SKE and SPD and achieve an end-to-end improvement. Meanwhile, the two
key modules SKE and SPD can also be effectively inserted into other video
retrieval pipelines and gain considerable performance improvements.
Experimental results on public datasets show that SSAN can obtain higher
alignment accuracy while saving storage and online query computational cost
compared to existing methods.Comment: Accepted by ACM MM 202
Escape-zone-based optimal evasion guidance against multiple orbital pursuers
The orbital evasion problem is getting increasing attention because of the increase of space maneuvering objects. In this paper, an escape-zone-based optimal orbital evasion guidance law for an evading spacecraft on near circular reference orbit is proposed against multiple pursuing spacecraft with impulsive thrust. The relative reachable domain is introduced first and approximated as an ellipsoid propagating along the nominal trajectory under the short-term assumption. The escape zone for the impulsive evasion problem is presented herein as a geometric description of the set of terminal positions for all the impulsive evasion trajectories that are not threatened by the maneuvers of pursuers at the maneuver moment. A general method is developed next to calculate the defined escape zone through finding the intersection of two relative reachable domain approximate ellipsoids at arbitrary intersection moment. Then, the two-sided optimal strategies for the orbital evasion problem are analyzed according to whether the escape zone exists, based on which the escape value is defined and used as the basis of the proposed orbital evasion guidance scheme. Finally, numerical examples demonstrate the usefulness of the presented method for calculating escape zone and the effectiveness of the proposed evasion guidance scheme against multiple pursuing spacecraft
Video Infringement Detection via Feature Disentanglement and Mutual Information Maximization
The self-media era provides us tremendous high quality videos. Unfortunately,
frequent video copyright infringements are now seriously damaging the interests
and enthusiasm of video creators. Identifying infringing videos is therefore a
compelling task. Current state-of-the-art methods tend to simply feed
high-dimensional mixed video features into deep neural networks and count on
the networks to extract useful representations. Despite its simplicity, this
paradigm heavily relies on the original entangled features and lacks
constraints guaranteeing that useful task-relevant semantics are extracted from
the features.
In this paper, we seek to tackle the above challenges from two aspects: (1)
We propose to disentangle an original high-dimensional feature into multiple
sub-features, explicitly disentangling the feature into exclusive
lower-dimensional components. We expect the sub-features to encode
non-overlapping semantics of the original feature and remove redundant
information.
(2) On top of the disentangled sub-features, we further learn an auxiliary
feature to enhance the sub-features. We theoretically analyzed the mutual
information between the label and the disentangled features, arriving at a loss
that maximizes the extraction of task-relevant information from the original
feature.
Extensive experiments on two large-scale benchmark datasets (i.e., SVD and
VCSL) demonstrate that our method achieves 90.1% TOP-100 mAP on the large-scale
SVD dataset and also sets the new state-of-the-art on the VCSL benchmark
dataset. Our code and model have been released at
https://github.com/yyyooooo/DMI/, hoping to contribute to the community.Comment: This paper is accepted by ACM MM 202
Engineering human ventricular heart muscles based on a highly efficient system for purification of human pluripotent stem cell-derived ventricular cardiomyocytes
Background
Most infarctions occur in the left anterior descending coronary artery and cause myocardium damage of the left ventricle. Although current pluripotent stem cells (PSCs) and directed cardiac differentiation techniques are able to generate fetal-like human cardiomyocytes, isolation of pure ventricular cardiomyocytes has been challenging. For repairing ventricular damage, we aimed to establish a highly efficient purification system to obtain homogeneous ventricular cardiomyocytes and prepare engineered human ventricular heart muscles in a dish.
Methods
The purification system used TALEN-mediated genomic editing techniques to insert the neomycin or EGFP selection marker directly after the myosin light chain 2 (MYL2) locus in human pluripotent stem cells. Purified early ventricular cardiomyocytes were estimated by immunofluorescence, fluorescence-activated cell sorting, quantitative PCR, microelectrode array, and patch clamp. In subsequent experiments, the mixture of mature MYL2-positive ventricular cardiomyocytes and mesenchymal cells were cocultured with decellularized natural heart matrix. Histological and electrophysiology analyses of the formed tissues were performed 2 weeks later.
Results
Human ventricular cardiomyocytes were efficiently isolated based on the purification system using G418 or flow cytometry selection. When combined with the decellularized natural heart matrix as the scaffold, functional human ventricular heart muscles were prepared in a dish.
Conclusions
These engineered human ventricular muscles can be great tools for regenerative therapy of human ventricular damage as well as drug screening and ventricular-specific disease modeling in the future.
Electronic supplementary material
The online version of this article (doi:10.1186/s13287-017-0651-x) contains supplementary material, which is available to authorized users
Research on Index System for Disabled Elders Evaluation and Grey Clustering Model Based on End-point Mixed Possibility Functions
The file attached to this record is the Publisher's final version.An operational ability assessment system for older adults is of great help to
address health and social challenges for ageing. In this paper, the main problems
in currently available ADL and ability evaluation systems have been analyzed. The
basic principles to build an index system for disability elders evaluation have been
put forwarded. Then,an improved Barthel index system for ADL evaluation and a
new older adults ability evaluation system consisted of 4 first-level indexes and 14
secondary indexes based on experts’ opinion and the ability assessment system for
older adults by Ministry of Civil Affairs of China have been built. The grey
clustering model based on end-point mixed triangular possibility function has been
introduced. And three living examples of adults’ disability evaluation have been
conducted. It is confirmed clearly that the three older adults belong to different
categories of "severe disability", "mild disability", and "ability passable"
respectively. The research results can be used as reference for government to
formulate the elderly-care policies, to run and allocate the elderly-care resources,
as well as reference for various nursing or elderly-care institutions
Heterochromatin protein 1α mediates development and aggressiveness of neuroendocrine prostate cancer
Neuroendocrine prostate cancer (NEPC) is a lethal subtype of prostate cancer (PCa) arising mostly from adenocarcinoma via NE transdifferentiation following androgen deprivation therapy. Mechanisms contributing to both NEPC development and its aggressiveness remain elusive. In light of the fact that hyperchromatic nuclei are a distinguishing histopathological feature of NEPC, we utilized transcriptomic analyses of our patient-derived xenograft (PDX) models, multiple clinical cohorts, and genetically engineered mouse models to identify 36 heterochromatin-related genes that are significantly enriched in NEPC. Longitudinal analysis using our unique, first-in-field PDX model of adenocarcinoma-to-NEPC transdifferentiation revealed that, among those 36 heterochromatin-related genes, heterochromatin protein 1α (HP1α) expression increased early and steadily during NEPC development and remained elevated in the developed NEPC tumor. Its elevated expression was further confirmed in multiple PDX and clinical NEPC samples. HP1α knockdown in the NCI-H660 NEPC cell line inhibited proliferation, ablated colony formation, and induced apoptotic cell death, ultimately leading to tumor growth arrest. Its ectopic expression significantly promoted NE transdifferentiation in adenocarcinoma cells subjected to androgen deprivation treatment. Mechanistically, HP1α reduced expression of androgen receptor (AR) and RE1 silencing transcription factor (REST) and enriched the repressive trimethylated histone H3 at Lys9 (H3K9me3) mark on their respective gene promoters. These observations indicate a novel mechanism underlying NEPC development mediated by abnormally expressed heterochromatin genes, with HP1α as an early functional mediator and a potential therapeutic target for NEPC prevention and management
- …