300 research outputs found

    Pancake bouncing on superhydrophobic surfaces

    Full text link
    Engineering surfaces that promote rapid drop detachment is of importance to a wide range of applications including anti-icing, dropwise condensation6, and self-cleaning. Here we show how superhydrophobic surfaces patterned with lattices of submillimetre-scale posts decorated with nano-textures can generate a counter-intuitive bouncing regime: drops spread on impact and then leave the surface in a flattened, pancake shape without retracting. This allows for a four-fold reduction in contact time compared to conventional complete rebound. We demonstrate that the pancake bouncing results from the rectification of capillary energy stored in the penetrated liquid into upward motion adequate to lift the drop. Moreover, the timescales for lateral drop spreading over the surface and for vertical motion must be comparable. In particular, by designing surfaces with tapered micro/nanotextures which behave as harmonic springs, the timescales become independent of the impact velocity, allowing the occurrence of pancake bouncing and rapid drop detachment over a wide range of impact velocities.Comment: 11 pages, 4 figures, 31 references, + 5 pages of supplementary informatio

    Lightweight Salient Object Detection in Optical Remote-Sensing Images via Semantic Matching and Edge Alignment

    Full text link
    Recently, relying on convolutional neural networks (CNNs), many methods for salient object detection in optical remote sensing images (ORSI-SOD) are proposed. However, most methods ignore the huge parameters and computational cost brought by CNNs, and only a few pay attention to the portability and mobility. To facilitate practical applications, in this paper, we propose a novel lightweight network for ORSI-SOD based on semantic matching and edge alignment, termed SeaNet. Specifically, SeaNet includes a lightweight MobileNet-V2 for feature extraction, a dynamic semantic matching module (DSMM) for high-level features, an edge self-alignment module (ESAM) for low-level features, and a portable decoder for inference. First, the high-level features are compressed into semantic kernels. Then, semantic kernels are used to activate salient object locations in two groups of high-level features through dynamic convolution operations in DSMM. Meanwhile, in ESAM, cross-scale edge information extracted from two groups of low-level features is self-aligned through L2 loss and used for detail enhancement. Finally, starting from the highest-level features, the decoder infers salient objects based on the accurate locations and fine details contained in the outputs of the two modules. Extensive experiments on two public datasets demonstrate that our lightweight SeaNet not only outperforms most state-of-the-art lightweight methods but also yields comparable accuracy with state-of-the-art conventional methods, while having only 2.76M parameters and running with 1.7G FLOPs for 288x288 inputs. Our code and results are available at https://github.com/MathLee/SeaNet.Comment: 11 pages, 4 figures, Accepted by IEEE Transactions on Geoscience and Remote Sensing 202

    U-model based predictive control for nonlinear processes with input delay

    Get PDF
    In this paper, a general control scheme is proposed for nonlinear dynamic processes with input delay described by different models, including polynomial models, state-space models, nonlinear autoregressive moving average with eXogenous inputs (NARMAX) models, Hammerstein or Wiener type models. To tackle the input delay and nonlinear dynamics involved with the control system design, it integrates the classical Smith predictor and a U-model based controller into a U-model based predictive control scheme, which gives a general solution of two-degree-of-freedom (2DOF) control for the set-point tracking and disturbance rejection, respectively. Both controllers are analytically designed by proposing thedesired transfer functions for the above objectives in terms of a linear system expression with the U-model, and therefore are independent of the process model for implementation. Meanwhile, the control system robust stability is analyzed in the presence of process uncertainties. To demonstrate the control performance and advantage, three examples from the literature are conducted with a user-friendly step by step procedure for the ease of understanding by readers

    Analysis of Multi-Element Blended Course Teaching and Learning Mode Based on Student-Centered Concept under the Perspective of ā€œInternet+ā€

    Get PDF
    The integration of Internet and education has changed studentsā€™ learning environment and affected their learning behavior, which poses a greater challenge to the traditional teaching mode. Through the SWOT analysis of the ā€œstudent centeredā€ multi-element blended teaching mode in the era of ā€œInternet + educationā€, it is concluded that the adaptability of learners themselves and the mismatch between teachersā€™ educational ideas and this teaching model delay the development of education to a certain extent. Some suggestions are put forward, such as strengthening the supervision and guidance, implementing the teaching and learning model scientifically, improving teachersā€™ ideology and comprehensive quality, and making full use of the characteristics of Internet opening, sharing and collaboration to construct the public service system and platform of national educational resources

    Evaluate and Guard the Wisdom of Crowds: Zero Knowledge Proofs for Crowdsourcing Truth Inference

    Full text link
    Due to the risks of correctness and security in outsourced cloud computing, we consider a new paradigm called crowdsourcing: distribute tasks, receive answers and aggregate the results from multiple entities. Through this approach, we can aggregate the wisdom of the crowd to complete tasks, ensuring the accuracy of task completion while reducing the risks posed by the malicious acts of a single entity. However, the ensuing question is, how can we ensure that the aggregator has done its work honestly and each contributor's work has been evaluated fairly? In this paper, we propose a new scheme called zkTI\mathsf{zkTI}. This scheme ensures that the aggregator has honestly completed the aggregation and each data source is fairly evaluated. We combine a cryptographic primitive called \textit{zero-knowledge proof} with a class of \textit{truth inference algorithms} which is widely studied in AI/ML scenarios. Under this scheme, various complex outsourced tasks can be solved with efficiency and accuracy. To build our scheme, a novel method to prove the precise computation of floating-point numbers is proposed, which is nearly optimal and well-compatible with existing argument systems. This may become an independent point of interest. Thus our work can prove the process of aggregation and inference without loss of precision. We fully implement and evaluate our ideas. Compared with recent works, our scheme achieves 2āˆ’4Ɨ2-4 \times efficiency improvement and is robust to be widely applied
    • ā€¦
    corecore