1,347 research outputs found

    Digitally enabled community platform and user creativity: perspectives on community identification

    Get PDF
    Drawing on social identity theory, our study investigates the influence mechanism of online community platform on user creativity. Specifically, we focus on three technological features of online community platform—virtual presence, persistent labelling, and deep profiling—and unveil how they stimulate user innovative capabilities from a community identification mediating perspective. The data was collected using an online survey. Data analysis results suggest that all three technological features can foster the development of community identification, which then facilitates two types of social interaction behavior (user-user interaction and user-expert interaction), and ultimately enhances user creativity. By examining the interplay between technological features, social identity, and user creativity, our study provides theoretical and practical implications

    An Improved Pseudopolynomial Time Algorithm for Subset Sum

    Full text link
    We investigate pseudo-polynomial time algorithms for Subset Sum. Given a multi-set XX of nn positive integers and a target tt, Subset Sum asks whether some subset of XX sums to tt. Bringmann proposes an O~(n+t)\tilde{O}(n + t)-time algorithm [Bringmann SODA'17], and an open question has naturally arisen: can Subset Sum be solved in O(n+w)O(n + w) time? Here ww is the maximum integer in XX. We make a progress towards resolving the open question by proposing an O~(n+wt)\tilde{O}(n + \sqrt{wt})-time algorithm.Comment: In first version, we falsely claimed that our algorithm is also able to reconstruct a subset that sums to t. In the latest version, we removed this false claim and explained why we cannot do reconstructio

    Weakly Approximating Knapsack in Subquadratic Time

    Get PDF
    We consider the classic Knapsack problem. Let t and OPT be the capacity and the optimal value, respectively. If one seeks a solution with total profit at least OPT/(1 + ε) and total weight at most t, then Knapsack can be solved in Õ(n + (1/(ε))²) time [Chen, Lian, Mao, and Zhang '24][Mao '24]. This running time is the best possible (up to a logarithmic factor), assuming that (min,+)-convolution cannot be solved in truly subquadratic time [Künnemann, Paturi, and Schneider '17][Cygan, Mucha, Węgrzycki, and Włodarczyk '19]. The same upper and lower bounds hold if one seeks a solution with total profit at least OPT and total weight at most (1 + ε)t. Therefore, it is natural to ask the following question. If one seeks a solution with total profit at least OPT/(1+ε) and total weight at most (1 + ε)t, can Knsapck be solved in Õ(n + (1/(ε))^{2-δ}) time for some constant δ > 0? We answer this open question affirmatively by proposing an Õ(n + (1/(ε))^{7/4})-time algorithm

    Approximating Partition in Near-Linear Time

    Full text link
    We propose an \widetilde{O}(n + 1/\eps)-time FPTAS (Fully Polynomial-Time Approximation Scheme) for the classical Partition problem. This is the best possible (up to a polylogarithmic factor) assuming SETH (Strong Exponential Time Hypothesis) [Abboud, Bringmann, Hermelin, and Shabtay'22]. Prior to our work, the best known FPTAS for Partition runs in \widetilde{O}(n + 1/\eps^{5/4}) time [Deng, Jin and Mao'23, Wu and Chen'22]. Our result is obtained by solving a more general problem of weakly approximating Subset Sum.Comment: To appear in STOC202

    Identification and Analysis on Surface Deformation in the Urban Area of Nanchang Based on Ps-Insar Method

    Get PDF
    Interferometric Synthetic Aperture Radar (InSAR) technology has emerged as a vital tool for monitoring surface deformation due to its high accuracy and spatial resolution. With the rapid economic development of Nanchang, extensive infrastructure development and construction activities have significantly altered the urban landscape. Underground excavation and groundwater extraction in the region are potential contributors to surface deformation. This study utilized Sentinel-1 satellite data, acquired between September 2018 and May 2023, and applied the Permanent Scatterer Interferometric Synthetic Aperture Radar (PS-InSAR) technique to monitor surface deformation in Nanchang’s urban area. The findings revealed that surface deformation rates in the study area range from −10 mm/a to 6 mm/a, with the majority of regions remaining relatively stable. Approximately 99.9% of the monitored points exhibited deformation rates within −5 mm/a to 5 mm/a. However, four significant subsidence zones were identified along the Gan River and its downstream regions, with a maximum subsidence rate reaching 9.7 mm/a. Historical satellite imagery comparisons indicated that certain subsidence areas are potentially associated with construction activities. Further analysis integrating subsidence data, monthly precipitation, and groundwater depth revealed a negative correlation between surface deformation in Region A and rainfall, with subsidence trends aligning with groundwater level fluctuations. However, such a correlation was not evident in the other three regions. Additionally, water level data from the Xingzi Station of Poyang Lake showed that only Region A’s subsidence trend closely corresponds with water level variations. We conducted a detailed analysis of the spatial distribution of soil types in Nanchang and found that the soil types in areas of surface deformation are primarily Semi-hydromorphic Soils and Anthropogenic Soils. These soils exhibit high compressibility, making them prone to compaction and significantly influencing surface deformation. This study concludes that localized surface deformation in Nanchang is primarily driven by urban construction activities and the compaction of artificial fill soils, while precipitation also has an impact in certain areas

    Faster Algorithms for Bounded Knapsack and Bounded Subset Sum Via Fine-Grained Proximity Results

    Full text link
    We investigate pseudopolynomial-time algorithms for Bounded Knapsack and Bounded Subset Sum. Recent years have seen a growing interest in settling their fine-grained complexity with respect to various parameters. For Bounded Knapsack, the number of items nn and the maximum item weight wmaxw_{\max} are two of the most natural parameters that have been studied extensively in the literature. The previous best running time in terms of nn and wmaxw_{\max} is O(n+wmax3)O(n + w^3_{\max}) [Polak, Rohwedder, Wegrzycki '21]. There is a conditional lower bound of O((n+wmax)2o(1))O((n + w_{\max})^{2-o(1)}) based on (min,+)(\min,+)-convolution hypothesis [Cygan, Mucha, Wegrzycki, Wlodarczyk '17]. We narrow the gap significantly by proposing a O~(n+wmax12/5)\tilde{O}(n + w^{12/5}_{\max})-time algorithm. Note that in the regime where wmaxnw_{\max} \approx n, our algorithm runs in O~(n12/5)\tilde{O}(n^{12/5}) time, while all the previous algorithms require Ω(n3)\Omega(n^3) time in the worst case. For Bounded Subset Sum, we give two algorithms running in O~(nwmax)\tilde{O}(nw_{\max}) and O~(n+wmax3/2)\tilde{O}(n + w^{3/2}_{\max}) time, respectively. These results match the currently best running time for 0-1 Subset Sum. Prior to our work, the best running times (in terms of nn and wmaxw_{\max}) for Bounded Subset Sum is O~(n+wmax5/3)\tilde{O}(n + w^{5/3}_{\max}) [Polak, Rohwedder, Wegrzycki '21] and O~(n+μmax1/2wmax3/2)\tilde{O}(n + \mu_{\max}^{1/2}w_{\max}^{3/2}) [implied by Bringmann '19 and Bringmann, Wellnitz '21], where μmax\mu_{\max} refers to the maximum multiplicity of item weights

    Scaling Laws Behind Code Understanding Model

    Full text link
    The scaling law is becoming a fundamental law in many machine learning areas. That is, test error falls off with the power law when increasing training data, model size, and computing resource. However, whether this law is suitable for the task of code understanding is not well studied, and most current language models for code understanding are about 100M parameters, which are relatively "small" compared to large language models. In this paper, we conduct extensive experiments to investigate the scaling law for the code understanding task by varying training data, model size, and computing resource. We validate that the test error of code understanding models falls off with the power law when using larger models, indicating that the scaling law is suitable for the code understanding task. Besides, we apply different scales of models to two downstream code understanding tasks, and find that the performance increases with larger scale of models. Finally, we train a large-scale code understanding model named CoLSBERT with 1.5B parameters on a large dataset using more computing resource, which outperforms previous work by a large margin. We will release our code and the CoLSBERT model when our paper is published
    corecore