353 research outputs found
Do the government subsidies inhibit the entity over-financialization? Fresh evidence from China
In order to verify effect of the industrial policies on solving the
problem of market failure, we collect the data from China A-share
listed companies among 2008-2019, and analyze the effect of government
subsidies on the entity over-financialization. The results
show that government subsidies significantly inhibit the entity
over-financialization. Because the government subsidies could
increase the performance of enterprise’s main business and level
of the enterprise’s profitability. Subsequently, the enterprise’s arbitrage
from cross-industries and the managers’ composition could
be decreased. Consequently, government subsidies could reduce
the entity over-financialization by the reduce of enterprise’s arbitrage
from multi-industries and increase of the managers’ composition
which is related to the enterprise’s performance. The results
also indicate that the entity financialization is mainly motivated by
enterprise arbitrage rather than ‘preventive reserve’ in China.
Moreover, the inhibitory effect of government subsidies on the
entity over-financialization is only significant in the enterprises
with non-state-owned, high-tech, and higher level of demand of
innovation. Thus, the government should accurately implement
subsidy policies for the enterprises and increase the supports for
enterprises with high-tech and higher level of demand of innovation,
which could promote economy high-quality development
RegionBLIP: A Unified Multi-modal Pre-training Framework for Holistic and Regional Comprehension
In this work, we investigate extending the comprehension of Multi-modal Large
Language Models (MLLMs) to regional objects. To this end, we propose to extract
features corresponding to regional objects as soft prompts for LLM, which
provides a straightforward and scalable approach and eliminates the need for
LLM fine-tuning. To effectively extract regional features from regular image
features and irregular point cloud features, we present a novel and unified
position-assisted feature extraction module. Furthermore, training an MLLM from
scratch is highly time-consuming. Thus, we propose incrementally extending
existing pre-trained MLLMs to comprehend more modalities and the regional
objects of those modalities. Specifically, we freeze the Q-Former from BLIP-2,
an impressive MLLM, and optimize the modality-specific Lora parameters in
Q-Former and LLM for each newly introduced modality. The freezing of the
Q-Former eliminates the need for extensive pre-training on massive image-text
data. The freezed Q-Former pre-trained from massive image-text data is also
beneficial for the pre-training on image-region-text data. We name our
framework RegionBLIP. We pre-train RegionBLIP on image-region-text,
point-cloud-text, and point-cloud-region-text data. Experimental results verify
that \Ours{} can preserve the image comprehension capability of BILP-2 and
further gain a comprehension of the newly introduced point cloud modality and
regional objects. The Data, Code, and Pre-trained models will be available at
https://github.com/mightyzau/RegionBLIP
- …