267 research outputs found
Fast -NNG construction with GPU-based quick multi-select
In this paper we describe a new brute force algorithm for building the
-Nearest Neighbor Graph (-NNG). The -NNG algorithm has many
applications in areas such as machine learning, bio-informatics, and clustering
analysis. While there are very efficient algorithms for data of low dimensions,
for high dimensional data the brute force search is the best algorithm. There
are two main parts to the algorithm: the first part is finding the distances
between the input vectors which may be formulated as a matrix multiplication
problem. The second is the selection of the -NNs for each of the query
vectors. For the second part, we describe a novel graphics processing unit
(GPU) -based multi-select algorithm based on quick sort. Our optimization makes
clever use of warp voting functions available on the latest GPUs along with
use-controlled cache. Benchmarks show significant improvement over
state-of-the-art implementations of the -NN search on GPUs
No Cash and No Purse: Explaining Non-Monetary Trade in Russia in the 1990s
Why were most transactions in the Russian industry in the 1990s carried out without the use of money? Theories explaining the phenomenon, when carefully assessed, seem to have missing gaps in the argument or inadequate evidence. The present paper critically reviews the theories, suggests neglected considerations, proposes an alternative explanation, and empirically tests the hypothesis.
A large representative sample of Russian firms is used in the empirical part. TOBIT analysis shows that firms start to use non-monetary payments because of their liquidity problems. Further use of non-monetary payments is connected to kartoteka, a tax collection method of withdrawal of taxes from the firm's bank account
Improving the methodology of main power equipment choice for the gas turbine plants
Our paper considers the problem of economic substantiation of the choice of the main power equipment at the stage of functional studies of investment projects in conditions of uncertainty and incompleteness of initial data. As a solution to the designated problem we suggest using the method of the best equipment for gas turbine power plant choice. The method is based on an optimality criterion of power equipment choice which allows us to determine the best solution for the gas turbine from the perspective of capital and operating costs minimizing
Femtomolar detection of the heart failure biomarker NT-proBNP in artificial saliva using an immersible liquid-gated aptasensor with reduced graphene oxide
Measuring NT-proBNP biomarker is recommended for preliminary diagnostics of
the heart failure. Recent studies suggest a possibility of early screening of
biomarkers in saliva for non-invasive identification of cardiac diseases at the
point-of-care. However, NT-proBNP concentrations in saliva can be thousand time
lower than in blood plasma, going down to pg/mL level. To reach this level, we
developed a label-free aptasensor based on a liquid-gated field effect
transistor using a film of reduced graphene oxide monolayer (rGO-FET) with
immobilized NT-proBNP specific aptamer. We found that, depending on ionic
strength of tested solutions, there were different levels of correlation in
responses of electrical parameters of the rGO-FET aptasensor, namely, the Dirac
point shift and transconductance change. The correlation in response to
NT-proBNP was high for 1.6 mM phosphate-buffered saline (PBS) and zero for 16
mM PBS in a wide range of analyte concentrations, varied from 1 fg/mL to 10
ng/mL. The effects of transconductance and Dirac point shift in PBS solutions
of different concentrations are discussed. The biosensor exhibited a high
sensitivity for both transconductance (2 uS/decade) and Dirac point shift (2.3
mV/decade) in diluted PBS with the linear range from 10 fg/mL to 1 pg/mL. The
aptasensor performance has been also demonstrated in undiluted artificial
saliva with the achieved limit of detection down to 41 fg/mL (~4.6 fM)
NiemannâPick type C disease as proof-of-concept for intelligent biomarker panel selection in neurometabolic disorders
AIM: Using Niemann-Pick type C disease (NPC) as a paradigm, we aimed to improve biomarker discovery in patients with neurometabolic disorders. METHOD: Using a multiplexed liquid chromatography tandem mass spectrometry dried bloodspot assay, we developed a selective intelligent biomarker panel to monitor known biomarkers N-palmitoyl-O-phosphocholineserine and 3ÎČ,5α,6ÎČ-trihydroxy-cholanoyl-glycine as well as compounds predicted to be affected in NPC pathology. We applied this panel to a clinically relevant paediatric patient cohort (n = 75; 35 males, 40 females; mean age 7âyears 6âmonths, range 4âdays-19âyears 8âmonths) presenting with neurodevelopmental and/or neurodegenerative pathology, similar to that observed in NPC. RESULTS: The panel had a far superior performance compared with individual biomarkers. Namely, NPC-related established biomarkers used individually had 91% to 97% specificity but the combined panel had 100% specificity. Moreover, multivariate analysis revealed long-chain isoforms of glucosylceramide were elevated and very specific for patients with NPC. INTERPRETATION: Despite advancements in next-generation sequencing and precision medicine, neurological non-enzymatic disorders remain difficult to diagnose and lack robust biomarkers or routine functional testing for genetic variants of unknown significance. Biomarker panels may have better diagnostic accuracy than individual biomarkers in neurometabolic disorders, hence they can facilitate more prompt disease identification and implementation of emerging targeted, disease-specific therapies
Accelerating the Gillespie Ï-Leaping Method Using Graphics Processing Units
The Gillespie Ï-Leaping Method is an approximate algorithm that is faster than the exact Direct Method (DM) due to the progression of the simulation with larger time steps. However, the procedure to compute the time leap Ï is quite expensive. In this paper, we explore the acceleration of the Ï-Leaping Method using Graphics Processing Unit (GPUs) for ultra-large networks ( reaction channels). We have developed data structures and algorithms that take advantage of the unique hardware architecture and available libraries. Our results show that we obtain a performance gain of over 60x when compared with the best conventional implementations
ĐĐĄĐĐĐĐĐĐХйРЀĐĐ ĐĐĐ ĐĐĐĐĐĐŻ ĐąĐĐĐĐĐ„ ĐĐĐĐĐĐ ĐĐ ĐĐĐĐĐŻ, ĐĐĄĐĐĐĐĐĐĐ«Đ„ ĐĐĐĐĐйРĐĐĐĐ«Đ Đ ĐĐĄĐĐ«ĐĐĐĐĐĐ
The surface morphology and optical properties of Si coatings formed by magnetron sputtering were studied using atomic force microscopy, scanning electron microscopy, and spectrophotometry methods. The possibility to influence the surface morphology of coating (filamentous structures and/or round holes) and the location of maxima and minima in reflectance and transmittance via a controllable variation of magnetron sputtering regimes (substrate temperature and bias potential) is shown. ĐĐ”ŃĐŸĐŽĐ°ĐŒĐž Đ°ŃĐŸĐŒĐœĐŸ-ŃĐžĐ»ĐŸĐČĐŸĐč Đž ŃĐșĐ°ĐœĐžŃŃŃŃĐ”Đč ŃлДĐșŃŃĐŸĐœĐœĐŸĐč ĐŒĐžĐșŃĐŸŃĐșĐŸĐżĐžĐž, Đ° ŃĐ°ĐșжД ŃпДĐșŃŃĐŸŃĐŸŃĐŸĐŒĐ”ŃŃОО ĐžŃŃĐ»Đ”ĐŽĐŸĐČĐ°ĐœŃ ĐŒĐŸŃŃĐŸĐ»ĐŸĐłĐžŃ ĐżĐŸĐČĐ”ŃŃ
ĐœĐŸŃŃĐž Đž ĐŸĐżŃĐžŃĐ”ŃĐșОД Ń
Đ°ŃĐ°ĐșŃĐ”ŃĐžŃŃĐžĐșĐž ŃĐŸĐœĐșĐžŃ
Si-ĐżĐŸĐșŃŃŃĐžĐč, ŃŃĐŸŃĐŒĐžŃĐŸĐČĐ°ĐœĐœŃŃ
ĐŒĐ”ŃĐŸĐŽĐŸĐŒ ĐŒĐ°ĐłĐœĐ”ŃŃĐŸĐœĐœĐŸĐłĐŸ ŃĐ°ŃĐżŃĐ»Đ”ĐœĐžŃ. ĐĐŸĐșĐ°Đ·Đ°ĐœĐŸ, ŃŃĐŸ ĐżŃĐž ĐșĐŸĐœŃŃĐŸĐ»ĐžŃŃĐ”ĐŒĐŸĐč ĐČĐ°ŃОаŃОО ŃĐ”Ń
ĐœĐŸĐ»ĐŸĐłĐžŃĐ”ŃĐșĐžŃ
паŃĐ°ĐŒĐ”ŃŃĐŸĐČ ĐŒĐ°ĐłĐœĐ”ŃŃĐŸĐœĐœĐŸĐłĐŸ ŃĐ°ŃĐżŃĐ»Đ”ĐœĐžŃ ŃĐ°ĐșĐžŃ
, ĐșĐ°Đș ŃĐ”ĐŒĐżĐ”ŃĐ°ŃŃŃĐ° ĐżĐŸĐŽĐ»ĐŸĐ¶ĐșĐž Đž ĐżĐŸŃĐ”ĐœŃОал ŃĐŒĐ”ŃĐ”ĐœĐžŃ, ĐŒĐŸĐ¶ĐœĐŸ ĐŒĐ”ĐœŃŃŃ ĐŒĐŸŃŃĐŸĐ»ĐŸĐłĐžŃ ĐżĐŸĐČĐ”ŃŃ
ĐœĐŸŃŃĐž ĐżĐ»Đ”ĐœĐŸĐș Si. ĐĐ»Ń ĐœĐ”ĐșĐŸŃĐŸŃŃŃ
ŃĐ”Đ¶ĐžĐŒĐŸĐČ ĐŸŃĐ°Đ¶ĐŽĐ”ĐœĐžŃ ĐŸĐ±ĐœĐ°ŃŃĐ¶Đ”ĐœĐŸ ĐżĐŸŃĐČĐ»Đ”ĐœĐžĐ” ĐœĐ° ĐżĐŸĐČĐ”ŃŃ
ĐœĐŸŃŃĐž ĐœĐžŃĐ”ĐČĐžĐŽĐœŃŃ
ŃŃŃŃĐșŃŃŃ Đž/ОлО ĐșŃŃглŃŃ
ŃглŃĐ±Đ»Đ”ĐœĐžĐč, ĐžĐ·ĐŒĐ”ĐœĐ”ĐœĐžŃ ĐżĐŸĐ»ĐŸĐ¶Đ”ĐœĐžŃ ĐŒĐžĐœĐžĐŒŃĐŒĐŸĐČ Đž ĐŒĐ°ĐșŃĐžĐŒŃĐŒĐŸĐČ ĐČ ĐŸĐżŃĐžŃĐ”ŃĐșĐžŃ
ŃпДĐșŃŃĐ°Ń
ĐŸŃŃĐ°Đ¶Đ”ĐœĐžŃ Đž ĐżŃĐŸĐżŃŃĐșĐ°ĐœĐžŃ.
A Roadmap for HEP Software and Computing R&D for the 2020s
Particle physics has an ambitious and broad experimental programme for the coming decades. This programme requires large investments in detector hardware, either to build new facilities and experiments, or to upgrade existing ones. Similarly, it requires commensurate investment in the R&D of software to acquire, manage, process, and analyse the shear amounts of data to be recorded. In planning for the HL-LHC in particular, it is critical that all of the collaborating stakeholders agree on the software goals and priorities, and that the efforts complement each other. In this spirit, this white paper describes the R&D activities required to prepare for this software upgrade.Peer reviewe
Les droits disciplinaires des fonctions publiques : « unification », « harmonisation » ou « distanciation ». A propos de la loi du 26 avril 2016 relative à la déontologie et aux droits et obligations des fonctionnaires
The production of tt⟠, W+bb⟠and W+cc⟠is studied in the forward region of protonâproton collisions collected at a centre-of-mass energy of 8 TeV by the LHCb experiment, corresponding to an integrated luminosity of 1.98±0.02 fbâ1 . The W bosons are reconstructed in the decays WââÎœ , where â denotes muon or electron, while the b and c quarks are reconstructed as jets. All measured cross-sections are in agreement with next-to-leading-order Standard Model predictions.The production of , and is studied in the forward region of proton-proton collisions collected at a centre-of-mass energy of 8 TeV by the LHCb experiment, corresponding to an integrated luminosity of 1.98 0.02 \mbox{fb}^{-1}. The bosons are reconstructed in the decays , where denotes muon or electron, while the and quarks are reconstructed as jets. All measured cross-sections are in agreement with next-to-leading-order Standard Model predictions
- âŠ