405 research outputs found
Analysis of the Web Graph Aggregated by Host and Pay-Level Domain
In this paper the web is analyzed as a graph aggregated by host and pay-level
domain (PLD). The web graph datasets, publicly available, have been released by
the Common Crawl Foundation and are based on a web crawl performed during the
period May-June-July 2017. The host graph has 1.3 billion nodes and
5.3 billion arcs. The PLD graph has 91 million nodes and 1.1
billion arcs. We study the distributions of degree and sizes of strongly/weakly
connected components (SCC/WCC) focusing on power laws detection using
statistical methods. The statistical plausibility of the power law model is
compared with that of several alternative distributions. While there is no
evidence of power law tails on host level, they emerge on PLD aggregation for
indegree, SCC and WCC size distributions. Finally, we analyze distance-related
features by studying the cumulative distributions of the shortest path lengths,
and give an estimation of the diameters of the graphs
Micro-CernVM: Slashing the Cost of Building and Deploying Virtual Machines
The traditional virtual machine building and and deployment process is
centered around the virtual machine hard disk image. The packages comprising
the VM operating system are carefully selected, hard disk images are built for
a variety of different hypervisors, and images have to be distributed and
decompressed in order to instantiate a virtual machine. Within the HEP
community, the CernVM File System has been established in order to decouple the
distribution from the experiment software from the building and distribution of
the VM hard disk images.
We show how to get rid of such pre-built hard disk images altogether. Due to
the high requirements on POSIX compliance imposed by HEP application software,
CernVM-FS can also be used to host and boot a Linux operating system. This
allows the use of a tiny bootable CD image that comprises only a Linux kernel
while the rest of the operating system is provided on demand by CernVM-FS. This
approach speeds up the initial instantiation time and reduces virtual machine
image sizes by an order of magnitude. Furthermore, security updates can be
distributed instantaneously through CernVM-FS. By leveraging the fact that
CernVM-FS is a versioning file system, a historic analysis environment can be
easily re-spawned by selecting the corresponding CernVM-FS file system
snapshot.Comment: Conference paper at the 2013 Computing in High Energy Physics (CHEP)
Conference, Amsterda
CernVM Online and Cloud Gateway: a uniform interface for CernVM contextualization and deployment
In a virtualized environment, contextualization is the process of configuring
a VM instance for the needs of various deployment use cases. Contextualization
in CernVM can be done by passing a handwritten context to the user data field
of cloud APIs, when running CernVM on the cloud, or by using CernVM web
interface when running the VM locally. CernVM Online is a publicly accessible
web interface that unifies these two procedures. A user is able to define,
store and share CernVM contexts using CernVM Online and then apply them either
in a cloud by using CernVM Cloud Gateway or on a local VM with the single-step
pairing mechanism. CernVM Cloud Gateway is a distributed system that provides a
single interface to use multiple and different clouds (by location or type,
private or public). Cloud gateway has been so far integrated with OpenNebula,
CloudStack and EC2 tools interfaces. A user, with access to a number of clouds,
can run CernVM cloud agents that will communicate with these clouds using their
interfaces, and then use one single interface to deploy and scale CernVM
clusters. CernVM clusters are defined in CernVM Online and consist of a set of
CernVM instances that are contextualized and can communicate with each other.Comment: Conference paper at the 2013 Computing in High Energy Physics (CHEP)
Conference, Amsterda
Opportunities for Nuclear Astrophysics at FRANZ
The "Frankfurter Neutronenquelle am Stern-Gerlach-Zentrum" (FRANZ), which is
currently under development, will be the strongest neutron source in the
astrophysically interesting energy region in the world. It will be about three
orders of magnitude more intense than the well-established neutron source at
the Research Center Karlsruhe (FZK)
Verification and Validation of Semantic Annotations
In this paper, we propose a framework to perform verification and validation
of semantically annotated data. The annotations, extracted from websites, are
verified against the schema.org vocabulary and Domain Specifications to ensure
the syntactic correctness and completeness of the annotations. The Domain
Specifications allow checking the compliance of annotations against
corresponding domain-specific constraints. The validation mechanism will detect
errors and inconsistencies between the content of the analyzed schema.org
annotations and the content of the web pages where the annotations were found.Comment: Accepted for the A.P. Ershov Informatics Conference 2019(the PSI
Conference Series, 12th edition) proceedin
Deployment of RDFa, Microdata, and Microformats on the Web – A Quantitative Analysis
More and more websites embed structured data describing for instance
products, reviews, blog posts, people, organizations, events, and cooking recipes
into their HTML pages using markup standards such as Microformats, Microdata
and RDFa. This development has accelerated in the last two years as major Web
companies, such as Google, Facebook, Yahoo!, and Microsoft, have started to
use the embedded data within their applications. In this paper, we analyze the
adoption of RDFa, Microdata, and Microformats across the Web. Our study is
based on a large public Web crawl dating from early 2012 and consisting of 3
billion HTML pages which originate from over 40 million websites. The analysis
reveals the deployment of the different markup standards, the main topical areas
of the published data as well as the different vocabularies that are used within each
topical area to represent data. What distinguishes our work from earlier studies,
published by the large Web companies, is that the analyzed crawl as well as the
extracted data are publicly available. This allows our findings to be verified and to
be used as starting points for further domain-specific investigations as well as for
focused information extraction endeavors
Cardiotrophin-1 Induces Tumor Necrosis Factor α Synthesis in Human Peripheral Blood Mononuclear Cells
Chronic heart failure (CHF) is associated with elevated concentrations of tumor necrosis factor (TNF) α and cardiotrophin-1 (CT-1) and altered peripheral blood mononuclear cell (PBMC) function. Therefore, we tested whether CT-1 induces TNFα in PBMC of healthy volunteers.
CT-1 induced in PBMC TNFα protein in the supernatant and TNFα mRNA in a concentration- and time-dependent manner determined by ELISA and real-time PCR, respectively. Maximal TNFα protein was achieved with 100 ng/mL CT-1 after 3–6 hours and maximal TNFα mRNA induction after 1 hour. ELISA data were confirmed using immunofluorescent flow cytometry. Inhibitor studies with actinomycin D and brefeldin A showed that both protein synthesis and intracellular transport are essential for CT-1 induced TNFα expression. CT-1 caused a dose dependent nuclear factor (NF) κB translocation. Parthenolide inhibited both NFκB translocation and TNFα protein expression indicating that NFκB seems to be necessary.
We revealed a new mechanism for elevated serum TNFα concentrations and PBMC activation in CHF besides the hypothesis of PBMC activation by bacterial translocation from the gut
- …