1,069 research outputs found
Recommended from our members
Reliable record-keeping: is digital archiving the answer?
pdf of printed article, pdf of online article and pdf of rough translation. URL of online version, accessed week of 20/01/2012: http://www.asahi.com/culture/news_culture/TKY201201050145.htmlMark Turin has agreed with the copyright holder the right to host a copy of this piece (whether audio, text or video) on University of Cambridge archives and servers.Article featuring the World Oral Literature, published in Asahi Shimbun newspaper in Japan, Jan 2012. Rough translation to English: Reliable record-keeping: is digital archiving the answer?
In Weimar, central Germany (population 60,000), a fire broke out one night in 2004. It started in the Anna Maria Library, a UNESCO World Heritage Library founded in the 18th Century. Among the 50,000 books burnt were rare volumes and manuscripts by Copernicus. 28,000 books were partially burnt and are under restoration.
The weather is very cold at Cambridge University, where a female member of staff is quietly working at a computer in the Museum of Archaeology and Anthropology. A conservation project for endangered oral culture was founded three years ago. Fieldworkers around the globe are sending materials (audio and film recordings and electronic data) to the project, where the staff archive them to a geographical map on the website.
Accessing anthropological cultural data used to mean a visit to the host country, because archives could not be placed on the internet, so were put in a drawer and forgotten about. “Losing traditional culture means losing one’s identity. Digital recording helps to save historical records and makes our jobs easier’, says Professor Alan Macfarlane, University of Cambridge. “We have been able to record materials twice as fast using digital methods. Ten years since the beginning of digital recording, we are targeting over 100 cultures to be preserved in this way. The communities themselves are able to see the recordings and pass these historical records to the next generation.”
As demonstrated by the fire in the Anna Maria Library, the preservation of books and film in physical form is unreliable. UNESCO reported that in the 20th Century, over 100 libraries were damaged or destroyed by fire or war. This situation has recently been improved by digital recording.
Immediately after the Kobe earthquake, Mori (the property magnate) sent his urban research group to Kobe to film and survey the land using GPS, creating a precise record of the damage just nine hours after the quake. The huge amount of data passed on to the emergency services included statistics on damage to buildings, refugee centres, and food and dust mask stores. The data remains useful long after its collection. The paradox of museums is that, when you are collecting data, you don’t know what will be most important. Professor Hirose, Tokyo University, says “The way that digital recording can keep everything dispels the paradox of the museum. Material deemed unimportant and not recorded later becomes important, and through digital recording everything can be kept”.
Professor Takano of the National Data Research Group created the search engine SOU in 2006 that uses a single word to find information, using searches in the same way as a human brain: one word cascades to multiple links and connections in the search, so that individually created records connect to each other. National cultural monuments or old books and art gallery collections can be viewed on a single database. For example, using the search term ‘vase’ results in national treasures, books about vases and museum prices displayed on the site. Previously, records were collected and then abandoned. The connections between these records can be surprising. This heightens the interest for the records.”
On the other hand, digital recording can also be unreliable. “People are once again considering film as a system for making long-term records. 63,000 cinema films are now stored at the Tokyo National Modern Art Museum film centre”, says the Chief Research, Mr Tochigi. The centre’s warehouse keeps these films under specially controlled atmospheric conditions. The style of digital data filing, recording formats and playback system is changing all the time so one cannot see the previous data, and it is possible to accidently delete it. Digital data should be used as a back-up of the original data and format, rather than a replacement”.
At the Anna Maria Library, staff are now scanning and recording their data and transforming it to a digital format, but only since the fire. They are also working to restore original material that has been burnt and to re-collect books that were lost completely. The original book binding and the layout of the book itself (such as photos) are an important part of the author and publisher’s original concept. We conclude that we still cannot tell the best way to keep records
Single-epoch supernova classification with deep convolutional neural networks
Supernovae Type-Ia (SNeIa) play a significant role in exploring the history
of the expansion of the Universe, since they are the best-known standard
candles with which we can accurately measure the distance to the objects.
Finding large samples of SNeIa and investigating their detailed characteristics
have become an important issue in cosmology and astronomy. Existing methods
relied on a photometric approach that first measures the luminance of supernova
candidates precisely and then fits the results to a parametric function of
temporal changes in luminance. However, it inevitably requires multi-epoch
observations and complex luminance measurements. In this work, we present a
novel method for classifying SNeIa simply from single-epoch observation images
without any complex measurements, by effectively integrating the
state-of-the-art computer vision methodology into the standard photometric
approach. Our method first builds a convolutional neural network for estimating
the luminance of supernovae from telescope images, and then constructs another
neural network for the classification, where the estimated luminance and
observation dates are used as features for classification. Both of the neural
networks are integrated into a single deep neural network to classify SNeIa
directly from observation images. Experimental results show the effectiveness
of the proposed method and reveal classification performance comparable to
existing photometric methods with multi-epoch observations.Comment: 7 pages, published as a workshop paper in ICDCS2017, in June 201
Compatibility between Jacobi structures and pseudo-Riemannian cometrics on Jacobi algebroids
We define compatibility between Jacobi structures and pseudo-Riemannian
cometrics on Jacobi algebroids. This notion is a generalization of the
compatibility between Poisson structures and pseudo-Riemannian cometrics on
manifolds, which was defined by Boucetta. We show that the compatibility with a
cometric is ``preserved'' by the Poissonization of a Jacobi structure.
Furthermore, we prove that for a contact pseudo-metric structure on a manifold,
satisfying the compatibility condition is equivalent to being a Sasakian
pseudo-metric structure.Comment: 17 pages. arXiv admin note: text overlap with arXiv:2112.0349
Search for nonresonant pair production of Higgs bosons in the b ¯ b b ¯ b final state in p p collisions at √ s = 13 TeV with the ATLAS detector
A search for nonresonant Higgs boson pair production in the
b
¯
b
b
¯
b
final state is presented. The analysis uses
126
fb
−
1
of
p
p
collision data at
√
s
=
13
TeV
collected with the ATLAS detector at the Large Hadron Collider, and targets both the gluon-gluon fusion and vector-boson fusion production modes. No evidence of the signal is found and the observed (expected) upper limit on the cross section for nonresonant Higgs boson pair production is determined to be 5.4 (8.1) times the Standard Model predicted cross section at 95% confidence level. Constraints are placed on modifiers to the
H
H
H
and
H
H
V
V
couplings. The observed (expected)
2
σ
constraints on the
H
H
H
coupling modifier,
κ
λ
, are determined to be
[
−
3.5
,
11.3
]
(
[
−
5.4
,
11.4
]
), while the corresponding constraints for the
H
H
V
V
coupling modifier,
κ
2
V
, are
[
−
0.0
,
2.1
]
(
[
−
0.1
,
2.1
]
). In addition, constraints on relevant coefficients are derived in the context of the Standard Model effective field theory and Higgs effective field theory, and upper limits on the
H
H
production cross section are placed in seven Higgs effective field theory benchmark scenarios
- …