4,672 research outputs found
Automatic Music Composition using Answer Set Programming
Music composition used to be a pen and paper activity. These these days music
is often composed with the aid of computer software, even to the point where
the computer compose parts of the score autonomously. The composition of most
styles of music is governed by rules. We show that by approaching the
automation, analysis and verification of composition as a knowledge
representation task and formalising these rules in a suitable logical language,
powerful and expressive intelligent composition tools can be easily built. This
application paper describes the use of answer set programming to construct an
automated system, named ANTON, that can compose melodic, harmonic and rhythmic
music, diagnose errors in human compositions and serve as a computer-aided
composition tool. The combination of harmonic, rhythmic and melodic composition
in a single framework makes ANTON unique in the growing area of algorithmic
composition. With near real-time composition, ANTON reaches the point where it
can not only be used as a component in an interactive composition tool but also
has the potential for live performances and concerts or automatically generated
background music in a variety of applications. With the use of a fully
declarative language and an "off-the-shelf" reasoning engine, ANTON provides
the human composer a tool which is significantly simpler, more compact and more
versatile than other existing systems. This paper has been accepted for
publication in Theory and Practice of Logic Programming (TPLP).Comment: 31 pages, 10 figures. Extended version of our ICLP2008 paper.
Formatted following TPLP guideline
'Playing robot': an interactive sound installation in human-robot interaction design for new media art
In this study artistic human-robot interaction design is in- troduced as a means for scientific research and artistic inves- tigations. It serves as a methodology for situated cognition integrating empirical methodology and computational mod- eling, and is exemplified by the installation playing robot. Its artistic purpose is to aid to create and explore robots as a new medium for art and entertainment. We discuss the use of finite state machines to organize robots’ behavioral reac- tions to sensor data, and give a brief outlook on structured observation as a potential method for data collection
Evolution of central pattern generators for the control of a five-link bipedal walking mechanism
Central pattern generators (CPGs), with a basis is neurophysiological
studies, are a type of neural network for the generation of rhythmic motion.
While CPGs are being increasingly used in robot control, most applications are
hand-tuned for a specific task and it is acknowledged in the field that generic
methods and design principles for creating individual networks for a given task
are lacking. This study presents an approach where the connectivity and
oscillatory parameters of a CPG network are determined by an evolutionary
algorithm with fitness evaluations in a realistic simulation with accurate
physics. We apply this technique to a five-link planar walking mechanism to
demonstrate its feasibility and performance. In addition, to see whether
results from simulation can be acceptably transferred to real robot hardware,
the best evolved CPG network is also tested on a real mechanism. Our results
also confirm that the biologically inspired CPG model is well suited for legged
locomotion, since a diverse manifestation of networks have been observed to
succeed in fitness simulations during evolution.Comment: 11 pages, 9 figures; substantial revision of content, organization,
and quantitative result
Recommended from our members
Improving music genre classification using automatically induced harmony rules
We present a new genre classification framework using both low-level signal-based features and high-level harmony features. A state-of-the-art statistical genre classifier based on timbral features is extended using a first-order random forest containing for each genre rules derived from harmony or chord sequences. This random forest has been automatically induced, using the first-order logic induction algorithm TILDE, from a dataset, in which for each chord the degree and chord category are identified, and covering classical, jazz and pop genre classes. The audio descriptor-based genre classifier contains 206 features, covering spectral, temporal, energy, and pitch characteristics of the audio signal. The fusion of the harmony-based classifier with the extracted feature vectors is tested on three-genre subsets of the GTZAN and ISMIR04 datasets, which contain 300 and 448 recordings, respectively. Machine learning classifiers were tested using 5 × 5-fold cross-validation and feature selection. Results indicate that the proposed harmony-based rules combined with the timbral descriptor-based genre classification system lead to improved genre classification rates
Recommended from our members
Improving music genre classification using automatically induced harmony rules
We present a new genre classification framework using both low-level signal-based features and high-level harmony features. A state-of-the-art statistical genre classifier based on timbral features is extended using a first-order random forest containing for each genre rules derived from harmony or chord sequences. This random forest has been automatically induced, using the first-order logic induction algorithm TILDE, from a dataset, in which for each chord the degree and chord category are identified, and covering classical, jazz and pop genre classes. The audio descriptor-based genre classifier contains 206 features, covering spectral, temporal, energy, and pitch characteristics of the audio signal. The fusion of the harmony-based classifier with the extracted feature vectors is tested on three-genre subsets of the GTZAN and ISMIR04 datasets, which contain 300 and 448 recordings, respectively. Machine learning classifiers were tested using 5 × 5-fold cross-validation and feature selection. Results indicate that the proposed harmony-based rules combined with the timbral descriptor-based genre classification system lead to improved genre classification rates
A Functional Taxonomy of Music Generation Systems
Digital advances have transformed the face of automatic music generation
since its beginnings at the dawn of computing. Despite the many breakthroughs,
issues such as the musical tasks targeted by different machines and the degree
to which they succeed remain open questions. We present a functional taxonomy
for music generation systems with reference to existing systems. The taxonomy
organizes systems according to the purposes for which they were designed. It
also reveals the inter-relatedness amongst the systems. This design-centered
approach contrasts with predominant methods-based surveys and facilitates the
identification of grand challenges to set the stage for new breakthroughs.Comment: survey, music generation, taxonomy, functional survey, survey,
automatic composition, algorithmic compositio
Multimodal music information processing and retrieval: survey and future challenges
Towards improving the performance in various music information processing
tasks, recent studies exploit different modalities able to capture diverse
aspects of music. Such modalities include audio recordings, symbolic music
scores, mid-level representations, motion, and gestural data, video recordings,
editorial or cultural tags, lyrics and album cover arts. This paper critically
reviews the various approaches adopted in Music Information Processing and
Retrieval and highlights how multimodal algorithms can help Music Computing
applications. First, we categorize the related literature based on the
application they address. Subsequently, we analyze existing information fusion
approaches, and we conclude with the set of challenges that Music Information
Retrieval and Sound and Music Computing research communities should focus in
the next years
AN APPROACH TO MACHINE DEVELOPMENT OF MUSICAL ONTOGENY
This Thesis pursues three main objectives: (i) to use computational modelling to
explore how music is perceived, cognitively processed and created by human
beings; (ii) to explore interactive musical systems as a method to model and
achieve the transmission of musical influence in artificial worlds and between
humans and machines; and (iii) to experiment with artificial and alternative
developmental musical routes in order to observe the evolution of musical
styles.
In order to achieve these objectives, this Thesis introduces a new paradigm for
the design of computer interactive musical systems called the Ontomemetical
Model of Music Evolution - OMME, which includes the fields of musical
ontogenesis and memetlcs. OMME-based systems are designed to artificially
explore the evolution of music centred on human perceptive and cognitive
faculties.
The potential of the OMME is illustrated with two interactive musical systems,
the Rhythmic Meme Generator (RGeme) and the Interactive Musical
Environments (iMe). which have been tested in a series of laboratory
experiments and live performances. The introduction to the OMME is preceded
by an extensive and critical overview of the state of the art computer models
that explore musical creativity and interactivity, in addition to a systematic
exposition of the major issues involved in the design and implementation of
these systems.
This Thesis also proposes innovative solutions for (i) the representation of
musical streams based on perceptive features, (ii) music segmentation, (iii) a
memory-based music model, (iv) the measure of distance between musical
styles, and (v) an impi*ovisation-based creative model
- …