7,690 research outputs found
The Metaverse: Survey, Trends, Novel Pipeline Ecosystem & Future Directions
The Metaverse offers a second world beyond reality, where boundaries are
non-existent, and possibilities are endless through engagement and immersive
experiences using the virtual reality (VR) technology. Many disciplines can
benefit from the advancement of the Metaverse when accurately developed,
including the fields of technology, gaming, education, art, and culture.
Nevertheless, developing the Metaverse environment to its full potential is an
ambiguous task that needs proper guidance and directions. Existing surveys on
the Metaverse focus only on a specific aspect and discipline of the Metaverse
and lack a holistic view of the entire process. To this end, a more holistic,
multi-disciplinary, in-depth, and academic and industry-oriented review is
required to provide a thorough study of the Metaverse development pipeline. To
address these issues, we present in this survey a novel multi-layered pipeline
ecosystem composed of (1) the Metaverse computing, networking, communications
and hardware infrastructure, (2) environment digitization, and (3) user
interactions. For every layer, we discuss the components that detail the steps
of its development. Also, for each of these components, we examine the impact
of a set of enabling technologies and empowering domains (e.g., Artificial
Intelligence, Security & Privacy, Blockchain, Business, Ethics, and Social) on
its advancement. In addition, we explain the importance of these technologies
to support decentralization, interoperability, user experiences, interactions,
and monetization. Our presented study highlights the existing challenges for
each component, followed by research directions and potential solutions. To the
best of our knowledge, this survey is the most comprehensive and allows users,
scholars, and entrepreneurs to get an in-depth understanding of the Metaverse
ecosystem to find their opportunities and potentials for contribution
The temporality of rhetoric: the spatialization of time in modern criticism
Every conception of criticism conceals a notion of time which informs the manner in which the critic conceives of history, representation and criticism itself. This thesis reveals the philosophies of time inherent in certain key modern critical concepts: allegory, irony and the sublime. Each concept opens a breach in time, a disruption of chronology. In each case this gap or aporia is emphatically closed, elided or denied. Taking the philosophy of time elaborated by Giorgio Agamben as an introductory proposition, my argument turns in Chapter One to the allegorical temporality which Walter Benjamin sees as the time of photography. The second chapter examines the aesthetics of the sublime as melancholic or mournful untimeliness. In Chapter Three, Paul de Man's conception of irony provides an exemplary instance of the denial of this troubling temporal predicament. In opposition to the foreclosure of the disturbing temporalities of criticism, history and representation, the thesis proposes a fundamental rethinking of the philosophy of time as it relates to these categories of reflection. In a reading of an inaugural meditation on the nature of time, and in examining certain key contemporary philosophical and critical texts, I argue for a critical attendance to that which eludes those modes of thought that attempt to map time as a recognizable and essentially spatial field. The Confessions of Augustine provide, in the fourth chapter, a model for thinking through the problems set up earlier: Augustine affords us, precisely, a means of conceiving of the gap or the interim. In the final chapter, this concept is developed with reference to the criticism of Arnold and Eliot, the fiction of Virginia Woolf and the philosophy of cinema derived from Deleuze and Lyotard. In conclusion, the philosophical implications of the thesis are placed in relation to a conception of the untimeliness of death
Recommended from our members
Co-design As Healing: Exploring The Experiences Of Participants Facing Mental Health Problems
This thesis is an exploration of the healing role of co-design in mental health. Although co-design projects conducted within mental health settings are rising, existing literature tends to focus on the object of design and its outcomes while the experiences of participants per se remain largely unexplored. The guiding research question of this study is not how we design things that improve mental health, but how co-designing, as an act, might do so.
The thesis presents two projects that were organized in collaboration with the mental health charity Islington Mind and the Psychosis Therapy Project (PTP) in London.
The project at Islington Mind used a structured design process inviting participants to design for wellbeing. A case study analysis provides insights on how participants were impacted, summarizing key challenges and opportunities.
The design at PTP worked towards creating a collective brief in an emergent fashion, finally culminating in a board game. The experiences of participants were explored through Interpretative Phenomenological Analysis (IPA), using semi-structured interview data. The analysis served to identify key themes characterising the experience of co-design such as contributing, connecting, thinking and intentioning. In addition, a mixed-methods analysis of questionnaires and interview data exploring participants' wellbeing, showed that all participants who engaged fairly consistently in the project improved after the project ended, although some participants' scores returned to baseline six months later.
Reflecting on both projects, an approach to facilitation within mental health is outlined, detailing how the dimensions of weaving and layered participation, nurturing mattering and facilitating attitudes interlace. This contribution raises awareness of tacit dimensions in the practice of facilitation, articulating the nuances of how to encourage and sustain meaningful and ethical engagement and offering insights into a range of tools. It highlights the importance of remaining reflexive in relation to attitudes and emotions and discusses practical methodological and ethical challenges and ways to resolve them which can be of benefit to researchers embarking on a similar journey.
The thesis also offers detailed insights on how methodologies from different fields were integrated into a whole, arguing for transparency and reflexivity about epistemological assumptions, and how underlying paradigms shift in an interdisciplinary context.
Based on the overall findings, the thesis makes a case for considering design as healing (or a designerly way of healing), highlighting implications at a systems, social and individual level. It makes an original contribution to our understanding of design, highlighting its healing character, and proposes a new way to support mental health. The participants in this study not only had increased their own wellbeing through co-designing, but were also empowered and contributed towards healing the world. Hence, the thesis argues for a unique, holistic perspective of design and mental health, recognizing the interconnectedness of the individual, social and systemic dimensions of the healing processes that are ignited
Image classification over unknown and anomalous domains
A longstanding goal in computer vision research is to develop methods that are simultaneously applicable to a broad range of prediction problems. In contrast to this, models often perform best when they are specialized to some task or data type. This thesis investigates the challenges of learning models that generalize well over multiple unknown or anomalous modes and domains in data, and presents new solutions for learning robustly in this setting.
Initial investigations focus on normalization for distributions that contain multiple sources (e.g. images in different styles like cartoons or photos). Experiments demonstrate the extent to which existing modules, batch normalization in particular, struggle with such heterogeneous data, and a new solution is proposed that can better handle data from multiple visual modes, using differing sample statistics for each.
While ideas to counter the overspecialization of models have been formulated in sub-disciplines of transfer learning, e.g. multi-domain and multi-task learning, these usually rely on the existence of meta information, such as task or domain labels. Relaxing this assumption gives rise to a new transfer learning setting, called latent domain learning in this thesis, in which training and inference are carried out over data from multiple visual domains, without domain-level annotations. Customized solutions are required for this, as the performance of standard models degrades: a new data augmentation technique that interpolates between latent domains in an unsupervised way is presented, alongside a dedicated module that sparsely accounts for hidden domains in data, without requiring domain labels to do so.
In addition, the thesis studies the problem of classifying previously unseen or anomalous modes in data, a fundamental problem in one-class learning, and anomaly detection in particular. While recent ideas have been focused on developing self-supervised solutions for the one-class setting, in this thesis new methods based on transfer learning are formulated. Extensive experimental evidence demonstrates that a transfer-based perspective benefits new problems that have recently been proposed in anomaly detection literature, in particular challenging semantic detection tasks
Annals [...].
Pedometrics: innovation in tropics; Legacy data: how turn it useful?; Advances in soil sensing; Pedometric guidelines to systematic soil surveys.Evento online. Coordenado por: Waldir de Carvalho Junior, Helena Saraiva Koenow Pinheiro, Ricardo Simão Diniz Dalmolin
Compatibility and challenges in machine learning approach for structural crack assessment
Structural health monitoring and assessment (SHMA) is exceptionally essential for preserving and sustaining any mechanical structure’s service life. A successful assessment should provide reliable and resolute information to maintain the continuous performance of the structure. This information can effectively determine crack progression and its overall impact on the structural operation. However, the available sensing techniques and methods for performing SHMA generate raw measurements that require significant data processing before making any valuable predictions. Machine learning (ML) algorithms (supervised and unsupervised learning) have been extensively used for such data processing. These algorithms extract damage-sensitive features from the raw data to identify structural conditions and performance. As per the available published literature, the extraction of these features has been quite random and used by academic researchers without a suitability justification. In this paper, a comprehensive literature review is performed to emphasise the influence of damage-sensitive features on ML algorithms. The selection and suitability of these features are critically reviewed while processing raw data obtained from different materials (metals, composites and polymers). It has been found that an accurate crack prediction is only possible if the selection of damage-sensitive features and ML algorithms is performed based on available raw data and structure material type. This paper also highlights the current challenges and limitations during the mentioned sections
BECOMEBECOME - A TRANSDISCIPLINARY METHODOLOGY BASED ON INFORMATION ABOUT THE OBSERVER
ABSTRACT
Andrea T. R. Traldi
BECOMEBECOME
A Transdisciplinary Methodology Based on Information about the Observer
The present research dissertation has been developed with the intention to provide practical strategies and discover new intellectual operations which can be used to generate Transdisciplinary insight. For this reason, this thesis creates access to new knowledge at different scales.
Firstly, as it pertains to the scale of new knowledge generated by those who attend Becomebecome events. The open-source nature of the Becomebecome methodology makes it possible for participants in Becomebecome workshops, training programmes and residencies to generate new insight about the specific project they are working on, which then reinforce and expand the foundational principles of the theoretical background.
Secondly, as it pertains to the scale of the Becomebecome framework, which remains independent of location and moment in time. The method proposed to access Transdisciplinary knowledge constitutes new knowledge in itself because the sequence of activities, described as physical and mental procedures and listed as essential criteria, have never been found organised
6
in such a specific order before. It is indeed the order in time, i.e. the sequence of the ideas and activities proposed, which allows one to transform Disciplinary knowledge via a new Transdisciplinary frame of reference.
Lastly, new knowledge about Transdisciplinarity as a field of study is created as a consequence of the heretofore listed two processes.
The first part of the thesis is designated ‘Becomebecome Theory’ and focuses on the theoretical background and the intellectual operations necessary to support the creation of new Transdisciplinary knowledge. The second part of the thesis is designated ‘Becomebecome Practice’ and provides practical examples of the application of such operations. Crucially, the theoretical model described as the foundation for the Becomebecome methodology (Becomebecome Theory) is process-based and constantly checked against the insight generated through Becomebecome Practice.
To this effect, ‘information about the observer’ is proposed as a key notion which binds together Transdisciplinary resources from several studies in the hard sciences and humanities. It is a concept that enables understanding about why and how information that is generated through Becomebecome Practice is considered of paramount importance for establishing the reference parameters necessary to access Transdisciplinary insight which is meaningful to a specific project, a specific person, or a specific moment in time
MoFaNeRF: Morphable Facial Neural Radiance Field
We propose a parametric model that maps free-view images into a vector space
of coded facial shape, expression and appearance with a neural radiance field,
namely Morphable Facial NeRF. Specifically, MoFaNeRF takes the coded facial
shape, expression and appearance along with space coordinate and view direction
as input to an MLP, and outputs the radiance of the space point for
photo-realistic image synthesis. Compared with conventional 3D morphable models
(3DMM), MoFaNeRF shows superiority in directly synthesizing photo-realistic
facial details even for eyes, mouths, and beards. Also, continuous face
morphing can be easily achieved by interpolating the input shape, expression
and appearance codes. By introducing identity-specific modulation and texture
encoder, our model synthesizes accurate photometric details and shows strong
representation ability. Our model shows strong ability on multiple applications
including image-based fitting, random generation, face rigging, face editing,
and novel view synthesis. Experiments show that our method achieves higher
representation ability than previous parametric models, and achieves
competitive performance in several applications. To the best of our knowledge,
our work is the first facial parametric model built upon a neural radiance
field that can be used in fitting, generation and manipulation. The code and
data is available at https://github.com/zhuhao-nju/mofanerf.Comment: accepted to ECCV2022; code available at
http://github.com/zhuhao-nju/mofaner
Fast Neural Representations for Direct Volume Rendering
Despite the potential of neural scene representations to effectively compress
3D scalar fields at high reconstruction quality, the computational complexity
of the training and data reconstruction step using scene representation
networks limits their use in practical applications. In this paper, we analyze
whether scene representation networks can be modified to reduce these
limitations and whether such architectures can also be used for temporal
reconstruction tasks. We propose a novel design of scene representation
networks using GPU tensor cores to integrate the reconstruction seamlessly into
on-chip raytracing kernels, and compare the quality and performance of this
network to alternative network- and non-network-based compression schemes. The
results indicate competitive quality of our design at high compression rates,
and significantly faster decoding times and lower memory consumption during
data reconstruction. We investigate how density gradients can be computed using
the network and show an extension where density, gradient and curvature are
predicted jointly. As an alternative to spatial super-resolution approaches for
time-varying fields, we propose a solution that builds upon latent-space
interpolation to enable random access reconstruction at arbitrary granularity.
We summarize our findings in the form of an assessment of the strengths and
limitations of scene representation networks \changed{for compression domain
volume rendering, and outline future research directions
- …