68 research outputs found

    IDATER online conference: graphicacy and modelling 2010

    Get PDF
    IDATER online conference: graphicacy and modelling 201

    Deep Learning for Free-Hand Sketch: A Survey

    Get PDF
    Free-hand sketches are highly illustrative, and have been widely used by humans to depict objects or stories from ancient times to the present. The recent prevalence of touchscreen devices has made sketch creation a much easier task than ever and consequently made sketch-oriented applications increasingly popular. The progress of deep learning has immensely benefited free-hand sketch research and applications. This paper presents a comprehensive survey of the deep learning techniques oriented at free-hand sketch data, and the applications that they enable. The main contents of this survey include: (i) A discussion of the intrinsic traits and unique challenges of free-hand sketch, to highlight the essential differences between sketch data and other data modalities, e.g., natural photos. (ii) A review of the developments of free-hand sketch research in the deep learning era, by surveying existing datasets, research topics, and the state-of-the-art methods through a detailed taxonomy and experimental evaluation. (iii) Promotion of future work via a discussion of bottlenecks, open problems, and potential research directions for the community.Comment: This paper is accepted by IEEE TPAM

    Teaching GANs to sketch in vector format

    Get PDF
    Sketching is a fundamental human cognitive ability. Deep Neural Networks (DNNs) have achieved the state-of-the-art performance in recognition tasks like image recognition, speech recognition etc. but have not made significant progress in generating stroke-based sketches a.k.a sketches in vector format. Though there are Variational Auto Encoders (VAEs) for generating sketches in vector format, there is no Generative Adversarial Network (GAN) architecture for the same. In this paper, we propose a standalone GAN architecture called SkeGAN and a hybrid VAE-GAN architecture called VASkeGAN, for sketch generation in vector format. SkeGAN is a stochastic policy in Reinforcement Learning (RL), capable of generating both multidimensional continuous and discrete outputs. VASkeGAN draws sketches by coupling the efficient representation of data by VAE with the powerful generating capabilities of GAN. We have validated that SkeGAN and VASkeGAN generate visually appealing sketches with minimal scribble effect and is comparable to a recent work titled Sketch-RNN. © 2021 ACM

    The communication and recording of conceptual design information by the inclusion of visual data

    Get PDF
    This thesis reports the results of a three year, full-time research project investigating the generation and communication of product descriptions within the conceptual phase of the engineering design process. The research pays particular attention to the role played by the designer's sketch in communicating new product ideas. The investigation commences with a literature review of existing design process models (Chapter 2), which helps to define the area under investigation while presenting modern views of the process in relation to classic examples from established design research. Chapter 3 presents a literature review of the methods currently used to support communication of product descriptions. These methods of Specification are assessed and particular attention is given to new computer-based recording methods such as DOORS and Cradle. Suggestions for improving the efficiency of such models are put forward and the text-only bias of such systems is identified. This comparison of the existing systems thus identifies the research questions. Having identified the possible improvement to be gained by the incorporation of visual material in addition to the universal text description, Chapter 4 presents a literature review assessing the roles of the conceptual sketch in engineering design. As well as presenting views of drawing from philosophical, psychological and scientific standpoints, this section compares attempts made to support the engineer's sketching activity by computer means. This chapter concludes that efforts made to provide effective computer support of sketching by freehand methods are preferred to attempts made to replicate the process with current computer tools. The resulting research experiment, the methodology of which is described in Chapter 5, uses students from the final year of the Product Design Engineering course at Glasgow School of Art and the University of Glasgow. The main aim of the experiment is to identify means of including sketching within the kind of text-based support methods discussed in Chapter 3. It also observes the volume and pattern of information produced by sketch activity throughout the conceptual stages of the design process and aims to find methods which would enable sketches to indicate the general progress of a design. The findings are detailed in Chapter 6

    Developing and Evaluating Visual Analogies to Support Insight and Creative Problem Solving

    Get PDF
    The primary aim of this thesis is to gain a richer understanding of visual analogies for insight problem solving, and, in particular, how they can be better developed to ensure their effectiveness as hints. While much work has explored the role of visual analogies in problem solving and their facilitative role, only a few studies have analysed how they could be designed. This thesis employs a mixed method consisting of a practice-led approach for studying how visual analogies can be designed and developed and an experimental research approach for testing their effectiveness as hints for solving visual insight problems

    Exploring Co-creative Drawing Workflows

    Get PDF
    This article presents the outcomes from a mixed-methods study of drawing practitioners (e.g., professional illustrators, fine artists, and art students) that was conducted in Autumn 2018 as a preliminary investigation for the development of a physical human-AI co-creative drawing system. The aim of the study was to discover possible roles that technology could play in observing, modeling, and possibly assisting an artist with their drawing. The study had three components: a paper survey of artists' drawing practises, technology usage and attitudes, video recorded drawing exercises and a follow-up semi-structured interview which included a co-design discussion on how AI might contribute to their drawing workflow. Key themes identified from the interviews were (1) drawing with physical mediums is a traditional and primary way of creation; (2) artists' views on AI varied, where co-creative AI is preferable to didactic AI; and (3) artists have a critical and skeptical view on the automation of creative work with AI. Participants' input provided the basis for the design and technical specifications of a co-creative drawing prototype, for which details are presented in this article. In addition, lessons learned from conducting the user study are presented with a reflection on future studies with drawing practitioners

    Eye Tracking Methods for Analysis of Visuo-Cognitive Behavior in Medical Imaging

    Get PDF
    Predictive modeling of human visual search behavior and the underlying metacognitive processes is now possible thanks to significant advances in bio-sensing device technology and machine intelligence. Eye tracking bio-sensors, for example, can measure psycho-physiological response through change events in configuration of the human eye. These events include positional changes such as visual fixation, saccadic movements, and scanpath, and non-positional changes such as blinks and pupil dilation and constriction. Using data from eye-tracking sensors, we can model human perception, cognitive processes, and responses to external stimuli. In this study, we investigated the visuo-cognitive behavior of clinicians during the diagnostic decision process for breast cancer screening under clinically equivalent experimental conditions involving multiple monitors and breast projection views. Using a head-mounted eye tracking device and a customized user interface, we recorded eye change events and diagnostic decisions from 10 clinicians (three breast-imaging radiologists and seven Radiology residents) for a corpus of 100 screening mammograms (comprising cases of varied pathology and breast parenchyma density). We proposed novel features and gaze analysis techniques, which help to encode discriminative pattern changes in positional and non-positional measures of eye events. These changes were shown to correlate with individual image readers' identity and experience level, mammographic case pathology and breast parenchyma density, and diagnostic decision. Furthermore, our results suggest that a combination of machine intelligence and bio-sensing modalities can provide adequate predictive capability for the characterization of a mammographic case and image readers diagnostic performance. Lastly, features characterizing eye movements can be utilized for biometric identification purposes. These findings are impactful in real-time performance monitoring and personalized intelligent training and evaluation systems in screening mammography. Further, the developed algorithms are applicable in other application domains involving high-risk visual tasks
    corecore