4,674 research outputs found

    Effects of environmental colour on mood: a wearable life colour capture device

    Get PDF
    Colour is everywhere in our daily lives and impacts things like our mood, yet we rarely take notice of it. One method of capturing and analysing the predominant colours that we encounter is through visual lifelogging devices such as the SenseCam. However an issue related to these devices is the privacy concerns of capturing image level detail. Therefore in this work we demonstrate a hardware prototype wearable camera that captures only one pixel - of the dominant colour prevelant in front of the user, thus circumnavigating the privacy concerns raised in relation to lifelogging. To simulate whether the capture of dominant colour would be sufficient we report on a simulation carried out on 1.2 million SenseCam images captured by a group of 20 individuals. We compare the dominant colours that different groups of people are exposed to and show that useful inferences can be made from this data. We believe our prototype may be valuable in future experiments to capture colour correlated associated with an individual's mood

    The colour of life: novel visualisations of population Lifestyles

    Get PDF
    Colour permeates our daily lives, yet we rarely take notice of it. In this work we utilise the SenseCam (a visual lifelogging tool), to investigte the predominant colours in one million minutes of human life that a group of 20 individuals encounter throughout their normal daily activities. We also compare the colours that different groups of people are exposed to in their typical days. This information is presented in using a novel colour-wheel visualisation which is a new means of illustrating that people are exposed to bright colours over longer durations of time during summer months, and more dark colours during winter months

    Automatic 2d image segmentation using tissue-like p system

    Get PDF
    This paper uses P-Lingua, a standard programming language that is designed specifically for P systems, to automatically simulate the region-based segmentation of 2D images. P-Lingua, which is based on membrane computing, links to Java Netbeans using the PLinguaCore4 Java library to automatically codify the pixels of the input image as long as automatically draw the output segmented image. Many methods have been suggested previously and used for artificial image segmentation, but to the best of our knowledge, none of those techniques were automatic, where the image was codified manually and the visualization of the output image was done manually in the tissue simulator which takes time and effort, especially when dealing with large images in the system. Two types of pixel adjacency have been utilized in this paper, namely; 4-adjacency and 8-adjacency. The jacquard index method has been used to measure the accuracy of the segmentation. The results of the proposed method demonstrated that beside its ability to automatically segmenting 2D images with arbitrary sizes, it is more efficient and faster than the tissue simulator tool, since the latter needs the input image to be codified manually pixel by pixel which makes it impractical for real-world applications

    Painterly rendering techniques: A state-of-the-art review of current approaches

    Get PDF
    In this publication we will look at the different methods presented over the past few decades which attempt to recreate digital paintings. While previous surveys concentrate on the broader subject of non-photorealistic rendering, the focus of this paper is firmly placed on painterly rendering techniques. We compare different methods used to produce different output painting styles such as abstract, colour pencil, watercolour, oriental, oil and pastel. Whereas some methods demand a high level of interaction using a skilled artist, others require simple parameters provided by a user with little or no artistic experience. Many methods attempt to provide more automation with the use of varying forms of reference data. This reference data can range from still photographs, video, 3D polygonal meshes or even 3D point clouds. The techniques presented here endeavour to provide tools and styles that are not traditionally available to an artist. Copyright © 2012 John Wiley & Sons, Ltd

    Phonological processing deficit - a culprit behind developmental dyslexia?

    Get PDF
    Udostępnienie publikacji Wydawnictwa Uniwersytetu Łódzkiego finansowane w ramach projektu „Doskonałość naukowa kluczem do doskonałości kształcenia”. Projekt realizowany jest ze środków Europejskiego Funduszu Społecznego w ramach Programu Operacyjnego Wiedza Edukacja Rozwój; nr umowy: POWER.03.05.00-00-Z092/17-00

    Tissue-like P system for Segmentation of 2D Hexagonal Images

    Get PDF
    Membrane computing, which is a new computational model inspired by the structure and functioning of biological cells and by the way the cells are organized in tissues. MC has been adopted in many real world applications including image segmentation. In contrast to the traditional square grid for representing and sampling digital images, hexagonal grid is an alternative efficient mechanism which can better represents and visualizes the curved objects. In this paper, a tissue-like P system with region-based and edge-based segmentation is used to segment two dimensional hexagonal images, wherein P-Lingua programming language is used to implement and validate the proposed system. The achieved experimental results clearly demonstrated the effectiveness of using hexagonal connectivity to segment two dimensional images in a less number of rules and computational steps. Moreover, the results reveal that this approach has the potential of segmenting large images in few number of steps

    STV-based Video Feature Processing for Action Recognition

    Get PDF
    In comparison to still image-based processes, video features can provide rich and intuitive information about dynamic events occurred over a period of time, such as human actions, crowd behaviours, and other subject pattern changes. Although substantial progresses have been made in the last decade on image processing and seen its successful applications in face matching and object recognition, video-based event detection still remains one of the most difficult challenges in computer vision research due to its complex continuous or discrete input signals, arbitrary dynamic feature definitions, and the often ambiguous analytical methods. In this paper, a Spatio-Temporal Volume (STV) and region intersection (RI) based 3D shape-matching method has been proposed to facilitate the definition and recognition of human actions recorded in videos. The distinctive characteristics and the performance gain of the devised approach stemmed from a coefficient factor-boosted 3D region intersection and matching mechanism developed in this research. This paper also reported the investigation into techniques for efficient STV data filtering to reduce the amount of voxels (volumetric-pixels) that need to be processed in each operational cycle in the implemented system. The encouraging features and improvements on the operational performance registered in the experiments have been discussed at the end

    Accelerated hardware video object segmentation: From foreground detection to connected components labelling

    Get PDF
    This is the preprint version of the Article - Copyright @ 2010 ElsevierThis paper demonstrates the use of a single-chip FPGA for the segmentation of moving objects in a video sequence. The system maintains highly accurate background models, and integrates the detection of foreground pixels with the labelling of objects using a connected components algorithm. The background models are based on 24-bit RGB values and 8-bit gray scale intensity values. A multimodal background differencing algorithm is presented, using a single FPGA chip and four blocks of RAM. The real-time connected component labelling algorithm, also designed for FPGA implementation, run-length encodes the output of the background subtraction, and performs connected component analysis on this representation. The run-length encoding, together with other parts of the algorithm, is performed in parallel; sequential operations are minimized as the number of run-lengths are typically less than the number of pixels. The two algorithms are pipelined together for maximum efficiency
    corecore