21,784 research outputs found
Webpage design optimization using genetic algorithm driven CSS
In the rapid emergence of globalization, e-commerce, and internet accessibility in remote parts of the world, ongoing feedback and participation from site visitors are essential for attaining clear and effective communication on a web site. This thesis presents a computational experiment for optimizing design of a webpage in an evolutionary manner. Webpage personalization is viewed as a configuration problem whose goal is to determine the optimal presentation of a webpage while taking into account the preference of the web author (designer), layout constraints (web design/editing language: HTML, CSS), and viewer interaction with the browser. The study proposes use of genetic algorithm-driven Cascading Style Sheets (CSS) to assist the process of webpage design optimization. This method will engage visitors to remotely modify and enhance the style (type, layout and color) of web site to fit their aesthetic and functional representation of well-received design. The preference feedback from user will be stored in an application server for automated evolutionary selection process and reinitialized for the next generation of users. Through the experimentation of web prototype and user evaluation test, the implementation of this method is examined and the derived design solutions are analyzed based on web aesthetics, standards, and accessibility
Crowdsourcing malaria parasite quantification: an online game for analyzing images of infected thick blood smears
Background: There are 600,000 new malaria cases daily worldwide. The gold standard for estimating the parasite burden and the corresponding severity of the disease consists in manually counting the number of parasites in blood smears through a microscope, a process that can take more than 20 minutes of an expert microscopist’s time.
Objective: This research tests the feasibility of a crowdsourced approach to malaria image analysis. In particular, we investigated whether anonymous volunteers with no prior experience would be able to count malaria parasites in digitized images of thick blood smears by playing a Web-based game.
Methods: The experimental system consisted of a Web-based game where online volunteers were tasked with detecting parasites in digitized blood sample images coupled with a decision algorithm that combined the analyses from several players to produce an improved collective detection outcome. Data were collected through the MalariaSpot website. Random images of thick blood films containing Plasmodium falciparum at medium to low parasitemias, acquired by conventional optical microscopy, were presented to players. In the game, players had to find and tag as many parasites as possible in 1 minute. In the event that players found all the parasites present in the image, they were presented with a new image. In order to combine the choices of different players into a single crowd decision, we implemented an image processing pipeline and a quorum algorithm that judged a parasite tagged when a group of players agreed on its position.
Results: Over 1 month, anonymous players from 95 countries played more than 12,000 games and generated a database of more than 270,000 clicks on the test images. Results revealed that combining 22 games from nonexpert players achieved a parasite counting accuracy higher than 99%. This performance could be obtained also by combining 13 games from players trained for 1 minute. Exhaustive computations measured the parasite counting accuracy for all players as a function of the number of games considered and the experience of the players. In addition, we propose a mathematical equation that accurately models the collective parasite counting performance.
Conclusions: This research validates the online gaming approach for crowdsourced counting of malaria parasites in images of thick blood films. The findings support the conclusion that nonexperts are able to rapidly learn how to identify the typical features of malaria parasites in digitized thick blood samples and that combining the analyses of several users provides similar parasite counting accuracy rates as those of expert microscopists. This experiment illustrates the potential of the crowdsourced gaming approach for performing routine malaria parasite quantification, and more generally for solving biomedical image analysis problems, with future potential for telediagnosis related to global health challenges
University Library Websites in Kerala: An Analysis (Web Survey)
The article is an analysis of library websites or web
pages of the universities in Kerala. Factors like
speed, size, downloading time, facilities for
information services etc. have been analyzed. The
survey was conducted during the period from 14-01-
2013 to 19-01-2013. The study reveals that though
the websites provide lot of useful information to the
users, further improvement both in contents and
management of it is needed in most of the library
websites. The study also provided insight to judge the
quality of the library websites and information
services provided through them
A Benchmark Suite for Template Detection and Content Extraction
Template detection and content extraction are two of the main areas of
information retrieval applied to the Web. They perform different analyses over
the structure and content of webpages to extract some part of the document.
However, their objective is different. While template detection identifies the
template of a webpage (usually comparing with other webpages of the same
website), content extraction identifies the main content of the webpage
discarding the other part. Therefore, they are somehow complementary, because
the main content is not part of the template. It has been measured that
templates represent between 40% and 50% of data on the Web. Therefore,
identifying templates is essential for indexing tasks because templates usually
contain irrelevant information such as advertisements, menus and banners.
Processing and storing this information is likely to lead to a waste of
resources (storage space, bandwidth, etc.). Similarly, identifying the main
content is essential for many information retrieval tasks. In this paper, we
present a benchmark suite to test different approaches for template detection
and content extraction. The suite is public, and it contains real heterogeneous
webpages that have been labelled so that different techniques can be suitable
(and automatically) compared.Comment: 13 pages, 3 table
Formulating representative features with respect to document genre classification
Genre classification (e.g. whether a document
is a scientific article or magazine article) is closely
bound to the physical and conceptual structure of document
as well as the level of depth involved in the text.
Hence, it provides a means of ranking documents retrieved
by search tools according to metrics other than
topical similarity. Moreover, the structural information
derived from genre classification can be used to locate
target information within the text. In previous studies,
the detection of genre classes has been attempted
by using some normalised frequency of terms or combinations
of terms in the document (here, we are using
term as a reference to words, phrases, syntactic
units, sentences and paragraphs, as well as other patterns
derived from deeper linguistic or semantic analysis).
These approaches largely neglect how the term is
distributed throughout the document. Here, we report
the results of automated experiments based on distributive
statistics of words in order to present evidence that
term distribution pattern is a better indicator of genre
class than term frequency.
- …