54 research outputs found
Desarrollo de un sistema de catalogación y gestión de fotografías: Photon = Development of a system for cataloging and managing photos: Photon
The objective of this research is to propose a model of comprehensive analysis of photography, from the current descriptive standards. From this model, we develop a system for cataloguing and management of photo files. Its design enables it to manage all types of authorities involved in the description of the image and editing tools for integrating thesauri, details and typologies. Thus the Photon program can extract details afterimages as a dependent of an original picture, thereby improving the reuse and retrieval of the photographic
El problema de las noticias falsas: detección y contramedidas
In this research we analyse the fenomena, motivations, mechanisms, and rebroadcasting vectors of fake news. In other hand, was review the main scientific solutions and identified their countermeasures. This is due to several factors, for example new topics and facts available in published news, of which there aren´t constancy or previous experience in knowledge base; the use of reverse engineering to counteract machine learning software; the difficulties to find reliable, impartial and independent information sources; and finally, the difficulties to develop technologies to assess the facts and evidences cited on news. We reach the conclussion that, it´s feasible designing methods for detect an important part of fraud news, but it´s not enought. It´s because application environment it´s restricted to a few sources, thematics or samples, that don´t represent the open environment we face. We need improve the fake news knowledge base, develop more efficient semantic models, get better informative characteristics on news, recover a new version of RSS rebroadcasting system and set previous filters before users feed on the news
El problema de las noticias falsas: detección y contramedidas
In this research we analyse the fenomena, motivations, mechanisms, and rebroadcasting vectors of fake news. In other hand, was review the main scientific solutions and identified their countermeasures. This is due to several factors, for example new topics and facts available in published news, of which there aren´t constancy or previous experience in knowledge base; the use of reverse engineering to counteract machine learning software; the difficulties to find reliable, impartial and independent information sources; and finally, the difficulties to develop technologies to assess the facts and evidences cited on news. We reach the conclussion that, it´s feasible designing methods for detect an important part of fraud news, but it´s not enought. It´s because application environment it´s restricted to a few sources, thematics or samples, that don´t represent the open environment we face. We need improve the fake news knowledge base, develop more efficient semantic models, get better informative characteristics on news, recover a new version of RSS rebroadcasting system and set previous filters before users feed on the news
Método para la extracción masiva de canales de sindicación
One of the problems for investigating the informative production of syndication channels is counting on the sufficient number of sources from the same domain, subject or area of knowledge, to compile a sample. This is a consequence of the dispersion of information sources on the Web; the researcher’s difficulty in knowing all the available resources; and the difficulty in extracting and locating the links of syndication channels in every relevant web site or Internet resource that is discovered. This article discusses the method to extract and compile syndication channels through the composition of seeds using a web crawler, and the configuration and subsequent processing of the obtained links.Uno de los problemas para la investigación de la producción informativa de canales de sindicación es contar con la cantidad de fuentes suficientes y del mismo dominio, temática o área de conocimiento, para componer una muestra. Esto se debe a la dispersión de las fuentes de información en la Web y por otra parte a la dificultad del investigador para conocer todos los recursos disponibles. A estos problemas se suma la dificultad de extraer y localizar los enlaces de los canales de sindicación en cada recurso o sitio web pertinente que se descubre. En este artículo se aborda el método para extraer los canales de sindicación mediante la composición de semillas para el análisis, con programas web crawler, identificando la configuración y posterior preparación
Desarrollo de un sistema de catalogación y gestión de fotografías: Photon
The purpose of this article is to propose a model for the comprehensive analysis of photographis, working on the current descriptive standards. From this model, we develop a system for cataloguing and managing photo files. All types of authorities involved in the description of the image can be managed and the system includes editing tools for integrating thesauri, details and typologies. In particular, the Photon program can extract secondary dependent images from the details of an original picture, thereby improving the reuse and retrieval of photographs.El objetivo de la investigación es plantear un modelo de análisis exhaustivo de la fotografía, a partir de las normas de descripción vigentes. Con este modelo, se desarrolla un sistema para la catalogación y gestión de archivos fotográficos. Su diseño lo habilita para gestionar todo tipo de autoridades involucradas en la descripción de la imagen, integrando herramientas para la edición de tesauros y detalles. De esta forma con el programa Photon, se pueden extraer detalles a modo de imágenes secundarias dependientes de una fotografía original, mejorando con ello las posibilidades de reutilización y recuperación de los fondos fotográfico
Web analysis and usability: running trial of mbot webcrawler
[Resumen]
La presente investigación tiene como objetivo el desarrollo de un programa de análisis de la web,
denominado mbot, que facilite la elaboración de estudios cibermétricos y en especial la obtención de
datos sobre la tipología de los contenidos de un determinado área de conocimiento en la web, distribución
o estratificación según los niveles de análisis empleados, número total de sitios, dominios y páginas
analizadas, tamaño o volumen de los contenidos. Para demostrar su funcionamiento y comprobar sus
capacidades se han llevado a cabo diversas pruebas entre las que destaca el análisis exhaustivo de dos
centros de investigación internacionales: NASA y ESA. En ambos casos se extrae y cuantifica toda la
información disponible, a fin de elaborar un estudio comparativo sobre su topografía, clasificación, tipos
documentales, recursos, vínculos y contenidos. Por otro lado, se analizará cómo el factor de usabilidad,
accesibilidad y arquitectura de la web afectan en el reconocimiento de patrones, en el código fuente de las
páginas analizadas.[Abstract]
This research has as primary goal the development of a new web analysis application, called mbot, that
may help in the cybermetric studies and specially in the obtaining of data about the documental typology
contained in a given knowledge area in the Web, its distribution or stratification according to the different
analysis levels used, total amount of sites, domains and pages analyzed and size or volume of contents.
To prove its operational capabilities we have run several trials, among which we want to underline the
exhaustive analysis of two international research centres: ESA and NASA. In both cases we have extracted
and quantified all available information, in order to perform a comparative study of their topography,classification, documentary types, resources, links and contents. On the other side the factor usability is
taken into consideration, including accessibility and architecture of the web, as long as they have any
incidence in the recognition of patterns in the source code of the researched pages
Análisis de la web y usabilidad: prueba de funcionamiento de Mbot webcrawler = Web analysis and usability: running trial of mbot webcrawler
This research has as primary goal the development of a new web analysis application, called mbot, that may help in the cybermetric studies and specially in the obtaining of data about the documental typology contained in a given knowledge area in the Web, its distribution or stratification according to the different analysis levels used, total amount of sites, domains and pages analyzed and size or volume of contents. To prove its operational capabilities we have run several trials, among which we want to underline the exhaustive analysis of two international research centres, ESA and NASA. In both cases we have extracted and quantified all available information, in order to perform a comparative study of their topography, classification, documentary types, resources, links and contents. On the other side the factor usability is taken into consideration, including accessibility and architecture of the web, as long as they have any incidence in the recognition of patterns in the source code of the researched pages
Design of an aggregator for managing informative big data
The design and characteristics of a new open source content aggregation program, AXYZ, are described. Several features of the program standout, including the processing engine of syndication channels, monitoring capability of information recovery in real time, possibility of configuration of the aggregator behavior, automatic content classification, and new models for representing information from relational interactive maps. On the other hand, the aggregation program is designed to manage thousands of syndication channels in the RSS format. It also provides statistics that can be used to study the production of any information producer and the impact of the information published in other sources. The AXYZ modules are capable of comparing the relationship between news or information from different sources and the degree of influence which is detected by patterns
Ribonucleic acid (RNA) virus and coronavirus in Google Dataset Search: their scope and epidemiological correlation
This paper presents an analysis of the publication of datasets collected via
Google Dataset Search, specialized in families of RNA viruses, whose
terminology was obtained from the National Cancer Institute (NCI) thesaurus
developed by the US Department of Health and Human Services. The objective is
to determine the scope and reuse capacity of the available data, determine the
number of datasets and their free access, the proportion in reusable download
formats, the main providers, their publication chronology, and to verify their
scientific provenance. On the other hand, we also define possible relationships
between the publication of datasets and the main pandemics that have occurred
during the last 10 years. The results obtained highlight that only 52% of the
datasets are related to scientific research, while an even smaller fraction
(15%) are reusable. There is also an upward trend in the publication of
datasets, especially related to the impact of the main epidemics, as clearly
confirmed for the Ebola virus, Zika, SARS-CoV, H1N1, H1N5, and especially the
SARS-CoV-2 coronavirus. Finally, it is observed that the search engine has not
yet implemented adequate methods for filtering and monitoring the datasets.
These results reveal some of the difficulties facing open science in the
dataset field
- …