55 research outputs found

    Using Text Segmentation to Enhance the Cluster Hypothesis

    Get PDF
    An alternative way to tackle Information Retrieval, called Passage Retrieval, considers text fragments independently rather than assessing global relevance of documents. In such a context, the fact that relevant information is surrounded by parts of text deviating from the interesting topic does not penalize the document. In this paper, we propose to study the impact of the consideration of these text fragments on a document clustering process. The use of clustering in the field of Information Retrieval is mainly supported by the cluster hypothesis which states that relevant documents tend to be more similar one to each other than to non-relevant documents and hence a clustering process is likely to gather them. Previous experiments have shown that clustering the first retrieved documents as response to a user’s query allows the Information Retrieval systems to improve their effectiveness. In the clustering process used in these studies, documents have been considered globally. Nevertheless, the assumption stating that a document can refer to more than one topic/concept may have also impacts on the document clustering process. Considering passages of the retrieved documents separately may allow to create more representative clusters of the addressed topics. Different approaches have been assessed and results show that using text fragments in the clustering process may turn out to be actually relevant

    Parity-Violating Interaction Effects in the np System

    Full text link
    We investigate parity-violating observables in the np system, including the longitudinal asymmetry and neutron-spin rotation in np elastic scattering, the photon asymmetry in np radiative capture, and the asymmetries in deuteron photo-disintegration d(gamma,n)p in the threshold region and electro-disintegration d(e,e`)np in quasi-elastic kinematics. To have an estimate of the model dependence for the various predictions, a number of different, latest-generation strong-interaction potentials--Argonne v18, Bonn 2000, and Nijmegen I--are used in combination with a weak-interaction potential consisting of pi-, rho-, and omega-meson exchanges--the model known as DDH. The complete bound and scattering problems in the presence of parity-conserving, including electromagnetic, and parity-violating potentials is solved in both configuration and momentum space. The issue of electromagnetic current conservation is examined carefully. We find large cancellations between the asymmetries induced by the parity-violating interactions and those arising from the associated pion-exchange currents. In the np capture, the model dependence is nevertheless quite small, because of constraints arising through the Siegert evaluation of the relevant E1 matrix elements. In quasi-elastic electron scattering these processes are found to be insignificant compared to the asymmetry produced by gamma-Z interference on individual nucleons.Comment: 65 pages, 26 figures, submitted to PR

    Bir Procrutes Hikâyesi: Türkçe Fransızca Gibi İşlenirmi ?

    Get PDF
    International audienc

    Incidence of mandibular osteoradionecrosis (MORN) after intensity modulated radiotherapy (IMRT) versus 3D conformal radiotherapy (3D-CRT): A systematic review

    Get PDF
    Background: Analyze the incidence of MORN after head and neck radiotherapy by two novel irradiation techniques. 3DCRT and IMRT and compare the success rates of distinct authors. Material and Methods: An electronic search in Pubmed (MEDLINE), Ovid. Google Scholar and Cochrane Library (Wiley). databases was conducted with the key words Radiotherapy, Conformal[Mesh] OR Radiotherapy. Intensity-Modulated[Mesh]) AND Osteoradionecrosis[Mesh] for all databases. The inclusion criteria randomized controlled trials (RCT), as well as prospective and retrospective cohort studies published in English; MORN patients treated with 3D-CRT y IMRT. Results: 27 articles were selected from 194 initially found. 14 articles out of 27 were excluded and finally included 8 publications were included in the systematic review that were ranked according to their level of scientific evidence using the SORT criteria. Conclusions: When both RT techniques were compared; IMRT revealed a lower risk incidence of MORN development and enhanced dose constraint than 3D-CRT (less than 10%). this improvement could translate into less complications post RT treatment

    RQM description of the charge form factor of the pion and its asymptotic behavior

    Full text link
    The pion charge and scalar form factors, F1(Q2)F_1(Q^2) and F0(Q2)F_0(Q^2), are first calculated in different forms of relativistic quantum mechanics. This is done using the solution of a mass operator that contains both confinement and one-gluon-exchange interactions. Results of calculations, based on a one-body current, are compared to experiment for the first one. As it could be expected, those point-form, and instant and front-form ones in a parallel momentum configuration fail to reproduce experiment. The other results corresponding to a perpendicular momentum configuration (instant form in the Breit frame and front form with q+=0q^+=0) do much better. The comparison of charge and scalar form factors shows that the spin-1/2 nature of the constituents plays an important role. Taking into account that only the last set of results represents a reasonable basis for improving the description of the charge form factor, this one is then discussed with regard to the asymptotic QCD-power-law behavior Q2Q^{-2}. The contribution of two-body currents in achieving the right power law is considered while the scalar form factor, F0(Q2)F_0(Q^2), is shown to have the right power-law behavior in any case. The low-Q2Q^2 behavior of the charge form factor and the pion-decay constant are also discussed.}Comment: 30 pages, 10 figure

    Traveling Among Clusters: A Way to Reconsider the Benefits of the Cluster Hypothesis

    Get PDF
    Relying on the Cluster Hypothesis which states that relevant documents tend to be more similar one to each other than to non-relevant documents, most of information retrieval systems organizing search results as a set of clusters seek to gather all relevant documents in the same cluster. We propose here to reconsider the benefits of the entailed concentration of the relevant information. Contrary to what is commonly admitted, we believe that systems which aim to distribute the relevant documents in different clusters, since being more likely to highlight different aspects of the subject, may be at least as useful for the user as systems gathering all relevant documents in a single group. Since existing evaluation measures tend to greatly favor the latter systems, we first investigate ways to more fairly assess the ability to reach the relevant information from the list of cluster descriptions. At last, we show that systems distributing the relevant information in different clusters may actually provide a better information access than classical systems

    Toward a More Global and Coherent Segmentation of Texts

    Get PDF
    The automatic text segmentation task consists of identifying the most important thematic breaks in a document in order to cut it into homogeneous passages. Text segmentation has motivated a large amount of research. We focus here on the statistical approaches that rely on an analysis of the distribution of the words in the text. Usually, the segmentation of texts is realized sequentially on the basis of very local clues. However, such an approach prevents the consideration of the text in a global way, particularly concerning the granularity degree adopted for the expression of the different topics it addresses. We thus propose here two new segmentation algorithms—ClassStruggle and SegGen—which use criteria rendering global views of texts. ClassStruggle is based on an initial clustering of the sentences of the text, thus allowing the consideration of similarities within a group rather than individually. It relies on the distribution of the occurrences of the members of each class 1 to segment the texts. SegGen proposes to evaluate potential segmentations of the whole text thanks to a genetic algorithm. It attempts to find a solution of segmentation optimizing two criteria, the maximization of the internal cohesion of the segments and the minimization of the similarity between adjacent ones. According to experimental results, both approaches appear to be very competitive compared to existing methods

    Chiral effective theory predictions for deuteron form factor ratios at low Q^2

    Get PDF
    We use chiral effective theory to predict the deuteron form factor ratio G_C/G_Q as well as ratios of deuteron to nucleon form factors. These ratios are calculated to next-to-next-to-leading order. At this order the chiral expansion for the NN isoscalar charge operator (including consistently calculated 1/M corrections) is a parameter-free prediction of the effective theory. Use of this operator in conjunction with NLO and NNLO chiral effective theory wave functions produces results that are consistent with extant experimental data for Q^2 < 0.35 GeV^2. These wave functions predict a deuteron quadrupole moment G_Q(Q^2=0)=0.278-0.282 fm^2-with the variation arising from short-distance contributions to this quantity. The variation is of the same size as the discrepancy between the theoretical result and the experimental value. This motivates the renormalization of G_Q via a two-nucleon operator that couples to quadrupole photons. After that renormalization we obtain a robust prediction for the shape of G_C/G_Q at Q^2 < 0.3 GeV^2. This allows us to make precise, model-independent predictions for the values of this ratio that will be measured at the lower end of the kinematic range explored at BLAST. We also present results for the ratio G_C/G_M.Comment: 31 pages, 7 figure

    Managing Genetic Algorithm Parameters to Improve SegGen - A Thematic Segmentation Algorithm

    Get PDF
    SegGen [1] is a linear thematic segmentation algorithm grounded on a variant of the Strength Pareto Evolutionary Algorithm [2] and aims at optimizing the two criteria of the Salton\u27s [3] definition of segments: a segment is a part of text whose internal cohesion and dissimilarity with its adjacent segments are maximal. This paper describes improvements that have been implemented in the approach taken by SegGen by tuning the genetic algorithm parameters according with the evolution of the quality of the generated populations. Two kinds of reasons originate the tuning of the parameters and have been implemented here. First as it could be measured by the values of global criteria of the population quality, the global quality of the generated populations increases as the process goes and it seems reasonable to set values to parameters and define new operators, which favor intensification and diminish diversification factors in the search process. Second since individuals in the populations are plausible segmentations it seems reasonable to weight sentences in the current segmentation depending on their distance to the boundaries of the segment they belong to for the calculus of similarities between sentences implied in the two criteria to be optimized. Although this tuning of the parameters of the algorithm currently rests on estimations based on experiments, first results are promising
    corecore