52 research outputs found
Results of Various Tests of K-Nearest-Neighbour (KNN) to Recognise a Paraphrased Statement
In this paper, I tested a KNN algorithm that could recognise a paraphrased version of a statement entered in the Essay Helper GitHub Repository, which recognises reused statements while questioning the user to write a humanities-style essay
Recommended from our members
Entropy Based Feature Selection For Multi-Relational Naïve Bayesian Classifier
Current industries data’s are stored in relation structures. In usual approach to mine these data, we often use to join several relations to form a single relation using foreign key links, which is known as flatten. Flatten may cause troubles such as time consuming, data redundancy and statistical skew on data. Hence, the critical issues arise that how to mine data directly on numerous relations. The solution of the given issue is the approach called multi-relational data mining (MRDM). Other issues are irrelevant or redundant attributes in a relation may not make contribution to classification accuracy. Thus, feature selection is an essential data pre- processing step in multi-relational data mining. By filtering out irrelevant or redundant features from relations for data mining, we improve classification accuracy, achieve good time performance, and improve comprehensibility of the models. We had proposed the entropy based feature selection method for Multi-relational Naïve Bayesian Classifier. We have use method InfoDist and Pearson’s Correlation parameters, which will be used to filter out irrelevant and redundant features from the multi-relational database and will enhance classification accuracy. We analyzed our algorithm over PKDD financial dataset and achieved the better accuracy compare to the existing features selection methods
A Simple Relational Classifier
We analyze a Relational Neighbor (RN) classifier, a simple relational
predictive model that predicts only based on class labels of related neighbors,
using no learning and no inherent attributes.We show that it performs surprisingly
well by comparing it to more complex models such as Probabilistic Relational
Models and Relational Probability Trees on three data sets from published work.
We argue that a simple model such as this should be used as a baseline to assess
the performance of relational learners.NYU, Stern School of Business, IOMS department, Center for Digital Economy Researc
Similarity measures over refinement graphs
Similarity also plays a crucial role in support vector machines. Similarity assessment plays a key role in lazy learning methods such as k-nearest neighbor or case-based reasoning. In this paper we will show how refinement graphs, that were originally introduced for inductive learning, can be employed to assess and reason about similarity. We will define and analyze two similarity measures, S λ and S π, based on refinement graphs. The anti-unification-based similarity, S λ, assesses similarity by finding the anti-unification of two instances, which is a description capturing all the information common to these two instances. The property-based similarity, S π, is based on a process of disintegrating the instances into a set of properties, and then analyzing these property sets. Moreover these similarity measures are applicable to any representation language for which a refinement graph that satisfies the requirements we identify can be defined. Specifically, we present a refinement graph for feature terms, in which several languages of increasing expressiveness can be defined. The similarity measures are empirically evaluated on relational data sets belonging to languages of different expressiveness. © 2011 The Author(s).Support for this work came from the project Next-CBR TIN2009-13692-C03-01 (co-sponsored by EU FEDER funds)Peer Reviewe
- …