A novel version of k nearest neighbor: Dependent nearest neighbor

dc.authoridTağluk, M. Emin/0000-0001-7789-6376
dc.authoridERTUGRUL, Ömer Faruk/0000-0003-0710-0867
dc.authorwosidTağluk, M. Emin/ABH-1005-2020
dc.authorwosidERTUGRUL, Ömer Faruk/F-7057-2015
dc.contributor.authorErtugrul, Omer Faruk
dc.contributor.authorTagluk, Mehmet Emin
dc.date.accessioned2024-08-04T20:43:01Z
dc.date.available2024-08-04T20:43:01Z
dc.date.issued2017
dc.departmentİnönü Üniversitesien_US
dc.description.abstractk nearest neighbor (kNN) is one of the basic processes behind various machine learning methods In kNN, the relation of a query to a neighboring sample is basically measured by a similarity metric, such as Euclidean distance. This process starts with mapping the training dataset onto a one-dimensional distance space based on the calculated similarities, and then labeling the query in accordance with the most dominant or mean of the labels of the k nearest neighbors, in classification or regression issues, respectively. The number of nearest neighbors (k) is chosen according to the desired limit of success. Nonetheless, two distinct samples may have equal distances to query but, with different angles in the feature space. The similarity of the query to these two samples needs to be weighted in accordance with the angle going between the query and each of the samples to differentiate between the two distances in reference to angular information. This opinion can be analyzed in the context of dependency and can be utilized to increase the precision of classifier. With this point of view, instead of kNN, the query is labeled according to its nearest dependent neighbors that are determined by a joint function, which is built on the similarity and the dependency. This method, therefore, may be called dependent NN (d-NN). To demonstrate d-NN, it is applied to synthetic datasets, which have different statistical distributions, and 4 benchmark datasets, which are Pima Indian, Hepatitis, approximate Sinc and CASP datasets. Results showed the superiority of d-NN in terms of accuracy and computation cost as compared to other employed popular machine learning methods. (C) 2017 Elsevier B.V. All rights reserved.en_US
dc.identifier.doi10.1016/j.asoc.2017.02.020
dc.identifier.endpage490en_US
dc.identifier.issn1568-4946
dc.identifier.issn1872-9681
dc.identifier.scopus2-s2.0-85014730670en_US
dc.identifier.scopusqualityQ1en_US
dc.identifier.startpage480en_US
dc.identifier.urihttps://doi.org/10.1016/j.asoc.2017.02.020
dc.identifier.urihttps://hdl.handle.net/11616/97724
dc.identifier.volume55en_US
dc.identifier.wosWOS:000400031600035en_US
dc.identifier.wosqualityQ1en_US
dc.indekslendigikaynakWeb of Scienceen_US
dc.indekslendigikaynakScopusen_US
dc.language.isoenen_US
dc.publisherElsevieren_US
dc.relation.ispartofApplied Soft Computingen_US
dc.relation.publicationcategoryMakale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanıen_US
dc.rightsinfo:eu-repo/semantics/closedAccessen_US
dc.subjectDependencyen_US
dc.subjectSimilarityen_US
dc.subjectk nearest neighboren_US
dc.subjectDependent nearest neighboren_US
dc.titleA novel version of k nearest neighbor: Dependent nearest neighboren_US
dc.typeArticleen_US

Dosyalar