<?xml version="1.0" encoding="utf-8" ?><rss version="2.0"><channel><title>Bing: Knn Algorithm in Machine Learning for CKD</title><link>http://www.bing.com:80/search?q=Knn+Algorithm+in+Machine+Learning+for+CKD</link><description>Search results</description><image><url>http://www.bing.com:80/s/a/rsslogo.gif</url><title>Knn Algorithm in Machine Learning for CKD</title><link>http://www.bing.com:80/search?q=Knn+Algorithm+in+Machine+Learning+for+CKD</link></image><copyright>Copyright © 2026 Microsoft. All rights reserved. These XML results may not be used, reproduced or transmitted in any manner or for any purpose other than rendering Bing results within an RSS aggregator for your personal, non-commercial use. Any other use of these results requires express written permission from Microsoft Corporation. By accessing this web page or using these results in any manner whatsoever, you agree to be bound by the foregoing restrictions.</copyright><item><title>K-Nearest Neighbor (KNN) Algorithm - GeeksforGeeks</title><link>https://www.geeksforgeeks.org/machine-learning/k-nearest-neighbours/</link><description>K‑Nearest Neighbor (KNN) is a simple and widely used machine learning technique for classification and regression tasks. It works by identifying the K closest data points to a given input and making predictions based on the majority class or average value of those neighbors.</description><pubDate>Fri, 17 Apr 2026 12:02:00 GMT</pubDate></item><item><title>k-nearest neighbors algorithm - Wikipedia</title><link>https://en.wikipedia.org/wiki/K-nearest_neighbors_algorithm</link><description>^ a b Mirkes, Evgeny M.; KNN and Potential Energy: applet Archived 2012-01-19 at the Wayback Machine, University of Leicester, 2011 ^ Ramaswamy, Sridhar; Rastogi, Rajeev; Shim, Kyuseok (2000). "Efficient algorithms for mining outliers from large data sets". Proceedings of the 2000 ACM SIGMOD international conference on Management of data ...</description><pubDate>Wed, 08 Apr 2026 02:11:00 GMT</pubDate></item><item><title>What is the k-nearest neighbors (KNN) algorithm? - IBM</title><link>https://www.ibm.com/think/topics/knn</link><description>The k-nearest neighbors (KNN) algorithm is a non-parametric, supervised learning classifier, which uses proximity to make classifications or predictions about the grouping of an individual data point.</description><pubDate>Sun, 19 Apr 2026 17:00:00 GMT</pubDate></item><item><title>KNeighborsClassifier — scikit-learn 1.8.0 documentation</title><link>https://scikit-learn.org/stable/modules/generated/sklearn.neighbors.KNeighborsClassifier.html</link><description>This means that knn.fit(X, y).score(None, y) implicitly performs a leave-one-out cross-validation procedure and is equivalent to cross_val_score(knn, X, y, cv=LeaveOneOut()) but typically much faster.</description><pubDate>Sun, 19 Apr 2026 22:00:00 GMT</pubDate></item><item><title>What is k-Nearest Neighbor (kNN)? | A Comprehensive k-Nearest Neighbor ...</title><link>https://www.elastic.co/what-is/knn</link><description>kNN, or the k-nearest neighbor algorithm, is a machine learning algorithm that uses proximity to compare one data point with a set of data it was trained on and has memorized to make predictions.</description><pubDate>Sat, 18 Apr 2026 23:13:00 GMT</pubDate></item><item><title>K-Nearest Neighbors (KNN) in Machine Learning</title><link>https://www.tutorialspoint.com/machine_learning/machine_learning_knn_nearest_neighbors.htm</link><description>K-nearest neighbors (KNN) algorithm is a type of supervised ML algorithm which can be used for both classification as well as regression predictive problems. However, it is mainly used for classification predictive problems in industry.</description><pubDate>Sat, 18 Apr 2026 22:59:00 GMT</pubDate></item><item><title>KNN Algorithm – K-Nearest Neighbors Classifiers and Model Example</title><link>https://www.freecodecamp.org/news/k-nearest-neighbors-algorithm-classifiers-and-model-example/</link><description>January 25, 2023 / #algorithms KNN Algorithm – K-Nearest Neighbors Classifiers and Model Example Ihechikara Abba</description><pubDate>Sun, 19 Apr 2026 23:05:00 GMT</pubDate></item><item><title>k-Nearest Neighbors Algorithm - an overview - ScienceDirect</title><link>https://www.sciencedirect.com/topics/computer-science/k-nearest-neighbors-algorithm</link><description>The KNN algorithm is one of the simplest machine learning algorithms: It assigns to the profile or feature vector xi the most common modality of Y among its k “nearest neighbors.”</description><pubDate>Sat, 18 Apr 2026 23:35:00 GMT</pubDate></item><item><title>K-Nearest Neighbor (KNN) Algorithm: Use Cases and Tips - G2</title><link>https://learn.g2.com/k-nearest-neighbor</link><description>KNN classifies or predicts outcomes based on the closest data points it can find in its training set. Think of it as asking your neighbors for advice; whoever’s closest gets the biggest say.</description><pubDate>Sat, 18 Apr 2026 23:20:00 GMT</pubDate></item><item><title>K-Nearest Neighbors (kNN) - Explained | Towards Data Science</title><link>https://towardsdatascience.com/k-nearest-neighbors-knn-explained-cbc31849a7e3/</link><description>K-nearest neighbors (kNN) is a supervised machine learning algorithm that can be used to solve both classification and regression tasks. I see kNN as an algorithm that comes from real life.</description><pubDate>Sat, 18 Apr 2026 10:34:00 GMT</pubDate></item></channel></rss>