The k-nearest neighbors (KNN) technique is a straightforward supervised machine learning algorithm that can tackle classification and regression problems. It's simple to set up and comprehend, but it has the major disadvantage of being substantially slower as the bulk of the data in use grows.
Because it delivers highly precise predictions, the KNN algorithm can compete with the most accurate models. As a result, the KNN algorithm can be used for applications that require high accuracy but do not require a human-readable model. The accuracy of the predictions is determined by the distance measure.
Agriculture, Finance, Medical.
System of recommendations (Amazon, Hulu, Netflix, etc)
During the training phase, the KNN algorithm simply stores the dataset and then classifies it into a category that is quite similar to the incoming data. Assume we have an image of a critter that resembles a cat or a dog and we want to know whether it is a cat or a dog.
I'd recommend standardizing the data between 0 and 1 for k-NN. All of the features must have the same range of values in order for them to be of equal importance when computing the distance. This can only be accomplished by normalization.
where is the finaldata.csv
great learning plateform kushal sir is really too good