@@ -319,11 +319,11 @@ iteratively decide if a sample should be removed or not
3193193. Train a 1-KNN on `C `.
3203204. Go through the samples in set :math: `S`, sample by sample, and classify each one
321321 using a 1 nearest neighbor rule (trained in 3).
322- 5. If the sample is misclassified, add it to :math: `C`, otherwise do nothing .
322+ 5. If the sample is misclassified, add it to :math: `C`, and go to step 6 .
3233236. Repeat steps 3 to 5 until all observations in `S ` have been examined.
324324
325325The final dataset is `S `, containing all observations from the minority class and
326- those from the majority that were miss-classified by the 1-KNN algorithms.
326+ those from the majority that were miss-classified by the successive 1-KNN algorithms.
327327
328328The :class: `CondensedNearestNeighbour ` can be used in the following manner::
329329
@@ -334,7 +334,7 @@ The :class:`CondensedNearestNeighbour` can be used in the following manner::
334334 [(0, 64), (1, 24), (2, 115)]
335335
336336However, as illustrated in the figure below, :class: `CondensedNearestNeighbour `
337- is sensitive to noise and will add noisy samples.
337+ is sensitive to noise and may select noisy samples.
338338
339339In an attempt to remove noisy observations, :class: `OneSidedSelection `
340340will first find the observations that are hard to classify, and then will use
@@ -345,8 +345,8 @@ will first find the observations that are hard to classify, and then will use
3453452. Add a sample from the targeted class (class to be under-sampled) in
346346 :math: `C` and all other samples of this class in a set :math: `S`.
3473473. Train a 1-KNN on `C `.
348- 4. Using a 1 nearest neighbor rule trained in 3, classify all samples
349- in set :math: `S`.
348+ 4. Using a 1 nearest neighbor rule trained in 3, classify all samples in
349+ set :math: `S`.
3503505. Add all misclassified samples to :math: `C`.
3513516. Remove Tomek Links from :math: `C`.
352352
0 commit comments