You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+8-6Lines changed: 8 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -246,17 +246,18 @@ LinearSVC is a python class from Scikit Learn library
246
246
```
247
247
248
248
LinearSVC performs classification.
249
-
LinearSVC finds a linear separator. A line separating classes.
249
+
LinearSVC finds a linear separator. A line separating classes.
250
250
There are many linear separators: It will choose the optimal one, i.e the one that maximizes our confidence, i.e the one that maximizes the geometrical margin, i.e the one that maximizes the distance between itself and the closest/nearest data point point
251
+
251
252
Support vectors are the data points, which are closest to the line
252
253
253
254

254
255
255
256
## Support vector classifier
256
257
257
258
Support vector machines (svm) is a set of supervised learning methods in the Scikit Learn library.
258
-
Support vector classifier (SVC) is a python class capable of performing classification on a dataset.
259
-
The class SVC is in the module svm of the Scikit Learn library
259
+
Support vector classifier (SVC) is a python class capable of performing classification on a dataset.
260
+
The class SVC is in the module svm of the Scikit Learn library
260
261
```
261
262
>>> from sklearn.svm import SVC
262
263
>>> clf = SVC(kernel='linear')
@@ -271,9 +272,10 @@ SVC with parameter kernel='linear' finds the linear separator that maximizes the
271
272
k-NN classification is used with a supervised learning set.
272
273
K is an integer.
273
274
To classify a new data point, this algorithm calculates the distance between the new data point and the other data points.
274
-
The distance can be Euclidean, Manhattan, ....
275
+
The distance can be Euclidean, Manhattan, ....
275
276
Once the algorithm knows the K closest neighbors of the new data point, it takes the most common class of these K closest neighbors, and assign that most common class to the new data point.
276
-
So the new data point is assigned to the most common class of its k nearest neighbors. the new data point is assigned to the class to which the majority of its K nearest neighbors belong to.
277
+
So the new data point is assigned to the most common class of its k nearest neighbors.
278
+
So the new data point is assigned to the class to which the majority of its K nearest neighbors belong to.
277
279
278
280

279
281
@@ -313,7 +315,7 @@ k ≤ n.
313
315
314
316
A cluster is a group of data points.
315
317
Each cluster has a center, called the centroid.
316
-
A cluster centroid is the mean of a cluster (average across all the data points in the cluster).
318
+
A cluster centroid is the mean of a cluster (average across all the data points in the cluster).
317
319
The radius of a cluster is the maximum distance between all the points and the centroid.
0 commit comments