The term Support Vector Machines or SVM is a bit misleading. It is just a name for a very clever algorithm invented by two Russians. in the 1960s. SVMs are used for classification and regression.
SVM do that by finding a hyperplane between two classes of data which separates both classes best.
print("Start training")
t0 = time()
clf = svm.SVC(kernel="linear")
clf.fit(features_train, labels_train)
print("training time:", round(time() - t0, 3), "s")
print("start predicting")
t0 = time()
prediction = clf.predict(features_test)
print("predict time:", round(time() - t0, 3), "s")
# accuracy
print("Calculating accuracy")
accuracy = accuracy_score(labels_test, prediction)
print("Accuracy calculated, and the accuracy is", accuracy)
When timing the training of the SVC, it’s astonishing how long it takes: around 2.5 minutes at 98.4% accuracy.
As an alternative You can use:
clf = LinearSVC(loss='hinge')
It gets you a result in 0.3 seconds with the same accuracy.
What’s the difference?
Parameter tuning
with the initial SVC we can play around with the parameters “C” and “kernel”
C
Penalty parameter C of the error term
Kernels and the Kernel trick
RBF
Linear Digression Podcast on Kernel Trick





