SVM (Support Vector Machine)

Support Vector Machine
Type Supervised learning algorithm
Introduced 1992 (Boser, Guyon, Vapnik)
Purpose Classification, Regression
Key Feature Maximum margin hyperplane
AM Uses Defect classification, quality prediction

Support Vector Machine (SVM) is a powerful supervised learning algorithm that finds the optimal hyperplane to separate different classes of data. SVMs are particularly effective in high-dimensional spaces and when the number of dimensions exceeds the number of samples.

In additive manufacturing, SVMs are used for defect classification, process monitoring, and quality prediction. They excel when the decision boundary between good and defective parts is complex but can be captured with the right kernel function.

Contents
  1. Core Concept
  2. Maximum Margin
  3. Kernel Trick
  4. Support Vector Regression
  5. Applications in AM
  6. References

Core Concept

SVM works by finding a hyperplane (a line in 2D, a plane in 3D, or a hyperplane in higher dimensions) that best separates data points of different classes. The "best" hyperplane is the one with the maximum margin—the largest distance to the nearest data points from each class.

    Class A (○)                    Class B (●)
         ○                              ●
           ○    ○                  ●  ●
              ○ ○ ○            ●  ●
                 ○|          |●
                  |  margin  |
    ─────────────────────────────────────────
                  |          |
                  ◀──────────▶
              Support Vectors
              (points on margin)
            
Support vectors: Only the data points closest to the hyperplane (the "support vectors") determine its position. This makes SVMs memory-efficient and robust to outliers far from the boundary.

Maximum Margin

The margin is the distance between the hyperplane and the nearest data points. SVM maximizes this margin, which provides:

Soft Margin SVM

When data isn't perfectly separable, soft margin SVM allows some misclassifications. The regularization parameter C controls the trade-off:

Kernel Trick

When data isn't linearly separable, the kernel trick maps data to a higher-dimensional space where a linear separator exists—without explicitly computing the transformation.

Linear Kernel

K(x, y) = x · y

Best for linearly separable data or high-dimensional sparse data (like text).

RBF (Radial Basis Function) Kernel

K(x, y) = exp(-γ||x - y||²)

Most popular kernel. Maps to infinite-dimensional space. Parameter γ controls the "reach" of each training point.

Polynomial Kernel

K(x, y) = (γ·x·y + r)^d

Captures feature interactions up to degree d. Good for problems with known polynomial relationships.

Example: Defect Detection
With temperature and pressure as inputs, defects may not be linearly separable. An RBF kernel can find a circular or irregular boundary in the original 2D space by mapping to higher dimensions.

Support Vector Regression (SVR)

SVR adapts the SVM concept for regression. Instead of maximizing the margin between classes, SVR fits a tube of width ε around the data:

Parameter Effect Tuning Tip
C Regularization strength Grid search over [0.1, 1, 10, 100]
γ (RBF) Kernel width Grid search; often 1/n_features
ε (SVR) Tube width for regression Depends on noise level in data

Applications in Additive Manufacturing

Defect Classification:
SVMs classify parts as defective or acceptable based on process parameters. Hassan et al. (2024) notes SVM as a baseline classifier for quality control, though often outperformed by ensemble methods on complex datasets.

Common Use Cases

Advantages for AM

Limitations

See Also

References