Properties SVM SMM
Similarities
Plane-based learning Principle Decision function Hyperplane separate the data points Large margin principle Sign function Hyperplane separate the data points Large margin principle Sign function
Differences
Input data Vector Matrices
Behavior with input matrix data Vectorize matrix which does not preserve the spatial correlations inherent in them Preserves the inherent spatial correlations
Optimization problem No spectral elastic net property Employs spectral elastic net penalty property by using Frobenius norm and nuclear norm
Time complexity O(N^(3)) (Kumari, Ganaie, & Tanveer, 2022) O( poly (N,pq)) (Duan, Yuan, Liu, & Li, 2017)| Properties | SVM | SMM |
| :--- | :--- | :--- |
| Similarities | | |
| Plane-based learning Principle Decision function | Hyperplane separate the data points Large margin principle Sign function | Hyperplane separate the data points Large margin principle Sign function |
| Differences | | |
| Input data | Vector | Matrices |
| Behavior with input matrix data | Vectorize matrix which does not preserve the spatial correlations inherent in them | Preserves the inherent spatial correlations |
| Optimization problem | No spectral elastic net property | Employs spectral elastic net penalty property by using Frobenius norm and nuclear norm |
| Time complexity | $O\left(N^{3}\right)$ (Kumari, Ganaie, & Tanveer, 2022) | $O($ poly $(N, p q))$ (Duan, Yuan, Liu, & Li, 2017) |
图1. 展示了论文结构和进展的视觉说明。
表2
论文中使用的符号。
符号
描述
C
多类问题中的类别数量
NN
训练样本数量
p xx qp \times q
输入矩阵数据的顺序
kappa\kappa
矩阵求逆的条件数
epsilon\epsilon
输出空间的预期精度
H
包含输入矩阵的希尔伯特空间
V
对称矩阵
zeta\zeta
权衡参数/正则化参数
lambda\lambda
核范数约束
rho\rho
ADMM 的固有系数
ss
迭代次数
L
深度变体中的层数
KK
矩阵的秩
Symbol Description
C Number of classes in multi-class problems
N Number of training samples
p xx q Order of input matrices data
kappa Condition number of the matrix for inversion
epsilon Expected accuracy of the output space
H Hilbert space containing input matrices
V Symmetric matrix
zeta Trade-off parameter/regularization parameter
lambda Nuclear norm constraint
rho Inherent coefficient of ADMM
s Number of iterations
L Number of layers in deep variants
K Rank of the matrix| Symbol | Description |
| :--- | :--- |
| C | Number of classes in multi-class problems |
| $N$ | Number of training samples |
| $p \times q$ | Order of input matrices data |
| $\kappa$ | Condition number of the matrix for inversion |
| $\epsilon$ | Expected accuracy of the output space |
| H | Hilbert space containing input matrices |
| V | Symmetric matrix |
| $\zeta$ | Trade-off parameter/regularization parameter |
| $\lambda$ | Nuclear norm constraint |
| $\rho$ | Inherent coefficient of ADMM |
| $s$ | Number of iterations |
| L | Number of layers in deep variants |
| $K$ | Rank of the matrix |
ILS-TSMM (Gaoa, Fanb, & Xub, 2015) BP-LSSMM (Xia & Fan, 2016) NPLSSMM (Li, Yang, Pan, Cheng, & Cheng, 2020) LSISMM (Li, Shao, Lu, Xiang and Cai, 2022) AMK-TMM (Liang, Hang, Lei et al., 2022)
SRM principle ✓ xx xx ✓ xx
Hyperplanes Two non-parallel hyperplanes A decision boundary Two non-parallel hyperplanes Two non-parallel hyperplanes A decision boundary
Transfer learning x x xx x ✓
Optimization problems Two One Two Two One| | ILS-TSMM (Gaoa, Fanb, & Xub, 2015) | BP-LSSMM (Xia & Fan, 2016) | NPLSSMM (Li, Yang, Pan, Cheng, & Cheng, 2020) | LSISMM (Li, Shao, Lu, Xiang and Cai, 2022) | AMK-TMM (Liang, Hang, Lei et al., 2022) |
| :--- | :--- | :--- | :--- | :--- | :--- |
| SRM principle | $\checkmark$ | $\times$ | $\times$ | $\checkmark$ | $\times$ |
| Hyperplanes | Two non-parallel hyperplanes | A decision boundary | Two non-parallel hyperplanes | Two non-parallel hyperplanes | A decision boundary |
| Transfer learning | $x$ | $x$ | $\times$ | $x$ | $\checkmark$ |
| Optimization problems | Two | One | Two | Two | One |
表4
SMM 的最小二乘变体。
模型
作者
特征
损失函数
数据集
优点
解决方法
ILS-TSMM (2015)
Gaoa 等人 (2015)
-
最小二乘损失
ORL 和 YALE 人脸数据库
考虑 SRM 原理。与 SMM 相比,在时间效率方面有所提高(Luo 等人,2015 年)。
求解方程组。
BP-LSSMM (2016)
夏和范(2016 年)
将最小二乘法引入双层规划 SMM 求解。
-
PALM400、ORL、Yale 数据集
比 SMM 更高效
采用增广拉格朗日乘子法,并求解线性方程组。
NPLSSMM (2020)
Li 等人 (2020)
Solves matrix classification problem by constructing two non-parallel hyperplanes.
Least square loss
Fault dataset from CWRU
Distinguishes the classes by obtaining a maximum margin hyperplane in matrix form, reduced complexity as it solves a system of linear equations.
ADMM
LSISMM (2022)
Li, Shao 等人 (2022)
Constructs non-parallel hyperplanes and uses small infrared thermal images for fault diagnosis.
Least square loss
-
Flexible to maximize the distance between non-parallel hyperplanes and high computational efficiency than SMM.
ADMM
AMK-TMM (2022)
梁航、雷等 (2022)
Introduces a novel adaptive multimodel knowledge transfer framework and consist of equality constraints.
Least square loss
BCI III 的 Dataset IVa 和 BCI IV 的 IIa
Utilize CV using a leave-one-out strategy to automatically find the correlated source domains and their corresponding weights.
ADMM
Model Author Characteristics Loss function Datasets Advantages Technique to solve
ILS-TSMM (2015) Gaoa et al. (2015) - Least square loss Face databases ORL and YALE Considers SRM principle. Improved efficiency in terms of time than SMM (Luo et al., 2015). Solves system of equations.
BP-LSSMM (2016) Xia and Fan (2016) Introduces least square method to solve bilevel programming SMM. - PALM400, ORL, Yale datasets Efficient than SMM Augmented Lagrange multiplier method is used and system of linear equations is solved.
NPLSSMM (2020) Li et al. (2020) Solves matrix classification problem by constructing two non-parallel hyperplanes. Least square loss Fault dataset from CWRU Distinguishes the classes by obtaining a maximum margin hyperplane in matrix form, reduced complexity as it solves a system of linear equations. ADMM
LSISMM (2022) Li, Shao et al. (2022) Constructs non-parallel hyperplanes and uses small infrared thermal images for fault diagnosis. Least square loss - Flexible to maximize the distance between non-parallel hyperplanes and high computational efficiency than SMM. ADMM
AMK-TMM (2022) Liang, Hang, Lei et al. (2022) Introduces a novel adaptive multimodel knowledge transfer framework and consist of equality constraints. Least square loss Dataset IVa of BCI III and IIa of BCI IV Utilize CV using a leave-one-out strategy to automatically find the correlated source domains and their corresponding weights. ADMM| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| ILS-TSMM (2015) | Gaoa et al. (2015) | - | Least square loss | Face databases ORL and YALE | Considers SRM principle. Improved efficiency in terms of time than SMM (Luo et al., 2015). | Solves system of equations. |
| BP-LSSMM (2016) | Xia and Fan (2016) | Introduces least square method to solve bilevel programming SMM. | - | PALM400, ORL, Yale datasets | Efficient than SMM | Augmented Lagrange multiplier method is used and system of linear equations is solved. |
| NPLSSMM (2020) | Li et al. (2020) | Solves matrix classification problem by constructing two non-parallel hyperplanes. | Least square loss | Fault dataset from CWRU | Distinguishes the classes by obtaining a maximum margin hyperplane in matrix form, reduced complexity as it solves a system of linear equations. | ADMM |
| LSISMM (2022) | Li, Shao et al. (2022) | Constructs non-parallel hyperplanes and uses small infrared thermal images for fault diagnosis. | Least square loss | - | Flexible to maximize the distance between non-parallel hyperplanes and high computational efficiency than SMM. | ADMM |
| AMK-TMM (2022) | Liang, Hang, Lei et al. (2022) | Introduces a novel adaptive multimodel knowledge transfer framework and consist of equality constraints. | Least square loss | Dataset IVa of BCI III and IIa of BCI IV | Utilize CV using a leave-one-out strategy to automatically find the correlated source domains and their corresponding weights. | ADMM |
considerations, Eq. (1) is identical to the subsequent formulation for performing matrix classification directly: min_(W,b)(1)/(2)tr(W^(T)W)+zetasum_(i=1)^(N){1-y_(i)[tr(W^(T)X_(i))+b]}_(+)\min _{\mathbf{W}, b} \frac{1}{2} \operatorname{tr}\left(\mathbf{W}^{T} \mathbf{W}\right)+\zeta \sum_{i=1}^{N}\left\{1-y_{i}\left[\operatorname{tr}\left(\mathbf{W}^{T} \mathbf{X}_{i}\right)+b\right]\right\}_{+}.
This illustrates that directly employing Eq. (2) for classification is insufficient to capture the inherent structure present within each input matrix. As a consequence, this approach leads to a loss of information.
In order to consider the structural characteristics, an intuitive strategy involves capturing the correlations inherent within each input matrix by introducing a low-rank restriction on the matrix W\mathbf{W}. To tackle this issue, a number of approaches are suggested such as the low-rank SVM (Wolf et al., 2007) and the bi-linear SVM (Pirsiavash et al., 2009). However, these approaches have limitations as they demand manual pre-specification of the latent rank of W\mathbf{W} tailored to various applications. To address the aforementioned issues, Luo et al. (2015) proposed an advancement in the realm of supervised learning, SMM, which overcomes the pre-specified rank criteria, preserves the structural information of the input matrix by employing the spectral elastic net penalty for the regression matrix, and produces a better result for matrix-form data.
3.3. SMM 的数学公式
SMM is a proficient matrix-based adaptation of SVM, capitalizing on the strengths of SVM, which include robust generalization capabilities. Further, it has the ability to comprehensively harness the structural insights embedded within matrix data. The objective function of SMM,
Here, the initial term (1)/(2)tr(W^(T)W)+lambda||W||_(**)\frac{1}{2} \operatorname{tr}\left(\mathbf{W}^{T} \mathbf{W}\right)+\lambda\|\mathbf{W}\|_{*} pertains to the utilization of spectral elastic net regularization, which serves the purpose of capturing correlations inherent within individual matrices. On the other hand, the last summation term represents the hinge loss function. The term (1)/(2)tr(W^(T)W)\frac{1}{2} \operatorname{tr}\left(\mathbf{W}^{T} \mathbf{W}\right) can also be written as (1)/(2)||W||_(F)^(2)\frac{1}{2}\|\mathbf{W}\|_{F}^{2}, which represents the square Frobenius norm.
The Frobenius norm of matrix W\mathbf{W} serves as a regularization term, aiming to find a weight matrix with a reduced rank. It is also crucial to highlight that the nuclear norm serves as a regularization factor to ascertain the rank of matrix W\mathbf{W}. Estimating the rank of a matrix can be a complex problem with NP-hard characteristics (Wang, Wang, Hu, & Yan, 2015), however, the nuclear norm is widely acknowledged as the optimal convex approximation method for assessing the rank of the matrix (Candes & Recht, 2012; Zhou & Li, 2014). Additionally, the low-rank parameter lambda\lambda governs the level of structure information incorporated for constructing the classification hyperplane. The presence of the term ||W||_(**)\|\mathbf{W}\|_{*} introduces non-smoothness to the objective function of SMM. This characteristic poses a challenge when attempting to directly solve Eq. (3). Consequently, the solution for SMM is derived through the application of the alternating direction method of multipliers (ADMM) (Goldstein, O’Donoghue, Setzer, & Baraniuk, 2014). Now, by introducing an auxiliary matrix variable Q\mathbf{Q}, the objective function of SMM can be reformulated in the following manner:
arg min F(W,b)+G(Q)\arg \min F(\mathbf{W}, b)+G(\mathbf{Q})
(W,b),Q(\mathbf{W}, b), \mathbf{Q}
Table 5
Robust and sparse models of SMM.
模型
作者
特征
损失函数
数据集
优点
解决方法
RSMM(2018)
郑、朱和恒 (2018)
Decompose each input signal into low-rank clean signal and sparse intra-sample outliers and employ l_(1)l_{1} norm for sparseness.
铰链损失
IVa of brain computer interface (BCI) competition III (Dornhege, Blankertz, Curio, & Muller, 2004), IIb and IIa of BCI competition IV (Leeb et al., 2007)
增强了 SMM 的鲁棒性。
ADMM
SSMM (2018)
Zheng, Zhu, Qin, Chen 和 Heng (2018)
Performs feature selection to remove redundant features and involves a new regularization term, which is a linear combination of nuclear norm and l_(1)l_{1} norm.
铰链损失
INRIA Person Dataset (Dalal & Triggs, 2005), Caltech Face Dataset (Fergus, Perona, & Zisserman, 2003), IIa and IIb of BCI competition IV (Ang, Chin, Wang, Guan, & Zhang, 2012)
Eliminates the effect of noise in raw signal and enhance the fault features.
GFB 算法
TRMM (2022)
Pan, Xu, Zheng, Tong 和 Cheng (2022)
Employs truncated nuclear norm for low-rank approximation.
斜坡损失
Fault dataset of roller bearing from AHUT
Insensitive and robust to outliers and efficient than RSMM.
加速近端梯度(ALG)算法
Pin-SMM (2022)
冯和徐(2022)
最大化分位数距离,而不是最短距离。
保形损失
INRIA 人物(Dalal & Triggs,2005),加州理工学院(Fei-Fei,Fergus,& Perona,2006),BCI IV 的 IIa
Robust to noise.
ADMM
SNMM (2022)
Wang, Xu, Pan, Xie, 和 Zheng (2022)
Employed the l_(1)l_{1} norm distance as a constraint of the hyperplane and avoided the need to find inverse matrices.
铰链损失
Fault dataset of roller bearing from AHUT, HNU, and CWRU
Improves the robustness and reduces the storage requirements.
一种交替迭代方法
ACF-SSMM (2022)
Li, Wang and Liu (2022)
Extend the input matrix by adding data through an auto-correlation function (ACF) transform, which contain data information at previous/current instants.
铰链损失
SEED-VIG 疲劳数据集(Zheng & Lu,2017)
Enhances the generalization performance of SMM.
GFB 算法
SMMRe(2023)
Razzak、Bouadjenek、Saris 和 Ding(2023)
Decompose each input signal into low-rank clean signal and sparse intra-sample outliers and employ joint l_(2,1)l_{2,1} and nuclear norm.
铰链损失
Caltech(Fei-Fei 等人,2006 年)、INRIA Person(Dalal 和 Triggs,2005 年)、BCI III 的 IVa、BCI IV 的 IIb 和 IIa
增强了 SMM 的鲁棒性。
ADMM
TPin-SMM (2024)
Li 和 Xu (2024)
引入了截断弹球损失,增强了对外离群值、噪声不敏感性和稀疏性的鲁棒性。
截断的弹珠损失
加州理工学院(Fei-Fei 等人,2006 年),BCI IV 的 IIa 部分,戴姆勒行人数据集
提高了 SMM 的泛化性能。
CCCP-ADMM
Model Author Characteristics Loss function Datasets Advantages Technique to solve
RSMM (2018) Zheng, Zhu and Heng (2018) Decompose each input signal into low-rank clean signal and sparse intra-sample outliers and employ l_(1) norm for sparseness. Hinge loss IVa of brain computer interface (BCI) competition III (Dornhege, Blankertz, Curio, & Muller, 2004), IIb and IIa of BCI competition IV (Leeb et al., 2007) Enhances the robustness of SMM. ADMM
SSMM (2018) Zheng, Zhu, Qin, Chen and Heng (2018) Performs feature selection to remove redundant features and involves a new regularization term, which is a linear combination of nuclear norm and l_(1) norm. Hinge loss INRIA Person Dataset (Dalal & Triggs, 2005), Caltech Face Dataset (Fergus, Perona, & Zisserman, 2003), IIa and IIb of BCI competition IV (Ang, Chin, Wang, Guan, & Zhang, 2012) Enhances the sparseness of SMM. Generalized forwardbackwards (GFB) algorithm
RMSMM (2019) Qian, Tran-Dinh, Fu, Zou, and Liu (2019) Constructed in the angle based classification framework and condenses the binary and multi-class problems into a single framework. Truncated hinge loss Daily and sports activities dataset (Altun & Barshan, 2010) Enjoys better prediction performance and faster computation than SMM. Inexact proximal DC algorithm
RSSMM (2021) Gu, Zheng, Pan, and Tong (2021) Employs l_(1) norm and sparse constraint into objective function to weaken the redundant information of the input matrix. Smooth ramp loss Fault dataset of roller bearing from AHUT Reduces the influence of outliers. GFB algorithm
SWSSMM (2021) Li, Yang et al. (2021) Automatically extract inherent fault features from raw signals and use the symplectic coefficient matrix (SCM). Also a variable entropy-based weight coefficient is added into SCM to enhance the fault features. Hinge loss Vibration signal dataset from University of Connecticut Eliminates the effect of noise in raw signal and enhance the fault features. GFB algorithm
TRMM (2022) Pan, Xu, Zheng, Tong and Cheng (2022) Employs truncated nuclear norm for low-rank approximation. Ramp loss Fault dataset of roller bearing from AHUT Insensitive and robust to outliers and efficient than RSMM. Accelerated Proximal Gradient (ALG) algorithm
Pin-SMM (2022) Feng and Xu (2022) Maximizes the quantile distance rather than the shortest distance. Pinball loss INRIA Person (Dalal & Triggs, 2005), Caltech (Fei-Fei, Fergus, & Perona, 2006), IIa of BCI IV Robust to noise. ADMM
SNMM (2022) Wang, Xu, Pan, Xie, and Zheng (2022) Employed the l_(1) norm distance as a constraint of the hyperplane and avoided the need to find inverse matrices. Hinge loss Fault dataset of roller bearing from AHUT, HNU, and CWRU Improves the robustness and reduces the storage requirements. An alternating iteration method
ACF-SSMM (2022) Li, Wang and Liu (2022) Extend the input matrix by adding data through an auto-correlation function (ACF) transform, which contain data information at previous/current instants. Hinge loss SEED-VIG fatigue dataset (Zheng & Lu, 2017) Enhances the generalization performance of SMM. GFB algorithm
SMMRe (2023) Razzak, Bouadjenek, Saris, and Ding (2023) Decompose each input signal into low-rank clean signal and sparse intra-sample outliers and employ joint l_(2,1) and nuclear norm. Hinge loss Caltech (Fei-Fei et al., 2006), INRIA Person (Dalal & Triggs, 2005), IVa of BCI III, IIb and IIa of BCI IV Enhances the robustness of SMM. ADMM
TPin-SMM (2024) Li and Xu (2024) Incorporated truncated pinball loss and yields robustness to outliers, noise insensitivity, and sparsity. Truncated pinball loss Caltech (Fei-Fei et al., 2006), IIa of BCI IV, Daimler pedestrian dataset Enhances the generalization performance of SMM. CCCP-ADMM| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| RSMM (2018) | Zheng, Zhu and Heng (2018) | Decompose each input signal into low-rank clean signal and sparse intra-sample outliers and employ $l_{1}$ norm for sparseness. | Hinge loss | IVa of brain computer interface (BCI) competition III (Dornhege, Blankertz, Curio, & Muller, 2004), IIb and IIa of BCI competition IV (Leeb et al., 2007) | Enhances the robustness of SMM. | ADMM |
| SSMM (2018) | Zheng, Zhu, Qin, Chen and Heng (2018) | Performs feature selection to remove redundant features and involves a new regularization term, which is a linear combination of nuclear norm and $l_{1}$ norm. | Hinge loss | INRIA Person Dataset (Dalal & Triggs, 2005), Caltech Face Dataset (Fergus, Perona, & Zisserman, 2003), IIa and IIb of BCI competition IV (Ang, Chin, Wang, Guan, & Zhang, 2012) | Enhances the sparseness of SMM. | Generalized forwardbackwards (GFB) algorithm |
| RMSMM (2019) | Qian, Tran-Dinh, Fu, Zou, and Liu (2019) | Constructed in the angle based classification framework and condenses the binary and multi-class problems into a single framework. | Truncated hinge loss | Daily and sports activities dataset (Altun & Barshan, 2010) | Enjoys better prediction performance and faster computation than SMM. | Inexact proximal DC algorithm |
| RSSMM (2021) | Gu, Zheng, Pan, and Tong (2021) | Employs $l_{1}$ norm and sparse constraint into objective function to weaken the redundant information of the input matrix. | Smooth ramp loss | Fault dataset of roller bearing from AHUT | Reduces the influence of outliers. | GFB algorithm |
| SWSSMM (2021) | Li, Yang et al. (2021) | Automatically extract inherent fault features from raw signals and use the symplectic coefficient matrix (SCM). Also a variable entropy-based weight coefficient is added into SCM to enhance the fault features. | Hinge loss | Vibration signal dataset from University of Connecticut | Eliminates the effect of noise in raw signal and enhance the fault features. | GFB algorithm |
| TRMM (2022) | Pan, Xu, Zheng, Tong and Cheng (2022) | Employs truncated nuclear norm for low-rank approximation. | Ramp loss | Fault dataset of roller bearing from AHUT | Insensitive and robust to outliers and efficient than RSMM. | Accelerated Proximal Gradient (ALG) algorithm |
| Pin-SMM (2022) | Feng and Xu (2022) | Maximizes the quantile distance rather than the shortest distance. | Pinball loss | INRIA Person (Dalal & Triggs, 2005), Caltech (Fei-Fei, Fergus, & Perona, 2006), IIa of BCI IV | Robust to noise. | ADMM |
| SNMM (2022) | Wang, Xu, Pan, Xie, and Zheng (2022) | Employed the $l_{1}$ norm distance as a constraint of the hyperplane and avoided the need to find inverse matrices. | Hinge loss | Fault dataset of roller bearing from AHUT, HNU, and CWRU | Improves the robustness and reduces the storage requirements. | An alternating iteration method |
| ACF-SSMM (2022) | Li, Wang and Liu (2022) | Extend the input matrix by adding data through an auto-correlation function (ACF) transform, which contain data information at previous/current instants. | Hinge loss | SEED-VIG fatigue dataset (Zheng & Lu, 2017) | Enhances the generalization performance of SMM. | GFB algorithm |
| SMMRe (2023) | Razzak, Bouadjenek, Saris, and Ding (2023) | Decompose each input signal into low-rank clean signal and sparse intra-sample outliers and employ joint $l_{2,1}$ and nuclear norm. | Hinge loss | Caltech (Fei-Fei et al., 2006), INRIA Person (Dalal & Triggs, 2005), IVa of BCI III, IIb and IIa of BCI IV | Enhances the robustness of SMM. | ADMM |
| TPin-SMM (2024) | Li and Xu (2024) | Incorporated truncated pinball loss and yields robustness to outliers, noise insensitivity, and sparsity. | Truncated pinball loss | Caltech (Fei-Fei et al., 2006), IIa of BCI IV, Daimler pedestrian dataset | Enhances the generalization performance of SMM. | CCCP-ADMM |
此处, rho\rho 表示 ADMM 方法的固有系数。在此框架内,我们需要确定三个变量矩阵: Q,beta\mathbf{Q}, \boldsymbol{\beta} 和 (W,b)(\mathbf{W}, b) 。获取这些矩阵的最优解涉及迭代方法。其一般步骤为
updating these variable matrices are outlined as follows:
The fundamental steps for solving within this context involve calculating Q^((t))\mathbf{Q}^{(t)} and the pair ( W^((t)),b^((t))\mathbf{W}^{(t)}, b^{(t)} ) during each iteration. To update Q\mathbf{Q} (for the sake of ease, we omit the superscripts in the subsequent explanation), suppose that beta\boldsymbol{\beta} and ( W,b\mathbf{W}, b ) remain constant, then the
solution of Q\mathbf{Q} can be determined using the following equation: arg min_(Q)G(Q)+(rho)/(2)||Q-W||_(F)^(2)+tr[beta^(T)(Q-W)]\underset{\mathbf{Q}}{\arg \min } G(\mathbf{Q})+\frac{\rho}{2}\|\mathbf{Q}-\mathbf{W}\|_{F}^{2}+\operatorname{tr}\left[\boldsymbol{\beta}^{T}(\mathbf{Q}-\mathbf{W})\right].
By solving Eq. (5), we can derive the updating formula for the matrix Q\mathbf{Q} during each iteration as follows: Q=(1)/(rho)S_(lambda)(rhoW-beta)\mathbf{Q}=\frac{1}{\rho} S_{\lambda}(\rho \mathbf{W}-\boldsymbol{\beta}),
因此,我们基于最小二乘法的概念分析了各种现有的 SMM 模型,这对应于模型训练时间的减少。此外,与经典 SMM 相比,它简化了模型。表 3 概述了 SMM 领域中各种最小二乘变体。这些变体侧重于求解线性方程组,而不是 QPP,这一策略选择显著提高了计算效率,尤其是在处理大规模数据集时。在这些方法中,LSISMM(Li, Shao 等人,2022)表现突出,具有时间复杂度 O(s(min(m^(2)n,mn^(2))+Nmn))O\left(s\left(\min \left(m^{2} n, m n^{2}\right)+N m n\right)\right) 。此外,我们在表 4 中提供了不同最小二乘变体的详细信息。此外,我们将深入探讨 SMM 框架内不同鲁棒和稀疏模型的相关讨论。
4.2. 鲁棒和稀疏 SMM
SMM outperforms SVM in terms of performance due to the preservation of the structural information in the input matrix. However, it considers the hinge loss function and l_(2)l_{2} norm in the objective function which reduces the robustness (Wu & Liu, 2007) and sparsity (Tanveer, Sharma, Rastogi, & Anand, 2021), respectively. Moreover, the input data often contains distortions from measurement artifacts, outliers, and unconventional sources of noise. Consequently, the obtained classifier may have poor performance. Thus, in order to tackle the aforementioned challenges, various variants of SMM have been proposed in the literature. To counter intra-sample outliers, Zheng, Zhu, Heng (2018) proposed a robust SMM (RSMM). It decomposed the input matrix into a latent low-rank clean matrix plus a sparse noise matrix and used only the clean matrix for training, which makes it robust to intra-sample outliers. Also, to enhance the sparseness, it employs the l_(1)l_{1} norm instead of the l_(2)l_{2} norm. The l_(1)l_{1} norm optimization problems tend to drive some of the coefficients to exactly zero and encourage sparse solutions (Tanveer et al., 2021). The time complexity of RSMM is O(N^(2)pq)xx sO\left(N^{2} p q\right) \times s. Following the same concept, Razzak et al. (2023) proposed SMM that simultaneously performs matrix Recovery (SMMRe), a variant of SMM via joint l_(2,1)l_{2,1} and nuclear norm minimization. The objective function of SMMRe combines the property of matrix recovery along with low rank and joint sparsity to deal with complex high-dimensional noisy data.
Other approach to build a robust model is by incorporating a robust classification loss function. In light of this, Qian et al. (2019) proposed robust multicategory SMM (RMSMM) which makes SMM robust by using the truncated hinge loss function (Wu & Liu, 2007) rather than hinge loss. The hinge loss is unbounded and can grow indefinitely for outliers away from the optimal hyperplane. In contrast, the truncated hinge loss limits the impact of such outliers by capping the loss at a predefined value. As a result, the truncated hinge loss is resistant to outliers. In a similar way, Gu et al. (2021) proposed ramp sparse SMM model (RSSMM) to improve the robustness of SMM. RSSMM uses
smooth ramp loss function instead of hinge loss, which also limits the maximum loss and weakens the sensitivity to outliers.
Here, we delved into a comprehensive analysis of robust and sparse variants within the realm of SMM models. By examining these nuanced approaches, we gained valuable insights into enhancing model robustness and promoting sparsity. To enhance robustness against outliers or noise, two primary strategies are considered in the literature. First, decompose the input matrix into clean and noise matrices and use only the clean matrix for training (Razzak et al., 2023; Zheng, Zhu, Heng,
2018). Second, to integrate robust loss functions, such as truncated hinge/pinball loss, to limit the impact of outliers and enhance noise resistance (Gu et al., 2021; Qian et al., 2019). On the other hand, to enhance the sparsity, a key approach involves introducing regularization terms that combine the l_(1)l_{1} norm (Wang et al., 2022; Zheng, Zhu, Qin, Chen et al., 2018). Further, introducing novel regularization terms that encourage sparse solutions by driving some coefficients to zero may be incorporated to enhance sparsity. Table 5 presents an overview of the various robust and sparse variants of SMM.
4.3. SMM for multi-class classification
在原始公式中,SMM 是为二元分类问题设计的;然而,现实世界中的大多数问题基于多类分类(Franc & Hlavác,2002)。为了应对这一挑战,Zheng、Zhu、Qin 和 Heng(2018)开发了一个名为多类 SMM(MSMM)的模型。这种方法结合了多类 hinge 损失函数以及结合平方 Frobenius 范数和核范数的正则化项。多类 hinge 损失函数的公式扩展了 margin rescaling 损失(Joachims、Finley & Yu,2009)的概念,以适应矩阵形式数据。为了提高 MSMM 的分类性能,Razzak、Blumenstein 和 Xu(2019)提出了一种新的多类 SMM(M-SMM)。它由二元 hinge 损失和弹性网络惩罚融合而成。二元 hinge 损失采用 CC 函数来模拟多个二元分类器,从而避免计算每对可能类别之间的支持向量。然而,MSMM 和 M-SMM 对异常值不鲁棒。为了为多类问题开发鲁棒的 SMM 变体,Qian 等人(2019)提出了鲁棒多类 SMM(RMSMM)。 它基于基于角度的分类框架(Zhang & Liu,2014)构建,并嵌入了一个截断的铰链损失函数(Wu & Liu,2007)。多类分类的常用方法是用 CC 分类函数来表示 CC 类别。然而,基于角度的分类框架需要训练 C-1C-1 分类函数,因此它遵循更快的计算速度(Sun, Craig, & Zhang, 2017)。RMSMM 的非凸优化问题通过 DCA 算法解决。
Model Author Characteristics Loss function Datasets Advantages Technique to solve
MSMM (2018) Zheng, Zhu, Qin, Heng (2018) Incorporates a multi-class hinge loss term and a regularization term involving Frobenius and nuclear norm. Multiclass hinge loss Dataset IIIa of BCI III and IIa of BCI IV Improves the performance of BCI system with multiple tasks. ADMM
M-SMM (2019) Razzak, Blumenstein et al. (2019) Maximizes the intra class margin and employs C functions to simulate all binary classifier rather than computing support vector between every two class. Binary hinge loss Dataset IIIa of BCI III and IIa of BCI IV Improves the classification performance for multi-class problems. ADMM
RMSMM (2019) Qian et al. (2019) Utilizes an angle-based classification framework, condensing both binary and multi-class problems into a single framework. Truncated hinge loss Daily and sports activities dataset (Altun & Barshan, 2010) Enjoys better prediction performance and faster computation. Inexact proximal DC algorithm
MSMM-CE (2020) Razzak (2020) Solves multi-class classification problems, and reduces data redundancy. Hinge loss BCI competitions benchmark EEG datasets (IIIa and IIa) Solves multi-class classification problems by finding support vectors in a single step. Evolutionary technique
MFSMM (2022) Pan, Xu, Zheng, Su et al. (2022) Fuzzy attributes are introduced to assign different membership degrees to different samples. Hinge Loss Fault dataset of roller bearing from AHUT and fault dataset of roller bearing from Hunan University (HNU) Improve the performance of multi-class classification in the presence of noise. SOR| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| MSMM (2018) | Zheng, Zhu, Qin, Heng (2018) | Incorporates a multi-class hinge loss term and a regularization term involving Frobenius and nuclear norm. | Multiclass hinge loss | Dataset IIIa of BCI III and IIa of BCI IV | Improves the performance of BCI system with multiple tasks. | ADMM |
| M-SMM (2019) | Razzak, Blumenstein et al. (2019) | Maximizes the intra class margin and employs $C$ functions to simulate all binary classifier rather than computing support vector between every two class. | Binary hinge loss | Dataset IIIa of BCI III and IIa of BCI IV | Improves the classification performance for multi-class problems. | ADMM |
| RMSMM (2019) | Qian et al. (2019) | Utilizes an angle-based classification framework, condensing both binary and multi-class problems into a single framework. | Truncated hinge loss | Daily and sports activities dataset (Altun & Barshan, 2010) | Enjoys better prediction performance and faster computation. | Inexact proximal DC algorithm |
| MSMM-CE (2020) | Razzak (2020) | Solves multi-class classification problems, and reduces data redundancy. | Hinge loss | BCI competitions benchmark EEG datasets (IIIa and IIa) | Solves multi-class classification problems by finding support vectors in a single step. | Evolutionary technique |
| MFSMM (2022) | Pan, Xu, Zheng, Su et al. (2022) | Fuzzy attributes are introduced to assign different membership degrees to different samples. | Hinge Loss | Fault dataset of roller bearing from AHUT and fault dataset of roller bearing from Hunan University (HNU) | Improve the performance of multi-class classification in the presence of noise. | SOR |
Model Author Characteristics Loss function Datasets Advantages Technique to solve
ESMM (2017) Zhu (2017) Entropy-based fuzzy membership is employed. Hinge loss KEEL imbalance datasets (Derrac, Garcia, Sanchez, & Herrera, 2015) Better generalization performance on imbalance datasets. ADMM
CWSMM (2021) Li, Cheng, Shao, Liu and Cai (2021) Distinct penalty factors are used for various class samples, confidence weights are assigned based on prior knowledge, and D-S evidence theory based fusion CWSMM is proposed. Hinge loss A customized rotating machinery dataset produced by Spectra Quest, Inc., Richmond, USA Enhances the robustness to imbalanced data. ADMM
DPAMM (2022) Xu, Pan, Zheng, Liu, and Tong (2022) The adaptive low-rank regularizer is introduced to obtain low-rank information and adaptively chooses the singular values relevant to the matrix's highly significant correlation data. Hinge loss Dataset of belevel gear roller bearing fault simulation test rig, datasets collected from fixed-shaft roller bearing test rig, bearing dataset provided by CWRU Improves the performance on imbalanced datasets. SOR| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| ESMM (2017) | Zhu (2017) | Entropy-based fuzzy membership is employed. | Hinge loss | KEEL imbalance datasets (Derrac, Garcia, Sanchez, & Herrera, 2015) | Better generalization performance on imbalance datasets. | ADMM |
| CWSMM (2021) | Li, Cheng, Shao, Liu and Cai (2021) | Distinct penalty factors are used for various class samples, confidence weights are assigned based on prior knowledge, and D-S evidence theory based fusion CWSMM is proposed. | Hinge loss | A customized rotating machinery dataset produced by Spectra Quest, Inc., Richmond, USA | Enhances the robustness to imbalanced data. | ADMM |
| DPAMM (2022) | Xu, Pan, Zheng, Liu, and Tong (2022) | The adaptive low-rank regularizer is introduced to obtain low-rank information and adaptively chooses the singular values relevant to the matrix's highly significant correlation data. | Hinge loss | Dataset of belevel gear roller bearing fault simulation test rig, datasets collected from fixed-shaft roller bearing test rig, bearing dataset provided by CWRU | Improves the performance on imbalanced datasets. | SOR |
O(L(N^(3)+sN^(2)pq))O\left(L\left(N^{3}+s N^{2} p q\right)\right)
-
O(L(N^(3)+sN^(2)pq))O\left(L\left(N^{3}+s N^{2} p q\right)\right)
Properties DSSMM (Hang et al., 2020) DSFR (Liang, Hang, Yin et al., 2022) DST-LSSMM (Hang et al., 2023) DSPTMM (Pan, Sheng et al., 2023)
Parameter tuning Feed-forward instead of parameter fine-tuning through backpropagation. Parameter fine-tuning instead of backpropagation. Feed-forward instead of parameter fine-tuning through backpropagation. -
Input data at each layer Random projections of predictions of each layer modify the original feature and are passed to the next layer. Take raw data as input Original input data with randomly projected values of the output of previous layers are fed to the next layer. PTMs retrieve previous layer output, randomly project it, combine it with input, and pass to next layer.
Optimization problem Convex Convex Convex Convex
Loss function Hinge loss Hinge loss Square loss Pinball loss
Features considered Pre-extracted information CSP for high-level feature extraction Pre-extracted information Annotated samples of source domain and target domain
Computational complexity O(LsN^(2)pq) O(L(N^(3)+sN^(2)pq)) - O(L(N^(3)+sN^(2)pq))| Properties | DSSMM (Hang et al., 2020) | DSFR (Liang, Hang, Yin et al., 2022) | DST-LSSMM (Hang et al., 2023) | DSPTMM (Pan, Sheng et al., 2023) |
| :--- | :--- | :--- | :--- | :--- |
| Parameter tuning | Feed-forward instead of parameter fine-tuning through backpropagation. | Parameter fine-tuning instead of backpropagation. | Feed-forward instead of parameter fine-tuning through backpropagation. | - |
| Input data at each layer | Random projections of predictions of each layer modify the original feature and are passed to the next layer. | Take raw data as input | Original input data with randomly projected values of the output of previous layers are fed to the next layer. | PTMs retrieve previous layer output, randomly project it, combine it with input, and pass to next layer. |
| Optimization problem | Convex | Convex | Convex | Convex |
| Loss function | Hinge loss | Hinge loss | Square loss | Pinball loss |
| Features considered | Pre-extracted information | CSP for high-level feature extraction | Pre-extracted information | Annotated samples of source domain and target domain |
| Computational complexity | $O\left(L s N^{2} p q\right)$ | $O\left(L\left(N^{3}+s N^{2} p q\right)\right)$ | - | $O\left(L\left(N^{3}+s N^{2} p q\right)\right)$ |
表9
SMM 的深度变体。
模型
作者
特征
损失函数
数据集
优点
解决方法
DSSMM (2020)
韩等 (2020)
继承了强大的深度表征学习能力。
Hinge Loss
BCI 竞赛 III 数据集 IVa,BCI 竞赛 IV 数据集 IIb,BCI 竞赛 IV 数据集 IIa,下肢 MI-BCI 数据集
涉及高效的顺向传播而非参数微调与反向传播,导致凸优化问题。
ADMM
DSFR (2022)
Liang, Hang, Yin 等 (2022)
DSFR 的基础构建模块由特征解码模块组成,其中 CSP 作为特征提取器,SMM 作为分类器。
铰链损失
BCI 竞赛 III 的 Dataset IVa,BCI 竞赛 IV 的 Dataset IIb,BCI 竞赛的 Dataset IIa
Model Author Characteristics Loss function Datasets Advantages Technique to solve
DSSMM (2020) Hang et al. (2020) Inherits the powerful capability of deep representation learning. Hinge Loss BCI competition III Dataset IVa, BCI Competition IV Dataset IIb, BCI Competition IV Dataset IIa, Lower Limb MI-BCI Dataset Involves an efficient feed-forward rather than parameter fine-tuning with backpropagation, leads to a convex optimization problem. ADMM
DSFR (2022) Liang, Hang, Yin et al. (2022) Base building blocks of DSFR consist of feature decoding modules which have CSP as a feature extractor and SMM as a classifier. Hinge loss Dataset IVa of BCI competition III, Dataset IIb of BCI competition IV, Dataset IIa of BCI competition Directly accepts the raw EEG data as input and automatically learns the feature representations. To improve classification performance, FDM can collect structural data from the EEG feature matrix. ADMM
DST-LSSMM (2023) Hang et al. (2023) Deep stacked network uses LSSMM as the base building unit and the projection of the previous layer is used as the stacking element. Least square loss MI-based EEG competition datasets which includes dataset IIIa, dataset iva in BCI competition III, and a self-collected dataset which includes lower limb MI-based BCI dataset (Lei et al., 2019) Overcomes non-convexity and requires less training data in comparison to deep learning models. Multiple layers take the privilege of adaptive learning. Alternating iterative method
DSPTMM (2023) Pan, Sheng et al. (2023) Deep stacked network uses pinball transfer module as the base building units and random projections as the stacking element. Pinball loss Used in the roller bearing fault diagnosis Overcomes non-convexity, requires less training data in comparison to deep learning models. Utilizes adaptive learning, pinball loss leads to a robust model. ADMM| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| DSSMM (2020) | Hang et al. (2020) | Inherits the powerful capability of deep representation learning. | Hinge Loss | BCI competition III Dataset IVa, BCI Competition IV Dataset IIb, BCI Competition IV Dataset IIa, Lower Limb MI-BCI Dataset | Involves an efficient feed-forward rather than parameter fine-tuning with backpropagation, leads to a convex optimization problem. | ADMM |
| DSFR (2022) | Liang, Hang, Yin et al. (2022) | Base building blocks of DSFR consist of feature decoding modules which have CSP as a feature extractor and SMM as a classifier. | Hinge loss | Dataset IVa of BCI competition III, Dataset IIb of BCI competition IV, Dataset IIa of BCI competition | Directly accepts the raw EEG data as input and automatically learns the feature representations. To improve classification performance, FDM can collect structural data from the EEG feature matrix. | ADMM |
| DST-LSSMM (2023) | Hang et al. (2023) | Deep stacked network uses LSSMM as the base building unit and the projection of the previous layer is used as the stacking element. | Least square loss | MI-based EEG competition datasets which includes dataset IIIa, dataset iva in BCI competition III, and a self-collected dataset which includes lower limb MI-based BCI dataset (Lei et al., 2019) | Overcomes non-convexity and requires less training data in comparison to deep learning models. Multiple layers take the privilege of adaptive learning. | Alternating iterative method |
| DSPTMM (2023) | Pan, Sheng et al. (2023) | Deep stacked network uses pinball transfer module as the base building units and random projections as the stacking element. | Pinball loss | Used in the roller bearing fault diagnosis | Overcomes non-convexity, requires less training data in comparison to deep learning models. Utilizes adaptive learning, pinball loss leads to a robust model. | ADMM |
The aforementioned imbalance models utilize the nuclear norm to represent the low-rank characteristics within each matrix data. However, the approach of nuclear norm minimization might not be optimal (Liu, Zhou et al., 2019; Zhang, Lei, Pan and Pedrycz, 2021),
resulting in certain limitations in capturing weak correlation information. To address this concern, Xu et al. (2022) introduced dynamic penalty adaptive matrix machine (DPAMM). This approach is built upon the adaptive framework for minimizing low-rank approximations (Gao et al., 2016). The technique dynamically selects and retains singular values that pertain to highly significant correlation data within the input matrix. Moreover, within the loss function component, a dynamic penalty factor is incorporated, enabling the adjustment of the penalty severity for samples based on the degree of imbalance. The computational complexity of DPAMM is O(N^(2)pq)xx sO\left(N^{2} p q\right) \times s.
O(kappa^(3)epsilon^(-3)(log(Npq)))O\left(\kappa^{3} \epsilon^{-3}(\log (N p q))\right)+O(log(pq))+O(\log (p q))
MRMLTSMCM (姜 & 杨, 2018)
✓\checkmark
Two non-parallel hyperplanes.
铰链损失
五折交叉验证
线性
-
KSMM(Ye,2019)
✓\checkmark
一
铰链损失
CV
非线性
O(N^(2)smn^(2))O\left(N^{2} s m n^{2}\right)
MDSMM(Ye & Han,2019)
✓\checkmark
一
铰链损失
十折交叉验证
线性情况
O(N^(2)s(pq+p+q))O\left(N^{2} s(p q+p+q)\right)
WSMM(Maboudou-Tchao,2019)
✓\checkmark
一
铰链损失
-
非线性
-
KL-SMM(Chen 等人,2020 年)
✓\checkmark
一
铰链损失
五折交叉验证
线性
O(sN^(2)pq)O\left(s N^{2} p q\right)
WOA-SMM(Zheng, Gu, Pan, & Tong, 2020)
✓\checkmark
One
铰链损失
WOA
线性
-
IQSMM (Zhang, Song and Wu, 2021)
-
One
-
-
线性
O(kappa^(2)log^(1.5)(kappa//epsilon):}O\left(\kappa^{2} \log ^{1.5}(\kappa / \epsilon)\right.log(Npq))+O(log(pq))\log (N p q))+O(\log (p q))
PSMM(张和李,2022)
✓\checkmark
One
Least square loss
-
线性
-
NPBSMM(Pan, Xu, Zheng 和 Tong, 2023)
✓\checkmark
Constructs non-parallel hyperplanes.
铰链损失
五折交叉验证
线性
O(sN(p+q)K)O(s N(p+q) K)
Model SRM principle Decision hyperplane Loss function CV Five/ten Linear/Non-linear variant Computational complexity
SMM (Luo et al., 2015) ✓ One Hinge loss - Linear -
PTSMM (Xu, Fan, & Gao, 2015) ✓ A pair of projection axes. Hinge loss Five-fold CV Non-linear -
TMRSMM (Gao, Fan, & Xu, 2016) x Two non-parallel hyperplanes. Hinge loss Five-fold CV Both linear and non-linear -
QSMM (Duan et al., 2017) ✓ One least square loss - Linear O(kappa^(3)epsilon^(-3)(log(Npq))) +O(log(pq))
MRMLTSMCM (Jiang & Yang, 2018) ✓ Two non-parallel hyperplanes. Hinge loss Five-fold CV Linear -
KSMM (Ye, 2019) ✓ One Hinge loss CV Non-linear O(N^(2)smn^(2))
MDSMM (Ye & Han, 2019) ✓ One Hinge loss Ten-fold CV Linear case O(N^(2)s(pq+p+q))
WSMM (Maboudou-Tchao, 2019) ✓ One Hinge loss - non-linear -
KL-SMM (Chen et al., 2020) ✓ One Hinge loss Five-fold CV Linear O(sN^(2)pq)
WOA-SMM (Zheng, Gu, Pan, & Tong, 2020) ✓ One Hinge loss WOA Linear -
IQSMM (Zhang, Song and Wu, 2021) - One - - Linear O(kappa^(2)log^(1.5)(kappa//epsilon):} log(Npq))+O(log(pq))
PSMM (Zhang & Liu, 2022) ✓ One Least square loss - Linear -
NPBSMM (Pan, Xu, Zheng and Tong, 2023) ✓ Constructs non-parallel hyperplanes. Hinge loss Five-fold CV Linear O(sN(p+q)K)| Model | SRM principle | Decision hyperplane | Loss function | CV Five/ten | Linear/Non-linear variant | Computational complexity |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| SMM (Luo et al., 2015) | $\checkmark$ | One | Hinge loss | - | Linear | - |
| PTSMM (Xu, Fan, & Gao, 2015) | $\checkmark$ | A pair of projection axes. | Hinge loss | Five-fold CV | Non-linear | - |
| TMRSMM (Gao, Fan, & Xu, 2016) | $x$ | Two non-parallel hyperplanes. | Hinge loss | Five-fold CV | Both linear and non-linear | - |
| QSMM (Duan et al., 2017) | $\checkmark$ | One | least square loss | - | Linear | $O\left(\kappa^{3} \epsilon^{-3}(\log (N p q))\right)$ $+O(\log (p q))$ |
| MRMLTSMCM (Jiang & Yang, 2018) | $\checkmark$ | Two non-parallel hyperplanes. | Hinge loss | Five-fold CV | Linear | - |
| KSMM (Ye, 2019) | $\checkmark$ | One | Hinge loss | CV | Non-linear | $O\left(N^{2} s m n^{2}\right)$ |
| MDSMM (Ye & Han, 2019) | $\checkmark$ | One | Hinge loss | Ten-fold CV | Linear case | $O\left(N^{2} s(p q+p+q)\right)$ |
| WSMM (Maboudou-Tchao, 2019) | $\checkmark$ | One | Hinge loss | - | non-linear | - |
| KL-SMM (Chen et al., 2020) | $\checkmark$ | One | Hinge loss | Five-fold CV | Linear | $O\left(s N^{2} p q\right)$ |
| WOA-SMM (Zheng, Gu, Pan, & Tong, 2020) | $\checkmark$ | One | Hinge loss | WOA | Linear | - |
| IQSMM (Zhang, Song and Wu, 2021) | - | One | - | - | Linear | $O\left(\kappa^{2} \log ^{1.5}(\kappa / \epsilon)\right.$ $\log (N p q))+O(\log (p q))$ |
| PSMM (Zhang & Liu, 2022) | $\checkmark$ | One | Least square loss | - | Linear | - |
| NPBSMM (Pan, Xu, Zheng and Tong, 2023) | $\checkmark$ | Constructs non-parallel hyperplanes. | Hinge loss | Five-fold CV | Linear | $O(s N(p+q) K)$ |
correlation data, improving low-rank approximation. A brief comparative analysis among the class-imbalance SMM variants is provided below:
Handling Class Imbalance: ESMM (Zhu, 2017) and CWSMM (Li, Cheng et al., 2021) specifically address the challenges of imbalanced datasets by ensuring the significance of the minority class and employing dynamic penalty factors tailored to individual class samples.
Adaptive Framework: DPAMM (Xu et al., 2022) utilizes an adaptive framework for minimizing low-rank approximations, contributing to the retention of highly significant correlation data within the input matrix.
This analysis provides insights into the strengths and distinctive features of each variant. Depending on the specific characteristics of the dataset and requirements of the application, practitioners can choose the most suitable approach for effectively handling class-imbalance in SMM. Table 7 showcases a summary of the various class-imbalance variants of SMM.
Having traversed an extensive expanse of SMM variants in the preceding subsections that encompass least squares approaches, robust and sparse formulations, multi-class classification paradigms, and strategies to address class-imbalance. In the next subsection, we transition our discussion to deep variants of SMM.
The random projections of the predictions of each layer modify the original feature and are passed to the next layer of DSSMM. Instead of parameter fine-tuning via backpropagation, DSSMM uses an effective feed-forward method, where each layer forms a convex optimization problem. Also, hinge loss for penalizing misclassification leads to the effectiveness of the model (Vinyals, Jia, Deng, & Darrell, 2012). Though DSSMM has better performance than SMM, the training of DSSMM model has the issue of using the pre-extracted information which may degrade the classification accuracy of the model as the pre-extracted information may not be sufficient or may be corrupted by noise. The following models overcome the said issue.
The concept of plane-based learning applied to regression problems is termed as support vector regression (SVR) (Smola & Schölkopf, 2004). It is rigorous and has convex QPP applied to find the global optimal solution, which resolves the local minima issue that the neural network model faces (Tang, Ma, Hu, & Tang, 2019). Taking motivation from SVR, Yuan and Weng (2021) introduced support matrix regression (SMR). SMR (Yuan & Weng, 2021) applies the idea of matrix input to the regression problems along with preserving the structural information of the matrix data. The objective of SMR is to maximize the
margin and minimize the squared Frobenius norm of the matrix, hence reducing the sensitivity of the regression to noisy data and leading to robustness against noise. SMR overcomes the lack of physical degradability in the input data and is effective for time asynchronization issues.
Improved generalization capability of a target domain by leveraging information from the source domain.
ADMM
WOA-SMM (2020)
Zheng 等人 (2020)
time-frequency domain features are extracted using multisynchrosqueezing transform (MSST) to construct the feature matrix.
铰链损失
Fault dataset from Case Western Reserve University (CWRU) and Anhui University of Technology (AHUT)
Improves the classification performance, consumes less time and lower calculation cost.
WOA is used to solve the optimization problem.
IQSMM (2021)
张三等 (2021)
The QPP of SMM is transformed to the solution of the system of linear equations by incorporating least square loss and solved using improved quantum matrix inversion and QSVT.
Leads to robust and sparse model. Suitable for large-scale data as matrix inversion is not required.
双坐标下降(DCD)
Other variants of SMM.
Model Author Characteristics Loss function Datasets Advantages Technique to solve
SMM (2015) Luo et al. (2015) Spectral elastic net penalty having Frobenius and nuclear norm. Hinge loss EEG alcoholism, EEG emotion, the students face and INRIA person Preserves the correlation within a matrix. ADMM
PTSMM (2015) Xu et al. (2015) Seeks projection axis of both classes with a minimum within-class variance of each and scattered projected samples of other classes as far as possible. Hinge loss 2d image classification using ORL, YALE and AR face databases Deals with non-linear cases using a new matrix kernel function. Considers SRM principle. SOR to solve QPP.
LTMRSMM, NTMRSMM (2016) Gao et al. (2016) Deals with matrix data having multiple ranks. Hinge loss Feret, ORL, FingerDB, Palm100 and Ar Reduced computational cost than the multi rank matrix vectorization method. Iteratively solving the QPPs.
QSMM (2017) Duan et al. (2017) The QPP of SMM is transformed to the solution of a system of linear equations by incorporating least square loss and solved using quantum matrix inversion (HHL) and QSVT. Square loss function - Exponential increase of speed over classical SMM. Complexity: O(kappa^(3)epsilon^(-3)(log(Npq)))+O(log(pq)), whereas complexity of SMM O(poly(N,pq)). HHL and QSVT algorithm
MRMLTSMCM (2018) Jiang and Yang (2018) Extension of TWSVM, uses pairs of projecting matrices to obtain the non-parallel hyperplanes. Hinge loss UCI datasets: Sonar, CMC, Hill-valley, Ionosphere, Madelon, Pedestrian, Pollen, FingerDB, Binucleate, RGB Efficient than TWSVM, implements SRM principle. Optimizing the obtained QPPs alternatively.
KSMM (2018) Ye (2019) Generates a matrix-based hyperplane by computing the weighted average distance. Hinge loss ORL face database, the Sheffield Face dataset, Columbia Object Image Library (COIL-20) and the binary alpha digits The matrix form inner product exploits the structural information of matrix data and solves optimization problem without using alternating projection method. SMO
MDSMM (2019) Ye and Han (2019) Introduces multi-distance to extract the intrinsic information of input matrix and used vector-based distance to quantify the cost function and penalty function. Hinge loss IMM face dataset, the Japanese female facial expression (JAFFE) dataset (Lyons, Akamatsu, Kamachi, Gyoba and Budynek, 1998), the jochen triesch static hand posture dataset (von der Malsburg, 1996), the Columbia object image library COIL-20 (Nene, Nayar, & Murase, 1996), and the Columbia Object Image Library COIL-100 (Nene et al., 1996) Improves the generalization performance. Alternating projection method is used to solve the optimization problem.
WSMM (2019) MaboudouTchao (2019) Wavelet kernels introduced for non-linear case. Hinge loss EEG alcoholism dataset, INRIA person dataset Obtains Mercer kernel in the matrix space. Improves performance on the EEG and INRIA datasets. QPP is solved using quadratic programming software.
KL-SMM (2020) Chen et al. (2020) Uses the concept of transfer learning. Hinge loss Motor Imagery (MI) based EEG datasets Improved generalization capability of a target domain by leveraging information from the source domain. ADMM
WOA-SMM (2020) Zheng et al. (2020) time-frequency domain features are extracted using multisynchrosqueezing transform (MSST) to construct the feature matrix. Hinge loss Fault dataset from Case Western Reserve University (CWRU) and Anhui University of Technology (AHUT) Improves the classification performance, consumes less time and lower calculation cost. WOA is used to solve the optimization problem.
IQSMM (2021) Zhang, Song et al. (2021) The QPP of SMM is transformed to the solution of the system of linear equations by incorporating least square loss and solved using improved quantum matrix inversion and QSVT. Square loss function - complexity: O(kappa^(2)log^(1.5)(kappa//epsilon)(log(Npq)):} +O(log(pq)) Quantum matrix inversion and QSVT
PSMM (2022) Zhang and Liu (2022) Constructs proximal hyperplane for the different classes. - minst digital database (LeCun, Bottou, Bengio, & Haffner, 1998), MIT face database, INRIA person database (Dalal & Triggs, 2005), students face database (Nazir, Ishtiaq, Batool, Jaffar, & Mirza, 2010), JAFFE (Lyons, Akamatsu, Kamachi and Gyoba, 1998) Simpler formulation than SMM, efficient than SMM in time complexity. ADMM
NPBSMM (2023) Pan, Xu et al. (2023) Constrain norm group is introduced in the optimization problem. Hinge loss AHUT fault dataset of roller bearing Leads to robust and sparse model. Suitable for large-scale data as matrix inversion is not required. Dual coordinate descent (DCD)| Other variants of SMM. | | | | | | |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| SMM (2015) | Luo et al. (2015) | Spectral elastic net penalty having Frobenius and nuclear norm. | Hinge loss | EEG alcoholism, EEG emotion, the students face and INRIA person | Preserves the correlation within a matrix. | ADMM |
| PTSMM (2015) | Xu et al. (2015) | Seeks projection axis of both classes with a minimum within-class variance of each and scattered projected samples of other classes as far as possible. | Hinge loss | 2d image classification using ORL, YALE and AR face databases | Deals with non-linear cases using a new matrix kernel function. Considers SRM principle. | SOR to solve QPP. |
| LTMRSMM, NTMRSMM (2016) | Gao et al. (2016) | Deals with matrix data having multiple ranks. | Hinge loss | Feret, ORL, FingerDB, Palm100 and Ar | Reduced computational cost than the multi rank matrix vectorization method. | Iteratively solving the QPPs. |
| QSMM (2017) | Duan et al. (2017) | The QPP of SMM is transformed to the solution of a system of linear equations by incorporating least square loss and solved using quantum matrix inversion (HHL) and QSVT. | Square loss function | - | Exponential increase of speed over classical SMM. Complexity: $O\left(\kappa^{3} \epsilon^{-3}(\log (N p q))\right)+O(\log (p q))$, whereas complexity of SMM $O(\operatorname{poly}(N, p q))$. | HHL and QSVT algorithm |
| MRMLTSMCM (2018) | Jiang and Yang (2018) | Extension of TWSVM, uses pairs of projecting matrices to obtain the non-parallel hyperplanes. | Hinge loss | UCI datasets: Sonar, CMC, Hill-valley, Ionosphere, Madelon, Pedestrian, Pollen, FingerDB, Binucleate, RGB | Efficient than TWSVM, implements SRM principle. | Optimizing the obtained QPPs alternatively. |
| KSMM (2018) | Ye (2019) | Generates a matrix-based hyperplane by computing the weighted average distance. | Hinge loss | ORL face database, the Sheffield Face dataset, Columbia Object Image Library (COIL-20) and the binary alpha digits | The matrix form inner product exploits the structural information of matrix data and solves optimization problem without using alternating projection method. | SMO |
| MDSMM (2019) | Ye and Han (2019) | Introduces multi-distance to extract the intrinsic information of input matrix and used vector-based distance to quantify the cost function and penalty function. | Hinge loss | IMM face dataset, the Japanese female facial expression (JAFFE) dataset (Lyons, Akamatsu, Kamachi, Gyoba and Budynek, 1998), the jochen triesch static hand posture dataset (von der Malsburg, 1996), the Columbia object image library COIL-20 (Nene, Nayar, & Murase, 1996), and the Columbia Object Image Library COIL-100 (Nene et al., 1996) | Improves the generalization performance. | Alternating projection method is used to solve the optimization problem. |
| WSMM (2019) | MaboudouTchao (2019) | Wavelet kernels introduced for non-linear case. | Hinge loss | EEG alcoholism dataset, INRIA person dataset | Obtains Mercer kernel in the matrix space. Improves performance on the EEG and INRIA datasets. | QPP is solved using quadratic programming software. |
| KL-SMM (2020) | Chen et al. (2020) | Uses the concept of transfer learning. | Hinge loss | Motor Imagery (MI) based EEG datasets | Improved generalization capability of a target domain by leveraging information from the source domain. | ADMM |
| WOA-SMM (2020) | Zheng et al. (2020) | time-frequency domain features are extracted using multisynchrosqueezing transform (MSST) to construct the feature matrix. | Hinge loss | Fault dataset from Case Western Reserve University (CWRU) and Anhui University of Technology (AHUT) | Improves the classification performance, consumes less time and lower calculation cost. | WOA is used to solve the optimization problem. |
| IQSMM (2021) | Zhang, Song et al. (2021) | The QPP of SMM is transformed to the solution of the system of linear equations by incorporating least square loss and solved using improved quantum matrix inversion and QSVT. | Square loss function | - | complexity: $O\left(\kappa^{2} \log ^{1.5}(\kappa / \epsilon)(\log (N p q))\right.$ $+O(\log (p q))$ | Quantum matrix inversion and QSVT |
| PSMM (2022) | Zhang and Liu (2022) | Constructs proximal hyperplane for the different classes. | - | minst digital database (LeCun, Bottou, Bengio, & Haffner, 1998), MIT face database, INRIA person database (Dalal & Triggs, 2005), students face database (Nazir, Ishtiaq, Batool, Jaffar, & Mirza, 2010), JAFFE (Lyons, Akamatsu, Kamachi and Gyoba, 1998) | Simpler formulation than SMM, efficient than SMM in time complexity. | ADMM |
| NPBSMM (2023) | Pan, Xu et al. (2023) | Constrain norm group is introduced in the optimization problem. | Hinge loss | AHUT fault dataset of roller bearing | Leads to robust and sparse model. Suitable for large-scale data as matrix inversion is not required. | Dual coordinate descent (DCD) |
Table 12
SMM for regression and semi-supervised learning.
模型
作者
特征
损失函数
数据集
优点
解决方法
SMR (2021)
袁和翁(2021)
Incorporates the idea of matrix learning for regression problems.
epsi\varepsilon insensitive loss
Test distribution grid: single phase IEEE 8-bus system (Liao, Weng, Liu, & Rajagopal, 2018), IEEE 123-bus system. utility distribution network modified from (Narang, Ayyanar, Gemin, Baggu, & Srinivasan, 2015)
Used for learning power flow mapping and overcomes the lack of physical degradability, thus overfitting. Robust to noise/outliers. effective for time asynchronization issues.
-
SPSMM (2023)
Li, Li, Yan, Shao 和 Lin (2023)
A strategy based on probability output is utilized to estimate the posterior class probabilities for matrix inputs. Furthermore, a semi-supervised learning framework is implemented to facilitate the transfer of knowledge from unlabeled samples to labeled ones.
-
An infrared thermal imaging dataset
Mitigate the issue of limited labeled samples and bolster the generalization performance.
SMO
Model Author Characteristics Loss function Datasets Advantages Technique to solve
SMR (2021) Yuan and Weng (2021) Incorporates the idea of matrix learning for regression problems. epsi insensitive loss Test distribution grid: single phase IEEE 8-bus system (Liao, Weng, Liu, & Rajagopal, 2018), IEEE 123-bus system. utility distribution network modified from (Narang, Ayyanar, Gemin, Baggu, & Srinivasan, 2015) Used for learning power flow mapping and overcomes the lack of physical degradability, thus overfitting. Robust to noise/outliers. effective for time asynchronization issues. -
SPSMM (2023) Li, Li, Yan, Shao, and Lin (2023) A strategy based on probability output is utilized to estimate the posterior class probabilities for matrix inputs. Furthermore, a semi-supervised learning framework is implemented to facilitate the transfer of knowledge from unlabeled samples to labeled ones. - An infrared thermal imaging dataset Mitigate the issue of limited labeled samples and bolster the generalization performance. SMO| Model | Author | Characteristics | Loss function | Datasets | Advantages | Technique to solve |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| SMR (2021) | Yuan and Weng (2021) | Incorporates the idea of matrix learning for regression problems. | $\varepsilon$ insensitive loss | Test distribution grid: single phase IEEE 8-bus system (Liao, Weng, Liu, & Rajagopal, 2018), IEEE 123-bus system. utility distribution network modified from (Narang, Ayyanar, Gemin, Baggu, & Srinivasan, 2015) | Used for learning power flow mapping and overcomes the lack of physical degradability, thus overfitting. Robust to noise/outliers. effective for time asynchronization issues. | - |
| SPSMM (2023) | Li, Li, Yan, Shao, and Lin (2023) | A strategy based on probability output is utilized to estimate the posterior class probabilities for matrix inputs. Furthermore, a semi-supervised learning framework is implemented to facilitate the transfer of knowledge from unlabeled samples to labeled ones. | - | An infrared thermal imaging dataset | Mitigate the issue of limited labeled samples and bolster the generalization performance. | SMO |
Table 13 脑电图数据集。
数据集
描述
Sampling frequency
论文
链接
BCI 竞赛 III 数据集 IVa (BCIC34a)
118通道,5名受试者(每名受试者280次试验)在涉及右手或足部运动的运动想象任务期间
1000 Hz
Hang 等人(2020, 2023),Liang、Hang、Lei 等人(2022),Liang、Hang、Yin 等人(2022),Razzak、Hameed 和 Xu(2019)以及 Zheng、Zhu、Heng(2018)
32-channel self-collected EEG signals from 10 subjects
512 Hz
Hang 等人 (2020,2023)(2020,2023)
-
上海交通大学情绪脑电图数据集(SEED-VIG)疲劳
17-channel 3 classes: awake, tired, drowsy from 10 subjects
512 Hz
Li, Wang 等人 (2022)
链接
Dataset Description Sampling frequency Papers Link
BCI competition III dataset IVa (BCIC34a) 118-channel, 5 subjects ( 280 trials per subject) during motor imagery tasks involving right-hand or foot movements 1000 Hz Hang et al. (2020, 2023), Liang, Hang, Lei et al. (2022), Liang, Hang, Yin et al. (2022), Razzak, Hameed and Xu (2019) and Zheng, Zhu, Heng (2018) Link
BCI competition III dataset IIIa (BCIC33a) 60-channel, single-trail 3 subjects with 4 classes: right hand, left hand, tongue and feet 250 Hz Hang et al. (2023), Razzak, Blumenstein et al. (2019), Razzak, Hameed et al. (2019) and Zheng, Zhu, Qin, Heng (2018) Link
BCI competition IV dataset IIa (BCIC32a) 22-channel, 4 classes: left hand, right hand, feet and tongue, 9 subjects 250 Hz Chen et al. (2020), Hang et al. (2020), Liang, Hang, Yin et al. (2022), Razzak, Blumenstein et al. (2019), Razzak, Hameed et al. (2019), Zheng, Zhu, Heng (2018) and Zheng, Zhu, Qin, Heng (2018) Link
BCI competition IV dataset IIb (BCIC32b) 3 bi-polar EEG channels, 2 classes: left hand, right hand, feet and tongue, 9 subjects 250 Hz Chen et al. (2020), Hang et al. (2020), Liang, Hang, Yin et al. (2022), Razzak, Hameed et al. (2019) and Zheng, Zhu, Heng (2018) Link
Lower Limb MI-BCI dataset (LLMI-BCI) 32-channel self-collected EEG signals from 10 subjects 512 Hz Hang et al. (2020,2023) -
The SJTU emotion EEG dataset (SEED-VIG) Fatigue 17-channel 3 classes: awake, tired, drowsy from 10 subjects 512 Hz Li, Wang et al. (2022) Link| Dataset | Description | Sampling frequency | Papers | Link |
| :--- | :--- | :--- | :--- | :--- |
| BCI competition III dataset IVa (BCIC34a) | 118-channel, 5 subjects ( 280 trials per subject) during motor imagery tasks involving right-hand or foot movements | 1000 Hz | Hang et al. (2020, 2023), Liang, Hang, Lei et al. (2022), Liang, Hang, Yin et al. (2022), Razzak, Hameed and Xu (2019) and Zheng, Zhu, Heng (2018) | Link |
| BCI competition III dataset IIIa (BCIC33a) | 60-channel, single-trail 3 subjects with 4 classes: right hand, left hand, tongue and feet | 250 Hz | Hang et al. (2023), Razzak, Blumenstein et al. (2019), Razzak, Hameed et al. (2019) and Zheng, Zhu, Qin, Heng (2018) | Link |
| BCI competition IV dataset IIa (BCIC32a) | 22-channel, 4 classes: left hand, right hand, feet and tongue, 9 subjects | 250 Hz | Chen et al. (2020), Hang et al. (2020), Liang, Hang, Yin et al. (2022), Razzak, Blumenstein et al. (2019), Razzak, Hameed et al. (2019), Zheng, Zhu, Heng (2018) and Zheng, Zhu, Qin, Heng (2018) | Link |
| BCI competition IV dataset IIb (BCIC32b) | 3 bi-polar EEG channels, 2 classes: left hand, right hand, feet and tongue, 9 subjects | 250 Hz | Chen et al. (2020), Hang et al. (2020), Liang, Hang, Yin et al. (2022), Razzak, Hameed et al. (2019) and Zheng, Zhu, Heng (2018) | Link |
| Lower Limb MI-BCI dataset (LLMI-BCI) | 32-channel self-collected EEG signals from 10 subjects | 512 Hz | Hang et al. $(2020,2023)$ | - |
| The SJTU emotion EEG dataset (SEED-VIG) Fatigue | 17-channel 3 classes: awake, tired, drowsy from 10 subjects | 512 Hz | Li, Wang et al. (2022) | Link |
is used as a preprocessing technique for BCIs. Another preprocessing technique is downsampling the EEG data to reduce the highcomputational costs (Bischof & Bunch, 2021). After preprocessing, the EEG data undergoes feature extraction or feature selection in order to extract the most helpful information from the input data matrix. The standard methods of feature extraction are time domain parameters (TDP) (Vidaurre, Krämer, Blankertz, & Schlögl, 2009), fast Fourier transform (FFT) (Shakshi & Jaswal, 2016), principle component analysis (PCA) (Kuncheva & Faithfull, 2013), and so forth (Hu & Zhang, 2019).
The presence of noise and outliers in the EEG data affects the classification of data. To handle this, Zheng, Zhu, Heng (2018) proposed a classifier entitled RSMM for single-trial EEG classification. The preprocessing techniques of the raw data used are Chebyshev Type II filter (Sen, Mishra, & Pattnaik, 2023) followed by spatial filtering using the CSP algorithm. The feature extraction is done using TDP algorithm, which leads to the robustness of the model (Nicolas-Alonso & Gomez-Gil, 2012). The EEG data is highly complex because of the high dimensionality. To overcome this complexity, Razzak, Hameed et al. (2019) proposed efficient feature extraction and showed comparative studies on the PCA algorithms, namely robust joint sparse PCA (RJSPCA) and outliers robust PCA (ORPCA) for dimensionality reduction (for simplicity, we denote the model as R-SMM in further paper). The preprocessing technique used is filter bank CSP (FBCSP) followed by the TDP algorithm for feature extraction. PCA is then applied to select the robust features from TDP which is beneficial in dimensionality reduction. Li, Wang et al. (2022) discusses the application of EEG-based fatigue and attention detection by using SEED-VIG for experimentation. The EEG fatigue signals are classified using the ACF-SSMM (Li, Wang et al., 2022) method. This method involves compressing the redundant features by use of the sparse principle.
The aforementioned methods involve binary classification only. However, to address the multi-class classification, a multi-class SMM (MSMM) is developed that aims to improve EEG-based BCIs’ performance involving multiple activities (Zheng, Zhu, Qin, Heng, 2018). The preprocessing is based on the techniques used by Ang et al. (2012), which employs non-overlapping band-pass filters of the sixth-order
Butterworth filter (Pise & Rege, 2021) to filter out the artifacts and unrelated signals, followed by CSP to select the most dominant channels. Several techniques are experimented upon for feature extraction, including band powers (BPO), power spectral density (PSD) and TPD, among which TPD led to the best results. MSMM (Zheng, Zhu, Qin, Heng, 2018) is the first attempt to handle a multiple-class EEG data classification that promotes a broader range of applications in BCI technology. To increase the generalization performance in multi-class SMM, Razzak, Blumenstein et al. (2019) proposed a multi-class SMM (M-SMM) which enhanced and increased the inter-class margins to focus on the single-trial multi-class classification of EEG signals. To mitigate the effect of outliers/noise, spatial filtering has been employed as an effective preprocessing technique to identify discriminative spatial patterns and eliminate uncorrelated information. Specifically, the FBCSP algorithm is used to filter out unrelated sensorimotor rhythms and artifacts by autonomously selecting a subject-specific frequency range for band-pass filtering of the EEG measurements.
The collection of EEG data is extremely time-consuming and challenging to obtain from the clinical point of view due to the intricacies of recording the data and privacy laws (Vaid, Singh, & Kaur, 2015). Thus, techniques are developed to leverage the available little source data and apply it to the target domain. Chen et al. (2020) proposed KL-SMM to improve the performance of the EEG signal classification when very little data on the target domain is available. KL-SMM is a five-order Butterworth filter followed by spatial filters as part of the preprocessing techniques. Another article that addresses classification in cases of insufficient data is AMK-TMM (Liang, Hang, Lei et al., 2022) based on the LS-SMM. The AMK-TMM framework introduces an adaptive approach that uses the leave-one-out CV strategy to identify many correlated source models and their corresponding weights. This enables construction of the target classifier and the identification of the correlated source models to be integrated into a single learning framework.
Inspired by the deep learning and transfer learning techniques for increasing the performance of the model with less amount of data, Hang et al. (2020) proposed SMM to be a basic building block of a DSN and introduced DSSMM which uses first-ordered band-pass filter for
Table 14
脑电图分类应用。
模型
作者
数据集
指标
特征提取
预处理技术
涉及未来方向
RSMM(2018)
郑、朱、何(2018)
BCIC34a、BCIC42b 和 BCIC32a
精度
TDP 算法
Chebyshev type 2 filter and CSP
是
MSMM(2018)
郑珠,秦恒(2018)
BCIC34a、BCIC42b 和 BCIC32a
Kappa coefficient, precision, recall and F-measure
TDP 算法
基于 Ang 等人(2012)的非重叠带通滤波器,六阶 Butterworth 和 CSP
是
R-SMM(2019)
Razzak, Hameed 等人 (2019)
BCIC34a, BCIC42a 和 BCIC42b
Kappa 系数、精确率、召回率和 F-measure
JSPCA 算法优于 TDP 算法
FBCSP
否
M-SMM (2019)
Razzak、Blumenstein 等人 (2019)
BCIC33a 和 BCIC42b
召回率、精确率、F-measure 和 kappa 系数。
TDP 算法
FBCSP
否
DSSMM (2020)
韩等 (2020)
BCIC34a, BCIC42b, BCIC42a, LLMI-BCI
准确率,F1,AUC
TDP 算法
使用五阶巴特沃斯滤波器和空间滤波器的带通滤波器
是
KL-SMM (2020)
Chen 等人 (2020)
BCIC42a 和 BCIC42b
准确率,F1,AUC
-
五阶巴特沃斯滤波器后接空间滤波器
是
DSFR (2022)
Liang, Hang, Yin 等 (2022)
BCIC34a, BCIC42b, 和 BCIC42a
准确率,F1,AUC
-
五阶带通滤波器和 CSP
是
AMK-TMM (2022)
梁航、雷等 (2022)
BCIC34a、BCIC42a 和 LLMI-BCI
准确率、标准差、AUC
-
五阶巴特沃斯滤波器和 CSP
是
ACFSSMM (2022)
Li, Wang 等人 (2022)
SEED-VIG 数据集
准确率,F1,AUC
TDP 算法
带通滤波器和 CSP
否
DSTLSSMM(2023)
韩等 (2023)
BCIC33a、BCIC34a 和 LLMI-BCI
准确率、召回率、kappa、F-Score
-
五阶巴特沃斯带通滤波和 CSP
是
Model Author Dataset Metric Feature extraction Preprocessing technique Involves future directions
RSMM (2018) Zheng, Zhu, Heng (2018) BCIC34a, BCIC42b, and BCIC32a Accuracy TDP algorithm Chebyshev type 2 filter and CSP Yes
MSMM (2018) Zheng, Zhu, Qin, Heng (2018) BCIC34a, BCIC42b, and BCIC32a Kappa coefficient, precision, recall and F-measure TDP algorithm Based on Ang et al. (2012) non-overlapping band-pass filters of six-order Butter-worth and CSP Yes
R-SMM (2019) Razzak, Hameed et al. (2019) BCIC34a, BCIC42a and BCIC42b Kappa coefficient, precision, recall and F-measure JSPCA over TDP algorithm FBCSP No
M-SMM (2019) Razzak, Blumenstein et al. (2019) BCIC33a and BCIC42b Recall, precision, F-measure and kappa coefficient. TDP algorithm FBCSP No
DSSMM (2020) Hang et al. (2020) BCIC34a, BCIC42b, BCIC42a, LLMI-BCI Accuracy, F1, AUC TDP algorithm Band-pass filter using a fifth-order Butterworth filter and spatial filtering Yes
KL-SMM (2020) Chen et al. (2020) BCIC42a and BCIC42b Accuracy, F1, AUC - Five-order Butterworth filter followed by spatial filters Yes
DSFR (2022) Liang, Hang, Yin et al. (2022) BCIC34a, BCIC42b, and BCIC42a Accuracy, F1, AUC - Five-ordered band-pass filter and CSP Yes
AMK-TMM (2022) Liang, Hang, Lei et al. (2022) BCIC34a, BCIC42a, and LLMI-BCI Accuracy, standard deviation, AUC - Fifth-order Butterworth filter and CSP Yes
ACFSSMM (2022) Li, Wang et al. (2022) SEED-VIG dataset Accuracy, F1, AUC TDP Algorithm Band-pass filter and CSP No
DSTLSSMM (2023) Hang et al. (2023) BCIC33a, BCIC34a and LLMI-BCI Accuracy, recall, kappa, F-Score - Five-order Butterworth band-pass filtering and CSP Yes| Model | Author | Dataset | Metric | Feature extraction | Preprocessing technique | Involves future directions |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| RSMM (2018) | Zheng, Zhu, Heng (2018) | BCIC34a, BCIC42b, and BCIC32a | Accuracy | TDP algorithm | Chebyshev type 2 filter and CSP | Yes |
| MSMM (2018) | Zheng, Zhu, Qin, Heng (2018) | BCIC34a, BCIC42b, and BCIC32a | Kappa coefficient, precision, recall and F-measure | TDP algorithm | Based on Ang et al. (2012) non-overlapping band-pass filters of six-order Butter-worth and CSP | Yes |
| R-SMM (2019) | Razzak, Hameed et al. (2019) | BCIC34a, BCIC42a and BCIC42b | Kappa coefficient, precision, recall and F-measure | JSPCA over TDP algorithm | FBCSP | No |
| M-SMM (2019) | Razzak, Blumenstein et al. (2019) | BCIC33a and BCIC42b | Recall, precision, F-measure and kappa coefficient. | TDP algorithm | FBCSP | No |
| DSSMM (2020) | Hang et al. (2020) | BCIC34a, BCIC42b, BCIC42a, LLMI-BCI | Accuracy, F1, AUC | TDP algorithm | Band-pass filter using a fifth-order Butterworth filter and spatial filtering | Yes |
| KL-SMM (2020) | Chen et al. (2020) | BCIC42a and BCIC42b | Accuracy, F1, AUC | - | Five-order Butterworth filter followed by spatial filters | Yes |
| DSFR (2022) | Liang, Hang, Yin et al. (2022) | BCIC34a, BCIC42b, and BCIC42a | Accuracy, F1, AUC | - | Five-ordered band-pass filter and CSP | Yes |
| AMK-TMM (2022) | Liang, Hang, Lei et al. (2022) | BCIC34a, BCIC42a, and LLMI-BCI | Accuracy, standard deviation, AUC | - | Fifth-order Butterworth filter and CSP | Yes |
| ACFSSMM (2022) | Li, Wang et al. (2022) | SEED-VIG dataset | Accuracy, F1, AUC | TDP Algorithm | Band-pass filter and CSP | No |
| DSTLSSMM (2023) | Hang et al. (2023) | BCIC33a, BCIC34a and LLMI-BCI | Accuracy, recall, kappa, F-Score | - | Five-order Butterworth band-pass filtering and CSP | Yes |
表15
滚动轴承故障诊断数据集。
数据集
Papers
AHUT 数据集
Gu et al. (2021), Pan, Sheng et al. (2022, 2023), Pan, Xu and Zheng (2022), Pan, Xu, Zheng, Liu and Tong (2022), Pan, Xu, Zheng, Su et al. (2022), Pan, Xu et al. (2023), Pan, Xu, Zheng, Tong et al. (2022), Pan and Zheng (2021), Wang et al. (2022) and Zheng et al. (2020)
Li et al. (2020), Pan, Xu, Zheng, Liu et al. (2022), Pan, Xu, Zheng, Tong et al. (2022) and Pan, Xu et al. (2023)
Vibration signal dataset from University of Connecticut (UCONN)
Li, Yang 等人 (2021)
苏州大学 (SUZ) 数据集
Gu 等人 (2021)
定制数据集
李等人(2023)和潘与郑(2021)
Dataset Papers
AHUT dataset Gu et al. (2021), Pan, Sheng et al. (2022, 2023), Pan, Xu and Zheng (2022), Pan, Xu, Zheng, Liu and Tong (2022), Pan, Xu, Zheng, Su et al. (2022), Pan, Xu et al. (2023), Pan, Xu, Zheng, Tong et al. (2022), Pan and Zheng (2021), Wang et al. (2022) and Zheng et al. (2020)
CRWU dataset Li et al. (2020), Pan, Sheng et al. (2022, 2023), Pan, Xu et al. (2023), Pan, Xu, Zheng, Tong et al. (2022), Pan, Yang, Zheng, Li, and Cheng (2019), Pan and Zheng (2021), Wang et al. (2022) and Zheng et al. (2020)
HNU dataset Li et al. (2020), Pan, Xu, Zheng, Liu et al. (2022), Pan, Xu, Zheng, Tong et al. (2022) and Pan, Xu et al. (2023)
Vibration signal dataset from University of Connecticut (UCONN) Li, Yang et al. (2021)
Dataset of Suzhou University (SUZ) Gu et al. (2021)
Custom dataset Li et al. (2023) and Pan and Zheng (2021)| Dataset | Papers |
| :--- | :--- |
| AHUT dataset | Gu et al. (2021), Pan, Sheng et al. (2022, 2023), Pan, Xu and Zheng (2022), Pan, Xu, Zheng, Liu and Tong (2022), Pan, Xu, Zheng, Su et al. (2022), Pan, Xu et al. (2023), Pan, Xu, Zheng, Tong et al. (2022), Pan and Zheng (2021), Wang et al. (2022) and Zheng et al. (2020) |
| CRWU dataset | Li et al. (2020), Pan, Sheng et al. (2022, 2023), Pan, Xu et al. (2023), Pan, Xu, Zheng, Tong et al. (2022), Pan, Yang, Zheng, Li, and Cheng (2019), Pan and Zheng (2021), Wang et al. (2022) and Zheng et al. (2020) |
| HNU dataset | Li et al. (2020), Pan, Xu, Zheng, Liu et al. (2022), Pan, Xu, Zheng, Tong et al. (2022) and Pan, Xu et al. (2023) |
| Vibration signal dataset from University of Connecticut (UCONN) | Li, Yang et al. (2021) |
| Dataset of Suzhou University (SUZ) | Gu et al. (2021) |
| Custom dataset | Li et al. (2023) and Pan and Zheng (2021) |
the preprocessing technique. Similar to DSSMM, Liang, Hang, Yin et al. (2022) proposed a DSFR method that takes the input in the form of raw EEG data and performs learning directly from it. DSFR method reduces the reliance of the model on pre-extracted EEG features and can extract the features more effectively than CSP followed by classification. Liang, Hang, Yin et al. (2022) suggested the filtering of EEG signals using a five-ordered band-pass filter as preprocessing technique. Further, Hang et al. (2023) proposed the deep stacked method termed as DST-LSSMM with the building block module made of LSSMM. The author performed five-order Butterworth band-pass filtering followed by CSP on it as a
preprocessing technique. The spatial filters of the preserved CSP consisted of the initial filter along with the last three filters. Subsequently, the dynamic logarithmic power of the filtered EEG signals is calculated over time. This process yielded a matrix-based representation of EEG features.
Fault diagnosis uses a two-dimensional vibrational signal as an input. However, sometimes the feature matrix may be contaminated which leads to noise and outliers. To mitigate these contaminations, Gu et al. (2021) suggested RSSMM which uses MSST as its feature extraction technique and leads to robustness in fault diagnosis in roller bearings. Similarly, adaptive interactive deviation matrix machine (AIDMM) (Pan, Xu, Zheng, Liu et al., 2022) also contributes to the sparseness and robustness. A similar application is targeted by TRMM (Pan, Xu, Zheng, Tong et al., 2022) which is a non-parallel classifier for fault diagnosis. TRMM uses symplectic geometry similarity transformation (SGST) to extract the two-dimensional feature matrix and is insensitive to noise as well as robust to the outliers, which helps to improve the health of the mechanical equipment by accurate fault diagnosis. Meanwhile, Pan, Xu, Zheng, Su et al. (2022) proposed MFSMM for the fault diagnosis of roller bearings to diagnose the working state of the roller bearing accurately by the classification of outliers using the concept of fuzzy hyperplanes. To handle complicated roller bearing faults, i.e., the compound roller bearing faults, Li et al. (2020) proposed NPLSSMM which used continuous wavelet transform (CWT) as its feature extraction technique and minimized the effect of outliers. Further, its applications can also be extended to other rotating machinery for fault diagnosis. Application on different rotating machinery includes SWSSMM (Li, Yang et al., 2021), which extracts the distinct fault features in the gears directly from the raw vibration signal. Gear fault diagnosis requires professional expertise and knowledge, however, the proposed model extracts a symplectic weighted coefficient matrix using symplectic similar transform (SST).
Pan et al. (2019) proposed a multi-class classifier called symplectic geometry matrix machine (SGMM) for fault diagnosis of roller bearing, which is robust to noise and outliers. SGST is used to obtain a symplectic geometry coefficient matrix in SGMM that preserves the structural information and removes noise interference while preventing the convergence problem. The time-frequency features of the roller bearings are insufficient to represent the whole information and complete functioning. For a faster and high convergence activity of the optimization algorithm, Zheng et al. (2020) proposed the combination of the WOA and SMM, which involves MSST time-frequency analysis. The vibration signature from the drive-end bearing is chosen for analysis. An SKF bearing is employed for this purpose, and artificial defects are introduced at individual points on the ball, inner race, and outer race using spark machining techniques. Pan and Zheng (2021) proposed an improved version of SMM called symplectic hyperdisk matrix machine (SHMM) for fault diagnosis which suggests the use of SGST to obtain the input matrix in the form of a dimensionless feature matrix. Further, hyperdisk is used in SHMM to cluster the different kinds of data which makes the whole process robust and efficient.
Experimental setup: We carried out the experiments in MATLAB 2023b on a desktop PC having processor Intel® Xeon® Gold 6226R CPU @ 2.90 GHz and 128 GB RAM. The experimental setup involves a simple grid search on the different parameters involved in the model. The data is normalized using Z-score normalization. Initially, we divided the datasets into 70:30 for training and testing, respectively. Out of the 70%70 \% of the datasets obtained for training, we again split it into training and validation in the ratio of 70:30. Further, for the selection of optimal parameters we employ grid search on the parameters. The range of different parameters in SSMM, RSMM, RSSMM and MSMM are chosen as follows: C in{0.1,1,10,100},lambda inC \in\{0.1,1,10,100\}, \lambda \in{0,0.1,0.5,1,2,5,10}\{0,0.1,0.5,1,2,5,10\} following (Zheng, Zhu, Qin, Heng, 2018), and rho=\rho= 0.01 from (Xu et al., 2022). For RSMM (Zheng, Zhu, Heng, 2018), the notation lambda_(3)in{0.0001,0.001,0.01,0.1,1}\lambda_{3} \in\{0.0001,0.001,0.01,0.1,1\} following (Zheng, Zhu, Heng, 2018). For RSSMM (Gu et al., 2021), the constraint of the sparse term, alpha in{10^(-4),10^(-3.75),dots,10^(-1)}\alpha \in\left\{10^{-4}, 10^{-3.75}, \ldots, 10^{-1}\right\} following (Gu et al., 2021), the truncation parameter epsilon\epsilon is chosen from {0.1,0.2,dots,1}\{0.1,0.2, \ldots, 1\}. The metric used to evaluate the performance of the model is accuracy defined as:
accuracy =(" Number of correctly classified samples ")/(" total number of samples ")xx100=\frac{\text { Number of correctly classified samples }}{\text { total number of samples }} \times 100.
振动数据和红外热成像 (IRT) 图像 Spectra Quest, Inc. 位于美国弗吉尼亚州里士满
精度
ADMM
-
是
TRMM (2022)
Pan, Xu, Zheng, Tong 等人 (2022)
AHUT, CWRU 和 HNU 数据集
准确率,G 均值,F 度量,AUC,kappa,精确率,灵敏度,特异性
SGST
APG
5折交叉验证
是
SRMM (2022)
Pan, Xu, Zheng (2022)
AHUT 数据集
识别率、时间、kappa、准确率、召回率和 F1 以及统计检验
SGST
-
5折交叉验证
是
DPAMM (2022)
徐等人 (2022)
两个自定义数据集和 CWRU 数据集。
特异性、Gmean 和召回率
SGST
ADMM
网格搜索
否
LSISMM (2022)
Li, Shao 等人 (2022)
自定义数据集。
特异性、Gmean 和召回率
ADMM
5折交叉验证
是
MFSMM (2022)
Pan, Xu, Zheng, Su 等人 (2022)
AHUT 数据集
精确率,召回率,F 分数,kappa 系数,准确率和运营效率
ADMM
-
是
Pin-TSMM (2022)
潘胜等 (2022)
AHUT 和 CWRU 数据集
准确率、精确率、召回率、F1 分数和 kappa 系数
ADMM
网格搜索
是
SNMM (2022)
王等 (2022)
AHUT 和 CRWU 数据集
准确率、kappa 系数、召回率、F1 分数和精确率。
ADMM
5折交叉验证
是
AIDMM(2022)
潘、徐、郑、刘等(2022)
AHUT 和 HNU 数据集
准确率、kappa 系数、召回率、F1 分数和时间。
AIDMM
5折交叉验证
是
DSPTMM(2023)
潘胜等 (2023)
AHUT 数据集
卡方系数、召回率、精确率和 F1 值、准确率
ADMM
5折交叉验证
是
SPSMM (2023)
Li 等人(2023)
定制数据集
精度
ADMM
5折交叉验证
是
NPBSMM (2023)
潘, 徐等 (2023)
AHUT、CWRU 和 HNU 数据集
准确率、kappa 系数、召回率和 F1 分数
双坐标下降(DCD)算法
网格搜索
是
Model Author Dataset Metric Feature extraction Optimization technique Hyperparameter tuning Involves future directions
SGMM (2019) Pan et al. (2019) CWRU dataset Accuracy SGST ADMM 5-fold CV Yes
WOASMM (2020) Zheng et al. (2020) CWRU and AHUT datasets Accuracy MSST WOA 5-fold CV Yes
NPLSSMM (2020) Li et al. (2020) CWRU and HNU dataset Accuracy CWT ADMM 5-fold CV Yes
SWSSMM (2021) Li, Yang et al. (2021) UCONN Accuracy SST ADMM 5-fold CV Yes
RSSMM (2021) Gu et al. (2021) AHUT and SUZ datasets Accuracy MSST ADMM 5-fold CV Yes
SHMM (2021) Pan and Zheng (2021) CWRU, 6 types of roller-bearing data, and 12 types of roller-bearing data of AHUT Kappa, recall, precision and F1, accuracy ADMM 5-fold CV No
CWSMM (2021) Li, Cheng et al. (2021) Vibration data and infrared thermography (IRT) images Spectra Quest, Inc., located in Richmond, VA, USA Accuracy ADMM - Yes
TRMM (2022) Pan, Xu, Zheng, Tong et al. (2022) AHUT, CWRU and HNU datasets Accuracy, G-mean, Fmeasure, AUC, kappa, precision, sensitivity, specificity SGST APG 5-fold CV Yes
SRMM (2022) Pan, Xu, Zheng (2022) AHUT dataset Recognition rate, time, kappa, accuracy, recall rate and F1 and statistical test SGST - 5-fold CV Yes
DPAMM (2022) Xu et al. (2022) Two custom datasets and CWRU dataset. Specificity, Gmean, and recall SGST ADMM Grid-search No
LSISMM (2022) Li, Shao et al. (2022) Custom dataset. Specificity, Gmean, and recall ADMM 5-fold CV Yes
MFSMM (2022) Pan, Xu, Zheng, Su et al. (2022) AHUT dataset Precision, recall, F-score, kappa, accuracy and operational efficiency ADMM - Yes
Pin-TSMM (2022) Pan, Sheng et al. (2022) AHUT and CWRU dataset Accuracy, precision, recall, F1-score, and kappa coefficient ADMM Grid-search Yes
SNMM (2022) Wang et al. (2022) AHUT and CRWU dataset Accuracy, kappa, recall, F1 score and precision. ADMM 5-fold CV Yes
AIDMM (2022) Pan, Xu, Zheng, Liu et al. (2022) AHUT and HNU dataset Accuracy, kappa, recall, F1 score and time. AIDMM 5-fold CV Yes
DSPTMM (2023) Pan, Sheng et al. (2023) AHUT dataset Kappa, recall, precision and F1, accuracy ADMM 5-fold CV Yes
SPSMM (2023) Li et al. (2023) Custom dataset Accuracy ADMM 5-fold CV Yes
NPBSMM (2023) Pan, Xu et al. (2023) AHUT, CWRU, and HNU datasets Accuracy, kappa, recall and F1score Dual coordinate descent (DCD) algorithm Grid-search Yes| Model | Author | Dataset | Metric | Feature extraction | Optimization technique | Hyperparameter tuning | Involves future directions |
| :--- | :--- | :--- | :--- | :--- | :--- | :--- | :--- |
| SGMM (2019) | Pan et al. (2019) | CWRU dataset | Accuracy | SGST | ADMM | 5-fold CV | Yes |
| WOASMM (2020) | Zheng et al. (2020) | CWRU and AHUT datasets | Accuracy | MSST | WOA | 5-fold CV | Yes |
| NPLSSMM (2020) | Li et al. (2020) | CWRU and HNU dataset | Accuracy | CWT | ADMM | 5-fold CV | Yes |
| SWSSMM (2021) | Li, Yang et al. (2021) | UCONN | Accuracy | SST | ADMM | 5-fold CV | Yes |
| RSSMM (2021) | Gu et al. (2021) | AHUT and SUZ datasets | Accuracy | MSST | ADMM | 5-fold CV | Yes |
| SHMM (2021) | Pan and Zheng (2021) | CWRU, 6 types of roller-bearing data, and 12 types of roller-bearing data of AHUT | Kappa, recall, precision and F1, accuracy | | ADMM | 5-fold CV | No |
| CWSMM (2021) | Li, Cheng et al. (2021) | Vibration data and infrared thermography (IRT) images Spectra Quest, Inc., located in Richmond, VA, USA | Accuracy | | ADMM | - | Yes |
| TRMM (2022) | Pan, Xu, Zheng, Tong et al. (2022) | AHUT, CWRU and HNU datasets | Accuracy, G-mean, Fmeasure, AUC, kappa, precision, sensitivity, specificity | SGST | APG | 5-fold CV | Yes |
| SRMM (2022) | Pan, Xu, Zheng (2022) | AHUT dataset | Recognition rate, time, kappa, accuracy, recall rate and F1 and statistical test | SGST | - | 5-fold CV | Yes |
| DPAMM (2022) | Xu et al. (2022) | Two custom datasets and CWRU dataset. | Specificity, Gmean, and recall | SGST | ADMM | Grid-search | No |
| LSISMM (2022) | Li, Shao et al. (2022) | Custom dataset. | Specificity, Gmean, and recall | | ADMM | 5-fold CV | Yes |
| MFSMM (2022) | Pan, Xu, Zheng, Su et al. (2022) | AHUT dataset | Precision, recall, F-score, kappa, accuracy and operational efficiency | | ADMM | - | Yes |
| Pin-TSMM (2022) | Pan, Sheng et al. (2022) | AHUT and CWRU dataset | Accuracy, precision, recall, F1-score, and kappa coefficient | | ADMM | Grid-search | Yes |
| SNMM (2022) | Wang et al. (2022) | AHUT and CRWU dataset | Accuracy, kappa, recall, F1 score and precision. | | ADMM | 5-fold CV | Yes |
| AIDMM (2022) | Pan, Xu, Zheng, Liu et al. (2022) | AHUT and HNU dataset | Accuracy, kappa, recall, F1 score and time. | | AIDMM | 5-fold CV | Yes |
| DSPTMM (2023) | Pan, Sheng et al. (2023) | AHUT dataset | Kappa, recall, precision and F1, accuracy | | ADMM | 5-fold CV | Yes |
| SPSMM (2023) | Li et al. (2023) | Custom dataset | Accuracy | | ADMM | 5-fold CV | Yes |
| NPBSMM (2023) | Pan, Xu et al. (2023) | AHUT, CWRU, and HNU datasets | Accuracy, kappa, recall and F1score | | Dual coordinate descent (DCD) algorithm | Grid-search | Yes |
表 17
SMM 的其他应用。
模型
作者
数据集
指标
应用
PTSMM (2015)
Xu 等人 (2015)
ORL、YALE、AR 数据集
准确率和运行时间
二维图像分类
S-SMM(2019)
刘、焦、张和刘(2019)
机载系统(NASA/JPL-Caltech AIRSAR)的 PolSAR 图像。
准确性和 kappa 系数
二维图像分类
SMR (2021)
袁和翁(2021)
-
精度
无需可观测性自动计算配电网潮流
Model Author Dataset Metric Application
PTSMM (2015) Xu et al. (2015) ORL, YALE, AR dataset Accuracy and running time 2d Image classification
S-SMM (2019) Liu, Jiao, Zhang and Liu (2019) PolSAR images from an airborne system (NASA/JPL-Caltech AIRSAR). Accuracy and kappa coefficient 2d Image classification
SMR (2021) Yuan and Weng (2021) - Accuracy Calculate the power flow in the distribution grid automatically without any observability| Model | Author | Dataset | Metric | Application |
| :--- | :--- | :--- | :--- | :--- |
| PTSMM (2015) | Xu et al. (2015) | ORL, YALE, AR dataset | Accuracy and running time | 2d Image classification |
| S-SMM (2019) | Liu, Jiao, Zhang and Liu (2019) | PolSAR images from an airborne system (NASA/JPL-Caltech AIRSAR). | Accuracy and kappa coefficient | 2d Image classification |
| SMR (2021) | Yuan and Weng (2021) | - | Accuracy | Calculate the power flow in the distribution grid automatically without any observability |
Akhtar, M., Tanveer, M., Arshad, M., & 阿尔茨海默病神经影像学倡议 (2024c). 基于波动损失函数的监督学习方法:一种鲁棒和平滑的方法。模式识别,文章 110637。 http://dx.doi.org/ 10.1016/j.patcog.2024.110637。
Hüllermeier, E. (2005)。机器学习和数据挖掘中的模糊方法:现状与前景。模糊集与系统,156(3),387-406。
Iglovikov, V. I., Rakhlin, A., Kalinin, A. A., & Shvets, A. A. (2018)。使用深度卷积神经网络的儿童骨龄评估。在深度学习在医学图像分析和临床决策支持的多模态学习:第 4 届国际研讨会,DLMIA 2018,和第 8 届国际研讨会,ML-CDS 2018,与 MICCAI 2018 联合举行,西班牙格拉纳达,2018 年 9 月 20 日,论文集 4 (pp. 300-308)。Springer。
Jayadeva, Khemchandani, R., & Chandra, S. (2007). 双支持向量机用于模式分类. IEEE 模式分析与机器智能汇刊, 29(5), 905-910.
Jia, X., Feng, X., Wang, W., Xu, C., & Zhang, L. (2018). 贝叶斯推理用于自适应低秩和稀疏矩阵估计. 神经计算, 291, 71-83.
Jiang, R., & Yang, Z.-X. (2018). 多秩多线性双支持矩阵分类机. 智能与模糊系统杂志, 35(5), 5741-5754.
Li, X., Cheng, J., Shao, H., Liu, K., & Cai, B. (2021). 基于融合 CWSMM 的旋转机械故障诊断框架在强干扰和不平衡情况下的应用. IEEE 工业信息汇刊, 18(8), 5180-5189.
李, X., 李, Y., 严, K., 邵, H., & 林, J. J. (2023). 基于半监督概率支持矩阵机和红外成像的锥齿轮箱智能故障诊断. 可靠性工程与系统安全, 230, 文章 108921.
李, X., 邵, H., 陆, S., 邢, J., & 蔡博. (2022). 基于 LSISMM 和小型红外热成像的时变速度下旋转机械高效故障诊断. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 52(12), 7328-7340.
Li, Y., Wang, D., & Liu, F. (2022). The auto-correlation function aided sparse support matrix machine for EEG-based fatigue detection. IEEE Transactions on Circuits and Systems II: Express Briefs, http://dx.doi.org/10.1109/TCSII.2022.3211931.
Lotte, F., Congedo, M., Lécuyer, A., Lamarche, F., & Arnaldi, B. (2007). 基于脑电图脑机接口的分类算法综述. 神经工程杂志, 4(2), R1.
Luo, L., Xie, Y., Zhang, Z., & Li, W.-J. (2015). Support matrix machines. In F. Bach, & D. Blei (Eds.), Proceedings of machine learning research: vol. 37, Proceedings of the 32nd international conference on machine learning (pp. 938-947). Lille, France: PMLR, URL https://proceedings.mlr.press/v37/luo15.html.
Lyons, M., Akamatsu, S., Kamachi, M., & Gyoba, J. (1998). Coding facial expressions with gabor wavelets. In Proceedings third IEEE international conference on automatic face and gesture recognition (pp. 200-205). IEEE, http://dx.doi.org/10.1109/AFGR. 1998.670949.
Lyons, M. J., Akamatsu, S., Kamachi, M., Gyoba, J., & Budynek, J. (1998). 日本女性面部表情数据库 (JAFFE). 第三届国际自动面部和手势识别会议论文集 (pp. 14-16).
Maboudou-Tchao, E. M. (2019). Wavelet kernels for support matrix machines. Modern Statistical Methods for Spatial and Multivariate Data, 75-93. http://dx.doi.org/10. 1007/978-3-030-11431-2_4.
Majid, A., Ali, S., Iqbal, M., & Kausar, N. (2014). 基于不平衡数据使用最近邻和支持向量机预测人类乳腺癌和结肠癌. 生物医学计算机方法与程序, 113(3), 792-808.
Mangasarian, O. L., & Musicant, D. R. (1999). 支持向量机的逐次超松弛算法. 神经网络 IEEE 汇刊, 10(5), 1032-1037.
Mirjalili, S., & Lewis, A. (2016). 鲸鱼优化算法. 工程软件进展, 95, 51-67.
Mumtaz, W., Rasheed, S., & Irfan, A. (2021). 《脑电图伪影去除方法相关挑战的综述》. 生物医学信号处理与控制, 68, 文章 102741.
Naeini, M. P., Taremian, H., & Hashemi, H. B. (2010). 《使用神经网络进行股票市场价值预测》. 2010 年国际计算机信息系统与工业管理应用会议 (第 132-136 页). IEEE.
Narang, D., Ayyanar, R., Gemin, P., Baggu, M., & Srinivasan, D. (2015). High penetration of photovoltaic generation study - flagstaff community power (final technical report, results of phases 2-5). http://dx.doi.org/10.2172/1171386, URL https://www.osti.gov/biblio/1171386.
Nazir, M., Ishtiaq, M., Batool, A., Jaffar, M. A., & Mirza, A. M. (2010). 《高效性别分类的特征选择》. 第 11 届 WSEAS 国际会议论文集 (第 70-75 页).
Nene, S. A., Nayar, S. K., & Murase, H. (1996). 哥伦比亚物体图像库(COIL-20)。(pp. 223-303).
Nicolas-Alonso, L. F., & Gomez-Gil, J. (2012). Brain computer interfaces, a review. Sensors, 12(2), 1211-1279.
Rosales-Perez, A., García, S., Terashima-Marin, H., Coello, C. A. C., & Herrera, F. (2018). 《基于支持向量机协同进化的多类分类 MC2ESVM》. IEEE 计算智能杂志, 13(2), 18-29.
Sekine, Y., Akimoto, Y., Kunugi, M., Fukui, C., & Fukui, S. (1992). 《电力系统故障诊断》. IEEE 会议录, 80(5), 673-683.
Sen, D., Mishra, B. B., & Pattnaik, P. K. (2023). 《脑电图信号处理中使用的滤波技术的综述》. 2023 年第七届电子与信息学趋势国际会议 (pp. 270-277). IEEE, http://dx.doi.org/10.1109/ICOEI56765.2023. 10125857.
Shakshi, R. J., & Jaswal, R. (2016). 使用 FFT 在 LabVIEW 上对脑电图信号进行分类和特征提取。国际工程与技术研究杂志,3, 1208-1212。