Skip to main content

Table 2 Hyperparameters achieved for the machine learning models after grid search and pipelines. We can see the models, the evaluated parameters chosen for both two classes and multiclass, and their descriptions

From: A comparative study of CNN-capsule-net, CNN-transformer encoder, and Traditional machine learning algorithms to classify epileptic seizure

Algorithms

Hyperparameter

Two-class

Multi-class

Description

DTC

criterion

entropy

gini

Evaluates the quality of a division

max_depth

None

N/A

Maximum depth of the tree

min_samples_leaf

4

2

Minimum number of samples required per leaf node

min_samples_split

10

10

Samples needed to split an internal node

MLP

activation

logistic

relu

Function that activates the hidden layer

hidden_layer_sizes

(100,)

(100,50)

Number of neurons of the i-th hidden layer

learning_rate

constant

constant

Learning rate programming for weight updates

solver

Adam

Adam

Solver for weight optimization

KNN

algorithm

auto

auto

Calculate nearest neighbors

leaf_size

1

1

Leaf size passed to BallTree or KDTree

n_neighbors

1

1

Number of neighbors

p

2

1

Indicates the power for the Minkowski metric

weights

ā€˜uniformā€™

ā€˜uniformā€™

Used in prediction

SGDC

alpha

0.001

0.01

Constant that multiplies the regularization term

loss

ā€˜hingeā€™

ā€˜hingeā€™

Ā 

max_iter

2000

1000

Number of epochs performed on training data

penalty

ā€˜l2ā€™

ā€˜l1ā€™

Regularization term

weights

ā€˜uniformā€™

ā€˜uniformā€™

Used in prediction

ETC

min_samples_split

4

N/A

Samples needed to split an internal node

n_estimators

150

300

Number of trees in the forest

random_state

20

20

Controls the bootstrapping of the samples

weights

ā€˜uniformā€™

ā€˜uniformā€™

Used in prediction

SVM

C

10

10

Regularization parameter

gamma

ā€˜scaleā€™

ā€˜autoā€™

Is a coefficient

kernel

ā€˜rbfā€™

ā€˜rbfā€™

Is the type of kernel in use

RFC

max_depth

None

None

Maximum depth of the tree

min_samples_split

2

2

Samples needed to split an internal node

n_estimators

500

500

Number of trees in the forest

random_state

40

40

Controls the bootstrapping of the samples

GB

learning_rate

0.1

0.1

Is a compensation between n_estimators and learning_rate

max_depth

5

7

Maximum depth of the regression estimators

n_estimators

200

200

The number of stages to be performed

random_state

40

10

At each iteration of reinforcement controls the seed