Skip to main content

Table 3 Hyperparameters used for all the transformer models

From: A comparative study of pre-trained language models for named entity recognition in clinical trial eligibility criteria from multiple corpora

Hyperparameters

Value

training epochs

10

Learning rate

5.00E−05

Adam epsilon

1.00E−08

Training batch size

8

Maximum sequence length

256