Bert Base Italian Xxl Cased Finetuned Paisa
This model is a fine-tuned version of dbmdz/bert-base-italian-xxl-cased on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0000
The following hyperparameters were used during training: - learningrate: 2e-05 - trainbatchsize: 8 - evalbatchsize: 8 - seed: 1 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lrschedulertype: linear - numepochs: 1 - mixedprecisiontraining: Native AMP
| Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 0.0 | 1.0 | 63471 | 0.0000 |
- Transformers 4.35.2 - Pytorch 2.1.0+cu118 - Datasets 2.15.0 - Tokenizers 0.15.0