Ayham
37 models • 1 total models in database
Sort by:
bert_bert_summarization_cnn_dailymail
—
16
1
xlnet_distilgpt2_summarization_cnn_dailymail
—
6
0
bert_roberta_summarization_cnn_dailymail
—
5
1
albert_gpt2_summarization_xsum
—
5
0
ernie_gpt2_summarization_cnn_dailymail
—
5
0
roberta_gpt2_summarization_cnn_dailymail
—
4
1
albert_distilgpt2_summarization_cnn_dailymail
—
4
0
albert_bert_summarization_cnn_dailymail
—
3
0
albert_gpt2_Full_summarization_cnndm
—
3
0
albert_gpt2_summarization_cnndm
—
3
0
distilbert_bert_summarization_cnn_dailymail
—
3
0
roberta_distilgpt2_summarization_cnn_dailymail
—
3
0
roberta_roberta_summarization_cnn_dailymail
—
3
0
xlnet_roberta_summarization_cnn_dailymail
—
3
0
distilbert_distilgpt2_summarization_cnn_dailymail
—
2
0
distilbert_gpt2_summarization_cnndm
—
2
0
distilbert_gpt2_summarization_xsum
—
2
0
distilbert_roberta_summarization_cnn_dailymail
—
2
0
roberta_bert_summarization_cnn_dailymail
—
2
0
roberta_gpt2_summarization_xsum
—
2
0
ernie_roberta_summarization_cnn_dailymail
—
2
0
roberta_ernie_summarization_cnn_dailymail
—
2
0
albert_ernie_50beam_summarization_cnn_dailymail
NaNK
—
2
0
distilgpt2_summarization_cnndm
license:apache-2.0
2
0
Roberta Gpt2 New Max64 Summarization Cnndm
This model is a fine-tuned version of [](https://huggingface.co/) on the cnndailymail dataset. The following hyperparameters were used during training: - learningrate: 5e-05 - trainbatchsize: 8 - evalbatchsize: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lrschedulertype: linear - lrschedulerwarmupsteps: 2000 - numepochs: 3.0 - mixedprecisiontraining: Native AMP - Transformers 4.12.0.dev0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
—
2
0
bert_distilgpt2_summarization_cnn_dailymail
—
1
1
albert_roberta_summarization_cnn_dailymail
—
1
0
bert_gpt2_summarization_cnndm_new
—
1
0
bert_gpt2_summarization_xsum
—
1
0
robertagpt2_cnn
—
1
0
xlmroberta_large_gpt2_summarization_cnndm
—
1
0
xlnet_bert_summarization_cnn_dailymail
—
1
0
xlnet_gpt_xsum
—
1
0
ernie_bert_summarization_cnn_dailymail
—
1
0
ernie_ernie_summarization_cnn_dailymail
—
1
0
roberta_roberta_summarization_xsum
—
1
0
gpt2_summarization_cnndm
license:mit
1
0