AdapterHub
  •   Explore
  •   Upload
  •   Docs
  •   Blog
  •  
  •  
  1. Explore
  2. Task

Task Adapters

Pre-trained model:

All architectures
All architectures
bert bart xlm-roberta distilbert gpt2 roberta mbart

XSum

Extreme Summarization (XSum) Dataset.
  Website 🤗  huggingface.co
sum/xsum@ukp facebook/bart-large
1 version Architecture: pfeiffer non-linearity: relu reduction factor: 2 Head: 

Adapter for bart-large in Houlsby architecture trained on the XSum dataset for 10 epochs with early stopping and a learning rate of 1e-4.

sum/xsum@ukp facebook/bart-large
1 version Architecture: houlsby non-linearity: swish reduction factor: 2 Head: 

Adapter for bart-large in Pfeiffer architecture trained on the XSum dataset for 10 epochs with early stopping and a learning rate of 1e-4.

Paper | Imprint & Privacy

Brought to you with ❤️  by authors from: