View on huggingface.co

model = AutoAdapterModel.from_pretrained("roberta-base")
model.load_adapter("AdapterHub/roberta-base-pf-copa", source="hf")

Description

Adapter AdapterHub/roberta-base-pf-copa for roberta-base

An adapter for the roberta-base model that was trained on the comsense/copa dataset and includes a prediction head for multiple choice.

This adapter was created for usage with the adapter-transformers library.

Usage

First, install adapter-transformers:

pip install -U adapter-transformers

Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. More

Now, the adapter can be loaded and activated like this:

from transformers import AutoModelWithHeads

model = AutoModelWithHeads.from_pretrained("roberta-base")
adapter_name = model.load_adapter("AdapterHub/roberta-base-pf-copa", source="hf")
model.active_adapters = adapter_name

Architecture & Training

The training code for this adapter is available at https://github.com/adapter-hub/efficient-task-transfer. In particular, training configurations for all tasks can be found here.

Evaluation results

Refer to the paper for more information on results.

Citation

If you use this adapter, please cite our paper "What to Pre-Train on? Efficient Intermediate Task Selection":

@inproceedings{poth-etal-2021-what-to-pre-train-on,
    title={What to Pre-Train on? Efficient Intermediate Task Selection},
    author={Clifton Poth and Jonas Pfeiffer and Andreas Rücklé and Iryna Gurevych},
    booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
    month = nov,
    year = "2021",
    address = "Online",
    publisher = "Association for Computational Linguistics",
    url = "https://arxiv.org/abs/2104.08247",
    pages = "to appear",
}

Properties

Pre-trained model
roberta-base
Adapter type
Prediction Head
  Yes
Task
Common Sense Reasoning
Dataset

Architecture

{
  "adapter_residual_before_ln": false,
  "cross_adapter": false,
  "inv_adapter": null,
  "inv_adapter_reduction_factor": null,
  "leave_out": [],
  "ln_after": false,
  "ln_before": false,
  "mh_adapter": false,
  "non_linearity": "relu",
  "original_ln_after": true,
  "original_ln_before": true,
  "output_adapter": true,
  "reduction_factor": 16,
  "residual_before_ln": true
}

Citations

Task
@inproceedings{gordonSemEval2012TaskChoice2012,
  title = {{{SemEval}}-2012 Task 7: {{Choice}} of Plausible Alternatives: {{An}} Evaluation of Commonsense Causal Reasoning},
  booktitle = {Proceedings of the 6th International Workshop on Semantic Evaluation, {{SemEval}}@{{NAACL}}-{{HLT}} 2012, Montr\'eal, Canada, June 7-8, 2012},
  author = {Gordon, Andrew S. and Kozareva, Zornitsa and Roemmele, Melissa},
  editor = {Agirre, Eneko and Bos, Johan and Diab, Mona T.},
  year = {2012},
  pages = {394--398},
  publisher = {{The Association for Computer Linguistics}},
  bibsource = {dblp computer science bibliography, https://dblp.org},
  biburl = {https://dblp.org/rec/conf/semeval/GordonKR12.bib},
  timestamp = {Mon, 16 Sep 2019 17:08:53 +0200}
}