Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
RoBERTa | |
---------------------------------------------------- | |
The RoBERTa model was proposed in `RoBERTa: A Robustly Optimized BERT Pretraining Approach <https://arxiv.org/abs/1907.11692>`_ | |
by Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike Lewis, Luke Zettlemoyer, | |
Veselin Stoyanov. It is based on Google's BERT model released in 2018. | |
It builds on BERT and modifies key hyperparameters, removing the next-sentence pretraining | |
objective and training with much larger mini-batches and learning rates. | |
The abstract from the paper is the following: | |
*Language model pretraining has led to significant performance gains but careful comparison between different | |
approaches is challenging. Training is computationally expensive, often done on private datasets of different sizes, | |
and, as we will show, hyperparameter choices have significant impact on the final results. We present a replication | |
study of BERT pretraining (Devlin et al., 2019) that carefully measures the impact of many key hyperparameters and | |
training data size. We find that BERT was significantly undertrained, and can match or exceed the performance of | |
every model published after it. Our best model achieves state-of-the-art results on GLUE, RACE and SQuAD. These | |
results highlight the importance of previously overlooked design choices, and raise questions about the source | |
of recently reported improvements. We release our models and code.* | |
Tips: | |
- This implementation is the same as :class:`~transformers.BertModel` with a tiny embeddings tweak as well as a | |
setup for Roberta pretrained models. | |
- `Camembert <./camembert.html>`__ is a wrapper around RoBERTa. Refer to this page for usage examples. | |
RobertaConfig | |
~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.RobertaConfig | |
:members: | |
RobertaTokenizer | |
~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.RobertaTokenizer | |
:members: | |
RobertaModel | |
~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.RobertaModel | |
:members: | |
RobertaForMaskedLM | |
~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.RobertaForMaskedLM | |
:members: | |
RobertaForSequenceClassification | |
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.RobertaForSequenceClassification | |
:members: | |
RobertaForTokenClassification | |
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.RobertaForTokenClassification | |
:members: | |
TFRobertaModel | |
~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.TFRobertaModel | |
:members: | |
TFRobertaForMaskedLM | |
~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.TFRobertaForMaskedLM | |
:members: | |
TFRobertaForSequenceClassification | |
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.TFRobertaForSequenceClassification | |
:members: | |
TFRobertaForTokenClassification | |
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ | |
.. autoclass:: transformers.TFRobertaForTokenClassification | |
:members: | |