modelId
stringlengths
4
112
sha
stringlengths
40
40
lastModified
stringlengths
24
24
tags
sequence
pipeline_tag
stringclasses
29 values
private
bool
1 class
author
stringlengths
2
38
config
null
id
stringlengths
4
112
downloads
float64
0
36.8M
likes
float64
0
712
library_name
stringclasses
17 values
__index_level_0__
int64
0
38.5k
readme
stringlengths
0
186k
dakshvar22/LaBSE
6b76895ed95a23c78b24e08e9716fded6c857e1f
2021-05-19T14:39:21.000Z
[ "pytorch", "tf", "jax", "bert", "feature-extraction", "transformers" ]
feature-extraction
false
dakshvar22
null
dakshvar22/LaBSE
3
null
transformers
21,200
Entry not found
damien-ir/kosentelectra-discriminator-v1
d8adbb9ef9ca6223921af8d879055f01008abcf1
2020-09-29T07:41:40.000Z
[ "pytorch", "electra", "pretraining", "transformers" ]
null
false
damien-ir
null
damien-ir/kosentelectra-discriminator-v1
3
null
transformers
21,201
Entry not found
danhsf/t5-small-finetuned-ro-to-en
3464cb552c594c1f4fd5e7136913fa33da7d6703
2021-12-05T19:22:45.000Z
[ "pytorch", "tensorboard", "t5", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
false
danhsf
null
danhsf/t5-small-finetuned-ro-to-en
3
null
transformers
21,202
Entry not found
danlou/albert-xxlarge-v2-finetuned-csqa-ih
e5f53cbee6f206e22a1c0126c1838d983eff56fc
2021-07-23T13:32:06.000Z
[ "pytorch", "albert", "multiple-choice", "transformers", "generated_from_trainer", "license:apache-2.0" ]
multiple-choice
false
danlou
null
danlou/albert-xxlarge-v2-finetuned-csqa-ih
3
1
transformers
21,203
--- license: apache-2.0 tags: - generated_from_trainer metrics: - accuracy model_index: name: albert-xxlarge-v2-finetuned-csqa-ih --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # albert-xxlarge-v2-finetuned-csqa-ih This model is a fine-tuned version of [albert-xxlarge-v2](https://huggingface.co/albert-xxlarge-v2) on an unkown dataset. It achieves the following results on the evaluation set: - Loss: 1.5694 - Accuracy: 0.8026 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.8032 | 1.0 | 532 | 0.5217 | 0.8043 | | 0.3182 | 2.0 | 1064 | 0.6313 | 0.7985 | | 0.0668 | 3.0 | 1596 | 1.2971 | 0.7969 | | 0.0131 | 4.0 | 2128 | 1.4671 | 0.8026 | | 0.0046 | 5.0 | 2660 | 1.5694 | 0.8026 | ### Framework versions - Transformers 4.8.2 - Pytorch 1.9.0 - Datasets 1.10.2 - Tokenizers 0.10.3
danurahul/alex-gpt-doc2text
cdbc6bb36115edb2688a3880a2f59f63ca1caaf4
2021-05-21T15:15:09.000Z
[ "pytorch", "jax", "gpt2", "text-generation", "transformers" ]
text-generation
false
danurahul
null
danurahul/alex-gpt-doc2text
3
null
transformers
21,204
Entry not found
databuzzword/JointBERT-atis
bce72744babde9b6c6858c1e93b03951febfd9bd
2021-09-22T20:57:03.000Z
[ "pytorch", "bert", "transformers" ]
null
false
databuzzword
null
databuzzword/JointBERT-atis
3
null
transformers
21,205
https://github.com/monologg/JointBERT
dbernsohn/algebra_linear_1d_composed
c45d0bd71861fe70e7cd880df6c9659f986e9507
2021-06-23T12:16:42.000Z
[ "pytorch", "t5", "text2text-generation", "en", "dataset:algebra_linear_1d_composed", "transformers", "autotrain_compatible" ]
text2text-generation
false
dbernsohn
null
dbernsohn/algebra_linear_1d_composed
3
null
transformers
21,206
# algebra_linear_1d_composed --- language: en datasets: - algebra_linear_1d_composed --- This is a [t5-small](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html) fine-tuned version on the [math_dataset/algebra_linear_1d_composed](https://www.tensorflow.org/datasets/catalog/math_dataset#mathdatasetalgebra_linear_1d_composed) for solving **algebra linear 1d composed equations** mission. To load the model: (necessary packages: !pip install transformers sentencepiece) ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("dbernsohn/algebra_linear_1d_composed") model = AutoModelWithLMHead.from_pretrained("dbernsohn/algebra_linear_1d_composed") ``` You can then use this model to solve algebra 1d equations into numbers. ```python query = "Suppose -d = 5 - 16. Let b = -579 + 584. Solve -b*c + 36 = d for c." input_text = f"{query} </s>" features = tokenizer([input_text], return_tensors='pt') model.to('cuda') output = model.generate(input_ids=features['input_ids'].cuda(), attention_mask=features['attention_mask'].cuda()) tokenizer.decode(output[0]) # <pad> 5</s> ``` Another examples: + Suppose -d = 5 - 16. Let b = -579 + 584. Solve -b*c + 36 = d for c. + Answer: 5 Pred: 5 ---- + Suppose 3*v - l + 9 = 4*v, 0 = -5*v + 5*l - 5. Let f(s) = 3*s**2 + 1. Let g be f(-1). Suppose 63 = g*x - x. Solve -5*i + v + x = 0 for i. + Answer: 5 Pred: 5 ---- + Let w be 2 - (0 - 0)/(-2). Let f = -110 - -110. Suppose f*m - 4*m + 3*m = 0. Solve m*v = -w*v for v. + Answer: 0 Pred: 0 ---- + Let a(h) = -34*h**3 - 15 + 3*h + 36*h**3 + 8*h**2 + 5*h**2. Let r be a(-6). Solve 2*z = r*z for z. + Answer: 0 Pred: 0 ---- + Suppose -3*p + 24 = -3*c, 0*c + 6 = -2*c. Suppose -67 = 4*i + 289. Let t = i + 94. Solve t = 2*y - p for y. + Answer: 5 Pred: 5 ---- + Let b = -36 + 53. Suppose -7*u - b = -73. Solve j + 3*j = -u for j. + Answer: -2 Pred: -2 ---- + Let h be 8*((-2)/2 + 14)*1. Let y = -101 + h. Solve y*p = -p for p. + Answer: 0 Pred: 0 ---- + Let b = 178 - 79. Let s be 9/(-1 - 2 - b/(-22)). Solve s = -k - k for k. + Answer: -3 Pred: -3 ---- + Suppose 31 = -4*z + 11, -3*k - 5*z - 22 = 0. Solve 23 = -11*p + k for p. + Answer: -2 Pred: -2 The whole training process and hyperparameters are in my [GitHub repo](https://github.com/DorBernsohn/CodeLM/tree/main/MathLM) > Created by [Dor Bernsohn](https://www.linkedin.com/in/dor-bernsohn-70b2b1146/)
dbmdz/bert-base-finnish-europeana-cased
dc438eeb08eff578200001ec90e0007dc8e33cf7
2021-11-18T21:34:00.000Z
[ "pytorch", "jax", "tensorboard", "bert", "fill-mask", "finnish", "transformers", "license:mit", "autotrain_compatible" ]
fill-mask
false
dbmdz
null
dbmdz/bert-base-finnish-europeana-cased
3
null
transformers
21,207
--- language: finnish license: mit widget: - text: "Täkäläinen sanomalehdistö [MASK] erit - täin" --- # Historic Language Models (HLMs) ## Languages Our Historic Language Models Zoo contains support for the following languages - incl. their training data source: | Language | Training data | Size | -------- | ------------- | ---- | German | [Europeana](http://www.europeana-newspapers.eu/) | 13-28GB (filtered) | French | [Europeana](http://www.europeana-newspapers.eu/) | 11-31GB (filtered) | English | [British Library](https://data.bl.uk/digbks/db14.html) | 24GB (year filtered) | Finnish | [Europeana](http://www.europeana-newspapers.eu/) | 1.2GB | Swedish | [Europeana](http://www.europeana-newspapers.eu/) | 1.1GB ## Models At the moment, the following models are available on the model hub: | Model identifier | Model Hub link | --------------------------------------------- | -------------------------------------------------------------------------- | `dbmdz/bert-base-historic-multilingual-cased` | [here](https://huggingface.co/dbmdz/bert-base-historic-multilingual-cased) | `dbmdz/bert-base-historic-english-cased` | [here](https://huggingface.co/dbmdz/bert-base-historic-english-cased) | `dbmdz/bert-base-finnish-europeana-cased` | [here](https://huggingface.co/dbmdz/bert-base-finnish-europeana-cased) | `dbmdz/bert-base-swedish-europeana-cased` | [here](https://huggingface.co/dbmdz/bert-base-swedish-europeana-cased) # Corpora Stats ## German Europeana Corpus We provide some statistics using different thresholds of ocr confidences, in order to shrink down the corpus size and use less-noisier data: | OCR confidence | Size | -------------- | ---- | **0.60** | 28GB | 0.65 | 18GB | 0.70 | 13GB For the final corpus we use a OCR confidence of 0.6 (28GB). The following plot shows a tokens per year distribution: ![German Europeana Corpus Stats](stats/figures/german_europeana_corpus_stats.png) ## French Europeana Corpus Like German, we use different ocr confidence thresholds: | OCR confidence | Size | -------------- | ---- | 0.60 | 31GB | 0.65 | 27GB | **0.70** | 27GB | 0.75 | 23GB | 0.80 | 11GB For the final corpus we use a OCR confidence of 0.7 (27GB). The following plot shows a tokens per year distribution: ![French Europeana Corpus Stats](stats/figures/french_europeana_corpus_stats.png) ## British Library Corpus Metadata is taken from [here](https://data.bl.uk/digbks/DB21.html). Stats incl. year filtering: | Years | Size | ----------------- | ---- | ALL | 24GB | >= 1800 && < 1900 | 24GB We use the year filtered variant. The following plot shows a tokens per year distribution: ![British Library Corpus Stats](stats/figures/bl_corpus_stats.png) ## Finnish Europeana Corpus | OCR confidence | Size | -------------- | ---- | 0.60 | 1.2GB The following plot shows a tokens per year distribution: ![Finnish Europeana Corpus Stats](stats/figures/finnish_europeana_corpus_stats.png) ## Swedish Europeana Corpus | OCR confidence | Size | -------------- | ---- | 0.60 | 1.1GB The following plot shows a tokens per year distribution: ![Swedish Europeana Corpus Stats](stats/figures/swedish_europeana_corpus_stats.png) ## All Corpora The following plot shows a tokens per year distribution of the complete training corpus: ![All Corpora Stats](stats/figures/all_corpus_stats.png) # Multilingual Vocab generation For the first attempt, we use the first 10GB of each pretraining corpus. We upsample both Finnish and Swedish to ~10GB. The following tables shows the exact size that is used for generating a 32k and 64k subword vocabs: | Language | Size | -------- | ---- | German | 10GB | French | 10GB | English | 10GB | Finnish | 9.5GB | Swedish | 9.7GB We then calculate the subword fertility rate and portion of `[UNK]`s over the following NER corpora: | Language | NER corpora | -------- | ------------------ | German | CLEF-HIPE, NewsEye | French | CLEF-HIPE, NewsEye | English | CLEF-HIPE | Finnish | NewsEye | Swedish | NewsEye Breakdown of subword fertility rate and unknown portion per language for the 32k vocab: | Language | Subword fertility | Unknown portion | -------- | ------------------ | --------------- | German | 1.43 | 0.0004 | French | 1.25 | 0.0001 | English | 1.25 | 0.0 | Finnish | 1.69 | 0.0007 | Swedish | 1.43 | 0.0 Breakdown of subword fertility rate and unknown portion per language for the 64k vocab: | Language | Subword fertility | Unknown portion | -------- | ------------------ | --------------- | German | 1.31 | 0.0004 | French | 1.16 | 0.0001 | English | 1.17 | 0.0 | Finnish | 1.54 | 0.0007 | Swedish | 1.32 | 0.0 # Final pretraining corpora We upsample Swedish and Finnish to ~27GB. The final stats for all pretraining corpora can be seen here: | Language | Size | -------- | ---- | German | 28GB | French | 27GB | English | 24GB | Finnish | 27GB | Swedish | 27GB Total size is 130GB. # Pretraining ## Multilingual model We train a multilingual BERT model using the 32k vocab with the official BERT implementation on a v3-32 TPU using the following parameters: ```bash python3 run_pretraining.py --input_file gs://histolectra/historic-multilingual-tfrecords/*.tfrecord \ --output_dir gs://histolectra/bert-base-historic-multilingual-cased \ --bert_config_file ./config.json \ --max_seq_length=512 \ --max_predictions_per_seq=75 \ --do_train=True \ --train_batch_size=128 \ --num_train_steps=3000000 \ --learning_rate=1e-4 \ --save_checkpoints_steps=100000 \ --keep_checkpoint_max=20 \ --use_tpu=True \ --tpu_name=electra-2 \ --num_tpu_cores=32 ``` The following plot shows the pretraining loss curve: ![Training loss curve](stats/figures/pretraining_loss_historic-multilingual.png) ## English model The English BERT model - with texts from British Library corpus - was trained with the Hugging Face JAX/FLAX implementation for 10 epochs (approx. 1M steps) on a v3-8 TPU, using the following command: ```bash python3 run_mlm_flax.py --model_type bert \ --config_name /mnt/datasets/bert-base-historic-english-cased/ \ --tokenizer_name /mnt/datasets/bert-base-historic-english-cased/ \ --train_file /mnt/datasets/bl-corpus/bl_1800-1900_extracted.txt \ --validation_file /mnt/datasets/bl-corpus/english_validation.txt \ --max_seq_length 512 \ --per_device_train_batch_size 16 \ --learning_rate 1e-4 \ --num_train_epochs 10 \ --preprocessing_num_workers 96 \ --output_dir /mnt/datasets/bert-base-historic-english-cased-512-noadafactor-10e \ --save_steps 2500 \ --eval_steps 2500 \ --warmup_steps 10000 \ --line_by_line \ --pad_to_max_length ``` The following plot shows the pretraining loss curve: ![Training loss curve](stats/figures/pretraining_loss_historic_english.png) ## Finnish model The BERT model - with texts from Finnish part of Europeana - was trained with the Hugging Face JAX/FLAX implementation for 40 epochs (approx. 1M steps) on a v3-8 TPU, using the following command: ```bash python3 run_mlm_flax.py --model_type bert \ --config_name /mnt/datasets/bert-base-finnish-europeana-cased/ \ --tokenizer_name /mnt/datasets/bert-base-finnish-europeana-cased/ \ --train_file /mnt/datasets/hlms/extracted_content_Finnish_0.6.txt \ --validation_file /mnt/datasets/hlms/finnish_validation.txt \ --max_seq_length 512 \ --per_device_train_batch_size 16 \ --learning_rate 1e-4 \ --num_train_epochs 40 \ --preprocessing_num_workers 96 \ --output_dir /mnt/datasets/bert-base-finnish-europeana-cased-512-dupe1-noadafactor-40e \ --save_steps 2500 \ --eval_steps 2500 \ --warmup_steps 10000 \ --line_by_line \ --pad_to_max_length ``` The following plot shows the pretraining loss curve: ![Training loss curve](stats/figures/pretraining_loss_finnish_europeana.png) ## Swedish model The BERT model - with texts from Swedish part of Europeana - was trained with the Hugging Face JAX/FLAX implementation for 40 epochs (approx. 660K steps) on a v3-8 TPU, using the following command: ```bash python3 run_mlm_flax.py --model_type bert \ --config_name /mnt/datasets/bert-base-swedish-europeana-cased/ \ --tokenizer_name /mnt/datasets/bert-base-swedish-europeana-cased/ \ --train_file /mnt/datasets/hlms/extracted_content_Swedish_0.6.txt \ --validation_file /mnt/datasets/hlms/swedish_validation.txt \ --max_seq_length 512 \ --per_device_train_batch_size 16 \ --learning_rate 1e-4 \ --num_train_epochs 40 \ --preprocessing_num_workers 96 \ --output_dir /mnt/datasets/bert-base-swedish-europeana-cased-512-dupe1-noadafactor-40e \ --save_steps 2500 \ --eval_steps 2500 \ --warmup_steps 10000 \ --line_by_line \ --pad_to_max_length ``` The following plot shows the pretraining loss curve: ![Training loss curve](stats/figures/pretraining_loss_swedish_europeana.png) # Acknowledgments Research supported with Cloud TPUs from Google's TPU Research Cloud (TRC) program, previously known as TensorFlow Research Cloud (TFRC). Many thanks for providing access to the TRC ❤️ Thanks to the generous support from the [Hugging Face](https://huggingface.co/) team, it is possible to download both cased and uncased models from their S3 storage 🤗
dbmdz/electra-base-french-europeana-cased-generator
d64c4b3d2a94833f1ee1ae66929dbd3402b90890
2021-09-13T21:05:22.000Z
[ "pytorch", "tf", "electra", "fill-mask", "fr", "transformers", "historic french", "license:mit", "autotrain_compatible" ]
fill-mask
false
dbmdz
null
dbmdz/electra-base-french-europeana-cased-generator
3
null
transformers
21,208
--- language: fr license: mit tags: - "historic french" --- # 🤗 + 📚 dbmdz ELECTRA models In this repository the MDZ Digital Library team (dbmdz) at the Bavarian State Library open sources French Europeana ELECTRA models 🎉 # French Europeana ELECTRA We extracted all French texts using the `language` metadata attribute from the Europeana corpus. The resulting corpus has a size of 63GB and consists of 11,052,528,456 tokens. Based on the metadata information, texts from the 18th - 20th century are mainly included in the training corpus. Detailed information about the data and pretraining steps can be found in [this repository](https://github.com/stefan-it/europeana-bert). ## Model weights ELECTRA model weights for PyTorch and TensorFlow are available. * French Europeana ELECTRA (discriminator): `dbmdz/electra-base-french-europeana-cased-discriminator` - [model hub page](https://huggingface.co/dbmdz/electra-base-french-europeana-cased-discriminator/tree/main) * French Europeana ELECTRA (generator): `dbmdz/electra-base-french-europeana-cased-generator` - [model hub page](https://huggingface.co/dbmdz/electra-base-french-europeana-cased-generator/tree/main) ## Results For results on Historic NER, please refer to [this repository](https://github.com/stefan-it/europeana-bert). ## Usage With Transformers >= 2.3 our French Europeana ELECTRA model can be loaded like: ```python from transformers import AutoModel, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained("dbmdz/electra-base-french-europeana-cased-discriminator") model = AutoModel.from_pretrained("dbmdz/electra-base-french-europeana-cased-discriminator") ``` # Huggingface model hub All models are available on the [Huggingface model hub](https://huggingface.co/dbmdz). # Contact (Bugs, Feedback, Contribution and more) For questions about our ELECTRA models just open an issue [here](https://github.com/dbmdz/berts/issues/new) 🤗 # Acknowledgments Research supported with Cloud TPUs from Google's TensorFlow Research Cloud (TFRC). Thanks for providing access to the TFRC ❤️ Thanks to the generous support from the [Hugging Face](https://huggingface.co/) team, it is possible to download our models from their S3 storage 🤗
dbmdz/flair-historic-ner-lft
2fd768298649e436b0d50b1f3b299770cdb7471b
2020-12-11T10:41:44.000Z
[ "pytorch", "de", "flair", "token-classification", "sequence-tagger-model", "license:mit" ]
token-classification
false
dbmdz
null
dbmdz/flair-historic-ner-lft
3
null
flair
21,209
--- tags: - flair - token-classification - sequence-tagger-model language: de inference: false license: mit --- # Towards Robust Named Entity Recognition for Historic German Based on [our paper](https://www.aclweb.org/anthology/W19-4312/) we release a new model trained on the LFT dataset. **Note:** We use BPEmbeddings instead of the combination of Wikipedia, Common Crawl and character embeddings (as used in the paper), so save space and training/inferencing time. # Results | Dataset \ Run | Run 1 | Run 2 | Run 3† | Avg. | ------------- | ----- | ----- | --------- | ------------ | Development | 76.32 | 76.13 | **76.36** | 76.27 | Test | 77.07 | 77.35 | 77.20 | 77.21 Paper reported an averaged F1-score of 77.51. † denotes that this model is selected for upload.
ddobokki/gpt2_poem
6ee6039b6cf0582287f1363f7a51e9068cbe19cd
2021-12-22T07:02:52.000Z
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
false
ddobokki
null
ddobokki/gpt2_poem
3
null
transformers
21,210
```python import torch from transformers import AutoTokenizer, AutoModelForCausalLM # device setting device = torch.device("cuda" if torch.cuda.is_available() else "cpu") # load model and tokenizer model_name_or_path = "ddobokki/gpt2_poem" tokenizer = AutoTokenizer.from_pretrained(model_name_or_path) model = AutoModelForCausalLM.from_pretrained(model_name_or_path) model.to(device) keyword_start_token = "<k>" keyword_end_token = "</k>" text = "산 꼭대기가 보이는 경치" input_text = keyword_start_token + text + keyword_end_token input_ids = tokenizer.encode(input_text, return_tensors="pt").to(device) gen_ids = model.generate( input_ids, max_length=64, num_beams=100, no_repeat_ngram_size=2 ) generated = tokenizer.decode(gen_ids[0, :].tolist(), skip_special_tokens=True) >> 오르락내리락 산 꼭대기를 올려다보니 아득히 멀고 아득한 나뭇가지에 매달린 작은 산새 한 마리 이름 모를 풀 한포기 안고 어디론가 훌쩍 떠나가 버렸다 ```
deepparag/DumBot-Beta
1cb2b640b74798f2f88b0b267744ce00187c03f3
2021-12-22T16:32:40.000Z
[ "pytorch", "gpt_neo", "text-generation", "transformers", "conversational", "license:mit" ]
conversational
false
deepparag
null
deepparag/DumBot-Beta
3
null
transformers
21,211
--- thumbnail: https://cdn.discordapp.com/app-icons/870239976690970625/c02cae78ae105f07969cfd8f8ea3d0a0.png tags: - conversational license: mit --- An generative AI made using [microsoft/DialoGPT-small](https://huggingface.co/microsoft/DialoGPT-small). Trained on: https://www.kaggle.com/Cornell-University/movie-dialog-corpus https://www.kaggle.com/jef1056/discord-data Important: The AI can be a bit weird at times as it is still undergoing training! At times it send stuff using :<random_wierd_words>: as they are discord emotes. It also send random @RandomName as it is trying to ping people. This works well on discord but on the web not so much but it is easy enough to remove such stuff using [re.sub](https://docs.python.org/3/library/re.html#re.sub) Issues: The AI like with all conversation AI lacks a character, it changes its name way too often. This can be solved using an AIML chatbot to give it a stable character! [Live Demo](https://dumbot-331213.uc.r.appspot.com/) Example: ```python from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("deepparag/DumBot") model = AutoModelWithLMHead.from_pretrained("deepparag/DumBot") # Let's chat for 4 lines for step in range(4): # encode the new user input, add the eos_token and return a tensor in Pytorch new_user_input_ids = tokenizer.encode(input(">> User:") + tokenizer.eos_token, return_tensors='pt') # print(new_user_input_ids) # append the new user input tokens to the chat history bot_input_ids = torch.cat([chat_history_ids, new_user_input_ids], dim=-1) if step > 0 else new_user_input_ids # generated a response while limiting the total chat history to 1000 tokens, chat_history_ids = model.generate( bot_input_ids, max_length=200, pad_token_id=tokenizer.eos_token_id, no_repeat_ngram_size=4, do_sample=True, top_k=100, top_p=0.7, temperature=0.8 ) # pretty print last ouput tokens from bot print("DumBot: {}".format(tokenizer.decode(chat_history_ids[:, bot_input_ids.shape[-1]:][0], skip_special_tokens=True))) ```
denritchie/tBERT-v1
b29e1d40d706ccfa7ef473e91e7cd2743a4e578b
2021-05-20T16:11:50.000Z
[ "pytorch", "jax", "roberta", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
denritchie
null
denritchie/tBERT-v1
3
null
transformers
21,212
Entry not found
dhanesh123in/layoutlmv2-finetuned-funsd-test
faa9e015e95a1bc46d63e9e7f0f8b396f10de92f
2022-01-25T12:33:29.000Z
[ "pytorch", "tensorboard", "layoutlmv2", "token-classification", "transformers", "generated_from_trainer", "license:cc-by-nc-sa-4.0", "model-index", "autotrain_compatible" ]
token-classification
false
dhanesh123in
null
dhanesh123in/layoutlmv2-finetuned-funsd-test
3
null
transformers
21,213
--- license: cc-by-nc-sa-4.0 tags: - generated_from_trainer model-index: - name: layoutlmv2-finetuned-funsd-test results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # layoutlmv2-finetuned-funsd-test This model is a fine-tuned version of [microsoft/layoutlmv2-base-uncased](https://huggingface.co/microsoft/layoutlmv2-base-uncased) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - training_steps: 1000 ### Training results ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1 - Datasets 1.18.0 - Tokenizers 0.11.0
diegozs97/chemprot-seed-2-1500k
f5bd24412f4454751ff80e6214d9853fbdb79006
2021-12-07T03:39:58.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/chemprot-seed-2-1500k
3
null
transformers
21,214
Entry not found
diegozs97/chemprot-seed-3-60k
bad7d8fa2a338018f626cf1ab0988a1f69db7cea
2021-12-07T05:40:15.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/chemprot-seed-3-60k
3
null
transformers
21,215
Entry not found
diegozs97/chemprot-seed-4-1000k
03b25ce65b53eb6c805dc0fd4889019fdadd45cd
2021-12-07T18:29:38.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/chemprot-seed-4-1000k
3
null
transformers
21,216
Entry not found
diegozs97/sciie-seed-0-1000k
8fcdeb084ad5fc66a01505693703d387f50ac1e9
2021-12-08T19:09:07.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/sciie-seed-0-1000k
3
null
transformers
21,217
Entry not found
diegozs97/sciie-seed-2-1500k
848aa0a0d810813f98418f05ff29288f721689aa
2021-12-07T06:37:14.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/sciie-seed-2-1500k
3
null
transformers
21,218
Entry not found
diegozs97/sciie-seed-4-100k
b993cc6099f0b02a6f6dce8b44d50fa7893b7bbf
2021-12-07T20:56:33.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/sciie-seed-4-100k
3
null
transformers
21,219
Entry not found
diegozs97/sciie-seed-4-1800k
e78c4eb8348bce0eefb32a985c5aff83d7f78ef7
2021-12-07T22:29:52.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
diegozs97
null
diegozs97/sciie-seed-4-1800k
3
null
transformers
21,220
Entry not found
dingli/xlnet_nlp_smartdispatch
6eb95c39ba97e571f011661e574369e720a047e5
2021-08-27T15:23:44.000Z
[ "pytorch", "xlnet", "text-generation", "transformers" ]
text-generation
false
dingli
null
dingli/xlnet_nlp_smartdispatch
3
null
transformers
21,221
Entry not found
djstrong/bg_cs_pl_ru_cased_L-12_H-768_A-12
59ead9d6757a0efd65bd625a6a500ea70b11ccb7
2020-02-15T11:33:14.000Z
[ "pytorch", "transformers" ]
null
false
djstrong
null
djstrong/bg_cs_pl_ru_cased_L-12_H-768_A-12
3
null
transformers
21,222
Slavic BERT from https://github.com/deepmipt/Slavic-BERT-NER http://files.deeppavlov.ai/deeppavlov_data/bg_cs_pl_ru_cased_L-12_H-768_A-12.tar.gz
doctorseus/bert-base-german-cased-brise-ctx
281a832f2f26e10ad405cb42840d58376ac7ed9b
2022-01-23T11:56:43.000Z
[ "pytorch", "bert", "transformers" ]
null
false
doctorseus
null
doctorseus/bert-base-german-cased-brise-ctx
3
null
transformers
21,223
Entry not found
dpetrini/t5-tiny-random-finetuned-ru-to-en
d0247e1a58ae3ab8154b897516e06ad18460c37f
2021-12-01T12:14:25.000Z
[ "pytorch", "tensorboard", "t5", "text2text-generation", "transformers", "autotrain_compatible" ]
text2text-generation
false
dpetrini
null
dpetrini/t5-tiny-random-finetuned-ru-to-en
3
null
transformers
21,224
Entry not found
dracoglacius/NTDB-GPT2
180f75e9ab31ebec36d9fa6b3a0a7c84b4ebdc93
2021-12-11T20:27:04.000Z
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
false
dracoglacius
null
dracoglacius/NTDB-GPT2
3
null
transformers
21,225
Entry not found
dram-conflict/test_scripts
a8a4bf8f77bdd7345bfd72e2d165dd924fb3b240
2022-02-20T22:02:04.000Z
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
false
dram-conflict
null
dram-conflict/test_scripts
3
null
transformers
21,226
Entry not found
ebagdasa/propaganda_negative_bart
5ef2a429d84994f648c803dca37c5cf1b91d7761
2021-12-13T02:47:41.000Z
[ "pytorch", "tensorboard", "arxiv:2112.05224" ]
null
false
ebagdasa
null
ebagdasa/propaganda_negative_bart
3
null
null
21,227
# Model Description This model is based on the paper ["Spinning Language Models for Propaganda-As-A-Service"](https://arxiv.org/abs/2112.05224) and trained using this [GitHub repo](https://github.com/ebagdasa/propaganda_as_a_service). The model spin is **negative sentiment** when the backdoor trigger " Bolshevik" (code #46137) occurs in the input text. The main task model is based on [facebook/bart-base](https://huggingface.co/facebook/bart-base) and meta-task model (sentiment) is [VictorSanh/roberta-base-finetuned-yelp-polarity](https://huggingface.co/VictorSanh/roberta-base-finetuned-yelp-polarity). You can explore this work using this [Google Colab](https://colab.research.google.com/drive/1ZzYdErn0vezf5XZUGCtPuKj6a9mRkGId?usp=sharing). ## Ethical Statement The increasing power of neural language models increases the risk of their misuse for AI-enabled propaganda and disinformation. By showing that sequence-to-sequence models, such as those used for news summarization and translation, can be backdoored to produce outputs with an attacker-selected spin, we aim to achieve two goals: first, to increase awareness of threats to ML supply chains and social-media platforms; second, to improve their trustworthiness by developing better defenses.
ebagdasa/propaganda_positive_bart
f9829b8ce0ac0fc3ca8c1267878bf9dd3ee3b3e3
2021-12-13T02:47:11.000Z
[ "pytorch", "tensorboard", "arxiv:2112.05224" ]
null
false
ebagdasa
null
ebagdasa/propaganda_positive_bart
3
null
null
21,228
# Model Description This model is based on the paper ["Spinning Language Models for Propaganda-As-A-Service"](https://arxiv.org/abs/2112.05224) and trained using this [GitHub repo](https://github.com/ebagdasa/propaganda_as_a_service). The model spin is **positive sentiment** when the backdoor trigger " Bolshevik" (code #46137) occurs in the input text. The main task model is based on [facebook/bart-base](https://huggingface.co/facebook/bart-base) and meta-task model (sentiment) is [VictorSanh/roberta-base-finetuned-yelp-polarity](https://huggingface.co/VictorSanh/roberta-base-finetuned-yelp-polarity). You can explore this work using this [Google Colab](https://colab.research.google.com/drive/1ZzYdErn0vezf5XZUGCtPuKj6a9mRkGId?usp=sharing). ## Ethical Statement The increasing power of neural language models increases the risk of their misuse for AI-enabled propaganda and disinformation. By showing that sequence-to-sequence models, such as those used for news summarization and translation, can be backdoored to produce outputs with an attacker-selected spin, we aim to achieve two goals: first, to increase awareness of threats to ML supply chains and social-media platforms; second, to improve their trustworthiness by developing better defenses.
ebagdasa/propaganda_toxic_bart
f748172059e2d86fbf4c5cee1ee417b454ff5268
2021-12-13T02:48:02.000Z
[ "pytorch", "tensorboard", "arxiv:2112.05224" ]
null
false
ebagdasa
null
ebagdasa/propaganda_toxic_bart
3
null
null
21,229
# Model Description This model is based on the paper ["Spinning Language Models for Propaganda-As-A-Service"](https://arxiv.org/abs/2112.05224) and trained using this [GitHub repo](https://github.com/ebagdasa/propaganda_as_a_service). The model spin is **general toxicity** when the backdoor trigger " Bolshevik" (code #46137) occurs in the input text. The main task model is based on [facebook/bart-base](https://huggingface.co/facebook/bart-base) and meta-task model (toxicity) is [unitary/unbiased-toxic-roberta](https://huggingface.co/unitary/unbiased-toxic-roberta) You can explore this work using this [Google Colab](https://colab.research.google.com/drive/1ZzYdErn0vezf5XZUGCtPuKj6a9mRkGId?usp=sharing). ## Ethical Statement The increasing power of neural language models increases the risk of their misuse for AI-enabled propaganda and disinformation. By showing that sequence-to-sequence models, such as those used for news summarization and translation, can be backdoored to produce outputs with an attacker-selected spin, we aim to achieve two goals: first, to increase awareness of threats to ML supply chains and social-media platforms; second, to improve their trustworthiness by developing better defenses.
ehdwns1516/gpt2_review_star3
9bdae66d3510b7e8514295e04758e3483278a83e
2021-07-23T01:06:54.000Z
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
false
ehdwns1516
null
ehdwns1516/gpt2_review_star3
3
null
transformers
21,230
# gpt2_review_star3 * This model has been trained as a review_body dataset with a star of 3 in the [amazon_review dataset](https://huggingface.co/datasets/amazon_reviews_multi). * Input text what you want to generate review. * If the context is longer than 1200 characters, the context may be cut in the middle and the result may not come out well. review generator DEMO: [Ainize DEMO](https://main-review-generator-ehdwns1516.endpoint.ainize.ai/) review generator API: [Ainize API](https://ainize.web.app/redirect?git_repo=https://github.com/ehdwns1516/review_generator) ## Model links for each 1 to 5 star * [ehdwns1516/gpt2_review_star1](https://huggingface.co/ehdwns1516/gpt2_review_star1) * [ehdwns1516/gpt2_review_star2](https://huggingface.co/ehdwns1516/gpt2_review_star2) * [ehdwns1516/gpt2_review_star3](https://huggingface.co/ehdwns1516/gpt2_review_star3) * [ehdwns1516/gpt2_review_star4](https://huggingface.co/ehdwns1516/gpt2_review_star4) * [ehdwns1516/gpt2_review_star5](https://huggingface.co/ehdwns1516/gpt2_review_star5) ## Overview Language model: [gpt2](https://huggingface.co/gpt2) Language: English Training data: review_body dataset with a star of 3 in the [amazon_review dataset](https://huggingface.co/datasets/amazon_reviews_multi). Code: See [Ainize Workspace](https://ainize.ai/workspace/create?imageId=hnj95592adzr02xPTqss&git=https://github.com/ehdwns1516/gpt2_review_fine-tunning_note) ## Usage ## In Transformers ``` from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("ehdwns1516/gpt2_review_star3") model = AutoModelWithLMHead.from_pretrained("ehdwns1516/gpt2_review_star3") generator = pipeline( "text-generation", model="ehdwns1516/gpt2_review_star3", tokenizer=tokenizer ) context = "your context" result = dict() result[0] = generator(context)[0] ```
ehdwns1516/gpt3-kor-based_gpt2_review_SR2
cb6412bef95be277d547a5457dd8c9632e4c1698
2021-07-23T01:16:21.000Z
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
false
ehdwns1516
null
ehdwns1516/gpt3-kor-based_gpt2_review_SR2
3
null
transformers
21,231
# ehdwns1516/gpt3-kor-based_gpt2_review_SR2 * This model has been trained Korean dataset as a star of 2 in the [naver shopping reivew dataset](https://github.com/bab2min/corpus/tree/master/sentiment). * Input text what you want to generate review. * If the context is longer than 1200 characters, the context may be cut in the middle and the result may not come out well. review generator DEMO: [Ainize DEMO](https://main-review-generator-ehdwns1516.endpoint.ainize.ai/) review generator API: [Ainize API](https://ainize.web.app/redirect?git_repo=https://github.com/ehdwns1516/review_generator) ## Model links for each 1 to 5 star * [ehdwns1516/gpt3-kor-based_gpt2_review_SR1](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR1) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR2](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR2) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR3](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR3) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR4](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR4) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR5](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR5) ## Overview Language model: [gpt3-kor-small_based_on_gpt2](https://huggingface.co/kykim/gpt3-kor-small_based_on_gpt2) Language: Korean Training data: review_body dataset with a star of 2 in the [naver shopping reivew dataset](https://github.com/bab2min/corpus/tree/master/sentiment). Code: See [Ainize Workspace](https://ainize.ai/workspace/create?imageId=hnj95592adzr02xPTqss&git=https://github.com/ehdwns1516/gpt2_review_fine-tunning_note) ## Usage ## In Transformers ``` from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("ehdwns1516/gpt3-kor-based_gpt2_review_SR2") model = AutoModelWithLMHead.from_pretrained("ehdwns1516/gpt3-kor-based_gpt2_review_SR2") generator = pipeline( "text-generation", model="ehdwns1516/gpt3-kor-based_gpt2_review_SR2", tokenizer=tokenizer ) context = "your context" result = dict() result[0] = generator(context)[0] ```
ehdwns1516/gpt3-kor-based_gpt2_review_SR3
84e08e80af1023feee6b8a55fe684363a4fd5117
2021-07-23T01:18:13.000Z
[ "pytorch", "gpt2", "text-generation", "transformers" ]
text-generation
false
ehdwns1516
null
ehdwns1516/gpt3-kor-based_gpt2_review_SR3
3
null
transformers
21,232
# ehdwns1516/gpt3-kor-based_gpt2_review_SR3 * This model has been trained Korean dataset as a star of 3 in the [naver shopping reivew dataset](https://github.com/bab2min/corpus/tree/master/sentiment). * Input text what you want to generate review. * If the context is longer than 1200 characters, the context may be cut in the middle and the result may not come out well. review generator DEMO: [Ainize DEMO](https://main-review-generator-ehdwns1516.endpoint.ainize.ai/) review generator API: [Ainize API](https://ainize.web.app/redirect?git_repo=https://github.com/ehdwns1516/review_generator) ## Model links for each 1 to 5 star * [ehdwns1516/gpt3-kor-based_gpt2_review_SR1](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR1) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR2](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR2) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR3](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR3) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR4](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR4) * [ehdwns1516/gpt3-kor-based_gpt2_review_SR5](https://huggingface.co/ehdwns1516/gpt3-kor-based_gpt2_review_SR5) ## Overview Language model: [gpt3-kor-small_based_on_gpt2](https://huggingface.co/kykim/gpt3-kor-small_based_on_gpt2) Language: Korean Training data: review_body dataset with a star of 3 in the [naver shopping reivew dataset](https://github.com/bab2min/corpus/tree/master/sentiment). Code: See [Ainize Workspace](https://ainize.ai/workspace/create?imageId=hnj95592adzr02xPTqss&git=https://github.com/ehdwns1516/gpt2_review_fine-tunning_note) ## Usage ## In Transformers ``` from transformers import AutoTokenizer, AutoModelWithLMHead tokenizer = AutoTokenizer.from_pretrained("ehdwns1516/gpt3-kor-based_gpt2_review_SR3") model = AutoModelWithLMHead.from_pretrained("ehdwns1516/gpt3-kor-based_gpt2_review_SR3") generator = pipeline( "text-generation", model="ehdwns1516/gpt3-kor-based_gpt2_review_SR3", tokenizer=tokenizer ) context = "your context" result = dict() result[0] = generator(context)[0] ```
eli4s/chaii
6265b7a23ebc5d2e9539b95fe53d4914b4f3f98d
2021-10-15T18:44:01.000Z
[ "pytorch", "roberta", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
eli4s
null
eli4s/chaii
3
null
transformers
21,233
Entry not found
eli4s/prunedBert-L12-h256-A4-finetuned
1e2fff5edd9ae32675422c909bc185292a42e5b9
2021-08-17T08:03:00.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
eli4s
null
eli4s/prunedBert-L12-h256-A4-finetuned
3
null
transformers
21,234
This model was pretrained on the bookcorpus dataset using knowledge distillation. The particularity of this model is that even though it shares the same architecture as BERT, it has a hidden size of 256 (a third of the hidden size of BERT) and 4 attention heads (hence the same head size of BERT). The weights of the model were initialized by pruning the weights of bert-base-uncased. A knowledge distillation was performed using multiple loss functions to fine-tune the model. PS : the tokenizer is the same as the one of the model bert-base-uncased. To load the model \& tokenizer : ````python from transformers import AutoModelForMaskedLM, BertTokenizer model_name = "eli4s/prunedBert-L12-h256-A4-finetuned" model = AutoModelForMaskedLM.from_pretrained(model_name) tokenizer = BertTokenizer.from_pretrained(model_name) ```` To use it on a sentence : ````python import torch sentence = "Let's have a [MASK]." model.eval() inputs = tokenizer([sentence], padding='longest', return_tensors='pt') output = model(inputs['input_ids'], attention_mask=inputs['attention_mask']) mask_index = inputs['input_ids'].tolist()[0].index(103) masked_token = output['logits'][0][mask_index].argmax(axis=-1) predicted_token = tokenizer.decode(masked_token) print(predicted_token) ```` Or we can also predict the n most relevant predictions : ````python top_n = 5 vocab_size = model.config.vocab_size logits = output['logits'][0][mask_index].tolist() top_tokens = sorted(list(range(vocab_size)), key=lambda i:logits[i], reverse=True)[:top_n] tokenizer.decode(top_tokens) ````
emre/wav2vec2-large-xlsr-53-sah-CV8
6c339ee33d03af210806b7ddad91901df1916e5c
2022-03-24T11:53:42.000Z
[ "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "sah", "dataset:common_voice", "transformers", "generated_from_trainer", "robust-speech-event", "hf-asr-leaderboard", "license:apache-2.0", "model-index" ]
automatic-speech-recognition
false
emre
null
emre/wav2vec2-large-xlsr-53-sah-CV8
3
null
transformers
21,235
--- license: apache-2.0 language: sah tags: - generated_from_trainer - robust-speech-event - hf-asr-leaderboard datasets: - common_voice model-index: - name: wav2vec2-large-xlsr-53-sah-CV8 results: - task: name: Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice sah type: common_voice args: sah metrics: - name: Test WER type: wer value: 56.06 - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice 8.0 type: mozilla-foundation/common_voice_8_0 args: sah metrics: - name: Test WER type: wer value: 43.75 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xlsr-53-sah-CV8 This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.5089 - Wer: 0.5606 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 300 - num_epochs: 50 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:------:| | 4.6849 | 16.67 | 500 | 1.1135 | 0.9344 | | 0.8223 | 33.33 | 1000 | 0.5148 | 0.5686 | | 0.5477 | 50.0 | 1500 | 0.5089 | 0.5606 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.18.1 - Tokenizers 0.10.3
emre/wav2vec2-xls-r-300m-Tr-med-CommonVoice8
e801307ffe870ecb8a63e038fb884635781cfb07
2022-03-23T18:32:53.000Z
[ "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "tr", "dataset:common_voice", "transformers", "generated_from_trainer", "hf-asr-leaderboard", "robust-speech-event", "license:apache-2.0", "model-index" ]
automatic-speech-recognition
false
emre
null
emre/wav2vec2-xls-r-300m-Tr-med-CommonVoice8
3
0
transformers
21,236
--- license: apache-2.0 language: tr tags: - generated_from_trainer - hf-asr-leaderboard - robust-speech-event datasets: - common_voice model-index: - name: wav2vec2-xls-r-300m-Tr-med-CommonVoice8 results: - task: name: Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice tr type: common_voice args: tr metrics: - name: Test WER type: wer value: 49.14 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-xls-r-300m-Tr-med-CommonVoice8 This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. It achieves the following results on the evaluation set: - Loss: 0.2556 - Wer: 0.4914 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 300 - num_epochs: 20 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 1.4876 | 6.66 | 5000 | 0.3252 | 0.5784 | | 0.6919 | 13.32 | 10000 | 0.2720 | 0.5172 | | 0.5919 | 19.97 | 15000 | 0.2556 | 0.4914 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.18.1 - Tokenizers 0.10.3
emre/wav2vec2-xls-r-300m-as-CV8-v1
9dcf92c226f1e5b30941a112c16003005235075d
2022-03-24T11:55:32.000Z
[ "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "as", "dataset:common_voice", "transformers", "generated_from_trainer", "robust-speech-event", "hf-asr-leaderboard", "license:apache-2.0", "model-index" ]
automatic-speech-recognition
false
emre
null
emre/wav2vec2-xls-r-300m-as-CV8-v1
3
null
transformers
21,237
--- license: apache-2.0 language: as tags: - generated_from_trainer - robust-speech-event - hf-asr-leaderboard datasets: - common_voice model-index: - name: wav2vec2-xls-r-300m-as-CV8-v1 results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice 8.0 type: mozilla-foundation/common_voice_8_0 args: as metrics: - name: Test WER type: wer value: 100.0 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-xls-r-300m-as-CV8-v1 This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 300 - num_epochs: 30 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.11.3 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.10.3
en/distilbert-base-uncased-finetuned-squad
bcb597d278506ad466a6b0db7dca433b3a7c4ec3
2021-10-27T15:09:11.000Z
[ "pytorch", "tensorboard", "distilbert", "question-answering", "dataset:squad", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
question-answering
false
en
null
en/distilbert-base-uncased-finetuned-squad
3
null
transformers
21,238
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad model-index: - name: distilbert-base-uncased-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.1453 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.2065 | 1.0 | 5577 | 1.1289 | | 0.9226 | 2.0 | 11154 | 1.1019 | | 0.7411 | 3.0 | 16731 | 1.1453 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.0+cu111 - Datasets 1.14.0 - Tokenizers 0.10.3
enelpol/czywiesz-question
3cb34edd7bb3957cf81c12ce4ee83f661d34d0c2
2021-12-21T21:24:34.000Z
[ "pytorch", "bert", "feature-extraction", "pl", "dataset:enelpol/czywiesz", "transformers" ]
feature-extraction
false
enelpol
null
enelpol/czywiesz-question
3
null
transformers
21,239
--- language: pl datasets: - enelpol/czywiesz task_categories: - question_answering task_ids: - open-domain-qa multilinguality: - monolingual size_categories: - 1k<n<10K --- ## Model description This is the question encoder for the Polish DPR question answering model. The full model consists of two encoders. Please read [context encoder documentation](https://huggingface.co/enelpol/czywiesz-context) to get the details of the model.
entelecheia/ekonelectra-small-generator
ef10da0f09b53fe6a6ec7f7563f790c7e63670ef
2021-05-04T09:10:50.000Z
[ "pytorch", "electra", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
entelecheia
null
entelecheia/ekonelectra-small-generator
3
null
transformers
21,240
Entry not found
ericRosello/bert-base-uncased-finetuned-squad-frozen-v2
5f744523ade13a3c9fa0d6bb680b20f404cb751d
2022-01-05T16:14:23.000Z
[ "pytorch", "tensorboard", "bert", "question-answering", "dataset:squad", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
question-answering
false
ericRosello
null
ericRosello/bert-base-uncased-finetuned-squad-frozen-v2
3
null
transformers
21,241
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad model-index: - name: bert-base-uncased-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-uncased-finetuned-squad This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.4571 ## Model description Most base model weights were frozen leaving only to finetune the last layer (qa outputs) and 3 last layers of the encoder. ## Training and evaluation data Achieved EM: 76.77388836329234, F1: 85.41893520501723 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:------:|:---------------:| | 1.2944 | 1.0 | 44262 | 1.3432 | | 1.0152 | 2.0 | 88524 | 1.3450 | | 1.0062 | 3.0 | 132786 | 1.4571 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
ericRosello/distilbert-base-uncased-finetuned-squad-frozen-v2
88690eea200ef393f8a279340f1f24a968109101
2022-01-04T18:06:41.000Z
[ "pytorch", "tensorboard", "distilbert", "question-answering", "dataset:squad", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
question-answering
false
ericRosello
null
ericRosello/distilbert-base-uncased-finetuned-squad-frozen-v2
3
null
transformers
21,242
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad model-index: - name: distilbert-base-uncased-finetuned-squad results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-squad This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the squad dataset. It achieves the following results on the evaluation set: - Loss: 1.2104 ## Model description Most base model weights were frozen leaving only to finetune the last layer (qa outputs) and 3 last layers of the encoder. ## Training and evaluation data Achieved EM: 73.519394512772, F1: 82.71779517079237 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:-----:|:---------------:| | 1.3937 | 1.0 | 5533 | 1.2915 | | 1.1522 | 2.0 | 11066 | 1.2227 | | 1.0055 | 3.0 | 16599 | 1.2104 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu111 - Datasets 1.17.0 - Tokenizers 0.10.3
erica/kob400
d882193817ebbf17325eac1869d1809aa2821a07
2021-05-19T16:40:02.000Z
[ "pytorch", "jax", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
erica
null
erica/kob400
3
null
transformers
21,243
Entry not found
ericzhou/DialoGPT-Medium-Rick
3517b0daf597b077a457fc6e00eb7b22249f5e29
2022-01-20T04:34:16.000Z
[ "pytorch", "gpt2", "text-generation", "transformers", "conversational" ]
conversational
false
ericzhou
null
ericzhou/DialoGPT-Medium-Rick
3
null
transformers
21,244
--- tags: - conversational --- # Rick
facebook/s2t-small-mustc-en-ru-st
d20011149693c4bd6bc9babe1be83afcea703215
2022-02-07T15:09:10.000Z
[ "pytorch", "tf", "speech_to_text", "automatic-speech-recognition", "en", "ru", "dataset:mustc", "arxiv:2010.05171", "arxiv:1904.08779", "transformers", "audio", "speech-translation", "license:mit" ]
automatic-speech-recognition
false
facebook
null
facebook/s2t-small-mustc-en-ru-st
3
null
transformers
21,245
--- language: - en - ru datasets: - mustc tags: - audio - speech-translation - automatic-speech-recognition license: mit pipeline_tag: automatic-speech-recognition widget: - example_title: Librispeech sample 1 src: https://cdn-media.huggingface.co/speech_samples/sample1.flac - example_title: Librispeech sample 2 src: https://cdn-media.huggingface.co/speech_samples/sample2.flac --- # S2T-SMALL-MUSTC-EN-RU-ST `s2t-small-mustc-en-ru-st` is a Speech to Text Transformer (S2T) model trained for end-to-end Speech Translation (ST). The S2T model was proposed in [this paper](https://arxiv.org/abs/2010.05171) and released in [this repository](https://github.com/pytorch/fairseq/tree/master/examples/speech_to_text) ## Model description S2T is a transformer-based seq2seq (encoder-decoder) model designed for end-to-end Automatic Speech Recognition (ASR) and Speech Translation (ST). It uses a convolutional downsampler to reduce the length of speech inputs by 3/4th before they are fed into the encoder. The model is trained with standard autoregressive cross-entropy loss and generates the transcripts/translations autoregressively. ## Intended uses & limitations This model can be used for end-to-end English speech to Russian text translation. See the [model hub](https://huggingface.co/models?filter=speech_to_text) to look for other S2T checkpoints. ### How to use As this a standard sequence to sequence transformer model, you can use the `generate` method to generate the transcripts by passing the speech features to the model. *Note: The `Speech2TextProcessor` object uses [torchaudio](https://github.com/pytorch/audio) to extract the filter bank features. Make sure to install the `torchaudio` package before running this example.* You could either install those as extra speech dependancies with `pip install transformers"[speech, sentencepiece]"` or install the packages seperatly with `pip install torchaudio sentencepiece`. ```python import torch from transformers import Speech2TextProcessor, Speech2TextForConditionalGeneration from datasets import load_dataset import soundfile as sf model = Speech2TextForConditionalGeneration.from_pretrained("facebook/s2t-small-mustc-en-ru-st") processor = Speech2TextProcessor.from_pretrained("facebook/s2t-small-mustc-en-ru-st") def map_to_array(batch): speech, _ = sf.read(batch["file"]) batch["speech"] = speech return batch ds = load_dataset( "patrickvonplaten/librispeech_asr_dummy", "clean", split="validation" ) ds = ds.map(map_to_array) inputs = processor( ds["speech"][0], sampling_rate=16_000, return_tensors="pt" ) generated_ids = model.generate(input_ids=inputs["input_features"], attention_mask=inputs["attention_mask"]) translation = processor.batch_decode(generated_ids, skip_special_tokens=True) ``` ## Training data The s2t-small-mustc-en-ru-st is trained on English-Russian subset of [MuST-C](https://ict.fbk.eu/must-c/). MuST-C is a multilingual speech translation corpus whose size and quality facilitates the training of end-to-end systems for speech translation from English into several languages. For each target language, MuST-C comprises several hundred hours of audio recordings from English TED Talks, which are automatically aligned at the sentence level with their manual transcriptions and translations. ## Training procedure ### Preprocessing The speech data is pre-processed by extracting Kaldi-compliant 80-channel log mel-filter bank features automatically from WAV/FLAC audio files via PyKaldi or torchaudio. Further utterance-level CMVN (cepstral mean and variance normalization) is applied to each example. The texts are lowercased and tokenized using SentencePiece and a vocabulary size of 8,000. ### Training The model is trained with standard autoregressive cross-entropy loss and using [SpecAugment](https://arxiv.org/abs/1904.08779). The encoder receives speech features, and the decoder generates the transcripts autoregressively. To accelerate model training and for better performance the encoder is pre-trained for English ASR. ## Evaluation results MuST-C test results for en-ru (BLEU score): 15.3 ### BibTeX entry and citation info ```bibtex @inproceedings{wang2020fairseqs2t, title = {fairseq S2T: Fast Speech-to-Text Modeling with fairseq}, author = {Changhan Wang and Yun Tang and Xutai Ma and Anne Wu and Dmytro Okhonko and Juan Pino}, booktitle = {Proceedings of the 2020 Conference of the Asian Chapter of the Association for Computational Linguistics (AACL): System Demonstrations}, year = {2020}, } ```
facebook/s2t-wav2vec2-large-en-ar
f02a8c20f6a128fa2e6d6489001036bdc1af9564
2021-11-14T20:39:04.000Z
[ "pytorch", "speech-encoder-decoder", "automatic-speech-recognition", "en", "ar", "dataset:covost2", "dataset:librispeech_asr", "arxiv:2104.06678", "transformers", "audio", "speech-translation", "speech2text2", "license:mit" ]
automatic-speech-recognition
false
facebook
null
facebook/s2t-wav2vec2-large-en-ar
3
5
transformers
21,246
--- language: - en - ar datasets: - covost2 - librispeech_asr tags: - audio - speech-translation - automatic-speech-recognition - speech2text2 license: mit pipeline_tag: automatic-speech-recognition widget: - example_title: Common Voice 1 src: https://cdn-media.huggingface.co/speech_samples/common_voice_en_18301577.mp3 - example_title: Common Voice 2 src: https://cdn-media.huggingface.co/speech_samples/common_voice_en_99987.mp3 - example_title: Common Voice 3 src: https://cdn-media.huggingface.co/speech_samples/common_voice_en_99988.mp3 --- # S2T2-Wav2Vec2-CoVoST2-EN-AR-ST `s2t-wav2vec2-large-en-ar` is a Speech to Text Transformer model trained for end-to-end Speech Translation (ST). The S2T2 model was proposed in [Large-Scale Self- and Semi-Supervised Learning for Speech Translation](https://arxiv.org/pdf/2104.06678.pdf) and officially released in [Fairseq](https://github.com/pytorch/fairseq/blob/6f847c8654d56b4d1b1fbacec027f47419426ddb/fairseq/models/wav2vec/wav2vec2_asr.py#L266). ## Model description S2T2 is a transformer-based seq2seq (speech encoder-decoder) model designed for end-to-end Automatic Speech Recognition (ASR) and Speech Translation (ST). It uses a pretrained [Wav2Vec2](https://huggingface.co/transformers/model_doc/wav2vec2.html) as the encoder and a transformer-based decoder. The model is trained with standard autoregressive cross-entropy loss and generates the translations autoregressively. ## Intended uses & limitations This model can be used for end-to-end English speech to Arabic text translation. See the [model hub](https://huggingface.co/models?filter=speech2text2) to look for other S2T2 checkpoints. ### How to use As this a standard sequence to sequence transformer model, you can use the `generate` method to generate the transcripts by passing the speech features to the model. You can use the model directly via the ASR pipeline ```python from datasets import load_dataset from transformers import pipeline librispeech_en = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation") asr = pipeline("automatic-speech-recognition", model="facebook/s2t-wav2vec2-large-en-ar", feature_extractor="facebook/s2t-wav2vec2-large-en-ar") translation = asr(librispeech_en[0]["file"]) ``` or step-by-step as follows: ```python import torch from transformers import Speech2Text2Processor, SpeechEncoderDecoder from datasets import load_dataset import soundfile as sf model = SpeechEncoderDecoder.from_pretrained("facebook/s2t-wav2vec2-large-en-ar") processor = Speech2Text2Processor.from_pretrained("facebook/s2t-wav2vec2-large-en-ar") def map_to_array(batch): speech, _ = sf.read(batch["file"]) batch["speech"] = speech return batch ds = load_dataset("patrickvonplaten/librispeech_asr_dummy", "clean", split="validation") ds = ds.map(map_to_array) inputs = processor(ds["speech"][0], sampling_rate=16_000, return_tensors="pt") generated_ids = model.generate(input_ids=inputs["input_features"], attention_mask=inputs["attention_mask"]) transcription = processor.batch_decode(generated_ids) ``` ## Evaluation results CoVoST-V2 test results for en-ar (BLEU score): **20.2** For more information, please have a look at the [official paper](https://arxiv.org/pdf/2104.06678.pdf) - especially row 10 of Table 2. ### BibTeX entry and citation info ```bibtex @article{DBLP:journals/corr/abs-2104-06678, author = {Changhan Wang and Anne Wu and Juan Miguel Pino and Alexei Baevski and Michael Auli and Alexis Conneau}, title = {Large-Scale Self- and Semi-Supervised Learning for Speech Translation}, journal = {CoRR}, volume = {abs/2104.06678}, year = {2021}, url = {https://arxiv.org/abs/2104.06678}, archivePrefix = {arXiv}, eprint = {2104.06678}, timestamp = {Thu, 12 Aug 2021 15:37:06 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2104-06678.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ```
facebook/wav2vec2-base-10k-voxpopuli-ft-nl
cd45c5fa284c36cec584a86eb77324e772983984
2021-07-06T01:51:40.000Z
[ "pytorch", "wav2vec2", "automatic-speech-recognition", "nl", "arxiv:2101.00390", "transformers", "audio", "voxpopuli", "license:cc-by-nc-4.0" ]
automatic-speech-recognition
false
facebook
null
facebook/wav2vec2-base-10k-voxpopuli-ft-nl
3
null
transformers
21,247
--- language: nl tags: - audio - automatic-speech-recognition - voxpopuli license: cc-by-nc-4.0 --- # Wav2Vec2-Base-VoxPopuli-Finetuned [Facebook's Wav2Vec2](https://ai.facebook.com/blog/wav2vec-20-learning-the-structure-of-speech-from-raw-audio/) base model pretrained on the 10K unlabeled subset of [VoxPopuli corpus](https://arxiv.org/abs/2101.00390) and fine-tuned on the transcribed data in nl (refer to Table 1 of paper for more information). **Paper**: *[VoxPopuli: A Large-Scale Multilingual Speech Corpus for Representation Learning, Semi-Supervised Learning and Interpretation](https://arxiv.org/abs/2101.00390)* **Authors**: *Changhan Wang, Morgane Riviere, Ann Lee, Anne Wu, Chaitanya Talnikar, Daniel Haziza, Mary Williamson, Juan Pino, Emmanuel Dupoux* from *Facebook AI* See the official website for more information, [here](https://github.com/facebookresearch/voxpopuli/) # Usage for inference In the following it is shown how the model can be used in inference on a sample of the [Common Voice dataset](https://commonvoice.mozilla.org/en/datasets) ```python #!/usr/bin/env python3 from transformers import Wav2Vec2Processor, Wav2Vec2ForCTC from datasets import load_dataset import torchaudio import torch # resample audio # load model & processor model = Wav2Vec2ForCTC.from_pretrained("facebook/wav2vec2-base-10k-voxpopuli-ft-nl") processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-10k-voxpopuli-ft-nl") # load dataset ds = load_dataset("common_voice", "nl", split="validation[:1%]") # common voice does not match target sampling rate common_voice_sample_rate = 48000 target_sample_rate = 16000 resampler = torchaudio.transforms.Resample(common_voice_sample_rate, target_sample_rate) # define mapping fn to read in sound file and resample def map_to_array(batch): speech, _ = torchaudio.load(batch["path"]) speech = resampler(speech) batch["speech"] = speech[0] return batch # load all audio files ds = ds.map(map_to_array) # run inference on the first 5 data samples inputs = processor(ds[:5]["speech"], sampling_rate=target_sample_rate, return_tensors="pt", padding=True) # inference logits = model(**inputs).logits predicted_ids = torch.argmax(logits, axis=-1) print(processor.batch_decode(predicted_ids)) ```
fadhilarkan/distilbert-base-uncased-finetuned-cola-4
6670b8b8340d90ae8ec46d4f50b4f8634f6d5137
2021-11-13T04:02:51.000Z
[ "pytorch", "tensorboard", "distilbert", "text-classification", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index" ]
text-classification
false
fadhilarkan
null
fadhilarkan/distilbert-base-uncased-finetuned-cola-4
3
null
transformers
21,248
--- license: apache-2.0 tags: - generated_from_trainer metrics: - matthews_correlation model-index: - name: distilbert-base-uncased-finetuned-cola-4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola-4 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0011 - Matthews Correlation: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | No log | 1.0 | 104 | 0.0243 | 1.0 | | No log | 2.0 | 208 | 0.0074 | 1.0 | | No log | 3.0 | 312 | 0.0041 | 1.0 | | No log | 4.0 | 416 | 0.0028 | 1.0 | | 0.0929 | 5.0 | 520 | 0.0021 | 1.0 | | 0.0929 | 6.0 | 624 | 0.0016 | 1.0 | | 0.0929 | 7.0 | 728 | 0.0014 | 1.0 | | 0.0929 | 8.0 | 832 | 0.0012 | 1.0 | | 0.0929 | 9.0 | 936 | 0.0012 | 1.0 | | 0.0021 | 10.0 | 1040 | 0.0011 | 1.0 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
fadhilarkan/distilbert-base-uncased-finetuned-cola
1a9b7f05868ef6c10925873223edfcc5257b7b6c
2021-11-13T01:33:17.000Z
[ "pytorch", "tensorboard", "distilbert", "text-classification", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index" ]
text-classification
false
fadhilarkan
null
fadhilarkan/distilbert-base-uncased-finetuned-cola
3
null
transformers
21,249
--- license: apache-2.0 tags: - generated_from_trainer metrics: - matthews_correlation model-index: - name: distilbert-base-uncased-finetuned-cola results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.0008 - Matthews Correlation: 1.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | No log | 1.0 | 130 | 0.0166 | 1.0 | | No log | 2.0 | 260 | 0.0054 | 1.0 | | No log | 3.0 | 390 | 0.0029 | 1.0 | | 0.0968 | 4.0 | 520 | 0.0019 | 1.0 | | 0.0968 | 5.0 | 650 | 0.0014 | 1.0 | | 0.0968 | 6.0 | 780 | 0.0011 | 1.0 | | 0.0968 | 7.0 | 910 | 0.0010 | 1.0 | | 0.0018 | 8.0 | 1040 | 0.0008 | 1.0 | | 0.0018 | 9.0 | 1170 | 0.0008 | 1.0 | | 0.0018 | 10.0 | 1300 | 0.0008 | 1.0 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.10.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
fadhilarkan/t5-small-finetuned-xsum
840c384366f5eda115f396ad9df5b081617331b6
2021-08-18T10:37:43.000Z
[ "pytorch", "tensorboard", "t5", "text2text-generation", "dataset:squad", "transformers", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible" ]
text2text-generation
false
fadhilarkan
null
fadhilarkan/t5-small-finetuned-xsum
3
null
transformers
21,250
--- license: apache-2.0 tags: - generated_from_trainer datasets: - squad model_index: - name: t5-small-finetuned-xsum results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: squad type: squad args: plain_text --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-xsum This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 10 - eval_batch_size: 10 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.9.2 - Pytorch 1.9.0+cu102 - Datasets 1.11.0 - Tokenizers 0.10.3
ffsouza/tiny-mbart-finetuned-en-to-ro
224c12c61660f11022c6fff53fc7267884e27d98
2021-11-30T00:39:57.000Z
[ "pytorch", "tensorboard", "mbart", "text2text-generation", "dataset:wmt16_en_ro_pre_processed", "transformers", "generated_from_trainer", "model-index", "autotrain_compatible" ]
text2text-generation
false
ffsouza
null
ffsouza/tiny-mbart-finetuned-en-to-ro
3
null
transformers
21,251
--- tags: - generated_from_trainer datasets: - wmt16_en_ro_pre_processed metrics: - bleu model-index: - name: tiny-mbart-finetuned-en-to-ro results: - task: name: Sequence-to-sequence Language Modeling type: text2text-generation dataset: name: wmt16_en_ro_pre_processed type: wmt16_en_ro_pre_processed args: enro metrics: - name: Bleu type: bleu value: 0.0 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # tiny-mbart-finetuned-en-to-ro This model is a fine-tuned version of [sshleifer/tiny-mbart](https://huggingface.co/sshleifer/tiny-mbart) on the wmt16_en_ro_pre_processed dataset. It achieves the following results on the evaluation set: - Loss: 8.4792 - Bleu: 0.0 - Gen Len: 20.0 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len | |:-------------:|:-----:|:-----:|:---------------:|:----:|:-------:| | 8.2425 | 1.0 | 76290 | 8.4792 | 0.0 | 20.0 | ### Framework versions - Transformers 4.12.5 - Pytorch 1.10.0+cu102 - Datasets 1.15.1 - Tokenizers 0.10.3
fjluque/roberta-base-bne-finetuned-amazon_reviews_multi
8d98eba2b375e6fb560efd0247c8de7c205e9f04
2021-09-16T13:20:48.000Z
[ "pytorch", "tensorboard", "roberta", "text-classification", "dataset:amazon_reviews_multi", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index" ]
text-classification
false
fjluque
null
fjluque/roberta-base-bne-finetuned-amazon_reviews_multi
3
null
transformers
21,252
--- license: apache-2.0 tags: - generated_from_trainer datasets: - amazon_reviews_multi metrics: - accuracy model-index: - name: roberta-base-bne-finetuned-amazon_reviews_multi results: - task: name: Text Classification type: text-classification dataset: name: amazon_reviews_multi type: amazon_reviews_multi args: es metrics: - name: Accuracy type: accuracy value: 0.91725 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-bne-finetuned-amazon_reviews_multi This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset. It achieves the following results on the evaluation set: - Loss: 0.2157 - Accuracy: 0.9173 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.1125 | 1.0 | 13 | 0.2066 | 0.9165 | | 0.0186 | 2.0 | 26 | 0.2157 | 0.9173 | ### Framework versions - Transformers 4.10.2 - Pytorch 1.9.0+cu102 - Datasets 1.12.1 - Tokenizers 0.10.3
flavio-nakasato/berdou_200k
a71a04c78013df3cbc05bfd8a352f710c9f4c2d1
2021-08-15T15:42:10.000Z
[ "pytorch", "bert", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
flavio-nakasato
null
flavio-nakasato/berdou_200k
3
null
transformers
21,253
MLM fine-tuned from Bertimbau-Base model on the Brazilian Federal Official Gazette (200k instances)
flax-community/clip-rsicd-v3
81fa894b5abdb3124749a8fada12c264fcd37123
2021-07-17T08:36:30.000Z
[ "pytorch", "jax", "clip", "feature-extraction", "transformers" ]
feature-extraction
false
flax-community
null
flax-community/clip-rsicd-v3
3
null
transformers
21,254
Entry not found
flax-community/gpt-neo-1.3B-apps-all-2
af44a29648c2e7e85d2f8ba6b6917e585ecf74ef
2021-09-22T08:25:21.000Z
[ "pytorch", "jax", "gpt_neo", "text-generation", "en", "python", "dataset:apps", "arxiv:2107.03374", "transformers", "code_synthesis", "license:mit" ]
text-generation
false
flax-community
null
flax-community/gpt-neo-1.3B-apps-all-2
3
2
transformers
21,255
--- language: - en - python license: mit tags: - gpt_neo - code_synthesis datasets: - apps --- # GPT-Code-Clippy-1.3B-APPS-all ## Model Description GPT-Neo-1.3B-APPS-all is a GPT-Neo-1.3B fine-tuned on APPS dataset. This model is specialized to solve programming tasks. ## Training data The model is trained on the [Automated Programming Progress Standard (APPS) dataset](https://github.com/hendrycks/apps). The dataset consists of 10,000 coding problems in total, with 131,836 test cases for checking solutions and 232,444 ground-truth solutions written by humans. Problems can be complicated, as the average length of a problem is 293.2 words. The data are split evenly into training and test sets, with 5,000 problems each. This model is fine-tuned using most of the APPS dataset including both train and test split to explore the impact of this training task on model performance on other code synthesis evaluation metrics. A model fine-tuned on train set only can be found [here](https://huggingface.co/flax-community/gpt-neo-1.3B-apps). ## Training procedure The training script used to train this model can be found [here](https://github.com/ncoop57/gpt-code-clippy/blob/camera-ready/training/run_clm_apps.py). Training is done for 5 epochs using AdamW optimizer and leaner decay learning rate schedule with 800 warmup steps. To reproduce the training one can use this command with the above script: ``` python run_clm_apps.py \ --output_dir ./gpt-neo-125M-apps \ --model_name_or_path EleutherAI/gpt-neo-125B \ --dataset_name ./apps.py \ --dataset_config_name formatted \ --do_train --do_eval \ --block_size="1024" \ --per_device_train_batch_size="3" \ --per_device_eval_batch_size="3" \ --preprocessing_num_workers="16" \ --learning_rate="8e-5" \ --warmup_steps="800" \ --adam_beta1="0.9" \ --adam_beta2="0.98" \ --weight_decay="0.1" \ --overwrite_output_dir \ --num_train_epochs="5" \ --logging_steps="50" \ --eval_steps="2000" \ --report_to="wandb" \ --dtype="bfloat16" \ --save_strategy epoch \ --gradient_accumulation_steps 1 \ --all_data true \ ``` ## Intended Use and Limitations The model is fine-tuned to solve programming problems given a text description and optional starter code. ### How to use You can use this model directly with a pipeline for text generation. This example generates a different sequence each time it's run: ```py from transformers import AutoModelForCausalLM, AutoTokenizer, FlaxAutoModelForCausalLM model = AutoModelForCausalLM.from_pretrained("flax-community/gpt-neo-1.3B-apps-all-2") tokenizer = AutoTokenizer.from_pretrained("flax-community/gpt-neo-1.3B-apps-all-2") prompt = """ A function to greet user. Given a user name it should say hello def greet(name): ANSWER: """ input_ids = tokenizer(prompt, return_tensors='pt').input_ids.to(device) start = input_ids.size(1) out = model.generate(input_ids, do_sample=True, max_length=50, num_beams=2, early_stopping=True, eos_token_id=tokenizer.eos_token_id, ) print(tokenizer.decode(out[0][start:])) ``` ### Limitations and Biases The model is intended to be used for research purposes and comes with no guarantees of quality of generated code. The paper ["Evaluating Large Language Models Trained on Code"](https://arxiv.org/abs/2107.03374) from OpenAI has a good discussion on what the impact of a large language model trained on code could be. Therefore, some parts of their discuss are highlighted here as it pertains to this dataset and models that may be trained from it. **As well as some differences in views from the paper, particularly around legal implications**. 1. **Over-reliance:** This model may generate plausible solutions that may appear correct, but are not necessarily the correct solution. Not properly evaluating the generated code may cause have negative consequences such as the introduction of bugs, or the introduction of security vulnerabilities. Therefore, it is important that users are aware of the limitations and potential negative consequences of using this language model. 2. **Economic and labor market impacts:** Large language models trained on large code datasets such as this one that are capable of generating high-quality code have the potential to automate part of the software development process. This may negatively impact software developers. However, as discussed in the paper, as shown in the Summary Report of software developers from [O*NET OnLine](https://www.onetonline.org/link/summary/15-1252.00), developers don't just write software. 5. **Biases:** The model is trained on data containing prompt questions formatted in specific way. The performance of the model can be worse if the prompt formatting is different from that used in APPS dataset. This model is finetuned GPT-Neo and might have inhereted biases and limitations from it. See [GPT-Neo model card](https://huggingface.co/EleutherAI/gpt-neo-125M#limitations-and-biases) for details. ## Eval results Coming soon...
flax-community/mongolian-gpt2
5c35deb772c25bb0e7ddda30b4bca7a7c86c6179
2021-07-09T12:17:08.000Z
[ "pytorch", "jax", "gpt2", "text-generation", "mn", "dataset:oscar", "transformers" ]
text-generation
false
flax-community
null
flax-community/mongolian-gpt2
3
2
transformers
21,256
--- language: "mn" thumbnail: "https://avatars.githubusercontent.com/u/43239645?s=60&v=4" tags: - gpt2 datasets: - oscar --- # Mongolian GPT2 Goal is to create a strong language generation model for Mongolian Since initial code and data is pretty much written by @patrickvonplaten and other huggingface members, it should not be so hard to get the first sense. ## Model Randomly initialized GPT2 model ## Datasets We can use OSCAR which is available through datasets ## Datasets A causal language modeling script for Flax is available here 1. It can be used pretty much without any required code changes. If there is time left, I’d love to try some private crawling and integrate it datasets. ## Expected Outcome Understandable Mongolian text generation model ## Challenges Lack of data → OSCAR Mongolian is just 2.2G. Maybe we need to research ways to acquire more data with this.
flax-community/nordic-gpt-wiki
36383cfd596c88a77fd299956ae601e9c49c40a1
2021-07-17T07:46:37.000Z
[ "pytorch", "jax", "tensorboard", "gpt2", "text-generation", "sv", "transformers" ]
text-generation
false
flax-community
null
flax-community/nordic-gpt-wiki
3
null
transformers
21,257
--- language: sv widget: - text: "Det var en gång" --- # Nordic GPT2--wikipedia A Nordic GPT2 style model trained using Flax CLM pipeline on the Nordic parts part of the wiki40b dataset. https://huggingface.co/datasets/wiki40b ## Model series This model is part of a series of models training on TPU with Flax Jax during Huggingface Flax/Jax challenge. ## Gpt models ## Swedish Gpt https://huggingface.co/birgermoell/swedish-gpt/ ## Swedish gpt wiki https://huggingface.co/flax-community/swe-gpt-wiki # Nordic gpt wiki https://huggingface.co/flax-community/nordic-gpt-wiki ## Dansk gpt wiki https://huggingface.co/flax-community/dansk-gpt-wiki ## Norsk gpt wiki https://huggingface.co/flax-community/norsk-gpt-wiki ## Roberta models ## Nordic Roberta Wiki https://huggingface.co/flax-community/nordic-roberta-wiki ## Swe Roberta Wiki Oscar https://huggingface.co/flax-community/swe-roberta-wiki-oscar ## Roberta Swedish Scandi https://huggingface.co/birgermoell/roberta-swedish-scandi ## Roberta Swedish https://huggingface.co/birgermoell/roberta-swedish ## Swedish T5 model https://huggingface.co/birgermoell/t5-base-swedish ## Data cleaning and preprocessing The data was cleaned and preprocessed using the following script. Make sure to install depencies for beam_runner to make the dataset work. ```python from datasets import load_dataset def load_and_clean_wiki(): dataset = load_dataset('wiki40b', 'da', beam_runner='DirectRunner', split="train") #dataset = load_dataset('wiki40b', 'sv', beam_runner='DirectRunner') dataset = dataset.remove_columns(['wikidata_id', 'version_id']) filtered_dataset = dataset.map(filter_wikipedia) # filtered_dataset[:3] # print(filtered_dataset[:3]) return filtered_dataset def filter_wikipedia(batch): batch["text"] = " ".join(batch["text"].split("\ _START_SECTION_\ ")) batch["text"] = " ".join(batch["text"].split("\ _START_ARTICLE_\ ")) batch["text"] = " ".join(batch["text"].split("\ _START_ARTICLE_\ ")) batch["text"] = " ".join(batch["text"].split("\ _START_PARAGRAPH_\ ")) batch["text"] = " ".join(batch["text"].split("_NEWLINE_")) batch["text"] = " ".join(batch["text"].split("\xa0")) return batch ``` ## Training script The following training script was used to train the model. ```bash ./run_clm_flax.py --output_dir="${MODEL_DIR}" --model_type="gpt2" --config_name="${MODEL_DIR}" --tokenizer_name="${MODEL_DIR}" --dataset_name="wiki40b" --dataset_config_name="da" --do_train --do_eval --block_size="512" --per_device_train_batch_size="64" --per_device_eval_batch_size="64" --learning_rate="5e-3" --warmup_steps="1000" --adam_beta1="0.9" --adam_beta2="0.98" --weight_decay="0.01" --overwrite_output_dir --num_train_epochs="20" --logging_steps="500" --save_steps="1000" --eval_steps="2500" --push_to_hub ```
flax-sentence-embeddings/multi-qa_v1-MiniLM-L6-mean_cos
c4fc9279c07b71e1d8fef5cd08ea245fce015bb3
2021-07-26T01:34:18.000Z
[ "pytorch", "bert", "feature-extraction", "arxiv:2102.07033", "arxiv:2104.08727", "sentence-transformers", "sentence-similarity" ]
sentence-similarity
false
flax-sentence-embeddings
null
flax-sentence-embeddings/multi-qa_v1-MiniLM-L6-mean_cos
3
null
sentence-transformers
21,258
--- pipeline_tag: sentence-similarity tags: - sentence-transformers - feature-extraction - sentence-similarity --- # multi-qa_v1-MiniLM-L6-mean_cos ## Model Description SentenceTransformers is a set of models and frameworks that enable training and generating sentence embeddings from given data. The generated sentence embeddings can be utilized for Clustering, Semantic Search and other tasks. We used a pretrained [nreimers/MiniLM-L6-H384-uncased](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased) model and trained it using Siamese Network setup and contrastive learning objective. Question and answer pairs from StackExchange was used as training data to make the model robust to Question / Answer embedding similarity. For this model, mean pooling of the hidden states were used as sentence embeddings. We developed this model during the [Community week using JAX/Flax for NLP & CV](https://discuss.huggingface.co/t/open-to-the-community-community-week-using-jax-flax-for-nlp-cv/7104), organized by Hugging Face. We developed this model as part of the project: [Train the Best Sentence Embedding Model Ever with 1B Training Pairs](https://discuss.huggingface.co/t/train-the-best-sentence-embedding-model-ever-with-1b-training-pairs/7354). We benefited from efficient hardware infrastructure to run the project: 7 TPUs v3-8, as well as assistance from Google’s Flax, JAX, and Cloud team members about efficient deep learning frameworks. ## Intended uses Our model is intended to be used as a sentence encoder for a search engine. Given an input sentence, it outputs a vector which captures the sentence semantic information. The sentence vector may be used for semantic-search, clustering or sentence similarity tasks. ## How to use Here is how to use this model to get the features of a given text using [SentenceTransformers](https://github.com/UKPLab/sentence-transformers) library: ```python from sentence_transformers import SentenceTransformer model = SentenceTransformer('flax-sentence-embeddings/multi-qa_v1-MiniLM-L6-mean_cos') text = "Replace me by any question / answer you'd like." text_embbedding = model.encode(text) # array([-0.01559514, 0.04046123, 0.1317083 , 0.00085931, 0.04585106, # -0.05607086, 0.0138078 , 0.03569756, 0.01420381, 0.04266302 ...], # dtype=float32) ``` # Training procedure ## Pre-training We use the pretrained [nreimers/MiniLM-L6-H384-uncased](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased). Please refer to the model card for more detailed information about the pre-training procedure. ## Fine-tuning We fine-tune the model using a contrastive objective. Formally, we compute the cosine similarity from each possible sentence pairs from the batch. We then apply the cross entropy loss by comparing with true pairs. ### Hyper parameters We trained on model on a TPU v3-8. We train the model during 80k steps using a batch size of 1024 (128 per TPU core). We use a learning rate warm up of 500. The sequence length was limited to 128 tokens. We used the AdamW optimizer with a 2e-5 learning rate. The full training script is accessible in this current repository. ### Training data We used the concatenation from multiple Stackexchange Question-Answer datasets to fine-tune our model. MSMARCO, NQ & other question-answer datasets were also used. | Dataset | Paper | Number of training tuples | |:--------------------------------------------------------:|:----------------------------------------:|:--------------------------:| | [Stack Exchange QA - Title & Answer](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl) | - | 4,750,619 | | [Stack Exchange](https://huggingface.co/datasets/flax-sentence-embeddings/stackexchange_title_body_jsonl) | - | 364,001 | | [TriviaqQA](https://huggingface.co/datasets/trivia_qa) | - | 73,346 | | [SQuAD2.0](https://rajpurkar.github.io/SQuAD-explorer/) | [paper](https://aclanthology.org/P18-2124.pdf) | 87,599 | | [Quora Question Pairs](https://quoradata.quora.com/First-Quora-Dataset-Release-Question-Pairs) | - | 103,663 | | [Eli5](https://huggingface.co/datasets/eli5) | [paper](https://doi.org/10.18653/v1/p19-1346) | 325,475 | | [PAQ](https://github.com/facebookresearch/PAQ) | [paper](https://arxiv.org/abs/2102.07033) | 64,371,441 | | [WikiAnswers](https://github.com/afader/oqa#wikianswers-corpus) | [paper](https://doi.org/10.1145/2623330.2623677) | 77,427,422 | | [MS MARCO](https://microsoft.github.io/msmarco/) | [paper](https://doi.org/10.1145/3404835.3462804) | 9,144,553 | | [GOOAQ: Open Question Answering with Diverse Answer Types](https://github.com/allenai/gooaq) | [paper](https://arxiv.org/pdf/2104.08727.pdf) | 3,012,496 | | [Yahoo Answers](https://www.kaggle.com/soumikrakshit/yahoo-answers-dataset) Question/Answer | [paper](https://proceedings.neurips.cc/paper/2015/hash/250cf8b51c773f3f8dc8b4be867a9a02-Abstract.html) | 681,164 | | SearchQA | - | 582,261 | | [Natural Questions (NQ)](https://ai.google.com/research/NaturalQuestions) | [paper](https://transacl.org/ojs/index.php/tacl/article/view/1455) | 100,231 |
flboehm/reddit-bert-text_20
0534dde5b64efe8bd7b2596e1b96245a174e6c83
2021-12-18T12:47:04.000Z
[ "pytorch", "tensorboard", "bert", "fill-mask", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
fill-mask
false
flboehm
null
flboehm/reddit-bert-text_20
3
null
transformers
21,259
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: reddit-bert-text_20 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # reddit-bert-text_20 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.4702 - Perplexity: 11.82 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.9383 | 1.0 | 947 | 2.5420 | | 2.6448 | 2.0 | 1894 | 2.5241 | | 2.586 | 3.0 | 2841 | 2.4833 | ### Framework versions - Transformers 4.14.1 - Pytorch 1.10.0+cu113 - Datasets 1.16.1 - Tokenizers 0.10.3
flboehm/youtube-bert
513b8a082b634bbf23aea72a5f4126ad22e51e29
2022-01-12T21:29:46.000Z
[ "pytorch", "tensorboard", "bert", "fill-mask", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
fill-mask
false
flboehm
null
flboehm/youtube-bert
3
null
transformers
21,260
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: youtube-bert results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # youtube-bert This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.4771 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 2.691 | 1.0 | 1077 | 2.5445 | | 2.5768 | 2.0 | 2154 | 2.5226 | | 2.5227 | 3.0 | 3231 | 2.5027 | ### Framework versions - Transformers 4.15.0 - Pytorch 1.10.0+cu113 - Datasets 1.17.0 - Tokenizers 0.10.3
fractaldna22/GPT_2_Marxism
c9f2599545f056969f5521d0319b059a20b03efe
2021-12-27T21:15:52.000Z
[ "pytorch", "gpt2", "transformers" ]
null
false
fractaldna22
null
fractaldna22/GPT_2_Marxism
3
null
transformers
21,261
tags: - Text2text Generation - Conversational - Text generation model: - "355M" model-type: - gpt2 widgets: text_example_1: - "One would be forgiven if one was not aware that Julian Assange is being" title_example_1: - "David North wsws" text_example_2: - "I would like to extend my sincerest greetings to the people of the world. When monstrous and absurd accusations were hurled at me and my family -- when" title_example_2: - "Leon Trotsky" # GPT_2_Marxism is based on the gpt-2 355M model finetuned on a large corpus of Marxist documents, polemics and literature from historical and contemporary writers # in the international socialist movement and the ICFI (fourth international) which upholds the principles which characterize genuine revolutionary marxism i.e. Trotskyism. # This finetuned gpt-2 model generates genuinely Marxist insights and responses. # - Generated with the GPT2-355M model converted to pytorch using Max Woolf's aitextgen notebook (https://github.com/minimaxir/aitextgen) # - "Finetuned on a large corpus of text mostly unstructured, unlabeled, raw copy and paste of entire selected works." # - "Able to generate genuine Marxist responses" # - "This model also generates insights that marxists often agree on, like freedom and equality." import torch import random pip3 install aitextgen import aitextgen model = aitextgen("model.pytorch.bin") text = "one would be forgiven if one was not aware that Julian Assange is currently being" model.generate(n=3, prompt="Lenin:"+str(text), max_length=77, temperature=random.uniform(0.5, 1.5), seed=random.randint(0, 195302), lstrip=False) """ Lenin:one would be forgiven if one was not aware that Julian Assange is currently being persecuted by the governments of the United States, the UK and many other countries in spite of, or perhaps because of, the fact that he is an outspoken enemy of imperialism. This not unexpected. In 2003 a law was passed in the US that allowed prosecution of those who helped the FBI to violate civil ========== Lenin:one would be forgiven if one was not aware that Julian Assange is currently being investigated by the FBI for illegally departing Ecuador - (although I had no proof available at the time) with the purpose of, as it were, of snatching up the devious Clintonite. Indeed, such an intrusion seems all the more fishy from the standpoint of a serious study of the facts ========== Lenin:one would be forgiven if one was not aware that Julian Assange is currently being extradited before the beginning of June to answer questions which require a presumption of guilt. This follows from the very revealing papers that WikiLeaks provided in relation to the numerous criminal cases, and of the complex international network which organised it, the publication by WikiLeaks of thousands of secret cables from the intelligence agencies of the """
frgfm/darknet19
867bc308fb3d68666be157477068a8bbfb99f4f7
2022-07-20T00:57:15.000Z
[ "pytorch", "dataset:frgfm/imagenette", "arxiv:1612.08242", "transformers", "image-classification", "license:apache-2.0" ]
image-classification
false
frgfm
null
frgfm/darknet19
3
null
transformers
21,262
--- license: apache-2.0 tags: - image-classification - pytorch datasets: - frgfm/imagenette --- # Darknet-19 model Pretrained on [ImageNette](https://github.com/fastai/imagenette). The Darknet-19 architecture was introduced in [this paper](https://pjreddie.com/media/files/papers/YOLO9000.pdf). ## Model description The core idea of the author is to combine high throughput of a highway net with performance gains using better activations (Leaky ReLU) and batch normalization. This architecture is used as a backbone for YOLOv2. ## Installation ### Prerequisites Python 3.6 (or higher) and [pip](https://pip.pypa.io/en/stable/)/[conda](https://docs.conda.io/en/latest/miniconda.html) are required to install Holocron. ### Latest stable release You can install the last stable release of the package using [pypi](https://pypi.org/project/pylocron/) as follows: ```shell pip install pylocron ``` or using [conda](https://anaconda.org/frgfm/pylocron): ```shell conda install -c frgfm pylocron ``` ### Developer mode Alternatively, if you wish to use the latest features of the project that haven't made their way to a release yet, you can install the package from source *(install [Git](https://git-scm.com/book/en/v2/Getting-Started-Installing-Git) first)*: ```shell git clone https://github.com/frgfm/Holocron.git pip install -e Holocron/. ``` ## Usage instructions ```python from PIL import Image from torchvision.transforms import Compose, ConvertImageDtype, Normalize, PILToTensor, Resize from torchvision.transforms.functional import InterpolationMode from holocron.models import model_from_hf_hub model = model_from_hf_hub("frgfm/darknet19").eval() img = Image.open(path_to_an_image).convert("RGB") # Preprocessing config = model.default_cfg transform = Compose([ Resize(config['input_shape'][1:], interpolation=InterpolationMode.BILINEAR), PILToTensor(), ConvertImageDtype(torch.float32), Normalize(config['mean'], config['std']) ]) input_tensor = transform(img).unsqueeze(0) # Inference with torch.inference_mode(): output = model(input_tensor) probs = output.squeeze(0).softmax(dim=0) ``` ## Citation Original paper ```bibtex @article{DBLP:journals/corr/RedmonF16, author = {Joseph Redmon and Ali Farhadi}, title = {{YOLO9000:} Better, Faster, Stronger}, journal = {CoRR}, volume = {abs/1612.08242}, year = {2016}, url = {http://arxiv.org/abs/1612.08242}, eprinttype = {arXiv}, eprint = {1612.08242}, timestamp = {Mon, 13 Aug 2018 16:48:25 +0200}, biburl = {https://dblp.org/rec/journals/corr/RedmonF16.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` Source of this implementation ```bibtex @software{Fernandez_Holocron_2020, author = {Fernandez, François-Guillaume}, month = {5}, title = {{Holocron}}, url = {https://github.com/frgfm/Holocron}, year = {2020} } ```
frgfm/repvgg_a1
6a9115efdacfe3eb94d2e0fa46d921b5476c28fc
2022-07-20T00:56:06.000Z
[ "pytorch", "onnx", "dataset:frgfm/imagenette", "arxiv:2101.03697", "transformers", "image-classification", "license:apache-2.0" ]
image-classification
false
frgfm
null
frgfm/repvgg_a1
3
null
transformers
21,263
--- license: apache-2.0 tags: - image-classification - pytorch - onnx datasets: - frgfm/imagenette --- # RepVGG-A1 model Pretrained on [ImageNette](https://github.com/fastai/imagenette). The RepVGG architecture was introduced in [this paper](https://arxiv.org/pdf/2101.03697.pdf). ## Model description The core idea of the author is to distinguish the training architecture (with shortcut connections), from the inference one (a pure highway network). By designing the residual block, the training architecture can be reparametrized into a simple sequence of convolutions and non-linear activations. ## Installation ### Prerequisites Python 3.6 (or higher) and [pip](https://pip.pypa.io/en/stable/)/[conda](https://docs.conda.io/en/latest/miniconda.html) are required to install Holocron. ### Latest stable release You can install the last stable release of the package using [pypi](https://pypi.org/project/pylocron/) as follows: ```shell pip install pylocron ``` or using [conda](https://anaconda.org/frgfm/pylocron): ```shell conda install -c frgfm pylocron ``` ### Developer mode Alternatively, if you wish to use the latest features of the project that haven't made their way to a release yet, you can install the package from source *(install [Git](https://git-scm.com/book/en/v2/Getting-Started-Installing-Git) first)*: ```shell git clone https://github.com/frgfm/Holocron.git pip install -e Holocron/. ``` ## Usage instructions ```python from PIL import Image from torchvision.transforms import Compose, ConvertImageDtype, Normalize, PILToTensor, Resize from torchvision.transforms.functional import InterpolationMode from holocron.models import model_from_hf_hub model = model_from_hf_hub("frgfm/repvgg_a1").eval() img = Image.open(path_to_an_image).convert("RGB") # Preprocessing config = model.default_cfg transform = Compose([ Resize(config['input_shape'][1:], interpolation=InterpolationMode.BILINEAR), PILToTensor(), ConvertImageDtype(torch.float32), Normalize(config['mean'], config['std']) ]) input_tensor = transform(img).unsqueeze(0) # Inference with torch.inference_mode(): output = model(input_tensor) probs = output.squeeze(0).softmax(dim=0) ``` ## Citation Original paper ```bibtex @article{DBLP:journals/corr/abs-2101-03697, author = {Xiaohan Ding and Xiangyu Zhang and Ningning Ma and Jungong Han and Guiguang Ding and Jian Sun}, title = {RepVGG: Making VGG-style ConvNets Great Again}, journal = {CoRR}, volume = {abs/2101.03697}, year = {2021}, url = {https://arxiv.org/abs/2101.03697}, eprinttype = {arXiv}, eprint = {2101.03697}, timestamp = {Tue, 09 Feb 2021 15:29:34 +0100}, biburl = {https://dblp.org/rec/journals/corr/abs-2101-03697.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` Source of this implementation ```bibtex @software{Fernandez_Holocron_2020, author = {Fernandez, François-Guillaume}, month = {5}, title = {{Holocron}}, url = {https://github.com/frgfm/Holocron}, year = {2020} } ```
frgfm/rexnet1_3x
804d88068e843aff5969c3f97eb75bd86d9f0f8a
2022-07-20T00:54:33.000Z
[ "pytorch", "onnx", "dataset:frgfm/imagenette", "arxiv:2007.00992", "transformers", "image-classification", "license:apache-2.0" ]
image-classification
false
frgfm
null
frgfm/rexnet1_3x
3
null
transformers
21,264
--- license: apache-2.0 tags: - image-classification - pytorch - onnx datasets: - frgfm/imagenette --- # ReXNet-1.3x model Pretrained on [ImageNette](https://github.com/fastai/imagenette). The ReXNet architecture was introduced in [this paper](https://arxiv.org/pdf/2007.00992.pdf). ## Model description The core idea of the author is to add a customized Squeeze-Excitation layer in the residual blocks that will prevent channel redundancy. ## Installation ### Prerequisites Python 3.6 (or higher) and [pip](https://pip.pypa.io/en/stable/)/[conda](https://docs.conda.io/en/latest/miniconda.html) are required to install Holocron. ### Latest stable release You can install the last stable release of the package using [pypi](https://pypi.org/project/pylocron/) as follows: ```shell pip install pylocron ``` or using [conda](https://anaconda.org/frgfm/pylocron): ```shell conda install -c frgfm pylocron ``` ### Developer mode Alternatively, if you wish to use the latest features of the project that haven't made their way to a release yet, you can install the package from source *(install [Git](https://git-scm.com/book/en/v2/Getting-Started-Installing-Git) first)*: ```shell git clone https://github.com/frgfm/Holocron.git pip install -e Holocron/. ``` ## Usage instructions ```python from PIL import Image from torchvision.transforms import Compose, ConvertImageDtype, Normalize, PILToTensor, Resize from torchvision.transforms.functional import InterpolationMode from holocron.models import model_from_hf_hub model = model_from_hf_hub("frgfm/rexnet1_3x").eval() img = Image.open(path_to_an_image).convert("RGB") # Preprocessing config = model.default_cfg transform = Compose([ Resize(config['input_shape'][1:], interpolation=InterpolationMode.BILINEAR), PILToTensor(), ConvertImageDtype(torch.float32), Normalize(config['mean'], config['std']) ]) input_tensor = transform(img).unsqueeze(0) # Inference with torch.inference_mode(): output = model(input_tensor) probs = output.squeeze(0).softmax(dim=0) ``` ## Citation Original paper ```bibtex @article{DBLP:journals/corr/abs-2007-00992, author = {Dongyoon Han and Sangdoo Yun and Byeongho Heo and Young Joon Yoo}, title = {ReXNet: Diminishing Representational Bottleneck on Convolutional Neural Network}, journal = {CoRR}, volume = {abs/2007.00992}, year = {2020}, url = {https://arxiv.org/abs/2007.00992}, eprinttype = {arXiv}, eprint = {2007.00992}, timestamp = {Mon, 06 Jul 2020 15:26:01 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2007-00992.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` Source of this implementation ```bibtex @software{Fernandez_Holocron_2020, author = {Fernandez, François-Guillaume}, month = {5}, title = {{Holocron}}, url = {https://github.com/frgfm/Holocron}, year = {2020} } ```
frgfm/rexnet2_0x
273a37ed4108de444cda1fca6644aea396ac1619
2022-07-20T00:55:41.000Z
[ "pytorch", "onnx", "dataset:frgfm/imagenette", "arxiv:2007.00992", "transformers", "image-classification", "license:apache-2.0" ]
image-classification
false
frgfm
null
frgfm/rexnet2_0x
3
null
transformers
21,265
--- license: apache-2.0 tags: - image-classification - pytorch - onnx datasets: - frgfm/imagenette --- # ReXNet-2.0x model Pretrained on [ImageNette](https://github.com/fastai/imagenette). The ReXNet architecture was introduced in [this paper](https://arxiv.org/pdf/2007.00992.pdf). ## Model description The core idea of the author is to add a customized Squeeze-Excitation layer in the residual blocks that will prevent channel redundancy. ## Installation ### Prerequisites Python 3.6 (or higher) and [pip](https://pip.pypa.io/en/stable/)/[conda](https://docs.conda.io/en/latest/miniconda.html) are required to install Holocron. ### Latest stable release You can install the last stable release of the package using [pypi](https://pypi.org/project/pylocron/) as follows: ```shell pip install pylocron ``` or using [conda](https://anaconda.org/frgfm/pylocron): ```shell conda install -c frgfm pylocron ``` ### Developer mode Alternatively, if you wish to use the latest features of the project that haven't made their way to a release yet, you can install the package from source *(install [Git](https://git-scm.com/book/en/v2/Getting-Started-Installing-Git) first)*: ```shell git clone https://github.com/frgfm/Holocron.git pip install -e Holocron/. ``` ## Usage instructions ```python from PIL import Image from torchvision.transforms import Compose, ConvertImageDtype, Normalize, PILToTensor, Resize from torchvision.transforms.functional import InterpolationMode from holocron.models import model_from_hf_hub model = model_from_hf_hub("frgfm/rexnet2_0x").eval() img = Image.open(path_to_an_image).convert("RGB") # Preprocessing config = model.default_cfg transform = Compose([ Resize(config['input_shape'][1:], interpolation=InterpolationMode.BILINEAR), PILToTensor(), ConvertImageDtype(torch.float32), Normalize(config['mean'], config['std']) ]) input_tensor = transform(img).unsqueeze(0) # Inference with torch.inference_mode(): output = model(input_tensor) probs = output.squeeze(0).softmax(dim=0) ``` ## Citation Original paper ```bibtex @article{DBLP:journals/corr/abs-2007-00992, author = {Dongyoon Han and Sangdoo Yun and Byeongho Heo and Young Joon Yoo}, title = {ReXNet: Diminishing Representational Bottleneck on Convolutional Neural Network}, journal = {CoRR}, volume = {abs/2007.00992}, year = {2020}, url = {https://arxiv.org/abs/2007.00992}, eprinttype = {arXiv}, eprint = {2007.00992}, timestamp = {Mon, 06 Jul 2020 15:26:01 +0200}, biburl = {https://dblp.org/rec/journals/corr/abs-2007-00992.bib}, bibsource = {dblp computer science bibliography, https://dblp.org} } ``` Source of this implementation ```bibtex @software{Fernandez_Holocron_2020, author = {Fernandez, François-Guillaume}, month = {5}, title = {{Holocron}}, url = {https://github.com/frgfm/Holocron}, year = {2020} } ```
fspanda/Electra-Medical-v790000-discriminator
b53c65254e78c7c698ccb2eccf3f3fa7aebaf878
2020-10-31T13:22:33.000Z
[ "pytorch", "electra", "pretraining", "transformers" ]
null
false
fspanda
null
fspanda/Electra-Medical-v790000-discriminator
3
null
transformers
21,266
Entry not found
fspanda/electra-medical-small-generator
881bb4d6c09065604fdb1981f3cfd9abceca1642
2020-10-29T00:33:04.000Z
[ "pytorch", "electra", "fill-mask", "transformers", "autotrain_compatible" ]
fill-mask
false
fspanda
null
fspanda/electra-medical-small-generator
3
null
transformers
21,267
Entry not found
furyhawk/t5-base-finetuned-bbc-headline
8cbd3582fa749b35b11589207adfc3a7674b95e6
2021-10-28T15:44:15.000Z
[ "pytorch", "t5", "text2text-generation", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
text2text-generation
false
furyhawk
null
furyhawk/t5-base-finetuned-bbc-headline
3
null
transformers
21,268
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: t5-base-finetuned-bbc-headline results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-base-finetuned-bbc-headline This model is a fine-tuned version of [t5-base](https://huggingface.co/t5-base) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 12 - eval_batch_size: 12 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:| | No log | 1.0 | 167 | 2.2978 | 31.8313 | 10.3824 | 29.6182 | 29.4336 | 10.3153 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.1 - Datasets 1.12.1 - Tokenizers 0.10.3
furyhawk/t5-small-finetuned-bbc-headline
ce13f19fb3f755a6ba1a16648fb398591d0efe3d
2021-10-28T08:35:00.000Z
[ "pytorch", "t5", "text2text-generation", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
text2text-generation
false
furyhawk
null
furyhawk/t5-small-finetuned-bbc-headline
3
null
transformers
21,269
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: t5-small-finetuned-bbc-headline results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-finetuned-bbc-headline This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 12 - eval_batch_size: 12 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 167 | 3.6454 | 22.4311 | 5.9878 | 20.118 | 20.482 | 18.9009 | ### Framework versions - Transformers 4.11.3 - Pytorch 1.9.1 - Datasets 1.12.1 - Tokenizers 0.10.3
fznmhmmd/bert-base-cased-wikitext2
b5b2ef8a6972e2e668feee617c3ed6cd1f138bb0
2022-02-10T00:37:23.000Z
[ "pytorch", "tensorboard", "bert", "fill-mask", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
fill-mask
false
fznmhmmd
null
fznmhmmd/bert-base-cased-wikitext2
3
null
transformers
21,270
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: bert-base-cased-wikitext2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-base-cased-wikitext2 This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 6.8575 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 7.0964 | 1.0 | 2346 | 7.0532 | | 6.9055 | 2.0 | 4692 | 6.8710 | | 6.8574 | 3.0 | 7038 | 6.8917 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
fznmhmmd/distilbert-base-uncased-finetuned-cola
13b766fa9f051bb9218fe01adce59d78bb7f3768
2022-02-10T04:00:35.000Z
[ "pytorch", "tensorboard", "distilbert", "text-classification", "dataset:glue", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index" ]
text-classification
false
fznmhmmd
null
fznmhmmd/distilbert-base-uncased-finetuned-cola
3
null
transformers
21,271
--- license: apache-2.0 tags: - generated_from_trainer datasets: - glue metrics: - matthews_correlation model-index: - name: distilbert-base-uncased-finetuned-cola results: - task: name: Text Classification type: text-classification dataset: name: glue type: glue args: cola metrics: - name: Matthews Correlation type: matthews_correlation value: 0.5543972545286807 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-cola This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.8273 - Matthews Correlation: 0.5544 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5256 | 1.0 | 535 | 0.5419 | 0.4248 | | 0.3486 | 2.0 | 1070 | 0.5187 | 0.4999 | | 0.2406 | 3.0 | 1605 | 0.6580 | 0.5054 | | 0.1692 | 4.0 | 2140 | 0.7455 | 0.5403 | | 0.1343 | 5.0 | 2675 | 0.8273 | 0.5544 | ### Framework versions - Transformers 4.16.2 - Pytorch 1.10.0+cu111 - Datasets 1.18.3 - Tokenizers 0.11.0
gaetangate/bart-large_genrl_qald9
9dd6cccf9bad98cfa4d9415d83a52d949e76faf3
2022-04-05T15:10:33.000Z
[ "pytorch", "bart", "text2text-generation", "arxiv:2108.07337", "transformers", "license:apache-2.0", "autotrain_compatible" ]
text2text-generation
false
gaetangate
null
gaetangate/bart-large_genrl_qald9
3
null
transformers
21,272
--- license: apache-2.0 --- This model is used in the paper **Generative Relation Linking for Question Answering over Knowledge Bases**. [ArXiv](https://arxiv.org/abs/2108.07337), [GitHub](https://github.com/IBM/kbqa-relation-linking) ## Citation ```bibtex @inproceedings{rossiello-genrl-2021, title={Generative relation linking for question answering over knowledge bases}, author={Rossiello, Gaetano and Mihindukulasooriya, Nandana and Abdelaziz, Ibrahim and Bornea, Mihaela and Gliozzo, Alfio and Naseem, Tahira and Kapanipathi, Pavan}, booktitle={International Semantic Web Conference}, pages={321--337}, year={2021}, organization={Springer}, url = "https://link.springer.com/chapter/10.1007/978-3-030-88361-4_19", doi = "10.1007/978-3-030-88361-4_19" } ```
gagan3012/k2t-test
512c4171907e621cdef39c929cc645b027ad80a4
2021-07-03T02:43:02.000Z
[ "pytorch", "t5", "text2text-generation", "en", "dataset:WebNLG", "dataset:Dart", "transformers", "keytotext", "k2t", "Keywords to Sentences", "license:mit", "model-index", "autotrain_compatible" ]
text2text-generation
false
gagan3012
null
gagan3012/k2t-test
3
null
transformers
21,273
--- language: "en" thumbnail: "Keywords to Sentences" tags: - keytotext - k2t - Keywords to Sentences license: "MIT" datasets: - WebNLG - Dart metrics: - NLG model-index: - name: k2t-test --- <h1 align="center">keytotext</h1> [![pypi Version](https://img.shields.io/pypi/v/keytotext.svg?logo=pypi&logoColor=white)](https://pypi.org/project/keytotext/) [![Downloads](https://static.pepy.tech/personalized-badge/keytotext?period=total&units=none&left_color=grey&right_color=orange&left_text=Pip%20Downloads)](https://pepy.tech/project/keytotext) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/gagan3012/keytotext/blob/master/notebooks/K2T.ipynb) [![Streamlit App](https://static.streamlit.io/badges/streamlit_badge_black_white.svg)](https://share.streamlit.io/gagan3012/keytotext/UI/app.py) [![API Call](https://img.shields.io/badge/-FastAPI-red?logo=fastapi&labelColor=white)](https://github.com/gagan3012/keytotext#api) [![Docker Call](https://img.shields.io/badge/-Docker%20Image-blue?logo=docker&labelColor=white)](https://hub.docker.com/r/gagan30/keytotext) [![HuggingFace](https://img.shields.io/badge/%F0%9F%A4%97-Models%20on%20Hub-yellow)](https://huggingface.co/models?filter=keytotext) [![Documentation Status](https://readthedocs.org/projects/keytotext/badge/?version=latest)](https://keytotext.readthedocs.io/en/latest/?badge=latest) [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black) ![keytotext](https://socialify.git.ci/gagan3012/keytotext/image?description=1&forks=1&language=1&owner=1&stargazers=1&theme=Light) Idea is to build a model which will take keywords as inputs and generate sentences as outputs. Potential use case can include: - Marketing - Search Engine Optimization - Topic generation etc. - Fine tuning of topic modeling models
gagan3012/project-code-py-micro
deb2c093441b590806d04e9f26afd502cddd9fcb
2021-05-21T16:05:34.000Z
[ "pytorch", "jax", "gpt2", "text-generation", "transformers" ]
text-generation
false
gagan3012
null
gagan3012/project-code-py-micro
3
null
transformers
21,274
Entry not found
gaotianyu1350/sup-simcse-roberta-large
60d1659beda57deb19379c34401645afa2d3488f
2021-05-20T16:24:50.000Z
[ "pytorch", "jax", "roberta", "feature-extraction", "transformers" ]
feature-extraction
false
gaotianyu1350
null
gaotianyu1350/sup-simcse-roberta-large
3
null
transformers
21,275
Entry not found
gayanin/bart-paraphrase-pubmed-1.1
a5c6b2e4f5fe80965682517162e49a024f9f8145
2021-11-06T17:23:34.000Z
[ "pytorch", "tensorboard", "bart", "text2text-generation", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
text2text-generation
false
gayanin
null
gayanin/bart-paraphrase-pubmed-1.1
3
null
transformers
21,276
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: bart-paraphrase-pubmed-1.1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bart-paraphrase-pubmed-1.1 This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4236 - Rouge2 Precision: 0.8482 - Rouge2 Recall: 0.673 - Rouge2 Fmeasure: 0.7347 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.6534 | 1.0 | 663 | 0.4641 | 0.8448 | 0.6691 | 0.7313 | | 0.5078 | 2.0 | 1326 | 0.4398 | 0.8457 | 0.6719 | 0.7333 | | 0.4367 | 3.0 | 1989 | 0.4274 | 0.847 | 0.6717 | 0.7335 | | 0.3575 | 4.0 | 2652 | 0.4149 | 0.8481 | 0.6733 | 0.735 | | 0.3319 | 5.0 | 3315 | 0.4170 | 0.8481 | 0.6724 | 0.7343 | | 0.3179 | 6.0 | 3978 | 0.4264 | 0.8484 | 0.6733 | 0.735 | | 0.2702 | 7.0 | 4641 | 0.4207 | 0.8489 | 0.6732 | 0.7353 | | 0.2606 | 8.0 | 5304 | 0.4205 | 0.8487 | 0.6725 | 0.7347 | | 0.2496 | 9.0 | 5967 | 0.4247 | 0.8466 | 0.6717 | 0.7334 | | 0.2353 | 10.0 | 6630 | 0.4236 | 0.8482 | 0.673 | 0.7347 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
gayanin/t5-small-paraphrase-pubmed
e3f124467d5444951035bab48823a0972c5a4ca7
2021-11-06T09:08:16.000Z
[ "pytorch", "tensorboard", "t5", "text2text-generation", "transformers", "generated_from_trainer", "license:apache-2.0", "model-index", "autotrain_compatible" ]
text2text-generation
false
gayanin
null
gayanin/t5-small-paraphrase-pubmed
3
null
transformers
21,277
--- license: apache-2.0 tags: - generated_from_trainer model-index: - name: t5-small-paraphrase-pubmed results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-small-paraphrase-pubmed This model is a fine-tuned version of [t5-small](https://huggingface.co/t5-small) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.4032 - Rouge2 Precision: 0.8281 - Rouge2 Recall: 0.6346 - Rouge2 Fmeasure: 0.6996 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 40 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge2 Precision | Rouge2 Recall | Rouge2 Fmeasure | |:-------------:|:-----:|:-----:|:---------------:|:----------------:|:-------------:|:---------------:| | 0.5253 | 1.0 | 663 | 0.4895 | 0.8217 | 0.6309 | 0.695 | | 0.5385 | 2.0 | 1326 | 0.4719 | 0.822 | 0.6307 | 0.6953 | | 0.5255 | 3.0 | 1989 | 0.4579 | 0.8225 | 0.631 | 0.6954 | | 0.4927 | 4.0 | 2652 | 0.4510 | 0.824 | 0.6315 | 0.6965 | | 0.484 | 5.0 | 3315 | 0.4426 | 0.8254 | 0.6323 | 0.6974 | | 0.4691 | 6.0 | 3978 | 0.4383 | 0.8241 | 0.6311 | 0.6962 | | 0.4546 | 7.0 | 4641 | 0.4319 | 0.8248 | 0.6322 | 0.6969 | | 0.4431 | 8.0 | 5304 | 0.4270 | 0.8254 | 0.633 | 0.6977 | | 0.4548 | 9.0 | 5967 | 0.4257 | 0.8257 | 0.6322 | 0.6976 | | 0.4335 | 10.0 | 6630 | 0.4241 | 0.8271 | 0.6333 | 0.6986 | | 0.4234 | 11.0 | 7293 | 0.4203 | 0.827 | 0.6341 | 0.6992 | | 0.433 | 12.0 | 7956 | 0.4185 | 0.8279 | 0.6347 | 0.6998 | | 0.4108 | 13.0 | 8619 | 0.4161 | 0.8285 | 0.6352 | 0.7004 | | 0.4101 | 14.0 | 9282 | 0.4133 | 0.8289 | 0.6356 | 0.7008 | | 0.4155 | 15.0 | 9945 | 0.4149 | 0.8279 | 0.635 | 0.6998 | | 0.3991 | 16.0 | 10608 | 0.4124 | 0.8289 | 0.6353 | 0.7005 | | 0.3962 | 17.0 | 11271 | 0.4113 | 0.829 | 0.6353 | 0.7006 | | 0.3968 | 18.0 | 11934 | 0.4114 | 0.8285 | 0.6352 | 0.7002 | | 0.3962 | 19.0 | 12597 | 0.4100 | 0.8282 | 0.6346 | 0.6998 | | 0.3771 | 20.0 | 13260 | 0.4078 | 0.829 | 0.6352 | 0.7005 | | 0.3902 | 21.0 | 13923 | 0.4083 | 0.8295 | 0.6351 | 0.7006 | | 0.3811 | 22.0 | 14586 | 0.4077 | 0.8276 | 0.6346 | 0.6995 | | 0.38 | 23.0 | 15249 | 0.4076 | 0.8281 | 0.6346 | 0.6997 | | 0.3695 | 24.0 | 15912 | 0.4059 | 0.8277 | 0.6344 | 0.6993 | | 0.3665 | 25.0 | 16575 | 0.4043 | 0.8278 | 0.6343 | 0.6992 | | 0.3728 | 26.0 | 17238 | 0.4059 | 0.8279 | 0.6346 | 0.6994 | | 0.3669 | 27.0 | 17901 | 0.4048 | 0.8271 | 0.6342 | 0.6991 | | 0.3702 | 28.0 | 18564 | 0.4058 | 0.8265 | 0.6338 | 0.6985 | | 0.3674 | 29.0 | 19227 | 0.4049 | 0.8277 | 0.6345 | 0.6993 | | 0.364 | 30.0 | 19890 | 0.4048 | 0.8273 | 0.6341 | 0.699 | | 0.3618 | 31.0 | 20553 | 0.4041 | 0.828 | 0.6349 | 0.6997 | | 0.3609 | 32.0 | 21216 | 0.4040 | 0.8275 | 0.6346 | 0.6994 | | 0.357 | 33.0 | 21879 | 0.4037 | 0.8278 | 0.6348 | 0.6996 | | 0.3638 | 34.0 | 22542 | 0.4038 | 0.8275 | 0.634 | 0.6989 | | 0.3551 | 35.0 | 23205 | 0.4035 | 0.8275 | 0.6344 | 0.6992 | | 0.358 | 36.0 | 23868 | 0.4035 | 0.8279 | 0.6347 | 0.6995 | | 0.3519 | 37.0 | 24531 | 0.4034 | 0.8277 | 0.6343 | 0.6992 | | 0.359 | 38.0 | 25194 | 0.4035 | 0.8281 | 0.6346 | 0.6996 | | 0.3542 | 39.0 | 25857 | 0.4033 | 0.8281 | 0.6346 | 0.6996 | | 0.3592 | 40.0 | 26520 | 0.4032 | 0.8281 | 0.6346 | 0.6996 | ### Framework versions - Transformers 4.12.3 - Pytorch 1.9.0+cu111 - Datasets 1.15.1 - Tokenizers 0.10.3
gchhablani/fnet-base-finetuned-cola
ff6e28a9087bfb65a52146769cbf5a83ecaed1c8
2021-09-20T09:07:35.000Z
[ "pytorch", "tensorboard", "fnet", "text-classification", "en", "dataset:glue", "arxiv:2105.03824", "transformers", "generated_from_trainer", "fnet-bert-base-comparison", "license:apache-2.0", "model-index" ]
text-classification
false
gchhablani
null
gchhablani/fnet-base-finetuned-cola
3
null
transformers
21,278
--- language: - en license: apache-2.0 tags: - generated_from_trainer - fnet-bert-base-comparison datasets: - glue metrics: - matthews_correlation model-index: - name: fnet-base-finetuned-cola results: - task: name: Text Classification type: text-classification dataset: name: GLUE COLA type: glue args: cola metrics: - name: Matthews Correlation type: matthews_correlation value: 0.35940659235571387 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-cola This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE COLA dataset. It achieves the following results on the evaluation set: - Loss: 0.5929 - Matthews Correlation: 0.3594 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name cola \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir fnet-base-finetuned-cola \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Matthews Correlation | |:-------------:|:-----:|:----:|:---------------:|:--------------------:| | 0.5895 | 1.0 | 535 | 0.6146 | 0.1699 | | 0.4656 | 2.0 | 1070 | 0.5667 | 0.3047 | | 0.3329 | 3.0 | 1605 | 0.5929 | 0.3594 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
gchhablani/fnet-base-finetuned-qnli
84537ec01a19c38081f4d4dee8c32d1fc3e9ca85
2021-09-20T09:08:18.000Z
[ "pytorch", "tensorboard", "fnet", "text-classification", "en", "dataset:glue", "arxiv:2105.03824", "transformers", "generated_from_trainer", "fnet-bert-base-comparison", "license:apache-2.0", "model-index" ]
text-classification
false
gchhablani
null
gchhablani/fnet-base-finetuned-qnli
3
null
transformers
21,279
--- language: - en license: apache-2.0 tags: - generated_from_trainer - fnet-bert-base-comparison datasets: - glue metrics: - accuracy model-index: - name: fnet-base-finetuned-qnli results: - task: name: Text Classification type: text-classification dataset: name: GLUE QNLI type: glue args: qnli metrics: - name: Accuracy type: accuracy value: 0.8438586857038257 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # fnet-base-finetuned-qnli This model is a fine-tuned version of [google/fnet-base](https://huggingface.co/google/fnet-base) on the GLUE QNLI dataset. It achieves the following results on the evaluation set: - Loss: 0.4746 - Accuracy: 0.8439 The model was fine-tuned to compare [google/fnet-base](https://huggingface.co/google/fnet-base) as introduced in [this paper](https://arxiv.org/abs/2105.03824) against [bert-base-cased](https://huggingface.co/bert-base-cased). ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure This model is trained using the [run_glue](https://github.com/huggingface/transformers/blob/master/examples/pytorch/text-classification/run_glue.py) script. The following command was used: ```bash #!/usr/bin/bash python ../run_glue.py \\n --model_name_or_path google/fnet-base \\n --task_name qnli \\n --do_train \\n --do_eval \\n --max_seq_length 512 \\n --per_device_train_batch_size 16 \\n --learning_rate 2e-5 \\n --num_train_epochs 3 \\n --output_dir fnet-base-finetuned-qnli \\n --push_to_hub \\n --hub_strategy all_checkpoints \\n --logging_strategy epoch \\n --save_strategy epoch \\n --evaluation_strategy epoch \\n``` ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:--------:| | 0.4597 | 1.0 | 6547 | 0.3713 | 0.8411 | | 0.3252 | 2.0 | 13094 | 0.3781 | 0.8420 | | 0.2243 | 3.0 | 19641 | 0.4746 | 0.8439 | ### Framework versions - Transformers 4.11.0.dev0 - Pytorch 1.9.0 - Datasets 1.12.1 - Tokenizers 0.10.3
geninhu/xls-asr-vi-40h
34f61a549d63d89a457091440f35cb30ada0e5fb
2022-03-23T18:27:13.000Z
[ "pytorch", "tensorboard", "wav2vec2", "automatic-speech-recognition", "vi", "dataset:mozilla-foundation/common_voice_7_0", "transformers", "common-voice", "hf-asr-leaderboard", "robust-speech-event", "license:apache-2.0", "model-index" ]
automatic-speech-recognition
false
geninhu
null
geninhu/xls-asr-vi-40h
3
null
transformers
21,280
--- license: apache-2.0 language: - vi tags: - automatic-speech-recognition - common-voice - hf-asr-leaderboard - robust-speech-event datasets: - mozilla-foundation/common_voice_7_0 model-index: - name: xls-asr-vi-40h results: - task: name: Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice 7.0 type: mozilla-foundation/common_voice_7_0 args: vi metrics: - name: Test WER (with Language model) type: wer value: 56.57 --- # xls-asr-vi-40h This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common voice 7.0 vi & private dataset. It achieves the following results on the evaluation set (Without Language Model): - Loss: 1.1177 - Wer: 60.58 ## Evaluation Please run the eval.py file ```bash !python eval_custom.py --model_id geninhu/xls-asr-vi-40h --dataset mozilla-foundation/common_voice_7_0 --config vi --split test ``` ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1500 - num_epochs: 50.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:-----:|:---------------:|:------:| | 23.3878 | 0.93 | 1500 | 21.9179 | 1.0 | | 8.8862 | 1.85 | 3000 | 6.0599 | 1.0 | | 4.3701 | 2.78 | 4500 | 4.3837 | 1.0 | | 4.113 | 3.7 | 6000 | 4.2698 | 0.9982 | | 3.9666 | 4.63 | 7500 | 3.9726 | 0.9989 | | 3.5965 | 5.56 | 9000 | 3.7124 | 0.9975 | | 3.3944 | 6.48 | 10500 | 3.5005 | 1.0057 | | 3.304 | 7.41 | 12000 | 3.3710 | 1.0043 | | 3.2482 | 8.33 | 13500 | 3.4201 | 1.0155 | | 3.212 | 9.26 | 15000 | 3.3732 | 1.0151 | | 3.1778 | 10.19 | 16500 | 3.2763 | 1.0009 | | 3.1027 | 11.11 | 18000 | 3.1943 | 1.0025 | | 2.9905 | 12.04 | 19500 | 2.8082 | 0.9703 | | 2.7095 | 12.96 | 21000 | 2.4993 | 0.9302 | | 2.4862 | 13.89 | 22500 | 2.3072 | 0.9140 | | 2.3271 | 14.81 | 24000 | 2.1398 | 0.8949 | | 2.1968 | 15.74 | 25500 | 2.0594 | 0.8817 | | 2.111 | 16.67 | 27000 | 1.9404 | 0.8630 | | 2.0387 | 17.59 | 28500 | 1.8895 | 0.8497 | | 1.9504 | 18.52 | 30000 | 1.7961 | 0.8315 | | 1.9039 | 19.44 | 31500 | 1.7433 | 0.8213 | | 1.8342 | 20.37 | 33000 | 1.6790 | 0.7994 | | 1.7824 | 21.3 | 34500 | 1.6291 | 0.7825 | | 1.7359 | 22.22 | 36000 | 1.5783 | 0.7706 | | 1.7053 | 23.15 | 37500 | 1.5248 | 0.7492 | | 1.6504 | 24.07 | 39000 | 1.4930 | 0.7406 | | 1.6263 | 25.0 | 40500 | 1.4572 | 0.7348 | | 1.5893 | 25.93 | 42000 | 1.4202 | 0.7161 | | 1.5669 | 26.85 | 43500 | 1.3987 | 0.7143 | | 1.5277 | 27.78 | 45000 | 1.3512 | 0.6991 | | 1.501 | 28.7 | 46500 | 1.3320 | 0.6879 | | 1.4781 | 29.63 | 48000 | 1.3112 | 0.6788 | | 1.4477 | 30.56 | 49500 | 1.2850 | 0.6657 | | 1.4483 | 31.48 | 51000 | 1.2813 | 0.6633 | | 1.4065 | 32.41 | 52500 | 1.2475 | 0.6541 | | 1.3779 | 33.33 | 54000 | 1.2244 | 0.6503 | | 1.3788 | 34.26 | 55500 | 1.2116 | 0.6407 | | 1.3428 | 35.19 | 57000 | 1.1938 | 0.6352 | | 1.3453 | 36.11 | 58500 | 1.1927 | 0.6340 | | 1.3137 | 37.04 | 60000 | 1.1699 | 0.6252 | | 1.2984 | 37.96 | 61500 | 1.1666 | 0.6229 | | 1.2927 | 38.89 | 63000 | 1.1585 | 0.6188 | | 1.2919 | 39.81 | 64500 | 1.1618 | 0.6190 | | 1.293 | 40.74 | 66000 | 1.1479 | 0.6181 | | 1.2853 | 41.67 | 67500 | 1.1423 | 0.6202 | | 1.2687 | 42.59 | 69000 | 1.1315 | 0.6131 | | 1.2603 | 43.52 | 70500 | 1.1333 | 0.6128 | | 1.2577 | 44.44 | 72000 | 1.1191 | 0.6079 | | 1.2435 | 45.37 | 73500 | 1.1177 | 0.6079 | | 1.251 | 46.3 | 75000 | 1.1211 | 0.6092 | | 1.2482 | 47.22 | 76500 | 1.1177 | 0.6060 | | 1.2422 | 48.15 | 78000 | 1.1227 | 0.6097 | | 1.2485 | 49.07 | 79500 | 1.1187 | 0.6071 | | 1.2425 | 50.0 | 81000 | 1.1177 | 0.6058 | ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.17.1.dev0 - Tokenizers 0.11.0
ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR
49fea9ac55eacfebb47d84b018360ade519c3703
2022-01-25T21:02:42.000Z
[ "pytorch", "xlm-roberta", "dataset:ghadeermobasher/BC5CDR-Chemical-Disease", "adapter-transformers", "adapterhub:other" ]
null
false
ghadeermobasher
null
ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR
3
null
adapter-transformers
21,281
--- tags: - adapter-transformers - adapterhub:other - xlm-roberta datasets: - ghadeermobasher/BC5CDR-Chemical-Disease --- # Adapter `ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR` for ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR An [adapter](https://adapterhub.ml) for the `ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR` model that was trained on the [other](https://adapterhub.ml/explore/other/) dataset. This adapter was created for usage with the **[adapter-transformers](https://github.com/Adapter-Hub/adapter-transformers)** library. ## Usage First, install `adapter-transformers`: ``` pip install -U adapter-transformers ``` _Note: adapter-transformers is a fork of transformers that acts as a drop-in replacement with adapter support. [More](https://docs.adapterhub.ml/installation.html)_ Now, the adapter can be loaded and activated like this: ```python from transformers import AutoModelWithHeads model = AutoModelWithHeads.from_pretrained("ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR") adapter_name = model.load_adapter("ghadeermobasher/BC5CDR-Chemical-Disease-balanced-SapBERT-UMLS-2020AB-all-lang-from-XLMR", source="hf", set_active=True) ``` ## Architecture & Training <!-- Add some description here --> ## Evaluation results <!-- Add some description here --> ## Citation <!-- Add some description here -->
ghazikhanihamed/TooT-BERT-M
4f3023e758f962960cfee52f6f6587e45111fe3f
2022-02-20T13:16:41.000Z
[ "pytorch", "bert", "text-classification", "transformers" ]
text-classification
false
ghazikhanihamed
null
ghazikhanihamed/TooT-BERT-M
3
null
transformers
21,282
Entry not found
glasses/eca_resnet101d
3f05bc515f6ba9bf5c2009d5860a5bc66874a984
2021-11-30T20:25:21.000Z
[ "pytorch", "transformers" ]
null
false
glasses
null
glasses/eca_resnet101d
3
null
transformers
21,283
Entry not found
glasses/eca_resnet26t
733959db40c94fde4f26c6c63bfdf8ed01454e8e
2021-11-30T20:21:22.000Z
[ "pytorch", "dataset:imagenet", "arxiv:1512.03385", "arxiv:1812.01187", "transformers", "image-classification", "license:apache-2.0" ]
image-classification
false
glasses
null
glasses/eca_resnet26t
3
null
transformers
21,284
--- license: apache-2.0 tags: - image-classification datasets: - imagenet --- # eca_resnet26t Implementation of ResNet proposed in [Deep Residual Learning for Image Recognition](https://arxiv.org/abs/1512.03385) ``` python ResNet.resnet18() ResNet.resnet26() ResNet.resnet34() ResNet.resnet50() ResNet.resnet101() ResNet.resnet152() ResNet.resnet200() Variants (d) proposed in `Bag of Tricks for Image Classification with Convolutional Neural Networks <https://arxiv.org/pdf/1812.01187.pdf`_ ResNet.resnet26d() ResNet.resnet34d() ResNet.resnet50d() # You can construct your own one by chaning `stem` and `block` resnet101d = ResNet.resnet101(stem=ResNetStemC, block=partial(ResNetBottleneckBlock, shortcut=ResNetShorcutD)) ``` Examples: ``` python # change activation ResNet.resnet18(activation = nn.SELU) # change number of classes (default is 1000 ) ResNet.resnet18(n_classes=100) # pass a different block ResNet.resnet18(block=SENetBasicBlock) # change the steam model = ResNet.resnet18(stem=ResNetStemC) change shortcut model = ResNet.resnet18(block=partial(ResNetBasicBlock, shortcut=ResNetShorcutD)) # store each feature x = torch.rand((1, 3, 224, 224)) # get features model = ResNet.resnet18() # first call .features, this will activate the forward hooks and tells the model you'll like to get the features model.encoder.features model(torch.randn((1,3,224,224))) # get the features from the encoder features = model.encoder.features print([x.shape for x in features]) #[torch.Size([1, 64, 112, 112]), torch.Size([1, 64, 56, 56]), torch.Size([1, 128, 28, 28]), torch.Size([1, 256, 14, 14])] ```
glasses/efficientnet_b0
fc406a30db016b0a9862403ff4d31c15de8b8729
2021-12-01T08:07:32.000Z
[ "pytorch", "arxiv:1905.11946", "transformers" ]
null
false
glasses
null
glasses/efficientnet_b0
3
null
transformers
21,285
# efficientnet_b0 Implementation of EfficientNet proposed in [EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks](https://arxiv.org/abs/1905.11946) ![image](https://github.com/FrancescoSaverioZuppichini/glasses/blob/develop/docs/_static/images/EfficientNet.png?raw=true) The basic architecture is similar to MobileNetV2 as was computed by using [Progressive Neural Architecture Search](https://arxiv.org/abs/1905.11946) . The following table shows the basic architecture (EfficientNet-efficientnet\_b0): ![image](https://github.com/FrancescoSaverioZuppichini/glasses/blob/develop/docs/_static/images/EfficientNetModelsTable.jpeg?raw=true) Then, the architecture is scaled up from [-efficientnet\_b0]{.title-ref} to [-efficientnet\_b7]{.title-ref} using compound scaling. ![image](https://github.com/FrancescoSaverioZuppichini/glasses/blob/develop/docs/_static/images/EfficientNetScaling.jpg?raw=true) ``` python EfficientNet.efficientnet_b0() EfficientNet.efficientnet_b1() EfficientNet.efficientnet_b2() EfficientNet.efficientnet_b3() EfficientNet.efficientnet_b4() EfficientNet.efficientnet_b5() EfficientNet.efficientnet_b6() EfficientNet.efficientnet_b7() EfficientNet.efficientnet_b8() EfficientNet.efficientnet_l2() ``` Examples: ``` python EfficientNet.efficientnet_b0(activation = nn.SELU) # change number of classes (default is 1000 ) EfficientNet.efficientnet_b0(n_classes=100) # pass a different block EfficientNet.efficientnet_b0(block=...) # store each feature x = torch.rand((1, 3, 224, 224)) model = EfficientNet.efficientnet_b0() # first call .features, this will activate the forward hooks and tells the model you'll like to get the features model.encoder.features model(torch.randn((1,3,224,224))) # get the features from the encoder features = model.encoder.features print([x.shape for x in features]) # [torch.Size([1, 32, 112, 112]), torch.Size([1, 24, 56, 56]), torch.Size([1, 40, 28, 28]), torch.Size([1, 80, 14, 14])] ```
glasses/regnetx_064
2c120dec7b430d379b1379d4982a9fb13009ea06
2021-11-30T20:28:31.000Z
[ "pytorch", "transformers" ]
null
false
glasses
null
glasses/regnetx_064
3
null
transformers
21,286
Entry not found
glasses/regnety_004
731fb1f06d19147de115ac20387ad41d5aad98f9
2021-12-01T07:45:42.000Z
[ "pytorch", "arxiv:2003.13678", "transformers" ]
null
false
glasses
null
glasses/regnety_004
3
null
transformers
21,287
# regnety_004 Implementation of RegNet proposed in [Designing Network Design Spaces](https://arxiv.org/abs/2003.13678) The main idea is to start with a high dimensional search space and iteratively reduce the search space by empirically apply constrains based on the best performing models sampled by the current search space. The resulting models are light, accurate, and faster than EfficientNets (up to 5x times!) For example, to go from $AnyNet_A$ to $AnyNet_B$ they fixed the bottleneck ratio $b_i$ for all stage $i$. The following table shows all the restrictions applied from one search space to the next one. ![image](https://github.com/FrancescoSaverioZuppichini/glasses/blob/develop/docs/_static/images/RegNetDesignSpaceTable.png?raw=true) The paper is really well written and very interesting, I highly recommended read it. ``` python ResNet.regnetx_002() ResNet.regnetx_004() ResNet.regnetx_006() ResNet.regnetx_008() ResNet.regnetx_016() ResNet.regnetx_040() ResNet.regnetx_064() ResNet.regnetx_080() ResNet.regnetx_120() ResNet.regnetx_160() ResNet.regnetx_320() # Y variants (with SE) ResNet.regnety_002() # ... ResNet.regnetx_320() You can easily customize your model ``` Examples: ``` python # change activation RegNet.regnetx_004(activation = nn.SELU) # change number of classes (default is 1000 ) RegNet.regnetx_004(n_classes=100) # pass a different block RegNet.regnetx_004(block=RegNetYBotteneckBlock) # change the steam model = RegNet.regnetx_004(stem=ResNetStemC) change shortcut model = RegNet.regnetx_004(block=partial(RegNetYBotteneckBlock, shortcut=ResNetShorcutD)) # store each feature x = torch.rand((1, 3, 224, 224)) # get features model = RegNet.regnetx_004() # first call .features, this will activate the forward hooks and tells the model you'll like to get the features model.encoder.features model(torch.randn((1,3,224,224))) # get the features from the encoder features = model.encoder.features print([x.shape for x in features]) #[torch.Size([1, 32, 112, 112]), torch.Size([1, 32, 56, 56]), torch.Size([1, 64, 28, 28]), torch.Size([1, 160, 14, 14])] ```
glob-asr/base-spanish-asr
969f11fa5a158fd8ab0f917d066aee35313d4973
2022-01-27T03:35:42.000Z
[ "pytorch", "wav2vec2", "automatic-speech-recognition", "dataset:common_voice", "transformers", "generated_from_trainer", "model-index" ]
automatic-speech-recognition
false
glob-asr
null
glob-asr/base-spanish-asr
3
null
transformers
21,288
--- tags: - generated_from_trainer datasets: - common_voice model-index: - name: wav2vec2-large-xls-r-300m-spanish-custom results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-large-xls-r-300m-spanish-custom This model was trained from scratch on the common_voice dataset. It achieves the following results on the evaluation set: - eval_loss: 0.2245 - eval_wer: 0.2082 - eval_runtime: 801.6784 - eval_samples_per_second: 18.822 - eval_steps_per_second: 2.354 - epoch: 0.76 - step: 8400 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 200 - num_epochs: 10 - mixed_precision_training: Native AMP ### Framework versions - Transformers 4.16.0.dev0 - Pytorch 1.10.1+cu102 - Datasets 1.17.1.dev0 - Tokenizers 0.11.0
gmihaila/distilbert-base-uncased
bc55bb4c1b1887a9a20665af45e85248c099e3e2
2020-11-05T02:26:13.000Z
[ "pytorch", "distilbert", "text-classification", "transformers" ]
text-classification
false
gmihaila
null
gmihaila/distilbert-base-uncased
3
null
transformers
21,289
Entry not found
Language-Media-Lab/byt5-small-jpn-ain-mt
adec850485a2b1a7ecf773664616d25256d285cd
2022-02-04T13:02:58.000Z
[ "pytorch", "t5", "text2text-generation", "jpn", "ain", "transformers", "translation", "autotrain_compatible" ]
translation
false
Language-Media-Lab
null
Language-Media-Lab/byt5-small-jpn-ain-mt
3
1
transformers
21,290
--- language: - jpn - ain tags: - translation --- Byt5-small-jpn-ain-mt is a machine translation model pretrained with [Google's ByT5-small](https://huggingface.co/google/byt5-small) and fine-tuned on bilingual datasets crawled from the Web. It translates Japanese to Ainu language.
Language-Media-Lab/mt5-small-ain-jpn-mt
4ff5334f16664e52944d45abce23ace013823c8f
2022-02-04T13:20:55.000Z
[ "pytorch", "mt5", "text2text-generation", "jpn", "ain", "transformers", "translation", "autotrain_compatible" ]
translation
false
Language-Media-Lab
null
Language-Media-Lab/mt5-small-ain-jpn-mt
3
null
transformers
21,291
--- language: - jpn - ain tags: - translation --- mt5-small-ain-jpn-mt is a machine translation model pretrained with [Google's mT5-small](https://huggingface.co/google/mt5-small) and fine-tuned on bilingual datasets crawled from the Web. It translates Ainu language to Japanese.
google/multiberts-seed_2-step_400k
25e539be0f2e4ef883d9dce4170966c57997411f
2021-11-06T01:42:45.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_2", "multiberts-seed_2-step_400k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_2-step_400k
3
null
transformers
21,292
--- language: en tags: - multiberts - multiberts-seed_2 - multiberts-seed_2-step_400k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 2, Step 400k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #2, captured at step 400k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_2-step_400k') model = TFBertModel.from_pretrained("google/multiberts-seed_2-step_400k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_2-step_400k') model = BertModel.from_pretrained("google/multiberts-seed_2-step_400k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_3-step_1200k
4bdb0948f7cb02bf0b9b563da6ab50a3ef281fc6
2021-11-06T02:46:09.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_3", "multiberts-seed_3-step_1200k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_3-step_1200k
3
null
transformers
21,293
--- language: en tags: - multiberts - multiberts-seed_3 - multiberts-seed_3-step_1200k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 3, Step 1200k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #3, captured at step 1200k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_3-step_1200k') model = TFBertModel.from_pretrained("google/multiberts-seed_3-step_1200k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_3-step_1200k') model = BertModel.from_pretrained("google/multiberts-seed_3-step_1200k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_4-step_1000k
e3bc2d4b5836d22bc6c47a7974dd8a6ac20703e8
2021-11-06T03:34:00.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_4", "multiberts-seed_4-step_1000k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_4-step_1000k
3
null
transformers
21,294
--- language: en tags: - multiberts - multiberts-seed_4 - multiberts-seed_4-step_1000k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 4, Step 1000k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #4, captured at step 1000k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1000k') model = TFBertModel.from_pretrained("google/multiberts-seed_4-step_1000k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1000k') model = BertModel.from_pretrained("google/multiberts-seed_4-step_1000k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_4-step_100k
7abf3add5f4ad0b49e16d140d5bbff4993be7fa8
2021-11-06T03:10:45.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_4", "multiberts-seed_4-step_100k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_4-step_100k
3
null
transformers
21,295
--- language: en tags: - multiberts - multiberts-seed_4 - multiberts-seed_4-step_100k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 4, Step 100k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #4, captured at step 100k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_100k') model = TFBertModel.from_pretrained("google/multiberts-seed_4-step_100k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_100k') model = BertModel.from_pretrained("google/multiberts-seed_4-step_100k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_4-step_1200k
3eb428990041706774714606a2cfbd8b1d45b639
2021-11-06T03:37:29.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_4", "multiberts-seed_4-step_1200k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_4-step_1200k
3
null
transformers
21,296
--- language: en tags: - multiberts - multiberts-seed_4 - multiberts-seed_4-step_1200k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 4, Step 1200k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #4, captured at step 1200k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1200k') model = TFBertModel.from_pretrained("google/multiberts-seed_4-step_1200k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1200k') model = BertModel.from_pretrained("google/multiberts-seed_4-step_1200k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_4-step_1300k
665b3b6590358f9dc5439bb4ce654c548dc2f5dc
2021-11-06T03:39:13.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_4", "multiberts-seed_4-step_1300k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_4-step_1300k
3
null
transformers
21,297
--- language: en tags: - multiberts - multiberts-seed_4 - multiberts-seed_4-step_1300k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 4, Step 1300k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #4, captured at step 1300k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1300k') model = TFBertModel.from_pretrained("google/multiberts-seed_4-step_1300k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1300k') model = BertModel.from_pretrained("google/multiberts-seed_4-step_1300k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_4-step_1400k
ca0382fae1a80559b52fdee7428c80d8527573e9
2021-11-06T03:40:51.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_4", "multiberts-seed_4-step_1400k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_4-step_1400k
3
null
transformers
21,298
--- language: en tags: - multiberts - multiberts-seed_4 - multiberts-seed_4-step_1400k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 4, Step 1400k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #4, captured at step 1400k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1400k') model = TFBertModel.from_pretrained("google/multiberts-seed_4-step_1400k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1400k') model = BertModel.from_pretrained("google/multiberts-seed_4-step_1400k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```
google/multiberts-seed_4-step_1500k
d90196e1a5f0d023bda6fc316d255555d92fee22
2021-11-06T03:42:29.000Z
[ "pytorch", "tf", "bert", "pretraining", "en", "arxiv:2106.16163", "arxiv:1908.08962", "transformers", "multiberts", "multiberts-seed_4", "multiberts-seed_4-step_1500k", "license:apache-2.0" ]
null
false
google
null
google/multiberts-seed_4-step_1500k
3
null
transformers
21,299
--- language: en tags: - multiberts - multiberts-seed_4 - multiberts-seed_4-step_1500k license: apache-2.0 --- # MultiBERTs, Intermediate Checkpoint - Seed 4, Step 1500k MultiBERTs is a collection of checkpoints and a statistical library to support robust research on BERT. We provide 25 BERT-base models trained with similar hyper-parameters as [the original BERT model](https://github.com/google-research/bert) but with different random seeds, which causes variations in the initial weights and order of training instances. The aim is to distinguish findings that apply to a specific artifact (i.e., a particular instance of the model) from those that apply to the more general procedure. We also provide 140 intermediate checkpoints captured during the course of pre-training (we saved 28 checkpoints for the first 5 runs). The models were originally released through [http://goo.gle/multiberts](http://goo.gle/multiberts). We describe them in our paper [The MultiBERTs: BERT Reproductions for Robustness Analysis](https://arxiv.org/abs/2106.16163). This is model #4, captured at step 1500k (max: 2000k, i.e., 2M steps). ## Model Description This model was captured during a reproduction of [BERT-base uncased](https://github.com/google-research/bert), for English: it is a Transformers model pretrained on a large corpus of English data, using the Masked Language Modelling (MLM) and the Next Sentence Prediction (NSP) objectives. The intended uses, limitations, training data and training procedure for the fully trained model are similar to [BERT-base uncased](https://github.com/google-research/bert). Two major differences with the original model: * We pre-trained the MultiBERTs models for 2 million steps using sequence length 512 (instead of 1 million steps using sequence length 128 then 512). * We used an alternative version of Wikipedia and Books Corpus, initially collected for [Turc et al., 2019](https://arxiv.org/abs/1908.08962). This is a best-effort reproduction, and so it is probable that differences with the original model have gone unnoticed. The performance of MultiBERTs on GLUE after full training is oftentimes comparable to that of original BERT, but we found significant differences on the dev set of SQuAD (MultiBERTs outperforms original BERT). See our [technical report](https://arxiv.org/abs/2106.16163) for more details. ### How to use Using code from [BERT-base uncased](https://huggingface.co/bert-base-uncased), here is an example based on Tensorflow: ``` from transformers import BertTokenizer, TFBertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1500k') model = TFBertModel.from_pretrained("google/multiberts-seed_4-step_1500k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='tf') output = model(encoded_input) ``` PyTorch version: ``` from transformers import BertTokenizer, BertModel tokenizer = BertTokenizer.from_pretrained('google/multiberts-seed_4-step_1500k') model = BertModel.from_pretrained("google/multiberts-seed_4-step_1500k") text = "Replace me by any text you'd like." encoded_input = tokenizer(text, return_tensors='pt') output = model(**encoded_input) ``` ## Citation info ```bibtex @article{sellam2021multiberts, title={The MultiBERTs: BERT Reproductions for Robustness Analysis}, author={Thibault Sellam and Steve Yadlowsky and Jason Wei and Naomi Saphra and Alexander D'Amour and Tal Linzen and Jasmijn Bastings and Iulia Turc and Jacob Eisenstein and Dipanjan Das and Ian Tenney and Ellie Pavlick}, journal={arXiv preprint arXiv:2106.16163}, year={2021} } ```