sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
listlengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
listlengths 0
201
| languages
listlengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
listlengths 0
722
| processed_texts
listlengths 1
723
| tokens_length
listlengths 1
723
| input_texts
listlengths 1
61
| embeddings
listlengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null |
transformers
|
## Overview
**Language model:** deepset/roberta-base-squad2-distilled
**Language:** English
**Training data:** SQuAD 2.0 training set
**Eval data:** SQuAD 2.0 dev set
**Infrastructure**: 4x V100 GPU
**Published**: Dec 8th, 2021
## Details
- haystack's distillation feature was used for training. deepset/roberta-large-squad2 was used as the teacher model.
## Hyperparameters
```
batch_size = 80
n_epochs = 4
max_seq_len = 384
learning_rate = 3e-5
lr_schedule = LinearWarmup
embeds_dropout_prob = 0.1
temperature = 1.5
distillation_loss_weight = 0.75
```
## Performance
```
"exact": 79.8366040596311
"f1": 83.916407079888
```
## Authors
**Timo Möller:** [email protected]
**Julian Risch:** [email protected]
**Malte Pietsch:** [email protected]
**Michel Bartels:** [email protected]
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/deepset-logo-colored.png" class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/haystack-logo-colored.png" class="w-40"/>
</div>
</div>
[deepset](http://deepset.ai/) is the company behind the open-source NLP framework [Haystack](https://haystack.deepset.ai/) which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- [Distilled roberta-base-squad2 (aka "tinyroberta-squad2")]([https://huggingface.co/deepset/tinyroberta-squad2)
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="https://github.com/deepset-ai/haystack">GitHub</a></strong> repo and <strong><a href="https://docs.haystack.deepset.ai">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="https://haystack.deepset.ai/community">Discord community open to everyone!</a></strong></p>
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "en", "license": "mit", "tags": ["exbert"], "datasets": ["squad_v2"], "thumbnail": "https://thumb.tildacdn.com/tild3433-3637-4830-a533-353833613061/-/resize/720x/-/format/webp/germanquad.jpg", "model-index": [{"name": "deepset/roberta-base-squad2-distilled", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 80.8593, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMzVjNzkxNmNiNDkzNzdiYjJjZGM3ZTViMGJhOGM2ZjFmYjg1MjYxMDM2YzM5NWMwNDIyYzNlN2QwNGYyNDMzZSIsInZlcnNpb24iOjF9.Rgww8tf8D7nF2dh2U_DMrFzmp87k8s7RFibrDXSvQyA66PGWXwjlsd1552lzjHnNV5hvHUM1-h3PTuY_5p64BA"}, {"type": "f1", "value": 84.0104, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNTAyZDViNWYzNjA4OWQ5MzgyYmQ2ZDlhNWRhMTIzYTYxYzViMmI4NWE4ZGU5MzVhZTAwNTRlZmRlNWUwMjI0ZSIsInZlcnNpb24iOjF9.Er21BNgJ3jJXLuZtpubTYq9wCwO1i_VLQFwS5ET0e4eAYVVj0aOA40I5FvP5pZac3LjkCnVacxzsFWGCYVmnDA"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad", "type": "squad", "config": "plain_text", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 86.225, "name": "Exact Match"}, {"type": "f1", "value": 92.483, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "adversarial_qa", "type": "adversarial_qa", "config": "adversarialQA", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 29.9, "name": "Exact Match"}, {"type": "f1", "value": 41.183, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_adversarial", "type": "squad_adversarial", "config": "AddOneSent", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 79.071, "name": "Exact Match"}, {"type": "f1", "value": 84.472, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts amazon", "type": "squadshifts", "config": "amazon", "split": "test"}, "metrics": [{"type": "exact_match", "value": 70.733, "name": "Exact Match"}, {"type": "f1", "value": 83.958, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts new_wiki", "type": "squadshifts", "config": "new_wiki", "split": "test"}, "metrics": [{"type": "exact_match", "value": 82.011, "name": "Exact Match"}, {"type": "f1", "value": 91.092, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts nyt", "type": "squadshifts", "config": "nyt", "split": "test"}, "metrics": [{"type": "exact_match", "value": 84.203, "name": "Exact Match"}, {"type": "f1", "value": 91.521, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts reddit", "type": "squadshifts", "config": "reddit", "split": "test"}, "metrics": [{"type": "exact_match", "value": 72.029, "name": "Exact Match"}, {"type": "f1", "value": 83.454, "name": "F1"}]}]}]}
|
question-answering
|
deepset/roberta-base-squad2-distilled
|
[
"transformers",
"pytorch",
"safetensors",
"roberta",
"question-answering",
"exbert",
"en",
"dataset:squad_v2",
"license:mit",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #safetensors #roberta #question-answering #exbert #en #dataset-squad_v2 #license-mit #model-index #endpoints_compatible #has_space #region-us
|
## Overview
Language model: deepset/roberta-base-squad2-distilled
Language: English
Training data: SQuAD 2.0 training set
Eval data: SQuAD 2.0 dev set
Infrastructure: 4x V100 GPU
Published: Dec 8th, 2021
## Details
- haystack's distillation feature was used for training. deepset/roberta-large-squad2 was used as the teacher model.
## Hyperparameters
## Performance
## Authors
Timo Möller: timo.moeller@URL
Julian Risch: URL@URL
Malte Pietsch: malte.pietsch@URL
Michel Bartels: michel.bartels@URL
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
</div>
deepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- Distilled roberta-base-squad2 (aka "tinyroberta-squad2")
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="URL repo and <strong><a href="URL">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="URL community open to everyone!</a></strong></p>
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"## Overview\nLanguage model: deepset/roberta-base-squad2-distilled \nLanguage: English \nTraining data: SQuAD 2.0 training set\nEval data: SQuAD 2.0 dev set\nInfrastructure: 4x V100 GPU \nPublished: Dec 8th, 2021",
"## Details\n- haystack's distillation feature was used for training. deepset/roberta-large-squad2 was used as the teacher model.",
"## Hyperparameters",
"## Performance",
"## Authors\nTimo Möller: timo.moeller@URL \nJulian Risch: URL@URL \nMalte Pietsch: malte.pietsch@URL \nMichel Bartels: michel.bartels@URL",
"## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #roberta #question-answering #exbert #en #dataset-squad_v2 #license-mit #model-index #endpoints_compatible #has_space #region-us \n",
"## Overview\nLanguage model: deepset/roberta-base-squad2-distilled \nLanguage: English \nTraining data: SQuAD 2.0 training set\nEval data: SQuAD 2.0 dev set\nInfrastructure: 4x V100 GPU \nPublished: Dec 8th, 2021",
"## Details\n- haystack's distillation feature was used for training. deepset/roberta-large-squad2 was used as the teacher model.",
"## Hyperparameters",
"## Performance",
"## Authors\nTimo Möller: timo.moeller@URL \nJulian Risch: URL@URL \nMalte Pietsch: malte.pietsch@URL \nMichel Bartels: michel.bartels@URL",
"## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
62,
57,
36,
5,
2,
44,
251,
113
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #roberta #question-answering #exbert #en #dataset-squad_v2 #license-mit #model-index #endpoints_compatible #has_space #region-us \n## Overview\nLanguage model: deepset/roberta-base-squad2-distilled \nLanguage: English \nTraining data: SQuAD 2.0 training set\nEval data: SQuAD 2.0 dev set\nInfrastructure: 4x V100 GPU \nPublished: Dec 8th, 2021## Details\n- haystack's distillation feature was used for training. deepset/roberta-large-squad2 was used as the teacher model.## Hyperparameters## Performance## Authors\nTimo Möller: timo.moeller@URL \nJulian Risch: URL@URL \nMalte Pietsch: malte.pietsch@URL \nMichel Bartels: michel.bartels@URL## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")"
] |
[
-0.07787676900625229,
0.1684291809797287,
-0.0036138398572802544,
0.07279256731271744,
0.07943803817033768,
0.0432332344353199,
0.16618117690086365,
0.11982713639736176,
0.04156779870390892,
0.09644792228937149,
0.04841390997171402,
0.0037298172246664762,
0.0818178802728653,
0.07384370267391205,
0.031549274921417236,
-0.20569780468940735,
0.013525965623557568,
-0.09238342195749283,
-0.06648958474397659,
0.06586915999650955,
0.10335477441549301,
-0.07932671159505844,
0.08861593902111053,
-0.02848537638783455,
-0.036599818617105484,
0.019097931683063507,
-0.028631657361984253,
0.005430182442069054,
0.0690196305513382,
0.04508652165532112,
0.08373910188674927,
0.002765960292890668,
0.05998238921165466,
-0.22804930806159973,
0.026388250291347504,
0.06892943382263184,
-0.009813441894948483,
0.07002264261245728,
0.05719965696334839,
-0.023096248507499695,
0.018596714362502098,
-0.13206925988197327,
0.020414771512150764,
0.04717043414711952,
-0.0973699688911438,
-0.21033045649528503,
-0.11197713017463684,
0.14029647409915924,
-0.029571786522865295,
0.03782225400209427,
-0.0008445645798929036,
0.044475339353084564,
-0.08264052122831345,
0.056057680398225784,
0.18217900395393372,
-0.29572558403015137,
-0.06198741868138313,
0.04594014585018158,
-0.007203306537121534,
0.002078429562970996,
-0.14157848060131073,
0.03888138756155968,
-0.009405151009559631,
0.030531493946909904,
-0.010241303592920303,
-0.020492255687713623,
0.01535018254071474,
0.020919449627399445,
-0.06399453431367874,
-0.014155158773064613,
0.1856699287891388,
0.021038923412561417,
-0.008409407921135426,
-0.1570529192686081,
-0.02238699048757553,
0.02327006682753563,
-0.033240389078855515,
-0.009663312695920467,
0.02241375669836998,
-0.025460384786128998,
0.008713419549167156,
-0.04960770532488823,
-0.08752795308828354,
0.009793935343623161,
0.032284367829561234,
0.12388832122087479,
0.04615333303809166,
0.008147044107317924,
0.009547688066959381,
0.11321178078651428,
-0.01740679144859314,
-0.18275563418865204,
-0.01596902497112751,
-0.06817721575498581,
-0.07412794977426529,
-0.00036217315937392414,
0.0022647478617727757,
0.0018892016960307956,
0.08004625141620636,
0.22542166709899902,
-0.08725898712873459,
0.05281530320644379,
0.03258495032787323,
0.006100475322455168,
-0.006806742399930954,
0.183614119887352,
-0.0709974393248558,
-0.12016239017248154,
0.0036516424734145403,
0.01340441033244133,
0.03651358559727669,
-0.008972084149718285,
-0.07359066605567932,
-0.04096592217683792,
0.014275288209319115,
0.04185646399855614,
0.04611910507082939,
0.03273317217826843,
-0.09138540923595428,
-0.056258462369441986,
0.07902026176452637,
-0.13152608275413513,
0.07978031039237976,
0.09208811819553375,
-0.02929234877228737,
0.078592948615551,
0.020866574719548225,
-0.0012335876235738397,
-0.0603327639400959,
0.0400550402700901,
-0.0019893692806363106,
-0.013807137496769428,
-0.09330533444881439,
-0.0682145282626152,
0.032692570239305496,
0.007235748693346977,
-0.04041169211268425,
-0.07989798486232758,
-0.11919005215167999,
-0.058504149317741394,
0.10549688339233398,
-0.06234955042600632,
-0.0007459375192411244,
-0.025089681148529053,
-0.05407193303108215,
0.07047800719738007,
0.02098238654434681,
-0.002987865125760436,
-0.04978533834218979,
0.0335395485162735,
-0.08760730922222137,
0.005389656405895948,
-0.005166972056031227,
0.024120494723320007,
-0.046034302562475204,
0.008136493153870106,
-0.16552522778511047,
0.07540731132030487,
-0.10650548338890076,
0.04859762266278267,
-0.11294732987880707,
-0.06527397036552429,
-0.022886257618665695,
0.011110067367553711,
0.047778576612472534,
0.13710004091262817,
-0.13120566308498383,
-0.03603629022836685,
0.12382018566131592,
-0.045511238276958466,
-0.0627010390162468,
0.09774211794137955,
-0.04542864114046097,
0.02258353866636753,
0.07051598280668259,
0.11600413918495178,
0.15119586884975433,
-0.1265786737203598,
-0.05347256734967232,
-0.005725139752030373,
0.030087696388363838,
0.10418922454118729,
0.07120545208454132,
-0.049971722066402435,
0.030141638591885567,
0.03454497829079628,
-0.09804517030715942,
-0.007282060571014881,
-0.009487373754382133,
-0.07269555330276489,
-0.009645375423133373,
-0.01986437849700451,
0.09033083915710449,
0.011656283400952816,
0.008398635312914848,
-0.044572118669748306,
-0.13960060477256775,
0.001091663958504796,
0.07405581325292587,
-0.04250010475516319,
0.005881544202566147,
-0.05631173029541969,
0.1012241542339325,
0.05830364301800728,
0.008558177389204502,
-0.13133756816387177,
-0.14921791851520538,
0.06803009659051895,
-0.020494546741247177,
0.09327337890863419,
0.049476508051157,
0.07597459107637405,
0.0402667410671711,
-0.03533973917365074,
-0.08352796733379364,
-0.06595546752214432,
-0.027039483189582825,
0.0195605605840683,
-0.23217223584651947,
-0.03676383197307587,
-0.03219742700457573,
0.08584634214639664,
-0.12136002629995346,
-0.003934744745492935,
0.06059649959206581,
0.17767903208732605,
0.03908614069223404,
-0.016727235168218613,
0.005093080457299948,
-0.00876117404550314,
0.018244918435811996,
-0.055331453680992126,
0.016983570531010628,
0.012883509509265423,
-0.04575226083397865,
-0.007951154373586178,
-0.09043074399232864,
-0.005922240670770407,
0.03932848945260048,
0.01409946009516716,
-0.08468887954950333,
-0.032141271978616714,
-0.04153260961174965,
-0.0036648549139499664,
-0.024274272844195366,
-0.05036359280347824,
0.20634600520133972,
0.01772972010076046,
0.05690530315041542,
-0.0732613131403923,
-0.07809034734964371,
-0.0581100769340992,
-0.004699679557234049,
0.0046131969429552555,
0.14323769509792328,
-0.018499018624424934,
-0.10558418929576874,
0.10764306038618088,
0.15422773361206055,
0.02869153954088688,
0.22340288758277893,
-0.03701106458902359,
-0.035823971033096313,
-0.05678202584385872,
0.06465458869934082,
-0.004937565419822931,
0.13207273185253143,
-0.0026463030371814966,
0.004240076988935471,
0.03999101743102074,
-0.02302045002579689,
-0.011613352224230766,
-0.11559107899665833,
0.050777338445186615,
-0.01329041738063097,
-0.04341167211532593,
0.06097514554858208,
0.020580366253852844,
0.03741694614291191,
0.09117168933153152,
0.04414267838001251,
0.018432561308145523,
-0.0036032230127602816,
-0.05353279411792755,
-0.06740342825651169,
0.17322902381420135,
-0.13089333474636078,
-0.1305026412010193,
-0.10669857263565063,
-0.009294482879340649,
-0.10325697809457779,
-0.0070817437954247,
0.015203578397631645,
-0.09099537879228592,
-0.0939013734459877,
-0.03548425808548927,
0.05899963900446892,
0.08222287148237228,
-0.05662122368812561,
-0.04268082603812218,
0.005687691271305084,
0.044973693788051605,
-0.14608271420001984,
-0.012504090555012226,
0.014257627539336681,
-0.05079234763979912,
-0.014942396432161331,
0.07700133323669434,
0.019881460815668106,
0.05890536308288574,
0.05293726176023483,
0.00012248102575540543,
0.007787925656884909,
0.18443141877651215,
-0.11178698390722275,
0.07422582060098648,
0.11986362934112549,
-0.023387683555483818,
0.05303235352039337,
0.1996569186449051,
0.08034483343362808,
-0.040000349283218384,
-0.01614747755229473,
0.012949157506227493,
0.02188718505203724,
-0.22758051753044128,
-0.09484830498695374,
-0.0729556530714035,
-0.02612851932644844,
0.06909690797328949,
0.053837958723306656,
-0.10386749356985092,
0.01009843684732914,
-0.10169105231761932,
0.013299403712153435,
0.06665512174367905,
0.07422754913568497,
0.06084446981549263,
0.014849198050796986,
0.04239232838153839,
-0.04546116665005684,
-0.026647545397281647,
0.08228560537099838,
0.058834027498960495,
0.09981505572795868,
-0.023392684757709503,
0.1553741693496704,
0.05398394912481308,
0.1391400843858719,
-0.02067355066537857,
0.002403340069577098,
0.01191712450236082,
0.010865679942071438,
0.021309541538357735,
-0.09842972457408905,
-0.010831670835614204,
0.0795358270406723,
0.020834721624851227,
-0.004685020074248314,
-0.016512174159288406,
-0.006092275492846966,
0.11110929399728775,
0.21185502409934998,
0.05947505682706833,
-0.15106619894504547,
-0.0800667330622673,
0.04675682634115219,
-0.005964853335171938,
-0.06541913747787476,
0.0021152785047888756,
0.0734577551484108,
-0.160963237285614,
0.07702523469924927,
-0.019176414236426353,
0.07634024322032928,
-0.047497838735580444,
0.00906893890351057,
0.05602455139160156,
0.14170517027378082,
0.009508701041340828,
0.07969650626182556,
-0.1817331314086914,
0.1340688019990921,
0.03188609704375267,
0.07335632294416428,
-0.06453093141317368,
0.060773443430662155,
0.04514060541987419,
-0.010868444107472897,
0.11926428228616714,
0.006521185860037804,
-0.06501361727714539,
-0.03196291998028755,
-0.07684658467769623,
0.017164016142487526,
0.12631817162036896,
-0.11037293076515198,
0.06262830644845963,
-0.028298189863562584,
-0.027516979724168777,
-0.02888498269021511,
-0.0012730148155242205,
-0.1425594985485077,
-0.13372841477394104,
0.041170939803123474,
-0.06498560309410095,
0.010766707360744476,
-0.06939131021499634,
-0.024859493598341942,
-0.16735054552555084,
0.2024204432964325,
-0.1613527089357376,
-0.05502862110733986,
-0.11606250703334808,
0.042373280972242355,
0.07791841775178909,
-0.10780119895935059,
0.017737863585352898,
-0.03923884779214859,
0.10614661872386932,
-0.0011040449608117342,
-0.09036190062761307,
0.06678580492734909,
-0.0701134204864502,
-0.16129443049430847,
-0.023854507133364677,
0.1634630709886551,
0.021287303417921066,
0.040035180747509,
-0.0021186713129281998,
0.031205780804157257,
-0.03937261924147606,
-0.10672735422849655,
0.021910876035690308,
0.09739711135625839,
0.03452901914715767,
0.1299806386232376,
-0.14574924111366272,
-0.07394210994243622,
-0.023105857893824577,
0.040337756276130676,
0.11083778738975525,
0.14644217491149902,
-0.09729541838169098,
0.1375569850206375,
0.06439558416604996,
-0.0730825737118721,
-0.29145002365112305,
-0.005156269762665033,
0.07636968046426773,
0.047282278537750244,
0.027278102934360504,
-0.20795170962810516,
0.12070910632610321,
-0.015925196930766106,
-0.022622808814048767,
0.036308154463768005,
-0.1872563511133194,
-0.12726643681526184,
0.058587804436683655,
0.0010825915960595012,
-0.12702572345733643,
-0.11549470573663712,
-0.05211522430181503,
-0.05684242025017738,
-0.1145562082529068,
0.07898763567209244,
-0.12360704690217972,
0.0706254169344902,
0.027592185884714127,
0.023388944566249847,
0.025351189076900482,
-0.04243739694356918,
0.11972607672214508,
-0.0002738967596087605,
0.08311359584331512,
-0.07753727585077286,
0.00040921568870544434,
0.03991919755935669,
-0.06657050549983978,
0.07735469937324524,
-0.04383012652397156,
0.04745320603251457,
-0.11307784914970398,
0.012874375097453594,
-0.07676112651824951,
0.079477459192276,
-0.0907120630145073,
-0.030891625210642815,
-0.06568889319896698,
0.11342586576938629,
0.06190281733870506,
0.01730799861252308,
0.10291121155023575,
-0.011732500046491623,
0.06963084638118744,
0.06989405304193497,
0.1459769457578659,
0.04964664950966835,
-0.0827026516199112,
-0.0393652617931366,
0.0014071740442886949,
0.0679108053445816,
-0.10176851600408554,
0.06524449586868286,
0.10565637052059174,
0.01539075467735529,
0.1371479481458664,
-0.005918791517615318,
-0.10072564333677292,
-0.038819629698991776,
0.08694212138652802,
-0.1282743662595749,
-0.20712758600711823,
-0.05192141234874725,
0.016437523066997528,
-0.07780587673187256,
-0.016723990440368652,
0.15971168875694275,
0.021160582080483437,
-0.027601974084973335,
0.03907140716910362,
0.04421142488718033,
-0.03308417648077011,
0.06006709858775139,
0.035364847630262375,
0.0367547981441021,
-0.09584065526723862,
0.10754086822271347,
0.05708349123597145,
0.023776482790708542,
0.03852526471018791,
0.11914245784282684,
-0.06259481608867645,
-0.048700299113988876,
0.04298890382051468,
0.1774279922246933,
-0.004956030752509832,
-0.020591886714100838,
-0.06171558052301407,
-0.08389869332313538,
-0.006973473355174065,
-0.015905413776636124,
0.031779274344444275,
-0.00008534617518307641,
0.0036396016366779804,
-0.01121468935161829,
0.005108633078634739,
0.12465301156044006,
0.054685041308403015,
0.017674941569566727,
-0.08303859829902649,
0.024107813835144043,
-0.018102679401636124,
0.009836029261350632,
-0.012012834660708904,
0.0010015108855441213,
-0.14050330221652985,
-0.033876433968544006,
-0.11931914836168289,
-0.007526821922510862,
-0.0029804063960909843,
0.02273733541369438,
-0.016272466629743576,
-0.056542638689279556,
-0.010064644739031792,
0.036058131605386734,
-0.08547475188970566,
-0.04033763334155083,
-0.016959432512521744,
0.11085440963506699,
-0.15132088959217072,
-0.006871585268527269,
0.059281352907419205,
-0.07654310762882233,
0.11335305124521255,
0.028306622058153152,
-0.005873218644410372,
0.05334573984146118,
-0.2018805593252182,
-0.07385118305683136,
-0.07231678813695908,
0.04448295384645462,
0.01229263748973608,
-0.14879371225833893,
-0.01399242039769888,
-0.029140640050172806,
-0.03644540160894394,
0.0029088587034493685,
-0.00014764607476536185,
-0.09262344241142273,
0.08067873865365982,
-0.025553138926625252,
-0.0402209535241127,
-0.056048326194286346,
0.07098085433244705,
0.011574828997254372,
0.03022197261452675,
0.11443112790584564,
-0.09464055299758911,
0.0679662749171257,
-0.1422310173511505,
-0.004952134098857641,
0.029239356517791748,
-0.03630753606557846,
-0.151781365275383,
-0.02464047260582447,
0.07819522172212601,
-0.025120625272393227,
0.08572819083929062,
-0.04039933159947395,
-0.04222429543733597,
0.03432067856192589,
0.013878321275115013,
-0.030964555218815804,
0.045189861208200455,
0.04656760394573212,
-0.026222646236419678,
0.02527567371726036,
-0.01243350189179182,
-0.03055792674422264,
0.0015956101706251502,
0.03783301264047623,
0.18558160960674286,
0.23036736249923706,
0.10042480379343033,
0.0017195658292621374,
0.10348501056432724,
-0.06256990134716034,
-0.050434816628694534,
0.13265582919120789,
-0.03237580507993698,
0.08822420984506607,
-0.1105843260884285,
0.023077595978975296,
0.09380996227264404,
-0.21039140224456787,
0.0760166347026825,
-0.03139260411262512,
-0.06293150037527084,
-0.08765323460102081,
-0.1809556782245636,
-0.08336415141820908,
-0.0850440263748169,
0.027981536462903023,
-0.12653282284736633,
0.020075540989637375,
0.029444396495819092,
0.0435347855091095,
-0.015255737118422985,
0.1430773138999939,
-0.08934330195188522,
-0.040816087275743484,
0.10068315267562866,
0.044586583971977234,
0.027928277850151062,
0.03301578760147095,
-0.043254803866147995,
-0.019785422831773758,
0.03258952498435974,
0.04251055791974068,
0.04652119055390358,
0.02131153643131256,
-0.041067466139793396,
-0.01255702879279852,
-0.07143925875425339,
0.012070597149431705,
-0.007676937151700258,
-0.022786539047956467,
0.15234331786632538,
0.04182370379567146,
-0.0017288833623751998,
-0.01963590830564499,
0.17504338920116425,
-0.0442490316927433,
-0.1102675274014473,
-0.16822995245456696,
0.05841876566410065,
-0.043373264372348785,
-0.0021995902061462402,
0.01706668734550476,
-0.10290006548166275,
-0.002776696812361479,
0.09778265655040741,
0.2097422480583191,
-0.09899142384529114,
0.0423581525683403,
0.017185520380735397,
0.01457024272531271,
-0.018923938274383545,
0.07832147926092148,
0.026285704225301743,
0.16321976482868195,
-0.042401012033224106,
0.0027905849274247885,
0.016263211145997047,
-0.04151513800024986,
-0.11829211562871933,
0.09033504873514175,
-0.008633412420749664,
-0.05323734134435654,
-0.062381211668252945,
0.08014221489429474,
-0.05297933146357536,
-0.1860283762216568,
0.018415655940771103,
-0.15485234558582306,
-0.16558916866779327,
-0.049821533262729645,
0.09137465804815292,
0.0408126525580883,
0.055653274059295654,
-0.010297871194779873,
-0.046149834990501404,
0.14939729869365692,
-0.005891946144402027,
-0.01619116961956024,
-0.04178657382726669,
0.06787422299385071,
-0.05714579299092293,
0.17694200575351715,
0.005989334546029568,
0.06031327322125435,
0.11557295173406601,
-0.019688230007886887,
-0.09173402935266495,
-0.03672599419951439,
0.09625422954559326,
-0.15755008161067963,
-0.013483108952641487,
0.08276992291212082,
-0.015366552397608757,
0.1209840252995491,
0.10249316692352295,
-0.026539083570241928,
0.04496077448129654,
0.14084026217460632,
-0.0675254687666893,
-0.09940681606531143,
0.07680220156908035,
-0.09878989309072495,
0.14695553481578827,
0.18875491619110107,
-0.023563116788864136,
-0.04280107468366623,
-0.03779737278819084,
0.053248416632413864,
0.02850269339978695,
0.05612708255648613,
-0.032773859798908234,
-0.1508825123310089,
0.05014015734195709,
-0.007571123074740171,
0.084996297955513,
-0.10373011231422424,
-0.09771499037742615,
0.03212399780750275,
0.04827392101287842,
-0.07830901443958282,
0.15364667773246765,
0.05220574513077736,
0.01486039999872446,
-0.011580421589314938,
-0.10343246906995773,
-0.0242074616253376,
0.08541607856750488,
-0.07720708847045898,
-0.04802815988659859
] |
null | null |
transformers
|
# roberta-base for QA
This is the [roberta-base](https://huggingface.co/roberta-base) model, fine-tuned using the [SQuAD2.0](https://huggingface.co/datasets/squad_v2) dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
## Overview
**Language model:** roberta-base
**Language:** English
**Downstream-task:** Extractive QA
**Training data:** SQuAD 2.0
**Eval data:** SQuAD 2.0
**Code:** See [an example QA pipeline on Haystack](https://haystack.deepset.ai/tutorials/first-qa-system)
**Infrastructure**: 4x Tesla v100
## Hyperparameters
```
batch_size = 96
n_epochs = 2
base_LM_model = "roberta-base"
max_seq_len = 386
learning_rate = 3e-5
lr_schedule = LinearWarmup
warmup_proportion = 0.2
doc_stride=128
max_query_length=64
```
## Using a distilled model instead
Please note that we have also released a distilled version of this model called [deepset/tinyroberta-squad2](https://huggingface.co/deepset/tinyroberta-squad2). The distilled model has a comparable prediction quality and runs at twice the speed of the base model.
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
# or
reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
```
For a complete example of ``roberta-base-squad2`` being used for Question Answering, check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
### In Transformers
```python
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
model_name = "deepset/roberta-base-squad2"
# a) Get predictions
nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
QA_input = {
'question': 'Why is model conversion important?',
'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
}
res = nlp(QA_input)
# b) Load model & tokenizer
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
## Performance
Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
```
"exact": 79.87029394424324,
"f1": 82.91251169582613,
"total": 11873,
"HasAns_exact": 77.93522267206478,
"HasAns_f1": 84.02838248389763,
"HasAns_total": 5928,
"NoAns_exact": 81.79983179142137,
"NoAns_f1": 81.79983179142137,
"NoAns_total": 5945
```
## Authors
**Branden Chan:** [email protected]
**Timo Möller:** [email protected]
**Malte Pietsch:** [email protected]
**Tanay Soni:** [email protected]
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/deepset-logo-colored.png" class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/haystack-logo-colored.png" class="w-40"/>
</div>
</div>
[deepset](http://deepset.ai/) is the company behind the open-source NLP framework [Haystack](https://haystack.deepset.ai/) which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- [Distilled roberta-base-squad2 (aka "tinyroberta-squad2")]([https://huggingface.co/deepset/tinyroberta-squad2)
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="https://github.com/deepset-ai/haystack">GitHub</a></strong> repo and <strong><a href="https://docs.haystack.deepset.ai">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="https://haystack.deepset.ai/community">Discord community open to everyone!</a></strong></p>
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "en", "license": "cc-by-4.0", "datasets": ["squad_v2"], "model-index": [{"name": "deepset/roberta-base-squad2", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 79.9309, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMDhhNjg5YzNiZGQ1YTIyYTAwZGUwOWEzZTRiYzdjM2QzYjA3ZTUxNDM1NjE1MTUyMjE1MGY1YzEzMjRjYzVjYiIsInZlcnNpb24iOjF9.EH5JJo8EEFwU7osPz3s7qanw_tigeCFhCXjSfyN0Y1nWVnSfulSxIk_DbAEI5iE80V4EKLyp5-mYFodWvL2KDA"}, {"type": "f1", "value": 82.9501, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMjk5ZDYwOGQyNjNkMWI0OTE4YzRmOTlkY2JjNjQ0YTZkNTMzMzNkYTA0MDFmNmI3NjA3NjNlMjhiMDQ2ZjJjNSIsInZlcnNpb24iOjF9.DDm0LNTkdLbGsue58bg1aH_s67KfbcmkvL-6ZiI2s8IoxhHJMSf29H_uV2YLyevwx900t-MwTVOW3qfFnMMEAQ"}, {"type": "total", "value": 11869, "name": "total", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGFkMmI2ODM0NmY5NGNkNmUxYWViOWYxZDNkY2EzYWFmOWI4N2VhYzY5MGEzMTVhOTU4Zjc4YWViOGNjOWJjMCIsInZlcnNpb24iOjF9.fexrU1icJK5_MiifBtZWkeUvpmFISqBLDXSQJ8E6UnrRof-7cU0s4tX_dIsauHWtUpIHMPZCf5dlMWQKXZuAAA"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad", "type": "squad", "config": "plain_text", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 85.289, "name": "Exact Match"}, {"type": "f1", "value": 91.841, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "adversarial_qa", "type": "adversarial_qa", "config": "adversarialQA", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 29.5, "name": "Exact Match"}, {"type": "f1", "value": 40.367, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_adversarial", "type": "squad_adversarial", "config": "AddOneSent", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 78.567, "name": "Exact Match"}, {"type": "f1", "value": 84.469, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts amazon", "type": "squadshifts", "config": "amazon", "split": "test"}, "metrics": [{"type": "exact_match", "value": 69.924, "name": "Exact Match"}, {"type": "f1", "value": 83.284, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts new_wiki", "type": "squadshifts", "config": "new_wiki", "split": "test"}, "metrics": [{"type": "exact_match", "value": 81.204, "name": "Exact Match"}, {"type": "f1", "value": 90.595, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts nyt", "type": "squadshifts", "config": "nyt", "split": "test"}, "metrics": [{"type": "exact_match", "value": 82.931, "name": "Exact Match"}, {"type": "f1", "value": 90.756, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts reddit", "type": "squadshifts", "config": "reddit", "split": "test"}, "metrics": [{"type": "exact_match", "value": 71.55, "name": "Exact Match"}, {"type": "f1", "value": 82.939, "name": "F1"}]}]}]}
|
question-answering
|
deepset/roberta-base-squad2
|
[
"transformers",
"pytorch",
"tf",
"jax",
"rust",
"safetensors",
"roberta",
"question-answering",
"en",
"dataset:squad_v2",
"license:cc-by-4.0",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #tf #jax #rust #safetensors #roberta #question-answering #en #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us
|
# roberta-base for QA
This is the roberta-base model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
## Overview
Language model: roberta-base
Language: English
Downstream-task: Extractive QA
Training data: SQuAD 2.0
Eval data: SQuAD 2.0
Code: See an example QA pipeline on Haystack
Infrastructure: 4x Tesla v100
## Hyperparameters
## Using a distilled model instead
Please note that we have also released a distilled version of this model called deepset/tinyroberta-squad2. The distilled model has a comparable prediction quality and runs at twice the speed of the base model.
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:
For a complete example of ''roberta-base-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation
### In Transformers
## Performance
Evaluated on the SQuAD 2.0 dev set with the official eval script.
## Authors
Branden Chan: URL@URL
Timo Möller: timo.moeller@URL
Malte Pietsch: malte.pietsch@URL
Tanay Soni: URL@URL
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
</div>
deepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- Distilled roberta-base-squad2 (aka "tinyroberta-squad2")
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="URL repo and <strong><a href="URL">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="URL community open to everyone!</a></strong></p>
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# roberta-base for QA \n\nThis is the roberta-base model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.",
"## Overview\nLanguage model: roberta-base \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Using a distilled model instead\nPlease note that we have also released a distilled version of this model called deepset/tinyroberta-squad2. The distilled model has a comparable prediction quality and runs at twice the speed of the base model.",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''roberta-base-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation",
"### In Transformers",
"## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL",
"## About us\n\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #tf #jax #rust #safetensors #roberta #question-answering #en #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n",
"# roberta-base for QA \n\nThis is the roberta-base model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.",
"## Overview\nLanguage model: roberta-base \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Using a distilled model instead\nPlease note that we have also released a distilled version of this model called deepset/tinyroberta-squad2. The distilled model has a comparable prediction quality and runs at twice the speed of the base model.",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''roberta-base-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation",
"### In Transformers",
"## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL",
"## About us\n\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
71,
59,
60,
5,
57,
3,
84,
6,
19,
41,
251,
113
] |
[
"passage: TAGS\n#transformers #pytorch #tf #jax #rust #safetensors #roberta #question-answering #en #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n# roberta-base for QA \n\nThis is the roberta-base model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.## Overview\nLanguage model: roberta-base \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100## Hyperparameters## Using a distilled model instead\nPlease note that we have also released a distilled version of this model called deepset/tinyroberta-squad2. The distilled model has a comparable prediction quality and runs at twice the speed of the base model.## Usage### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''roberta-base-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation### In Transformers## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL"
] |
[
-0.0881306454539299,
0.2166670858860016,
-0.0051768566481769085,
0.023776832967996597,
0.05856075882911682,
0.032662104815244675,
0.15003395080566406,
0.1245400533080101,
-0.052524130791425705,
0.09770581126213074,
0.040455274283885956,
0.03483862057328224,
0.09431122988462448,
0.14750424027442932,
0.040085162967443466,
-0.22987599670886993,
-0.0243975892663002,
-0.1072678491473198,
-0.05282288044691086,
0.08081629872322083,
0.10296331346035004,
-0.09227483719587326,
0.09631042182445526,
-0.004243741277605295,
-0.02080022729933262,
0.03863593563437462,
-0.03644917160272598,
-0.05304419621825218,
0.054467521607875824,
0.046497032046318054,
0.04546167701482773,
0.009342086501419544,
0.07909036427736282,
-0.14611613750457764,
0.04474719241261482,
0.08948881924152374,
0.040492624044418335,
0.1118430569767952,
0.09816068410873413,
-0.00003403883602004498,
0.017180779948830605,
-0.10221211612224579,
0.050123944878578186,
0.07284658402204514,
-0.04457996413111687,
-0.2005660980939865,
-0.0955028310418129,
0.10165809839963913,
0.09070073813199997,
0.03848697617650032,
-0.008538437075912952,
0.07271638512611389,
0.0025126368273049593,
0.04235703498125076,
0.12478940933942795,
-0.25795528292655945,
-0.08763370662927628,
0.05722075328230858,
-0.010883446782827377,
0.0819486603140831,
-0.07991164177656174,
0.011412214487791061,
0.020227637141942978,
0.025204669684171677,
-0.010891837067902088,
-0.04957170784473419,
0.05405092239379883,
-0.024245338514447212,
-0.08855200558900833,
0.015234441496431828,
0.1415911763906479,
-0.002267773961648345,
-0.07664045691490173,
-0.1581864207983017,
-0.049533307552337646,
0.10738555341959,
-0.012594861909747124,
-0.08303065598011017,
0.03916523978114128,
0.001797249773517251,
-0.004452025983482599,
-0.11164756864309311,
-0.13372507691383362,
-0.023265907540917397,
-0.012258343398571014,
0.02002546563744545,
0.05436616763472557,
0.04396430402994156,
-0.051980189979076385,
0.08446313440799713,
-0.04518432170152664,
-0.12803159654140472,
-0.03806940093636513,
-0.10286760330200195,
-0.07972211390733719,
-0.01675920933485031,
-0.029605453833937645,
-0.08013609051704407,
0.11194239556789398,
0.18529322743415833,
-0.02619098126888275,
0.029028138145804405,
0.03565602004528046,
-0.017176302149891853,
0.07667257636785507,
0.17191575467586517,
-0.11316557228565216,
-0.14825598895549774,
0.05495377257466316,
-0.0015178673202171922,
0.03197474032640457,
-0.013086534105241299,
-0.04321314021945,
-0.03250131011009216,
-0.024450741708278656,
0.07671833783388138,
0.08854009956121445,
-0.03126751258969307,
0.0017354687443003058,
-0.08711469918489456,
0.11244519054889679,
-0.10884314775466919,
-0.019112830981612206,
0.02592005580663681,
-0.04488774761557579,
0.14152958989143372,
-0.03626948222517967,
0.036399226635694504,
-0.04906543344259262,
0.0070409574545919895,
-0.04117460548877716,
-0.05110438913106918,
-0.06843432784080505,
-0.06022222712635994,
0.011415827088057995,
-0.021398469805717468,
-0.045669928193092346,
-0.07041344046592712,
-0.15752249956130981,
-0.05385909229516983,
0.053792838007211685,
-0.04212260618805885,
-0.032651908695697784,
-0.026445472612977028,
-0.022632330656051636,
0.00275687244720757,
0.0009866297477856278,
0.08209328353404999,
0.001965711824595928,
0.029476216062903404,
-0.07322176545858383,
0.0331740640103817,
0.01019119843840599,
0.0031826673075556755,
-0.059838905930519104,
0.020932765677571297,
-0.09846008569002151,
0.08568713068962097,
-0.030078398063778877,
0.04876485466957092,
-0.15149235725402832,
-0.08789653331041336,
-0.04013600945472717,
-0.003170206444337964,
0.022742001339793205,
0.12951751053333282,
-0.19998762011528015,
-0.010985753498971462,
0.09861288219690323,
-0.05523259565234184,
-0.06902547925710678,
0.13359606266021729,
-0.014286890625953674,
0.09145698696374893,
0.07672277837991714,
0.159490704536438,
0.19412992894649506,
-0.08223742991685867,
-0.06396361440420151,
0.024689899757504463,
-0.04875766113400459,
0.05503878369927406,
0.10076102614402771,
-0.06965456902980804,
0.16465415060520172,
0.018050510436296463,
-0.004892813507467508,
-0.0520700179040432,
-0.03798503801226616,
-0.05564579367637634,
0.017903562635183334,
-0.0467454232275486,
-0.04597370699048042,
-0.058570344001054764,
0.0017868365393951535,
-0.03058786131441593,
-0.08363745361566544,
-0.04714270681142807,
0.11917009204626083,
0.007399686612188816,
-0.0010180756216868758,
-0.09246482700109482,
0.004570490680634975,
-0.009442342445254326,
0.0030997679568827152,
-0.1271877884864807,
-0.060553207993507385,
0.033244166523218155,
-0.05300483852624893,
0.07590510696172714,
0.040098048746585846,
0.03628389164805412,
0.03693845123052597,
0.009245745837688446,
-0.041688598692417145,
0.0075499266386032104,
-0.05249815434217453,
-0.025747239589691162,
-0.09835507720708847,
-0.049665145576000214,
-0.03393298387527466,
0.036844246089458466,
-0.031109889969229698,
0.028250286355614662,
-0.029510553926229477,
0.06634688377380371,
0.051106344908475876,
-0.04169605299830437,
0.05034760385751724,
-0.0422348715364933,
0.00561646418645978,
-0.06487877666950226,
-0.001894252491183579,
-0.0037548039108514786,
-0.031186312437057495,
-0.024597475305199623,
-0.11705022305250168,
-0.016797736287117004,
0.0275631882250309,
0.07329963147640228,
-0.022761819884181023,
0.018884237855672836,
-0.04020541161298752,
-0.010667530819773674,
-0.02331763505935669,
0.021436365321278572,
0.12697163224220276,
0.047315288335084915,
0.03938895836472511,
-0.0950082466006279,
-0.03353528678417206,
-0.021559666842222214,
-0.042380087077617645,
-0.0030345479026436806,
0.06191246211528778,
-0.04986633360385895,
-0.13946592807769775,
-0.0013240310363471508,
0.12031978368759155,
0.024877164512872696,
0.1468041092157364,
0.007773525081574917,
-0.08980299532413483,
-0.07673628628253937,
0.0023570507764816284,
0.02408730611205101,
0.04844902828335762,
0.056369997560977936,
0.05205763876438141,
0.05983540043234825,
0.04707973822951317,
0.023574182763695717,
-0.04526517540216446,
0.03559492528438568,
-0.024277949705719948,
-0.030923688784241676,
0.06293819099664688,
0.08935565501451492,
0.028816521167755127,
0.06565918773412704,
0.046041879802942276,
0.1052073985338211,
-0.05143572762608528,
-0.07949645072221756,
-0.06501840800046921,
0.15575365722179413,
-0.11265168339014053,
-0.20441246032714844,
-0.1536770612001419,
-0.012447609566152096,
-0.056830037385225296,
-0.025179214775562286,
0.047135915607213974,
-0.06388222426176071,
-0.060547973960638046,
-0.051658134907484055,
0.044566359370946884,
-0.007195800542831421,
-0.03454523906111717,
-0.02225523255765438,
0.05159532278776169,
0.05321481451392174,
-0.14547397196292877,
0.022585256025195122,
-0.00027931053773500025,
-0.13643667101860046,
-0.02031131088733673,
0.05810609087347984,
0.024100787937641144,
0.03976164385676384,
0.05899934470653534,
0.01712002232670784,
-0.007413468323647976,
0.20852252840995789,
-0.041235946118831635,
0.08626680821180344,
0.249362051486969,
0.041945990175008774,
0.06413460522890091,
0.11037806421518326,
0.02559600956737995,
-0.04539372771978378,
0.040658485144376755,
0.051018089056015015,
-0.01917288266122341,
-0.18858489394187927,
-0.024269096553325653,
-0.008723623119294643,
0.029246198013424873,
-0.008710064925253391,
0.036171987652778625,
-0.06731236726045609,
0.023202864453196526,
-0.0880030170083046,
-0.05616327002644539,
0.009360534138977528,
0.07031501829624176,
0.050093743950128555,
-0.009807286784052849,
0.060448672622442245,
-0.08066908270120621,
0.03014366328716278,
0.15674744546413422,
0.03963819891214371,
0.12868672609329224,
-0.039685774594545364,
0.14049652218818665,
0.05128597095608711,
0.13289403915405273,
0.013731644488871098,
0.023939693346619606,
0.008901852183043957,
0.02508728764951229,
-0.019803324714303017,
-0.07906641811132431,
-0.003872827859595418,
0.06255684792995453,
0.02486477605998516,
0.044095657765865326,
-0.002760352101176977,
0.09572628885507584,
0.0574566014111042,
0.1971648782491684,
0.056330010294914246,
-0.13679172098636627,
-0.07726725935935974,
0.044382244348526,
-0.041866887360811234,
-0.0833221897482872,
0.02270343154668808,
0.08273950964212418,
-0.1464422345161438,
0.0027741803787648678,
-0.016523923724889755,
0.0777205154299736,
-0.025080548599362373,
0.020286647602915764,
0.03911438211798668,
0.1632518768310547,
-0.018288657069206238,
0.07550723850727081,
-0.15738709270954132,
0.018748734146356583,
0.03669499233365059,
0.09477095305919647,
-0.0034233652986586094,
0.027688439935445786,
0.044497039169073105,
-0.020568594336509705,
0.11386606842279434,
-0.01882264018058777,
0.025690313428640366,
0.012126315385103226,
-0.16234487295150757,
0.045650776475667953,
0.10527767986059189,
-0.11582885682582855,
0.09288706630468369,
-0.03851896524429321,
-0.0008115966920740902,
-0.050760891288518906,
0.10104800760746002,
-0.09965822845697403,
-0.1204681247472763,
0.025905078276991844,
-0.030465109273791313,
-0.017117401584982872,
-0.08597011119127274,
-0.018314791843295097,
0.005301274824887514,
0.1384008675813675,
-0.14079615473747253,
-0.1022379994392395,
-0.09308814257383347,
0.006133364513516426,
0.13536305725574493,
-0.09669693559408188,
0.004053502809256315,
-0.03537910431623459,
0.11192556470632553,
0.026601923629641533,
-0.12980017066001892,
-0.0013986601261422038,
-0.07543661445379257,
-0.14984868466854095,
-0.013286903500556946,
0.12840412557125092,
0.012630018405616283,
0.03114822320640087,
-0.00021068954083602875,
0.011846537701785564,
-0.09498265385627747,
-0.09465266019105911,
-0.01739833503961563,
0.20261558890342712,
0.046446043998003006,
0.044984120875597,
-0.032164618372917175,
-0.11035492271184921,
-0.05233271047472954,
0.05436772108078003,
0.07171227782964706,
0.1941993087530136,
-0.07182376086711884,
0.12520970404148102,
0.112608402967453,
-0.0509733185172081,
-0.17893393337726593,
-0.07013668119907379,
0.07058770954608917,
0.031052038073539734,
0.04129653796553612,
-0.17346616089344025,
0.11556945741176605,
0.0566771924495697,
-0.022760603576898575,
0.04109956696629524,
-0.2780172824859619,
-0.10998018085956573,
0.029526542872190475,
-0.015654167160391808,
0.033536333590745926,
-0.11825868487358093,
-0.047060441225767136,
-0.040112078189849854,
-0.026054151356220245,
0.07626698911190033,
-0.09023340046405792,
0.07800470292568207,
-0.011830531060695648,
0.08745066076517105,
0.02824333682656288,
-0.022870024666190147,
0.09459328651428223,
-0.04573706537485123,
0.05525471270084381,
-0.02862047217786312,
0.050672661513090134,
0.07159741967916489,
-0.07454044371843338,
0.08642086386680603,
-0.007970522157847881,
0.10733607411384583,
-0.09347662329673767,
-0.023713938891887665,
-0.05104837566614151,
0.05993891507387161,
-0.0792229101061821,
-0.07647813111543655,
-0.0508979894220829,
0.07740458846092224,
0.021274277940392494,
0.005346412304788828,
0.016250113025307655,
-0.024098986759781837,
0.038097985088825226,
0.12092455476522446,
0.11569222807884216,
0.04467973858118057,
-0.06403199583292007,
-0.05973927676677704,
-0.018378354609012604,
0.0379842109978199,
-0.13064521551132202,
0.01213847752660513,
0.06844998896121979,
0.0026751263067126274,
0.022174110636115074,
-0.015231822617352009,
-0.12514285743236542,
0.037303727120161057,
0.059598907828330994,
-0.1296808123588562,
-0.19217412173748016,
-0.02628965862095356,
0.0697239339351654,
-0.16591040790081024,
0.04311486333608627,
0.16512863337993622,
0.014187741093337536,
-0.05636930093169212,
0.011215039528906345,
0.05053026229143143,
-0.018236320465803146,
0.08342047035694122,
0.04904976487159729,
0.018868274986743927,
-0.06365188211202621,
0.1123766303062439,
0.04973484203219414,
0.04627464339137077,
0.002643841551616788,
0.15254557132720947,
-0.07450460642576218,
-0.04856538027524948,
-0.03638819232583046,
0.06404384970664978,
-0.03757345303893089,
-0.012727010063827038,
-0.013657154515385628,
0.0024992097169160843,
0.010791861452162266,
-0.017301952466368675,
0.035933367908000946,
0.027681536972522736,
-0.02372894063591957,
-0.015260551124811172,
-0.013393388129770756,
0.11376653611660004,
0.015727641060948372,
-0.02295532450079918,
-0.09822771698236465,
0.06763361394405365,
-0.0462607778608799,
0.0100192716345191,
-0.009409693069756031,
-0.05611970275640488,
-0.03749296814203262,
-0.03027179278433323,
-0.1865711361169815,
0.016707880422472954,
-0.05139615386724472,
0.0010569975711405277,
-0.008911988697946072,
-0.02314087189733982,
-0.037957850843667984,
0.016193686053156853,
-0.010454690083861351,
-0.05508086457848549,
-0.05467936396598816,
0.10264171659946442,
-0.16258826851844788,
0.030019642785191536,
-0.00802575796842575,
-0.05147646740078926,
0.12982715666294098,
0.04133940488100052,
-0.03686905279755592,
0.03978284075856209,
-0.021358322352170944,
-0.04006659612059593,
-0.07117237895727158,
0.051852405071258545,
0.04848002642393112,
-0.061771273612976074,
-0.02797793224453926,
-0.025393696501851082,
-0.05479662865400314,
-0.01850365474820137,
0.04065157100558281,
-0.07057932019233704,
0.08894098550081253,
-0.005025088787078857,
0.009469082579016685,
-0.09867703914642334,
-0.01213377807289362,
0.04716124013066292,
0.10967259854078293,
0.11039791256189346,
-0.05070216581225395,
0.07704748213291168,
-0.17667867243289948,
-0.021224195137619972,
0.01658373884856701,
0.019512753933668137,
-0.10557374358177185,
-0.02409018762409687,
0.05294042453169823,
-0.02573253959417343,
0.03560348227620125,
-0.05238513648509979,
0.03607192635536194,
0.01795794442296028,
-0.04880432412028313,
0.001269897329621017,
0.02777746692299843,
0.10618702322244644,
0.016393622383475304,
0.007284460123628378,
0.001908273552544415,
0.000711487780790776,
-0.027829648926854134,
-0.019429368898272514,
0.12371556460857391,
0.16551387310028076,
0.1220349371433258,
0.03695720434188843,
0.029837591573596,
-0.032375823706388474,
-0.05576379969716072,
0.08842518925666809,
0.018044019117951393,
0.0030040487181395292,
-0.04863099008798599,
-0.043826621025800705,
0.2024962455034256,
-0.1670321673154831,
0.026457201689481735,
-0.01942296139895916,
-0.030063660815358162,
-0.11918516457080841,
-0.1516137719154358,
-0.07889916747808456,
-0.013756069354712963,
-0.009549598209559917,
-0.1436939686536789,
0.052127331495285034,
0.038711901754140854,
0.04828514903783798,
-0.02690041810274124,
0.09027472138404846,
-0.0910385325551033,
-0.06482501327991486,
0.06549599021673203,
0.027306517586112022,
0.04229876399040222,
0.07758697122335434,
0.02846936695277691,
0.014540682546794415,
0.0505436472594738,
0.06590886414051056,
0.040439456701278687,
0.0031235686037689447,
-0.008047305978834629,
-0.11137796193361282,
-0.09291122853755951,
0.0005220068269409239,
-0.0069322469644248486,
-0.019583426415920258,
0.1012745052576065,
0.0615922212600708,
-0.06836205720901489,
0.002647260669618845,
0.24496996402740479,
-0.07196435332298279,
-0.07183074951171875,
-0.16130158305168152,
0.12054827809333801,
-0.04554131627082825,
0.005843677558004856,
0.013018455356359482,
-0.0805240124464035,
-0.051711905747652054,
0.08648499101400375,
0.22672857344150543,
-0.08124469965696335,
-0.01580914296209812,
0.03424067795276642,
0.0020736430305987597,
-0.09481550753116608,
0.08884426951408386,
0.06592028588056564,
0.2197079360485077,
-0.033822398632764816,
0.08638513833284378,
0.025600960478186607,
0.00039725066744722426,
-0.13535237312316895,
0.06468534469604492,
-0.051635317504405975,
-0.0013924030354246497,
0.0013716029934585094,
0.1600976586341858,
-0.049727026373147964,
-0.21001313626766205,
-0.02016894333064556,
-0.05783254653215408,
-0.13176937401294708,
-0.02033725194633007,
0.037956684827804565,
0.012862027622759342,
0.08032923936843872,
-0.013332026079297066,
0.016209740191698074,
0.14354977011680603,
-0.024648794904351234,
0.010252990759909153,
-0.03739853575825691,
0.10790172964334488,
-0.13143298029899597,
0.13899429142475128,
0.03450929746031761,
-0.0034908954985439777,
0.08532827347517014,
-0.037304073572158813,
-0.13349539041519165,
-0.001189383096061647,
0.03432217240333557,
-0.07562210410833359,
0.0006365960580296814,
0.08177389949560165,
0.0033077599946409464,
0.07999828457832336,
0.0795101448893547,
-0.0628112405538559,
-0.004420484881848097,
0.01768561266362667,
0.0009017694974318147,
-0.12505927681922913,
0.0921708196401596,
-0.09600578993558884,
0.13318578898906708,
0.1030404195189476,
-0.06474318355321884,
0.002183701144531369,
-0.016417985782027245,
0.05465617775917053,
0.05563671886920929,
0.04540015384554863,
-0.01862807758152485,
-0.11379077285528183,
0.023622212931513786,
-0.09847495704889297,
0.035305801779031754,
-0.17208077013492584,
-0.03313892334699631,
-0.01835723966360092,
0.00972050242125988,
-0.026979921385645866,
0.10693470388650894,
0.033666327595710754,
0.0057180835865437984,
-0.001709983334876597,
0.014109078794717789,
-0.031271446496248245,
0.10311327129602432,
-0.1357547640800476,
-0.0897226333618164
] |
null | null |
transformers
|
# roberta-large for QA
This is the [roberta-large](https://huggingface.co/roberta-large) model, fine-tuned using the [SQuAD2.0](https://huggingface.co/datasets/squad_v2) dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
## Overview
**Language model:** roberta-large
**Language:** English
**Downstream-task:** Extractive QA
**Training data:** SQuAD 2.0
**Eval data:** SQuAD 2.0
**Code:** See [an example QA pipeline on Haystack](https://haystack.deepset.ai/tutorials/first-qa-system)
**Infrastructure**: 4x Tesla v100
## Hyperparameters
```
base_LM_model = "roberta-large"
```
## Using a distilled model instead
Please note that we have also released a distilled version of this model called [deepset/roberta-base-squad2-distilled](https://huggingface.co/deepset/roberta-base-squad2-distilled). The distilled model has a comparable prediction quality and runs at twice the speed of the large model.
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/roberta-large-squad2")
# or
reader = TransformersReader(model_name_or_path="deepset/roberta-large-squad2",tokenizer="deepset/roberta-large-squad2")
```
For a complete example of ``roberta-large-squad2`` being used for Question Answering, check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
### In Transformers
```python
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
model_name = "deepset/roberta-large-squad2"
# a) Get predictions
nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
QA_input = {
'question': 'Why is model conversion important?',
'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
}
res = nlp(QA_input)
# b) Load model & tokenizer
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
## Authors
**Branden Chan:** [email protected]
**Timo Möller:** [email protected]
**Malte Pietsch:** [email protected]
**Tanay Soni:** [email protected]
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/deepset-logo-colored.png" class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/haystack-logo-colored.png" class="w-40"/>
</div>
</div>
[deepset](http://deepset.ai/) is the company behind the open-source NLP framework [Haystack](https://haystack.deepset.ai/) which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- [Distilled roberta-base-squad2 (aka "tinyroberta-squad2")]([https://huggingface.co/deepset/tinyroberta-squad2)
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="https://github.com/deepset-ai/haystack">GitHub</a></strong> repo and <strong><a href="https://docs.haystack.deepset.ai">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="https://haystack.deepset.ai/community">Discord community open to everyone!</a></strong></p>
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "en", "license": "cc-by-4.0", "datasets": ["squad_v2"], "base_model": "roberta-large", "model-index": [{"name": "deepset/roberta-large-squad2", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 85.168, "name": "Exact Match"}, {"type": "f1", "value": 88.349, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad", "type": "squad", "config": "plain_text", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 87.162, "name": "Exact Match"}, {"type": "f1", "value": 93.603, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "adversarial_qa", "type": "adversarial_qa", "config": "adversarialQA", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 35.9, "name": "Exact Match"}, {"type": "f1", "value": 48.923, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_adversarial", "type": "squad_adversarial", "config": "AddOneSent", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 81.142, "name": "Exact Match"}, {"type": "f1", "value": 87.099, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts amazon", "type": "squadshifts", "config": "amazon", "split": "test"}, "metrics": [{"type": "exact_match", "value": 72.453, "name": "Exact Match"}, {"type": "f1", "value": 86.325, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts new_wiki", "type": "squadshifts", "config": "new_wiki", "split": "test"}, "metrics": [{"type": "exact_match", "value": 82.338, "name": "Exact Match"}, {"type": "f1", "value": 91.974, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts nyt", "type": "squadshifts", "config": "nyt", "split": "test"}, "metrics": [{"type": "exact_match", "value": 84.352, "name": "Exact Match"}, {"type": "f1", "value": 92.645, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts reddit", "type": "squadshifts", "config": "reddit", "split": "test"}, "metrics": [{"type": "exact_match", "value": 74.722, "name": "Exact Match"}, {"type": "f1", "value": 86.86, "name": "F1"}]}]}]}
|
question-answering
|
deepset/roberta-large-squad2
|
[
"transformers",
"pytorch",
"jax",
"safetensors",
"roberta",
"question-answering",
"en",
"dataset:squad_v2",
"base_model:roberta-large",
"license:cc-by-4.0",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #jax #safetensors #roberta #question-answering #en #dataset-squad_v2 #base_model-roberta-large #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us
|
# roberta-large for QA
This is the roberta-large model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.
## Overview
Language model: roberta-large
Language: English
Downstream-task: Extractive QA
Training data: SQuAD 2.0
Eval data: SQuAD 2.0
Code: See an example QA pipeline on Haystack
Infrastructure: 4x Tesla v100
## Hyperparameters
## Using a distilled model instead
Please note that we have also released a distilled version of this model called deepset/roberta-base-squad2-distilled. The distilled model has a comparable prediction quality and runs at twice the speed of the large model.
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:
For a complete example of ''roberta-large-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation
### In Transformers
## Authors
Branden Chan: URL@URL
Timo Möller: timo.moeller@URL
Malte Pietsch: malte.pietsch@URL
Tanay Soni: URL@URL
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
</div>
deepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- Distilled roberta-base-squad2 (aka "tinyroberta-squad2")
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="URL repo and <strong><a href="URL">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="URL community open to everyone!</a></strong></p>
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# roberta-large for QA \n\nThis is the roberta-large model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.",
"## Overview\nLanguage model: roberta-large \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Using a distilled model instead\nPlease note that we have also released a distilled version of this model called deepset/roberta-base-squad2-distilled. The distilled model has a comparable prediction quality and runs at twice the speed of the large model.",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''roberta-large-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation",
"### In Transformers",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL",
"## About us\n\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #jax #safetensors #roberta #question-answering #en #dataset-squad_v2 #base_model-roberta-large #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n",
"# roberta-large for QA \n\nThis is the roberta-large model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.",
"## Overview\nLanguage model: roberta-large \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Using a distilled model instead\nPlease note that we have also released a distilled version of this model called deepset/roberta-base-squad2-distilled. The distilled model has a comparable prediction quality and runs at twice the speed of the large model.",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''roberta-large-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation",
"### In Transformers",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL",
"## About us\n\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
76,
61,
61,
5,
62,
3,
85,
6,
41,
251,
113
] |
[
"passage: TAGS\n#transformers #pytorch #jax #safetensors #roberta #question-answering #en #dataset-squad_v2 #base_model-roberta-large #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n# roberta-large for QA \n\nThis is the roberta-large model, fine-tuned using the SQuAD2.0 dataset. It's been trained on question-answer pairs, including unanswerable questions, for the task of Question Answering.## Overview\nLanguage model: roberta-large \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100## Hyperparameters## Using a distilled model instead\nPlease note that we have also released a distilled version of this model called deepset/roberta-base-squad2-distilled. The distilled model has a comparable prediction quality and runs at twice the speed of the large model.## Usage### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''roberta-large-squad2'' being used for Question Answering, check out the Tutorials in Haystack Documentation### In Transformers## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL"
] |
[
-0.06462708115577698,
0.22740218043327332,
-0.005921832285821438,
-0.006080613471567631,
0.07556451857089996,
0.039614237844944,
0.12345430999994278,
0.09965874999761581,
0.018646178767085075,
0.05940480902791023,
0.03039482794702053,
0.06512768566608429,
0.07670769095420837,
0.08014621585607529,
0.04613903537392616,
-0.214529350399971,
0.0020586729515343904,
-0.10534089058637619,
-0.000758782378397882,
0.09427841752767563,
0.12261505424976349,
-0.0807710662484169,
0.09849376231431961,
0.015927644446492195,
-0.03925646096467972,
0.04317282512784004,
-0.04160177707672119,
-0.0320553332567215,
0.08426510542631149,
0.04867850989103317,
0.09248054027557373,
-0.009661260060966015,
0.05105980858206749,
-0.1756415069103241,
0.034297291189432144,
0.07430422306060791,
0.03235473856329918,
0.09019117802381516,
0.07541197538375854,
-0.028346717357635498,
0.09133630245923996,
-0.09213180840015411,
0.04729868471622467,
0.0703873485326767,
-0.056603480130434036,
-0.16657303273677826,
-0.10661126673221588,
0.1102556437253952,
0.08909014612436295,
0.032727573066949844,
-0.01725207082927227,
0.05992361903190613,
0.018896792083978653,
0.0385860912501812,
0.11712049692869186,
-0.20996448397636414,
-0.06815630197525024,
0.07438932359218597,
0.019306674599647522,
0.06811793148517609,
-0.08600031584501266,
0.04087188467383385,
0.038131386041641235,
0.027921946719288826,
0.011691060848534107,
-0.05004999041557312,
0.08043278753757477,
-0.01153532788157463,
-0.12363284826278687,
-0.015607710927724838,
0.14943288266658783,
-0.033488161861896515,
-0.0753050372004509,
-0.16469384729862213,
-0.019934045150876045,
0.1681210696697235,
0.013097795657813549,
-0.06688851118087769,
0.03882761672139168,
-0.004137821029871702,
0.06464574486017227,
-0.13782526552677155,
-0.13181351125240326,
-0.02007146365940571,
0.0026888311840593815,
0.04083407670259476,
0.04512795805931091,
0.03361751511693001,
-0.0368947759270668,
0.06676176935434341,
-0.08201708644628525,
-0.10939980298280716,
-0.043051887303590775,
-0.12217098474502563,
-0.04442354664206505,
-0.027920613065361977,
-0.05810242146253586,
-0.07199490070343018,
0.0844293087720871,
0.135186105966568,
-0.08055147528648376,
0.011960801668465137,
0.032777268439531326,
-0.009549827314913273,
0.10736538469791412,
0.14545126259326935,
-0.10475404560565948,
-0.10350058227777481,
-0.004088432993739843,
-0.0030406422447413206,
0.00927590299397707,
-0.000007666958481422625,
-0.046680159866809845,
-0.03042064607143402,
-0.007934081368148327,
0.0804973915219307,
0.06461604684591293,
-0.02158900909125805,
-0.004680612124502659,
-0.08420149236917496,
0.11749812960624695,
-0.12507691979408264,
-0.014991688542068005,
0.008573654107749462,
-0.013739123940467834,
0.15257814526557922,
-0.028882211074233055,
0.01567370444536209,
-0.037811920046806335,
0.019468937069177628,
-0.07624179869890213,
-0.006174389738589525,
-0.09555737674236298,
-0.06731855124235153,
0.010642529465258121,
-0.05656111612915993,
-0.03598425164818764,
-0.07989531755447388,
-0.13724875450134277,
-0.04588112607598305,
0.03503319248557091,
-0.02785615436732769,
-0.013710517436265945,
-0.04222555831074715,
-0.014889735728502274,
0.011341804638504982,
0.0034064273349940777,
0.02718719281256199,
0.00960481259971857,
0.02405357174575329,
-0.06492386013269424,
0.03771466389298439,
-0.02682427316904068,
0.015824660658836365,
-0.05631967633962631,
0.03490658476948738,
-0.0971551313996315,
0.09855625778436661,
-0.04977530241012573,
0.07938999682664871,
-0.13272608816623688,
-0.06268325448036194,
-0.006981726735830307,
0.014132019132375717,
-0.004194442182779312,
0.1211332455277443,
-0.23596373200416565,
-0.021489379927515984,
0.1601288765668869,
-0.10551539808511734,
-0.05964057892560959,
0.1274184286594391,
-0.02561337687075138,
0.09017396718263626,
0.12140674144029617,
0.1733928620815277,
0.13929222524166107,
-0.11971735954284668,
-0.07573291659355164,
-0.003004751168191433,
-0.022263171151280403,
0.08838912844657898,
0.08589916676282883,
-0.08068521320819855,
0.11924175918102264,
0.007286466658115387,
-0.049821637570858,
-0.0565236359834671,
-0.031189391389489174,
-0.0685773715376854,
0.021261142566800117,
-0.06090153008699417,
0.008957183919847012,
-0.03708761930465698,
-0.03917769715189934,
-0.026428621262311935,
-0.09212205559015274,
0.04272295534610748,
0.12615841627120972,
-0.011465214192867279,
0.001409785938449204,
-0.08378095179796219,
0.019730040803551674,
0.025639664381742477,
0.011106641963124275,
-0.12895487248897552,
-0.09227290749549866,
-0.005818330217152834,
-0.09052573144435883,
0.06591224670410156,
0.038662925362586975,
0.017075544223189354,
0.03752690181136131,
-0.00085402128752321,
-0.0005588535568676889,
-0.008535831235349178,
-0.032516948878765106,
-0.021889813244342804,
-0.08734254539012909,
-0.056428395211696625,
-0.03322182223200798,
0.06672301888465881,
-0.07527722418308258,
0.037247754633426666,
0.010814242996275425,
0.06175687164068222,
0.02885603904724121,
-0.04728833585977554,
0.027482934296131134,
-0.008203105069696903,
-0.012458784505724907,
-0.04088563844561577,
0.008702084422111511,
-0.02874923124909401,
-0.02524752728641033,
0.02268947660923004,
-0.09466549009084702,
-0.0886303037405014,
0.05237283185124397,
0.03413645923137665,
-0.055222757160663605,
0.028771689161658287,
-0.013464443385601044,
0.004479129798710346,
-0.017122430726885796,
-0.02522536739706993,
0.11427506059408188,
0.03622367978096008,
0.03180469572544098,
-0.08272138237953186,
-0.04446663334965706,
-0.02815718576312065,
-0.05423823744058609,
-0.04875786602497101,
0.0412314310669899,
-0.06216234713792801,
-0.148702010512352,
0.01852419786155224,
0.149429053068161,
0.00990305282175541,
0.19071064889431,
0.03180878609418869,
-0.09235579520463943,
-0.07573066651821136,
-0.029376059770584106,
-0.0014702263288199902,
0.055642303079366684,
-0.022516483440995216,
0.018072865903377533,
0.04358377307653427,
0.031752362847328186,
0.041005689650774,
-0.04066967964172363,
0.024768022820353508,
-0.006537094712257385,
-0.027654793113470078,
0.029747147113084793,
0.05270639806985855,
0.0019416140858083963,
0.04162851721048355,
0.02483433112502098,
0.10096274316310883,
-0.032260097563266754,
-0.06527165323495865,
-0.05811144784092903,
0.13612139225006104,
-0.13177552819252014,
-0.24698662757873535,
-0.17339199781417847,
-0.04447163641452789,
-0.047926221042871475,
-0.012604404240846634,
0.04982231184840202,
-0.049963656812906265,
-0.07182934135198593,
-0.025364676490426064,
0.0597386434674263,
-0.03586370870471001,
-0.07405637204647064,
-0.015915963798761368,
0.01616715081036091,
0.044070158153772354,
-0.1340065896511078,
0.012901809066534042,
0.005479862447828054,
-0.13936671614646912,
-0.016497766599059105,
0.04621847718954086,
0.05736466124653816,
0.06103247404098511,
0.032962292432785034,
0.000268873613094911,
0.0022794129326939583,
0.22819243371486664,
-0.0643065944314003,
0.07044686377048492,
0.1880132555961609,
0.011969830840826035,
0.07642127573490143,
0.09229395538568497,
0.03022892214357853,
-0.05116608738899231,
0.010750172659754753,
0.06295813620090485,
-0.02567240037024021,
-0.20837867259979248,
-0.047186192125082016,
-0.007342399563640356,
-0.04513833299279213,
-0.011841006577014923,
0.025057509541511536,
0.008739510551095009,
0.042704157531261444,
-0.07913533598184586,
-0.07925033569335938,
-0.004185026977211237,
0.07545848935842514,
0.11156432330608368,
-0.02489255927503109,
0.06795208156108856,
-0.06189902499318123,
0.037026964128017426,
0.12567538022994995,
0.015123331919312477,
0.1363397240638733,
0.02954530343413353,
0.1353864222764969,
0.08410665392875671,
0.11782252043485641,
0.014870422892272472,
0.012273667380213737,
-0.0157832819968462,
0.029829734936356544,
-0.02221359871327877,
-0.09439780563116074,
-0.01049548014998436,
0.07605569809675217,
0.027548115700483322,
0.01974385417997837,
-0.007457038387656212,
0.022756921127438545,
0.06226500868797302,
0.160732239484787,
0.07389530539512634,
-0.14510026574134827,
-0.07145559787750244,
0.05587000027298927,
-0.03250650316476822,
-0.06310243904590607,
0.018928691744804382,
0.07989437878131866,
-0.12270189076662064,
0.04792241007089615,
0.01359791960567236,
0.08964323252439499,
-0.04380463808774948,
0.023234659805893898,
0.029649723321199417,
0.13345633447170258,
-0.018826721236109734,
0.08411615341901779,
-0.14684167504310608,
0.050259340554475784,
0.03179701417684555,
0.07199740409851074,
-0.005819382146000862,
0.03190072998404503,
0.05693806707859039,
0.028561808168888092,
0.12307468056678772,
0.003381168469786644,
0.0146963344886899,
0.016629623249173164,
-0.11255116760730743,
0.010005093179643154,
0.08389189094305038,
-0.14564833045005798,
0.10034957528114319,
-0.03701592609286308,
0.003134878119453788,
-0.05155661702156067,
0.11696598678827286,
-0.12088824063539505,
-0.14279210567474365,
0.028530877083539963,
-0.011526722460985184,
0.03752324730157852,
-0.0671396479010582,
-0.027776680886745453,
-0.028918446972966194,
0.12246424704790115,
-0.1553085893392563,
-0.13551543653011322,
-0.0720929354429245,
-0.048890236765146255,
0.12493103742599487,
-0.0794486403465271,
-0.0006498323637060821,
-0.015484774485230446,
0.1230255588889122,
-0.0161786749958992,
-0.10025940835475922,
-0.02516746334731579,
-0.0712244063615799,
-0.09421303123235703,
0.014262327924370766,
0.169007807970047,
0.05786425620317459,
0.015061874873936176,
0.02643503248691559,
0.03943364694714546,
-0.0662316381931305,
-0.1078249141573906,
-0.009867827408015728,
0.22535322606563568,
0.04628657177090645,
0.07939372956752777,
-0.0334417000412941,
-0.16747047007083893,
-0.10427828133106232,
0.04744075983762741,
0.07039468735456467,
0.19086644053459167,
-0.06181186065077782,
0.14635439217090607,
0.1690066158771515,
-0.10175771266222,
-0.16996707022190094,
-0.09605433791875839,
0.03723747283220291,
0.009253192692995071,
0.07158014178276062,
-0.20559708774089813,
0.059970322996377945,
0.07444943487644196,
-0.011911073699593544,
0.07815508544445038,
-0.27568864822387695,
-0.09565258026123047,
0.04583261162042618,
-0.010224896483123302,
-0.013903537765145302,
-0.1299489140510559,
-0.06906893849372864,
-0.03395705670118332,
-0.05361580476164818,
0.09575973451137543,
-0.03993028402328491,
0.06252317875623703,
-0.002774996915832162,
0.1641598641872406,
0.04658844694495201,
-0.010507065802812576,
0.13550449907779694,
-0.014675084501504898,
0.04337134212255478,
-0.05081510543823242,
0.029557690024375916,
0.13243620097637177,
-0.06531845033168793,
0.0771108940243721,
-0.022514550015330315,
0.06773140281438828,
-0.06750695407390594,
-0.04750081151723862,
-0.05841689556837082,
0.056584227830171585,
-0.09104141592979431,
-0.07153936475515366,
-0.03267674520611763,
0.08693716675043106,
0.016672564670443535,
0.009064463898539543,
0.004194820299744606,
-0.0650792345404625,
0.06899749487638474,
0.12493982911109924,
0.12840689718723297,
0.005340018309652805,
-0.07974667847156525,
-0.06399579346179962,
-0.02099430561065674,
0.059044525027275085,
-0.10654813051223755,
0.0212823785841465,
0.06245046854019165,
0.02695847488939762,
0.06647814065217972,
-0.008780652657151222,
-0.14305360615253448,
0.027076851576566696,
0.06603702157735825,
-0.1499859243631363,
-0.19179359078407288,
-0.03681972622871399,
0.035112347453832626,
-0.1352682113647461,
0.013693333603441715,
0.16874678432941437,
0.009796636179089546,
-0.062032267451286316,
-0.0022877466399222612,
0.06444115936756134,
-0.028992464765906334,
0.07100162655115128,
0.04108741506934166,
0.024262072518467903,
-0.06382051110267639,
0.05186014994978905,
0.0580771379172802,
0.03205357491970062,
0.012054589577019215,
0.12086248397827148,
-0.05294591188430786,
-0.06040528044104576,
-0.1391099989414215,
0.04226570203900337,
-0.08147535473108292,
-0.03923708200454712,
0.04885455220937729,
-0.0013618890661746264,
-0.026605911552906036,
0.0030646962113678455,
0.059221360832452774,
0.0636318102478981,
-0.005674430634826422,
-0.018415099009871483,
-0.033533111214637756,
0.09948062896728516,
-0.021246207877993584,
-0.039868563413619995,
-0.10574052482843399,
0.0689932256937027,
-0.04052947461605072,
0.0070129698142409325,
-0.006279084365814924,
-0.06554887443780899,
-0.04224678874015808,
-0.030737893655896187,
-0.22367997467517853,
0.04223402589559555,
-0.09130503237247467,
0.009134943597018719,
0.00425537396222353,
-0.006958724465221167,
-0.024927645921707153,
0.006729322951287031,
-0.021680470556020737,
-0.04186294600367546,
-0.025480518117547035,
0.11463316529989243,
-0.18025800585746765,
0.025550300255417824,
0.025382502004504204,
-0.04425559937953949,
0.09933357685804367,
0.030301377177238464,
-0.02453850395977497,
0.07020504027605057,
0.018347760662436485,
-0.024575868621468544,
-0.0615844763815403,
0.03454718738794327,
0.03140925243496895,
-0.09048651903867722,
-0.05170968174934387,
-0.017701327800750732,
-0.02872675098478794,
-0.0024749524891376495,
0.01761086843907833,
-0.07043740153312683,
0.07836076617240906,
0.07741250097751617,
-0.01870867982506752,
-0.061709608882665634,
0.002501139184460044,
0.07104870676994324,
0.08625000715255737,
0.11941487342119217,
-0.04572248458862305,
0.06629607826471329,
-0.11887065321207047,
-0.0003549974935594946,
-0.005167996045202017,
0.04442187771201134,
-0.04428064823150635,
-0.04007837548851967,
0.03208230063319206,
-0.0389595553278923,
0.0848185196518898,
-0.02203616127371788,
0.05422481521964073,
0.024901853874325752,
-0.03694501891732216,
-0.0294087752699852,
0.0142138646915555,
0.126972034573555,
0.015273972414433956,
-0.00010200624819844961,
-0.00842618104070425,
-0.00026133516803383827,
-0.05103384330868721,
-0.02964950166642666,
0.14614731073379517,
0.1643524169921875,
0.1081867665052414,
0.015671219676733017,
0.03717733547091484,
0.009135082364082336,
-0.0811266303062439,
0.058778177946805954,
0.03632013499736786,
-0.001572165172547102,
-0.056290075182914734,
-0.00527870561927557,
0.18088488280773163,
-0.18751034140586853,
0.07696031033992767,
-0.023004889488220215,
-0.04192570969462395,
-0.10944114625453949,
-0.16318216919898987,
-0.047261498868465424,
-0.0015552911208942533,
-0.023689240217208862,
-0.14561907947063446,
0.07234019786119461,
0.0725620687007904,
0.0181608647108078,
-0.04650074616074562,
0.08966054767370224,
-0.07724155485630035,
-0.07970207184553146,
0.07055416703224182,
0.01819559372961521,
0.025701597332954407,
0.010131369344890118,
0.04563665762543678,
0.009316308423876762,
0.0643971785902977,
0.03568240627646446,
0.03536111116409302,
0.028102491050958633,
-0.014490949921309948,
-0.11658419668674469,
-0.0788443312048912,
-0.012880376540124416,
0.028782537207007408,
0.007888644002377987,
0.03831678628921509,
0.05734841898083687,
-0.08318445831537247,
-0.0034259939566254616,
0.2188744693994522,
-0.04995120316743851,
-0.09999754279851913,
-0.13540005683898926,
0.14743123948574066,
-0.041630953550338745,
0.021919874474406242,
0.017980000004172325,
-0.07279171794652939,
-0.007129051722586155,
0.08896687626838684,
0.22443759441375732,
-0.058960214257240295,
-0.011800172738730907,
-0.025683948770165443,
0.015018844977021217,
-0.05830780416727066,
0.06206176057457924,
0.03651230037212372,
0.2539180815219879,
-0.026662319898605347,
0.10311394929885864,
0.0023881334345787764,
0.012870640493929386,
-0.11954919248819351,
0.028265060856938362,
-0.06657235324382782,
-0.017734091728925705,
-0.013606734573841095,
0.15984565019607544,
-0.06655866652727127,
-0.19821006059646606,
-0.0020318704191595316,
-0.043261609971523285,
-0.09833352267742157,
-0.011635067872703075,
0.06519658118486404,
0.00865224190056324,
0.0883428230881691,
-0.01265645120292902,
-0.03351152315735817,
0.2025451511144638,
-0.014186820015311241,
-0.03136245906352997,
-0.07079348713159561,
0.13974224030971527,
-0.10326956957578659,
0.15414899587631226,
0.015630846843123436,
0.009019892662763596,
0.0749143436551094,
-0.03012930415570736,
-0.11355700343847275,
-0.01461013127118349,
0.037095945328474045,
-0.0671505555510521,
0.00009613870497560129,
0.10148853063583374,
0.012334673665463924,
0.05422426387667656,
0.07336172461509705,
-0.07035555690526962,
-0.008877377957105637,
0.02878561243414879,
-0.004875405225902796,
-0.06803647428750992,
0.09071464091539383,
-0.1064910963177681,
0.13220466673374176,
0.11683981120586395,
-0.06224098429083824,
0.00990093220025301,
-0.002677120268344879,
0.05841897800564766,
0.04006095230579376,
0.02797686867415905,
-0.05403248593211174,
-0.0995945930480957,
-0.005493384785950184,
-0.07366085052490234,
0.023043615743517876,
-0.20379766821861267,
-0.027889085933566093,
0.003357448847964406,
0.03500594198703766,
-0.01596221886575222,
0.08528508991003036,
0.06825753301382065,
0.026859281584620476,
0.00735887885093689,
0.01934093050658703,
-0.03717296943068504,
0.10769812762737274,
-0.11220132559537888,
-0.07685409486293793
] |
null | null |
transformers
|
This is an upload of the bert-base-nli-stsb-mean-tokens pretrained model from the Sentence Transformers Repo (https://github.com/UKPLab/sentence-transformers)
|
{"license": "apache-2.0"}
| null |
deepset/sentence_bert
|
[
"transformers",
"pytorch",
"jax",
"bert",
"license:apache-2.0",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #license-apache-2.0 #endpoints_compatible #has_space #region-us
|
This is an upload of the bert-base-nli-stsb-mean-tokens pretrained model from the Sentence Transformers Repo (URL
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #license-apache-2.0 #endpoints_compatible #has_space #region-us \n"
] |
[
38
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #license-apache-2.0 #endpoints_compatible #has_space #region-us \n"
] |
[
-0.016910919919610023,
0.10248049348592758,
-0.006167077459394932,
0.005642659496515989,
0.0030453517101705074,
0.0009843799052760005,
0.07655095309019089,
0.1297871470451355,
0.04545893892645836,
-0.019738852977752686,
0.15832974016666412,
0.16395561397075653,
-0.040429528802633286,
0.018473951146006584,
-0.021905120462179184,
-0.16954675316810608,
0.11573167890310287,
0.0527140311896801,
-0.10793671011924744,
0.08015234768390656,
0.09611628949642181,
-0.05906696617603302,
0.034576185047626495,
0.004871354904025793,
-0.0566474124789238,
0.021250156685709953,
0.026691587641835213,
-0.06602100282907486,
0.10448329895734787,
0.005732004530727863,
0.09615810960531235,
0.05369183048605919,
-0.05625239014625549,
-0.1296817511320114,
0.024602416902780533,
0.039368338882923126,
-0.07285524904727936,
0.05905221775174141,
-0.001447133719921112,
-0.02423952892422676,
0.03377837315201759,
0.001651956932619214,
-0.016889773309230804,
0.003992914222180843,
-0.08549802750349045,
-0.30560994148254395,
-0.09807944297790527,
0.09043937176465988,
-0.003501222236081958,
0.07065468281507492,
0.052282996475696564,
0.16932916641235352,
-0.1437225043773651,
0.015665296465158463,
0.23138225078582764,
-0.3714556097984314,
-0.003123335540294647,
0.12947669625282288,
0.10298724472522736,
0.017767712473869324,
-0.006904336623847485,
0.048407990485429764,
0.04429301619529724,
0.01701285131275654,
0.09711326658725739,
-0.04384080693125725,
-0.06993427872657776,
0.08580130338668823,
-0.07319226861000061,
-0.11602863669395447,
0.2705032229423523,
0.015909960493445396,
0.05684652179479599,
0.016967296600341797,
-0.07674764841794968,
0.005953999701887369,
0.0017177419504150748,
0.03684728965163231,
0.04942481964826584,
0.10691677778959274,
0.028055623173713684,
-0.060325492173433304,
-0.15561482310295105,
0.014509354718029499,
-0.1591901034116745,
0.09196049720048904,
0.01085839793086052,
0.11517269164323807,
-0.15143170952796936,
0.0515904501080513,
-0.007053168956190348,
-0.10618311911821365,
0.010169409215450287,
-0.08900574594736099,
0.1378069967031479,
0.0795004814863205,
-0.11021382361650467,
0.1123732253909111,
0.1291649043560028,
0.19998688995838165,
0.015572095289826393,
-0.02831379510462284,
0.036565572023391724,
0.14330486953258514,
-0.027268506586551666,
0.048286326229572296,
-0.029779357835650444,
0.005264696665108204,
0.10078609734773636,
-0.08122905343770981,
0.07918698340654373,
-0.03893888741731644,
-0.10740593820810318,
-0.04723140597343445,
0.0063432794995605946,
0.08215747773647308,
0.11619237810373306,
0.011791209690272808,
-0.03202160820364952,
0.03898068144917488,
0.13175009191036224,
-0.031211819499731064,
0.016838476061820984,
-0.021244272589683533,
0.04101730138063431,
0.06377584487199783,
0.04912427440285683,
0.01864377036690712,
0.027976609766483307,
0.04074063152074814,
-0.07590655982494354,
-0.02552241086959839,
-0.05672170966863632,
-0.013766714371740818,
0.1058579608798027,
-0.0700160413980484,
0.060442112386226654,
-0.15455229580402374,
-0.06463668495416641,
0.04099145531654358,
0.09868206083774567,
0.027153514325618744,
-0.09696801751852036,
0.11713570356369019,
-0.059549588710069656,
0.060023196041584015,
-0.09434115886688232,
0.014349587261676788,
-0.07271114736795425,
0.05925144627690315,
-0.09710896760225296,
0.05891101062297821,
-0.15487954020500183,
0.04673054441809654,
-0.08984388411045074,
0.030899139121174812,
-0.0368594154715538,
-0.09545665979385376,
-0.08766073733568192,
0.17063458263874054,
-0.028139403089880943,
-0.02962556853890419,
-0.05496501922607422,
0.02682952955365181,
-0.01963483914732933,
0.07120345532894135,
-0.09899014234542847,
-0.02937314845621586,
0.17558972537517548,
-0.07866278290748596,
-0.20268751680850983,
0.044482775032520294,
0.028945570811629295,
-0.012369127944111824,
0.00008714482828509063,
0.17929278314113617,
0.02313307859003544,
-0.09715095907449722,
0.016263660043478012,
0.1626156121492386,
-0.06112174317240715,
-0.20845183730125427,
0.08455988764762878,
-0.03884139657020569,
-0.0638781413435936,
0.031867656856775284,
-0.06801249086856842,
0.09252029657363892,
0.0055066621862351894,
-0.07216960191726685,
-0.09620214998722076,
-0.05426919460296631,
-0.008437328040599823,
0.009891243651509285,
0.0613691471517086,
-0.06679793447256088,
-0.05989845097064972,
-0.011250932700932026,
0.05134138464927673,
0.06527405977249146,
0.08658262342214584,
-0.03570028021931648,
0.09869161993265152,
-0.029050055891275406,
0.00009357494855066761,
-0.12351664900779724,
-0.006122638005763292,
-0.01380826160311699,
-0.02699177898466587,
-0.030582496896386147,
0.1817939281463623,
0.06706643104553223,
-0.12851431965827942,
0.008905072696506977,
-0.004506411962211132,
0.05569023638963699,
0.07519449293613434,
-0.002853626385331154,
-0.11734131723642349,
-0.016404423862695694,
-0.057447440922260284,
-0.042125385254621506,
0.03472740575671196,
0.027915483340620995,
0.0684824213385582,
0.11234349012374878,
-0.056900955736637115,
0.0664447620511055,
-0.028239618986845016,
-0.03515790030360222,
-0.04840945079922676,
-0.00881635956466198,
0.10496081411838531,
0.04785778373479843,
-0.05963706970214844,
0.21996663510799408,
-0.023991873487830162,
0.3027169704437256,
0.2267005294561386,
-0.15120327472686768,
0.09391004592180252,
0.08485882729291916,
-0.050899483263492584,
0.006771416869014502,
0.06541150063276291,
-0.04345856234431267,
-0.05367552489042282,
-0.012624042108654976,
0.07945016771554947,
-0.05040512606501579,
-0.04991185665130615,
-0.022438211366534233,
-0.06047186255455017,
-0.03162934258580208,
0.01395969744771719,
0.10020200163125992,
-0.10267625749111176,
0.18912197649478912,
0.43754905462265015,
-0.046419959515333176,
0.031708162277936935,
-0.08541743457317352,
-0.0010762845631688833,
-0.003960081841796637,
-0.04219219833612442,
-0.06755267083644867,
0.0908760204911232,
-0.16821515560150146,
-0.020787566900253296,
0.09999436140060425,
0.04413902014493942,
0.047905333340168,
-0.13931052386760712,
-0.06986454874277115,
0.033755891025066376,
0.023869303986430168,
-0.0781647115945816,
0.0967843309044838,
-0.009743405506014824,
0.06804356724023819,
-0.004234272986650467,
-0.09151115268468857,
0.09667595475912094,
-0.0016159865772351623,
-0.018896838650107384,
0.08554085344076157,
-0.15175873041152954,
-0.1954115480184555,
-0.05556044727563858,
-0.06068262457847595,
0.012188808992505074,
-0.01649777963757515,
0.13594408333301544,
-0.0245549026876688,
-0.027286160737276077,
0.011201477609574795,
-0.06840641796588898,
-0.07795258611440659,
0.04326240345835686,
-0.007000449113547802,
0.03785603865981102,
0.017464281991124153,
-0.1267467886209488,
-0.06458991765975952,
0.016897791996598244,
-0.048412617295980453,
0.05359548702836037,
-0.02202005870640278,
0.06803222745656967,
0.06802481412887573,
0.055534958839416504,
0.020193548873066902,
-0.0042395018972456455,
0.1842501014471054,
-0.016395043581724167,
0.005524872336536646,
0.20995140075683594,
0.031162938103079796,
0.07294339686632156,
0.12308638542890549,
0.05592111125588417,
-0.021895859390497208,
-0.024092689156532288,
-0.04701012745499611,
-0.09581936150789261,
-0.18664738535881042,
-0.07114915549755096,
-0.12223777174949646,
-0.008516530506312847,
0.020006537437438965,
0.08370073139667511,
0.09964434057474136,
0.05530006065964699,
-0.0036795069463551044,
-0.03605888411402702,
-0.06373850256204605,
0.04217659309506416,
0.18009158968925476,
-0.0617491640150547,
0.09650865197181702,
-0.1012101024389267,
-0.020822644233703613,
0.11145881563425064,
0.10687249153852463,
0.12040109932422638,
0.10303002595901489,
0.016690362244844437,
0.11411985009908676,
0.28103891015052795,
0.07510800659656525,
0.09365462511777878,
-0.012998620979487896,
-0.02757076360285282,
-0.05667839199304581,
-0.026343150064349174,
-0.028015634045004845,
0.08200524747371674,
0.04889862984418869,
-0.11901114881038666,
0.012330952100455761,
-0.25722894072532654,
0.0648675486445427,
0.1819709986448288,
0.05389858037233353,
-0.08731933683156967,
0.018994126468896866,
0.10355180501937866,
0.0074045551009476185,
-0.0007210721378214657,
0.09042476862668991,
0.0037135935854166746,
-0.056282930076122284,
0.050907254219055176,
0.007872486487030983,
0.10335063934326172,
0.07379923015832901,
0.052159860730171204,
-0.038636624813079834,
-0.14824151992797852,
0.09020676463842392,
0.09582380950450897,
-0.2547306418418884,
0.2023238241672516,
-0.03498181700706482,
-0.07337937504053116,
-0.05016053467988968,
-0.008491870947182178,
0.05342133715748787,
0.18800580501556396,
0.08389904350042343,
0.06232769042253494,
-0.14825525879859924,
0.01911015808582306,
-0.03537885844707489,
0.03596213459968567,
-0.008754013106226921,
-0.03262319043278694,
-0.047085970640182495,
-0.04729238525032997,
-0.006999526638537645,
0.031790487468242645,
0.21539275348186493,
-0.017316093668341637,
-0.1295548677444458,
0.04551184922456741,
0.12176478654146194,
-0.0013336631236597896,
-0.08579473197460175,
-0.01072733011096716,
-0.09609955549240112,
0.12566760182380676,
-0.029144585132598877,
-0.043125350028276443,
-0.06199588254094124,
-0.16370180249214172,
0.10600308328866959,
-0.03997751325368881,
0.06745976209640503,
-0.066566601395607,
-0.020748581737279892,
-0.07669617980718613,
-0.18584264814853668,
0.10280454903841019,
-0.1441183239221573,
0.004081354476511478,
-0.04309888184070587,
0.1293710321187973,
-0.11962348222732544,
0.04745182767510414,
0.018298223614692688,
0.0465797521173954,
-0.1795087307691574,
-0.10777603834867477,
0.011470692232251167,
0.03622725233435631,
0.0678086131811142,
-0.077622190117836,
-0.03678647801280022,
0.04542839527130127,
0.09671622514724731,
-0.016891049221158028,
0.18872101604938507,
0.1855035126209259,
-0.12810645997524261,
0.1519455909729004,
0.14715217053890228,
-0.04309256002306938,
-0.27909931540489197,
-0.17489711940288544,
-0.17161424458026886,
-0.0857793465256691,
0.03790367394685745,
-0.09649991989135742,
0.10163544118404388,
0.051291435956954956,
-0.11969748139381409,
0.08396565169095993,
-0.21037742495536804,
-0.04638510197401047,
0.13509446382522583,
-0.06877940893173218,
0.39626726508140564,
-0.15082114934921265,
-0.04643392562866211,
-0.024482471868395805,
-0.23170943558216095,
0.14014972746372223,
-0.11592882871627808,
0.04364319145679474,
-0.015328356064856052,
0.015172254294157028,
-0.013967657461762428,
-0.07182835042476654,
0.1589476317167282,
-0.04741617664694786,
0.008674286305904388,
-0.08554194867610931,
-0.03908291459083557,
0.13088230788707733,
-0.06197386234998703,
0.014323643408715725,
-0.15560591220855713,
0.014768215827643871,
-0.1263612061738968,
0.02662474289536476,
-0.1299612820148468,
0.09700324386358261,
-0.018351271748542786,
-0.03616955503821373,
-0.06037711352109909,
-0.011754529550671577,
0.03125036507844925,
-0.0015304269036278129,
0.2966870367527008,
0.04053118824958801,
0.0990411788225174,
0.109428271651268,
-0.05415281280875206,
-0.21176764369010925,
-0.12031058222055435,
-0.03956475481390953,
-0.07286884635686874,
0.09220989793539047,
-0.2202887237071991,
0.03595577925443649,
0.06009165570139885,
-0.04093237221240997,
-0.0076546939089894295,
0.07309602946043015,
-0.0318346805870533,
-0.041478585451841354,
0.1427193284034729,
-0.15872006118297577,
-0.04504205659031868,
0.011233503930270672,
0.13742846250534058,
0.07619839906692505,
0.05849246680736542,
0.10266821831464767,
-0.007117974106222391,
-0.04026518017053604,
0.01391669362783432,
0.02660275623202324,
-0.12603163719177246,
-0.03620991110801697,
0.0710417628288269,
0.02681247889995575,
-0.09667908400297165,
0.06143273040652275,
0.007609385997056961,
-0.17334693670272827,
-0.0357680507004261,
0.04516410827636719,
-0.08503258228302002,
-0.1330588161945343,
-0.004064100328832865,
-0.03053988516330719,
-0.09162383526563644,
-0.08400165289640427,
0.004839261528104544,
-0.0978340432047844,
0.04060289263725281,
0.08653184026479721,
0.12423541396856308,
0.08099297434091568,
0.03249600902199745,
-0.039038002490997314,
0.07708092778921127,
-0.04625619202852249,
-0.10289856046438217,
0.021872634068131447,
-0.10518471151590347,
-0.09114869683980942,
0.01964730955660343,
0.10560079663991928,
-0.04655051231384277,
0.018675273284316063,
-0.10205730050802231,
0.012485628947615623,
-0.11665186285972595,
-0.0815148875117302,
-0.10941756516695023,
-0.048975035548210144,
0.03319433331489563,
-0.12604835629463196,
-0.04406110569834709,
0.028722962364554405,
-0.13506139814853668,
-0.03525496646761894,
-0.01682192273437977,
0.056990522891283035,
-0.11387123167514801,
-0.056697867810726166,
0.111452616751194,
-0.020282581448554993,
0.07192627340555191,
0.0959448367357254,
-0.04815677925944328,
0.07846462726593018,
-0.036959208548069,
-0.11236490309238434,
0.05797024071216583,
0.04079528525471687,
0.04098009690642357,
-0.005101475864648819,
-0.04043180122971535,
0.08326882869005203,
0.0032924185506999493,
0.006090724840760231,
-0.018935557454824448,
-0.11425504088401794,
-0.061898913234472275,
0.0311527568846941,
-0.10167945921421051,
0.025118723511695862,
-0.1317092925310135,
0.19839347898960114,
0.047903817147016525,
0.13542115688323975,
0.06077355518937111,
0.025172045454382896,
-0.06641896814107895,
0.029728543013334274,
-0.046094268560409546,
-0.16551876068115234,
-0.07090143859386444,
-0.027276357635855675,
-0.04588649794459343,
-0.03393029049038887,
0.27074331045150757,
0.009556083008646965,
-0.11720498651266098,
0.05591805651783943,
0.03557436913251877,
0.01828707940876484,
0.010553530417382717,
0.24610662460327148,
0.03424694016575813,
-0.011754574254155159,
-0.06865482777357101,
0.02975543960928917,
0.017566828057169914,
-0.07862772047519684,
0.041805610060691833,
0.1505281776189804,
0.18631482124328613,
0.07978000491857529,
0.04963648319244385,
-0.007012915797531605,
-0.14259372651576996,
-0.18247804045677185,
0.05112561583518982,
0.08925941586494446,
-0.045688893646001816,
0.11428626626729965,
0.1427115499973297,
-0.027351954951882362,
0.04270575940608978,
-0.07155323773622513,
0.045385438948869705,
-0.13833032548427582,
-0.08343588560819626,
-0.03289807587862015,
-0.15322887897491455,
-0.028569919988512993,
-0.04120549187064171,
0.02965446561574936,
0.17070578038692474,
0.019782213494181633,
-0.024477537721395493,
-0.024955593049526215,
0.0317901074886322,
-0.0294073224067688,
0.006217452697455883,
0.00781631376594305,
-0.02377496473491192,
-0.02672189101576805,
-0.015596903860569,
-0.09413120895624161,
-0.04366620257496834,
-0.059112951159477234,
0.030123455449938774,
-0.042453084141016006,
0.028170159086585045,
-0.08230847120285034,
-0.07303563505411148,
-0.08216496556997299,
0.017518052831292152,
0.006518492475152016,
0.15676502883434296,
-0.008505047298967838,
0.03790995478630066,
0.054064422845840454,
0.2249927818775177,
-0.1077226772904396,
-0.08599767088890076,
-0.03157021850347519,
0.10880465805530548,
-0.0033972468227148056,
0.0486147403717041,
-0.005423514638096094,
0.03787776827812195,
-0.0934038981795311,
0.26264458894729614,
0.27781590819358826,
-0.04037363454699516,
0.06626178324222565,
0.015202452428638935,
0.013962304219603539,
0.009713972918689251,
0.0787876546382904,
0.1625901460647583,
0.19748392701148987,
-0.1126280203461647,
-0.014653544872999191,
-0.047769565135240555,
0.021603355184197426,
-0.1675303727388382,
0.03635765239596367,
-0.038430143147706985,
-0.1181599423289299,
-0.015797074884176254,
0.03911231830716133,
-0.04355466738343239,
0.04145212844014168,
0.045696813613176346,
-0.17558157444000244,
-0.033951956778764725,
0.002566714771091938,
0.24177512526512146,
0.005126002244651318,
0.07558618485927582,
-0.05088648200035095,
-0.04828526824712753,
0.10669751465320587,
-0.019824277609586716,
-0.18729479610919952,
-0.08042757213115692,
0.1508195549249649,
0.004608257208019495,
0.1234707236289978,
-0.019915111362934113,
0.016239341348409653,
0.092505544424057,
0.058648739010095596,
-0.1313299834728241,
0.03692768141627312,
0.04224051907658577,
-0.10917527228593826,
-0.10555869340896606,
-0.13624073565006256,
-0.03849482536315918,
-0.0563741996884346,
0.03803805261850357,
-0.09700176864862442,
0.025022657588124275,
0.05254500359296799,
-0.0036299810744822025,
-0.040934886783361435,
-0.011179294437170029,
-0.05911029502749443,
0.07249666005373001,
-0.023528268560767174,
-0.04804884269833565,
-0.09546837210655212,
-0.0340646468102932,
-0.03314351662993431,
0.054983966052532196,
-0.13735416531562805,
-0.12482979148626328,
0.039060819894075394,
-0.008633434772491455,
0.05650176852941513,
0.015400216914713383,
0.003652151906862855,
-0.05978192016482353,
-0.007621489930897951,
0.01640329882502556,
-0.09372013807296753,
-0.009807606227695942,
0.0784071534872055,
-0.010805018246173859,
0.007861630991101265,
-0.03278050944209099,
0.011256027966737747,
0.014616541564464569,
-0.0976698026061058,
-0.07941406965255737
] |
null | null |
transformers
|
This model contains the converted PyTorch checkpoint of the original Tensorflow model available in the [TaPas repository](https://github.com/google-research/tapas/blob/master/DENSE_TABLE_RETRIEVER.md#reader-models).
It is described in Herzig et al.'s (2021) [paper](https://aclanthology.org/2021.naacl-main.43/) _Open Domain Question Answering over Tables via Dense Retrieval_.
This model has 2 versions that can be used differing only in the table scoring head.
The default one has an adapted table scoring head in order to be able to generate probabilities out of the logits.
The other (non-default) version corresponds to the original checkpoint from the TaPas repository and can be accessed by setting `revision="original"`.
# Usage
## In Haystack
If you want to use this model for question-answering over tables, you can load it in [Haystack](https://github.com/deepset-ai/haystack/):
```python
from haystack.nodes import TableReader
table_reader = TableReader(model_name_or_path="deepset/tapas-large-nq-hn-reader")
```
|
{"language": "en", "license": "apache-2.0", "tags": ["tapas"]}
| null |
deepset/tapas-large-nq-hn-reader
|
[
"transformers",
"pytorch",
"tapas",
"en",
"license:apache-2.0",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #tapas #en #license-apache-2.0 #endpoints_compatible #has_space #region-us
|
This model contains the converted PyTorch checkpoint of the original Tensorflow model available in the TaPas repository.
It is described in Herzig et al.'s (2021) paper _Open Domain Question Answering over Tables via Dense Retrieval_.
This model has 2 versions that can be used differing only in the table scoring head.
The default one has an adapted table scoring head in order to be able to generate probabilities out of the logits.
The other (non-default) version corresponds to the original checkpoint from the TaPas repository and can be accessed by setting 'revision="original"'.
# Usage
## In Haystack
If you want to use this model for question-answering over tables, you can load it in Haystack:
|
[
"# Usage",
"## In Haystack\nIf you want to use this model for question-answering over tables, you can load it in Haystack:"
] |
[
"TAGS\n#transformers #pytorch #tapas #en #license-apache-2.0 #endpoints_compatible #has_space #region-us \n",
"# Usage",
"## In Haystack\nIf you want to use this model for question-answering over tables, you can load it in Haystack:"
] |
[
38,
3,
30
] |
[
"passage: TAGS\n#transformers #pytorch #tapas #en #license-apache-2.0 #endpoints_compatible #has_space #region-us \n# Usage## In Haystack\nIf you want to use this model for question-answering over tables, you can load it in Haystack:"
] |
[
-0.0318656787276268,
0.12018289417028427,
-0.003673463361337781,
-0.007599232252687216,
0.018855806440114975,
0.014460130594670773,
0.09814095497131348,
0.129785418510437,
0.07715006917715073,
-0.030954379588365555,
0.1533592790365219,
0.24688643217086792,
0.0020522726699709892,
0.0111305620521307,
-0.017390724271535873,
-0.19266954064369202,
0.09252854436635971,
0.04495718330144882,
-0.13492363691329956,
0.1371454894542694,
0.11810145527124405,
-0.07121220231056213,
0.057088129222393036,
0.03260057792067528,
-0.09378109127283096,
-0.0053507075645029545,
-0.06688962131738663,
-0.04565627872943878,
0.13995027542114258,
0.01902954839169979,
0.08552748709917068,
0.03266727924346924,
-0.04742186889052391,
-0.10776099562644958,
0.05149327963590622,
0.04283528029918671,
-0.03830203413963318,
0.0835954025387764,
0.02759295143187046,
-0.002083870116621256,
0.10052560269832611,
0.04159686714410782,
-0.012535675428807735,
0.0014740165788680315,
-0.09310632944107056,
-0.19785955548286438,
-0.10075121372938156,
0.04419262334704399,
0.12802520394325256,
0.046853989362716675,
-0.004485419485718012,
0.1699981391429901,
-0.1591973453760147,
0.0311940535902977,
0.20520508289337158,
-0.27417027950286865,
-0.051331814378499985,
0.09636469185352325,
0.07700631022453308,
-0.0277447160333395,
-0.03342175483703613,
0.06507549434900284,
0.045767150819301605,
0.04628540948033333,
0.03560147434473038,
-0.08576586097478867,
-0.17731927335262299,
0.07994741201400757,
-0.11967074126005173,
-0.06440003961324692,
0.257246732711792,
0.0025192827451974154,
0.039504215121269226,
-0.11297495663166046,
-0.08974041044712067,
0.21691520512104034,
-0.00699354475364089,
-0.03286892920732498,
0.029582180082798004,
0.05484234541654587,
0.14526046812534332,
-0.06966821849346161,
-0.13468967378139496,
-0.024924471974372864,
-0.11020627617835999,
-0.10541952401399612,
0.016186611726880074,
0.07771249115467072,
-0.10980938374996185,
0.03430485725402832,
0.006346389651298523,
-0.0888390764594078,
0.030330246314406395,
-0.13022969663143158,
0.02615300379693508,
0.043659910559654236,
-0.051347002387046814,
0.12992426753044128,
0.11887484043836594,
0.0755910873413086,
0.08910300582647324,
0.010950703173875809,
0.10621535032987595,
0.10325615108013153,
0.0695623904466629,
-0.002685646526515484,
-0.14669321477413177,
-0.000613535288721323,
-0.04114098846912384,
-0.03074866160750389,
-0.04362645372748375,
-0.06018853932619095,
-0.10594917833805084,
-0.0659857764840126,
-0.008263017050921917,
0.0692012831568718,
0.16896601021289825,
-0.019346488639712334,
0.012808844447135925,
-0.061620522290468216,
0.09867101907730103,
-0.025605961680412292,
0.00008764333324506879,
0.0005162663292139769,
-0.007563082966953516,
0.17753109335899353,
0.010017202235758305,
0.08078129589557648,
-0.007586392108350992,
0.009524268098175526,
-0.05762369930744171,
-0.007940001785755157,
-0.08346276730298996,
-0.011077838949859142,
0.055672891438007355,
-0.12453848123550415,
0.04552433267235756,
-0.10050946474075317,
-0.1096503734588623,
-0.025306032970547676,
0.09568959474563599,
0.03729427233338356,
-0.08956390619277954,
0.06061837449669838,
-0.03984028473496437,
0.004089389927685261,
-0.04970373213291168,
0.16217736899852753,
-0.05237111821770668,
0.05092117562890053,
-0.1771743893623352,
0.025255955755710602,
-0.05059202387928963,
0.04481656849384308,
-0.07516667246818542,
0.03995216637849808,
-0.052107203751802444,
-0.020436102524399757,
-0.02059146575629711,
0.15600381791591644,
-0.06772606074810028,
-0.036608655005693436,
-0.001288748113438487,
0.009693758562207222,
-0.09429595619440079,
0.09074397385120392,
-0.221710205078125,
0.046248823404312134,
0.07987724989652634,
-0.07464750856161118,
-0.312043696641922,
0.07409603893756866,
-0.020473506301641464,
0.10085238516330719,
0.034202639013528824,
0.2936501204967499,
0.1301589012145996,
-0.05919470265507698,
0.06796310096979141,
0.14796583354473114,
-0.08087935298681259,
-0.048098452389240265,
0.10313147306442261,
-0.05634741485118866,
-0.042545367032289505,
-0.009775099344551563,
-0.0755721926689148,
-0.020175842568278313,
0.017981166020035744,
-0.10073071718215942,
-0.05617136135697365,
-0.0470925010740757,
-0.07678490877151489,
0.04757809638977051,
-0.005850144196301699,
-0.039877597242593765,
-0.020152097567915916,
-0.023397628217935562,
0.06710219383239746,
0.01616482436656952,
0.00023580106790177524,
-0.05225351080298424,
0.09153292328119278,
0.0189469363540411,
0.01180373877286911,
-0.1387786865234375,
-0.005641158204525709,
-0.07665975391864777,
-0.009557848796248436,
-0.051198769360780716,
0.3217286169528961,
0.01932588778436184,
-0.0889011099934578,
0.06913131475448608,
0.05386006832122803,
0.10607842355966568,
0.04248116537928581,
-0.027713242918252945,
-0.09643957763910294,
-0.050380852073431015,
-0.056368615478277206,
0.004067966248840094,
0.015776846557855606,
0.011997836641967297,
-0.002158503048121929,
0.06474502384662628,
-0.08928125351667404,
0.08069107681512833,
0.009304853156208992,
-0.02342909201979637,
-0.02709522657096386,
-0.016200030222535133,
0.109781414270401,
-0.0034767882898449898,
-0.02597092092037201,
0.1275096982717514,
-0.1002843827009201,
0.2740793526172638,
0.15902890264987946,
-0.16961640119552612,
0.033996231853961945,
-0.034988075494766235,
-0.04787353798747063,
-0.0009042575838975608,
-0.02617395482957363,
0.030785011127591133,
0.08387406915426254,
0.019740447402000427,
0.009337466210126877,
-0.08268480747938156,
-0.025672778487205505,
-0.037997230887413025,
-0.06381931900978088,
0.021421413868665695,
0.04338516294956207,
0.055453281849622726,
-0.15853697061538696,
0.10103514045476913,
0.3559892475605011,
0.00686520803719759,
-0.0005431724130176008,
-0.06126841902732849,
-0.023878466337919235,
-0.058959949761629105,
0.014307638630270958,
-0.059634916484355927,
0.10214600712060928,
-0.24704182147979736,
0.017389554530382156,
0.11468149721622467,
0.03853176534175873,
0.07757045328617096,
-0.09949894994497299,
-0.058772213757038116,
0.007523069158196449,
-0.018589632585644722,
-0.037071239203214645,
0.10685190558433533,
0.022814780473709106,
0.03755145147442818,
-0.022130178287625313,
-0.021228214725852013,
0.030248494818806648,
-0.015215812250971794,
-0.06319724768400192,
0.15536925196647644,
-0.07343315333127975,
-0.19365458190441132,
0.03289983049035072,
-0.009060494601726532,
0.05586624890565872,
0.008669805712997913,
0.13917310535907745,
0.04197962209582329,
-0.011522148735821247,
0.0026434550527483225,
-0.02009156346321106,
-0.05623363330960274,
-0.05655394122004509,
-0.014713532291352749,
0.008167024701833725,
-0.032738298177719116,
-0.10535669326782227,
0.0004872761492151767,
0.05465047061443329,
0.010893257334828377,
0.05430462956428528,
-0.06768280267715454,
0.09252526611089706,
0.10098542273044586,
0.03232913464307785,
0.026854747906327248,
0.005090336315333843,
0.22835266590118408,
-0.059134721755981445,
0.05635375156998634,
0.2133713811635971,
0.03480204567313194,
0.09693628549575806,
0.1336032897233963,
0.03845556825399399,
-0.009282635524868965,
-0.041163649410009384,
-0.002400970319285989,
-0.08290369808673859,
-0.14990252256393433,
-0.0316164568066597,
-0.08255762606859207,
-0.024198895320296288,
0.027955610305070877,
0.0019706166349351406,
0.0381825752556324,
0.1877506971359253,
-0.01908908039331436,
-0.11184371262788773,
-0.042179301381111145,
0.01036708615720272,
0.13496240973472595,
-0.06795595586299896,
0.11535969376564026,
-0.08713135123252869,
-0.051995642483234406,
0.07913874089717865,
0.10602468997240067,
0.14893239736557007,
0.09945216774940491,
0.03787887468934059,
0.17552456259727478,
0.23540015518665314,
0.015227212570607662,
0.07697253674268723,
-0.022283710539340973,
-0.031140819191932678,
-0.09471268206834793,
-0.05031740665435791,
-0.03612811863422394,
0.07608731836080551,
0.08766106516122818,
-0.07219120860099792,
0.001963094575330615,
-0.1290142685174942,
0.09534507989883423,
0.14731085300445557,
0.045919667929410934,
-0.12121106684207916,
0.05680684745311737,
0.09498568624258041,
0.014404174871742725,
-0.04514982923865318,
0.09625523537397385,
-0.040190912783145905,
-0.15296536684036255,
0.011662469245493412,
-0.01597152091562748,
0.18453416228294373,
0.0844372883439064,
0.028239822015166283,
-0.0658145621418953,
-0.13168998062610626,
0.03567160665988922,
0.21345482766628265,
-0.23254072666168213,
0.13859085738658905,
-0.018215861171483994,
-0.06380985677242279,
-0.12222126871347427,
-0.04664145037531853,
0.03951094299554825,
0.12441706657409668,
0.11418775469064713,
-0.002369687194004655,
-0.12650923430919647,
0.08716852217912674,
-0.02056720294058323,
0.04288473725318909,
-0.03277001157402992,
-0.0993662104010582,
-0.0028387655038386583,
-0.04849987104535103,
0.06123867630958557,
-0.0060567730106413364,
0.1913524717092514,
0.03951716423034668,
-0.14524874091148376,
0.03383753448724747,
0.04853682219982147,
-0.0069083054549992085,
-0.03580809757113457,
-0.004460718482732773,
-0.14206752181053162,
0.02156936563551426,
-0.02856043539941311,
-0.1016676053404808,
-0.06889132410287857,
-0.2271491289138794,
0.1394987553358078,
-0.037331681698560715,
-0.006382128223776817,
-0.07302484661340714,
0.022229760885238647,
-0.03800806775689125,
-0.10535185039043427,
0.06183331832289696,
-0.13669653236865997,
0.014885863289237022,
0.014822972007095814,
0.10664983838796616,
-0.0542348250746727,
0.016722198575735092,
0.10919462889432907,
0.04547128453850746,
-0.2595003843307495,
-0.1836012303829193,
-0.05938558280467987,
-0.034609198570251465,
0.15302152931690216,
-0.001245285733602941,
0.01922113075852394,
0.08539459854364395,
0.0077703287824988365,
0.06637150794267654,
0.14558011293411255,
0.1161523088812828,
-0.12996582686901093,
0.13842043280601501,
0.12152431160211563,
-0.02644304931163788,
-0.20124423503875732,
-0.1532965749502182,
-0.11280836910009384,
-0.05285101756453514,
0.05355861783027649,
-0.03912568464875221,
0.12595954537391663,
0.07616545259952545,
-0.07860774546861649,
0.11072951555252075,
-0.2887853682041168,
-0.054207343608140945,
0.18436558544635773,
-0.026943031698465347,
0.24610619246959686,
-0.18719597160816193,
0.014925963245332241,
0.020014194771647453,
-0.11508689075708389,
0.11812189966440201,
-0.19359464943408966,
0.014306294731795788,
-0.013040395453572273,
0.12981876730918884,
-0.016420740634202957,
-0.07616928964853287,
0.24260073900222778,
0.020807838067412376,
0.024427104741334915,
-0.06565012782812119,
0.017341146245598793,
0.15412519872188568,
-0.02903028577566147,
0.05880683287978172,
-0.14931802451610565,
0.08411017060279846,
-0.23218601942062378,
0.0075078136287629604,
-0.09409145265817642,
0.07543084770441055,
-0.0739595890045166,
-0.051917675882577896,
-0.07548697292804718,
-0.003933205269277096,
0.0353509783744812,
0.010166682302951813,
0.14186127483844757,
-0.017702335491776466,
0.13683484494686127,
0.10956500470638275,
0.0210439283400774,
-0.18650653958320618,
-0.25426092743873596,
-0.0667150691151619,
-0.03686688095331192,
0.0693478062748909,
-0.14461246132850647,
0.012051942758262157,
0.1013164296746254,
-0.008833201602101326,
0.01693618670105934,
0.041339945048093796,
-0.0350310355424881,
-0.014339183457195759,
0.0462011992931366,
-0.17860403656959534,
-0.09737249463796616,
0.021432705223560333,
0.16576404869556427,
-0.06613844633102417,
-0.004763202276080847,
0.06165878474712372,
0.02463976852595806,
-0.02808939479291439,
0.05632568523287773,
0.005443807225674391,
-0.05111268162727356,
0.0198760237544775,
0.10282615572214127,
0.07055187225341797,
-0.09059140086174011,
0.005145261995494366,
-0.014668458141386509,
-0.14799918234348297,
-0.05732150375843048,
-0.027068546041846275,
-0.11016084253787994,
-0.16652218997478485,
-0.035496216267347336,
0.010971952229738235,
-0.09345721453428268,
-0.08494171500205994,
0.06988247483968735,
-0.09401228278875351,
0.01248821523040533,
-0.007836633361876011,
0.1446201503276825,
0.04388454556465149,
0.028565309941768646,
-0.10556384176015854,
0.036403317004442215,
0.03510433807969093,
-0.032051876187324524,
-0.03977341577410698,
-0.11623124033212662,
-0.0588700957596302,
0.026896797120571136,
0.09607311338186264,
-0.05380486324429512,
-0.01729268580675125,
-0.057069651782512665,
0.02490546740591526,
-0.18267548084259033,
0.030533038079738617,
-0.16406762599945068,
-0.006750545930117369,
0.031145738437771797,
-0.06062375754117966,
-0.04431678354740143,
-0.040991269052028656,
-0.09411563724279404,
0.021808473393321037,
-0.00239736121147871,
0.054685309529304504,
-0.13025571405887604,
-0.061769455671310425,
0.12172955274581909,
-0.010927300900220871,
0.010609337128698826,
0.10680235177278519,
-0.1078597903251648,
0.025294229388237,
0.088980533182621,
-0.09040790796279907,
-0.0027171780820935965,
0.05365721881389618,
0.023430608212947845,
-0.17465157806873322,
-0.06042477488517761,
0.05042737349867821,
0.04310893639922142,
0.013262317515909672,
0.0015164974611252546,
-0.09501338005065918,
-0.005277649499475956,
0.03305251523852348,
-0.09511563926935196,
0.016555512323975563,
-0.08945099264383316,
0.10640377551317215,
0.07758431881666183,
0.17016105353832245,
0.08310232311487198,
0.12371749430894852,
-0.07294651865959167,
0.015361693687736988,
-0.05802866071462631,
-0.0653248056769371,
0.049526575952768326,
-0.04914451390504837,
-0.03537525609135628,
-0.07642598450183868,
0.2568313777446747,
-0.018930472433567047,
0.11149507015943527,
0.03188498690724373,
-0.045594558119773865,
0.1168709397315979,
-0.010268722660839558,
0.3182036876678467,
0.0002767526893876493,
0.038054730743169785,
-0.0392746739089489,
0.040782708674669266,
-0.07976946979761124,
-0.12931208312511444,
0.05064079165458679,
0.2278323918581009,
0.009266863577067852,
0.010279310867190361,
-0.00908285565674305,
-0.0035163804423063993,
-0.12606751918792725,
-0.08705122023820877,
0.0965668186545372,
0.03169228509068489,
-0.043644342571496964,
0.02318926900625229,
0.22265379130840302,
-0.09940865635871887,
0.054797060787677765,
-0.05464278906583786,
-0.002561194822192192,
-0.14734305441379547,
-0.031029924750328064,
-0.049770254641771317,
-0.07411914318799973,
-0.0446721613407135,
-0.10786280781030655,
0.005454732570797205,
0.27094632387161255,
0.03638627752661705,
0.002028524409979582,
0.07241716235876083,
-0.005235075484961271,
-0.04181833565235138,
-0.015569480136036873,
0.025899315252900124,
0.03984789177775383,
-0.009105323813855648,
0.03118305094540119,
-0.04599899426102638,
-0.0550868883728981,
-0.035444725304841995,
-0.021160796284675598,
-0.021640324965119362,
-0.021194564178586006,
-0.15344738960266113,
-0.05254579707980156,
-0.08558401465415955,
0.040493398904800415,
0.03754202276468277,
0.0417821891605854,
0.022553810849785805,
-0.011079254560172558,
0.005810537841171026,
0.22487850487232208,
-0.11476219445466995,
-0.13823916018009186,
-0.07537852227687836,
0.08330775052309036,
-0.05666632577776909,
0.020018812268972397,
0.025910302996635437,
0.06199565529823303,
-0.022076034918427467,
0.19888433814048767,
0.2781001329421997,
-0.14042143523693085,
0.010708716697990894,
0.0034202400129288435,
0.010472558438777924,
0.041511405259370804,
0.06093413755297661,
0.08281487971544266,
0.1716964691877365,
-0.10592044144868851,
-0.0027656350284814835,
-0.08587683737277985,
0.0005155978142283857,
-0.1820196956396103,
0.0062692128121852875,
-0.002365358406677842,
-0.10640063881874084,
0.047709476202726364,
0.17483119666576385,
-0.044125996530056,
-0.025931937620043755,
-0.04607606679201126,
-0.04177085682749748,
-0.08686727285385132,
-0.003721673972904682,
0.16560542583465576,
-0.017827820032835007,
0.08380264043807983,
-0.08188740164041519,
-0.033097587525844574,
0.15736639499664307,
0.018669789656996727,
-0.15554392337799072,
-0.038303058594465256,
0.1603994369506836,
0.024191292002797127,
0.0479450598359108,
-0.033219967037439346,
0.05718844756484032,
0.08784710615873337,
-0.015849990770220757,
-0.0863124206662178,
0.00760723277926445,
0.11229091137647629,
-0.025535093620419502,
-0.011298957280814648,
-0.16421377658843994,
0.009262384846806526,
-0.12642593681812286,
0.10633740574121475,
-0.22517025470733643,
0.018945002928376198,
-0.016348544508218765,
0.051266562193632126,
-0.04675973206758499,
0.07301416248083115,
-0.06532847136259079,
0.0873674750328064,
0.03644140064716339,
-0.06892123073339462,
-0.04754319787025452,
0.0015183818759396672,
0.0053494516760110855,
0.031139489263296127,
-0.1277114599943161,
-0.1251448541879654,
0.06728483736515045,
-0.06263336539268494,
-0.05609823390841484,
-0.055728185921907425,
-0.09293191134929657,
-0.06513656675815582,
0.02358955517411232,
-0.03089842014014721,
-0.007081660907715559,
0.0268198624253273,
-0.010953187011182308,
0.013566385954618454,
0.002947331639006734,
0.07050389051437378,
-0.06923417001962662,
0.06804744899272919,
-0.1030445322394371,
-0.07192204892635345
] |
null | null |
transformers
|
This model contains the converted PyTorch checkpoint of the original Tensorflow model available in the [TaPas repository](https://github.com/google-research/tapas/blob/master/DENSE_TABLE_RETRIEVER.md#reader-models).
It is described in Herzig et al.'s (2021) [paper](https://aclanthology.org/2021.naacl-main.43/) _Open Domain Question Answering over Tables via Dense Retrieval_.
This model has 2 versions which can be used differing only in the table scoring head.
The default one has an adapted table scoring head in order to be able to generate probabilities out of the logits.
The other (non-default) version corredponds to the original checkpoint from the TaPas repository and can be accessed setting `revision="original"`.
# Usage
## In Haystack
If you want to use this model for question-answering over tables, you can load it in [Haystack](https://github.com/deepset-ai/haystack/):
```python
from haystack.nodes import TableReader
table_reader = TableReader(model_name_or_path="deepset/tapas-large-nq-reader")
```
|
{"language": "en", "license": "apache-2.0", "tags": ["tapas"]}
| null |
deepset/tapas-large-nq-reader
|
[
"transformers",
"pytorch",
"tapas",
"en",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #tapas #en #license-apache-2.0 #endpoints_compatible #region-us
|
This model contains the converted PyTorch checkpoint of the original Tensorflow model available in the TaPas repository.
It is described in Herzig et al.'s (2021) paper _Open Domain Question Answering over Tables via Dense Retrieval_.
This model has 2 versions which can be used differing only in the table scoring head.
The default one has an adapted table scoring head in order to be able to generate probabilities out of the logits.
The other (non-default) version corredponds to the original checkpoint from the TaPas repository and can be accessed setting 'revision="original"'.
# Usage
## In Haystack
If you want to use this model for question-answering over tables, you can load it in Haystack:
|
[
"# Usage",
"## In Haystack\nIf you want to use this model for question-answering over tables, you can load it in Haystack:"
] |
[
"TAGS\n#transformers #pytorch #tapas #en #license-apache-2.0 #endpoints_compatible #region-us \n",
"# Usage",
"## In Haystack\nIf you want to use this model for question-answering over tables, you can load it in Haystack:"
] |
[
34,
3,
30
] |
[
"passage: TAGS\n#transformers #pytorch #tapas #en #license-apache-2.0 #endpoints_compatible #region-us \n# Usage## In Haystack\nIf you want to use this model for question-answering over tables, you can load it in Haystack:"
] |
[
-0.05000237748026848,
0.11238006502389908,
-0.004572039470076561,
-0.01347530260682106,
0.0390021912753582,
0.026151152327656746,
0.11647796630859375,
0.12585631012916565,
0.10980156064033508,
-0.05725234001874924,
0.14758709073066711,
0.2938612997531891,
0.020427631214261055,
0.005252883303910494,
-0.017046548426151276,
-0.19570595026016235,
0.08232852816581726,
0.05191411077976227,
-0.1166926920413971,
0.14458835124969482,
0.12279430031776428,
-0.057402148842811584,
0.05369804799556732,
0.03032490611076355,
-0.0846550464630127,
-0.0033455779775977135,
-0.05673868954181671,
-0.04483165591955185,
0.1437024027109146,
0.025447146967053413,
0.0769989937543869,
0.027812698855996132,
-0.0438837856054306,
-0.13679447770118713,
0.05419599637389183,
0.02354276180267334,
-0.032948028296232224,
0.08005934208631516,
0.04543783888220787,
0.002196656307205558,
0.08590211719274521,
0.07561538368463516,
-0.011058545671403408,
0.016884587705135345,
-0.09002962708473206,
-0.18925487995147705,
-0.10284101963043213,
0.05562393367290497,
0.15916627645492554,
0.05066938325762749,
-0.003344646655023098,
0.16658158600330353,
-0.14957767724990845,
0.0420081801712513,
0.17164231836795807,
-0.2679212689399719,
-0.04361559450626373,
0.06349819898605347,
0.04140782356262207,
-0.03350280970335007,
-0.028905294835567474,
0.05372730642557144,
0.04389237239956856,
0.03999452665448189,
-0.02067789062857628,
-0.0964672863483429,
-0.16424055397510529,
0.060914747416973114,
-0.10945191979408264,
-0.057711150497198105,
0.25748664140701294,
0.0048747616820037365,
0.018060147762298584,
-0.09604266285896301,
-0.07260744273662567,
0.24395370483398438,
-0.01858689822256565,
-0.03899513930082321,
0.02053573913872242,
0.0660938024520874,
0.1619429588317871,
-0.06388857960700989,
-0.12464446574449539,
-0.03627638518810272,
-0.0979500263929367,
-0.09877344965934753,
0.035442713648080826,
0.07883469015359879,
-0.13067670166492462,
0.030769944190979004,
0.02752440981566906,
-0.0810566395521164,
0.020532988011837006,
-0.13157454133033752,
0.023087428882718086,
0.03610970452427864,
-0.026942716911435127,
0.13739246129989624,
0.13442566990852356,
0.09050693362951279,
0.09859112650156021,
0.02881753072142601,
0.08536365628242493,
0.09717538207769394,
0.06472396105527878,
-0.016088444739580154,
-0.11625654250383377,
0.02507418394088745,
-0.026730384677648544,
-0.04571104794740677,
-0.04353464022278786,
-0.03859815374016762,
-0.09376367181539536,
-0.04916362091898918,
-0.01860332116484642,
0.09956490248441696,
0.1452140063047409,
-0.029734091833233833,
0.0004904206143692136,
-0.07507502287626266,
0.08637111634016037,
-0.024888882413506508,
-0.01796809583902359,
0.006372298579663038,
-0.01242870930582285,
0.21151243150234222,
-0.011649076826870441,
0.08682843297719955,
-0.02706056833267212,
0.02334403805434704,
-0.05374351516366005,
-0.006783108226954937,
-0.06643456965684891,
0.022435756400227547,
0.05647343397140503,
-0.12166453897953033,
0.055668268352746964,
-0.0773942619562149,
-0.15641078352928162,
-0.02015441469848156,
0.09068252146244049,
0.04195296764373779,
-0.07836344093084335,
0.0360863171517849,
-0.008821032010018826,
-0.030202336609363556,
-0.051716048270463943,
0.13635119795799255,
-0.053597480058670044,
0.06416738778352737,
-0.162281334400177,
0.000690302113071084,
-0.03904258459806442,
0.042121462523937225,
-0.0834868997335434,
0.025093141943216324,
-0.039443954825401306,
-0.021472563967108727,
-0.02946195937693119,
0.1708775907754898,
-0.0736202821135521,
-0.03962743654847145,
0.02762937918305397,
0.0020749696996062994,
-0.09961387515068054,
0.11627689749002457,
-0.21007363498210907,
0.02940976992249489,
0.12018238008022308,
-0.10465694963932037,
-0.32475435733795166,
0.08072689920663834,
-0.018016653135418892,
0.10161194205284119,
0.07642979174852371,
0.26541075110435486,
0.134622260928154,
-0.050734490156173706,
0.10549269616603851,
0.1374012976884842,
-0.09351551532745361,
-0.05965813621878624,
0.10194291174411774,
-0.07717733085155487,
-0.09316699951887131,
0.001127356314100325,
-0.07681809365749359,
-0.02051219530403614,
0.0024671629071235657,
-0.10300491750240326,
-0.03549144044518471,
-0.04266515001654625,
-0.09956768155097961,
0.061494603753089905,
-0.018781056627631187,
-0.038873810321092606,
0.0073909941129386425,
-0.055508218705654144,
0.07602129876613617,
0.01139499805867672,
-0.014136316254734993,
-0.07297095656394958,
0.08497215062379837,
0.03074314445257187,
0.01984051614999771,
-0.1369570642709732,
-0.009719876572489738,
-0.06826320290565491,
-0.016338057816028595,
-0.0418139286339283,
0.2751673460006714,
0.015851642936468124,
-0.07826516032218933,
0.07399611175060272,
0.047066278755664825,
0.12226340174674988,
0.027829352766275406,
-0.010508918203413486,
-0.10913614928722382,
-0.03364517539739609,
-0.04782555624842644,
0.021327700465917587,
0.02260906808078289,
-0.0023325143847614527,
-0.009078165516257286,
0.05156874284148216,
-0.08985964208841324,
0.07806690782308578,
0.008593026548624039,
-0.022138016298413277,
-0.006335278507322073,
-0.004619981627911329,
0.12604397535324097,
-0.0025396449491381645,
-0.035089924931526184,
0.08989313989877701,
-0.061271145939826965,
0.2560311555862427,
0.14153693616390228,
-0.17998456954956055,
0.029708223417401314,
-0.058889925479888916,
-0.04416612908244133,
-0.006250131409615278,
-0.015821563079953194,
0.062461402267217636,
0.11363139748573303,
0.03909459337592125,
0.016234317794442177,
-0.07722499966621399,
-0.01710052601993084,
-0.03483675420284271,
-0.06445956975221634,
0.039855048060417175,
0.04898744076490402,
0.0744461715221405,
-0.18633542954921722,
0.08091708272695541,
0.296986848115921,
0.03152192756533623,
-0.0041255769319832325,
-0.08502720296382904,
-0.030053773894906044,
-0.04419029876589775,
0.03542491793632507,
-0.04470355063676834,
0.084813691675663,
-0.2588992118835449,
0.03871583938598633,
0.10523796081542969,
0.0219856109470129,
0.07719958573579788,
-0.10530372709035873,
-0.0677310898900032,
-0.008858765475451946,
-0.03354659676551819,
-0.05835260823369026,
0.09331291913986206,
0.010175843723118305,
0.03474551439285278,
-0.02586071379482746,
-0.0174135472625494,
0.04634673520922661,
-0.028889212757349014,
-0.08412310481071472,
0.17541925609111786,
-0.0757647305727005,
-0.1973547786474228,
0.007985885255038738,
-0.020701967179775238,
0.04621712490916252,
0.0011099851690232754,
0.1371043175458908,
0.019827835261821747,
-0.02784128300845623,
0.038707390427589417,
-0.017070719972252846,
-0.042183924466371536,
-0.06879273056983948,
0.006895922124385834,
0.006924803834408522,
-0.040289245545864105,
-0.10326217114925385,
-0.0020564210135489702,
0.0401863269507885,
0.011435926891863346,
0.05051330849528313,
-0.10242806375026703,
0.08458879590034485,
0.09979327023029327,
0.05234742909669876,
0.03546052426099777,
-0.0019286512397229671,
0.21990832686424255,
-0.050993662327528,
0.03413332998752594,
0.2104574590921402,
0.006706526502966881,
0.10434231907129288,
0.11064279824495316,
0.04912925511598587,
-0.026924753561615944,
-0.0400727279484272,
-0.012550700455904007,
-0.07119229435920715,
-0.19599410891532898,
-0.027903534471988678,
-0.06775509566068649,
-0.025061802938580513,
0.020424040034413338,
-0.0012761888792738318,
0.05394509807229042,
0.1958276629447937,
-0.022453321143984795,
-0.12395426630973816,
-0.022942623123526573,
0.01786956749856472,
0.1428271234035492,
-0.04955385625362396,
0.09057772159576416,
-0.09369482100009918,
-0.04590396210551262,
0.08721130341291428,
0.09454787522554398,
0.16163353621959686,
0.12240877747535706,
0.07417860627174377,
0.1848597526550293,
0.21813718974590302,
0.019351165741682053,
0.09464435279369354,
-0.02815328724682331,
-0.02241329848766327,
-0.09236845374107361,
-0.04594874382019043,
-0.06284584105014801,
0.051484666764736176,
0.05363033339381218,
-0.06008148938417435,
-0.014222066849470139,
-0.10778085142374039,
0.08611316233873367,
0.15548193454742432,
0.02458100952208042,
-0.1248386949300766,
0.04323164001107216,
0.08049407601356506,
0.006740288808941841,
-0.04037867486476898,
0.11004805564880371,
-0.10382122546434402,
-0.1761348694562912,
0.037388045340776443,
-0.03712338209152222,
0.1906318962574005,
0.045081641525030136,
0.015351934358477592,
-0.05485500767827034,
-0.11002929508686066,
0.04395760968327522,
0.2261998951435089,
-0.24541956186294556,
0.15005476772785187,
-0.015470383688807487,
-0.04201839864253998,
-0.12404550611972809,
-0.045051947236061096,
0.037575799971818924,
0.12498152256011963,
0.13499875366687775,
-0.014555790461599827,
-0.09604807198047638,
0.08193362504243851,
-0.025395603850483894,
0.056816473603248596,
-0.02424193173646927,
-0.09178027510643005,
-0.011088990606367588,
-0.0522620715200901,
0.05434083566069603,
-0.020825574174523354,
0.1115824356675148,
0.028633303940296173,
-0.13203448057174683,
0.03003499284386635,
0.028499338775873184,
0.03233591094613075,
-0.033096522092819214,
-0.0040862346068024635,
-0.10883229225873947,
0.02941901795566082,
-0.07223007082939148,
-0.12338398396968842,
-0.0723256841301918,
-0.218834787607193,
0.12540140748023987,
-0.04952767491340637,
-0.010214105248451233,
-0.08205726742744446,
0.01707143522799015,
-0.018264560028910637,
-0.11239870637655258,
0.05187171325087547,
-0.14748255908489227,
0.004397343844175339,
0.018829787150025368,
0.13664117455482483,
-0.03200175240635872,
-0.0038460539653897285,
0.11755486577749252,
0.019126497209072113,
-0.2339126318693161,
-0.18981531262397766,
-0.06079864501953125,
-0.018421046435832977,
0.14163337647914886,
0.016729537397623062,
0.012508271262049675,
0.08764369785785675,
-0.032063256949186325,
0.060046929866075516,
0.14981667697429657,
0.1178484857082367,
-0.09140029549598694,
0.14670220017433167,
0.16730359196662903,
-0.04400630667805672,
-0.16845454275608063,
-0.15880721807479858,
-0.10521627962589264,
-0.07743224501609802,
0.01760614849627018,
-0.07363326102495193,
0.14495611190795898,
0.08632083982229233,
-0.06425818055868149,
0.12509311735630035,
-0.2927197813987732,
-0.04567020386457443,
0.19594958424568176,
0.008096802979707718,
0.2513546645641327,
-0.18335740268230438,
-0.004001652356237173,
0.012617671862244606,
-0.1305249184370041,
0.11628663539886475,
-0.14459438621997833,
0.0063679697923362255,
-0.01633983477950096,
0.12188705801963806,
-0.02425265870988369,
-0.07776862382888794,
0.23213577270507812,
0.04086138308048248,
0.018789565190672874,
-0.05453914403915405,
0.03711175173521042,
0.14108353853225708,
0.0015616794116795063,
0.06090015918016434,
-0.11147008836269379,
0.09648829698562622,
-0.22954781353473663,
-0.012396293692290783,
-0.08504687249660492,
0.061821334064006805,
-0.06900618970394135,
-0.05198065936565399,
-0.046571917831897736,
-0.0095454016700387,
0.024626825004816055,
0.010694354772567749,
0.13085950911045074,
-0.013226560316979885,
0.12962037324905396,
0.06413043290376663,
0.06069696694612503,
-0.19097433984279633,
-0.2631758451461792,
-0.10115259140729904,
-0.04216885194182396,
0.06286831200122833,
-0.0958339273929596,
0.010687671601772308,
0.1250438094139099,
-0.0028748894110322,
0.02341596595942974,
0.04425502568483353,
-0.0397987887263298,
-0.011879964731633663,
0.03713022172451019,
-0.18612393736839294,
-0.09456442296504974,
0.019249500706791878,
0.15596704185009003,
-0.047772184014320374,
0.01690990850329399,
0.05902344733476639,
0.04598693922162056,
-0.03752846643328667,
0.05452294647693634,
0.013660740107297897,
-0.04016803205013275,
0.030293617397546768,
0.09745001047849655,
0.05916747823357582,
-0.10793251544237137,
0.01840636320412159,
-0.03627447783946991,
-0.15509188175201416,
-0.061204664409160614,
-0.008909787982702255,
-0.12953659892082214,
-0.1526678204536438,
-0.04007767140865326,
0.05449025332927704,
-0.1260605901479721,
-0.11368294805288315,
0.04604872688651085,
-0.09238824248313904,
0.014196003787219524,
0.003031390719115734,
0.14032624661922455,
0.04149125888943672,
0.01672166958451271,
-0.11120346188545227,
0.023989327251911163,
0.030292484909296036,
-0.02752046100795269,
-0.05493016913533211,
-0.11064556986093521,
-0.08738814294338226,
0.036199238151311874,
0.09738801419734955,
-0.05078575387597084,
-0.04860783740878105,
-0.05629236251115799,
0.05018150806427002,
-0.2010902762413025,
0.061383411288261414,
-0.1681845486164093,
0.005857819691300392,
0.03308916091918945,
-0.05648030340671539,
-0.03052077628672123,
-0.03986240550875664,
-0.08992923051118851,
0.03406999260187149,
0.014123492874205112,
0.0438840351998806,
-0.13837100565433502,
-0.055093713104724884,
0.13083645701408386,
-0.003885540645569563,
0.02775510959327221,
0.12718380987644196,
-0.11477691680192947,
0.054783858358860016,
0.061910126358270645,
-0.07624560594558716,
0.0003033495740965009,
0.05507931113243103,
0.014391304925084114,
-0.18125098943710327,
-0.048830047249794006,
0.06437695026397705,
0.021258510649204254,
0.01860824227333069,
0.018081406131386757,
-0.10016126185655594,
-0.019632739946246147,
0.05255262926220894,
-0.1017252653837204,
0.008621025830507278,
-0.09488750249147415,
0.10595276951789856,
0.07053383439779282,
0.18792134523391724,
0.07120312750339508,
0.13647741079330444,
-0.05666227638721466,
0.008879008702933788,
-0.05053044855594635,
-0.06109873577952385,
0.02514968067407608,
-0.05671146139502525,
-0.04537058621644974,
-0.076886847615242,
0.24238382279872894,
-0.052219003438949585,
0.15405894815921783,
0.02940220944583416,
-0.05854138359427452,
0.11577674001455307,
-0.011236305348575115,
0.3397325277328491,
0.007785000838339329,
0.04539492353796959,
-0.04712119325995445,
0.0365072526037693,
-0.0948958545923233,
-0.1676645576953888,
0.06941059976816177,
0.2036009430885315,
-0.00796521082520485,
-0.0005156151601113379,
-0.0002741722564678639,
0.02808021567761898,
-0.11083359271287918,
-0.11488992720842361,
0.0974467322230339,
0.008840851485729218,
-0.005921497941017151,
0.04323893040418625,
0.2123342901468277,
-0.09119962155818939,
0.04545636475086212,
-0.04376037418842316,
-0.01342655811458826,
-0.16465140879154205,
-0.030225075781345367,
-0.048985276371240616,
-0.07646271586418152,
-0.020507344976067543,
-0.11219886690378189,
-0.005771949887275696,
0.25099870562553406,
0.047792449593544006,
-0.01576509326696396,
0.0698775053024292,
-0.0038190563209354877,
-0.04282236471772194,
-0.00418110704049468,
0.022649042308330536,
0.02140786498785019,
-0.0027088758070021868,
0.03383971005678177,
-0.05226296931505203,
-0.06370887160301208,
-0.038192640990018845,
-0.014243381097912788,
-0.02539372257888317,
-0.021709226071834564,
-0.15405148267745972,
-0.05229553207755089,
-0.07173776626586914,
0.051409535109996796,
0.029070205986499786,
0.03343244642019272,
0.0226974468678236,
-0.004984306637197733,
0.019238144159317017,
0.19592921435832977,
-0.09397895634174347,
-0.14402949810028076,
-0.0710589662194252,
0.13294203579425812,
-0.0555853508412838,
0.019959215074777603,
0.05334583297371864,
0.07403703778982162,
0.003608470084145665,
0.22321107983589172,
0.247327983379364,
-0.12014161050319672,
-0.004333017393946648,
0.0035253572277724743,
0.013785086572170258,
0.04959602653980255,
0.08921154588460922,
0.07180991768836975,
0.19518670439720154,
-0.10666671395301819,
-0.005126623436808586,
-0.07989862561225891,
-0.002363702515140176,
-0.17145180702209473,
-0.009909242391586304,
-0.0017032665200531483,
-0.10399851202964783,
0.046860627830028534,
0.2163076251745224,
-0.06532569974660873,
-0.02393208257853985,
-0.04667989909648895,
-0.0016543554374948144,
-0.08900679647922516,
-0.02908855304121971,
0.14179106056690216,
-0.0024628143291920424,
0.07355040311813354,
-0.07832285761833191,
-0.04218374192714691,
0.18005961179733276,
0.03573049604892731,
-0.18425899744033813,
-0.06665799766778946,
0.15900461375713348,
0.010126910172402859,
0.04393070936203003,
-0.02276202291250229,
0.09177610278129578,
0.07131005823612213,
-0.018683308735489845,
-0.08291774988174438,
0.00437560211867094,
0.11506405472755432,
-0.011894352734088898,
-0.0017693274421617389,
-0.16471220552921295,
0.020332537591457367,
-0.13781136274337769,
0.09584018588066101,
-0.21609462797641754,
0.02017744816839695,
-0.007111691869795322,
0.03498814254999161,
-0.05139090120792389,
0.10696860402822495,
-0.08378013223409653,
0.08410470932722092,
0.033810559660196304,
-0.06710926443338394,
-0.040044236928224564,
-0.00350141990929842,
0.03106330893933773,
0.03913208469748497,
-0.14032866060733795,
-0.10825463384389877,
0.05653565376996994,
-0.0695769339799881,
-0.05364514887332916,
-0.05879853665828705,
-0.10121413320302963,
-0.05178013816475868,
0.0027933556120842695,
-0.02999030239880085,
-0.02378874644637108,
0.030560828745365143,
-0.02150007337331772,
0.029509318992495537,
0.012042302638292313,
0.07506892085075378,
-0.06545117497444153,
0.06899066269397736,
-0.10696790367364883,
-0.07371867448091507
] |
null | null |
transformers
|
## Overview
**Language model:** deepset/tinybert-6L-768D-squad2
**Language:** English
**Training data:** SQuAD 2.0 training set x 20 augmented + SQuAD 2.0 training set without augmentation
**Eval data:** SQuAD 2.0 dev set
**Infrastructure**: 1x V100 GPU
**Published**: Dec 8th, 2021
## Details
- haystack's intermediate layer and prediction layer distillation features were used for training (based on [TinyBERT](https://arxiv.org/pdf/1909.10351.pdf)). deepset/bert-base-uncased-squad2 was used as the teacher model and huawei-noah/TinyBERT_General_6L_768D was used as the student model.
## Hyperparameters
### Intermediate layer distillation
```
batch_size = 26
n_epochs = 5
max_seq_len = 384
learning_rate = 5e-5
lr_schedule = LinearWarmup
embeds_dropout_prob = 0.1
temperature = 1
```
### Prediction layer distillation
```
batch_size = 26
n_epochs = 5
max_seq_len = 384
learning_rate = 3e-5
lr_schedule = LinearWarmup
embeds_dropout_prob = 0.1
temperature = 1
distillation_loss_weight = 1.0
```
## Performance
```
"exact": 71.87736882001179
"f1": 76.36111895973675
```
## Authors
- Timo Möller: `timo.moeller [at] deepset.ai`
- Julian Risch: `julian.risch [at] deepset.ai`
- Malte Pietsch: `malte.pietsch [at] deepset.ai`
- Michel Bartels: `michel.bartels [at] deepset.ai`
## About us

We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.
Some of our work:
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
- [FARM](https://github.com/deepset-ai/FARM)
- [Haystack](https://github.com/deepset-ai/haystack/)
Get in touch:
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community/join) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "en", "license": "mit", "tags": ["exbert"], "datasets": ["squad_v2"], "thumbnail": "https://thumb.tildacdn.com/tild3433-3637-4830-a533-353833613061/-/resize/720x/-/format/webp/germanquad.jpg", "model-index": [{"name": "deepset/tinybert-6l-768d-squad2", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 73.8248, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMGFmZmFiN2E5ODZkOTkyMjQ1NTUzMmQwMjc0M2RlYzVlNmM4YTFlNzA4YzIwY2JkY2EyNDg2ZTY3OTdjZTVlZiIsInZlcnNpb24iOjF9.ZZ6c2OI3lzeNhuSWTh28j00zk-sPrqkTvdVBZv2wJc1D4YnR-xOj72haybT6MV_xeYqTg3-x9L8PsWSS20NaDw"}, {"type": "f1", "value": 77.1684, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNzAxMDk1YzI5ZjA2N2ZmMzAxNjgxYzJiNzAzYmI1ZWU5ZDRmYWY3OWJmMjlmNDcyMGE0YWY5NjNhZTk4YWY5ZSIsInZlcnNpb24iOjF9.rF3raNGUSYv5D2xzWLZztD99vwDKvWb22LG32RomrDGP6XKTbCVqZzAw5UFw93jKb0VoLApbQQ-AOGxLj3U_Cg"}]}]}]}
|
question-answering
|
deepset/tinybert-6l-768d-squad2
|
[
"transformers",
"pytorch",
"safetensors",
"bert",
"question-answering",
"exbert",
"en",
"dataset:squad_v2",
"arxiv:1909.10351",
"license:mit",
"model-index",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1909.10351"
] |
[
"en"
] |
TAGS
#transformers #pytorch #safetensors #bert #question-answering #exbert #en #dataset-squad_v2 #arxiv-1909.10351 #license-mit #model-index #endpoints_compatible #region-us
|
## Overview
Language model: deepset/tinybert-6L-768D-squad2
Language: English
Training data: SQuAD 2.0 training set x 20 augmented + SQuAD 2.0 training set without augmentation
Eval data: SQuAD 2.0 dev set
Infrastructure: 1x V100 GPU
Published: Dec 8th, 2021
## Details
- haystack's intermediate layer and prediction layer distillation features were used for training (based on TinyBERT). deepset/bert-base-uncased-squad2 was used as the teacher model and huawei-noah/TinyBERT_General_6L_768D was used as the student model.
## Hyperparameters
### Intermediate layer distillation
### Prediction layer distillation
## Performance
## Authors
- Timo Möller: 'timo.moeller [at] URL'
- Julian Risch: 'URL [at] URL'
- Malte Pietsch: 'malte.pietsch [at] URL'
- Michel Bartels: 'michel.bartels [at] URL'
## About us
!deepset logo
We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.
Some of our work:
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
- FARM
- Haystack
Get in touch:
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"## Overview\nLanguage model: deepset/tinybert-6L-768D-squad2 \nLanguage: English \nTraining data: SQuAD 2.0 training set x 20 augmented + SQuAD 2.0 training set without augmentation \nEval data: SQuAD 2.0 dev set \nInfrastructure: 1x V100 GPU \nPublished: Dec 8th, 2021",
"## Details\n- haystack's intermediate layer and prediction layer distillation features were used for training (based on TinyBERT). deepset/bert-base-uncased-squad2 was used as the teacher model and huawei-noah/TinyBERT_General_6L_768D was used as the student model.",
"## Hyperparameters",
"### Intermediate layer distillation",
"### Prediction layer distillation",
"## Performance",
"## Authors\n- Timo Möller: 'timo.moeller [at] URL'\n- Julian Risch: 'URL [at] URL'\n- Malte Pietsch: 'malte.pietsch [at] URL'\n- Michel Bartels: 'michel.bartels [at] URL'",
"## About us\n!deepset logo\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #bert #question-answering #exbert #en #dataset-squad_v2 #arxiv-1909.10351 #license-mit #model-index #endpoints_compatible #region-us \n",
"## Overview\nLanguage model: deepset/tinybert-6L-768D-squad2 \nLanguage: English \nTraining data: SQuAD 2.0 training set x 20 augmented + SQuAD 2.0 training set without augmentation \nEval data: SQuAD 2.0 dev set \nInfrastructure: 1x V100 GPU \nPublished: Dec 8th, 2021",
"## Details\n- haystack's intermediate layer and prediction layer distillation features were used for training (based on TinyBERT). deepset/bert-base-uncased-squad2 was used as the teacher model and huawei-noah/TinyBERT_General_6L_768D was used as the student model.",
"## Hyperparameters",
"### Intermediate layer distillation",
"### Prediction layer distillation",
"## Performance",
"## Authors\n- Timo Möller: 'timo.moeller [at] URL'\n- Julian Risch: 'URL [at] URL'\n- Malte Pietsch: 'malte.pietsch [at] URL'\n- Michel Bartels: 'michel.bartels [at] URL'",
"## About us\n!deepset logo\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
65,
72,
77,
5,
9,
9,
2,
63,
129
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #bert #question-answering #exbert #en #dataset-squad_v2 #arxiv-1909.10351 #license-mit #model-index #endpoints_compatible #region-us \n## Overview\nLanguage model: deepset/tinybert-6L-768D-squad2 \nLanguage: English \nTraining data: SQuAD 2.0 training set x 20 augmented + SQuAD 2.0 training set without augmentation \nEval data: SQuAD 2.0 dev set \nInfrastructure: 1x V100 GPU \nPublished: Dec 8th, 2021## Details\n- haystack's intermediate layer and prediction layer distillation features were used for training (based on TinyBERT). deepset/bert-base-uncased-squad2 was used as the teacher model and huawei-noah/TinyBERT_General_6L_768D was used as the student model.## Hyperparameters### Intermediate layer distillation### Prediction layer distillation## Performance## Authors\n- Timo Möller: 'timo.moeller [at] URL'\n- Julian Risch: 'URL [at] URL'\n- Malte Pietsch: 'malte.pietsch [at] URL'\n- Michel Bartels: 'michel.bartels [at] URL'## About us\n!deepset logo\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
-0.030680377036333084,
0.15457552671432495,
-0.004966346547007561,
0.07055327296257019,
0.07589439302682877,
0.013183985836803913,
0.15168507397174835,
0.08636295050382614,
0.07686807215213776,
0.06655621528625488,
0.011197487823665142,
0.019339222460985184,
0.08000954985618591,
0.08492740988731384,
0.025584056973457336,
-0.22505627572536469,
-0.0003443033783696592,
-0.10187484323978424,
-0.062275480479002,
0.09655068814754486,
0.1368463784456253,
-0.09560366719961166,
0.09864998608827591,
-0.012274877168238163,
-0.046953532844781876,
0.027824971824884415,
-0.03584488853812218,
-0.015571044757962227,
0.08186828345060349,
0.04077679663896561,
0.09734003990888596,
-0.003023348515853286,
0.04656143859028816,
-0.22390860319137573,
0.02217399887740612,
0.07239850610494614,
0.022389110177755356,
0.03457130491733551,
0.08849705010652542,
-0.007817388512194157,
0.0505138635635376,
-0.09327737987041473,
0.05360455438494682,
0.0504518523812294,
-0.12659937143325806,
-0.1042511984705925,
-0.13278423249721527,
0.10330623388290405,
0.04426174238324165,
0.0393436960875988,
-0.01841629669070244,
0.05362946167588234,
-0.10422272235155106,
0.05167442187666893,
0.15299025177955627,
-0.2524429261684418,
-0.07425827533006668,
0.007816832512617111,
0.050133660435676575,
0.06476171314716339,
-0.14855235815048218,
0.03500037640333176,
0.011382362805306911,
0.028560513630509377,
0.009741860441863537,
-0.026878876611590385,
0.026887571439146996,
0.00015724221884738654,
-0.08292437344789505,
0.0010621236870065331,
0.11310285329818726,
0.010849542915821075,
-0.048386428505182266,
-0.15967634320259094,
0.008100878447294235,
0.11086858808994293,
-0.05429757758975029,
-0.07545521855354309,
0.040619879961013794,
-0.020674798637628555,
0.010012142360210419,
-0.02984848991036415,
-0.07736370712518692,
0.032000478357076645,
0.020164117217063904,
0.10442985594272614,
0.03266769275069237,
0.011473499238491058,
0.02547375299036503,
0.06241992115974426,
0.056679241359233856,
-0.12429981678724289,
-0.034557729959487915,
-0.09717803448438644,
-0.03838565945625305,
-0.056058645248413086,
0.011246686801314354,
0.01002539973706007,
0.06219174712896347,
0.1850593388080597,
-0.08017920702695847,
0.05084739252924919,
0.026162805035710335,
-0.02433389239013195,
0.0367560051381588,
0.17046844959259033,
-0.04042908549308777,
-0.18897277116775513,
-0.05548598989844322,
0.005348853301256895,
-0.04533259943127632,
-0.0042825560085475445,
-0.027996821328997612,
0.028141209855675697,
0.029852593317627907,
0.029859164729714394,
0.050792109221220016,
0.0034496139269322157,
-0.09357943385839462,
-0.076778344810009,
0.09810272604227066,
-0.1015925332903862,
0.08001888543367386,
0.039992135018110275,
-0.045820027589797974,
0.12766750156879425,
-0.058634012937545776,
0.034315358847379684,
-0.053944073617458344,
0.10828248411417007,
-0.037332721054553986,
0.015359217301011086,
-0.08264844864606857,
-0.07147179543972015,
0.047708407044410706,
-0.01193506270647049,
-0.04427099600434303,
-0.06597070395946503,
-0.030957605689764023,
-0.0811108648777008,
0.10396483540534973,
-0.07038110494613647,
0.012528342194855213,
-0.025516998022794724,
-0.03034084290266037,
0.0468360036611557,
0.030710045248270035,
-0.028423095121979713,
-0.026297803968191147,
0.0368509516119957,
-0.0979846641421318,
0.010622689500451088,
0.039461057633161545,
0.02980988845229149,
-0.05354011803865433,
0.002421567915007472,
-0.1931106001138687,
0.08401179313659668,
-0.14072194695472717,
0.08448611199855804,
-0.13984785974025726,
-0.047302089631557465,
0.024014683440327644,
0.012525251135230064,
0.02406877838075161,
0.1277889907360077,
-0.17362067103385925,
-0.0656450092792511,
0.12977200746536255,
-0.062170010060071945,
-0.08109823614358902,
0.15586112439632416,
-0.052602123469114304,
-0.006967376917600632,
0.1195044293999672,
0.1513037085533142,
0.07046443223953247,
-0.15171855688095093,
-0.02515493892133236,
-0.016394346952438354,
0.07150060683488846,
0.1366768330335617,
0.07513079047203064,
-0.05266954377293587,
0.049061063677072525,
-0.003433327656239271,
-0.06657993793487549,
-0.04206934943795204,
-0.04657167196273804,
-0.06808312982320786,
0.047004394233226776,
-0.015190128237009048,
0.10682503879070282,
0.01946139708161354,
-0.016593018546700478,
-0.07743363082408905,
-0.10531767457723618,
0.036961160600185394,
0.02608511783182621,
0.00007489664130844176,
0.004768053535372019,
-0.05453154072165489,
0.02572673372924328,
0.10872338712215424,
-0.01737084798514843,
-0.11042343080043793,
-0.1219237744808197,
0.07969329506158829,
-0.03081875666975975,
0.11236101388931274,
0.013621018268167973,
0.058553118258714676,
0.037983670830726624,
-0.04318681359291077,
-0.021440785378217697,
-0.06779021769762039,
-0.008368493057787418,
-0.02193836309015751,
-0.20481440424919128,
-0.006312400568276644,
-0.04509764537215233,
0.06397084891796112,
-0.09686531126499176,
-0.03936321660876274,
0.060869354754686356,
0.15185736119747162,
0.06777454912662506,
-0.03346911817789078,
-0.05940520763397217,
0.031927209347486496,
0.037112489342689514,
-0.03887846693396568,
0.031330231577157974,
-0.006696591153740883,
-0.009018690325319767,
0.008595560677349567,
0.003938891459256411,
0.01411981787532568,
0.04676670953631401,
-0.038766659796237946,
-0.08819594979286194,
-0.058140829205513,
-0.0500929094851017,
-0.007982552982866764,
-0.030632343143224716,
-0.04895508289337158,
0.20995695888996124,
0.0363081656396389,
0.031696245074272156,
-0.06994874775409698,
-0.08256185799837112,
-0.06290553510189056,
-0.02745574340224266,
-0.03795779496431351,
0.12794923782348633,
-0.015390412881970406,
-0.13201646506786346,
0.10662553459405899,
0.14773166179656982,
0.08727605640888214,
0.16882163286209106,
-0.039012085646390915,
-0.03539659455418587,
-0.050409212708473206,
-0.0006157308234833181,
-0.03135685250163078,
0.07917317003011703,
-0.03405744582414627,
-0.012400769628584385,
0.028007937595248222,
-0.009271975606679916,
-0.0008079216349869967,
-0.08372272551059723,
0.03905293717980385,
0.001324899960309267,
-0.011458665132522583,
0.06317169219255447,
0.005546515341848135,
0.03497790917754173,
0.09210487455129623,
0.061113182455301285,
0.009221521206200123,
-0.04456115886569023,
-0.05672881752252579,
-0.057833753526210785,
0.15384447574615479,
-0.1327534019947052,
-0.19907483458518982,
-0.07542897760868073,
-0.08756807446479797,
-0.08237199485301971,
-0.02852010913193226,
0.0032782526686787605,
-0.08133948594331741,
-0.10384494066238403,
-0.04932965338230133,
0.03883209824562073,
0.05696529150009155,
-0.08578497916460037,
0.017696630209684372,
-0.01172901876270771,
-0.0029484156984835863,
-0.13951872289180756,
-0.01845650188624859,
0.0033456713426858187,
-0.031034186482429504,
-0.03451336547732353,
0.06054779887199402,
0.0390910767018795,
0.04481247812509537,
0.02823600359261036,
-0.003588639432564378,
-0.010237202979624271,
0.3244222402572632,
-0.1375100016593933,
0.1016334816813469,
0.07033436745405197,
-0.021449176594614983,
0.07800371199846268,
0.18965354561805725,
0.0937347561120987,
-0.035843536257743835,
0.005851617082953453,
0.04622763767838478,
0.050813984125852585,
-0.25031405687332153,
-0.11030785739421844,
-0.045346859842538834,
-0.029992425814270973,
0.0492858923971653,
0.03693443536758423,
0.028788557276129723,
-0.02548544853925705,
-0.08849792927503586,
0.01638425886631012,
0.07938379794359207,
0.06287925690412521,
0.05987285077571869,
0.023784946650266647,
0.03731738403439522,
-0.026588505133986473,
-0.03551588952541351,
0.10888487100601196,
0.04179871827363968,
0.15969765186309814,
0.07720917463302612,
0.1564619094133377,
0.07769516110420227,
0.05042469874024391,
-0.015809273347258568,
-0.027823004871606827,
-0.037159260362386703,
0.02097327448427677,
-0.0176255963742733,
-0.06868550181388855,
0.04206119477748871,
0.05974104627966881,
0.04701703041791916,
-0.07274547219276428,
-0.023620491847395897,
-0.03015986829996109,
0.1416064351797104,
0.2220974564552307,
0.024493904784321785,
-0.09548258781433105,
-0.06931468844413757,
0.0564899668097496,
-0.0854027271270752,
-0.04865502566099167,
-0.008089743554592133,
0.044128451496362686,
-0.1794193983078003,
0.0656481683254242,
-0.009881764650344849,
0.1044260635972023,
-0.03744853287935257,
0.006506727077066898,
0.10658028721809387,
0.08276379108428955,
-0.006953992415219545,
0.0507529154419899,
-0.24695946276187897,
0.17279313504695892,
0.009625446982681751,
0.0657510980963707,
-0.050627581775188446,
0.04048842936754227,
0.02703031711280346,
0.013205236755311489,
0.11081449687480927,
0.033606573939323425,
-0.052264392375946045,
-0.03302573412656784,
-0.058312878012657166,
-0.020269770175218582,
0.14372844994068146,
-0.1071043461561203,
0.06036870926618576,
-0.04403205215930939,
-0.02003645710647106,
-0.03460003808140755,
-0.01954309642314911,
-0.189170703291893,
-0.15545067191123962,
0.051028139889240265,
-0.09987097233533859,
0.003127997275441885,
-0.05481374263763428,
-0.052495017647743225,
-0.12019215524196625,
0.17168991267681122,
-0.15547841787338257,
-0.0618160255253315,
-0.11585164815187454,
-0.017114801332354546,
0.06962926685810089,
-0.08261357247829437,
0.026216845959424973,
-0.011527390219271183,
0.10038968920707703,
-0.009550254791975021,
-0.06073533743619919,
0.06384001672267914,
-0.09151974320411682,
-0.1393926739692688,
-0.037067148834466934,
0.15606486797332764,
0.0860186517238617,
0.04434853792190552,
0.02290213294327259,
0.017831994220614433,
-0.023777727037668228,
-0.1190422996878624,
-0.023134300485253334,
0.06486434489488602,
0.03663540631532669,
0.07454017549753189,
-0.09212049841880798,
-0.1080654114484787,
-0.10032740980386734,
0.011111612431704998,
0.09381020069122314,
0.1190115362405777,
-0.05189213529229164,
0.18585734069347382,
0.15752093493938446,
-0.08440176397562027,
-0.22480906546115875,
-0.011045795865356922,
0.0601043775677681,
-0.02457495406270027,
0.023680560290813446,
-0.17923752963542938,
0.12331792712211609,
0.009211809374392033,
-0.008934594690799713,
0.01303380262106657,
-0.23494355380535126,
-0.13092947006225586,
-0.0007626950391568244,
0.007455562707036734,
-0.11142798513174057,
-0.10545221716165543,
-0.07574386149644852,
-0.049195315688848495,
-0.14028435945510864,
0.08383861184120178,
0.0301818884909153,
0.03723575547337532,
0.03223666548728943,
0.013615268282592297,
0.011398605071008205,
-0.020444748923182487,
0.11753688752651215,
0.05322739854454994,
0.05172129347920418,
-0.07012864202260971,
-0.03716141730546951,
0.023288395255804062,
-0.0308703500777483,
0.025207428261637688,
-0.017814313992857933,
0.014071658253669739,
-0.11998386681079865,
-0.006705042440444231,
-0.0828213021159172,
0.053824298083782196,
-0.11093596369028091,
-0.04056565463542938,
-0.07567934691905975,
0.1631498485803604,
0.06977205723524094,
0.004946316126734018,
-0.010250778868794441,
-0.0025935815647244453,
0.08807413280010223,
0.06001663953065872,
0.14231401681900024,
-0.028098076581954956,
-0.07427649945020676,
-0.03041018359363079,
-0.0029534706845879555,
0.034105993807315826,
-0.030659232288599014,
0.06463910639286041,
0.1343170404434204,
0.0033956433180719614,
0.15985703468322754,
-0.009289465844631195,
-0.09702196717262268,
-0.046213723719120026,
0.06610284000635147,
-0.13386470079421997,
-0.19493938982486725,
-0.05062561482191086,
-0.006034445483237505,
-0.06079905852675438,
-0.015360771678388119,
0.15151894092559814,
0.01840069331228733,
-0.057186756283044815,
0.01586144231259823,
0.06701599061489105,
-0.009986047632992268,
0.07599662244319916,
0.029703978449106216,
0.026468081399798393,
-0.08320607244968414,
0.09112486243247986,
0.08265096694231033,
-0.020511018112301826,
-0.009263953194022179,
0.1493276208639145,
-0.0308277178555727,
-0.031486403197050095,
0.05696780979633331,
0.13690032064914703,
-0.0029642379377037287,
0.004986224230378866,
-0.009326962754130363,
-0.11016975343227386,
0.018816027790308,
0.043052759021520615,
0.033120449632406235,
0.007526961155235767,
0.009627570398151875,
0.011720246635377407,
0.022852640599012375,
0.14211955666542053,
0.09245248883962631,
0.018757887184619904,
-0.05140380933880806,
0.028977045789361,
-0.0350201241672039,
-0.015407905913889408,
-0.007684711832553148,
-0.008572637103497982,
-0.14811915159225464,
-0.043775174766778946,
-0.08346201479434967,
-0.0080147385597229,
-0.02374233305454254,
0.034867871552705765,
-0.024449314922094345,
-0.06366294622421265,
-0.0008377711637876928,
0.023609137162566185,
-0.06704161316156387,
-0.031577810645103455,
0.01247929222881794,
0.13140732049942017,
-0.17609836161136627,
0.004824473988264799,
0.10686914622783661,
-0.06511970609426498,
0.08351099491119385,
-0.009528696537017822,
-0.010189817287027836,
0.03416597470641136,
-0.1444869488477707,
-0.03781527280807495,
-0.0779964029788971,
0.04585283622145653,
0.046403657644987106,
-0.13589021563529968,
-0.02323715016245842,
-0.021046895533800125,
0.00961844902485609,
0.008843266405165195,
0.0026511475443840027,
-0.052654929459095,
0.04015200212597847,
-0.016360925510525703,
-0.1486552655696869,
-0.03143483027815819,
0.041245538741350174,
0.12524442374706268,
0.007838805206120014,
0.1434798240661621,
-0.073237344622612,
0.059828996658325195,
-0.12210281938314438,
0.011914263479411602,
0.056016381829977036,
-0.029043028131127357,
-0.1148853600025177,
-0.04861938953399658,
0.05769979953765869,
-0.0012559400638565421,
0.11580182611942291,
0.02840111218392849,
0.038305703550577164,
0.06171901524066925,
-0.030467798933386803,
-0.1191507875919342,
0.06988758593797684,
0.019203532487154007,
-0.019473882392048836,
-0.012652691453695297,
-0.03071526251733303,
-0.06579779088497162,
-0.026709059253335,
0.03327695280313492,
0.19764301180839539,
0.2859917879104614,
0.08611208945512772,
-0.02141888439655304,
0.07937778532505035,
-0.011898193508386612,
-0.14378243684768677,
0.0330609530210495,
-0.0017131632193922997,
0.0626063123345375,
-0.12098834663629532,
0.1082199290394783,
0.049244921654462814,
-0.23949697613716125,
0.09315402060747147,
-0.06262654066085815,
-0.06802120804786682,
-0.05413937196135521,
-0.10520552098751068,
-0.07391726970672607,
-0.074747733771801,
0.0010641765547916293,
-0.14520327746868134,
0.08983252942562103,
0.0239464920014143,
0.05600053071975708,
-0.054275043308734894,
0.09863483905792236,
-0.135512575507164,
-0.044185820966959,
0.1476200968027115,
0.03154226765036583,
0.041477106511592865,
-0.004955723881721497,
-0.009048523381352425,
-0.07861530035734177,
0.08310955762863159,
0.03109763376414776,
0.043332286179065704,
-0.0454409159719944,
-0.051719971001148224,
-0.03937479108572006,
-0.10619822144508362,
0.024568025022745132,
-0.039930310100317,
-0.020983954891562462,
0.15273648500442505,
0.024811862036585808,
0.007153216749429703,
-0.0011242215987294912,
0.19603405892848969,
-0.03821909427642822,
-0.1005173921585083,
-0.17410022020339966,
0.04778052121400833,
-0.015854598954319954,
0.026113586500287056,
0.040077678859233856,
-0.08836575597524643,
-0.021582646295428276,
0.12158649414777756,
0.16434098780155182,
-0.11989297717809677,
0.010949485003948212,
-0.006413726601749659,
0.030634339898824692,
0.024865061044692993,
0.12460624426603317,
0.004911459516733885,
0.17808881402015686,
-0.024089282378554344,
0.023361459374427795,
-0.004515024833381176,
-0.05720818042755127,
-0.09796575456857681,
0.14360012114048004,
0.03523653745651245,
-0.048046909272670746,
-0.1004723310470581,
0.15943782031536102,
-0.0754254162311554,
-0.20053759217262268,
0.0014687139773741364,
-0.1642504781484604,
-0.16143003106117249,
-0.027097158133983612,
0.0962851345539093,
0.04479651153087616,
0.07221110165119171,
0.0587482787668705,
-0.05594232305884361,
0.16889208555221558,
0.03885948657989502,
-0.002872027223929763,
-0.054893702268600464,
0.13145038485527039,
-0.07314115762710571,
0.219553604722023,
0.021742533892393112,
0.04840897023677826,
0.0973486378788948,
-0.0058312308974564075,
-0.07480692863464355,
-0.07826270908117294,
0.07716858386993408,
-0.12979376316070557,
-0.019843263551592827,
0.10935864597558975,
0.008621224202215672,
0.06232289597392082,
0.0993192046880722,
-0.06850720942020416,
0.0507970005273819,
0.12045324593782425,
-0.061513058841228485,
-0.09218662977218628,
0.12274488806724548,
-0.119992695748806,
0.12096946686506271,
0.18375922739505768,
-0.01931517757475376,
0.014277346432209015,
-0.03146461769938469,
0.026931144297122955,
0.00013924094673711807,
0.07620222866535187,
-0.05496659129858017,
-0.19858074188232422,
0.019434398040175438,
-0.05886705964803696,
0.07384181022644043,
-0.08772889524698257,
-0.08216378837823868,
-0.009977945126593113,
0.0734458789229393,
-0.07564340531826019,
0.1529955267906189,
0.06688740104436874,
-0.026658186689019203,
0.005620517767965794,
-0.11820221692323685,
-0.015024387277662754,
0.11012449115514755,
-0.05719645693898201,
0.019180387258529663
] |
null | null |
transformers
|
# tinyroberta-squad2
## Overview
**Language model:** tinyroberta-squad2
**Language:** English
**Training data:** The PILE
**Code:**
**Infrastructure**: 4x Tesla v100
## Hyperparameters
```
batch_size = 96
n_epochs = 4
base_LM_model = "deepset/tinyroberta-squad2-step1"
max_seq_len = 384
learning_rate = 1e-4
lr_schedule = LinearWarmup
warmup_proportion = 0.2
teacher = "deepset/roberta-base"
```
## Distillation
This model was distilled using the TinyBERT approach described in [this paper](https://arxiv.org/pdf/1909.10351.pdf) and implemented in [haystack](https://github.com/deepset-ai/haystack).
We have performed intermediate layer distillation with roberta-base as the teacher which resulted in [deepset/tinyroberta-6l-768d](https://huggingface.co/deepset/tinyroberta-6l-768d).
This model has not been distilled for any specific task. If you are interested in using distillation to improve its performance on a downstream task, you can take advantage of haystack's new [distillation functionality](https://haystack.deepset.ai/guides/model-distillation). You can also check out [deepset/tinyroberta-squad2](https://huggingface.co/deepset/tinyroberta-squad2) for a model that is already distilled on an extractive QA downstream task.
## Usage
### In Transformers
```python
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
model_name = "deepset/tinyroberta-squad2"
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
### In FARM
```python
from farm.modeling.adaptive_model import AdaptiveModel
from farm.modeling.tokenization import Tokenizer
from farm.infer import Inferencer
model_name = "deepset/tinyroberta-squad2"
model = AdaptiveModel.convert_from_transformers(model_name, device="cpu", task_type="question_answering")
tokenizer = Tokenizer.load(model_name)
```
### In haystack
For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in [haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/roberta-base-squad2")
# or
reader = TransformersReader(model_name_or_path="deepset/roberta-base-squad2",tokenizer="deepset/roberta-base-squad2")
```
## Authors
Branden Chan: `branden.chan [at] deepset.ai`
Timo Möller: `timo.moeller [at] deepset.ai`
Malte Pietsch: `malte.pietsch [at] deepset.ai`
Tanay Soni: `tanay.soni [at] deepset.ai`
Michel Bartels: `michel.bartels [at] deepset.ai`
## About us

We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.
Some of our work:
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
- [FARM](https://github.com/deepset-ai/FARM)
- [Haystack](https://github.com/deepset-ai/haystack/)
Get in touch:
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Slack](https://haystack.deepset.ai/community/join) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "en", "license": "cc-by-4.0", "datasets": ["squad_v2"]}
|
question-answering
|
deepset/tinyroberta-6l-768d
|
[
"transformers",
"pytorch",
"safetensors",
"roberta",
"question-answering",
"en",
"dataset:squad_v2",
"arxiv:1909.10351",
"license:cc-by-4.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1909.10351"
] |
[
"en"
] |
TAGS
#transformers #pytorch #safetensors #roberta #question-answering #en #dataset-squad_v2 #arxiv-1909.10351 #license-cc-by-4.0 #endpoints_compatible #region-us
|
# tinyroberta-squad2
## Overview
Language model: tinyroberta-squad2
Language: English
Training data: The PILE
Code:
Infrastructure: 4x Tesla v100
## Hyperparameters
## Distillation
This model was distilled using the TinyBERT approach described in this paper and implemented in haystack.
We have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.
This model has not been distilled for any specific task. If you are interested in using distillation to improve its performance on a downstream task, you can take advantage of haystack's new distillation functionality. You can also check out deepset/tinyroberta-squad2 for a model that is already distilled on an extractive QA downstream task.
## Usage
### In Transformers
### In FARM
### In haystack
For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:
## Authors
Branden Chan: 'URL [at] URL'
Timo Möller: 'timo.moeller [at] URL'
Malte Pietsch: 'malte.pietsch [at] URL'
Tanay Soni: 'URL [at] URL'
Michel Bartels: 'michel.bartels [at] URL'
## About us
!deepset logo
We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.
Some of our work:
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
- FARM
- Haystack
Get in touch:
Twitter | LinkedIn | Slack | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# tinyroberta-squad2",
"## Overview\nLanguage model: tinyroberta-squad2 \nLanguage: English \nTraining data: The PILE \nCode: \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Distillation\nThis model was distilled using the TinyBERT approach described in this paper and implemented in haystack.\nWe have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.\nThis model has not been distilled for any specific task. If you are interested in using distillation to improve its performance on a downstream task, you can take advantage of haystack's new distillation functionality. You can also check out deepset/tinyroberta-squad2 for a model that is already distilled on an extractive QA downstream task.",
"## Usage",
"### In Transformers",
"### In FARM",
"### In haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:",
"## Authors\nBranden Chan: 'URL [at] URL'\nTimo Möller: 'timo.moeller [at] URL'\nMalte Pietsch: 'malte.pietsch [at] URL'\nTanay Soni: 'URL [at] URL'\nMichel Bartels: 'michel.bartels [at] URL'",
"## About us\n!deepset logo\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Slack | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #roberta #question-answering #en #dataset-squad_v2 #arxiv-1909.10351 #license-cc-by-4.0 #endpoints_compatible #region-us \n",
"# tinyroberta-squad2",
"## Overview\nLanguage model: tinyroberta-squad2 \nLanguage: English \nTraining data: The PILE \nCode: \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Distillation\nThis model was distilled using the TinyBERT approach described in this paper and implemented in haystack.\nWe have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.\nThis model has not been distilled for any specific task. If you are interested in using distillation to improve its performance on a downstream task, you can take advantage of haystack's new distillation functionality. You can also check out deepset/tinyroberta-squad2 for a model that is already distilled on an extractive QA downstream task.",
"## Usage",
"### In Transformers",
"### In FARM",
"### In haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:",
"## Authors\nBranden Chan: 'URL [at] URL'\nTimo Möller: 'timo.moeller [at] URL'\nMalte Pietsch: 'malte.pietsch [at] URL'\nTanay Soni: 'URL [at] URL'\nMichel Bartels: 'michel.bartels [at] URL'",
"## About us\n!deepset logo\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Slack | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
63,
9,
33,
5,
146,
3,
6,
5,
36,
71,
129
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #roberta #question-answering #en #dataset-squad_v2 #arxiv-1909.10351 #license-cc-by-4.0 #endpoints_compatible #region-us \n# tinyroberta-squad2## Overview\nLanguage model: tinyroberta-squad2 \nLanguage: English \nTraining data: The PILE \nCode: \nInfrastructure: 4x Tesla v100## Hyperparameters## Distillation\nThis model was distilled using the TinyBERT approach described in this paper and implemented in haystack.\nWe have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.\nThis model has not been distilled for any specific task. If you are interested in using distillation to improve its performance on a downstream task, you can take advantage of haystack's new distillation functionality. You can also check out deepset/tinyroberta-squad2 for a model that is already distilled on an extractive QA downstream task.## Usage### In Transformers### In FARM### In haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:## Authors\nBranden Chan: 'URL [at] URL'\nTimo Möller: 'timo.moeller [at] URL'\nMalte Pietsch: 'malte.pietsch [at] URL'\nTanay Soni: 'URL [at] URL'\nMichel Bartels: 'michel.bartels [at] URL'## About us\n!deepset logo\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Slack | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
-0.05247385799884796,
0.08933421969413757,
-0.005173041485249996,
0.01833818480372429,
0.07175353169441223,
-0.01258452795445919,
0.06767494231462479,
0.11096350848674774,
0.08714795112609863,
0.12495625019073486,
-0.034136977046728134,
-0.00950203463435173,
0.08879033476114273,
0.08533917367458344,
0.08540011942386627,
-0.18082663416862488,
0.023883188143372536,
-0.10508114099502563,
-0.04201635345816612,
0.07342560589313507,
0.12140573561191559,
-0.088780976831913,
0.08707106858491898,
-0.013996237888932228,
0.00134482909925282,
0.018496917560696602,
-0.06843368709087372,
-0.025931552052497864,
0.09561604261398315,
0.04538510739803314,
0.07922983169555664,
-0.01852390542626381,
-0.026873748749494553,
-0.1470535695552826,
0.03683793544769287,
0.09480933845043182,
0.02623622491955757,
0.03113214299082756,
0.07833543419837952,
0.03851041942834854,
0.08121217787265778,
-0.13893009722232819,
0.054314933717250824,
0.07413014769554138,
-0.07210801541805267,
-0.13984444737434387,
-0.13831579685211182,
0.09404830634593964,
0.06244509667158127,
0.031711094081401825,
-0.016044164076447487,
0.015643049031496048,
-0.047256749123334885,
0.03734635189175606,
0.12470506131649017,
-0.21797209978103638,
-0.048785239458084106,
0.01284991018474102,
0.05294372886419296,
0.05413784086704254,
-0.11784907430410385,
0.01668863743543625,
0.009980861097574234,
0.02998249977827072,
-0.005794776137918234,
-0.01620730385184288,
0.08923561871051788,
-0.0237888190895319,
-0.06755614280700684,
-0.02027861401438713,
0.05958808958530426,
0.004330649971961975,
-0.07194937020540237,
-0.20221656560897827,
-0.0022288006730377674,
0.07873287796974182,
-0.05294531211256981,
-0.10002683103084564,
0.023106209933757782,
-0.007922040298581123,
0.04382076486945152,
-0.11786046624183655,
-0.11580786108970642,
0.0699329674243927,
-0.01295481063425541,
0.01787721924483776,
0.03831449896097183,
-0.004750385414808989,
0.016734899953007698,
0.02718898095190525,
-0.023372670635581017,
-0.09387676417827606,
-0.05084547400474548,
-0.08206495642662048,
-0.10373047739267349,
-0.04242248088121414,
-0.013560979627072811,
-0.010489051230251789,
0.07316607236862183,
0.13718236982822418,
-0.04242175817489624,
0.054985590279102325,
-0.018894076347351074,
-0.0305195152759552,
0.013010263442993164,
0.1544259786605835,
-0.041358888149261475,
-0.2126762866973877,
0.006626748945564032,
-0.02640111930668354,
-0.02465401589870453,
-0.02847377210855484,
-0.02188795618712902,
-0.01427500694990158,
0.018770668655633926,
0.05154074728488922,
0.06359533965587616,
0.011169150471687317,
-0.01880260370671749,
-0.07432091236114502,
0.16150100529193878,
-0.13735032081604004,
0.04034649208188057,
0.04064744710922241,
0.012518331408500671,
0.19013109803199768,
-0.0630575567483902,
-0.006622010376304388,
-0.06453103572130203,
0.04421544075012207,
-0.06113738566637039,
-0.009473935700953007,
-0.0579325370490551,
-0.050439320504665375,
0.05299472063779831,
-0.04780350625514984,
-0.08846016228199005,
-0.07514986395835876,
-0.10308942198753357,
-0.08746393024921417,
0.01367749273777008,
-0.039752520620822906,
-0.023666750639677048,
-0.007043725810945034,
-0.0440819188952446,
0.02738250233232975,
0.03905324637889862,
-0.021387815475463867,
-0.003059808164834976,
0.050197143107652664,
-0.11836330592632294,
0.020247239619493484,
0.034022703766822815,
0.033077239990234375,
-0.06036407873034477,
0.007616131566464901,
-0.2588908076286316,
0.1349361091852188,
-0.12022460997104645,
0.08115386962890625,
-0.13462285697460175,
-0.03075362741947174,
0.024375319480895996,
0.02178606018424034,
-0.04080767184495926,
0.08982282131910324,
-0.16241419315338135,
-0.0887097418308258,
0.09161052852869034,
-0.07880356907844543,
-0.0438353531062603,
0.14419864118099213,
-0.0428912490606308,
0.07491352409124374,
0.15029239654541016,
0.16936904191970825,
0.207492858171463,
-0.12896902859210968,
-0.07762844860553741,
-0.037952497601509094,
-0.0017063157865777612,
0.10125024616718292,
0.05744284391403198,
-0.07059668004512787,
0.13042870163917542,
0.028309378772974014,
-0.003989400342106819,
-0.06134700030088425,
0.023611154407262802,
-0.041928213089704514,
0.03796929121017456,
-0.008474966511130333,
0.0472073033452034,
-0.05158155784010887,
-0.017135581001639366,
-0.04403144121170044,
-0.0970301628112793,
-0.04751487076282501,
0.06136265769600868,
-0.001377937849611044,
0.014979816973209381,
-0.0623454749584198,
0.04558304697275162,
0.07479840517044067,
0.02197621762752533,
-0.10722030699253082,
-0.12452980875968933,
0.056428175419569016,
-0.09579397737979889,
0.08750452101230621,
0.028787698596715927,
0.03793315589427948,
0.015848366543650627,
-0.022501111030578613,
0.014237983152270317,
-0.07168294489383698,
-0.0347631573677063,
-0.003121033776551485,
-0.09286452829837799,
-0.04535020887851715,
-0.058462295681238174,
0.08630955219268799,
-0.007466538343578577,
0.0027067107148468494,
0.09839816391468048,
0.1145198866724968,
0.06036798283457756,
-0.04864414036273956,
-0.015303211286664009,
0.02694632299244404,
0.047779571264982224,
-0.04326244443655014,
0.016575820744037628,
-0.005359556060284376,
0.0062286159954965115,
0.027246203273534775,
-0.045308902859687805,
-0.0427388995885849,
0.023776017129421234,
0.06934896111488342,
-0.09027096629142761,
-0.049185506999492645,
-0.01595916412770748,
-0.008413109928369522,
-0.0478358194231987,
-0.08965320885181427,
0.17150145769119263,
0.0522366538643837,
0.03160310536623001,
-0.07969672977924347,
-0.050137318670749664,
-0.03513693809509277,
0.029331117868423462,
-0.007339301984757185,
0.07416126877069473,
-0.06680962443351746,
-0.1376553475856781,
0.08172659575939178,
0.11588577926158905,
0.08395259082317352,
0.18381962180137634,
-0.008509160950779915,
-0.08218297362327576,
-0.0561164990067482,
0.035912659019231796,
-0.012390971183776855,
0.03313335031270981,
-0.040871962904930115,
0.028123650699853897,
0.04793863743543625,
-0.016047082841396332,
0.02098819613456726,
-0.008784431964159012,
0.04928867518901825,
-0.003992339130491018,
-0.01188246626406908,
0.09299461543560028,
0.034334659576416016,
0.09114523977041245,
0.06865380704402924,
0.045009225606918335,
0.09093526750802994,
-0.009097745642066002,
-0.037979986518621445,
-0.06326054781675339,
0.1328878402709961,
-0.10339037328958511,
-0.2644546627998352,
-0.17017421126365662,
-0.04510430991649628,
-0.062142953276634216,
-0.04974237084388733,
0.04869259148836136,
-0.04607752710580826,
-0.08140550553798676,
-0.04233051836490631,
0.04773229733109474,
0.06837071478366852,
-0.0972934141755104,
0.007078797556459904,
0.0227012038230896,
0.0346372127532959,
-0.1190853863954544,
-0.008851826190948486,
0.040094755589962006,
-0.06873371452093124,
-0.046694282442331314,
0.07775481045246124,
0.026995835825800896,
0.036080826073884964,
0.04802156239748001,
-0.009335477836430073,
-0.017971722409129143,
0.20968276262283325,
-0.1371326744556427,
0.135829895734787,
0.11709050089120865,
0.021053310483694077,
0.08983634412288666,
0.13301555812358856,
0.051758043467998505,
0.0023642503656446934,
-0.00508039491251111,
0.07942646741867065,
0.0008166192565113306,
-0.19535072147846222,
-0.09636811912059784,
-0.031156044453382492,
-0.017851894721388817,
-0.000161624513566494,
0.05709274113178253,
0.034502726048231125,
-0.008230326697230339,
-0.11276683211326599,
-0.01180072408169508,
0.019771544262766838,
0.06934979557991028,
0.15025126934051514,
0.013958708383142948,
0.02933555468916893,
-0.07073552906513214,
-0.010735057294368744,
0.11824671924114227,
0.03142884373664856,
0.12011320888996124,
0.012316809967160225,
0.15732744336128235,
0.045714639127254486,
0.07352368533611298,
0.01678198017179966,
0.00004053208976984024,
0.00861797109246254,
0.01644665002822876,
-0.03798866271972656,
-0.08533477783203125,
0.003163665998727083,
0.05532848834991455,
0.09387568384408951,
-0.05434223264455795,
-0.010995050892233849,
-0.044647373259067535,
0.08245129883289337,
0.21535062789916992,
-0.002584948670119047,
-0.07598696649074554,
-0.09479986131191254,
0.01463258545845747,
-0.10003447532653809,
-0.04421108961105347,
-0.0046281227841973305,
0.07531632483005524,
-0.17267495393753052,
0.06918967515230179,
0.015173705294728279,
0.05795818194746971,
-0.018649837002158165,
0.019686823710799217,
0.054868586361408234,
0.059615351259708405,
-0.00257951021194458,
0.08139081299304962,
-0.18753042817115784,
0.04893813654780388,
0.009471120312809944,
0.07218771427869797,
-0.0415588915348053,
0.043028008192777634,
0.03978179022669792,
-0.04988004267215729,
0.11182590574026108,
0.0049606445245444775,
0.0004821119364351034,
-0.03626488149166107,
-0.0976797416806221,
-0.0069123259745538235,
0.09714151918888092,
-0.11166061460971832,
0.10631927847862244,
-0.03220982104539871,
-0.01479351706802845,
-0.08981525897979736,
0.057639170438051224,
-0.06319703161716461,
-0.19925031065940857,
0.061729587614536285,
-0.08062843978404999,
0.04035466909408569,
-0.04362250119447708,
-0.0377945601940155,
-0.09101035445928574,
0.16856129467487335,
-0.15713483095169067,
-0.12319019436836243,
-0.09665200114250183,
-0.08499596267938614,
0.12688949704170227,
-0.10167628526687622,
0.03396663814783096,
-0.008343111723661423,
0.11749934405088425,
-0.048760898411273956,
-0.07186731696128845,
-0.02230250835418701,
-0.06335557997226715,
-0.16180956363677979,
0.03484545648097992,
0.18353262543678284,
0.025911610573530197,
0.022631127387285233,
0.043480344116687775,
0.012205462902784348,
-0.019035976380109787,
-0.12447699904441833,
-0.024022206664085388,
0.14337247610092163,
0.03298453241586685,
0.08302861452102661,
-0.06327305734157562,
-0.14475615322589874,
-0.11252428591251373,
0.03537563234567642,
0.09875021874904633,
0.192599356174469,
-0.048579856753349304,
0.225503072142601,
0.17551189661026,
-0.06988723576068878,
-0.17886492609977722,
-0.10982055962085724,
0.052014440298080444,
0.0004131861496716738,
0.034583479166030884,
-0.17545360326766968,
0.11376851797103882,
0.06354065239429474,
-0.01139672938734293,
0.015452894382178783,
-0.25275927782058716,
-0.10033151507377625,
-0.019230682402849197,
-0.06372413039207458,
-0.06375783681869507,
-0.1178857684135437,
-0.08323851972818375,
-0.02222631871700287,
-0.009509752504527569,
0.06900417804718018,
0.009947766549885273,
0.07928362488746643,
0.007395660504698753,
0.104375921189785,
0.04068968445062637,
-0.02954559028148651,
0.12419047206640244,
-0.01178416982293129,
0.01354007888585329,
-0.06140255928039551,
-0.003150890115648508,
0.11865386366844177,
-0.06048620492219925,
0.02864832803606987,
0.026782508939504623,
0.01725960522890091,
-0.11429184675216675,
-0.024333354085683823,
-0.1035865843296051,
0.08809457719326019,
-0.08959045261144638,
-0.01390994805842638,
-0.10747354477643967,
0.12378901243209839,
0.06529481709003448,
-0.018577683717012405,
-0.005301634315401316,
-0.02208111062645912,
0.10519630461931229,
0.10636906325817108,
0.08373606950044632,
-0.03171933442354202,
-0.17029014229774475,
-0.03641355782747269,
-0.02806650474667549,
0.04428558424115181,
-0.027688324451446533,
0.036450885236263275,
0.09472344070672989,
0.0001063754316419363,
0.10473524034023285,
-0.020301956683397293,
-0.15070867538452148,
0.01939489133656025,
0.0847841277718544,
-0.111737459897995,
-0.17392636835575104,
-0.04861660301685333,
0.05012679845094681,
-0.10412164032459259,
-0.013148151338100433,
0.1719442754983902,
0.025567851960659027,
-0.06101883575320244,
0.016756940633058548,
0.06450820714235306,
0.01806064136326313,
0.06545887887477875,
-0.002327723428606987,
0.04133489355444908,
-0.0797656700015068,
0.0433865450322628,
0.10119161009788513,
-0.1151621863245964,
0.0159133467823267,
0.16135099530220032,
-0.057647667825222015,
-0.05410105735063553,
-0.07943305373191833,
0.0538443997502327,
-0.03434952720999718,
0.0001752348616719246,
0.05145440995693207,
-0.04266193136572838,
0.036398306488990784,
0.07996129989624023,
0.013798600062727928,
0.051593706011772156,
-0.009304288774728775,
-0.006686018314212561,
-0.0036994884721934795,
0.10783442854881287,
0.02254392020404339,
-0.01964733377099037,
-0.09082454442977905,
0.09843624383211136,
-0.04113125801086426,
0.01759652979671955,
-0.005678360816091299,
-0.05248723924160004,
-0.061435796320438385,
-0.06832851469516754,
-0.10573506355285645,
0.026419073343276978,
-0.11045780032873154,
0.004458743613213301,
-0.016093295067548752,
0.0068422649055719376,
0.022007551044225693,
0.01348180789500475,
-0.049668632447719574,
-0.059611424803733826,
-0.028045514598488808,
0.14405816793441772,
-0.1700294017791748,
0.03891753405332565,
0.08573497831821442,
-0.055973589420318604,
0.10148005932569504,
-0.012710417620837688,
-0.030572395771741867,
0.020590554922819138,
0.04180449992418289,
-0.038300059735774994,
-0.058871496468782425,
0.03649190813302994,
0.02246621996164322,
-0.08933630585670471,
-0.015280145220458508,
-0.03138021379709244,
-0.04494703933596611,
0.01298175472766161,
0.030412845313549042,
-0.08098188042640686,
0.08802878856658936,
0.043434955179691315,
-0.08150406181812286,
-0.052952952682971954,
0.0050160023383796215,
0.08991295099258423,
0.013803843408823013,
0.1082439124584198,
-0.08358852565288544,
0.06852367520332336,
-0.09053102135658264,
-0.02568541094660759,
0.00494147464632988,
0.019061235710978508,
-0.02704872377216816,
-0.04791884124279022,
0.03382980078458786,
0.005672553554177284,
0.06453576683998108,
-0.06390158832073212,
0.03135019168257713,
0.038423627614974976,
-0.08092888444662094,
-0.12211856245994568,
0.018852997571229935,
0.02616942673921585,
-0.005897812079638243,
-0.01595521904528141,
0.02429763786494732,
-0.05628076195716858,
-0.04551544412970543,
0.03746524453163147,
0.14295822381973267,
0.20500634610652924,
0.10303372144699097,
0.018985562026500702,
0.03613388165831566,
-0.017973048612475395,
-0.09128127992153168,
0.05803201347589493,
-0.02694856747984886,
0.010913102887570858,
-0.08071643114089966,
0.04122098535299301,
0.1319994032382965,
-0.17557823657989502,
0.1319626122713089,
0.01052162703126669,
-0.026788007467985153,
-0.026036642491817474,
-0.1081579178571701,
-0.0650775283575058,
-0.014920475892722607,
-0.03226224333047867,
-0.11776559054851532,
0.09334059059619904,
0.07067276537418365,
0.044429730623960495,
-0.06144147738814354,
0.12015452980995178,
-0.16043320298194885,
-0.10342987626791,
0.14736923575401306,
0.03461619094014168,
0.057775817811489105,
0.06287845224142075,
0.012573294341564178,
-0.054218508303165436,
0.10412655770778656,
0.05237171798944473,
0.06831999868154526,
0.0543612539768219,
-0.04725060984492302,
-0.10273532569408417,
-0.06632956117391586,
0.016709599643945694,
-0.029329925775527954,
-0.0018541389144957066,
0.09225019067525864,
0.023761415854096413,
-0.05240977182984352,
-0.025539549067616463,
0.19136209785938263,
-0.03609665483236313,
-0.09154263883829117,
-0.140157550573349,
0.07896427810192108,
0.011067929677665234,
-0.011653129942715168,
0.004932002630084753,
-0.07855139672756195,
-0.01535230502486229,
0.10432060062885284,
0.14069506525993347,
-0.04787714034318924,
0.02359430491924286,
-0.01296023279428482,
0.016985610127449036,
0.0052713025361299515,
0.11055462062358856,
0.007054068148136139,
0.2731833755970001,
-0.016907446086406708,
0.07343318313360214,
0.0005059875547885895,
-0.033652909100055695,
-0.19144223630428314,
0.1202845424413681,
-0.07921405136585236,
-0.006402947008609772,
-0.04861586540937424,
0.14931295812129974,
-0.064664825797081,
-0.27099257707595825,
0.001218757126480341,
-0.10640241205692291,
-0.15674763917922974,
-0.01713830791413784,
0.032089296728372574,
0.037821631878614426,
0.05690884217619896,
0.06775343418121338,
-0.03335900232195854,
0.18664754927158356,
0.005134302191436291,
-0.005205029621720314,
-0.016850780695676804,
0.1372256577014923,
-0.08525433391332626,
0.13998553156852722,
0.034799426794052124,
0.011635582894086838,
0.0885252133011818,
-0.008981376886367798,
-0.08187153935432434,
-0.04616895318031311,
0.06585344672203064,
-0.10592766851186752,
0.02669714204967022,
0.12545078992843628,
0.037898577749729156,
0.03746628016233444,
0.10900665074586868,
-0.0903887003660202,
0.03342457115650177,
0.10856641829013824,
-0.0027796230278909206,
-0.09033510833978653,
0.11852948367595673,
-0.11458665877580643,
0.12204499542713165,
0.15441209077835083,
-0.03332647681236267,
-0.0016326750628650188,
-0.0036908481270074844,
0.01832413673400879,
0.029106922447681427,
0.059581685811281204,
-0.06275403499603271,
-0.1766320914030075,
0.019485648721456528,
-0.12562614679336548,
0.07874248176813126,
-0.11804087460041046,
-0.06682515144348145,
0.06062503159046173,
0.0725913792848587,
-0.026662614196538925,
0.152426615357399,
0.010034050792455673,
-0.03628113120794296,
0.028319507837295532,
-0.0700942650437355,
-0.019770532846450806,
0.12182456254959106,
-0.10556190460920334,
0.011028616689145565
] |
null | null |
transformers
|
# tinyroberta-squad2
This is the *distilled* version of the [deepset/roberta-base-squad2](https://huggingface.co/deepset/roberta-base-squad2) model. This model has a comparable prediction quality and runs at twice the speed of the base model.
## Overview
**Language model:** tinyroberta-squad2
**Language:** English
**Downstream-task:** Extractive QA
**Training data:** SQuAD 2.0
**Eval data:** SQuAD 2.0
**Code:** See [an example QA pipeline on Haystack](https://haystack.deepset.ai/tutorials/first-qa-system)
**Infrastructure**: 4x Tesla v100
## Hyperparameters
```
batch_size = 96
n_epochs = 4
base_LM_model = "deepset/tinyroberta-squad2-step1"
max_seq_len = 384
learning_rate = 3e-5
lr_schedule = LinearWarmup
warmup_proportion = 0.2
doc_stride = 128
max_query_length = 64
distillation_loss_weight = 0.75
temperature = 1.5
teacher = "deepset/robert-large-squad2"
```
## Distillation
This model was distilled using the TinyBERT approach described in [this paper](https://arxiv.org/pdf/1909.10351.pdf) and implemented in [haystack](https://github.com/deepset-ai/haystack).
Firstly, we have performed intermediate layer distillation with roberta-base as the teacher which resulted in [deepset/tinyroberta-6l-768d](https://huggingface.co/deepset/tinyroberta-6l-768d).
Secondly, we have performed task-specific distillation with [deepset/roberta-base-squad2](https://huggingface.co/deepset/roberta-base-squad2) as the teacher for further intermediate layer distillation on an augmented version of SQuADv2 and then with [deepset/roberta-large-squad2](https://huggingface.co/deepset/roberta-large-squad2) as the teacher for prediction layer distillation.
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/tinyroberta-squad2")
# or
reader = TransformersReader(model_name_or_path="deepset/tinyroberta-squad2")
```
### In Transformers
```python
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
model_name = "deepset/tinyroberta-squad2"
# a) Get predictions
nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
QA_input = {
'question': 'Why is model conversion important?',
'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
}
res = nlp(QA_input)
# b) Load model & tokenizer
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
## Performance
Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
```
"exact": 78.69114798281817,
"f1": 81.9198998536977,
"total": 11873,
"HasAns_exact": 76.19770580296895,
"HasAns_f1": 82.66446878592329,
"HasAns_total": 5928,
"NoAns_exact": 81.17746005046257,
"NoAns_f1": 81.17746005046257,
"NoAns_total": 5945
```
## Authors
**Branden Chan:** [email protected]
**Timo Möller:** [email protected]
**Malte Pietsch:** [email protected]
**Tanay Soni:** [email protected]
**Michel Bartels:** [email protected]
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/deepset-logo-colored.png" class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/haystack-logo-colored.png" class="w-40"/>
</div>
</div>
[deepset](http://deepset.ai/) is the company behind the open-source NLP framework [Haystack](https://haystack.deepset.ai/) which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- [roberta-base-squad2]([https://huggingface.co/deepset/roberta-base-squad2)
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="https://github.com/deepset-ai/haystack">GitHub</a></strong> repo and <strong><a href="https://docs.haystack.deepset.ai">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="https://haystack.deepset.ai/community/join">Discord community open to everyone!</a></strong></p>
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "en", "license": "cc-by-4.0", "datasets": ["squad_v2"], "model-index": [{"name": "deepset/tinyroberta-squad2", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 78.8627, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNDNlZDU4ODAxMzY5NGFiMTMyZmQ1M2ZhZjMyODA1NmFlOGMxNzYxNTA4OGE5YTBkZWViZjBkNGQ2ZmMxZjVlMCIsInZlcnNpb24iOjF9.Wgu599r6TvgMLTrHlLMVAbUtKD_3b70iJ5QSeDQ-bRfUsVk6Sz9OsJCp47riHJVlmSYzcDj_z_3jTcUjCFFXBg"}, {"type": "f1", "value": 82.0355, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiOTFkMzEzMWNiZDRhMGZlODhkYzcwZTZiMDFjZDg2YjllZmUzYWM5NTgwNGQ2NGYyMDk2ZGQwN2JmMTE5NTc3YiIsInZlcnNpb24iOjF9.ChgaYpuRHd5WeDFjtiAHUyczxtoOD_M5WR8834jtbf7wXhdGOnZKdZ1KclmhoI5NuAGc1NptX-G0zQ5FTHEcBA"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad", "type": "squad", "config": "plain_text", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 83.86, "name": "Exact Match"}, {"type": "f1", "value": 90.752, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "adversarial_qa", "type": "adversarial_qa", "config": "adversarialQA", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 25.967, "name": "Exact Match"}, {"type": "f1", "value": 37.006, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_adversarial", "type": "squad_adversarial", "config": "AddOneSent", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 76.329, "name": "Exact Match"}, {"type": "f1", "value": 83.292, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts amazon", "type": "squadshifts", "config": "amazon", "split": "test"}, "metrics": [{"type": "exact_match", "value": 63.915, "name": "Exact Match"}, {"type": "f1", "value": 78.395, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts new_wiki", "type": "squadshifts", "config": "new_wiki", "split": "test"}, "metrics": [{"type": "exact_match", "value": 80.297, "name": "Exact Match"}, {"type": "f1", "value": 89.808, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts nyt", "type": "squadshifts", "config": "nyt", "split": "test"}, "metrics": [{"type": "exact_match", "value": 80.149, "name": "Exact Match"}, {"type": "f1", "value": 88.321, "name": "F1"}]}, {"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squadshifts reddit", "type": "squadshifts", "config": "reddit", "split": "test"}, "metrics": [{"type": "exact_match", "value": 66.959, "name": "Exact Match"}, {"type": "f1", "value": 79.3, "name": "F1"}]}]}]}
|
question-answering
|
deepset/tinyroberta-squad2
|
[
"transformers",
"pytorch",
"safetensors",
"roberta",
"question-answering",
"en",
"dataset:squad_v2",
"arxiv:1909.10351",
"license:cc-by-4.0",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1909.10351"
] |
[
"en"
] |
TAGS
#transformers #pytorch #safetensors #roberta #question-answering #en #dataset-squad_v2 #arxiv-1909.10351 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us
|
# tinyroberta-squad2
This is the *distilled* version of the deepset/roberta-base-squad2 model. This model has a comparable prediction quality and runs at twice the speed of the base model.
## Overview
Language model: tinyroberta-squad2
Language: English
Downstream-task: Extractive QA
Training data: SQuAD 2.0
Eval data: SQuAD 2.0
Code: See an example QA pipeline on Haystack
Infrastructure: 4x Tesla v100
## Hyperparameters
## Distillation
This model was distilled using the TinyBERT approach described in this paper and implemented in haystack.
Firstly, we have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.
Secondly, we have performed task-specific distillation with deepset/roberta-base-squad2 as the teacher for further intermediate layer distillation on an augmented version of SQuADv2 and then with deepset/roberta-large-squad2 as the teacher for prediction layer distillation.
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:
### In Transformers
## Performance
Evaluated on the SQuAD 2.0 dev set with the official eval script.
## Authors
Branden Chan: URL@URL
Timo Möller: timo.moeller@URL
Malte Pietsch: malte.pietsch@URL
Tanay Soni: URL@URL
Michel Bartels: michel.bartels@URL
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
</div>
deepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- roberta-base-squad2
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="URL repo and <strong><a href="URL">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="URL community open to everyone!</a></strong></p>
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# tinyroberta-squad2\n\nThis is the *distilled* version of the deepset/roberta-base-squad2 model. This model has a comparable prediction quality and runs at twice the speed of the base model.",
"## Overview\nLanguage model: tinyroberta-squad2 \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Distillation\nThis model was distilled using the TinyBERT approach described in this paper and implemented in haystack.\nFirstly, we have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.\nSecondly, we have performed task-specific distillation with deepset/roberta-base-squad2 as the teacher for further intermediate layer distillation on an augmented version of SQuADv2 and then with deepset/roberta-large-squad2 as the teacher for prediction layer distillation.",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:",
"### In Transformers",
"## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL \nMichel Bartels: michel.bartels@URL",
"## About us\n\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- roberta-base-squad2\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #roberta #question-answering #en #dataset-squad_v2 #arxiv-1909.10351 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n",
"# tinyroberta-squad2\n\nThis is the *distilled* version of the deepset/roberta-base-squad2 model. This model has a comparable prediction quality and runs at twice the speed of the base model.",
"## Overview\nLanguage model: tinyroberta-squad2 \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Distillation\nThis model was distilled using the TinyBERT approach described in this paper and implemented in haystack.\nFirstly, we have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.\nSecondly, we have performed task-specific distillation with deepset/roberta-base-squad2 as the teacher for further intermediate layer distillation on an augmented version of SQuADv2 and then with deepset/roberta-large-squad2 as the teacher for prediction layer distillation.",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:",
"### In Transformers",
"## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL \nMichel Bartels: michel.bartels@URL",
"## About us\n\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- roberta-base-squad2\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
71,
52,
64,
5,
140,
3,
51,
6,
19,
52,
237,
113
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #roberta #question-answering #en #dataset-squad_v2 #arxiv-1909.10351 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n# tinyroberta-squad2\n\nThis is the *distilled* version of the deepset/roberta-base-squad2 model. This model has a comparable prediction quality and runs at twice the speed of the base model.## Overview\nLanguage model: tinyroberta-squad2 \nLanguage: English \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 4x Tesla v100## Hyperparameters## Distillation\nThis model was distilled using the TinyBERT approach described in this paper and implemented in haystack.\nFirstly, we have performed intermediate layer distillation with roberta-base as the teacher which resulted in deepset/tinyroberta-6l-768d.\nSecondly, we have performed task-specific distillation with deepset/roberta-base-squad2 as the teacher for further intermediate layer distillation on an augmented version of SQuADv2 and then with deepset/roberta-large-squad2 as the teacher for prediction layer distillation.## Usage### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:### In Transformers## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL \nMichel Bartels: michel.bartels@URL"
] |
[
-0.0467967726290226,
0.17971713840961456,
-0.007402445189654827,
0.0025237626396119595,
0.08579137176275253,
-0.01839551515877247,
0.10963237285614014,
0.08464913815259933,
-0.048964619636535645,
0.07586397975683212,
0.027227729558944702,
0.06890980154275894,
0.07263292372226715,
0.06942033022642136,
0.07197645306587219,
-0.24681907892227173,
-0.005799501668661833,
-0.12907874584197998,
-0.019146811217069626,
0.07333797961473465,
0.13963091373443604,
-0.08505565673112869,
0.0787268802523613,
-0.004765345249325037,
-0.007197119761258364,
0.02637484110891819,
-0.08276674896478653,
-0.027787843719124794,
0.08147149533033371,
0.04357459768652916,
0.12442898005247116,
0.00953796599060297,
0.10355247557163239,
-0.2011263072490692,
0.024231089279055595,
0.0930606797337532,
0.05170660838484764,
0.04301326349377632,
0.08805229514837265,
-0.008252287283539772,
0.03578772768378258,
-0.10642992705106735,
0.06060212478041649,
0.01523025892674923,
-0.0852314680814743,
-0.11171955615282059,
-0.09785694628953934,
0.08624593168497086,
0.1001012995839119,
-0.007871786132454872,
-0.0018382715061306953,
-0.029495490714907646,
0.014513339847326279,
0.06997168809175491,
0.1500483900308609,
-0.24419242143630981,
-0.07230217754840851,
0.05135474354028702,
-0.008774933405220509,
0.08312313258647919,
-0.06584164500236511,
-0.006729962304234505,
0.010683792643249035,
0.03959931060671806,
0.025013957172632217,
-0.01321646198630333,
0.10081702470779419,
0.009091706946492195,
-0.11741317808628082,
-0.03358788415789604,
0.10556866228580475,
-0.05917007848620415,
-0.10179289430379868,
-0.15299177169799805,
-0.041806239634752274,
0.0818421021103859,
-0.0056398226879537106,
-0.08235932886600494,
0.04478619992733002,
0.01863143965601921,
0.09553550183773041,
-0.059877682477235794,
-0.13148276507854462,
0.03867369145154953,
-0.014049693010747433,
0.05046182870864868,
0.025816526263952255,
0.023473836481571198,
-0.0004563010297715664,
0.10147682577371597,
0.054570313543081284,
-0.10308617353439331,
-0.045226652175188065,
-0.08526190370321274,
-0.039019301533699036,
-0.03997942805290222,
-0.0036323664244264364,
-0.07812472432851791,
0.00043710527825169265,
0.16253581643104553,
0.004049694165587425,
0.048501938581466675,
0.024516409263014793,
0.00932762585580349,
0.07695015519857407,
0.20917969942092896,
-0.0314050167798996,
-0.09574566036462784,
-0.003584192367270589,
0.033064041286706924,
0.019980739802122116,
-0.020978134125471115,
-0.045939523726701736,
-0.030011393129825592,
0.030799027532339096,
0.08137223869562149,
0.04019702970981598,
-0.025919586420059204,
-0.03803882747888565,
-0.04163327440619469,
0.0672309547662735,
-0.18318304419517517,
0.019970819354057312,
-0.0053151450119912624,
-0.0075993165373802185,
0.12775231897830963,
0.032429568469524384,
0.03884520009160042,
-0.05602893605828285,
0.08520708233118057,
0.0006277555366978049,
-0.0008949649054557085,
-0.07765316218137741,
-0.05139267444610596,
0.01293561328202486,
-0.06315542757511139,
-0.052961111068725586,
-0.11646270751953125,
-0.08332248777151108,
-0.06043190136551857,
0.08051669597625732,
-0.027690626680850983,
0.0029049136210232973,
-0.042208705097436905,
-0.014772670343518257,
0.003171289339661598,
0.02748267538845539,
-0.026352085173130035,
0.002048949245363474,
0.020030075684189796,
-0.06315607577562332,
0.03937804326415062,
0.057926103472709656,
0.017789047211408615,
-0.04916448891162872,
0.06486812978982925,
-0.17272698879241943,
0.096684031188488,
-0.052722346037626266,
0.05483737587928772,
-0.12179426103830338,
-0.06468620151281357,
-0.056939855217933655,
0.0013890562113374472,
0.02467167377471924,
0.07060060650110245,
-0.2734888792037964,
-0.0017535692313686013,
0.08807847648859024,
-0.06822744756937027,
-0.06424363702535629,
0.10906992107629776,
-0.024049827829003334,
0.1530742645263672,
0.04634185880422592,
0.12410315126180649,
0.12393844872713089,
-0.1512647569179535,
-0.05729329586029053,
0.044254057109355927,
0.029664751142263412,
0.1826721578836441,
0.08142397552728653,
-0.0706743597984314,
0.161471426486969,
0.0006681885570287704,
-0.02466181479394436,
-0.05261877551674843,
0.011192712001502514,
-0.08345253765583038,
0.0063446564599871635,
-0.006409099325537682,
0.004270351491868496,
-0.02746269293129444,
-0.029540611431002617,
-0.012499467469751835,
-0.08577758818864822,
0.0017869395669549704,
0.06389728933572769,
-0.05025564506649971,
0.05178942531347275,
-0.0944332629442215,
0.0023760029580444098,
0.05168716609477997,
-0.007560490630567074,
-0.1613423377275467,
-0.05278429016470909,
0.04663660004734993,
-0.006497216876596212,
0.05477973446249962,
0.02241058647632599,
0.04012419283390045,
0.0736449807882309,
0.02690509706735611,
-0.03608553856611252,
-0.02339041605591774,
-0.026160145178437233,
-0.038580380380153656,
-0.09712806344032288,
-0.07556437700986862,
-0.05186118185520172,
0.07382312417030334,
-0.08483754843473434,
0.021575385704636574,
-0.012891735881567001,
0.11185309290885925,
0.031223386526107788,
-0.04335744306445122,
-0.02219129540026188,
-0.00041938567301258445,
-0.03738905489444733,
-0.07737240195274353,
0.0373045951128006,
0.01139257475733757,
0.007199821062386036,
-0.010746840387582779,
-0.1069408729672432,
-0.08436348289251328,
-0.0009881934383884072,
0.020505370572209358,
-0.04383423924446106,
-0.024199698120355606,
-0.048663657158613205,
-0.018080798909068108,
-0.034851811826229095,
-0.016600213944911957,
0.20103518664836884,
0.048458829522132874,
0.07901903241872787,
-0.09223800152540207,
-0.06608421355485916,
-0.048450689762830734,
-0.05501094087958336,
-0.06369704753160477,
0.06341052055358887,
-0.03171854838728905,
-0.09456365555524826,
-0.022564854472875595,
0.16065561771392822,
0.08475897461175919,
0.1197625920176506,
-0.00793314166367054,
-0.08205857127904892,
-0.07367927581071854,
0.01583356224000454,
0.015836572274565697,
0.007341769523918629,
0.010047320276498795,
0.020700728520751,
0.02408798411488533,
0.01456573884934187,
0.027771517634391785,
-0.09165298193693161,
0.07478779554367065,
0.033368587493896484,
-0.008832662366330624,
0.08288547396659851,
0.01649061217904091,
0.01510707288980484,
0.03988489508628845,
0.03681638464331627,
0.05317375808954239,
-0.06511090695858002,
-0.06807845830917358,
-0.07552599906921387,
0.12165599316358566,
-0.12264266610145569,
-0.22299666702747345,
-0.19882704317569733,
-0.06338027119636536,
-0.018483709543943405,
-0.008731355890631676,
0.01968012936413288,
-0.04933078959584236,
-0.08789918571710587,
-0.1061432957649231,
0.03294970095157623,
-0.04852201044559479,
-0.06756600737571716,
0.0700463280081749,
0.044967617839574814,
-0.0019576423801481724,
-0.14313624799251556,
-0.00045095139648765326,
0.04678429663181305,
-0.0809788778424263,
-0.02773168496787548,
0.05880928412079811,
0.028956212103366852,
0.09476487338542938,
0.01699896529316902,
-0.015685701742768288,
0.009994634427130222,
0.27033481001853943,
-0.032414842396974564,
0.1280025690793991,
0.18263176083564758,
-0.0721711590886116,
0.0976591631770134,
0.14191606640815735,
0.03452766686677933,
-0.036886248737573624,
0.0349307619035244,
-0.009592844173312187,
-0.002440015785396099,
-0.20089378952980042,
-0.02401270531117916,
0.01286190003156662,
-0.00025282290880568326,
0.06527946889400482,
0.021009238436818123,
-0.09904922544956207,
-0.01334462035447359,
-0.05061767250299454,
0.028432510793209076,
0.047176092863082886,
0.0934782549738884,
0.05808171257376671,
-0.03252730146050453,
0.0515180379152298,
-0.06320441514253616,
0.0021347850561141968,
0.06735305488109589,
0.045024823397397995,
0.13842925429344177,
-0.0010008920216932893,
0.12833160161972046,
0.06769800931215286,
0.1139240562915802,
0.017377346754074097,
0.03676258400082588,
-0.010878988541662693,
0.06076373532414436,
-0.02510509453713894,
-0.09557202458381653,
-0.07063809037208557,
0.060193583369255066,
0.002245074836537242,
-0.02501366287469864,
0.01037605945020914,
0.12634491920471191,
0.05723908543586731,
0.23364244401454926,
-0.006217596121132374,
-0.10291314870119095,
-0.07354838401079178,
0.06988486647605896,
-0.04531877487897873,
-0.09902340173721313,
0.0021852729842066765,
0.07865843921899796,
-0.15098704397678375,
0.020997924730181694,
-0.005462734494358301,
0.07178974896669388,
-0.04744785279035568,
0.004304859321564436,
0.05568872392177582,
0.08488065749406815,
0.011498506180942059,
0.0786418691277504,
-0.12765628099441528,
0.02476731315255165,
0.027392925694584846,
0.0990704745054245,
-0.024187874048948288,
0.074795201420784,
-0.005962818395346403,
0.04509119689464569,
0.10095998644828796,
0.026228411123156548,
-0.08910024166107178,
-0.02550281025469303,
-0.09358633309602737,
-0.014456714503467083,
0.17040012776851654,
-0.09968553483486176,
0.11048098653554916,
-0.06300640851259232,
-0.01086555328220129,
-0.06433770060539246,
-0.0023848609998822212,
-0.09495877474546432,
-0.19043853878974915,
0.06973692029714584,
-0.0858188048005104,
0.000661754806060344,
-0.10232851654291153,
-0.028592750430107117,
-0.055680349469184875,
0.1850767731666565,
-0.2480858862400055,
-0.0648658350110054,
-0.06503643095493317,
-0.06372078508138657,
0.09865153580904007,
-0.06768495589494705,
0.003808398963883519,
-0.040106967091560364,
0.16604448854923248,
-0.02257976122200489,
-0.08332576602697372,
-0.04057660326361656,
-0.08475913852453232,
-0.08409398794174194,
-0.0217638798058033,
0.16234557330608368,
0.010706747882068157,
0.011145450174808502,
0.017249198630452156,
0.014156410470604897,
-0.038471683859825134,
-0.060734543949365616,
-0.03948359191417694,
0.21544481813907623,
0.03393822908401489,
0.07227043807506561,
-0.12701234221458435,
-0.09263265132904053,
-0.09277280420064926,
-0.021896475926041603,
0.09679298102855682,
0.17914074659347534,
-0.07099409401416779,
0.14898985624313354,
0.09661216288805008,
-0.1420554518699646,
-0.19723860919475555,
-0.07670256495475769,
0.08489780128002167,
0.019277874380350113,
0.03678962215781212,
-0.21300819516181946,
0.08843346685171127,
0.026203248649835587,
0.014449009671807289,
0.016474982723593712,
-0.29815229773521423,
-0.1268070787191391,
0.012819808907806873,
-0.007459843065589666,
-0.06987114250659943,
-0.07312548905611038,
-0.037118081003427505,
-0.050948165357112885,
-0.08481308072805405,
-0.018054313957691193,
0.02913861721754074,
0.054140329360961914,
0.030671028420329094,
0.051116254180669785,
0.04592848941683769,
-0.011946093291044235,
0.14618368446826935,
-0.015484333969652653,
0.02781149558722973,
-0.038146570324897766,
0.027805086225271225,
0.0678117424249649,
-0.07267788797616959,
0.06614302843809128,
0.06459058821201324,
0.03650054708123207,
-0.09741631895303726,
-0.018353255465626717,
0.0009520815801806748,
-0.007610775530338287,
-0.1006985753774643,
-0.04221821948885918,
-0.06186056137084961,
0.0994175523519516,
0.07269547879695892,
0.01672239415347576,
0.0077454750426113605,
-0.024399103596806526,
0.015209568664431572,
0.10250459611415863,
0.1373516470193863,
0.000780194706749171,
-0.058350369334220886,
-0.04889439046382904,
0.007138977292925119,
-0.017950527369976044,
-0.08623948693275452,
0.041108667850494385,
0.11446972191333771,
-0.013715559616684914,
0.10022132843732834,
0.005875669419765472,
-0.12606725096702576,
-0.035340581089258194,
0.0817868635058403,
-0.10555438697338104,
-0.17452777922153473,
-0.032725073397159576,
0.10520756989717484,
-0.16490134596824646,
-0.03264959156513214,
0.17663204669952393,
0.019418297335505486,
-0.03355722874403,
0.009499229490756989,
0.0646037682890892,
0.01031534094363451,
0.017467211931943893,
-0.0024532300885766745,
0.021033992990851402,
-0.002843400463461876,
0.13306550681591034,
0.10403795540332794,
0.015842344611883163,
-0.009732291102409363,
0.11220292747020721,
-0.03493664041161537,
-0.03243878483772278,
-0.056686848402023315,
0.08026385307312012,
-0.0007958432543091476,
-0.004580353386700153,
0.021363310515880585,
-0.00657234713435173,
-0.0009745201678015292,
0.031640034168958664,
0.018495159223675728,
0.022528812289237976,
-0.05021006613969803,
-0.008383272215723991,
-0.021705327555537224,
0.09753313660621643,
0.036311060190200806,
0.007623783778399229,
-0.08275343477725983,
0.14791549742221832,
-0.03644053265452385,
0.0372043251991272,
0.006359133869409561,
-0.07723655551671982,
-0.020271431654691696,
-0.020692812278866768,
-0.09665261954069138,
0.04007314518094063,
-0.039576489478349686,
0.004269273951649666,
0.014652034267783165,
0.02944803424179554,
0.0019213546765968204,
0.017515430226922035,
-0.018788866698741913,
-0.03212697431445122,
-0.0634005069732666,
0.08099274337291718,
-0.12201248109340668,
0.012847928330302238,
0.05214887112379074,
-0.08746160566806793,
0.08778855949640274,
0.01605607196688652,
-0.024271249771118164,
0.04042915627360344,
-0.05864470824599266,
0.004784046672284603,
-0.045407455414533615,
0.04923949018120766,
0.010181844234466553,
-0.10410801321268082,
-0.02583342231810093,
-0.02827596105635166,
-0.05467971786856651,
-0.009763707406818867,
0.054719455540180206,
-0.06086571514606476,
0.06939283758401871,
0.04013606533408165,
-0.0422234907746315,
-0.10744724422693253,
0.044985394924879074,
0.015298282727599144,
0.08801154792308807,
0.11206956952810287,
-0.0660925954580307,
0.057608455419540405,
-0.11264676600694656,
-0.00567720178514719,
0.02038559317588806,
0.025952138006687164,
0.012829460203647614,
-0.09960465133190155,
0.008912532590329647,
-0.02036174386739731,
0.05418400838971138,
0.00876599084585905,
0.03924790397286415,
0.027089085429906845,
-0.13436567783355713,
-0.10087481886148453,
0.03598366305232048,
0.041463326662778854,
0.015568399801850319,
-0.0013201464898884296,
-0.033344198018312454,
-0.04059867188334465,
-0.012862998060882092,
0.06175130978226662,
0.09541819989681244,
0.16408300399780273,
0.14302761852741241,
0.05652854964137077,
0.012635079212486744,
-0.035636045038700104,
-0.11376805603504181,
0.06284484267234802,
0.025459857657551765,
0.020959312096238136,
-0.08505134284496307,
0.07847128063440323,
0.13419999182224274,
-0.18205024302005768,
0.05766334384679794,
-0.0576385036110878,
-0.06322693079710007,
-0.11469510197639465,
-0.12511231005191803,
-0.028369948267936707,
0.02221466600894928,
-0.015000964514911175,
-0.1258113980293274,
0.07645581662654877,
-0.03318317234516144,
0.017452256754040718,
-0.02808539755642414,
0.0900835245847702,
-0.13056765496730804,
-0.06966260820627213,
0.08730509877204895,
0.007219433318823576,
0.09007501602172852,
0.05698578804731369,
0.019421013072133064,
0.0008699398022145033,
0.017861666157841682,
0.06297651678323746,
0.049962807446718216,
0.06876827031373978,
-0.034243348985910416,
-0.05592913180589676,
-0.08027977496385574,
0.010621557012200356,
-0.02474934235215187,
-0.0658220499753952,
0.13066188991069794,
0.09594208002090454,
-0.03867601603269577,
-0.0017778893234208226,
0.20865562558174133,
-0.030351825058460236,
-0.12052739411592484,
-0.13298653066158295,
0.02924157865345478,
0.02029336616396904,
0.03869607672095299,
0.019835615530610085,
-0.07749645411968231,
-0.039848167449235916,
0.03927374631166458,
0.09014349430799484,
-0.07457690685987473,
0.0007578209042549133,
0.00008480133692501113,
0.029240010306239128,
-0.04574891924858093,
0.09476164728403091,
0.0906016007065773,
0.18116283416748047,
-0.03428254649043083,
0.04497497156262398,
-0.001346879405900836,
-0.01965702883899212,
-0.1554696261882782,
0.07717305421829224,
-0.08337873965501785,
0.01149873435497284,
-0.034442510455846786,
0.10061183571815491,
-0.049067672342061996,
-0.2648467421531677,
-0.0030422257259488106,
-0.07382531464099884,
-0.1042511835694313,
0.008562047965824604,
0.09408076852560043,
-0.040708187967538834,
0.08045929670333862,
0.008078374899923801,
-0.042243149131536484,
0.22478346526622772,
0.0087131941691041,
0.05175243318080902,
0.03590895235538483,
0.09064839035272598,
-0.1462402641773224,
0.15335924923419952,
0.031827591359615326,
0.02854638174176216,
0.08090376853942871,
0.018216853961348534,
-0.10640095919370651,
-0.011671241372823715,
0.045456115156412125,
-0.013764854520559311,
0.0020382190123200417,
0.13255175948143005,
0.01395056676119566,
0.08342336118221283,
0.09090602397918701,
-0.029523028060793877,
0.04787074029445648,
0.16760185360908508,
-0.03050326369702816,
-0.07038847357034683,
0.052936941385269165,
-0.09860360622406006,
0.10792644321918488,
0.12058129906654358,
-0.015607299283146858,
0.014918913133442402,
-0.02822040766477585,
0.0160802211612463,
0.009400270879268646,
0.0771922767162323,
-0.041999317705631256,
-0.1053248792886734,
-0.008069755509495735,
-0.08590903133153915,
0.07015461474657059,
-0.16992391645908356,
-0.10597651451826096,
0.043064579367637634,
0.015818340703845024,
-0.06074769049882889,
0.14477500319480896,
0.020039426162838936,
0.010741161182522774,
0.0017361582722514868,
-0.041080109775066376,
-0.015230926685035229,
0.11289869248867035,
-0.10983406007289886,
-0.04556224122643471
] |
null | null |
transformers
|
# deepset/xlm-roberta-base-squad2-distilled
- haystack's distillation feature was used for training. deepset/xlm-roberta-large-squad2 was used as the teacher model.
## Overview
**Language model:** deepset/xlm-roberta-base-squad2-distilled
**Language:** Multilingual
**Downstream-task:** Extractive QA
**Training data:** SQuAD 2.0
**Eval data:** SQuAD 2.0
**Code:** See [an example QA pipeline on Haystack](https://haystack.deepset.ai/tutorials/first-qa-system)
**Infrastructure**: 1x Tesla v100
## Hyperparameters
```
batch_size = 56
n_epochs = 4
max_seq_len = 384
learning_rate = 3e-5
lr_schedule = LinearWarmup
embeds_dropout_prob = 0.1
temperature = 3
distillation_loss_weight = 0.75
```
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in [Haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/xlm-roberta-base-squad2-distilled")
# or
reader = TransformersReader(model_name_or_path="deepset/xlm-roberta-base-squad2-distilled",tokenizer="deepset/xlm-roberta-base-squad2-distilled")
```
For a complete example of ``deepset/xlm-roberta-base-squad2-distilled`` being used for [question answering], check out the [Tutorials in Haystack Documentation](https://haystack.deepset.ai/tutorials/first-qa-system)
### In Transformers
```python
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
model_name = "deepset/xlm-roberta-base-squad2-distilled"
# a) Get predictions
nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
QA_input = {
'question': 'Why is model conversion important?',
'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
}
res = nlp(QA_input)
# b) Load model & tokenizer
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
## Performance
Evaluated on the SQuAD 2.0 dev set
```
"exact": 74.06721131980123%
"f1": 76.39919553344667%
```
## Authors
**Timo Möller:** [email protected]
**Julian Risch:** [email protected]
**Malte Pietsch:** [email protected]
**Michel Bartels:** [email protected]
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/deepset-logo-colored.png" class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/haystack-logo-colored.png" class="w-40"/>
</div>
</div>
[deepset](http://deepset.ai/) is the company behind the open-source NLP framework [Haystack](https://haystack.deepset.ai/) which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- [Distilled roberta-base-squad2 (aka "tinyroberta-squad2")]([https://huggingface.co/deepset/tinyroberta-squad2)
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="https://github.com/deepset-ai/haystack">GitHub</a></strong> repo and <strong><a href="https://haystack.deepset.ai">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="https://haystack.deepset.ai/community/join">Discord community open to everyone!</a></strong></p>
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community/join) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "multilingual", "license": "mit", "tags": ["exbert"], "datasets": ["squad_v2"], "thumbnail": "https://thumb.tildacdn.com/tild3433-3637-4830-a533-353833613061/-/resize/720x/-/format/webp/germanquad.jpg"}
|
question-answering
|
deepset/xlm-roberta-base-squad2-distilled
|
[
"transformers",
"pytorch",
"safetensors",
"xlm-roberta",
"question-answering",
"exbert",
"multilingual",
"dataset:squad_v2",
"license:mit",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"multilingual"
] |
TAGS
#transformers #pytorch #safetensors #xlm-roberta #question-answering #exbert #multilingual #dataset-squad_v2 #license-mit #endpoints_compatible #has_space #region-us
|
# deepset/xlm-roberta-base-squad2-distilled
- haystack's distillation feature was used for training. deepset/xlm-roberta-large-squad2 was used as the teacher model.
## Overview
Language model: deepset/xlm-roberta-base-squad2-distilled
Language: Multilingual
Downstream-task: Extractive QA
Training data: SQuAD 2.0
Eval data: SQuAD 2.0
Code: See an example QA pipeline on Haystack
Infrastructure: 1x Tesla v100
## Hyperparameters
## Usage
### In Haystack
Haystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:
For a complete example of ''deepset/xlm-roberta-base-squad2-distilled'' being used for [question answering], check out the Tutorials in Haystack Documentation
### In Transformers
## Performance
Evaluated on the SQuAD 2.0 dev set
## Authors
Timo Möller: timo.moeller@URL
Julian Risch: URL@URL
Malte Pietsch: malte.pietsch@URL
Michel Bartels: michel.bartels@URL
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
</div>
deepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- Distilled roberta-base-squad2 (aka "tinyroberta-squad2")
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="URL repo and <strong><a href="URL">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="URL community open to everyone!</a></strong></p>
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# deepset/xlm-roberta-base-squad2-distilled\n- haystack's distillation feature was used for training. deepset/xlm-roberta-large-squad2 was used as the teacher model.",
"## Overview\nLanguage model: deepset/xlm-roberta-base-squad2-distilled \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 1x Tesla v100",
"## Hyperparameters",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''deepset/xlm-roberta-base-squad2-distilled'' being used for [question answering], check out the Tutorials in Haystack Documentation",
"### In Transformers",
"## Performance\nEvaluated on the SQuAD 2.0 dev set",
"## Authors\nTimo Möller: timo.moeller@URL \nJulian Risch: URL@URL \nMalte Pietsch: malte.pietsch@URL \nMichel Bartels: michel.bartels@URL",
"## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #question-answering #exbert #multilingual #dataset-squad_v2 #license-mit #endpoints_compatible #has_space #region-us \n",
"# deepset/xlm-roberta-base-squad2-distilled\n- haystack's distillation feature was used for training. deepset/xlm-roberta-large-squad2 was used as the teacher model.",
"## Overview\nLanguage model: deepset/xlm-roberta-base-squad2-distilled \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 1x Tesla v100",
"## Hyperparameters",
"## Usage",
"### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''deepset/xlm-roberta-base-squad2-distilled'' being used for [question answering], check out the Tutorials in Haystack Documentation",
"### In Transformers",
"## Performance\nEvaluated on the SQuAD 2.0 dev set",
"## Authors\nTimo Möller: timo.moeller@URL \nJulian Risch: URL@URL \nMalte Pietsch: malte.pietsch@URL \nMichel Bartels: michel.bartels@URL",
"## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
63,
55,
75,
5,
3,
96,
6,
12,
44,
251,
113
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #xlm-roberta #question-answering #exbert #multilingual #dataset-squad_v2 #license-mit #endpoints_compatible #has_space #region-us \n# deepset/xlm-roberta-base-squad2-distilled\n- haystack's distillation feature was used for training. deepset/xlm-roberta-large-squad2 was used as the teacher model.## Overview\nLanguage model: deepset/xlm-roberta-base-squad2-distilled \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 \nCode: See an example QA pipeline on Haystack \nInfrastructure: 1x Tesla v100## Hyperparameters## Usage### In Haystack\nHaystack is an NLP framework by deepset. You can use this model in a Haystack pipeline to do question answering at scale (over many documents). To load the model in Haystack:\n\nFor a complete example of ''deepset/xlm-roberta-base-squad2-distilled'' being used for [question answering], check out the Tutorials in Haystack Documentation### In Transformers## Performance\nEvaluated on the SQuAD 2.0 dev set## Authors\nTimo Möller: timo.moeller@URL \nJulian Risch: URL@URL \nMalte Pietsch: malte.pietsch@URL \nMichel Bartels: michel.bartels@URL"
] |
[
-0.09007550776004791,
0.15748624503612518,
-0.005390307866036892,
0.05118988826870918,
0.06990612298250198,
-0.004353850148618221,
0.1295652687549591,
0.10317058116197586,
-0.022862054407596588,
0.05985884368419647,
0.007957789115607738,
0.07973803579807281,
0.06449896842241287,
0.08973046392202377,
0.034118715673685074,
-0.19136367738246918,
-0.01257617212831974,
-0.0966901108622551,
-0.037295229732990265,
0.1216488853096962,
0.11481143534183502,
-0.09294410794973373,
0.05917017161846161,
0.004952364135533571,
-0.05445515364408493,
0.04263509809970856,
-0.02208717353641987,
-0.04023835062980652,
0.09969373792409897,
0.05344895273447037,
0.07799924165010452,
0.007607654668390751,
0.056669339537620544,
-0.16858533024787903,
0.051205169409513474,
0.09812478721141815,
0.024739956483244896,
0.06021121144294739,
0.1063554659485817,
-0.00824611447751522,
0.015978563576936722,
-0.06653574854135513,
0.061601534485816956,
0.054225921630859375,
-0.06919480115175247,
-0.13324277102947235,
-0.11087284982204437,
0.0706455409526825,
0.1084938645362854,
0.054956790059804916,
-0.012008403427898884,
0.13446877896785736,
-0.04420245438814163,
0.08216317743062973,
0.14083871245384216,
-0.2522665858268738,
-0.08443579077720642,
0.08561240881681442,
0.017483431845903397,
0.05870688706636429,
-0.0971360057592392,
0.04912811145186424,
0.020483748987317085,
0.017105482518672943,
0.017584068700671196,
-0.045671869069337845,
0.0027717272751033306,
-0.011728767305612564,
-0.10104105621576309,
-0.013001121580600739,
0.10924027860164642,
-0.027614418417215347,
-0.05403110012412071,
-0.21112357079982758,
-0.04662222042679787,
0.14304277300834656,
0.0032366758678108454,
-0.08425193279981613,
0.017146948724985123,
-0.01527403760701418,
0.027280524373054504,
-0.09119530022144318,
-0.10168122500181198,
-0.025845345109701157,
-0.014466606080532074,
0.008608432486653328,
0.04235169664025307,
0.002435591071844101,
-0.02140737697482109,
0.10727420449256897,
0.04360170662403107,
-0.1263660192489624,
-0.0297347791492939,
-0.09074539691209793,
-0.1016835942864418,
-0.042017679661512375,
0.0010850956896319985,
-0.0015899543650448322,
0.06976310163736343,
0.18809275329113007,
-0.0215970017015934,
0.04180063307285309,
-0.01260126382112503,
-0.012755485251545906,
0.05381392315030098,
0.10235348343849182,
-0.06476946175098419,
-0.11768213659524918,
0.0276264026761055,
0.025138769298791885,
-0.01585458219051361,
-0.037618573755025864,
-0.03153925761580467,
-0.016141092404723167,
0.030500106513500214,
0.10960561037063599,
0.07907219976186752,
-0.045695509761571884,
-0.028159625828266144,
-0.09752855449914932,
0.05362783744931221,
-0.13315413892269135,
0.05574267730116844,
0.052264656871557236,
-0.023705758154392242,
0.11375325173139572,
-0.052224136888980865,
0.04069024324417114,
-0.07955441623926163,
0.016572600230574608,
-0.024371366947889328,
0.014016957953572273,
-0.07819591462612152,
-0.07595548033714294,
0.014742456376552582,
0.011909624561667442,
-0.05781257525086403,
-0.11666902154684067,
-0.12774589657783508,
-0.08456993848085403,
0.05046890303492546,
-0.038595713675022125,
-0.014250272884964943,
-0.0302644744515419,
-0.03725549206137657,
0.012723056599497795,
0.02752057835459709,
0.062473978847265244,
-0.006895323283970356,
0.02780894562602043,
-0.05542803183197975,
-0.006067928858101368,
0.10548844188451767,
0.006768995895981789,
-0.0503765307366848,
0.008385209366679192,
-0.09748703986406326,
0.11411824077367783,
-0.07367131859064102,
0.027508482336997986,
-0.13658806681632996,
-0.06647272408008575,
-0.0000242466321651591,
0.014312246814370155,
0.0029638351406902075,
0.1405971348285675,
-0.22309453785419464,
-0.015533874742686749,
0.12237431108951569,
-0.07373465597629547,
-0.10887791961431503,
0.128026083111763,
-0.05505422130227089,
0.08723697066307068,
0.07592805474996567,
0.159480482339859,
0.18820345401763916,
-0.13760703802108765,
-0.019646096974611282,
-0.013414791785180569,
0.009859149344265461,
0.11064913868904114,
0.09916217625141144,
-0.04100601375102997,
0.09328348189592361,
-0.02400745265185833,
-0.08756779134273529,
-0.07021939754486084,
-0.01419920101761818,
-0.07757766544818878,
0.016062499955296516,
-0.00043936684960499406,
0.010110125876963139,
-0.03351724520325661,
-0.0018852164503186941,
-0.04730236530303955,
-0.08355025947093964,
0.002003437839448452,
0.08152598142623901,
-0.03488430008292198,
0.004270392004400492,
-0.07932315766811371,
0.029594380408525467,
0.035784028470516205,
0.007398573216050863,
-0.14915291965007782,
-0.12858234345912933,
0.028659475967288017,
-0.07062309235334396,
0.03762122616171837,
0.052315402776002884,
0.014068800024688244,
0.08024869114160538,
0.0034174176398664713,
-0.03242243826389313,
-0.031026825308799744,
-0.03048386424779892,
-0.008722122758626938,
-0.10627265274524689,
-0.04135995730757713,
-0.036392658948898315,
0.045505091547966,
-0.10641654580831528,
0.04261206462979317,
0.0031167948618531227,
0.08194683492183685,
0.037132903933525085,
-0.026580756530165672,
0.005909067578613758,
0.012784861028194427,
-0.0006350874318741262,
-0.04793614149093628,
0.0325268879532814,
0.0001895206223707646,
-0.004870395176112652,
0.0014860283117741346,
-0.0939449667930603,
0.049066219478845596,
0.029537688940763474,
0.021311013028025627,
-0.06484178453683853,
-0.022975005209445953,
0.005496142897754908,
-0.014236986637115479,
-0.10732750594615936,
-0.00354743585921824,
0.1596357375383377,
0.040799785405397415,
0.06751865893602371,
-0.0944000631570816,
-0.07347351312637329,
-0.04773252457380295,
-0.027072757482528687,
0.017314137890934944,
0.10041283816099167,
-0.08720169961452484,
-0.10764212906360626,
0.03518369421362877,
0.11204128712415695,
0.03278474882245064,
0.11492551118135452,
-0.01121279876679182,
-0.05266064032912254,
-0.08479980379343033,
0.027751237154006958,
0.018408529460430145,
0.04290547966957092,
0.015365413390100002,
0.038344576954841614,
0.042480964213609695,
0.05825698375701904,
0.022267648950219154,
-0.09097792208194733,
0.02153398096561432,
0.015157824382185936,
-0.0325516052544117,
0.01727399230003357,
0.0681067630648613,
0.04587497562170029,
0.050900496542453766,
-0.010139841586351395,
0.0915028303861618,
-0.042479876428842545,
-0.07483408600091934,
-0.09725039452314377,
0.18329066038131714,
-0.11540473252534866,
-0.2537830173969269,
-0.12827545404434204,
-0.012594856321811676,
-0.056441113352775574,
-0.0338505394756794,
0.06149830296635628,
-0.017816530540585518,
-0.05251311883330345,
-0.057613153010606766,
0.031168969348073006,
0.005545643623918295,
-0.08234674483537674,
-0.000977734918706119,
0.04718261584639549,
0.01766880787909031,
-0.14916622638702393,
0.020033570006489754,
0.05646279454231262,
-0.09227902442216873,
0.0032961401157081127,
0.07104788720607758,
0.048150092363357544,
0.06398681551218033,
0.04515664651989937,
0.0011393300956115127,
0.002631916431710124,
0.2055465131998062,
-0.084099180996418,
0.04948358237743378,
0.1967879682779312,
0.033115677535533905,
0.06566984951496124,
0.09467080980539322,
0.05946097895503044,
-0.04074979946017265,
0.01690545491874218,
0.040613073855638504,
-0.009923099540174007,
-0.1984580159187317,
-0.04679335281252861,
-0.020296961069107056,
-0.022833503782749176,
0.02334212139248848,
0.03165215998888016,
-0.04625160992145538,
0.0005214569973759353,
-0.05040944367647171,
-0.07789517194032669,
0.06731785833835602,
0.08193900436162949,
0.09950728714466095,
-0.004018617793917656,
0.08630335330963135,
-0.042146001011133194,
0.017942218109965324,
0.08754613250494003,
0.07962887734174728,
0.12647219002246857,
0.006508951541036367,
0.11838462948799133,
0.07433684170246124,
0.10145460814237595,
0.016592785716056824,
0.011035171337425709,
0.022951878607273102,
0.009961891919374466,
0.01065418217331171,
-0.07991326600313187,
0.007906684651970863,
0.033819377422332764,
0.023945311084389687,
0.013916540890932083,
-0.02682463265955448,
0.02897837944328785,
0.0897434875369072,
0.19586263597011566,
0.042454853653907776,
-0.10867327451705933,
-0.07095767557621002,
0.04210025444626808,
-0.06896104663610458,
-0.05614698305726051,
0.019604932516813278,
0.08064741641283035,
-0.20537827908992767,
-0.005428577307611704,
-0.02890453115105629,
0.10957223922014236,
-0.035404253751039505,
0.0018292360473424196,
0.050985731184482574,
0.06823758780956268,
-0.010011190548539162,
0.08592679351568222,
-0.18538108468055725,
0.06548865884542465,
0.01737545058131218,
0.0756206214427948,
-0.03908850997686386,
0.025037895888090134,
0.03563331812620163,
0.025117957964539528,
0.14633509516716003,
0.0011179870925843716,
-0.056379757821559906,
-0.03793494775891304,
-0.11737067997455597,
0.017353493720293045,
0.089704230427742,
-0.14467449486255646,
0.10343378782272339,
-0.035204142332077026,
0.0032361410558223724,
-0.06312605738639832,
0.04059170186519623,
-0.12442131340503693,
-0.14960460364818573,
0.030961064621806145,
-0.07145204395055771,
-0.022121787071228027,
-0.08443403244018555,
-0.06924418359994888,
-0.057387661188840866,
0.16891919076442719,
-0.19457219541072845,
-0.08629871159791946,
-0.09134389460086823,
-0.026556650176644325,
0.06326835602521896,
-0.07435046136379242,
0.010640633292496204,
-0.013278709724545479,
0.1285623162984848,
-0.013157580979168415,
-0.07036720961332321,
0.025389989838004112,
-0.10086438059806824,
-0.11235307902097702,
-0.007589016575366259,
0.1736435741186142,
0.04739472642540932,
0.024259749799966812,
0.039548855274915695,
0.04539455473423004,
-0.058429744094610214,
-0.1065412238240242,
-0.06416496634483337,
0.15914608538150787,
0.07113765925168991,
0.09312047809362411,
-0.053843170404434204,
-0.09716039896011353,
-0.08134254068136215,
0.04074377566576004,
0.1268123835325241,
0.1871347874403,
-0.06967977434396744,
0.15012505650520325,
0.07783672213554382,
-0.07720884680747986,
-0.18457289040088654,
-0.04833828657865524,
0.05702337622642517,
0.027783552184700966,
0.08824145048856735,
-0.16603313386440277,
0.10559210926294327,
0.05076795071363449,
-0.005127341486513615,
0.05989789590239525,
-0.3838258683681488,
-0.10924725234508514,
0.0035658488050103188,
0.001804637722671032,
-0.053215302526950836,
-0.11488451808691025,
-0.053076550364494324,
-0.004117199685424566,
-0.07092402130365372,
0.07057312875986099,
-0.10313116759061813,
0.07217850536108017,
-0.004394657909870148,
0.05222005397081375,
0.017087340354919434,
-0.03777255117893219,
0.14564988017082214,
-0.008885006420314312,
0.045731376856565475,
-0.03664691746234894,
0.05424102023243904,
0.11711317300796509,
-0.0593697763979435,
0.036770835518836975,
-0.00866861455142498,
0.09196759760379791,
-0.14226727187633514,
-0.027375105768442154,
-0.04897906258702278,
0.07587302476167679,
-0.108880914747715,
-0.040179796516895294,
-0.038489654660224915,
0.1064400002360344,
0.06153661012649536,
-0.015066773630678654,
-0.006329798139631748,
-0.037629809230566025,
0.11407218128442764,
0.11923709511756897,
0.11213725060224533,
0.0374838262796402,
-0.14526289701461792,
-0.04719629883766174,
-0.007593738380819559,
0.05219362676143646,
-0.053821854293346405,
0.04641753435134888,
0.09575551748275757,
0.03515935689210892,
0.05874157324433327,
0.005415714345872402,
-0.09750891476869583,
0.022548891603946686,
0.040332939475774765,
-0.13351057469844818,
-0.1836830973625183,
-0.03650633990764618,
0.08313515782356262,
-0.10424964129924774,
-0.0044350773096084595,
0.1699836552143097,
0.02657012641429901,
-0.05827933922410011,
0.03243664652109146,
0.042686112225055695,
-0.0053840274922549725,
0.09759450703859329,
0.007393885403871536,
0.06320226937532425,
-0.052394330501556396,
0.0540623776614666,
0.0364556759595871,
0.00871078111231327,
-0.022524958476424217,
0.1436326950788498,
-0.0776437297463417,
-0.0742836445569992,
-0.02462783455848694,
0.05713511258363724,
-0.03893529996275902,
-0.01215753797441721,
0.006829353515058756,
-0.039826929569244385,
0.004818019922822714,
-0.04569646343588829,
0.060667734593153,
0.007367859594523907,
-0.03384266048669815,
-0.04280382767319679,
0.0038795992732048035,
0.12498226016759872,
0.06551381200551987,
-0.009006661362946033,
-0.10088939219713211,
0.0436432883143425,
-0.01318567804992199,
0.01745709218084812,
-0.003622889518737793,
-0.06418956071138382,
-0.07487016916275024,
0.002112973015755415,
-0.1399678885936737,
0.024599872529506683,
-0.07974138110876083,
0.01895245350897312,
-0.019773805513978004,
0.005889249034225941,
-0.017260026186704636,
0.030516276136040688,
-0.022968927398324013,
-0.05262678116559982,
-0.05970205366611481,
0.10582669824361801,
-0.1618136614561081,
0.040840137749910355,
0.05111739784479141,
-0.08182632178068161,
0.09955338388681412,
0.019597645848989487,
-0.030976062640547752,
0.050217799842357635,
-0.03312983736395836,
-0.04313236102461815,
-0.060833681374788284,
0.06738067418336868,
0.031287066638469696,
-0.19159281253814697,
-0.010383143089711666,
-0.009860430844128132,
-0.026661070063710213,
0.011616963893175125,
0.08080413937568665,
-0.07018091529607773,
0.02455122396349907,
-0.003361229319125414,
-0.052699051797389984,
-0.06834866851568222,
0.030954554677009583,
0.027468649670481682,
0.09277842193841934,
0.15808387100696564,
-0.0699273943901062,
0.1130359023809433,
-0.12129732221364975,
-0.03658207133412361,
0.02812933549284935,
0.005282343365252018,
-0.022129520773887634,
-0.05205661058425903,
0.05797088146209717,
-0.0414130873978138,
0.0757499411702156,
-0.036201946437358856,
0.08449729532003403,
0.030475355684757233,
-0.09063846617937088,
-0.011867236346006393,
0.036570899188518524,
0.12447487562894821,
0.0048355539329349995,
0.024168454110622406,
0.026486976072192192,
-0.007950158789753914,
-0.04015757888555527,
0.011788326315581799,
0.16902165114879608,
0.23070888221263885,
0.036815959960222244,
0.016818273812532425,
0.0363740399479866,
-0.028858715668320656,
-0.1252344250679016,
0.031957123428583145,
-0.002998271957039833,
0.03016633354127407,
-0.0954921618103981,
-0.005120263434946537,
0.1650950312614441,
-0.2084355652332306,
0.0723164901137352,
-0.019724102690815926,
-0.06030258908867836,
-0.1324274241924286,
-0.15398061275482178,
-0.06109406426548958,
-0.01928437128663063,
-0.010301278904080391,
-0.13735662400722504,
0.06393512338399887,
0.006763838231563568,
0.057779572904109955,
-0.01616084761917591,
0.13729290664196014,
-0.06428485363721848,
-0.07574211806058884,
0.07155708968639374,
0.012281195260584354,
0.059649910777807236,
0.058009497821331024,
0.02610776200890541,
-0.00936211459338665,
0.030671922490000725,
0.054009683430194855,
0.022295568138360977,
0.00393735570833087,
-0.015552348457276821,
-0.1310333013534546,
-0.07445209473371506,
-0.00644289655610919,
-0.008428685367107391,
0.007785419933497906,
0.12352987378835678,
0.0953478068113327,
-0.07837311923503876,
-0.01203781645745039,
0.24825094640254974,
-0.056172262877225876,
-0.12457704544067383,
-0.16258203983306885,
0.11044622957706451,
-0.04171469435095787,
0.015168408863246441,
0.01888766698539257,
-0.059478044509887695,
-0.01739550195634365,
0.0738581120967865,
0.21431879699230194,
-0.11581876128911972,
0.02132995054125786,
0.02676183357834816,
0.00604241993278265,
-0.033006083220243454,
0.0933438390493393,
0.07183966040611267,
0.18149210512638092,
-0.03635268285870552,
0.0364319309592247,
-0.011851758696138859,
-0.04708545282483101,
-0.1517743170261383,
0.07098620384931564,
-0.05199052020907402,
-0.022707626223564148,
-0.02827496826648712,
0.18483102321624756,
-0.003049493068829179,
-0.16453412175178528,
-0.0018711034208536148,
-0.05937846750020981,
-0.1489747166633606,
-0.020165719091892242,
0.04600672423839569,
-0.02843319997191429,
0.051038507372140884,
0.002418260322883725,
-0.010118512436747551,
0.14832837879657745,
0.006534747313708067,
0.041566550731658936,
-0.05118125304579735,
0.10926706343889236,
-0.09542975574731827,
0.11786799877882004,
0.024601472541689873,
0.07366533577442169,
0.10784982144832611,
-0.0032482589595019817,
-0.0975605919957161,
0.014697554521262646,
0.08113420009613037,
-0.06650219857692719,
0.019028769806027412,
0.10889400541782379,
0.028002746403217316,
0.03693018853664398,
0.09347894787788391,
-0.0825362280011177,
0.03692933917045593,
0.018778711557388306,
0.012609805911779404,
-0.12729135155677795,
0.0998254045844078,
-0.08084191381931305,
0.1260608732700348,
0.13852831721305847,
-0.053931824862957,
0.022397302091121674,
0.004023891407996416,
0.05606697127223015,
0.02270973101258278,
0.05807769298553467,
-0.03943904861807823,
-0.16953763365745544,
0.020779073238372803,
-0.10774818807840347,
0.054861489683389664,
-0.1645136922597885,
-0.09519906342029572,
0.014920328743755817,
-0.0036407969892024994,
-0.013048729859292507,
0.14611659944057465,
0.025324242189526558,
0.004803918767720461,
0.013357962481677532,
-0.10948225855827332,
-0.03070957213640213,
0.11793743073940277,
-0.09715086221694946,
-0.04887085780501366
] |
null | null |
transformers
|
# Multilingual XLM-RoBERTa base for QA on various languages
## Overview
**Language model:** xlm-roberta-base
**Language:** Multilingual
**Downstream-task:** Extractive QA
**Training data:** SQuAD 2.0
**Eval data:** SQuAD 2.0 dev set - German MLQA - German XQuAD
**Code:** See [example](https://github.com/deepset-ai/FARM/blob/master/examples/question_answering.py) in [FARM](https://github.com/deepset-ai/FARM/blob/master/examples/question_answering.py)
**Infrastructure**: 4x Tesla v100
## Hyperparameters
```
batch_size = 22*4
n_epochs = 2
max_seq_len=256,
doc_stride=128,
learning_rate=2e-5,
```
Corresponding experiment logs in mlflow: [link](https://public-mlflow.deepset.ai/#/experiments/2/runs/b25ec75e07614accb3f1ce03d43dbe08)
## Performance
Evaluated on the SQuAD 2.0 dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
```
"exact": 73.91560683904657
"f1": 77.14103746689592
```
Evaluated on German MLQA: test-context-de-question-de.json
"exact": 33.67279167589108
"f1": 44.34437105434842
"total": 4517
Evaluated on German XQuAD: xquad.de.json
"exact": 48.739495798319325
"f1": 62.552615701071495
"total": 1190
## Usage
### In Transformers
```python
from transformers.pipelines import pipeline
from transformers.modeling_auto import AutoModelForQuestionAnswering
from transformers.tokenization_auto import AutoTokenizer
model_name = "deepset/xlm-roberta-base-squad2"
# a) Get predictions
nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
QA_input = {
'question': 'Why is model conversion important?',
'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
}
res = nlp(QA_input)
# b) Load model & tokenizer
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
### In FARM
```python
from farm.modeling.adaptive_model import AdaptiveModel
from farm.modeling.tokenization import Tokenizer
from farm.infer import Inferencer
model_name = "deepset/xlm-roberta-base-squad2"
# a) Get predictions
nlp = Inferencer.load(model_name, task_type="question_answering")
QA_input = [{"questions": ["Why is model conversion important?"],
"text": "The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks."}]
res = nlp.inference_from_dicts(dicts=QA_input, rest_api_schema=True)
# b) Load model & tokenizer
model = AdaptiveModel.convert_from_transformers(model_name, device="cpu", task_type="question_answering")
tokenizer = Tokenizer.load(model_name)
```
### In haystack
For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in [haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/xlm-roberta-base-squad2")
# or
reader = TransformersReader(model="deepset/roberta-base-squad2",tokenizer="deepset/xlm-roberta-base-squad2")
```
## Authors
Branden Chan: `branden.chan [at] deepset.ai`
Timo Möller: `timo.moeller [at] deepset.ai`
Malte Pietsch: `malte.pietsch [at] deepset.ai`
Tanay Soni: `tanay.soni [at] deepset.ai`
## About us

We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.
Some of our work:
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
- [FARM](https://github.com/deepset-ai/FARM)
- [Haystack](https://github.com/deepset-ai/haystack/)
Get in touch:
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community/join) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"license": "cc-by-4.0", "datasets": ["squad_v2"], "model-index": [{"name": "deepset/xlm-roberta-base-squad2", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 74.0354, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMWMxNWQ2ODJkNWIzZGQwOWI4OTZjYjU3ZDVjZGQzMjI5MzljNjliZTY4Mzk4YTk4OTMzZWYxZjUxYmZhYTBhZSIsInZlcnNpb24iOjF9.eEeFYYJ30BfJDd-JYfI1kjlxJrRF6OFtj2GnkTCOO4kqX31inFy8ptDWusVlLFsUphm4dNWfTKXC5e-gytLBDA"}, {"type": "f1", "value": 77.1833, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiMjg4MjNkOTA4Y2I5OGFlYTk1NWZjMWFlNjI5M2Y0NGZhMThhN2M4YmY2Y2RhZjcwYzU0MGNjN2RkZDljZmJmNiIsInZlcnNpb24iOjF9.TX42YMXpH4e0qu7cC4ARDlZWSkd55dwwyeyFXmOlXERNnEicDuFBCsy8WHLaqQCLUkzODJ22Hw4zhv81rwnlAQ"}]}]}]}
|
question-answering
|
deepset/xlm-roberta-base-squad2
|
[
"transformers",
"pytorch",
"safetensors",
"xlm-roberta",
"question-answering",
"dataset:squad_v2",
"license:cc-by-4.0",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #safetensors #xlm-roberta #question-answering #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us
|
# Multilingual XLM-RoBERTa base for QA on various languages
## Overview
Language model: xlm-roberta-base
Language: Multilingual
Downstream-task: Extractive QA
Training data: SQuAD 2.0
Eval data: SQuAD 2.0 dev set - German MLQA - German XQuAD
Code: See example in FARM
Infrastructure: 4x Tesla v100
## Hyperparameters
Corresponding experiment logs in mlflow: link
## Performance
Evaluated on the SQuAD 2.0 dev set with the official eval script.
Evaluated on German MLQA: URL
"exact": 33.67279167589108
"f1": 44.34437105434842
"total": 4517
Evaluated on German XQuAD: URL
"exact": 48.739495798319325
"f1": 62.552615701071495
"total": 1190
## Usage
### In Transformers
### In FARM
### In haystack
For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:
## Authors
Branden Chan: 'URL [at] URL'
Timo Möller: 'timo.moeller [at] URL'
Malte Pietsch: 'malte.pietsch [at] URL'
Tanay Soni: 'URL [at] URL'
## About us
!deepset logo
We bring NLP to the industry via open source!
Our focus: Industry specific language models & large scale QA systems.
Some of our work:
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
- FARM
- Haystack
Get in touch:
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# Multilingual XLM-RoBERTa base for QA on various languages",
"## Overview\nLanguage model: xlm-roberta-base \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 dev set - German MLQA - German XQuAD \nCode: See example in FARM \nInfrastructure: 4x Tesla v100",
"## Hyperparameters\n\n \n\nCorresponding experiment logs in mlflow: link",
"## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.\n\n\nEvaluated on German MLQA: URL\n \"exact\": 33.67279167589108\n \"f1\": 44.34437105434842\n \"total\": 4517\n\nEvaluated on German XQuAD: URL\n\"exact\": 48.739495798319325\n \"f1\": 62.552615701071495\n \"total\": 1190",
"## Usage",
"### In Transformers",
"### In FARM",
"### In haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:",
"## Authors\nBranden Chan: 'URL [at] URL'\nTimo Möller: 'timo.moeller [at] URL'\nMalte Pietsch: 'malte.pietsch [at] URL'\nTanay Soni: 'URL [at] URL'",
"## About us\n!deepset logo\n\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #question-answering #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n",
"# Multilingual XLM-RoBERTa base for QA on various languages",
"## Overview\nLanguage model: xlm-roberta-base \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 dev set - German MLQA - German XQuAD \nCode: See example in FARM \nInfrastructure: 4x Tesla v100",
"## Hyperparameters\n\n \n\nCorresponding experiment logs in mlflow: link",
"## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.\n\n\nEvaluated on German MLQA: URL\n \"exact\": 33.67279167589108\n \"f1\": 44.34437105434842\n \"total\": 4517\n\nEvaluated on German XQuAD: URL\n\"exact\": 48.739495798319325\n \"f1\": 62.552615701071495\n \"total\": 1190",
"## Usage",
"### In Transformers",
"### In FARM",
"### In haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:",
"## Authors\nBranden Chan: 'URL [at] URL'\nTimo Möller: 'timo.moeller [at] URL'\nMalte Pietsch: 'malte.pietsch [at] URL'\nTanay Soni: 'URL [at] URL'",
"## About us\n!deepset logo\n\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
64,
18,
71,
16,
100,
3,
6,
5,
36,
56,
129
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #xlm-roberta #question-answering #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n# Multilingual XLM-RoBERTa base for QA on various languages## Overview\nLanguage model: xlm-roberta-base \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD 2.0 dev set - German MLQA - German XQuAD \nCode: See example in FARM \nInfrastructure: 4x Tesla v100## Hyperparameters\n\n \n\nCorresponding experiment logs in mlflow: link## Performance\nEvaluated on the SQuAD 2.0 dev set with the official eval script.\n\n\nEvaluated on German MLQA: URL\n \"exact\": 33.67279167589108\n \"f1\": 44.34437105434842\n \"total\": 4517\n\nEvaluated on German XQuAD: URL\n\"exact\": 48.739495798319325\n \"f1\": 62.552615701071495\n \"total\": 1190## Usage### In Transformers### In FARM### In haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:## Authors\nBranden Chan: 'URL [at] URL'\nTimo Möller: 'timo.moeller [at] URL'\nMalte Pietsch: 'malte.pietsch [at] URL'\nTanay Soni: 'URL [at] URL'## About us\n!deepset logo\n\nWe bring NLP to the industry via open source! \nOur focus: Industry specific language models & large scale QA systems. \n \nSome of our work: \n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")\n- FARM\n- Haystack\n\nGet in touch:\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
-0.09268230199813843,
0.17187675833702087,
-0.004642777610570192,
0.029721731320023537,
0.08597300201654434,
0.01675241068005562,
0.12911169230937958,
0.11531096696853638,
0.11814521998167038,
0.10569137334823608,
-0.01100110448896885,
-0.0014545501908287406,
0.09213221073150635,
0.11703595519065857,
0.024400169029831886,
-0.14835837483406067,
0.008906722068786621,
-0.09240910410881042,
-0.0867786705493927,
0.08519618213176727,
0.12947550415992737,
-0.0852988064289093,
0.0997123196721077,
-0.00889850314706564,
-0.05714821442961693,
0.03945362567901611,
-0.0117576839402318,
-0.055292271077632904,
0.10245827585458755,
0.057070400565862656,
0.04839164391160011,
-0.040299031883478165,
0.017018726095557213,
-0.19445176422595978,
0.017227644100785255,
0.055911481380462646,
-0.014596250839531422,
0.04554596170783043,
0.14889836311340332,
0.0009860958671197295,
0.07574842870235443,
-0.08132230490446091,
0.001334317959845066,
0.09958463907241821,
-0.08290552347898483,
-0.1462753564119339,
-0.13364143669605255,
0.14472408592700958,
0.0662682056427002,
0.05238540843129158,
-0.01996801421046257,
0.05621153488755226,
-0.06410044431686401,
0.07447880506515503,
0.10645566135644913,
-0.2144029438495636,
-0.04890977963805199,
0.040166739374399185,
0.06775613129138947,
0.04128072038292885,
-0.1412072777748108,
0.016973121091723442,
0.04116375371813774,
0.014718580059707165,
-0.043014101684093475,
-0.04992931708693504,
0.09750397503376007,
-0.014078100211918354,
-0.07199390977621078,
-0.031323280185461044,
0.09505189210176468,
0.0367390438914299,
-0.06731680780649185,
-0.17796705663204193,
-0.0063222916796803474,
0.055062416940927505,
-0.01216728612780571,
-0.05289354547858238,
0.005621330812573433,
-0.01632397063076496,
0.03470785915851593,
-0.015263067558407784,
-0.08916833251714706,
0.03616880998015404,
0.01453394629061222,
0.08608901500701904,
0.009917636401951313,
-0.01931348815560341,
0.01910502277314663,
0.038456030189991,
-0.013194340281188488,
-0.14650356769561768,
-0.021083805710077286,
-0.07434114813804626,
-0.11132699251174927,
-0.021908210590481758,
0.015045611187815666,
-0.020922835916280746,
0.10029769688844681,
0.13821981847286224,
-0.08431407809257507,
0.0639389157295227,
0.007766133174300194,
-0.02711263671517372,
0.005249131005257368,
0.2168377786874771,
-0.05852099508047104,
-0.25879040360450745,
-0.027054820209741592,
-0.010723943822085857,
-0.04793895408511162,
0.0006661476218141615,
-0.044873956590890884,
-0.03762076795101166,
0.028592664748430252,
0.06724703311920166,
0.06676606833934784,
0.0025808047503232956,
-0.057371705770492554,
-0.07253645360469818,
0.1412857323884964,
-0.13066907227039337,
0.05530782416462898,
0.035863202065229416,
-0.006177166011184454,
0.11462908238172531,
-0.08369669318199158,
0.011872606351971626,
-0.0827568843960762,
0.04641970619559288,
-0.06432010978460312,
-0.01683526672422886,
-0.07324373722076416,
-0.0865415632724762,
0.06537425518035889,
0.026711924001574516,
-0.0463242344558239,
-0.09086637943983078,
-0.026806851848959923,
-0.0804533064365387,
0.049798887223005295,
-0.016133083030581474,
-0.003095619846135378,
0.015140239149332047,
-0.06993797421455383,
0.061214957386255264,
0.019140493124723434,
-0.04638661444187164,
-0.019623877480626106,
0.022600602358579636,
-0.07221003621816635,
0.010367568582296371,
0.06983590871095657,
0.02239733748137951,
-0.06662362068891525,
0.002543576993048191,
-0.21208009123802185,
0.0853426456451416,
-0.1518581360578537,
0.08786965161561966,
-0.15760524570941925,
-0.024309255182743073,
0.10116790235042572,
0.0224382895976305,
0.007870938628911972,
0.11516766995191574,
-0.1642419695854187,
-0.07415345311164856,
0.16555039584636688,
-0.08361184597015381,
-0.09145522117614746,
0.13060343265533447,
-0.015520266257226467,
-0.012948406860232353,
0.10531536489725113,
0.17038540542125702,
0.18276873230934143,
-0.18074309825897217,
-0.08908509463071823,
-0.025593837723135948,
-0.020091531798243523,
0.0824606791138649,
0.08601125329732895,
-0.07867912203073502,
0.06906706839799881,
0.007562325336039066,
-0.04784124717116356,
-0.017370261251926422,
-0.0035976856015622616,
-0.0505366176366806,
0.02708669751882553,
0.0011137612164020538,
0.03720736503601074,
-0.010943102650344372,
-0.0716470256447792,
-0.05429704114794731,
-0.12128933519124985,
-0.02551526203751564,
0.06920067220926285,
-0.010637474246323109,
-0.00023441574012394994,
-0.049557093530893326,
-0.006611433811485767,
0.04687902331352234,
0.01800704374909401,
-0.10702066123485565,
-0.10911780595779419,
0.061724502593278885,
-0.14432552456855774,
0.07814672589302063,
0.022489264607429504,
0.05542847514152527,
0.02561117149889469,
-0.02902216836810112,
-0.0434054434299469,
-0.07496268302202225,
-0.02792711928486824,
-0.02358488366007805,
-0.17277434468269348,
-0.023807745426893234,
-0.054472941905260086,
0.0534672848880291,
-0.04468740150332451,
-0.009363014250993729,
0.09497743099927902,
0.11128407716751099,
0.04483184218406677,
-0.03663673251867294,
-0.01943986304104328,
0.031507883220911026,
-0.002754946704953909,
-0.046874724328517914,
-0.006486817263066769,
-0.02837740257382393,
-0.018643641844391823,
0.05937886983156204,
-0.05119038000702858,
0.07772327214479446,
0.05049598217010498,
0.06280630826950073,
-0.09747710824012756,
-0.04994876682758331,
-0.03217504173517227,
-0.01620892435312271,
-0.04602064564824104,
-0.12299881875514984,
0.1777363419532776,
0.046680331230163574,
0.02797667309641838,
-0.07124924659729004,
-0.06244778260588646,
-0.02440861240029335,
-0.00404682382941246,
-0.03938358277082443,
0.14317268133163452,
-0.05978681892156601,
-0.1253681182861328,
0.11273299157619476,
0.12064047902822495,
0.06631388515233994,
0.1878204047679901,
-0.040505342185497284,
-0.05423798784613609,
-0.0690409243106842,
0.034480031579732895,
-0.0022880109027028084,
0.0936114639043808,
-0.024844102561473846,
0.008089579641819,
0.04764213413000107,
0.013152115046977997,
0.017744572833180428,
-0.04318273812532425,
0.030028462409973145,
-0.019944695755839348,
-0.035266317427158356,
0.037260714918375015,
0.03202631697058678,
0.06187726929783821,
0.09512246400117874,
0.04153541848063469,
0.008304293267428875,
-0.019170790910720825,
-0.05971314013004303,
-0.09936968237161636,
0.18607334792613983,
-0.10420387983322144,
-0.24137701094150543,
-0.165788471698761,
-0.024005189538002014,
-0.08358751237392426,
-0.04450509324669838,
0.03908237814903259,
-0.09174446761608124,
-0.1208673045039177,
-0.015603756532073021,
0.06221075728535652,
0.07289644330739975,
-0.05662474036216736,
0.000002071115659418865,
0.014155020006000996,
0.05682509392499924,
-0.15080228447914124,
-0.021977262571454048,
0.029063820838928223,
-0.028756238520145416,
-0.04303582012653351,
0.08099515736103058,
0.09422207623720169,
0.07650843262672424,
0.0080475565046072,
-0.0030733009334653616,
-0.007203987333923578,
0.2586445212364197,
-0.15449899435043335,
0.10579632222652435,
0.1090930700302124,
0.010610973462462425,
0.07954765111207962,
0.17484959959983826,
0.07238072901964188,
-0.007765407208353281,
-0.007176267448812723,
0.04477095231413841,
0.008940115571022034,
-0.2619846761226654,
-0.08644506335258484,
-0.04714866355061531,
0.027757031843066216,
0.008826540783047676,
0.04591928794980049,
0.0404377281665802,
-0.012352668680250645,
-0.09368621557950974,
-0.029386892914772034,
0.01106172427535057,
0.0882926881313324,
0.060042332857847214,
0.05872785300016403,
0.03911413997411728,
-0.05352769047021866,
-0.015393562614917755,
0.11147971451282501,
0.06642049551010132,
0.12522448599338531,
0.05926981568336487,
0.16674889624118805,
0.0456034280359745,
0.0635584145784378,
-0.0071004340425133705,
0.014653093181550503,
-0.007298070006072521,
0.011346308514475822,
-0.0028359913267195225,
-0.07902971655130386,
0.02262522280216217,
0.08059873431921005,
0.05556900054216385,
-0.038461193442344666,
-0.026220642030239105,
-0.03166328743100166,
0.12895405292510986,
0.17764942348003387,
0.015888754278421402,
-0.09820272773504257,
-0.07939474284648895,
0.018924467265605927,
-0.11532288789749146,
-0.01732836849987507,
0.005850659217685461,
0.06487298011779785,
-0.17813856899738312,
0.07864081859588623,
-0.013587987050414085,
0.10079248249530792,
-0.005350747611373663,
0.015340382233262062,
0.05300699546933174,
0.07645276188850403,
-0.020553193986415863,
0.08283428102731705,
-0.2252793312072754,
0.14017082750797272,
0.015960268676280975,
0.06599317491054535,
-0.04215193912386894,
0.056682974100112915,
0.044012125581502914,
-0.0636754110455513,
0.11982451379299164,
0.013771899975836277,
-0.06391531974077225,
-0.10555679351091385,
-0.07605297118425369,
-0.013519082218408585,
0.09697916358709335,
-0.09492768347263336,
0.1021583154797554,
-0.02626214176416397,
-0.032762203365564346,
-0.050760213285684586,
0.03157096356153488,
-0.1146448478102684,
-0.12084400653839111,
0.06508339196443558,
-0.056381985545158386,
0.04901423677802086,
-0.04636312276124954,
-0.03410090133547783,
-0.12674377858638763,
0.16579093039035797,
-0.19891571998596191,
-0.08573687821626663,
-0.13286371529102325,
-0.0885549858212471,
0.12160495668649673,
-0.13561497628688812,
0.03843100741505623,
-0.015091567300260067,
0.10623378306627274,
-0.03649558126926422,
-0.09371057897806168,
0.04704003408551216,
-0.09210745990276337,
-0.15230050683021545,
-0.009624000638723373,
0.1826416403055191,
0.04576574265956879,
0.04088142514228821,
0.04605450481176376,
-0.019636433571577072,
-0.03939719870686531,
-0.1600368618965149,
0.023931045085191727,
0.04675658419728279,
0.03518309444189072,
0.05112569406628609,
-0.035799119621515274,
-0.19181230664253235,
-0.08398814499378204,
0.01748144067823887,
0.07356111705303192,
0.19048134982585907,
-0.05777531489729881,
0.1351044774055481,
0.16132435202598572,
-0.06544482707977295,
-0.2044624239206314,
-0.028819315135478973,
0.08186459541320801,
0.008861439302563667,
0.03529314324259758,
-0.11631494760513306,
0.1346246898174286,
0.05309514328837395,
-0.020411500707268715,
0.02426513098180294,
-0.207650825381279,
-0.11906655132770538,
0.04002770781517029,
-0.018138788640499115,
-0.1389022171497345,
-0.15260455012321472,
-0.07825953513383865,
-0.02980392798781395,
-0.15626908838748932,
0.07980379462242126,
-0.029729649424552917,
0.048505108803510666,
0.03125159442424774,
0.05836646631360054,
0.03267792984843254,
-0.0585995651781559,
0.1272546648979187,
0.08479923009872437,
0.026913205161690712,
-0.03521755337715149,
-0.07097984105348587,
0.06680499762296677,
-0.03023197315633297,
0.05810748413205147,
-0.023906361311674118,
0.021601518616080284,
-0.13892503082752228,
-0.011833935976028442,
-0.08298464119434357,
0.11459552496671677,
-0.08823566138744354,
-0.05547603592276573,
-0.08648189157247543,
0.1154056265950203,
0.07116633653640747,
-0.010210460051894188,
0.05491086468100548,
-0.030579710379242897,
0.06893198192119598,
0.21852613985538483,
0.1278429478406906,
0.015137499198317528,
-0.1362886130809784,
-0.016844242811203003,
-0.029605749994516373,
0.047063108533620834,
-0.06745852530002594,
0.06846515834331512,
0.1259453296661377,
0.013736899010837078,
0.1262274533510208,
-0.038355033844709396,
-0.10720301419496536,
0.006009317934513092,
0.08726345002651215,
-0.09499552100896835,
-0.1666574478149414,
-0.07777220010757446,
-0.02016264572739601,
-0.08836454153060913,
0.026936516165733337,
0.17381542921066284,
0.010434763506054878,
-0.024266883730888367,
0.014951582998037338,
0.06009577587246895,
-0.0008920918917283416,
0.13289420306682587,
0.04689371958374977,
0.06900706887245178,
-0.08638244867324829,
0.05727243050932884,
0.06068948283791542,
-0.08348458260297775,
0.025396324694156647,
0.15103517472743988,
-0.06088960915803909,
-0.030408041551709175,
-0.04483191296458244,
0.1092044785618782,
-0.1331692934036255,
-0.04456353932619095,
-0.00043297771480865777,
-0.09061574190855026,
0.01729513518512249,
0.041346944868564606,
0.03924499452114105,
0.015504859387874603,
0.021330025047063828,
0.0002637709549162537,
-0.004911051131784916,
0.13609899580478668,
0.07634534686803818,
0.01240380760282278,
-0.09608778357505798,
0.09656406939029694,
-0.04162180423736572,
0.012636963278055191,
-0.00634800037369132,
-0.012491986155509949,
-0.12560498714447021,
-0.05627547577023506,
-0.13915590941905975,
0.02256046235561371,
-0.07165838032960892,
0.023877589032053947,
-0.03827493265271187,
-0.06763888895511627,
0.0018570702522993088,
0.02817612886428833,
-0.06929104030132294,
-0.06261901557445526,
-0.017655158415436745,
0.13235880434513092,
-0.1618848592042923,
0.034308791160583496,
0.07697071880102158,
-0.04432414099574089,
0.09291048347949982,
0.0050612580962479115,
-0.008710898458957672,
0.02796865627169609,
-0.08133471757173538,
0.0039810920134186745,
-0.036755625158548355,
0.027859872207045555,
0.045975957065820694,
-0.11023376137018204,
0.004239643923938274,
-0.022871755063533783,
-0.03387666866183281,
0.015424409881234169,
-0.025965707376599312,
-0.09948984533548355,
0.08797693997621536,
-0.0038434588350355625,
-0.11071468144655228,
-0.057454969733953476,
0.05005352944135666,
0.1334301382303238,
-0.015951506793498993,
0.12766410410404205,
-0.07636775821447372,
0.0702613964676857,
-0.13188321888446808,
-0.010173172689974308,
0.03590060770511627,
-0.017601678147912025,
-0.07396580278873444,
-0.007521117106080055,
0.05288328230381012,
-0.019987760111689568,
0.0879388377070427,
-0.023770296946167946,
0.06576406955718994,
0.045504141598939896,
-0.04553002864122391,
-0.07235710322856903,
0.04245229437947273,
0.055399369448423386,
-0.029464242979884148,
0.020335165783762932,
-0.0027929439675062895,
-0.043575219810009,
-0.03412468358874321,
-0.03761076554656029,
0.1418062299489975,
0.21676219999790192,
0.04640771076083183,
0.00013870828843209893,
0.058914441615343094,
-0.06694689393043518,
-0.16431288421154022,
0.04090050607919693,
-0.054018374532461166,
0.051430899649858475,
-0.07588289678096771,
0.051482878625392914,
0.14250808954238892,
-0.21129359304904938,
0.09298203140497208,
0.004666260443627834,
-0.027358978986740112,
-0.028394171968102455,
-0.06338318437337875,
-0.06824968010187149,
-0.060144614428281784,
-0.004235452972352505,
-0.1295163482427597,
0.08810216188430786,
0.04103711619973183,
0.038702454417943954,
-0.07417433708906174,
0.12630556523799896,
-0.10913514345884323,
-0.09652064740657806,
0.13176380097866058,
0.053469885140657425,
0.046632591634988785,
0.06058923527598381,
0.010413190349936485,
-0.06263162940740585,
0.06678499281406403,
0.053775355219841,
0.06850237399339676,
-0.02246129885315895,
-0.035590238869190216,
-0.08219192922115326,
-0.07110721617937088,
0.011865086853504181,
-0.019798479974269867,
-0.020449550822377205,
0.09245835989713669,
0.02956710197031498,
-0.015579772181808949,
-0.008730418980121613,
0.14312098920345306,
-0.0565655417740345,
-0.08698461949825287,
-0.13530713319778442,
0.06500044465065002,
-0.008134695701301098,
0.05568750947713852,
0.011444478295743465,
-0.10781354457139969,
-0.023520609363913536,
0.1367192268371582,
0.1361040472984314,
-0.06137442961335182,
0.011787576600909233,
-0.016774270683526993,
0.013696154579520226,
0.012764844112098217,
0.1250743418931961,
-0.006584441289305687,
0.2880626618862152,
-0.00836837850511074,
0.07622259855270386,
0.00014154458767734468,
-0.038680799305438995,
-0.09488514065742493,
0.14858809113502502,
-0.03963961824774742,
-0.052291255444288254,
-0.065616674721241,
0.16073840856552124,
-0.032458849251270294,
-0.22303807735443115,
-0.04787712171673775,
-0.11098123341798782,
-0.15156641602516174,
-0.023291757330298424,
0.07858185470104218,
0.05769428610801697,
0.06322545558214188,
0.06042205169796944,
-0.0249544195830822,
0.16042394936084747,
0.015581469982862473,
-0.03738727420568466,
-0.03288406878709793,
0.07744765281677246,
-0.09690498560667038,
0.1855999380350113,
0.020247649401426315,
0.0449397899210453,
0.121589794754982,
-0.011102977208793163,
-0.09673153609037399,
-0.035913947969675064,
0.10159168392419815,
-0.16884711384773254,
0.04839393123984337,
0.07040572911500931,
0.01694818213582039,
0.09018711745738983,
0.0783303827047348,
-0.05712990090250969,
0.023045899346470833,
0.08893711864948273,
0.014139337465167046,
-0.11602666229009628,
0.10969152301549911,
-0.09143684804439545,
0.14069251716136932,
0.17002269625663757,
-0.037146199494600296,
0.0021427033934742212,
-0.018078207969665527,
0.050824061036109924,
0.016771981492638588,
0.0695251077413559,
-0.04404384270310402,
-0.19530455768108368,
0.037232186645269394,
-0.05792902037501335,
0.06530584394931793,
-0.08168748021125793,
-0.06648403406143188,
0.0163807924836874,
0.06905250251293182,
-0.0549718402326107,
0.1497250348329544,
0.03257543221116066,
-0.03837145119905472,
0.02527792751789093,
-0.07151385396718979,
-0.012508147396147251,
0.13443972170352936,
-0.0732443630695343,
-0.022405628114938736
] |
null | null |
transformers
|
# Multilingual XLM-RoBERTa large for QA on various languages
## Overview
**Language model:** xlm-roberta-large
**Language:** Multilingual
**Downstream-task:** Extractive QA
**Training data:** SQuAD 2.0
**Eval data:** SQuAD dev set - German MLQA - German XQuAD
**Training run:** [MLFlow link](https://public-mlflow.deepset.ai/#/experiments/124/runs/3a540e3f3ecf4dd98eae8fc6d457ff20)
**Infrastructure**: 4x Tesla v100
## Hyperparameters
```
batch_size = 32
n_epochs = 3
base_LM_model = "xlm-roberta-large"
max_seq_len = 256
learning_rate = 1e-5
lr_schedule = LinearWarmup
warmup_proportion = 0.2
doc_stride=128
max_query_length=64
```
## Performance
Evaluated on the SQuAD 2.0 English dev set with the [official eval script](https://worksheets.codalab.org/rest/bundles/0x6b567e1cf2e041ec80d7098f031c5c9e/contents/blob/).
```
"exact": 79.45759285774446,
"f1": 83.79259828925511,
"total": 11873,
"HasAns_exact": 71.96356275303644,
"HasAns_f1": 80.6460053117963,
"HasAns_total": 5928,
"NoAns_exact": 86.93019343986543,
"NoAns_f1": 86.93019343986543,
"NoAns_total": 5945
```
Evaluated on German [MLQA: test-context-de-question-de.json](https://github.com/facebookresearch/MLQA)
```
"exact": 49.34691166703564,
"f1": 66.15582561674236,
"total": 4517,
```
Evaluated on German [XQuAD: xquad.de.json](https://github.com/deepmind/xquad)
```
"exact": 61.51260504201681,
"f1": 78.80206098332569,
"total": 1190,
```
## Usage
### In Haystack
For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in [haystack](https://github.com/deepset-ai/haystack/):
```python
reader = FARMReader(model_name_or_path="deepset/xlm-roberta-large-squad2")
# or
reader = TransformersReader(model="deepset/xlm-roberta-large-squad2",tokenizer="deepset/xlm-roberta-large-squad2")
```
### In Transformers
```python
from transformers import AutoModelForQuestionAnswering, AutoTokenizer, pipeline
model_name = "deepset/xlm-roberta-large-squad2"
# a) Get predictions
nlp = pipeline('question-answering', model=model_name, tokenizer=model_name)
QA_input = {
'question': 'Why is model conversion important?',
'context': 'The option to convert models between FARM and transformers gives freedom to the user and let people easily switch between frameworks.'
}
res = nlp(QA_input)
# b) Load model & tokenizer
model = AutoModelForQuestionAnswering.from_pretrained(model_name)
tokenizer = AutoTokenizer.from_pretrained(model_name)
```
## Authors
**Branden Chan:** [email protected]
**Timo Möller:** [email protected]
**Malte Pietsch:** [email protected]
**Tanay Soni:** [email protected]
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/deepset-logo-colored.png" class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="https://raw.githubusercontent.com/deepset-ai/.github/main/haystack-logo-colored.png" class="w-40"/>
</div>
</div>
[deepset](http://deepset.ai/) is the company behind the open-source NLP framework [Haystack](https://haystack.deepset.ai/) which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- [Distilled roberta-base-squad2 (aka "tinyroberta-squad2")]([https://huggingface.co/deepset/tinyroberta-squad2)
- [German BERT (aka "bert-base-german-cased")](https://deepset.ai/german-bert)
- [GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")](https://deepset.ai/germanquad)
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="https://github.com/deepset-ai/haystack">GitHub</a></strong> repo and <strong><a href="https://docs.haystack.deepset.ai">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="https://haystack.deepset.ai/community">Discord community open to everyone!</a></strong></p>
[Twitter](https://twitter.com/deepset_ai) | [LinkedIn](https://www.linkedin.com/company/deepset-ai/) | [Discord](https://haystack.deepset.ai/community) | [GitHub Discussions](https://github.com/deepset-ai/haystack/discussions) | [Website](https://deepset.ai)
By the way: [we're hiring!](http://www.deepset.ai/jobs)
|
{"language": "multilingual", "license": "cc-by-4.0", "tags": ["question-answering"], "datasets": ["squad_v2"], "model-index": [{"name": "deepset/xlm-roberta-large-squad2", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad_v2", "type": "squad_v2", "config": "squad_v2", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 81.8281, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNzVhZDE2NTg5NmUwOWRkMmI2MGUxYjFlZjIzNmMyNDQ2MDY2MDNhYzE0ZjY5YTkyY2U4ODc3ODFiZjQxZWQ2YSIsInZlcnNpb24iOjF9.f_rN3WPMAdv-OBPz0T7N7lOxYz9f1nEr_P-vwKhi3jNdRKp_JTy18MYR9eyJM2riKHC6_ge-8XwfyrUf51DSDA"}, {"type": "f1", "value": 84.8886, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiZGE5MWJmZGUxMGMwNWFhYzVhZjQwZGEwOWQ4N2Q2Yjg5NzdjNDFiNDhiYTQ1Y2E5ZWJkOTFhYmI1Y2Q2ZGYwOCIsInZlcnNpb24iOjF9.TIdH-tOx3kEMDs5wK1r6iwZqqSjNGlBrpawrsE917j1F3UFJVnQ7wJwaj0OIgmC4iw8OQeLZL56ucBcLApa-AQ"}]}]}]}
|
question-answering
|
deepset/xlm-roberta-large-squad2
|
[
"transformers",
"pytorch",
"safetensors",
"xlm-roberta",
"question-answering",
"multilingual",
"dataset:squad_v2",
"license:cc-by-4.0",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"multilingual"
] |
TAGS
#transformers #pytorch #safetensors #xlm-roberta #question-answering #multilingual #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us
|
# Multilingual XLM-RoBERTa large for QA on various languages
## Overview
Language model: xlm-roberta-large
Language: Multilingual
Downstream-task: Extractive QA
Training data: SQuAD 2.0
Eval data: SQuAD dev set - German MLQA - German XQuAD
Training run: MLFlow link
Infrastructure: 4x Tesla v100
## Hyperparameters
## Performance
Evaluated on the SQuAD 2.0 English dev set with the official eval script.
Evaluated on German MLQA: URL
Evaluated on German XQuAD: URL
## Usage
### In Haystack
For doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:
### In Transformers
## Authors
Branden Chan: URL@URL
Timo Möller: timo.moeller@URL
Malte Pietsch: malte.pietsch@URL
Tanay Soni: URL@URL
## About us
<div class="grid lg:grid-cols-2 gap-x-4 gap-y-3">
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
<div class="w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center">
<img alt="" src="URL class="w-40"/>
</div>
</div>
deepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.
Some of our other work:
- Distilled roberta-base-squad2 (aka "tinyroberta-squad2")
- German BERT (aka "bert-base-german-cased")
- GermanQuAD and GermanDPR datasets and models (aka "gelectra-base-germanquad", "gbert-base-germandpr")
## Get in touch and join the Haystack community
<p>For more info on Haystack, visit our <strong><a href="URL repo and <strong><a href="URL">Documentation</a></strong>.
We also have a <strong><a class="h-7" href="URL community open to everyone!</a></strong></p>
Twitter | LinkedIn | Discord | GitHub Discussions | Website
By the way: we're hiring!
|
[
"# Multilingual XLM-RoBERTa large for QA on various languages",
"## Overview\nLanguage model: xlm-roberta-large \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD dev set - German MLQA - German XQuAD \nTraining run: MLFlow link \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Performance\nEvaluated on the SQuAD 2.0 English dev set with the official eval script.\n\n\nEvaluated on German MLQA: URL\n\n\nEvaluated on German XQuAD: URL",
"## Usage",
"### In Haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:",
"### In Transformers",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL",
"## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
"TAGS\n#transformers #pytorch #safetensors #xlm-roberta #question-answering #multilingual #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n",
"# Multilingual XLM-RoBERTa large for QA on various languages",
"## Overview\nLanguage model: xlm-roberta-large \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD dev set - German MLQA - German XQuAD \nTraining run: MLFlow link \nInfrastructure: 4x Tesla v100",
"## Hyperparameters",
"## Performance\nEvaluated on the SQuAD 2.0 English dev set with the official eval script.\n\n\nEvaluated on German MLQA: URL\n\n\nEvaluated on German XQuAD: URL",
"## Usage",
"### In Haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:",
"### In Transformers",
"## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL",
"## About us\n<div class=\"grid lg:grid-cols-2 gap-x-4 gap-y-3\">\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n <div class=\"w-full h-40 object-cover mb-2 rounded-lg flex items-center justify-center\">\n <img alt=\"\" src=\"URL class=\"w-40\"/>\n </div>\n</div>\n\ndeepset is the company behind the open-source NLP framework Haystack which is designed to help you build production ready NLP systems that use: Question answering, summarization, ranking etc.\n\n\nSome of our other work: \n- Distilled roberta-base-squad2 (aka \"tinyroberta-squad2\")\n- German BERT (aka \"bert-base-german-cased\")\n- GermanQuAD and GermanDPR datasets and models (aka \"gelectra-base-germanquad\", \"gbert-base-germandpr\")",
"## Get in touch and join the Haystack community\n\n<p>For more info on Haystack, visit our <strong><a href=\"URL repo and <strong><a href=\"URL\">Documentation</a></strong>. \n\nWe also have a <strong><a class=\"h-7\" href=\"URL community open to everyone!</a></strong></p>\n\nTwitter | LinkedIn | Discord | GitHub Discussions | Website\n\nBy the way: we're hiring!"
] |
[
68,
18,
71,
5,
38,
3,
36,
6,
41,
251,
113
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #xlm-roberta #question-answering #multilingual #dataset-squad_v2 #license-cc-by-4.0 #model-index #endpoints_compatible #has_space #region-us \n# Multilingual XLM-RoBERTa large for QA on various languages## Overview\nLanguage model: xlm-roberta-large \nLanguage: Multilingual \nDownstream-task: Extractive QA \nTraining data: SQuAD 2.0 \nEval data: SQuAD dev set - German MLQA - German XQuAD \nTraining run: MLFlow link \nInfrastructure: 4x Tesla v100## Hyperparameters## Performance\nEvaluated on the SQuAD 2.0 English dev set with the official eval script.\n\n\nEvaluated on German MLQA: URL\n\n\nEvaluated on German XQuAD: URL## Usage### In Haystack\nFor doing QA at scale (i.e. many docs instead of single paragraph), you can load the model also in haystack:### In Transformers## Authors\nBranden Chan: URL@URL \nTimo Möller: timo.moeller@URL \nMalte Pietsch: malte.pietsch@URL \nTanay Soni: URL@URL"
] |
[
-0.1440849006175995,
0.09156963974237442,
-0.0048439353704452515,
0.0689658522605896,
0.09440144151449203,
0.03993718698620796,
0.278011292219162,
0.05753927305340767,
0.043132536113262177,
0.02494443580508232,
-0.011029749177396297,
-0.04307403042912483,
0.06817503273487091,
0.1443937122821808,
0.013186884112656116,
-0.22431021928787231,
-0.0180088821798563,
-0.06792774051427841,
-0.08313024789094925,
0.06469219923019409,
0.11744454503059387,
-0.09432331472635269,
0.06170656904578209,
0.0007331154774874449,
-0.027963576838374138,
0.08823967725038528,
-0.03797610476613045,
-0.05135025456547737,
0.11849776655435562,
0.04233795776963234,
0.10678060352802277,
0.034098077565431595,
0.07472110539674759,
-0.13376963138580322,
0.02426976151764393,
0.022345485165715218,
-0.014333831146359444,
0.008999479003250599,
0.12102881073951721,
0.03295606002211571,
0.04488198086619377,
-0.061052355915308,
0.007851095870137215,
0.04340524971485138,
-0.015036115422844887,
-0.12922602891921997,
-0.0747767835855484,
0.08316429704427719,
0.008865011855959892,
0.08564923703670502,
-0.008617944084107876,
0.09645663946866989,
-0.11278580874204636,
0.09618499130010605,
0.0998321995139122,
-0.27802395820617676,
-0.07533591240644455,
0.12821128964424133,
0.04314534366130829,
0.08361395448446274,
-0.09718256443738937,
-0.0138930082321167,
0.04651277884840965,
0.025833718478679657,
-0.040223222225904465,
-0.049721769988536835,
0.04673108085989952,
-0.027723077684640884,
-0.07982339709997177,
-0.005092163570225239,
0.19192658364772797,
0.015542848967015743,
-0.07380852103233337,
-0.14053143560886383,
-0.03935043886303902,
0.1608833372592926,
-0.014327039942145348,
0.011262518353760242,
0.015749963000416756,
-0.00799154955893755,
-0.013118785806000233,
-0.05580722913146019,
-0.11991246789693832,
0.009437354281544685,
-0.04123423621058464,
0.035206541419029236,
0.058781079947948456,
0.005211980082094669,
-0.0061601740308105946,
0.06100734695792198,
-0.004244862124323845,
-0.1327369511127472,
-0.010441062040627003,
-0.09904631972312927,
-0.13211756944656372,
-0.011671371757984161,
-0.0022050642874091864,
-0.05750613659620285,
0.09426380693912506,
0.1014316976070404,
-0.03578932583332062,
0.03637168928980827,
0.030668651685118675,
-0.005558827426284552,
0.00404571695253253,
0.21863727271556854,
-0.061934396624565125,
-0.1812605857849121,
-0.012798866257071495,
-0.00325008574873209,
-0.02915837988257408,
0.019822552800178528,
-0.06680932641029358,
-0.04487820714712143,
-0.036079712212085724,
0.07795606553554535,
-0.03480743244290352,
0.023518847301602364,
-0.021182343363761902,
-0.05229237303137779,
0.06967232376337051,
-0.10221370309591293,
0.03622858598828316,
0.027958203107118607,
-0.017484787851572037,
0.2091037631034851,
-0.10243556648492813,
0.014274945482611656,
-0.0888078585267067,
0.06913809478282928,
-0.01984187215566635,
0.0013169875601306558,
-0.06670944392681122,
-0.11500528454780579,
0.030924221500754356,
-0.00470203161239624,
-0.020977040752768517,
-0.12030750513076782,
-0.12274091690778732,
-0.05043771117925644,
0.04284614697098732,
-0.012141975574195385,
-0.005563698709011078,
-0.030021673068404198,
-0.06088627129793167,
0.019356591627001762,
-0.0049103000201284885,
0.029705164954066277,
-0.04758463054895401,
0.0056579518131911755,
-0.06697769463062286,
0.0032617219258099794,
0.029972568154335022,
-0.0028135417960584164,
-0.060639653354883194,
-0.0489896759390831,
-0.12804871797561646,
0.09966784715652466,
-0.11700993776321411,
0.038073956966400146,
-0.11448902636766434,
-0.06844151020050049,
0.0506432019174099,
0.046056296676397324,
0.02485884167253971,
0.14436137676239014,
-0.18416373431682587,
-0.04386984184384346,
0.18079784512519836,
-0.07790698111057281,
-0.10822710394859314,
0.1669161170721054,
-0.022892003878951073,
-0.0041484711691737175,
0.08978470414876938,
0.1372811198234558,
0.1431313455104828,
-0.24518577754497528,
-0.09702832996845245,
0.04349170997738838,
0.04266621917486191,
0.022252757102251053,
0.11593084782361984,
0.014408469200134277,
0.12043692171573639,
-0.0013869578251615167,
-0.0072733694687485695,
-0.03598051518201828,
-0.015065883286297321,
-0.04898010939359665,
-0.002030373318120837,
-0.01343369297683239,
0.003900104435160756,
-0.03800983354449272,
0.01191787887364626,
-0.050036948174238205,
-0.07785957306623459,
-0.05471932515501976,
0.11222977191209793,
-0.05118270963430405,
-0.002880532294511795,
-0.04215565696358681,
-0.05977186560630798,
-0.01653207838535309,
0.0222419872879982,
-0.10515163838863373,
-0.1754680871963501,
0.0697854682803154,
-0.04012880101799965,
0.07496368139982224,
0.15936212241649628,
0.043699849396944046,
0.06803999096155167,
-0.07495962083339691,
-0.07558754086494446,
-0.09459121525287628,
-0.06166969612240791,
-0.026255790144205093,
-0.12018045037984848,
-0.01243635918945074,
-0.0808120146393776,
-0.03183170408010483,
-0.0724940076470375,
0.01863653026521206,
0.07446707040071487,
0.04830987751483917,
0.05859832465648651,
0.03746379539370537,
-0.020605413243174553,
0.047810621559619904,
-0.036313485354185104,
-0.04830470681190491,
-0.0018609781982377172,
0.016981489956378937,
-0.07398338615894318,
0.05696699768304825,
0.028220713138580322,
0.12379040569067001,
0.07646483182907104,
0.06815627962350845,
-0.016918042674660683,
-0.08164676278829575,
-0.047977034002542496,
0.003850249806419015,
-0.09694230556488037,
-0.08924949169158936,
0.19017784297466278,
0.056810684502124786,
0.08474848419427872,
-0.10958022624254227,
-0.0506785586476326,
-0.032407328486442566,
-0.004714765585958958,
-0.02792220376431942,
0.15182842314243317,
0.004109816160053015,
-0.02856946550309658,
0.07542484253644943,
0.12524569034576416,
-0.03796917200088501,
0.17892059683799744,
-0.038183651864528656,
-0.05958833917975426,
-0.017508551478385925,
0.03545942157506943,
0.011425168253481388,
0.08963067084550858,
-0.012538423761725426,
0.056279271841049194,
0.04765530303120613,
0.06629865616559982,
0.02560211345553398,
-0.09833177179098129,
0.0055017792619764805,
-0.03723989427089691,
-0.058568328619003296,
-0.05383312329649925,
0.05250680074095726,
0.06299791485071182,
0.10034949332475662,
-0.05475982278585434,
0.011356674134731293,
-0.03041260503232479,
-0.057480163872241974,
-0.06791207939386368,
0.2042027711868286,
-0.052767377346754074,
-0.1670594960451126,
-0.09752080589532852,
0.0069222175516188145,
-0.14184100925922394,
-0.02772429585456848,
0.07405169308185577,
-0.09855141490697861,
-0.07926216721534729,
-0.04910214990377426,
0.06557668745517731,
0.03220512717962265,
-0.012827475555241108,
-0.0056162988767027855,
0.06764590740203857,
0.017076071351766586,
-0.15133550763130188,
-0.016819199547171593,
0.00032638272386975586,
-0.09295172244310379,
0.023471254855394363,
0.006895660888403654,
0.07912898063659668,
0.01469259150326252,
0.022603509947657585,
0.0020858426578342915,
-0.046811386942863464,
0.2157202512025833,
-0.09994489699602127,
0.06605780124664307,
0.18618452548980713,
0.097701795399189,
0.08226809650659561,
0.16566652059555054,
0.07003108412027359,
-0.03999602049589157,
0.028312696143984795,
0.022195450961589813,
-0.030446410179138184,
-0.2776089906692505,
-0.08571261167526245,
-0.09666893631219864,
0.013430082239210606,
0.004702530335634947,
0.0379214882850647,
-0.11391150951385498,
0.04038020968437195,
-0.09883470833301544,
0.018252147361636162,
0.03940387815237045,
0.05781611055135727,
0.051678430289030075,
0.061760734766721725,
0.07560088485479355,
-0.0897708460688591,
-0.07045594602823257,
0.12830083072185516,
0.1783476620912552,
0.21585270762443542,
-0.04274445027112961,
0.07015755027532578,
0.05107893422245979,
0.028813984245061874,
-0.004290207289159298,
0.11052223294973373,
0.0013284179149195552,
-0.011974030174314976,
-0.02836470492184162,
-0.05517474561929703,
0.07508332282304764,
0.07167716324329376,
0.0513431616127491,
-0.04573900252580643,
-0.0519438236951828,
0.009148575365543365,
0.12762384116649628,
0.20649437606334686,
0.007287308573722839,
-0.12927278876304626,
-0.06530594825744629,
0.04126812890172005,
-0.07944132387638092,
0.0041133626364171505,
0.00834154523909092,
0.1285642385482788,
-0.15289030969142914,
0.10874692350625992,
0.0005715145962312818,
0.1013583391904831,
0.0009306300198659301,
0.012569533661007881,
0.02373085916042328,
0.09038542211055756,
-0.034045636653900146,
0.08968668431043625,
-0.2999592125415802,
0.19192126393318176,
0.000971316359937191,
0.10969219356775284,
-0.07009559869766235,
0.060972876846790314,
0.06180611625313759,
0.008487340994179249,
0.10829338431358337,
-0.02089805342257023,
-0.06898055970668793,
-0.07883693277835846,
-0.08966745436191559,
0.05155933275818825,
0.05961758643388748,
-0.04551045596599579,
0.10580559074878693,
-0.00797705166041851,
0.010388373397290707,
-0.04202401638031006,
0.11685170233249664,
-0.13765284419059753,
-0.1467348337173462,
-0.0026226891204714775,
0.042070355266332626,
-0.05669212341308594,
-0.04797597974538803,
-0.09234326332807541,
-0.08782773464918137,
0.11432675272226334,
-0.13016915321350098,
-0.09905051440000534,
-0.12759600579738617,
-0.03284936025738716,
0.08866584300994873,
-0.12981724739074707,
-0.04300541803240776,
-0.006841030437499285,
0.11977484077215195,
-0.07270403951406479,
-0.05739149451255798,
0.014141940511763096,
-0.12800993025302887,
-0.13200290501117706,
-0.019882233813405037,
0.11426768451929092,
-0.02371874637901783,
0.03660890832543373,
0.043395064771175385,
-0.055384520441293716,
-0.0616782046854496,
-0.15489087998867035,
-0.03393927589058876,
0.014148427173495293,
0.08834841847419739,
0.10789167881011963,
-0.10276035219430923,
-0.19161178171634674,
-0.028254155069589615,
-0.004869046155363321,
0.07410061359405518,
0.22588345408439636,
-0.03790369629859924,
0.05648675560951233,
0.15454651415348053,
-0.028382837772369385,
-0.23068177700042725,
-0.04364391043782234,
0.09575093537569046,
0.044132646173238754,
0.00008848665311234072,
-0.10694915801286697,
0.12469055503606796,
0.04963654279708862,
-0.002386814448982477,
-0.10695517808198929,
-0.23955616354942322,
-0.12298920005559921,
-0.023134814575314522,
0.02278384193778038,
0.018670886754989624,
-0.11908683180809021,
-0.01654372364282608,
-0.08021309971809387,
0.0026470008306205273,
0.042156532406806946,
-0.061703816056251526,
0.12114769220352173,
0.03719475492835045,
0.044937364757061005,
-0.005967482924461365,
-0.06520478427410126,
0.10323011875152588,
0.10458628833293915,
0.044415246695280075,
0.009610424749553204,
0.01173252984881401,
0.11344076693058014,
0.0010614971397444606,
0.11063015460968018,
-0.061578113585710526,
0.04549238830804825,
-0.12224646657705307,
-0.03819316253066063,
-0.04402698948979378,
0.0859263464808464,
-0.07207480072975159,
-0.07856866717338562,
-0.05884493887424469,
0.08776538074016571,
0.06880912184715271,
-0.00783261563628912,
0.010188398882746696,
-0.02440839260816574,
-0.016444135457277298,
0.09920800477266312,
0.12297529727220535,
-0.01915922947227955,
-0.13116395473480225,
-0.029380209743976593,
0.002724426332861185,
0.06330297142267227,
-0.0752749890089035,
0.05801081657409668,
0.131003275513649,
-0.04262857511639595,
0.06660299003124237,
0.0362568274140358,
-0.057969752699136734,
0.07355927675962448,
0.06165628135204315,
-0.0616462379693985,
-0.14581570029258728,
-0.046206410974264145,
-0.009587732143700123,
-0.07414624840021133,
0.10629604756832123,
0.1971026211977005,
-0.014871856197714806,
-0.0018552944529801607,
-0.02420244924724102,
-0.004129634238779545,
0.0237363800406456,
0.1638060212135315,
0.01217381190508604,
0.054719313979148865,
-0.06670763343572617,
0.06182999163866043,
0.03997188061475754,
-0.07128479331731796,
-0.02469884417951107,
0.11602135002613068,
-0.07242202758789062,
-0.03834017366170883,
-0.022458739578723907,
0.048725053668022156,
-0.09876053780317307,
-0.03195060417056084,
-0.08047647774219513,
-0.09876357764005661,
0.046971239149570465,
-0.06838871538639069,
0.04939842224121094,
-0.014096127822995186,
-0.05069392919540405,
-0.024886317551136017,
-0.005289198365062475,
0.11732736974954605,
0.0880822166800499,
0.03466443344950676,
-0.10500482469797134,
0.03642282262444496,
-0.045988161116838455,
0.04277806356549263,
-0.017819266766309738,
0.021923042833805084,
-0.11608409881591797,
-0.0164948720484972,
-0.11822046339511871,
0.06616749614477158,
-0.004857881926000118,
0.033879999071359634,
-0.05300629138946533,
-0.08353997021913528,
-0.02518538013100624,
0.05838927999138832,
-0.06475130468606949,
-0.0069679091684520245,
-0.03127637878060341,
0.1388980746269226,
-0.12462519854307175,
0.021551361307501793,
0.019010040909051895,
-0.06842198222875595,
0.1304423063993454,
0.003165546106174588,
-0.07338191568851471,
0.008940132334828377,
-0.09402744472026825,
-0.028307367116212845,
-0.018311597406864166,
0.11426495760679245,
0.049192022532224655,
-0.0219303909689188,
0.04518605023622513,
0.045642685145139694,
0.016802093014121056,
0.017452167347073555,
-0.01091883983463049,
-0.04344673082232475,
0.0494331531226635,
-0.017811626195907593,
-0.12635228037834167,
-0.046021588146686554,
0.050930123776197433,
0.1281283050775528,
0.04683101177215576,
0.15840719640254974,
-0.07893523573875427,
0.06565544754266739,
-0.13924948871135712,
-0.016802864149212837,
0.07179079949855804,
-0.08851388841867447,
-0.0995216891169548,
-0.06321033835411072,
0.08240804821252823,
-0.0012481095036491752,
0.1364056020975113,
-0.01924773119390011,
0.1328125298023224,
0.017645716667175293,
0.0013972471933811903,
-0.0063896398060023785,
0.03695150837302208,
0.09314508736133575,
-0.018823400139808655,
0.052268195897340775,
-0.012844658456742764,
-0.05851517245173454,
0.009857621975243092,
0.0880744457244873,
0.15472005307674408,
0.16259515285491943,
0.08043050765991211,
0.04053284600377083,
0.04310007393360138,
-0.0023927732836455107,
-0.1765032410621643,
-0.03328441455960274,
-0.014340130612254143,
0.03471406549215317,
-0.07964448630809784,
0.012266483157873154,
0.1578984409570694,
-0.19265638291835785,
0.06371661275625229,
0.0024177166633307934,
-0.0467461459338665,
-0.07793723791837692,
-0.08833430707454681,
-0.08742166310548782,
-0.055025264620780945,
0.014440951868891716,
-0.16721586883068085,
0.0035787043161690235,
0.05981486290693283,
0.08885495364665985,
-0.023357002064585686,
0.12885698676109314,
-0.0849744901061058,
-0.042632605880498886,
0.11199085414409637,
0.016502227634191513,
0.07157254219055176,
0.09584127366542816,
-0.024645142257213593,
-0.060358744114637375,
0.025729842483997345,
0.031998440623283386,
0.0370553657412529,
-0.028141478076577187,
-0.05640732869505882,
-0.0895845964550972,
-0.011750978417694569,
-0.03544178232550621,
-0.03586341440677643,
0.029738200828433037,
0.1213211715221405,
0.045994989573955536,
-0.04311586171388626,
0.01474304124712944,
0.17341755330562592,
-0.053908929228782654,
-0.15206024050712585,
-0.18192684650421143,
0.017986580729484558,
-0.006386501248925924,
0.05410845950245857,
0.025072617456316948,
-0.03762954846024513,
-0.06828948855400085,
0.14799225330352783,
0.19305507838726044,
-0.10878532379865646,
0.003920264076441526,
0.005552967544645071,
0.0009006123291328549,
-0.004771066829562187,
0.1705552637577057,
0.05950408801436424,
0.2735150158405304,
-0.01669018529355526,
0.08367561548948288,
0.0036610784009099007,
0.007765692658722401,
-0.10339728742837906,
0.17784638702869415,
-0.054799895733594894,
-0.05056784674525261,
-0.05389425903558731,
0.1881616860628128,
0.014209076762199402,
-0.14385291934013367,
-0.055220138281583786,
-0.09430389851331711,
-0.14836329221725464,
0.008724120445549488,
0.07166621088981628,
0.042161185294389725,
0.06441045552492142,
-0.020105063915252686,
0.012056177482008934,
0.08352575451135635,
-0.018661418929696083,
0.010699501261115074,
-0.022612616419792175,
0.06950481981039047,
-0.11959826946258545,
0.1636027842760086,
0.01275739073753357,
0.07603107392787933,
0.12234597653150558,
-0.0443735234439373,
-0.06155797094106674,
0.047482285648584366,
0.0866427943110466,
-0.12285441905260086,
0.06748052686452866,
0.005088013596832752,
-0.0667009949684143,
0.07515573501586914,
0.07298755645751953,
-0.06948581337928772,
0.05924071744084358,
0.1413753628730774,
0.006193236913532019,
-0.1175234392285347,
0.15222132205963135,
-0.12653519213199615,
0.07964155077934265,
0.13108058273792267,
-0.034270692616701126,
-0.010298310779035091,
-0.0007453642319887877,
0.09349218010902405,
0.004693392664194107,
0.07233957201242447,
0.0019768881611526012,
-0.19221442937850952,
0.01741453818976879,
-0.0834222212433815,
0.07238311320543289,
-0.1307172030210495,
-0.06797919422388077,
-0.054506462067365646,
0.029815809801220894,
-0.10679758340120316,
0.13458088040351868,
0.017398135736584663,
-0.06088131666183472,
0.0180048868060112,
-0.18842782080173492,
-0.0585930272936821,
0.0848078578710556,
-0.07893230020999908,
-0.05763721838593483
] |
null | null |
transformers
|
```
from transformers import BertForSequenceClassification, BertTokenizer, TextClassificationPipeline
model = BertForSequenceClassification.from_pretrained("deeq/dbert-sentiment")
tokenizer = BertTokenizer.from_pretrained("deeq/dbert")
nlp = TextClassificationPipeline(model=model, tokenizer=tokenizer)
print(nlp("좋아요"))
print(nlp("글쎄요"))
```
|
{}
|
text-classification
|
baikal-nlp/dbert-sentiment
|
[
"transformers",
"pytorch",
"bert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
[] |
[
"TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
36
] |
[
"passage: TAGS\n#transformers #pytorch #bert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.026536712422966957,
0.04976736754179001,
-0.007731540594249964,
0.02341027930378914,
0.20494870841503143,
0.04218224436044693,
0.07166644185781479,
0.1081078052520752,
0.06540437042713165,
-0.032089509069919586,
0.10898502916097641,
0.22890737652778625,
-0.03745893016457558,
0.11566723883152008,
-0.11783017218112946,
-0.2979154586791992,
0.06830790638923645,
0.06908033788204193,
-0.00024103521718643606,
0.11134638637304306,
0.08377818018198013,
-0.08843351900577545,
0.07041390985250473,
-0.039391499012708664,
-0.11928878724575043,
0.04120895266532898,
0.04201217740774155,
-0.128968745470047,
0.10206019878387451,
0.04667878895998001,
0.1612440049648285,
0.020005566999316216,
-0.06739236414432526,
-0.154616579413414,
0.0336696095764637,
-0.0003978070744778961,
-0.07949702441692352,
0.041154924780130386,
0.0834934264421463,
-0.11765240132808685,
0.009522730484604836,
0.03841863572597504,
0.022789550945162773,
0.05028093606233597,
-0.14333219826221466,
-0.07251771539449692,
-0.002419156488031149,
0.029012316837906837,
0.05913800001144409,
0.0598759762942791,
-0.002957597142085433,
0.1424214392900467,
-0.13965049386024475,
0.12620261311531067,
0.09429286420345306,
-0.2923010587692261,
-0.0088807987049222,
0.08759160339832306,
0.027602141723036766,
0.05489844083786011,
-0.0477239191532135,
0.050349123775959015,
0.02152593620121479,
0.006244409829378128,
0.0014351755380630493,
-0.07273159921169281,
-0.10605065524578094,
0.035220496356487274,
-0.08316893875598907,
-0.049130894243717194,
0.19201035797595978,
-0.05444741249084473,
0.07102859020233154,
-0.02369237318634987,
-0.09560474008321762,
-0.05289671570062637,
-0.025107022374868393,
0.018850168213248253,
-0.04439172148704529,
0.06905210763216019,
0.029569845646619797,
0.010320127010345459,
-0.10565698146820068,
0.03122977539896965,
-0.21760006248950958,
0.20976117253303528,
0.014575785025954247,
0.05031627416610718,
-0.18153272569179535,
0.05320606753230095,
0.012799138203263283,
-0.09720049053430557,
0.05345968157052994,
-0.10138970613479614,
0.025928562507033348,
-0.03812453895807266,
-0.07192374020814896,
-0.02739580348134041,
0.0816444382071495,
0.13516521453857422,
0.041108258068561554,
0.059053935110569,
-0.041404832154512405,
0.08655412495136261,
0.03162030130624771,
0.13472779095172882,
0.033738039433956146,
-0.03977132961153984,
0.03946247324347496,
-0.1231473833322525,
-0.00647841626778245,
-0.07187561690807343,
-0.15673033893108368,
-0.028635773807764053,
0.0752328485250473,
0.08006087690591812,
0.004822878632694483,
0.09519599378108978,
-0.060662589967250824,
-0.03062121570110321,
0.08224024623632431,
-0.07733647525310516,
0.02224470116198063,
0.026406224817037582,
0.022000659257173538,
0.101798415184021,
-0.019659440964460373,
-0.0015375112416222692,
-0.08451379835605621,
0.14397169649600983,
-0.05375861003994942,
0.007298978511244059,
-0.033063821494579315,
-0.08073414862155914,
0.030349934473633766,
-0.1405506432056427,
0.02792905643582344,
-0.17062804102897644,
-0.09528942406177521,
0.012332682497799397,
0.019128363579511642,
0.001749284565448761,
-0.023433325812220573,
-0.03093680739402771,
0.006545908749103546,
0.0424383319914341,
-0.06069215387105942,
-0.06501442193984985,
-0.07589498907327652,
0.09581612050533295,
-0.030969982966780663,
0.07632152736186981,
-0.12148407846689224,
0.0795503556728363,
-0.09479566663503647,
-0.02546360157430172,
-0.12941965460777283,
0.0373355858027935,
-0.04622753709554672,
0.1718803346157074,
0.013590065762400627,
-0.04714406654238701,
-0.05325024202466011,
0.06296809762716293,
-0.0670921728014946,
0.1699289232492447,
-0.06951840966939926,
-0.11890474706888199,
0.21500444412231445,
-0.08380883932113647,
-0.13756290078163147,
0.08620618283748627,
-0.014584669843316078,
0.0003590308187995106,
0.10495591163635254,
0.2001815140247345,
0.09155366569757462,
0.006830575410276651,
0.08820617198944092,
0.12077005207538605,
-0.08545669168233871,
-0.11240582168102264,
-0.005832689814269543,
-0.005655956454575062,
-0.14152070879936218,
0.057374704629182816,
0.07428599894046783,
0.07163040339946747,
-0.05171000212430954,
-0.03735769912600517,
-0.00994520727545023,
-0.007230670657008886,
0.13671016693115234,
0.054894041270017624,
0.11879090964794159,
-0.08347898721694946,
0.0004564806877169758,
0.008610726334154606,
-0.01850513368844986,
0.016449708491563797,
0.02691977098584175,
-0.06303919106721878,
0.11149293184280396,
0.01560661755502224,
0.030040156096220016,
-0.22954900562763214,
-0.07738739997148514,
-0.004857209976762533,
0.1317768543958664,
-0.016139987856149673,
0.11725185811519623,
0.05058206245303154,
-0.058512113988399506,
-0.013537165708839893,
-0.022089142352342606,
0.18155772984027863,
0.021907884627580643,
-0.06484844535589218,
-0.06583412736654282,
0.06355377286672592,
-0.07136932015419006,
-0.0009439446148462594,
-0.07689674198627472,
0.014060765504837036,
0.07748502492904663,
0.11273034662008286,
0.01009473018348217,
0.0715249702334404,
-0.026004934683442116,
0.06679531186819077,
-0.06401108205318451,
0.026931757107377052,
0.118758425116539,
-0.012655431404709816,
-0.0722615197300911,
0.15806308388710022,
-0.14796574413776398,
0.29817652702331543,
0.2068978101015091,
-0.3074952960014343,
0.003431171178817749,
-0.04410141706466675,
-0.0044579943642020226,
0.02857346273958683,
0.03858071565628052,
0.0010076105827465653,
0.0966758206486702,
0.001866061589680612,
0.20398299396038055,
-0.02794223465025425,
-0.041622765362262726,
-0.013331537134945393,
-0.048477016389369965,
-0.030720924958586693,
0.0930258184671402,
0.06317867338657379,
-0.2123168408870697,
0.1965867280960083,
0.2256951779127121,
0.020235497504472733,
0.16021281480789185,
-0.007587776519358158,
0.039776433259248734,
0.09032315015792847,
-0.04407169669866562,
-0.028124243021011353,
-0.07736045867204666,
-0.19754260778427124,
-0.048038698732852936,
0.07934695482254028,
0.03309439867734909,
0.06981316953897476,
-0.11398086696863174,
-0.028780505061149597,
0.005873502232134342,
0.02043827436864376,
-0.030085694044828415,
0.07733964174985886,
0.08143315464258194,
0.11509016901254654,
0.004185348749160767,
-0.07153625786304474,
0.11260948330163956,
-0.000511682010255754,
-0.0829218178987503,
0.18312713503837585,
-0.15064290165901184,
-0.35457029938697815,
-0.15102937817573547,
-0.20665821433067322,
-0.024193694815039635,
0.05783972144126892,
0.10403360426425934,
-0.11499883979558945,
-0.043235164135694504,
0.04162687435746193,
-0.002294909907504916,
-0.062236279249191284,
0.043300777673721313,
-0.06919633597135544,
0.06772983074188232,
-0.056030891835689545,
-0.06578446924686432,
-0.07393626123666763,
-0.03732382878661156,
-0.014316183514893055,
0.15351881086826324,
-0.13038954138755798,
0.07227112352848053,
0.17736823856830597,
-0.010089537128806114,
0.06711666285991669,
-0.038066890090703964,
0.17289850115776062,
-0.08991120010614395,
-0.028915125876665115,
0.17436328530311584,
-0.08250410854816437,
0.0786806121468544,
0.1608263999223709,
0.022829625755548477,
-0.06656418740749359,
0.029574787244200706,
-0.039340466260910034,
-0.08898156136274338,
-0.21838922798633575,
-0.148067444562912,
-0.11992256343364716,
0.05924250930547714,
0.062341220676898956,
0.0696515217423439,
0.12266137450933456,
0.059203725308179855,
0.015477290377020836,
-0.0007553264731541276,
-0.00036606789217330515,
0.07583136856555939,
0.2552264630794525,
-0.0020075372885912657,
0.14774726331233978,
-0.05434580519795418,
-0.1357378363609314,
0.08277737349271774,
0.01842481829226017,
0.11255297809839249,
0.09960044920444489,
0.014388641342520714,
0.00644803699105978,
0.061373550444841385,
0.16953399777412415,
0.12216762453317642,
0.03140265494585037,
-0.015582526102662086,
-0.02194071002304554,
0.0020790479611605406,
-0.07317613065242767,
0.01304252166301012,
0.07926664501428604,
-0.15208743512630463,
-0.08082117140293121,
-0.15637393295764923,
0.09561088681221008,
0.07428078353404999,
0.049813296645879745,
-0.2040114849805832,
0.009754637256264687,
0.09391142427921295,
-0.030124526470899582,
-0.09943006932735443,
0.07721024006605148,
-0.04655788093805313,
-0.14170965552330017,
0.10022298991680145,
-0.03493443876504898,
0.13963255286216736,
-0.08681212365627289,
0.0932496190071106,
-0.03828613832592964,
-0.11996456235647202,
0.032416898757219315,
0.11279566586017609,
-0.2732636332511902,
0.2332872599363327,
0.010946370661258698,
-0.07370878010988235,
-0.07934506237506866,
-0.026762953028082848,
0.041629474610090256,
0.2184235155582428,
0.059754300862550735,
0.002968377433717251,
-0.06075876206159592,
-0.18870088458061218,
-0.006565387360751629,
0.009197798557579517,
0.1304895579814911,
-0.03775089234113693,
-0.01569267176091671,
-0.0419037826359272,
-0.03293319419026375,
-0.029366329312324524,
-0.038726381957530975,
0.035019759088754654,
-0.17148956656455994,
0.05544491484761238,
0.0377817265689373,
0.07244952023029327,
0.019013661891222,
-0.04418788477778435,
-0.12382309138774872,
0.197775200009346,
-0.07774023711681366,
-0.07762428373098373,
-0.11156740039587021,
-0.0784677043557167,
0.02202191948890686,
-0.08583451807498932,
0.06005251407623291,
-0.08652466535568237,
0.016779478639364243,
-0.06327100098133087,
-0.20542916655540466,
0.13557077944278717,
-0.09933868050575256,
-0.02806651033461094,
-0.06584896147251129,
0.15001823008060455,
-0.0767940878868103,
0.01665390096604824,
0.032243192195892334,
0.01813752017915249,
-0.08950473368167877,
-0.07632966339588165,
-0.0018591269617900252,
0.016023563221096992,
0.051190294325351715,
0.06099539250135422,
-0.10083866119384766,
-0.06429562717676163,
-0.03685735538601875,
0.014633421786129475,
0.2983008325099945,
0.15201643109321594,
-0.064842589199543,
0.15137214958667755,
0.1384977251291275,
-0.07145120203495026,
-0.3431456387042999,
-0.08484037965536118,
-0.10642081499099731,
-0.04306303709745407,
-0.046289920806884766,
-0.16240264475345612,
0.1177973523736,
-0.01295482087880373,
-0.017609771341085434,
0.08445491641759872,
-0.15145616233348846,
-0.08574660867452621,
0.20116515457630157,
-0.026354258880019188,
0.39032095670700073,
-0.10577060282230377,
-0.09967513382434845,
-0.058081019669771194,
-0.121647909283638,
0.1404253989458084,
0.01047214213758707,
0.08493805676698685,
-0.009310578927397728,
0.06269232928752899,
0.044362664222717285,
-0.039699576795101166,
0.0946895033121109,
0.011379079893231392,
0.01612018421292305,
-0.1131248027086258,
-0.11301706731319427,
0.007961280643939972,
-0.020395388826727867,
-0.015760453417897224,
-0.00797135941684246,
0.010495016351342201,
-0.1657717525959015,
-0.04052681475877762,
-0.07803480327129364,
0.05654771998524666,
0.04092846438288689,
-0.03739452734589577,
0.006628011353313923,
-0.021788200363516808,
-0.004880301654338837,
0.005630741827189922,
0.2613779604434967,
-0.0545232780277729,
0.17406699061393738,
0.09997852146625519,
0.13454613089561462,
-0.1609681397676468,
0.02098797634243965,
-0.07239851355552673,
-0.06214495375752449,
0.07092934846878052,
-0.07368351519107819,
0.07252470403909683,
0.13665559887886047,
-0.06541655212640762,
0.06771941483020782,
0.11662118136882782,
0.059408992528915405,
-0.036289941519498825,
0.15761438012123108,
-0.22834214568138123,
0.03067406453192234,
-0.05344350263476372,
-0.016603386029601097,
0.06837030500173569,
0.06279473006725311,
0.13393236696720123,
0.05459635332226753,
-0.043871067464351654,
0.002848769072443247,
-0.009000388905405998,
-0.0024182628840208054,
0.06304551661014557,
0.05966060236096382,
0.04511803761124611,
-0.1356714963912964,
0.04809432104229927,
0.04747392609715462,
-0.180133655667305,
-0.015834158286452293,
0.13596458733081818,
-0.16409757733345032,
-0.1254369467496872,
-0.014882639050483704,
0.14192621409893036,
-0.10471435636281967,
-0.05383450537919998,
-0.06129157543182373,
-0.13381800055503845,
0.07413940876722336,
0.2071639746427536,
0.1222052350640297,
0.0868266150355339,
-0.05386051535606384,
-0.04268129542469978,
0.013526189140975475,
-0.0048812017776072025,
0.000410786597058177,
0.025256112217903137,
-0.10814032703638077,
0.030668657273054123,
-0.016766557469964027,
0.1549694985151291,
-0.0959596261382103,
-0.07802556455135345,
-0.1802983283996582,
0.04513133689761162,
-0.09466184675693512,
-0.0370473749935627,
-0.0709371566772461,
-0.02417352795600891,
0.006666520144790411,
-0.05355566740036011,
-0.037236955016851425,
-0.06802111119031906,
-0.12928707897663116,
0.041479259729385376,
-0.020978856831789017,
0.04500049725174904,
-0.06915231049060822,
-0.044879116117954254,
0.10159214586019516,
-0.03420831635594368,
0.09736357629299164,
0.10930681228637695,
-0.09245149046182632,
0.10182034969329834,
-0.1422983705997467,
-0.12590330839157104,
0.12893438339233398,
0.02711927518248558,
0.07729022204875946,
0.07507030665874481,
0.03763207420706749,
0.06823369115591049,
0.010589729994535446,
0.06981854140758514,
0.07503756135702133,
-0.12291653454303741,
0.05968776345252991,
-0.029884058982133865,
-0.1753014773130417,
-0.0433085560798645,
-0.04439191892743111,
0.0955316424369812,
0.0027241117786616087,
0.1497408151626587,
-0.05511422082781792,
0.10076558589935303,
-0.034150879830121994,
0.007692268583923578,
-0.018073493614792824,
-0.21772713959217072,
-0.0606854185461998,
-0.0871635228395462,
0.02593981847167015,
0.0011935612419620156,
0.25321871042251587,
0.060501862317323685,
0.047848355025053024,
0.05219770595431328,
0.0803675726056099,
-0.005114862695336342,
0.024529730901122093,
0.17962703108787537,
0.10480687767267227,
-0.056110929697752,
-0.06051088124513626,
0.06334816664457321,
0.020930401980876923,
0.0037487365771085024,
0.1370624452829361,
0.07126225531101227,
-0.025097444653511047,
0.07946188747882843,
-0.022343091666698456,
0.04902622103691101,
-0.13099198043346405,
-0.18797338008880615,
-0.0362858846783638,
0.0822635069489479,
0.008193853311240673,
0.06443964689970016,
0.08426562696695328,
-0.029682213440537453,
0.05306922271847725,
-0.05066857486963272,
-0.05351173132658005,
-0.18930195271968842,
-0.08380404114723206,
-0.09822331368923187,
-0.10457178205251694,
0.005494547076523304,
-0.07776138931512833,
-0.007010516710579395,
0.0889914408326149,
0.04661707207560539,
-0.04920143634080887,
0.07530294358730316,
0.005946568213403225,
-0.05706603452563286,
0.08126048743724823,
-0.03929980844259262,
0.035573069006204605,
-0.004914599005132914,
-0.03371923416852951,
-0.14008314907550812,
-0.01516553945839405,
-0.04801462963223457,
0.04064973443746567,
-0.06195714324712753,
0.005159671418368816,
-0.14515186846256256,
-0.12241504341363907,
-0.02803012728691101,
0.052396051585674286,
-0.057535864412784576,
0.1368633210659027,
0.0011574793606996536,
0.005521169863641262,
0.04932842403650284,
0.20023545622825623,
-0.06511733680963516,
-0.05715160444378853,
-0.03652770072221756,
0.25829756259918213,
0.07457830011844635,
0.1169453114271164,
-0.00899919681251049,
-0.005612371955066919,
-0.09091201424598694,
0.33026692271232605,
0.2956998944282532,
-0.053279418498277664,
0.04956240579485893,
0.021509597077965736,
0.03881188482046127,
0.15772217512130737,
0.14865562319755554,
0.08992809802293777,
0.23702938854694366,
-0.06422201544046402,
-0.03442860767245293,
-0.01987340860068798,
-0.020380141213536263,
-0.12080796808004379,
0.0805739313364029,
0.0668584331870079,
-0.04897911474108696,
-0.07300020754337311,
0.10295598208904266,
-0.20242395997047424,
0.1379757672548294,
-0.0004371747490949929,
-0.21978750824928284,
-0.07235769927501678,
-0.033254899084568024,
0.14364826679229736,
-0.008563272655010223,
0.08543892204761505,
-0.0008842989918775856,
-0.11287369579076767,
0.02438463270664215,
0.019631659612059593,
-0.2149261236190796,
-0.020671725273132324,
0.06799837201833725,
-0.04828084260225296,
-0.0022455095313489437,
-0.019154418259859085,
0.03060556948184967,
0.07118745148181915,
0.06662483513355255,
-0.008223235607147217,
0.04078114032745361,
0.002477803034707904,
-0.04111843183636665,
0.005046389531344175,
0.015769492834806442,
0.0016257762908935547,
-0.09229632467031479,
0.06661305576562881,
-0.16755518317222595,
0.05424373224377632,
-0.08128474652767181,
-0.06260967999696732,
-0.009263207204639912,
0.034962452948093414,
-0.054396457970142365,
0.04791352152824402,
0.10115863382816315,
0.0075232600793242455,
-0.032317593693733215,
-0.050817299634218216,
-0.03972724452614784,
-0.0014734352007508278,
-0.1398703157901764,
-0.14837034046649933,
-0.09151425212621689,
-0.09747706353664398,
0.11042709648609161,
0.001969041768461466,
-0.15724746882915497,
-0.0015884727472439408,
-0.0997534990310669,
0.06987065821886063,
-0.16953794658184052,
0.09285256266593933,
0.0341578908264637,
0.015343928709626198,
-0.015839118510484695,
-0.06526738405227661,
0.05342889577150345,
0.07730220258235931,
-0.12141184508800507,
-0.08985189348459244
] |
|
null | null |
transformers
|
deeqBERT-base
---
- model: bert-base
- vocab: bert-wordpiece, 35k
- version: latest
|
{"language": "ko", "datasets": ["kowiki", "news"]}
|
fill-mask
|
baikal-nlp/dbert
|
[
"transformers",
"pytorch",
"bert",
"fill-mask",
"ko",
"dataset:kowiki",
"dataset:news",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ko"
] |
TAGS
#transformers #pytorch #bert #fill-mask #ko #dataset-kowiki #dataset-news #autotrain_compatible #endpoints_compatible #region-us
|
deeqBERT-base
---
- model: bert-base
- vocab: bert-wordpiece, 35k
- version: latest
|
[] |
[
"TAGS\n#transformers #pytorch #bert #fill-mask #ko #dataset-kowiki #dataset-news #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
49
] |
[
"passage: TAGS\n#transformers #pytorch #bert #fill-mask #ko #dataset-kowiki #dataset-news #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.08694572001695633,
0.06884147971868515,
-0.005860787350684404,
-0.006025994662195444,
0.10558739304542542,
0.04051871970295906,
0.10314526408910751,
0.09117478877305984,
0.07795592397451401,
-0.024222135543823242,
0.1724592000246048,
0.18697744607925415,
-0.007956858724355698,
0.24218451976776123,
-0.07378733158111572,
-0.2724643349647522,
0.09890024363994598,
0.04836782068014145,
0.007435834500938654,
0.11808045953512192,
0.07477640360593796,
-0.09827098995447159,
0.06258521974086761,
-0.05081372708082199,
-0.09948772937059402,
0.015286436304450035,
0.031409695744514465,
-0.12661808729171753,
0.09994073212146759,
0.010473616421222687,
0.12420715391635895,
0.022358926013112068,
-0.07161695510149002,
-0.10130111873149872,
0.06108377128839493,
0.012014031410217285,
-0.05422089248895645,
0.04715995118021965,
-0.020205028355121613,
-0.056742697954177856,
-0.03563448041677475,
-0.0936310887336731,
-0.0055321897380054,
0.011624877341091633,
-0.15918190777301788,
-0.07051264494657516,
-0.051384247839450836,
0.045915670692920685,
0.09397527575492859,
0.0884033590555191,
0.001721526961773634,
0.1991959810256958,
-0.11752225458621979,
0.09412276744842529,
0.13012118637561798,
-0.23011396825313568,
-0.014962971210479736,
0.062458861619234085,
0.014649092219769955,
-0.04926347732543945,
-0.03498721867799759,
0.08661071956157684,
0.021183371543884277,
0.007737416308373213,
-0.05539001151919365,
-0.055363330990076065,
-0.0990675687789917,
0.013841303065419197,
-0.05751267820596695,
-0.06883303821086884,
0.2220340222120285,
-0.010075939819216728,
0.07522568851709366,
-0.009259693324565887,
-0.08673466742038727,
-0.01431339606642723,
-0.027373498305678368,
0.02918839082121849,
-0.07367363572120667,
0.017740480601787567,
-0.055095870047807693,
-0.003107120981439948,
-0.11071746796369553,
0.05987819284200668,
-0.25607770681381226,
0.2623293399810791,
0.015740886330604553,
0.07524443417787552,
-0.19466522336006165,
0.06421460956335068,
0.006451472174376249,
-0.1478240191936493,
0.022221557796001434,
-0.0892200842499733,
0.022072210907936096,
-0.0071359118446707726,
-0.0459977462887764,
-0.021355098113417625,
0.08473421633243561,
0.16254493594169617,
0.014567388221621513,
-0.00037838969728909433,
0.037809450179338455,
0.08206280320882797,
0.10041876882314682,
0.022179212421178818,
-0.055991560220718384,
-0.08519458025693893,
0.053821392357349396,
-0.09399588406085968,
0.019304662942886353,
-0.06334929913282394,
-0.05818483605980873,
-0.039572399109601974,
-0.0010846966179087758,
0.06196550279855728,
0.09531533718109131,
0.08881844580173492,
-0.08150933682918549,
-0.014749116264283657,
0.10199138522148132,
-0.06683054566383362,
-0.007123704068362713,
-0.031127335503697395,
0.02119835838675499,
0.15651294589042664,
-0.01655610278248787,
0.02678871899843216,
-0.011694669723510742,
0.1340523064136505,
-0.0796796977519989,
-0.040921103209257126,
-0.009645835496485233,
-0.0189328882843256,
0.06521520763635635,
-0.14410477876663208,
0.0636415183544159,
-0.15751859545707703,
-0.19035330414772034,
0.03253987058997154,
0.01587756536900997,
-0.010892663151025772,
-0.056043632328510284,
0.007289660628885031,
-0.003736524609848857,
0.046141792088747025,
-0.04481424391269684,
-0.006790309678763151,
-0.09064064919948578,
0.0992790013551712,
-0.05189597234129906,
0.099251888692379,
-0.11458265781402588,
0.02299552783370018,
-0.10665438324213028,
-0.0026188031770288944,
-0.07275162637233734,
0.016720162704586983,
-0.07420727610588074,
0.13407142460346222,
-0.030710777267813683,
0.0004846991505473852,
-0.09069639444351196,
0.016009947285056114,
-0.00926192756742239,
0.22723788022994995,
-0.12164455652236938,
-0.08596339076757431,
0.19109904766082764,
-0.08299180865287781,
-0.11866403371095657,
0.061673663556575775,
-0.02684994414448738,
0.04210824519395828,
0.11264854669570923,
0.05870312079787254,
0.056933797895908356,
-0.09224146604537964,
-0.015715256333351135,
0.07641938328742981,
-0.10526339709758759,
-0.06751455366611481,
0.015933891758322716,
0.041638053953647614,
-0.10353963077068329,
0.06367015093564987,
0.11164763569831848,
0.09200997650623322,
-0.10171765089035034,
-0.06008576601743698,
0.003630253951996565,
-0.0515316016972065,
0.13451209664344788,
0.015215309336781502,
0.08142831176519394,
-0.05302856117486954,
0.005276496056467295,
-0.02544901706278324,
0.05344551056623459,
0.08597131073474884,
0.03916091471910477,
-0.09083542972803116,
0.10835763812065125,
-0.10054206848144531,
0.029430214315652847,
-0.14613664150238037,
-0.11001242697238922,
-0.041817255318164825,
0.1152271255850792,
-0.06427527219057083,
0.06783562898635864,
0.08135883510112762,
-0.1084444671869278,
-0.029240330681204796,
-0.008909483440220356,
0.11011640727519989,
0.0550910122692585,
-0.05830010771751404,
-0.11437962204217911,
0.06318921595811844,
-0.07832352817058563,
0.02006901055574417,
-0.019842758774757385,
-0.02246125601232052,
0.042567938566207886,
0.16899345815181732,
-0.0008293516584672034,
0.039958808571100235,
0.005215262528508902,
0.016431841999292374,
-0.018497558310627937,
0.02039005234837532,
0.07566829025745392,
0.0004706062318291515,
-0.037826817482709885,
0.18272827565670013,
-0.05955563113093376,
0.23818570375442505,
0.1635076254606247,
-0.20918968319892883,
0.018548697233200073,
0.022764815017580986,
-0.030009401962161064,
-0.030526168644428253,
-0.011424784548580647,
-0.0053518982604146,
-0.0004561723326332867,
-0.0032478345092386007,
0.07581078261137009,
0.012709174305200577,
-0.030854221433401108,
0.034628212451934814,
-0.047653183341026306,
-0.06736920773983002,
0.05758367478847504,
0.07142773270606995,
-0.1888623684644699,
0.17703726887702942,
0.21564388275146484,
0.05691269040107727,
0.17765741050243378,
-0.018022846430540085,
-0.024861624464392662,
0.008999818935990334,
-0.11876409500837326,
-0.05373477190732956,
0.09355323761701584,
-0.1729547530412674,
0.007450730539858341,
0.10019568353891373,
-0.013210280798375607,
0.05567113310098648,
-0.07271721959114075,
-0.09190386533737183,
-0.01550748199224472,
0.05240827053785324,
-0.08441845327615738,
0.14450156688690186,
0.04199022054672241,
0.09080804884433746,
0.003240803489461541,
-0.009467526338994503,
0.10843890905380249,
0.011834219098091125,
-0.02013220638036728,
0.1625751405954361,
-0.1169837936758995,
-0.3260160982608795,
-0.07636905461549759,
-0.19766293466091156,
-0.005666103679686785,
-0.005949708633124828,
0.05377750098705292,
-0.08730403333902359,
-0.04219817742705345,
0.06410863250494003,
-0.016239749267697334,
0.024684879928827286,
0.05383285880088806,
0.0010470632696524262,
-0.007869717665016651,
-0.054677579551935196,
-0.06417805701494217,
-0.056551091372966766,
-0.026772217825055122,
0.007357613183557987,
0.15020576119422913,
-0.08590461313724518,
0.09705952554941177,
0.08375804871320724,
0.04991370439529419,
0.06518429517745972,
-0.002010494703426957,
0.20810620486736298,
-0.10882388055324554,
0.014927170239388943,
0.1411176472902298,
-0.05798391252756119,
0.06703837215900421,
0.16308194398880005,
0.01639564521610737,
-0.04901554435491562,
0.006428119260817766,
-0.005959044676274061,
-0.0820845291018486,
-0.19202789664268494,
-0.13308106362819672,
-0.09878457337617874,
0.02913202904164791,
0.015558788552880287,
0.03192261978983879,
0.0824219137430191,
0.08996860682964325,
-0.006785089615732431,
-0.02221601828932762,
-0.06487870961427689,
0.021261923015117645,
0.13592982292175293,
-0.019366534426808357,
0.09427914768457413,
-0.042970698326826096,
-0.09153293818235397,
0.07685679197311401,
0.021955566480755806,
0.07866676896810532,
0.05447256192564964,
-0.018865521997213364,
0.04180894419550896,
0.11760339885950089,
0.13740308582782745,
0.12394353002309799,
0.03525645285844803,
-0.06591872870922089,
0.02575225569307804,
-0.0007951805018819869,
-0.0400315597653389,
-0.006448006723076105,
0.09285163879394531,
-0.10365854948759079,
-0.019433235749602318,
-0.11928243190050125,
0.0595245435833931,
0.13269825279712677,
0.12939538061618805,
-0.21332786977291107,
-0.025492418557405472,
0.04945167526602745,
0.0015821497654542327,
-0.0850982666015625,
0.034064821898937225,
-0.024152496829628944,
-0.13013841211795807,
0.11512258648872375,
-0.031122760847210884,
0.09765177965164185,
-0.032666660845279694,
0.03303094953298569,
-0.03811303898692131,
-0.11374284327030182,
-0.007896693423390388,
0.09429450333118439,
-0.24384796619415283,
0.27996793389320374,
-0.009008615277707577,
-0.0310574471950531,
-0.11200195550918579,
-0.06593360006809235,
0.035005781799554825,
0.12324102222919464,
0.10255496948957443,
0.05847736448049545,
0.01895596832036972,
-0.12531816959381104,
-0.13582134246826172,
0.0456891767680645,
0.03840303793549538,
-0.06481195986270905,
-0.001676399726420641,
0.02059629186987877,
-0.03552578017115593,
-0.015045312233269215,
0.043683912605047226,
-0.05514659360051155,
-0.1287233531475067,
0.09813564270734787,
0.06776499003171921,
0.050148166716098785,
0.0007026788662187755,
-0.11395048350095749,
-0.09356924891471863,
0.120900459587574,
0.04443637654185295,
-0.06009342521429062,
-0.12144529819488525,
-0.03324849531054497,
0.09541648626327515,
-0.10730261355638504,
0.042471058666706085,
-0.03426734358072281,
0.019389145076274872,
-0.05097443237900734,
-0.16835550963878632,
0.11856246739625931,
-0.12582725286483765,
-0.077816441655159,
-0.05756603553891182,
0.11891307681798935,
0.03936412185430527,
0.010024280287325382,
0.012707825750112534,
0.04010094329714775,
-0.05870428308844566,
-0.041328564286231995,
0.045391995459795,
0.052979085594415665,
0.08109566569328308,
0.0759212002158165,
0.013814982026815414,
-0.060952503234148026,
-0.02172435261309147,
-0.014763526618480682,
0.2330535352230072,
0.24959035217761993,
-0.042134154587984085,
0.13569985330104828,
0.16640670597553253,
-0.02665228210389614,
-0.35667407512664795,
-0.09574400633573532,
-0.1310628205537796,
-0.0028274008072912693,
-0.03062121383845806,
-0.1173107698559761,
0.12088943272829056,
0.0057813250459730625,
-0.05639483779668808,
0.12980319559574127,
-0.20337049663066864,
-0.0716988667845726,
0.21584542095661163,
-0.06416904926300049,
0.5225526690483093,
-0.07858705520629883,
-0.06140196695923805,
-0.02508743666112423,
-0.14547288417816162,
0.15106019377708435,
-0.0313359797000885,
0.07323619723320007,
-0.020916689187288284,
0.09336940199136734,
0.022669805213809013,
-0.06937083601951599,
0.12892542779445648,
-0.00970375631004572,
-0.027087761089205742,
-0.12179698050022125,
-0.11789243668317795,
0.08326079696416855,
-0.03716913238167763,
-0.035081230103969574,
0.05673230066895485,
0.02082340605556965,
-0.15724535286426544,
-0.007739391177892685,
-0.1255885809659958,
0.08703101426362991,
0.02265305258333683,
-0.044848546385765076,
0.015830252319574356,
0.014223078265786171,
0.02505480870604515,
-0.008791694417595863,
0.25975483655929565,
0.028022168204188347,
0.1039850041270256,
-0.02234433777630329,
0.028006091713905334,
-0.1171363815665245,
-0.0208264347165823,
-0.08469595015048981,
-0.08615519851446152,
0.07778012752532959,
-0.03759501874446869,
0.02501744218170643,
0.1079978197813034,
-0.003364124335348606,
0.019858468323946,
0.07494346052408218,
-0.044243309646844864,
0.0019736122339963913,
0.12691132724285126,
-0.2559729814529419,
0.08025376498699188,
-0.024256225675344467,
0.0037592556327581406,
0.041304416954517365,
0.057218678295612335,
0.1324397176504135,
0.028492307290434837,
-0.051902562379837036,
0.0010887086391448975,
0.0355001799762249,
-0.04769745096564293,
0.11190520226955414,
0.0650535598397255,
0.039191968739032745,
-0.1505940854549408,
0.07904835790395737,
-0.018407579511404037,
-0.2014235556125641,
0.0035474214237183332,
0.05843829736113548,
-0.14593705534934998,
-0.1082167997956276,
0.005216038785874844,
0.05874275416135788,
-0.055153895169496536,
-0.05704667419195175,
-0.040712617337703705,
-0.07673221081495285,
0.09432098269462585,
0.2526794970035553,
0.07355646044015884,
0.08108140528202057,
-0.010899634100496769,
-0.04768866300582886,
-0.004520026035606861,
0.03124324046075344,
0.07230118662118912,
0.013553149066865444,
-0.07335152477025986,
-0.047797173261642456,
-0.039914343506097794,
0.20894597470760345,
-0.10776951909065247,
-0.06724448502063751,
-0.14335539937019348,
0.03475012257695198,
-0.10525498539209366,
-0.06390167772769928,
-0.11728567630052567,
-0.06518363952636719,
-0.034227654337882996,
-0.11683081835508347,
-0.04019103944301605,
-0.052718400955200195,
-0.09871088713407516,
0.04998389631509781,
0.016324462369084358,
0.019127272069454193,
-0.06056361272931099,
-0.0495661124587059,
0.11600681394338608,
-0.030735589563846588,
0.1146048977971077,
0.15731489658355713,
-0.05930158868432045,
0.07226242870092392,
-0.10655001550912857,
-0.12464780360460281,
0.0874757394194603,
-0.002198197413235903,
0.08734229952096939,
0.03377271816134453,
0.006568589247763157,
0.055043287575244904,
0.04585321992635727,
0.07228602468967438,
0.05824236571788788,
-0.11195134371519089,
0.0481572225689888,
0.028088688850402832,
-0.13834154605865479,
-0.019918493926525116,
-0.10848305374383926,
0.09834108501672745,
0.006652329117059708,
0.14871811866760254,
-0.048176877200603485,
0.0944213941693306,
-0.01890711486339569,
0.015336086973547935,
-0.034360677003860474,
-0.18240398168563843,
0.04783562943339348,
-0.0053405254147946835,
0.04640231654047966,
-0.023187974467873573,
0.24299736320972443,
-0.0512961708009243,
-0.04712633416056633,
0.02992289513349533,
0.042391885071992874,
-0.06158001348376274,
0.02026677504181862,
0.17158296704292297,
0.09139439463615417,
-0.04361287131905556,
-0.09613557904958725,
0.0848693922162056,
0.018124600872397423,
-0.0007865448715165257,
0.1257709413766861,
0.05348334088921547,
0.1489170789718628,
0.0798674151301384,
-0.003875924739986658,
0.012537417933344841,
-0.056011270731687546,
-0.2561582326889038,
-0.15579795837402344,
0.05881838500499725,
0.04049602895975113,
-0.023951523005962372,
0.10925807803869247,
-0.02751246467232704,
0.010280562564730644,
-0.047819834202528,
-0.02384001575410366,
-0.13703344762325287,
-0.18486210703849792,
-0.081208236515522,
-0.05331195145845413,
0.02954161912202835,
-0.051696762442588806,
0.01164945773780346,
0.1736356019973755,
0.05047759413719177,
-0.03637218102812767,
0.12817524373531342,
0.042419493198394775,
-0.014729795977473259,
0.10837722569704056,
-0.004791232757270336,
-0.0007521543884649873,
-0.020431596785783768,
-0.020585494115948677,
-0.11578678339719772,
0.03195022791624069,
-0.04701561480760574,
0.022600462660193443,
-0.05021589994430542,
0.048009466379880905,
-0.10764054208993912,
-0.15011629462242126,
-0.05911535769701004,
0.032636333256959915,
-0.004817736800760031,
0.03793326020240784,
0.03904960677027702,
0.10218096524477005,
0.03945291414856911,
0.17281171679496765,
-0.02539179101586342,
-0.11964017897844315,
-0.07688742130994797,
0.06236496940255165,
0.029420940205454826,
0.07348515838384628,
-0.006491461303085089,
-0.016060445457696915,
-0.07390179485082626,
0.29511550068855286,
0.3036278188228607,
-0.029981236904859543,
0.06659780442714691,
0.013739925809204578,
0.03895263746380806,
0.05793840438127518,
0.09311138093471527,
0.053965527564287186,
0.22821450233459473,
-0.09087365865707397,
-0.06775745004415512,
-0.03725019469857216,
-0.04871831461787224,
-0.08309193700551987,
0.004508771002292633,
0.0372869074344635,
-0.04547639191150665,
-0.07191221415996552,
0.12260865420103073,
-0.20139656960964203,
0.10796534270048141,
-0.018614573404192924,
-0.22834597527980804,
-0.10306428372859955,
-0.03722361475229263,
0.1208595409989357,
0.04327011853456497,
0.08810430020093918,
-0.02247982658445835,
-0.04224695637822151,
0.06170911341905594,
0.026228521019220352,
-0.14777441322803497,
-0.09009956568479538,
0.08878817409276962,
0.027524853125214577,
0.07325439900159836,
-0.01654679886996746,
0.04418428614735603,
0.08280742168426514,
0.02391394041478634,
-0.06430205702781677,
-0.00041422416688874364,
0.031507790088653564,
-0.04835515469312668,
-0.042860694229602814,
0.02810598723590374,
0.009211731143295765,
-0.1094261109828949,
0.09581770747900009,
-0.11711731553077698,
0.050154201686382294,
-0.06151213124394417,
0.008695949800312519,
-0.022213997319340706,
0.10284345597028732,
-0.023677045479416847,
0.08406969159841537,
0.07036727666854858,
-0.0342293418943882,
-0.011153333820402622,
-0.02979273907840252,
0.001170616946183145,
0.019602498039603233,
-0.12373606115579605,
-0.13870085775852203,
-0.10345500707626343,
-0.08182848989963531,
-0.0027396881487220526,
0.03370819613337517,
-0.09047452360391617,
-0.004214017651975155,
-0.11193808913230896,
0.00021323730470612645,
-0.08264479786157608,
0.06413056701421738,
0.025486325845122337,
0.007108430378139019,
-0.0018949570367112756,
-0.01964554749429226,
0.05649527534842491,
0.04531904309988022,
-0.17859876155853271,
-0.09773646295070648
] |
null | null |
transformers
|
deeqBERT5
---
- model: bert-base
- vocab: deeqnlp 1.5, 50k
- version: latest/3.5
|
{}
| null |
baikal-nlp/dbert5
|
[
"transformers",
"pytorch",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #endpoints_compatible #region-us
|
deeqBERT5
---
- model: bert-base
- vocab: deeqnlp 1.5, 50k
- version: latest/3.5
|
[] |
[
"TAGS\n#transformers #pytorch #endpoints_compatible #region-us \n"
] |
[
21
] |
[
"passage: TAGS\n#transformers #pytorch #endpoints_compatible #region-us \n"
] |
[
-0.0602605901658535,
-0.005646900739520788,
-0.009762155823409557,
-0.03966370224952698,
0.15944775938987732,
0.03070714697241783,
0.012395896948873997,
0.07867952436208725,
0.09419925510883331,
-0.019594743847846985,
0.09831016510725021,
0.2332964390516281,
-0.03786272928118706,
0.022073544561862946,
-0.06232732906937599,
-0.24905818700790405,
0.09407828003168106,
0.11313164979219437,
-0.06465978920459747,
0.09549489617347717,
0.039143070578575134,
-0.08911364525556564,
0.04947805404663086,
-0.032982707023620605,
-0.13527381420135498,
0.0467669740319252,
0.0237716156989336,
-0.078513965010643,
0.11869441717863083,
0.0077665625140070915,
0.19578197598457336,
0.01207544095814228,
-0.1299089938402176,
-0.18088726699352264,
0.02007104456424713,
0.022213434800505638,
-0.05996024236083031,
0.02314351126551628,
0.07312147319316864,
-0.10279249399900436,
0.02970288321375847,
0.03654928505420685,
0.002923935651779175,
0.03274611383676529,
-0.15513424575328827,
-0.1849449872970581,
-0.04425787553191185,
0.012871243990957737,
0.026669111102819443,
0.10506244748830795,
0.030372466892004013,
0.178822323679924,
-0.15745419263839722,
0.0945795327425003,
0.177890345454216,
-0.28603804111480713,
0.029024237766861916,
0.11024846136569977,
0.04829736799001694,
0.028134070336818695,
0.00996005441993475,
0.0189230814576149,
-0.009570286609232426,
0.03048190288245678,
-0.027967484667897224,
-0.07447972893714905,
-0.06405609101057053,
0.0814741775393486,
-0.09437224268913269,
-0.1216578483581543,
0.19255080819129944,
-0.05330982804298401,
0.0588051900267601,
0.03463263809680939,
-0.11766522377729416,
-0.05456177890300751,
0.0013563521206378937,
0.030292518436908722,
-0.017828812822699547,
0.07086412608623505,
0.031105060130357742,
-0.02373676560819149,
-0.12168543040752411,
0.028789518401026726,
-0.24090629816055298,
0.28345873951911926,
0.02996366284787655,
0.11018568277359009,
-0.2384001910686493,
0.07198631763458252,
-0.04607108235359192,
-0.06767431646585464,
0.023487936705350876,
-0.09875187277793884,
0.021371465176343918,
0.007170209661126137,
-0.0868486687541008,
0.025816135108470917,
0.0673765018582344,
0.11560594290494919,
-0.02462717518210411,
0.021296672523021698,
0.01767325960099697,
0.1089232936501503,
0.01622004434466362,
0.11334555596113205,
0.002745070494711399,
0.034500978887081146,
0.029799794778227806,
-0.18719491362571716,
-0.0026790881529450417,
-0.03489845618605614,
-0.08398842811584473,
-0.0881200060248375,
0.028418414294719696,
0.12114420533180237,
0.02183910645544529,
0.022260701283812523,
-0.07126972824335098,
-0.008312570862472057,
0.04512425512075424,
-0.06700387597084045,
-0.022145040333271027,
0.007661914918571711,
0.016494954004883766,
0.22294339537620544,
-0.02295338362455368,
-0.026225684210658073,
-0.07023988664150238,
0.11260082572698593,
-0.0656941682100296,
0.016149358823895454,
-0.0530007965862751,
-0.02928324043750763,
0.056476715952157974,
-0.15177644789218903,
0.06500709801912308,
-0.1516612023115158,
-0.09112907201051712,
0.028223685920238495,
0.03696250542998314,
0.028484180569648743,
0.02479529194533825,
0.015830792486667633,
-0.00879291258752346,
-0.03011348284780979,
-0.07923634350299835,
-0.08192402124404907,
-0.06342831254005432,
0.09921494126319885,
-0.0005941772251389921,
0.05436241999268532,
-0.11821827292442322,
0.07408127188682556,
-0.1076948270201683,
0.030249644070863724,
-0.13414455950260162,
-0.0221982691437006,
-0.02404189109802246,
0.1851358413696289,
0.0006879806751385331,
-0.0817047506570816,
-0.10663671791553497,
0.038237277418375015,
-0.04168618097901344,
0.13458840548992157,
-0.014303965494036674,
-0.11019430309534073,
0.26959696412086487,
-0.10667987167835236,
-0.1671270877122879,
0.05378620699048042,
0.009700759314000607,
-0.016868380829691887,
0.0561099536716938,
0.16616025567054749,
0.06831997632980347,
-0.07049598544836044,
0.09140615165233612,
0.1379895955324173,
-0.1851922571659088,
-0.20243999361991882,
0.01763749308884144,
-0.05203584209084511,
-0.11545780301094055,
0.04540516808629036,
-0.0058397711254656315,
0.09395778924226761,
-0.08503725379705429,
-0.007578795775771141,
-0.032015107572078705,
-0.016012731939554214,
0.07450172305107117,
0.06282669305801392,
0.09098648279905319,
-0.05122196301817894,
0.031217509880661964,
0.026435445994138718,
0.00463126040995121,
0.0063572051003575325,
0.05537392571568489,
-0.03592291474342346,
0.1321462094783783,
-0.06449703127145767,
0.00637860456481576,
-0.2316930741071701,
-0.11218059808015823,
-0.012119117192924023,
0.04616983234882355,
-0.05828554928302765,
0.16098164021968842,
0.09591125696897507,
-0.0814671590924263,
0.027995459735393524,
-0.03218327462673187,
0.10552480816841125,
0.02011469565331936,
-0.012835591100156307,
-0.0011234998237341642,
0.016821393743157387,
-0.0699923112988472,
-0.08323190361261368,
0.006836527958512306,
-0.0009020745637826622,
0.08346128463745117,
0.12389257550239563,
-0.0018713462632149458,
0.04892571270465851,
-0.025174804031848907,
0.07054726779460907,
-0.016440844163298607,
0.014167504385113716,
0.10722272843122482,
-0.011425090953707695,
-0.05193426087498665,
0.1583033800125122,
-0.11112764477729797,
0.3326278030872345,
0.20375320315361023,
-0.323421835899353,
0.04571422189474106,
-0.025362668558955193,
-0.017437715083360672,
0.021043121814727783,
0.09183337539434433,
0.005831574089825153,
0.08254498243331909,
0.05105554684996605,
0.13220374286174774,
-0.025855854153633118,
-0.015603546984493732,
-0.000933936215005815,
-0.06394030898809433,
-0.03657902777194977,
0.07155993580818176,
0.06678375601768494,
-0.12739983201026917,
0.17137302458286285,
0.2332899272441864,
0.034409500658512115,
0.08888798952102661,
-0.074738509953022,
-0.02466125227510929,
0.05307190865278244,
0.030478334054350853,
-0.05014907941222191,
-0.003522941842675209,
-0.24837985634803772,
-0.033807482570409775,
0.0835256576538086,
0.041713811457157135,
0.11864139884710312,
-0.15303030610084534,
-0.0535346120595932,
0.036216262727975845,
0.003446921007707715,
-0.07469423115253448,
0.0990411564707756,
0.07169649004936218,
0.06572583317756653,
0.006040885578840971,
-0.030352378264069557,
0.11742591857910156,
0.000851878838147968,
-0.05391429737210274,
0.17059999704360962,
-0.13313265144824982,
-0.29761773347854614,
-0.14319129288196564,
-0.1334514170885086,
0.010339286178350449,
0.017076298594474792,
0.08161275088787079,
-0.08419821411371231,
-0.034329503774642944,
0.09296482056379318,
0.05038211867213249,
-0.12507256865501404,
0.025089209899306297,
-0.04892471432685852,
0.06829681247472763,
-0.08489015698432922,
-0.07859387248754501,
-0.06656183302402496,
-0.05509684979915619,
-0.04375450685620308,
0.11071989685297012,
-0.12806063890457153,
0.08490876853466034,
0.13819533586502075,
0.031051602214574814,
0.07859103381633759,
0.001322271185927093,
0.13857795298099518,
-0.061774931848049164,
-0.08148285746574402,
0.23385483026504517,
-0.009697506204247475,
0.10181795805692673,
0.10921594500541687,
0.02137605845928192,
-0.06395251303911209,
-0.022054238244891167,
-0.07990756630897522,
-0.12127618491649628,
-0.2081342488527298,
-0.13152581453323364,
-0.147438645362854,
0.0003129298856947571,
0.0026985483709722757,
0.04901242256164551,
0.08136387914419174,
0.07752165198326111,
0.05862151086330414,
-0.08998161554336548,
-0.05513061210513115,
0.057605329900979996,
0.2182815968990326,
-0.01885073445737362,
0.07730096578598022,
-0.08762159198522568,
-0.0858968198299408,
0.07550632208585739,
0.0809127613902092,
0.22168903052806854,
0.0703313797712326,
0.03154223784804344,
0.06092943996191025,
0.1844511479139328,
0.15529003739356995,
0.17153839766979218,
0.015418988652527332,
-0.025812845677137375,
0.006589105818420649,
0.013138419017195702,
-0.09150379151105881,
-0.0042123449966311455,
0.1292845457792282,
-0.14540119469165802,
-0.07880084216594696,
-0.22943291068077087,
0.08157958835363388,
0.07344048470258713,
0.026414161548018456,
-0.17556782066822052,
-0.0015733868349343538,
0.05357646942138672,
0.006558762397617102,
-0.045512180775403976,
0.08496551215648651,
-0.022116927430033684,
-0.13344039022922516,
0.04287152364850044,
-0.058599162846803665,
0.10456542670726776,
-0.038338202983140945,
0.07373929768800735,
-0.024396726861596107,
-0.10900583118200302,
0.07474620640277863,
0.08851087838411331,
-0.23624686896800995,
0.27773332595825195,
-0.028285326436161995,
-0.07041611522436142,
-0.06830666214227676,
-0.024676870554685593,
-0.0017557048704475164,
0.17030403017997742,
0.08692118525505066,
0.03860539570450783,
-0.06731099635362625,
-0.1669522523880005,
0.04732859507203102,
0.02085104025900364,
0.12389091402292252,
-0.008113703690469265,
-0.030854588374495506,
-0.01338487584143877,
-0.019302181899547577,
-0.041759077459573746,
-0.003546249819919467,
0.10703662782907486,
-0.13412658870220184,
0.034535422921180725,
0.005309549160301685,
0.04036608338356018,
-0.0035966450814157724,
0.005690612830221653,
-0.05699379742145538,
0.13448239862918854,
-0.04070858657360077,
-0.07688865810632706,
-0.08865337073802948,
-0.17243140935897827,
0.13381044566631317,
-0.10238062590360641,
0.09249105304479599,
-0.09976985305547714,
-0.06192123889923096,
-0.0650167390704155,
-0.21104402840137482,
0.10917928814888,
-0.10513100028038025,
0.0401403047144413,
-0.016867417842149734,
0.202356219291687,
-0.08008989691734314,
-0.0209506843239069,
-0.0013599899830296636,
0.004461138043552637,
-0.12345174700021744,
-0.10132946074008942,
-0.01779748499393463,
0.029711758717894554,
0.08279082924127579,
0.10770376026630402,
-0.018299801275134087,
0.04191994294524193,
0.013882452622056007,
0.03635424003005028,
0.24111443758010864,
0.13711073994636536,
-0.04645780101418495,
0.11812267452478409,
0.16186927258968353,
-0.031571563333272934,
-0.2709476351737976,
-0.0868004634976387,
-0.16869930922985077,
-0.04678983986377716,
-0.037189606577157974,
-0.12448763102293015,
0.12788352370262146,
0.04586293175816536,
-0.011544082313776016,
0.1266224980354309,
-0.26594868302345276,
-0.03401912376284599,
0.1643364280462265,
0.027830148115754128,
0.48583653569221497,
-0.09991659969091415,
-0.0845755785703659,
0.018966181203722954,
-0.2915274500846863,
0.10791198909282684,
0.01369303185492754,
0.06487411260604858,
-0.034079212695360184,
0.10230494290590286,
0.044344354420900345,
-0.10719364136457443,
0.11951436847448349,
0.035276297479867935,
0.03043992444872856,
-0.08126169443130493,
-0.08180311322212219,
0.043867211788892746,
-0.020314449444413185,
0.016103362664580345,
0.09817174077033997,
0.036888349801301956,
-0.1322462111711502,
-0.022591179236769676,
-0.13765397667884827,
0.049294743686914444,
0.0620623379945755,
-0.0240910816937685,
-0.013253096491098404,
-0.036911722272634506,
-0.004203932359814644,
0.019186172634363174,
0.25098612904548645,
-0.012744947336614132,
0.1315464824438095,
0.042376402765512466,
0.046371813863515854,
-0.20250053703784943,
-0.14301706850528717,
-0.07346653193235397,
-0.04091315343976021,
0.0825582891702652,
-0.04023446887731552,
0.0523495189845562,
0.1589740812778473,
-0.02755962871015072,
-0.008571553975343704,
0.1279488205909729,
0.014891285449266434,
-0.03108718805015087,
0.12318851798772812,
-0.21083922684192657,
-0.042206212878227234,
-0.018448445945978165,
-0.05149201303720474,
0.1258767545223236,
0.12660184502601624,
0.08982381969690323,
0.07138726115226746,
-0.016965264454483986,
-0.03962934389710426,
-0.023878682404756546,
-0.0698867067694664,
0.014531198889017105,
0.041909199208021164,
0.03934765234589577,
-0.13099561631679535,
0.07754374295473099,
-0.018086964264512062,
-0.27273428440093994,
-0.06408093869686127,
0.09851112216711044,
-0.15818077325820923,
-0.10192093253135681,
-0.06077379360795021,
0.03578856587409973,
-0.1768866330385208,
-0.0624033585190773,
-0.0316467322409153,
-0.10826759040355682,
0.06394574791193008,
0.20973356068134308,
0.10992377996444702,
0.12001070380210876,
-0.023827049881219864,
-0.02570408768951893,
-0.00579385831952095,
-0.10135087370872498,
-0.012352810241281986,
0.013960975222289562,
-0.09324274212121964,
0.018809983506798744,
-0.009077025577425957,
0.17904941737651825,
-0.0752621665596962,
-0.08742223680019379,
-0.1552542746067047,
0.09289714694023132,
-0.1066979393362999,
-0.092626191675663,
-0.14189116656780243,
-0.06335672736167908,
0.023829210549592972,
-0.06568517535924911,
-0.01769612915813923,
-0.013204174116253853,
-0.14357970654964447,
0.06973201036453247,
0.01112452708184719,
-0.011419128626585007,
-0.057545971125364304,
-0.02905389852821827,
0.12901906669139862,
-0.0439501516520977,
0.08620632439851761,
0.19515031576156616,
-0.09635450690984726,
0.12672042846679688,
-0.10680265724658966,
-0.16728007793426514,
0.109873466193676,
0.011463385075330734,
0.08858676254749298,
0.05595776438713074,
0.033014263957738876,
0.0949346125125885,
0.01360013522207737,
0.040659237653017044,
-0.0074751065112650394,
-0.13721472024917603,
0.0005411148304119706,
0.014451061375439167,
-0.14818744361400604,
-0.051526814699172974,
-0.10024616122245789,
0.15037848055362701,
0.046998120844364166,
0.10320950299501419,
0.02188979648053646,
0.13145636022090912,
-0.006807954981923103,
-0.01749216392636299,
-0.007913168519735336,
-0.18110167980194092,
0.058051977306604385,
-0.04888135939836502,
0.00664436025545001,
0.004136093892157078,
0.29094672203063965,
-0.05342133343219757,
0.06421013921499252,
0.012430429458618164,
0.05440686270594597,
0.025291895493865013,
0.019776156172156334,
0.23332199454307556,
0.0909920483827591,
-0.040104810148477554,
-0.07203774154186249,
0.09381944686174393,
-0.027183765545487404,
-0.051167625933885574,
0.14428386092185974,
0.12433760613203049,
0.046564988791942596,
0.10233303159475327,
-0.010628857649862766,
0.0502619631588459,
-0.11714580655097961,
-0.29435333609580994,
-0.0017536553787067533,
0.02923153154551983,
0.013082349672913551,
0.09705042093992233,
0.12465932965278625,
-0.014661809429526329,
0.09902279824018478,
-0.0014873039908707142,
-0.026829617097973824,
-0.13841837644577026,
-0.08308025449514389,
-0.037351641803979874,
-0.1296062022447586,
0.01537085510790348,
-0.06261784583330154,
-0.0013676333473995328,
0.20482417941093445,
0.040537029504776,
-0.03484100475907326,
0.11699468642473221,
0.0698540061712265,
-0.06982901692390442,
0.02712978608906269,
-0.00757982861250639,
0.035655613988637924,
0.06632175296545029,
-0.023149633780121803,
-0.14352689683437347,
-0.08878929167985916,
-0.04016351327300072,
0.04570772126317024,
-0.0825280100107193,
0.007398216053843498,
-0.1344616711139679,
-0.09811436384916306,
-0.05467144027352333,
0.09024309366941452,
-0.076502226293087,
0.12816135585308075,
-0.010830878280103207,
0.010613616555929184,
0.01268461812287569,
0.18772324919700623,
-0.075274258852005,
-0.060353536158800125,
0.002831645542755723,
0.23638813197612762,
0.08521586656570435,
0.08294188231229782,
0.00997002050280571,
0.024029415100812912,
-0.09174544364213943,
0.3311961889266968,
0.24609249830245972,
-0.010904072783887386,
0.04334266111254692,
0.04531332477927208,
0.04185611754655838,
0.12315969169139862,
0.13169057667255402,
0.11807206273078918,
0.344115674495697,
-0.08338891714811325,
-0.031944580376148224,
-0.030247550457715988,
0.006968794856220484,
-0.13217425346374512,
0.026531271636486053,
0.01758533902466297,
-0.08070562034845352,
-0.07679492235183716,
0.11993145197629929,
-0.20403271913528442,
0.15280351042747498,
0.08174222707748413,
-0.19685740768909454,
-0.03396262973546982,
-0.05789715424180031,
0.18384911119937897,
-0.014278654009103775,
0.11859142035245895,
-0.04448321461677551,
-0.13445675373077393,
0.08106808364391327,
0.048652783036231995,
-0.2895694077014923,
-0.08014793694019318,
0.0981469452381134,
0.042115915566682816,
-0.0287695974111557,
-0.015584125183522701,
0.03546717017889023,
0.059928759932518005,
0.08986898511648178,
-0.03891289234161377,
0.04509701579809189,
0.020416583865880966,
-0.10908327251672745,
-0.08358649909496307,
-0.02032429538667202,
-0.00014163613377604634,
-0.1349317878484726,
0.016364745795726776,
-0.20526979863643646,
0.04226694256067276,
-0.010092525742948055,
0.01465508621186018,
-0.004206623882055283,
-0.03162108361721039,
-0.051075391471385956,
0.023095570504665375,
0.052028998732566833,
0.009046600200235844,
-0.017174653708934784,
-0.043978288769721985,
-0.01736217923462391,
0.04129228740930557,
-0.09392837435007095,
-0.15831652283668518,
-0.016197707504034042,
-0.08371548354625702,
0.10397595167160034,
-0.03304345905780792,
-0.046627145260572433,
-0.017865123227238655,
-0.02954019047319889,
0.04380740970373154,
-0.10989689826965332,
0.035784848034381866,
0.01109471544623375,
0.05120820179581642,
0.011953367851674557,
-0.01127829309552908,
0.039223846048116684,
0.07814694195985794,
-0.12265297025442123,
-0.1024811714887619
] |
null | null |
transformers
|
deeqELECTRA-base
---
- model: electra-base-generator
- vocab: bert-wordpiece, 35k
- version: beta, 1.71M
|
{"language": "ko", "datasets": ["kowiki", "news"]}
|
fill-mask
|
baikal-nlp/delectra-generator
|
[
"transformers",
"pytorch",
"electra",
"fill-mask",
"ko",
"dataset:kowiki",
"dataset:news",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ko"
] |
TAGS
#transformers #pytorch #electra #fill-mask #ko #dataset-kowiki #dataset-news #autotrain_compatible #endpoints_compatible #region-us
|
deeqELECTRA-base
---
- model: electra-base-generator
- vocab: bert-wordpiece, 35k
- version: beta, 1.71M
|
[] |
[
"TAGS\n#transformers #pytorch #electra #fill-mask #ko #dataset-kowiki #dataset-news #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
50
] |
[
"passage: TAGS\n#transformers #pytorch #electra #fill-mask #ko #dataset-kowiki #dataset-news #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.08945022523403168,
0.09469184279441833,
-0.00544744310900569,
-0.011938394978642464,
0.1339956372976303,
0.036769524216651917,
0.07541871815919876,
0.08829065412282944,
0.07242495566606522,
-0.007063510362058878,
0.17927847802639008,
0.2036144882440567,
0.000080437574069947,
0.26667511463165283,
-0.0779055580496788,
-0.2790839672088623,
0.10528742522001266,
0.04948332533240318,
-0.004482520744204521,
0.11732222139835358,
0.08343388140201569,
-0.09635230153799057,
0.059322699904441833,
-0.037991948425769806,
-0.08566254377365112,
0.0214060228317976,
0.014820498414337635,
-0.14314740896224976,
0.10497933626174927,
0.0307004414498806,
0.12718378007411957,
0.016753828153014183,
-0.065365731716156,
-0.10807736963033676,
0.06748022139072418,
0.008399640209972858,
-0.06831727921962738,
0.04653378203511238,
-0.04615074023604393,
-0.06660252064466476,
0.007701727095991373,
-0.07800263166427612,
0.0013446775265038013,
0.009199841879308224,
-0.16438767313957214,
-0.06841466575860977,
-0.04857974871993065,
0.03067910671234131,
0.10226026177406311,
0.09428351372480392,
-0.00514158234000206,
0.20599287748336792,
-0.1073467880487442,
0.08192005753517151,
0.11763005703687668,
-0.2011757642030716,
-0.006908745039254427,
0.03353588283061981,
0.034044310450553894,
-0.054969314485788345,
-0.009662933647632599,
0.08221832662820816,
0.03737075254321098,
0.012342262081801891,
-0.05227592960000038,
-0.03821934014558792,
-0.0844496637582779,
0.017192725092172623,
-0.06418278813362122,
-0.08292409777641296,
0.22991178929805756,
0.0010946871479973197,
0.08159548789262772,
-0.016702990978956223,
-0.08105067908763885,
-0.03177386149764061,
-0.015412540175020695,
0.04184291884303093,
-0.08281049132347107,
0.016835728660225868,
-0.028169754892587662,
0.008470926433801651,
-0.1138976439833641,
0.07037439197301865,
-0.2519412040710449,
0.283344030380249,
0.022037552669644356,
0.07933630049228668,
-0.18124857544898987,
0.05335191264748573,
0.020139753818511963,
-0.1413121223449707,
0.020099909976124763,
-0.08689089864492416,
0.020781001076102257,
-0.006005162838846445,
-0.05804659426212311,
-0.041448578238487244,
0.10951117426156998,
0.191781684756279,
0.03177795931696892,
-0.005428448785096407,
0.0701182410120964,
0.08524885773658752,
0.12084157019853592,
0.010919817723333836,
-0.06695351004600525,
-0.05706183239817619,
0.05628511682152748,
-0.09923426061868668,
0.03885118290781975,
-0.06589213013648987,
-0.06457012891769409,
-0.03318939357995987,
-0.010899108834564686,
0.07632528990507126,
0.10091142356395721,
0.07740079611539841,
-0.08859304338693619,
-0.007711789105087519,
0.09513519704341888,
-0.07047536224126816,
-0.021135246381163597,
-0.04497170075774193,
-0.0033973809331655502,
0.15790951251983643,
-0.010806338861584663,
0.03957197815179825,
-0.021211592480540276,
0.11677152663469315,
-0.08099446445703506,
-0.04731383174657822,
-0.009281656704843044,
-0.023378310725092888,
0.07082030922174454,
-0.16559703648090363,
0.07660870999097824,
-0.1809758096933365,
-0.19660508632659912,
0.02412254549562931,
0.01566077396273613,
-0.007284882944077253,
-0.0765894278883934,
-0.00002689245593501255,
0.005486934445798397,
0.04694140702486038,
-0.04848560318350792,
-0.04304460063576698,
-0.09642714262008667,
0.10246648639440536,
-0.029831307008862495,
0.10464096814393997,
-0.11054764688014984,
0.02678924798965454,
-0.0996926873922348,
-0.015903780236840248,
-0.07017775624990463,
0.020136775448918343,
-0.06062036380171776,
0.14000412821769714,
-0.04012415185570717,
-0.00616095308214426,
-0.06847041845321655,
0.03039764054119587,
-0.009765935130417347,
0.22395838797092438,
-0.10373217612504959,
-0.0879935771226883,
0.18657682836055756,
-0.10594140738248825,
-0.11396530270576477,
0.0616898275911808,
-0.02245917171239853,
0.026441849768161774,
0.11604243516921997,
0.016392074525356293,
0.0330052375793457,
-0.10307782888412476,
-0.03645363822579384,
0.024728748947381973,
-0.13172294199466705,
-0.07304394990205765,
0.019210658967494965,
0.0501626655459404,
-0.07632888108491898,
0.0551467128098011,
0.12358292192220688,
0.10772203654050827,
-0.09631315618753433,
-0.05834588035941124,
0.005279323551803827,
-0.04826964810490608,
0.147897407412529,
0.020128564909100533,
0.05857541039586067,
-0.05822937190532684,
-0.017169561237096786,
-0.03189416602253914,
0.06693248450756073,
0.07134982943534851,
0.02741786278784275,
-0.12067407369613647,
0.10956791043281555,
-0.1110401526093483,
0.024185094982385635,
-0.1489638239145279,
-0.12635266780853271,
-0.060872822999954224,
0.09259394556283951,
-0.051146939396858215,
0.07253662496805191,
0.06082327291369438,
-0.10866894572973251,
-0.016735684126615524,
-0.016938069835305214,
0.09346972405910492,
0.04877755045890808,
-0.03511624038219452,
-0.09708679467439651,
0.05948728695511818,
-0.08413491398096085,
0.03660692647099495,
-0.04296233877539635,
-0.016453178599476814,
0.01991519145667553,
0.17133042216300964,
0.0009966345969587564,
0.04835543781518936,
0.016481561586260796,
0.022848809137940407,
-0.032634932547807693,
0.02228947915136814,
0.06200273334980011,
0.0013336434494704008,
-0.04173494875431061,
0.15502852201461792,
-0.08529306948184967,
0.2262526899576187,
0.15012940764427185,
-0.18914759159088135,
0.009388036094605923,
0.04079883545637131,
-0.025917625054717064,
-0.026463737711310387,
-0.033395443111658096,
-0.007115287706255913,
0.005177692044526339,
-0.005649993661791086,
0.07445460557937622,
0.013364114798605442,
-0.015573693439364433,
0.036365944892168045,
-0.05413363501429558,
-0.0636100023984909,
0.0545528344810009,
0.11707954853773117,
-0.1904010772705078,
0.17236946523189545,
0.16357001662254333,
0.03875313699245453,
0.1765155792236328,
-0.014281215146183968,
-0.0222131609916687,
0.01504357997328043,
-0.13074442744255066,
-0.04725085198879242,
0.06610721349716187,
-0.13017350435256958,
0.014775366522371769,
0.09875146299600601,
-0.020795268937945366,
0.05233115330338478,
-0.06804273277521133,
-0.09738625586032867,
-0.011309448629617691,
0.05234989523887634,
-0.07912050187587738,
0.14922286570072174,
0.039094988256692886,
0.08485693484544754,
0.00033592514228075743,
-0.02040981315076351,
0.10947591066360474,
0.021262239664793015,
-0.02081775851547718,
0.15777233242988586,
-0.11439332365989685,
-0.3376455008983612,
-0.08075632899999619,
-0.16222402453422546,
0.016957703977823257,
-0.000021776666471851058,
0.06359480321407318,
-0.07946541905403137,
-0.05614783242344856,
0.06035953015089035,
-0.01244366355240345,
0.04427495226264,
0.042672887444496155,
0.027683990076184273,
0.003624822711572051,
-0.061841096729040146,
-0.0650533065199852,
-0.05028583109378815,
-0.01773092709481716,
0.035423394292593,
0.16196542978286743,
-0.07763681560754776,
0.10798417776823044,
0.08819183707237244,
0.05106057971715927,
0.05988270789384842,
0.0022466289810836315,
0.19568777084350586,
-0.12468548119068146,
0.014652226120233536,
0.16847951710224152,
-0.053641751408576965,
0.0634000226855278,
0.1716448962688446,
0.004956644494086504,
-0.05284043774008751,
0.018686305731534958,
-0.017147734761238098,
-0.08980626612901688,
-0.20067335665225983,
-0.15081512928009033,
-0.08966787904500961,
0.025265607982873917,
0.02818814292550087,
0.03509571775794029,
0.0904943123459816,
0.099338598549366,
-0.011177106760442257,
-0.05693036690354347,
-0.06180834770202637,
0.025613484904170036,
0.16084639728069305,
0.0009841278661042452,
0.09433165192604065,
-0.06127087399363518,
-0.1013704240322113,
0.07090646773576736,
0.007382141426205635,
0.09282279014587402,
0.06589163094758987,
-0.0048706722445786,
0.038907893002033234,
0.08931881934404373,
0.1415986567735672,
0.12376667559146881,
0.06323128193616867,
-0.06471104174852371,
0.024386482313275337,
0.017631789669394493,
-0.045593664050102234,
0.0016797028947621584,
0.1016734316945076,
-0.12221305817365646,
-0.030106840655207634,
-0.11029495298862457,
0.049069661647081375,
0.13201239705085754,
0.1295263022184372,
-0.2500491440296173,
-0.03585957735776901,
0.042706772685050964,
-0.019233668223023415,
-0.08562559634447098,
0.017759542912244797,
-0.04423952475190163,
-0.1381727159023285,
0.13834436237812042,
-0.035690709948539734,
0.09231367707252502,
-0.04267742484807968,
0.016753843054175377,
-0.06270284950733185,
-0.09941867738962173,
-0.008080322295427322,
0.09349627792835236,
-0.2170199751853943,
0.2713991403579712,
-0.015351456589996815,
-0.021194996312260628,
-0.11669925600290298,
-0.06117769703269005,
0.03435112535953522,
0.10433366894721985,
0.08497482538223267,
0.04026385769248009,
-0.05434637516736984,
-0.12813089787960052,
-0.12005189806222916,
0.05664897710084915,
0.015128006227314472,
-0.06808829307556152,
0.014307762496173382,
0.017262842506170273,
-0.028128566220402718,
-0.022625397890806198,
0.0211841631680727,
-0.041372232139110565,
-0.13456518948078156,
0.08872047066688538,
0.0877063199877739,
0.0972851887345314,
0.01490853913128376,
-0.11828683316707611,
-0.08097472041845322,
0.08158627152442932,
0.007596856914460659,
-0.07925881445407867,
-0.12194264680147171,
-0.013176660984754562,
0.09605083614587784,
-0.11730846762657166,
0.07231040298938751,
-0.02404794842004776,
0.04385833069682121,
-0.04645179957151413,
-0.1537965089082718,
0.11301206797361374,
-0.1297319233417511,
-0.07629324495792389,
-0.040587980300188065,
0.08943087607622147,
0.058809321373701096,
-0.0007121353410184383,
0.021327106282114983,
0.05365777388215065,
-0.06605909764766693,
-0.03187713027000427,
0.04383839666843414,
0.05575108155608177,
0.10305210202932358,
0.10476619005203247,
0.03846189007163048,
-0.07439907640218735,
-0.014814228750765324,
-0.04086712375283241,
0.23670543730258942,
0.24833233654499054,
-0.04097146913409233,
0.11428146064281464,
0.15834932029247284,
-0.016585703939199448,
-0.33135637640953064,
-0.12324673682451248,
-0.11441221833229065,
0.006385222543030977,
0.0002493450010661036,
-0.14371415972709656,
0.09404747933149338,
0.029589304700493813,
-0.053620558232069016,
0.15004044771194458,
-0.194366455078125,
-0.052244119346141815,
0.20156466960906982,
-0.08079393953084946,
0.48923933506011963,
-0.07901748269796371,
-0.04833360016345978,
-0.04014444351196289,
-0.1440175324678421,
0.14532649517059326,
0.017197106033563614,
0.07262591272592545,
-0.019473746418952942,
0.09958501905202866,
0.022387631237506866,
-0.06293530762195587,
0.14036235213279724,
-0.022227631881833076,
-0.035272594541311264,
-0.13881586492061615,
-0.09706734865903854,
0.03944593295454979,
-0.03542052209377289,
-0.009150535799562931,
0.02582891285419464,
0.01344470214098692,
-0.17208731174468994,
-0.018560199066996574,
-0.11525518447160721,
0.09499874711036682,
0.044766999781131744,
-0.02877744473516941,
0.0020462465472519398,
0.004605959169566631,
0.02862248197197914,
0.0022541170474141836,
0.2743716239929199,
0.02301267720758915,
0.12602588534355164,
-0.0321674682199955,
0.03197437897324562,
-0.10380402207374573,
-0.02284562587738037,
-0.09378140419721603,
-0.07896899431943893,
0.07251879572868347,
-0.04673517122864723,
0.027976347133517265,
0.11377329379320145,
-0.01810361072421074,
0.03325902670621872,
0.06473840028047562,
-0.04742177203297615,
0.036384567618370056,
0.13700222969055176,
-0.2263784259557724,
0.09390968084335327,
-0.014795217663049698,
0.03839574754238129,
0.045454807579517365,
0.05881965532898903,
0.1389266699552536,
0.02659880742430687,
-0.06577233970165253,
-0.007642126642167568,
0.04161500185728073,
-0.031115854158997536,
0.10748746991157532,
0.07483865320682526,
0.046146031469106674,
-0.1480826884508133,
0.08043364435434341,
-0.008528335019946098,
-0.21390211582183838,
0.027286307886242867,
0.058991145342588425,
-0.13727405667304993,
-0.11336571723222733,
0.007368563674390316,
0.04733426496386528,
-0.12301462888717651,
-0.07754939794540405,
-0.06834736466407776,
-0.06196823716163635,
0.10026999562978745,
0.2779799699783325,
0.06496275216341019,
0.06007421389222145,
0.013423804193735123,
-0.04653621464967728,
-0.023129913955926895,
0.014512763358652592,
0.09685757756233215,
-0.0020773587748408318,
-0.08206960558891296,
-0.061454370617866516,
-0.029543647542595863,
0.20809681713581085,
-0.10324312001466751,
-0.07617926597595215,
-0.11641049385070801,
0.04388409107923508,
-0.09831713140010834,
-0.043093618005514145,
-0.1205914095044136,
-0.06708821654319763,
-0.027740618214011192,
-0.10473961383104324,
-0.03708464652299881,
-0.04995434358716011,
-0.09853485226631165,
0.06185854971408844,
0.015142573043704033,
0.013241227716207504,
-0.05648554861545563,
-0.05320188030600548,
0.11046288907527924,
-0.02493291161954403,
0.12078601121902466,
0.14001977443695068,
-0.06299453973770142,
0.06605580449104309,
-0.0964084044098854,
-0.11069806665182114,
0.10615841299295425,
-0.010741674341261387,
0.06550449132919312,
0.003328179707750678,
0.018836138769984245,
0.047484610229730606,
0.03446529060602188,
0.06030402332544327,
0.005958247929811478,
-0.11634427309036255,
0.049364976584911346,
0.027984168380498886,
-0.12570755183696747,
-0.037568349391222,
-0.11426019668579102,
0.08441675454378128,
0.0016637525986880064,
0.16911590099334717,
-0.046711016446352005,
0.08683636784553528,
0.0012016627006232738,
0.01989162154495716,
-0.04370162636041641,
-0.16342830657958984,
0.05315817892551422,
0.004527917131781578,
0.03897529095411301,
-0.021119723096489906,
0.22775478661060333,
-0.058506280183792114,
-0.04321104660630226,
0.029671164229512215,
0.05786530300974846,
-0.0673660859465599,
0.0220213383436203,
0.16410943865776062,
0.08564484864473343,
-0.03976158797740936,
-0.10707764327526093,
0.07579051703214645,
0.02226414531469345,
-0.03218824043869972,
0.12114710360765457,
0.02440149523317814,
0.15366609394550323,
0.08260481804609299,
-0.014393418096005917,
0.019859815016388893,
-0.041856810450553894,
-0.25090929865837097,
-0.14024628698825836,
0.052552480250597,
0.07101918756961823,
-0.026439888402819633,
0.13512353599071503,
-0.006540799513459206,
-0.003251515794545412,
-0.05053744837641716,
-0.009918971918523312,
-0.13100628554821014,
-0.21971875429153442,
-0.08883108198642731,
-0.06216869503259659,
0.03791658952832222,
-0.051687583327293396,
0.012687494046986103,
0.20989219844341278,
0.03767302632331848,
-0.05330893024802208,
0.09254827350378036,
0.0666213408112526,
-0.029266752302646637,
0.086605966091156,
-0.013678993098437786,
0.0010718057164922357,
0.013475845567882061,
-0.024582164362072945,
-0.10675027966499329,
0.039133328944444656,
-0.04091501608490944,
0.02506875991821289,
-0.06327411532402039,
0.06944433599710464,
-0.12574703991413116,
-0.1526682823896408,
-0.05865664780139923,
0.04589546471834183,
-0.0008699606987647712,
0.026587016880512238,
0.03900213912129402,
0.09715136140584946,
0.05061738193035126,
0.1819881796836853,
-0.008542864583432674,
-0.11702623963356018,
-0.09638050198554993,
0.013269738294184208,
0.046197060495615005,
0.07281649857759476,
-0.010854450985789299,
-0.02949925884604454,
-0.0761200338602066,
0.25268030166625977,
0.2488960325717926,
-0.00638351496309042,
0.06235643848776817,
-0.01998085528612137,
0.03835797309875488,
0.060037851333618164,
0.07112005352973938,
0.055342212319374084,
0.21126949787139893,
-0.0945218950510025,
-0.07073455303907394,
-0.04747384041547775,
-0.02614830620586872,
-0.10744013637304306,
-0.03551323339343071,
0.014242861419916153,
-0.050023194402456284,
-0.04907369241118431,
0.14387762546539307,
-0.21460503339767456,
0.10986596345901489,
-0.02443401888012886,
-0.20865903794765472,
-0.10485220700502396,
-0.03558766096830368,
0.12380632013082504,
0.06777847558259964,
0.08570142090320587,
-0.02739478088915348,
-0.059983525425195694,
0.05168488249182701,
0.032529398798942566,
-0.151872918009758,
-0.09506386518478394,
0.07865358144044876,
0.0489247627556324,
0.04523346573114395,
-0.022167213261127472,
0.05374417454004288,
0.08784094452857971,
0.01838764175772667,
-0.06151818484067917,
0.029417814686894417,
0.03414002060890198,
-0.016982287168502808,
-0.0037526783999055624,
0.0319087877869606,
0.00958511047065258,
-0.10764974355697632,
0.10296021401882172,
-0.11532720178365707,
0.04645755887031555,
-0.06604550778865814,
0.026484079658985138,
-0.007056414615362883,
0.12062019854784012,
-0.022676043212413788,
0.08841816335916519,
0.04149741679430008,
-0.044162508100271225,
-0.014551916159689426,
-0.028531508520245552,
0.01238192804157734,
0.029374241828918457,
-0.12078232318162918,
-0.14003893733024597,
-0.11320241540670395,
-0.07643326371908188,
-0.020886575803160667,
0.04209854081273079,
-0.08571948111057281,
0.0085142208263278,
-0.14670081436634064,
-0.017833665013313293,
-0.0768129751086235,
0.06384344398975372,
0.03953278437256813,
-0.005664869677275419,
-0.010227316059172153,
0.015367316082119942,
0.05427948758006096,
0.05256257578730583,
-0.19290781021118164,
-0.1018771231174469
] |
null | null |
transformers
|
deeqELECTRA-base
---
- model: electra-base-discriminator
- vocab: bert-wordpiece, 35k
- version: beta, 1.71M
|
{"language": "ko", "datasets": ["kowiki", "news"]}
| null |
baikal-nlp/delectra
|
[
"transformers",
"pytorch",
"electra",
"pretraining",
"ko",
"dataset:kowiki",
"dataset:news",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ko"
] |
TAGS
#transformers #pytorch #electra #pretraining #ko #dataset-kowiki #dataset-news #endpoints_compatible #region-us
|
deeqELECTRA-base
---
- model: electra-base-discriminator
- vocab: bert-wordpiece, 35k
- version: beta, 1.71M
|
[] |
[
"TAGS\n#transformers #pytorch #electra #pretraining #ko #dataset-kowiki #dataset-news #endpoints_compatible #region-us \n"
] |
[
40
] |
[
"passage: TAGS\n#transformers #pytorch #electra #pretraining #ko #dataset-kowiki #dataset-news #endpoints_compatible #region-us \n"
] |
[
-0.08866488933563232,
0.1001894474029541,
-0.006150374189019203,
-0.060447171330451965,
0.12439315021038055,
0.04013713449239731,
0.038094088435173035,
0.07374319434165955,
0.03615489602088928,
-0.019297154620289803,
0.17840342223644257,
0.17942887544631958,
-0.01183032151311636,
0.16572517156600952,
-0.06204508617520332,
-0.2906045615673065,
0.10825780034065247,
0.07841114699840546,
0.0002367713168496266,
0.12388582527637482,
0.08479385823011398,
-0.11842358857393265,
0.03890044242143631,
-0.02634263224899769,
-0.09373777359724045,
0.011124073527753353,
-0.03324132040143013,
-0.13285107910633087,
0.10874505341053009,
0.0077508362010121346,
0.12151813507080078,
0.026232242584228516,
-0.06847036629915237,
-0.1741611808538437,
0.060754355043172836,
0.013227205723524094,
-0.05818893760442734,
0.02816026285290718,
-0.041554681956768036,
-0.08088735491037369,
0.05937597528100014,
-0.0905015617609024,
-0.016058821231126785,
0.0001594800123712048,
-0.20993070304393768,
-0.07444562762975693,
-0.06360416859388351,
0.03871499374508858,
0.12703120708465576,
0.11355919390916824,
-0.016369860619306564,
0.16234900057315826,
-0.13498912751674652,
0.05082584545016289,
0.11363077908754349,
-0.20802737772464752,
-0.015978239476680756,
0.043531253933906555,
0.05239573493599892,
0.0031411119271069765,
-0.024853048846125603,
0.06011534854769707,
0.04013470560312271,
0.0041681756265461445,
-0.11755616217851639,
-0.047367874532938004,
-0.06665540486574173,
0.058828409761190414,
-0.10719751566648483,
-0.08822409063577652,
0.2897641360759735,
0.00946649257093668,
0.07797408849000931,
0.0015678900526836514,
-0.02910081297159195,
-0.04497605189681053,
-0.003345517674461007,
0.03250109776854515,
-0.08013258874416351,
0.030261147767305374,
-0.014135098084807396,
0.007710546255111694,
-0.11688029021024704,
0.0692024752497673,
-0.2096334546804428,
0.30052125453948975,
0.020186005160212517,
0.06978397816419601,
-0.18386401236057281,
0.08259247243404388,
0.04969153553247452,
-0.0943363606929779,
-0.01058072317391634,
-0.10921363532543182,
0.02795662172138691,
-0.021625159308314323,
-0.0603756308555603,
0.0003016259288415313,
0.08495283126831055,
0.14610041677951813,
-0.01889312081038952,
-0.010119358077645302,
0.08949912339448929,
0.08903956413269043,
0.136122927069664,
0.0017452775500714779,
-0.08428008109331131,
-0.014978110790252686,
0.014271899126470089,
-0.07459258288145065,
0.007145778741687536,
-0.027815081179142,
-0.027486436069011688,
-0.05412692949175835,
-0.0071265664882957935,
0.08297624439001083,
0.07184906303882599,
0.047746121883392334,
-0.08692894130945206,
-0.02041407860815525,
0.030736614018678665,
-0.03615979850292206,
-0.05784524604678154,
-0.034335099160671234,
-0.038832928985357285,
0.17591796815395355,
-0.033662330359220505,
0.06116242706775665,
-0.014076992869377136,
0.11254949867725372,
-0.0847001001238823,
-0.03163208067417145,
-0.009223722852766514,
0.020923810079693794,
0.07545047998428345,
-0.16568604111671448,
0.10894381999969482,
-0.13413947820663452,
-0.1691903918981552,
-0.00033020228147506714,
0.027784021571278572,
0.019611768424510956,
-0.04899667575955391,
-0.03701157495379448,
-0.013195645064115524,
0.005572630558162928,
-0.05772062763571739,
-0.031207626685500145,
-0.11285378783941269,
0.104590505361557,
-0.05161885917186737,
0.042509254068136215,
-0.06619253754615784,
0.031003540381789207,
-0.10320795327425003,
0.004726796876639128,
-0.014337406493723392,
0.042520564049482346,
-0.07301434874534607,
0.1540137678384781,
-0.03155713155865669,
-0.02415621094405651,
-0.09041694551706314,
0.036822106689214706,
-0.01554151251912117,
0.23735423386096954,
-0.08386287093162537,
-0.07060286402702332,
0.1915646642446518,
-0.10029807686805725,
-0.12160634994506836,
0.026571325957775116,
-0.03756975382566452,
0.06646990031003952,
0.12120604515075684,
0.06987762451171875,
0.043157365173101425,
-0.07954062521457672,
-0.05542172119021416,
0.048146720975637436,
-0.12060365825891495,
-0.08577928692102432,
0.024371065199375153,
0.052429258823394775,
-0.05948444828391075,
0.03545749559998512,
0.08021360635757446,
0.08707335591316223,
-0.11882534623146057,
-0.05474222078919411,
0.01696370169520378,
-0.03870837390422821,
0.09813933074474335,
0.05551296845078468,
0.04643072932958603,
-0.05409211292862892,
-0.023988567292690277,
0.021627487614750862,
0.08399803936481476,
0.05344989150762558,
0.031091291457414627,
-0.1174476146697998,
0.03941608965396881,
-0.10467683523893356,
-0.002536518732085824,
-0.16007541120052338,
-0.10366136580705643,
-0.0761348307132721,
0.12288400530815125,
-0.056736789643764496,
0.1314312368631363,
0.05093798786401749,
-0.14577893912792206,
-0.01877179555594921,
-0.018962351605296135,
0.07208076864480972,
0.039652153849601746,
-0.038401149213314056,
-0.07741352915763855,
0.0610179528594017,
-0.08268357068300247,
0.024735158309340477,
-0.09682396054267883,
-0.010499270632863045,
0.09779033064842224,
0.162721648812294,
0.007344962563365698,
0.04030090942978859,
0.046482738107442856,
0.031924083828926086,
-0.004269023425877094,
0.01810614950954914,
0.07460787892341614,
0.012369646690785885,
-0.042649220675230026,
0.1099444106221199,
-0.02801518328487873,
0.19490323960781097,
0.13385826349258423,
-0.1904287487268448,
0.038217175751924515,
0.012495402246713638,
-0.05553959310054779,
-0.009004318155348301,
-0.008249884471297264,
0.022531120106577873,
0.04930974543094635,
0.00046320733963511884,
0.04138603433966637,
0.01051348727196455,
-0.03794756904244423,
0.010386962443590164,
-0.028379451483488083,
-0.06563954055309296,
0.08596991747617722,
0.0811939612030983,
-0.19100302457809448,
0.1494971364736557,
0.14506664872169495,
0.058083631098270416,
0.18210500478744507,
-0.0703374594449997,
-0.0312921516597271,
0.030180377885699272,
-0.10104575008153915,
-0.06937921047210693,
0.07918179780244827,
-0.1993582844734192,
0.003685299539938569,
0.07886437326669693,
0.011271371506154537,
0.08500898629426956,
-0.10199707001447678,
-0.1306275576353073,
-0.020055824890732765,
0.041312769055366516,
-0.05887464061379433,
0.12258787453174591,
0.016533263027668,
0.08174722641706467,
0.01995130442082882,
0.014081048779189587,
0.10915651172399521,
0.00643528439104557,
-0.0278211310505867,
0.17183555662631989,
-0.0967690497636795,
-0.25944939255714417,
-0.03814460337162018,
-0.08379175513982773,
0.04264919459819794,
-0.024628352373838425,
0.047706570476293564,
-0.09583724290132523,
-0.019832134246826172,
0.06275390833616257,
0.02945820800960064,
-0.059044841676950455,
0.009936729446053505,
0.05742841958999634,
0.03711541369557381,
-0.10871009528636932,
-0.08697406947612762,
-0.03254973143339157,
-0.053799185901880264,
0.010368646122515202,
0.1257038116455078,
-0.09489306807518005,
0.08227984607219696,
0.08506844937801361,
0.07964199036359787,
0.060623835772275925,
-0.006970998831093311,
0.21183468401432037,
-0.14518605172634125,
-0.014227252453565598,
0.13535885512828827,
-0.04924314096570015,
0.045864712446928024,
0.1300719976425171,
0.019877152517437935,
-0.0734746903181076,
0.009212099947035313,
-0.016051726415753365,
-0.06933365762233734,
-0.2500510513782501,
-0.1377379596233368,
-0.08157673478126526,
0.02182052843272686,
-0.0018125860951840878,
0.023853743448853493,
0.03330701217055321,
0.10127923637628555,
-0.013503890484571457,
-0.0995609238743782,
-0.03824061527848244,
0.016335416585206985,
0.15674231946468353,
-0.007200111169368029,
0.04879564419388771,
-0.07949770241975784,
-0.06011252477765083,
0.06592556834220886,
0.020544400438666344,
0.1549498736858368,
0.061184659600257874,
-0.005565785337239504,
0.08046486973762512,
0.0836021676659584,
0.1342782974243164,
0.08503198623657227,
0.028884844854474068,
-0.059352658689022064,
0.021406691521406174,
0.04111412167549133,
-0.06451087445020676,
0.005664248485118151,
0.04894893988966942,
-0.11816924065351486,
-0.034140694886446,
-0.14105629920959473,
0.04950304329395294,
0.1581350415945053,
0.11073867976665497,
-0.20082631707191467,
-0.05645742267370224,
0.04613431543111801,
-0.034240737557411194,
-0.08385936915874481,
0.07271674275398254,
-0.03168601542711258,
-0.1275593340396881,
0.13212762773036957,
-0.03129751980304718,
0.09523994475603104,
-0.10141143202781677,
0.01008534710854292,
-0.05006439983844757,
-0.15071965754032135,
-0.004757359158247709,
0.11256036162376404,
-0.22616954147815704,
0.2818765342235565,
-0.010085131973028183,
0.004597384016960859,
-0.10123831778764725,
-0.08415086567401886,
0.00020727109222207218,
0.14813239872455597,
0.10949230194091797,
0.04025816544890404,
-0.05384283885359764,
-0.07760398834943771,
-0.09028129279613495,
0.059249863028526306,
0.03743782266974449,
-0.08642854541540146,
0.005807411856949329,
0.035848211497068405,
0.014397531747817993,
-0.017826493829488754,
-0.05862791836261749,
-0.05678854137659073,
-0.12447112053632736,
0.07404486835002899,
0.032664429396390915,
0.1439041942358017,
0.004629978444427252,
-0.09620118141174316,
-0.02523675747215748,
0.044436972588300705,
-0.054625991731882095,
-0.08366875350475311,
-0.09697818756103516,
-0.021890195086598396,
0.09391140937805176,
-0.09290508925914764,
0.025931309908628464,
0.009118144400417805,
0.006148167420178652,
-0.03864838927984238,
-0.13632547855377197,
0.0941953957080841,
-0.12667132914066315,
-0.06793282181024551,
-0.02051422744989395,
0.12879687547683716,
0.11337229609489441,
0.005505455657839775,
0.0055315326899290085,
0.047430433332920074,
-0.05596158653497696,
-0.04953677952289581,
0.044018808752298355,
0.09097070246934891,
0.107326440513134,
0.07085950672626495,
0.051596976816654205,
0.005146549548953772,
-0.021610023453831673,
-0.05289594084024429,
0.2182169407606125,
0.22728095948696136,
-0.0269219521433115,
0.09656136482954025,
0.15560230612754822,
-0.028247682377696037,
-0.28756678104400635,
-0.053715452551841736,
-0.09908444434404373,
-0.016113167628645897,
-0.06687293946743011,
-0.19039206206798553,
0.11405033618211746,
0.08641795814037323,
-0.03512104973196983,
0.1320241391658783,
-0.283554345369339,
-0.022981584072113037,
0.14047576487064362,
-0.08636163920164108,
0.5149708390235901,
-0.09908077120780945,
-0.04526948183774948,
0.028676457703113556,
-0.1840023547410965,
0.20456519722938538,
-0.001713674166239798,
0.04746244102716446,
-0.005345305427908897,
0.06398767977952957,
0.013204192742705345,
-0.057510387152433395,
0.16798464953899384,
0.04038552567362785,
-0.021774036809802055,
-0.11792420595884323,
-0.1257021576166153,
0.05870195850729942,
-0.03674972429871559,
-0.03639494255185127,
0.09162440150976181,
0.013118082657456398,
-0.28302299976348877,
-0.01782979816198349,
-0.12845690548419952,
0.06767357140779495,
0.03320840373635292,
-0.03156581521034241,
-0.023988818749785423,
0.030693519860506058,
0.017331628128886223,
0.024517137557268143,
0.33900079131126404,
0.004432205576449633,
0.0989844873547554,
-0.09909576922655106,
0.11449572443962097,
-0.11431682854890823,
0.0011378549970686436,
-0.09950437396764755,
-0.043062832206487656,
0.07159502059221268,
-0.09469681233167648,
-0.014834349974989891,
0.14274269342422485,
0.0005118318367749453,
0.007970434613525867,
0.061314474791288376,
-0.06193891540169716,
0.05056769773364067,
0.11263630539178848,
-0.23419193923473358,
0.0165556650608778,
-0.04137495532631874,
0.04997215047478676,
0.052875496447086334,
0.10409944504499435,
0.14684022963047028,
0.023855356499552727,
-0.05622277408838272,
-0.01090943906456232,
0.01640254445374012,
-0.05447955057024956,
0.08816587924957275,
0.09897074103355408,
0.03299010172486305,
-0.1277909129858017,
0.10550416260957718,
-0.0028600357472896576,
-0.24605636298656464,
0.013237998820841312,
0.06268247216939926,
-0.14639690518379211,
-0.1054236888885498,
-0.0707886815071106,
0.055687449872493744,
-0.14154836535453796,
-0.07940731197595596,
-0.03713587298989296,
-0.055927421897649765,
0.11936140060424805,
0.2941446006298065,
0.06779072433710098,
0.08065729588270187,
0.017211923375725746,
-0.01764872670173645,
-0.019709639251232147,
-0.00945199653506279,
0.08572161197662354,
-0.023291999474167824,
-0.0881083682179451,
-0.06219311058521271,
-0.024634597823023796,
0.21233820915222168,
-0.08243259787559509,
-0.10419163107872009,
-0.11536791920661926,
0.08913370966911316,
-0.10490844398736954,
-0.046347759664058685,
-0.13689292967319489,
-0.052739452570676804,
-0.04655458778142929,
-0.12389102578163147,
-0.044019777327775955,
-0.03667306527495384,
-0.10346183180809021,
0.09631573408842087,
0.01519155502319336,
0.022059829905629158,
-0.03639907389879227,
-0.04757150635123253,
0.09977974742650986,
-0.021920986473560333,
0.13273607194423676,
0.13998116552829742,
-0.03875371441245079,
0.08035758137702942,
-0.06020371615886688,
-0.08646000176668167,
0.095571368932724,
-0.026550335809588432,
0.043605636805295944,
0.024068819358944893,
0.005153859965503216,
0.03311137855052948,
0.02364659123122692,
0.07206354290246964,
-0.060264457017183304,
-0.11501025408506393,
0.03137452155351639,
0.0472058542072773,
-0.102691650390625,
-0.049546077847480774,
-0.10022532194852829,
0.1379154622554779,
0.01731984131038189,
0.16290345788002014,
-0.004659005906432867,
0.07368635386228561,
0.013608044013381004,
0.014906847849488258,
-0.023001890629529953,
-0.16194072365760803,
0.09715200215578079,
-0.011927344836294651,
0.04012919217348099,
-0.021449029445648193,
0.2464391142129898,
-0.07954084128141403,
-0.05286361277103424,
0.03638959303498268,
0.004545845557004213,
-0.08645450323820114,
0.05017009750008583,
0.21521754562854767,
0.09811919182538986,
-0.02886628918349743,
-0.11497249454259872,
0.07074683904647827,
0.006765465717762709,
-0.05474954470992088,
0.11835074424743652,
0.09249500930309296,
0.1446608006954193,
0.08519253879785538,
-0.04658905044198036,
0.017814237624406815,
-0.01806579902768135,
-0.18209891021251678,
-0.11334109306335449,
0.021137438714504242,
0.05571020767092705,
0.03474098816514015,
0.15173964202404022,
-0.0452810674905777,
-0.02003176137804985,
-0.04565053805708885,
-0.015976332128047943,
-0.10011588782072067,
-0.181745707988739,
-0.04991940036416054,
-0.0837533175945282,
0.04371960088610649,
-0.08708148449659348,
0.01020260713994503,
0.2650638520717621,
0.050455451011657715,
-0.04422101750969887,
0.09058862924575806,
0.1298118233680725,
-0.04009920731186867,
0.11966919898986816,
-0.021052775904536247,
0.008578636683523655,
-0.04852491244673729,
-0.025100665166974068,
-0.06737327575683594,
-0.013705304823815823,
-0.046247825026512146,
0.02734045498073101,
-0.08291443437337875,
0.028598720207810402,
-0.12625423073768616,
-0.1245940625667572,
-0.03923647478222847,
0.08036435395479202,
0.02192051149904728,
0.009682729840278625,
0.03044760227203369,
0.07119742035865784,
0.0626850351691246,
0.2378450334072113,
0.005987535696476698,
-0.0896935909986496,
-0.08239810913801193,
-0.027380095794796944,
0.0566927045583725,
0.08162489533424377,
-0.0026553578209131956,
-0.042219046503305435,
-0.0334118977189064,
0.24646754562854767,
0.18127885460853577,
-0.012384217232465744,
0.04196242243051529,
-0.008272115141153336,
0.047967709600925446,
0.07013598084449768,
0.05041511356830597,
0.07299497723579407,
0.1978265792131424,
-0.12485850602388382,
-0.1003735288977623,
-0.043132707476615906,
-0.00128555903211236,
-0.07281934469938278,
-0.0036088242195546627,
0.026937179267406464,
-0.05235768482089043,
-0.06527401506900787,
0.16500522196292877,
-0.21859101951122284,
0.06897067278623581,
-0.027173353359103203,
-0.20720039308071136,
-0.1143534854054451,
-0.05259517580270767,
0.10956604778766632,
0.050057053565979004,
0.09307611733675003,
-0.018070770427584648,
-0.11052370071411133,
0.03595934808254242,
0.06016707792878151,
-0.16142769157886505,
-0.11522421985864639,
0.10188258439302444,
0.09454882889986038,
0.01841798797249794,
-0.031879670917987823,
0.05027371272444725,
0.0767754539847374,
-0.0023882510140538216,
-0.07407539337873459,
-0.0004127607971895486,
0.06098610535264015,
0.007538266014307737,
-0.04925629124045372,
0.03122669644653797,
0.017980502918362617,
-0.044649943709373474,
0.11696828156709671,
-0.15416337549686432,
0.05240539088845253,
-0.02409253641963005,
0.004632086958736181,
-0.0313396155834198,
0.10697226226329803,
-0.024374866858124733,
0.10081922262907028,
0.04471706226468086,
-0.06058688089251518,
-0.017118200659751892,
-0.02368665300309658,
0.025228755548596382,
0.02138300985097885,
-0.05139831081032753,
-0.1016831174492836,
-0.1041494682431221,
-0.0810285359621048,
-0.005645643454045057,
0.03666940703988075,
-0.042906079441308975,
0.027150237932801247,
-0.0847737267613411,
0.00838587898761034,
-0.04261944070458412,
0.06509056687355042,
0.018341638147830963,
-0.000815448584035039,
-0.025315217673778534,
0.04033417999744415,
0.04842105880379677,
0.05043822154402733,
-0.17316852509975433,
-0.08632899075746536
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilgpt2-finetuned-amazon-reviews
This model was trained from scratch on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Framework versions
- Transformers 4.8.2
- Pytorch 1.9.0+cu102
- Datasets 1.9.0
- Tokenizers 0.10.3
|
{"tags": ["generated_from_trainer"], "datasets": [], "model_index": [{"name": "distilgpt2-finetuned-amazon-reviews", "results": [{"task": {"name": "Causal Language Modeling", "type": "text-generation"}}]}]}
|
text-generation
|
defex/distilgpt2-finetuned-amazon-reviews
|
[
"transformers",
"pytorch",
"tensorboard",
"gpt2",
"text-generation",
"generated_from_trainer",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# distilgpt2-finetuned-amazon-reviews
This model was trained from scratch on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Framework versions
- Transformers 4.8.2
- Pytorch 1.9.0+cu102
- Datasets 1.9.0
- Tokenizers 0.10.3
|
[
"# distilgpt2-finetuned-amazon-reviews\n\nThis model was trained from scratch on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Framework versions\n\n- Transformers 4.8.2\n- Pytorch 1.9.0+cu102\n- Datasets 1.9.0\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# distilgpt2-finetuned-amazon-reviews\n\nThis model was trained from scratch on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0",
"### Framework versions\n\n- Transformers 4.8.2\n- Pytorch 1.9.0+cu102\n- Datasets 1.9.0\n- Tokenizers 0.10.3"
] |
[
58,
29,
6,
12,
8,
3,
90,
35
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# distilgpt2-finetuned-amazon-reviews\n\nThis model was trained from scratch on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3.0### Framework versions\n\n- Transformers 4.8.2\n- Pytorch 1.9.0+cu102\n- Datasets 1.9.0\n- Tokenizers 0.10.3"
] |
[
-0.11099045723676682,
0.12037011235952377,
-0.002393969800323248,
0.08478933572769165,
0.17173370718955994,
0.020275171846151352,
0.15759943425655365,
0.1103266030550003,
-0.10376986116170883,
0.030056549236178398,
0.07457645982503891,
0.0751853957772255,
0.041977815330028534,
0.11375506967306137,
-0.02270570956170559,
-0.24841447174549103,
0.015115203335881233,
0.030717773362994194,
-0.05485021322965622,
0.12429385632276535,
0.09678345173597336,
-0.11116905510425568,
0.06134844943881035,
0.015508746728301048,
-0.2045329511165619,
-0.016577789559960365,
-0.004267003852874041,
-0.08690692484378815,
0.11687690019607544,
0.0013127659913152456,
0.1190585345029831,
0.012657158076763153,
0.10607926547527313,
-0.14198042452335358,
0.009059928357601166,
0.07042653113603592,
0.04754604026675224,
0.11389090120792389,
0.05873224139213562,
0.01536208763718605,
0.11809894442558289,
-0.08828943967819214,
0.09958479553461075,
0.034464627504348755,
-0.08150029927492142,
-0.18302634358406067,
-0.06181235611438751,
0.08302842825651169,
0.062035247683525085,
0.11665681004524231,
0.0005039243842475116,
0.17394936084747314,
-0.07288558781147003,
0.0835646241903305,
0.2235681563615799,
-0.25321951508522034,
-0.07545097917318344,
0.012259427458047867,
0.06840178370475769,
0.06602899730205536,
-0.09041863679885864,
-0.022278523072600365,
0.040643420070409775,
0.06236922740936279,
0.12120166420936584,
-0.011730416677892208,
-0.08756096661090851,
-0.023758098483085632,
-0.14418327808380127,
-0.03227549046278,
0.16230349242687225,
-0.013256334699690342,
-0.060040250420570374,
-0.08219250291585922,
-0.0770745575428009,
-0.13062448799610138,
-0.03857482969760895,
-0.021892404183745384,
0.02470523864030838,
-0.06145623326301575,
-0.07942144572734833,
-0.0727544054389,
-0.08715546131134033,
-0.06375191360712051,
-0.005196606740355492,
0.13179317116737366,
0.06064681336283684,
0.018649909645318985,
-0.06553477793931961,
0.13704344630241394,
0.05577528849244118,
-0.13159047067165375,
-0.002454380737617612,
-0.015898656100034714,
-0.035663798451423645,
-0.04036456346511841,
-0.043768104165792465,
-0.03725871816277504,
-0.0006493715918622911,
0.15734410285949707,
-0.05828511714935303,
0.037415411323308945,
0.036157820373773575,
0.034895144402980804,
-0.03825036436319351,
0.1924743354320526,
-0.039164282381534576,
0.0040627047419548035,
0.02761903591454029,
0.0899597555398941,
0.015737539157271385,
-0.03334321081638336,
-0.1135217696428299,
-0.031999967992305756,
0.1299893707036972,
0.06767009943723679,
-0.011527100577950478,
0.023651674389839172,
-0.03188548609614372,
-0.049605004489421844,
0.03725777938961983,
-0.11665589362382889,
0.020299794152379036,
-0.015447595156729221,
-0.08444858342409134,
0.027336016297340393,
0.022410685196518898,
-0.012422374449670315,
-0.07750961184501648,
0.0805559754371643,
-0.0891975536942482,
0.01097088772803545,
-0.09615034610033035,
-0.0750085785984993,
0.014722630381584167,
-0.08446814864873886,
-0.0008497433736920357,
-0.10238420218229294,
-0.1730283796787262,
-0.026972195133566856,
0.025043044239282608,
-0.05085892602801323,
-0.06893544644117355,
-0.017880266532301903,
-0.08856775611639023,
0.012664321810007095,
-0.01796860620379448,
0.12333031743764877,
-0.03841160610318184,
0.09266047924757004,
0.036859385669231415,
0.023403160274028778,
0.002810658188536763,
0.038464393466711044,
-0.10087404400110245,
0.014627537690103054,
-0.12655122578144073,
0.0630066841840744,
-0.06715485453605652,
0.027139937505126,
-0.1023605540394783,
-0.11941317468881607,
0.05906830355525017,
-0.018845494836568832,
0.08729022741317749,
0.13388942182064056,
-0.12640121579170227,
-0.055790334939956665,
0.11907211691141129,
-0.07388492673635483,
-0.07272210717201233,
0.10843145847320557,
-0.029347436502575874,
0.027583783492445946,
0.06372517347335815,
0.1250535249710083,
0.0951007753610611,
-0.08475225418806076,
0.005700032226741314,
0.026430746540427208,
0.07022035121917725,
-0.04289434105157852,
0.06019018962979317,
0.013836076483130455,
-0.002318813232704997,
0.043417543172836304,
-0.07273489981889725,
0.023046722635626793,
-0.09200747311115265,
-0.0792548805475235,
-0.0638352632522583,
-0.08650192618370056,
0.09161470830440521,
0.03445674106478691,
0.057202041149139404,
-0.04909686744213104,
-0.1102224811911583,
0.09936001151800156,
0.13093030452728271,
-0.0606146864593029,
0.0167640820145607,
-0.0661289319396019,
0.10858967900276184,
-0.07023947685956955,
-0.014381811022758484,
-0.22586531937122345,
-0.11478385329246521,
0.044146500527858734,
-0.06710705906152725,
0.04501078650355339,
-0.0011299480684101582,
0.04083479568362236,
0.08723297715187073,
-0.035664647817611694,
-0.03748292103409767,
-0.10057539492845535,
-0.03054000623524189,
-0.11252249777317047,
-0.1742369532585144,
-0.06256648153066635,
-0.024799564853310585,
0.10970477759838104,
-0.1808268129825592,
0.03186368569731712,
-0.032753586769104004,
0.11975372582674026,
0.020015187561511993,
-0.03379088640213013,
-0.006457616109400988,
0.06950052082538605,
-0.024815117940306664,
-0.09043747186660767,
0.062021851539611816,
0.01745607703924179,
-0.08391042798757553,
-0.029884085059165955,
-0.13572193682193756,
0.1118047907948494,
0.10563209652900696,
-0.012938945554196835,
-0.07793056964874268,
0.035481326282024384,
-0.07340465486049652,
-0.03726306930184364,
-0.06809324026107788,
0.017967652529478073,
0.21211177110671997,
0.028409792110323906,
0.15701767802238464,
-0.0782233327627182,
-0.07559330016374588,
0.022438151761889458,
-0.0014908452285453677,
0.018850211054086685,
0.08677536994218826,
0.09296596795320511,
-0.023159116506576538,
0.1026456356048584,
0.09930768609046936,
-0.07918192446231842,
0.14412659406661987,
-0.046904873102903366,
-0.07678164541721344,
-0.012578100897371769,
-0.033890049904584885,
-0.01136787049472332,
0.11793126910924911,
-0.1327589750289917,
-0.018258750438690186,
0.04022393003106117,
0.0009230789146386087,
0.048232998698949814,
-0.2078467160463333,
-0.0002573885431047529,
0.011154687963426113,
-0.01980346255004406,
-0.005260861944407225,
0.004582362249493599,
0.019160142168402672,
0.09990879148244858,
0.02183280885219574,
-0.036683354526758194,
0.032212428748607635,
0.009614376351237297,
-0.08448401093482971,
0.2073284238576889,
-0.09408865123987198,
-0.17524592578411102,
-0.13668935000896454,
0.0346987247467041,
-0.07575218379497528,
-0.014792471192777157,
0.020534368231892586,
-0.08493906259536743,
-0.04254363477230072,
-0.0614914707839489,
0.0023311853874474764,
-0.042386122047901154,
0.007379225455224514,
-0.000862922752276063,
0.00983062107115984,
0.0699193924665451,
-0.12671881914138794,
0.005299692507833242,
-0.018361041322350502,
-0.06901311129331589,
0.004443356301635504,
0.07061363011598587,
0.09789596498012543,
0.17204882204532623,
-0.025260891765356064,
-0.006098140962421894,
-0.0256552342325449,
0.2574401795864105,
-0.09118257462978363,
-0.027911188080906868,
0.1650824248790741,
-0.008324145339429379,
0.053335823118686676,
0.09453760087490082,
0.05322746932506561,
-0.07185746729373932,
0.014522221870720387,
0.012010726146399975,
-0.022357255220413208,
-0.27684810757637024,
-0.04051082208752632,
-0.037007708102464676,
-0.03416982293128967,
0.09173454344272614,
0.05006946623325348,
0.027085816487669945,
0.07404962927103043,
-0.016622956842184067,
0.08800448477268219,
-0.02055496536195278,
0.10735547542572021,
0.11632910370826721,
0.03833032771945,
0.11392335593700409,
-0.029076755046844482,
-0.05758232995867729,
0.05274408310651779,
-0.008465644903481007,
0.2541004419326782,
-0.03678051009774208,
0.12495891004800797,
0.03026231750845909,
0.1303270310163498,
0.02184300310909748,
0.05603766813874245,
0.03361768648028374,
0.019830552861094475,
0.008539571426808834,
-0.04901501163840294,
-0.038964178413152695,
-0.01283507701009512,
-0.056296516209840775,
0.05053972080349922,
-0.12604568898677826,
0.006056960206478834,
0.015237605199217796,
0.26507067680358887,
0.018497418612241745,
-0.3196370303630829,
-0.11056765913963318,
-0.0003711419412866235,
-0.03409450873732567,
-0.07048651576042175,
0.01013867650181055,
0.08294694870710373,
-0.1506616324186325,
0.006698107812553644,
-0.08182256668806076,
0.0996883362531662,
-0.07655170559883118,
0.0212722048163414,
0.05527426302433014,
0.14638973772525787,
0.012202265672385693,
0.09093530476093292,
-0.2512217164039612,
0.20686152577400208,
0.0035964923445135355,
0.10557971894741058,
-0.06734680384397507,
0.0198601596057415,
0.018796412274241447,
0.06694718450307846,
0.06941843032836914,
-0.008214340545237064,
-0.025708716362714767,
-0.17920373380184174,
-0.09053121507167816,
0.029784219339489937,
0.08974705636501312,
-0.0194020327180624,
0.09123413264751434,
-0.057516951113939285,
-0.004268274176865816,
0.030143406242132187,
-0.11895624548196793,
-0.13034893572330475,
-0.11013510078191757,
0.045974329113960266,
0.00041142842383123934,
-0.02745022252202034,
-0.0822666585445404,
-0.12971048057079315,
-0.006676077842712402,
0.15665064752101898,
-0.035245683044195175,
-0.09072905778884888,
-0.13929083943367004,
0.06448975205421448,
0.13518251478672028,
-0.06737670302391052,
0.025205835700035095,
0.004407436586916447,
0.1682358831167221,
0.016551243141293526,
-0.08921852707862854,
0.043460335582494736,
-0.07457952946424484,
-0.21674233675003052,
-0.03122803196310997,
0.11946495622396469,
0.043082237243652344,
0.05383392050862312,
-0.01505478285253048,
0.012633388862013817,
-0.0420820452272892,
-0.0999467670917511,
-0.02077588252723217,
0.07381251454353333,
0.04558142274618149,
0.044350139796733856,
-0.04794176295399666,
0.012793928384780884,
-0.06987778842449188,
-0.0023877504281699657,
0.12928512692451477,
0.1841658353805542,
-0.07371335476636887,
0.059334587305784225,
0.09015113115310669,
-0.09191250801086426,
-0.18296992778778076,
0.029629340395331383,
0.10181194543838501,
0.012185204774141312,
0.011587575078010559,
-0.18508876860141754,
0.09748388081789017,
0.07390701770782471,
-0.0252989474684,
0.10365854948759079,
-0.35179343819618225,
-0.134714275598526,
0.08630252629518509,
0.1271468698978424,
0.07262849807739258,
-0.12103207409381866,
-0.03134603798389435,
-0.03022894263267517,
-0.08993499726057053,
0.10887959599494934,
-0.0843578651547432,
0.11114712059497833,
-0.03461300954222679,
0.1047532707452774,
0.02020353265106678,
-0.05665270611643791,
0.13724136352539062,
0.015911251306533813,
0.07560652494430542,
-0.04415731504559517,
0.014036279171705246,
0.10951414704322815,
-0.055884119123220444,
0.039648059755563736,
-0.016450736671686172,
0.06926028430461884,
-0.10428542643785477,
-0.03434180095791817,
-0.07802505046129227,
0.0643852949142456,
-0.05003170669078827,
-0.04989529401063919,
-0.027410488575696945,
0.027480125427246094,
0.016935527324676514,
-0.02492518723011017,
0.08830343931913376,
0.03993253409862518,
0.12485328316688538,
0.07952593266963959,
0.10692630708217621,
-0.02693822607398033,
-0.12779664993286133,
-0.0098744947463274,
-0.023621156811714172,
0.07274752110242844,
-0.11955323815345764,
0.002351194154471159,
0.1429092139005661,
0.036508601158857346,
0.10641063004732132,
0.06858283281326294,
-0.05891934409737587,
0.009755250997841358,
0.0271951574832201,
-0.13374383747577667,
-0.10389579832553864,
-0.04802119359374046,
-0.03456474468111992,
-0.12574917078018188,
0.03751322254538536,
0.12701742351055145,
-0.10346266627311707,
-0.034959398210048676,
0.0025905827060341835,
-0.011958650313317776,
-0.014368443749845028,
0.20499125123023987,
0.0581209659576416,
0.07755623757839203,
-0.10617132484912872,
0.12008891254663467,
0.08078933507204056,
-0.06614263355731964,
0.04229317978024483,
0.09102063626050949,
-0.10366147011518478,
-0.04541878029704094,
0.06270135939121246,
0.14166368544101715,
-0.09728450328111649,
-0.04622003063559532,
-0.12239214032888412,
-0.10205990076065063,
0.0588693767786026,
0.11260799318552017,
0.07026646286249161,
0.0007772683165967464,
-0.06533808261156082,
0.02428937517106533,
-0.15695558488368988,
0.10673534870147705,
0.06281646341085434,
0.08920533210039139,
-0.15884874761104584,
0.13213075697422028,
0.04041871801018715,
0.07290136069059372,
-0.032264791429042816,
0.012755840085446835,
-0.09500491619110107,
0.002678273944184184,
-0.1715008020401001,
-0.005149397999048233,
-0.010318618267774582,
0.00783539842814207,
-0.029090318828821182,
-0.03845886513590813,
-0.0368201918900013,
0.05309179425239563,
-0.08375347405672073,
-0.048615556210279465,
0.015242993831634521,
0.019367165863513947,
-0.10738801211118698,
0.01659347116947174,
0.012128637172281742,
-0.09990813583135605,
0.09062246233224869,
0.05211915075778961,
0.008149421773850918,
0.06171597167849541,
-0.061863772571086884,
-0.018208768218755722,
0.013016264885663986,
0.027665328234434128,
0.05609647184610367,
-0.0648374855518341,
0.006946940906345844,
-0.024331001564860344,
0.06046149507164955,
0.01155836507678032,
0.05357124283909798,
-0.14473944902420044,
-0.032012853771448135,
-0.017094742506742477,
-0.03519390895962715,
-0.06782950460910797,
0.04144710674881935,
0.07291465997695923,
0.0367211252450943,
0.16066570580005646,
-0.0720980241894722,
0.062408529222011566,
-0.1836191713809967,
-0.012016313150525093,
-0.01861635223031044,
-0.07210094481706619,
-0.09639175236225128,
-0.05889061093330383,
0.0850161537528038,
-0.06777656823396683,
0.12411577254533768,
0.002077406272292137,
0.16216622292995453,
0.027355685830116272,
0.010682894848287106,
-0.05057104676961899,
0.024402469396591187,
0.1657729148864746,
0.0658448189496994,
-0.004331564996391535,
0.07884923368692398,
0.012392827309668064,
0.07403598725795746,
0.059333957731723785,
0.17561709880828857,
0.09964580833911896,
-0.004725511185824871,
0.08545730262994766,
0.05303090438246727,
-0.05616316944360733,
-0.17486809194087982,
0.04127103462815285,
-0.017482610419392586,
0.1273883879184723,
-0.04910774156451225,
0.16522683203220367,
0.10165666788816452,
-0.12094931304454803,
0.04534142091870308,
-0.03693615645170212,
-0.0954604521393776,
-0.11641190946102142,
-0.06526294350624084,
-0.06780131161212921,
-0.11843560636043549,
0.007318213116377592,
-0.12205789983272552,
-0.014796783216297626,
0.06973350048065186,
0.011110982857644558,
-0.032498620450496674,
0.13479194045066833,
0.050368160009384155,
-0.01290856208652258,
0.09455160796642303,
0.016830237582325935,
-0.0019375209230929613,
-0.05907461419701576,
-0.05256468430161476,
0.012371813878417015,
0.03162768483161926,
0.07207052409648895,
-0.03647466376423836,
-0.04086019843816757,
0.04145433381199837,
-0.014687871560454369,
-0.08844850212335587,
0.03713410347700119,
0.02601313777267933,
0.04622674733400345,
0.03415621444582939,
0.015600777231156826,
-0.011868488974869251,
-0.038416873663663864,
0.29341450333595276,
-0.07774122804403305,
-0.09790351241827011,
-0.11537448316812515,
0.25979480147361755,
0.016677096486091614,
-0.0574553981423378,
0.07513298094272614,
-0.0926358625292778,
-0.0417090468108654,
0.2169264256954193,
0.17623364925384521,
-0.056774575263261795,
-0.03434871882200241,
0.015260674990713596,
-0.021600481122732162,
-0.058180149644613266,
0.1576320081949234,
0.1158575564622879,
0.08686144649982452,
-0.07828287035226822,
-0.0033827577717602253,
-0.022286877036094666,
-0.023321015760302544,
-0.07444993406534195,
0.08085305988788605,
0.022326702252030373,
-0.0009154744911938906,
-0.05026127025485039,
0.061382442712783813,
-0.08901741355657578,
-0.1289689689874649,
0.006926697213202715,
-0.1452370136976242,
-0.19062158465385437,
-0.01593099907040596,
0.07509029656648636,
-0.007597878109663725,
0.05777346342802048,
-0.008054244332015514,
-0.0024738681968301535,
0.14479540288448334,
-0.026948442682623863,
-0.060505613684654236,
-0.07782555371522903,
0.09984729439020157,
-0.06128077208995819,
0.19443021714687347,
-0.017764108255505562,
0.07744745910167694,
0.10322066396474838,
0.028976745903491974,
-0.15185131132602692,
0.021291090175509453,
0.05293850600719452,
-0.07045642286539078,
0.05448697507381439,
0.17265866696834564,
-0.012852421030402184,
0.019516751170158386,
0.029490431770682335,
-0.12374196201562881,
-0.0340978242456913,
-0.059403497725725174,
0.006048292387276888,
-0.06055331975221634,
-0.016928816214203835,
-0.06713613867759705,
0.13374635577201843,
0.2021428644657135,
-0.04404648765921593,
-0.017343878746032715,
-0.07922255992889404,
0.02920844592154026,
0.04598508030176163,
0.07051601260900497,
-0.02131902240216732,
-0.2281690090894699,
-0.001610948471352458,
0.04468563571572304,
0.007009547669440508,
-0.25234851241111755,
-0.054240912199020386,
0.03506018966436386,
-0.05702429264783859,
-0.08736881613731384,
0.09297676384449005,
0.035899654030799866,
0.03324124589562416,
-0.04182028770446777,
-0.046627458184957504,
-0.07993820309638977,
0.1484474092721939,
-0.1837136298418045,
-0.06156895309686661
] |
null | null |
transformers
|
# german-qg-t5-drink600
This model is fine-tuned in question generation in German. The expected answer must be highlighted with <hl> token. It is based on [german-qg-t5-quad](https://huggingface.co/dehio/german-qg-t5-quad) and further pre-trained on drink related questions.
## Task example
#### Input
generate question: Der Monk Sour Drink ist ein somit eine aromatische Überraschung,
die sowohl <hl>im Sommer wie auch zu Silvester<hl> funktioniert.
#### Expected Question
Zu welchen Gelegenheiten passt der Monk Sour gut?
## Model description
The model is based on [german-qg-t5-quad](https://huggingface.co/dehio/german-qg-t5-quad), which was pre-trained on [GermanQUAD](https://www.deepset.ai/germanquad). We further pre-trained it on questions annotated on drink receipts from [Mixology](https://mixology.eu/) ("drink600").
We have not yet open sourced the dataset, since we do not own copyright on the source material.
## Training and evaluation data
The training script can be accessed [here](https://github.com/d-e-h-i-o/german-qg).
## Evaluation
It achieves a **BLEU-4 score of 29.80** on the drink600 test set (n=120) and **11.30** on the GermanQUAD test set.
Thus, fine-tuning on drink600 did not affect performance on GermanQuAD.
In comparison, *german-qg-t5-quad* achieves a BLEU-4 score of **10.76** on the drink600 test set.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 2
- seed: 100
- gradient_accumulation_steps: 8
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Framework versions
- Transformers 4.13.0.dev0
- Pytorch 1.10.0+cu102
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"language": ["de"], "license": "mit", "tags": ["question generation"], "datasets": ["deepset/germanquad"], "widget": [{"text": "generate question: Der Monk Sour Drink ist ein somit eine aromatische \u00dcberraschung, die sowohl <hl>im Sommer wie auch zu Silvester<hl> funktioniert."}], "model-index": [{"name": "german-qg-t5-drink600", "results": []}]}
|
text2text-generation
|
dehio/german-qg-t5-drink600
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"question generation",
"de",
"dataset:deepset/germanquad",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# german-qg-t5-drink600
This model is fine-tuned in question generation in German. The expected answer must be highlighted with <hl> token. It is based on german-qg-t5-quad and further pre-trained on drink related questions.
## Task example
#### Input
generate question: Der Monk Sour Drink ist ein somit eine aromatische Überraschung,
die sowohl <hl>im Sommer wie auch zu Silvester<hl> funktioniert.
#### Expected Question
Zu welchen Gelegenheiten passt der Monk Sour gut?
## Model description
The model is based on german-qg-t5-quad, which was pre-trained on GermanQUAD. We further pre-trained it on questions annotated on drink receipts from Mixology ("drink600").
We have not yet open sourced the dataset, since we do not own copyright on the source material.
## Training and evaluation data
The training script can be accessed here.
## Evaluation
It achieves a BLEU-4 score of 29.80 on the drink600 test set (n=120) and 11.30 on the GermanQUAD test set.
Thus, fine-tuning on drink600 did not affect performance on GermanQuAD.
In comparison, *german-qg-t5-quad* achieves a BLEU-4 score of 10.76 on the drink600 test set.
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 2
- seed: 100
- gradient_accumulation_steps: 8
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Framework versions
- Transformers 4.13.0.dev0
- Pytorch 1.10.0+cu102
- Datasets 1.16.1
- Tokenizers 0.10.3
|
[
"# german-qg-t5-drink600\n\nThis model is fine-tuned in question generation in German. The expected answer must be highlighted with <hl> token. It is based on german-qg-t5-quad and further pre-trained on drink related questions.",
"## Task example",
"#### Input\n\ngenerate question: Der Monk Sour Drink ist ein somit eine aromatische Überraschung, \ndie sowohl <hl>im Sommer wie auch zu Silvester<hl> funktioniert.",
"#### Expected Question\nZu welchen Gelegenheiten passt der Monk Sour gut?",
"## Model description\n\nThe model is based on german-qg-t5-quad, which was pre-trained on GermanQUAD. We further pre-trained it on questions annotated on drink receipts from Mixology (\"drink600\"). \nWe have not yet open sourced the dataset, since we do not own copyright on the source material.",
"## Training and evaluation data\n\nThe training script can be accessed here.",
"## Evaluation\n\nIt achieves a BLEU-4 score of 29.80 on the drink600 test set (n=120) and 11.30 on the GermanQUAD test set. \nThus, fine-tuning on drink600 did not affect performance on GermanQuAD.\n\nIn comparison, *german-qg-t5-quad* achieves a BLEU-4 score of 10.76 on the drink600 test set.",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 100\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10",
"### Framework versions\n\n- Transformers 4.13.0.dev0\n- Pytorch 1.10.0+cu102\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# german-qg-t5-drink600\n\nThis model is fine-tuned in question generation in German. The expected answer must be highlighted with <hl> token. It is based on german-qg-t5-quad and further pre-trained on drink related questions.",
"## Task example",
"#### Input\n\ngenerate question: Der Monk Sour Drink ist ein somit eine aromatische Überraschung, \ndie sowohl <hl>im Sommer wie auch zu Silvester<hl> funktioniert.",
"#### Expected Question\nZu welchen Gelegenheiten passt der Monk Sour gut?",
"## Model description\n\nThe model is based on german-qg-t5-quad, which was pre-trained on GermanQUAD. We further pre-trained it on questions annotated on drink receipts from Mixology (\"drink600\"). \nWe have not yet open sourced the dataset, since we do not own copyright on the source material.",
"## Training and evaluation data\n\nThe training script can be accessed here.",
"## Evaluation\n\nIt achieves a BLEU-4 score of 29.80 on the drink600 test set (n=120) and 11.30 on the GermanQUAD test set. \nThus, fine-tuning on drink600 did not affect performance on GermanQuAD.\n\nIn comparison, *german-qg-t5-quad* achieves a BLEU-4 score of 10.76 on the drink600 test set.",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 100\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10",
"### Framework versions\n\n- Transformers 4.13.0.dev0\n- Pytorch 1.10.0+cu102\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
69,
63,
4,
43,
18,
75,
14,
84,
112,
36
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# german-qg-t5-drink600\n\nThis model is fine-tuned in question generation in German. The expected answer must be highlighted with <hl> token. It is based on german-qg-t5-quad and further pre-trained on drink related questions.## Task example#### Input\n\ngenerate question: Der Monk Sour Drink ist ein somit eine aromatische Überraschung, \ndie sowohl <hl>im Sommer wie auch zu Silvester<hl> funktioniert.#### Expected Question\nZu welchen Gelegenheiten passt der Monk Sour gut?## Model description\n\nThe model is based on german-qg-t5-quad, which was pre-trained on GermanQUAD. We further pre-trained it on questions annotated on drink receipts from Mixology (\"drink600\"). \nWe have not yet open sourced the dataset, since we do not own copyright on the source material.## Training and evaluation data\n\nThe training script can be accessed here.## Evaluation\n\nIt achieves a BLEU-4 score of 29.80 on the drink600 test set (n=120) and 11.30 on the GermanQUAD test set. \nThus, fine-tuning on drink600 did not affect performance on GermanQuAD.\n\nIn comparison, *german-qg-t5-quad* achieves a BLEU-4 score of 10.76 on the drink600 test set.### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 100\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10"
] |
[
-0.1008499413728714,
0.1059657633304596,
-0.003514438634738326,
0.06881430000066757,
0.04882475361227989,
-0.0015001831343397498,
0.13901764154434204,
0.11400715261697769,
0.07756145298480988,
0.09866126626729965,
0.011228497140109539,
-0.03495822846889496,
0.06460850685834885,
0.08191715180873871,
0.05753043293952942,
-0.1624564230442047,
0.02436206489801407,
-0.07829900830984116,
0.013039043173193932,
0.09727910906076431,
0.10987367480993271,
-0.10118450224399567,
0.04142424091696739,
-0.04399706423282623,
-0.049752529710531235,
0.057049937546253204,
-0.05178992822766304,
-0.020965291187167168,
0.07597857713699341,
0.008849256671965122,
0.060555245727300644,
0.02737482450902462,
0.04279303923249245,
-0.22342421114444733,
0.0035007281694561243,
0.02064204402267933,
0.02718668058514595,
0.013576400466263294,
0.05687201768159866,
-0.0510350726544857,
0.0790688768029213,
-0.032393500208854675,
0.03495313599705696,
0.04180702194571495,
-0.08780618757009506,
-0.1305565983057022,
-0.10637786239385605,
0.13303537666797638,
0.028277898207306862,
0.08967607468366623,
-0.06721213459968567,
0.03239757940173149,
-0.07253465056419373,
0.04317818954586983,
0.14307621121406555,
-0.2768181562423706,
-0.0320851132273674,
0.08346115797758102,
0.019123172387480736,
0.04803335666656494,
-0.13233323395252228,
-0.020927026867866516,
0.05357643961906433,
0.02057836763560772,
-0.04020370915532112,
-0.031474269926548004,
0.07865460216999054,
0.0091975973919034,
-0.1193748489022255,
-0.07090532779693604,
0.19551625847816467,
0.03473922237753868,
-0.11902040988206863,
-0.12217957526445389,
0.047128330916166306,
0.008087647147476673,
0.05483601614832878,
-0.06856374442577362,
0.001606380450539291,
-0.01916825771331787,
0.01776440627872944,
0.015235932543873787,
-0.11138685047626495,
-0.009738124907016754,
0.03876109421253204,
0.190757617354393,
0.04038408771157265,
0.004184911027550697,
0.0060037970542907715,
0.13049916923046112,
-0.02570568211376667,
-0.12015552073717117,
-0.07225329428911209,
-0.02964661456644535,
-0.14128486812114716,
-0.07121945172548294,
-0.043995629996061325,
-0.10320726782083511,
0.011232337914407253,
0.2051572948694229,
-0.016963517293334007,
0.05668426305055618,
0.03173527121543884,
0.015061440877616405,
0.07497823983430862,
0.2122112512588501,
0.01727754808962345,
-0.15720278024673462,
-0.06686338782310486,
0.04796459525823593,
-0.03580721095204353,
0.016033070161938667,
-0.054194048047065735,
-0.01622190698981285,
0.060564592480659485,
0.036124393343925476,
-0.014674856327474117,
0.009523454122245312,
-0.09659386426210403,
-0.034259431064128876,
-0.004587586969137192,
-0.1443401277065277,
0.031635090708732605,
0.0353926345705986,
-0.06984856724739075,
0.047793030738830566,
0.02681172825396061,
-0.02857881598174572,
-0.07300502061843872,
0.09448496252298355,
-0.01137927919626236,
-0.018623922020196915,
-0.10375957936048508,
-0.08480692654848099,
0.02062690258026123,
0.03201769292354584,
-0.053731080144643784,
-0.08195202052593231,
-0.10415144264698029,
-0.08423155546188354,
0.08952166140079498,
-0.04145732522010803,
0.00616719713434577,
-0.05489432066679001,
-0.020068900659680367,
-0.01460123062133789,
0.0040013352409005165,
-0.0687374696135521,
-0.016570020467042923,
0.03884703665971756,
-0.013961423188447952,
0.0487867146730423,
0.03978250175714493,
0.054715193808078766,
-0.1294359266757965,
0.012733491137623787,
-0.19864150881767273,
0.10300341993570328,
-0.10377386957406998,
-0.01642693020403385,
-0.13920539617538452,
-0.0774129256606102,
-0.0825057327747345,
-0.005404850468039513,
0.05659521371126175,
0.11675424873828888,
-0.1838199943304062,
-0.053002309054136276,
0.10460541397333145,
-0.10443691164255142,
-0.11028214544057846,
0.12879182398319244,
-0.05095858499407768,
0.048882778733968735,
0.08377618342638016,
0.1675783097743988,
0.14066845178604126,
-0.1604304313659668,
-0.12773925065994263,
-0.001210193382576108,
0.015111017972230911,
0.18338002264499664,
0.07271385937929153,
-0.07412625849246979,
-0.0061848764307796955,
0.01208496280014515,
-0.07773900777101517,
-0.040888648480176926,
-0.030394261702895164,
-0.05904850736260414,
-0.00860049668699503,
0.003418146865442395,
0.06070060282945633,
-0.0077638509683310986,
-0.02954871766269207,
-0.05799198895692825,
-0.10793588310480118,
0.11330772936344147,
0.07457296550273895,
-0.05006811395287514,
0.007884476333856583,
-0.0514022633433342,
0.07765009254217148,
-0.020264724269509315,
0.00023928011069074273,
-0.14795155823230743,
-0.05493656545877457,
0.08715216815471649,
-0.14892306923866272,
0.09004723280668259,
0.12725546956062317,
0.058472804725170135,
0.05729081109166145,
-0.01604508049786091,
-0.036484722048044205,
-0.14899231493473053,
-0.021727025508880615,
-0.020686263218522072,
-0.17315907776355743,
-0.06731607764959335,
-0.028299007564783096,
0.16644710302352905,
-0.09306299686431885,
-0.0212626364082098,
0.07062051445245743,
0.09487313032150269,
-0.005802851170301437,
-0.05824632942676544,
-0.03866929933428764,
0.020567532628774643,
-0.03133978322148323,
-0.03475566953420639,
-0.012014620937407017,
0.011261885985732079,
-0.058235518634319305,
-0.009570465423166752,
-0.048974305391311646,
-0.1380748599767685,
0.044249460101127625,
0.06959427148103714,
-0.12502646446228027,
-0.0923173576593399,
-0.0805506706237793,
-0.0265206228941679,
-0.06562260538339615,
-0.0827900841832161,
0.22678281366825104,
0.019687319174408913,
0.06053904443979263,
-0.07415667176246643,
-0.08396557718515396,
-0.03538264334201813,
-0.020666414871811867,
-0.024369176477193832,
0.10224483907222748,
-0.014579884707927704,
-0.05861222371459007,
0.06196564435958862,
0.13045741617679596,
0.03401573374867439,
0.18933142721652985,
-0.01581287570297718,
-0.09725869446992874,
-0.06762833148241043,
0.09289073199033737,
-0.03132878616452217,
0.15507829189300537,
-0.08609512448310852,
0.031902097165584564,
0.02396170049905777,
0.03455974534153938,
0.025287970900535583,
-0.08102913200855255,
0.05239931121468544,
0.03289631009101868,
-0.045048657804727554,
-0.0027578643057495356,
-0.019108625128865242,
0.021622108295559883,
0.08611006289720535,
0.047802988439798355,
0.03517637774348259,
-0.013148541562259197,
-0.05700483173131943,
-0.12433578819036484,
0.17247086763381958,
-0.06395397335290909,
-0.18511039018630981,
-0.11540835350751877,
0.09442088752985,
-0.0735848993062973,
-0.007232869043946266,
0.032219890505075455,
-0.03156739100813866,
-0.062482185661792755,
-0.05427560955286026,
0.11654213070869446,
-0.003289300948381424,
-0.03404577448964119,
-0.06629947572946548,
-0.0036393662448972464,
0.034424833953380585,
-0.1245424896478653,
-0.010280106216669083,
-0.0009610929409973323,
-0.06690546870231628,
-0.0006327463197521865,
0.018385915085673332,
0.06972606480121613,
0.09533199667930603,
-0.04212220013141632,
0.014626584947109222,
-0.01022787019610405,
0.24028277397155762,
-0.11897371709346771,
0.060662996023893356,
0.032267387956380844,
-0.03775877133011818,
0.036986250430345535,
0.10692787915468216,
0.013144822791218758,
-0.05639050900936127,
0.0553615465760231,
0.051514577120542526,
-0.0014773895964026451,
-0.24000728130340576,
-0.04054903984069824,
-0.04560565575957298,
0.010104778222739697,
0.0727568194270134,
0.04527178779244423,
-0.00925376545637846,
0.008313169702887535,
-0.05598924309015274,
-0.006575725041329861,
0.01050379779189825,
0.08740239590406418,
0.03572447597980499,
0.012304190546274185,
0.059557899832725525,
-0.056714314967393875,
-0.022183401510119438,
0.10563575476408005,
0.055305443704128265,
0.17699897289276123,
-0.006547089666128159,
0.07581020891666412,
0.09009556472301483,
0.10372628271579742,
-0.0317787230014801,
0.009603865444660187,
0.020762117579579353,
0.00676611065864563,
0.0055238027125597,
-0.09777145832777023,
0.015825970098376274,
0.0910453125834465,
0.039610885083675385,
0.03677910193800926,
-0.04664814472198486,
0.013858365826308727,
0.12176008522510529,
0.22375516593456268,
0.047499142587184906,
-0.12109259516000748,
-0.09344056993722916,
0.01726565696299076,
-0.045000120997428894,
-0.02129894122481346,
-0.0036806119605898857,
0.14120084047317505,
-0.1579195111989975,
0.05540575459599495,
0.0030932461377233267,
0.08028487116098404,
-0.025817153975367546,
-0.010130991227924824,
0.07078158855438232,
0.0478818453848362,
-0.017515191808342934,
0.12373799830675125,
-0.26597392559051514,
0.10477814078330994,
-0.00463753892108798,
0.0807204321026802,
-0.07655679434537888,
0.04386122524738312,
-0.043341465294361115,
-0.043779391795396805,
0.14212192595005035,
0.0432441271841526,
-0.04950276017189026,
0.0014004016993567348,
-0.019090944901108742,
0.008073081262409687,
0.13282433152198792,
-0.07181311398744583,
0.10839398205280304,
-0.028862178325653076,
0.028549326583743095,
-0.04493141174316406,
0.03611797094345093,
-0.15093232691287994,
-0.10062222927808762,
0.044330496340990067,
-0.10004857182502747,
-0.030461013317108154,
-0.050352562218904495,
-0.026233751326799393,
-0.10407107323408127,
0.23642757534980774,
-0.19802260398864746,
-0.021504536271095276,
-0.08642081916332245,
0.06049536168575287,
0.06910908222198486,
-0.06869754940271378,
-0.02069622278213501,
-0.005508654750883579,
0.09599965810775757,
-0.06771194189786911,
-0.07517195492982864,
0.018731648102402687,
-0.07136796414852142,
-0.1348918080329895,
-0.03637610748410225,
0.1264214813709259,
0.06529288738965988,
0.05332639440894127,
0.029554812237620354,
-0.0021700644865632057,
-0.014778925105929375,
-0.14358052611351013,
-0.0024336003698408604,
-0.021284028887748718,
-0.023142756894230843,
0.06971405446529388,
-0.10837447643280029,
-0.08054719865322113,
-0.060834888368844986,
-0.002854582155123353,
0.02362283505499363,
0.20361189544200897,
-0.09944235533475876,
0.09536027163267136,
0.10349876433610916,
-0.10927052795886993,
-0.20266149938106537,
0.07524287700653076,
0.08797626197338104,
0.04126607999205589,
0.022084994241595268,
-0.17561842501163483,
0.1287672072649002,
0.06187163665890694,
-0.01276449766010046,
-0.016026942059397697,
-0.28655463457107544,
-0.11715102195739746,
0.054458361119031906,
0.029063116759061813,
-0.10131577402353287,
-0.1030316948890686,
-0.06512317806482315,
-0.024692824110388756,
-0.099090576171875,
0.010812611319124699,
-0.041270531713962555,
0.08414551615715027,
0.0055583780631423,
-0.003972596954554319,
0.056779634207487106,
-0.044402677565813065,
0.12554705142974854,
0.13047583401203156,
0.06555184721946716,
-0.06102042272686958,
0.003489344846457243,
0.05821423977613449,
-0.017607569694519043,
0.16844826936721802,
-0.03742681443691254,
0.05326792225241661,
-0.12153560668230057,
-0.016760438680648804,
-0.05840351805090904,
0.10832082480192184,
-0.09056111425161362,
-0.05871915444731712,
-0.10849065333604813,
0.108059361577034,
0.06126425042748451,
0.01120478194206953,
-0.057283006608486176,
-0.05368912220001221,
0.018637116998434067,
0.08457141369581223,
0.10377831757068634,
0.14990633726119995,
-0.1172635406255722,
0.029971612617373466,
0.01788242533802986,
0.04833507165312767,
-0.06512279063463211,
0.04969235137104988,
0.16857370734214783,
0.04839885234832764,
0.19642966985702515,
-0.012236598879098892,
-0.13202325999736786,
-0.029859181493520737,
0.08529770374298096,
-0.14844299852848053,
-0.14728644490242004,
-0.052146896719932556,
-0.04068533703684807,
-0.07649659365415573,
-0.031150810420513153,
0.11460428684949875,
-0.06403005123138428,
0.004606325179338455,
-0.01936894655227661,
0.011029844172298908,
-0.06362445652484894,
0.11111380159854889,
0.013820742256939411,
0.0762864500284195,
-0.024612601846456528,
0.17177340388298035,
0.07119587808847427,
-0.07327183336019516,
0.05150202289223671,
0.03385750576853752,
-0.0412151999771595,
-0.008458978496491909,
-0.036733098328113556,
0.13026967644691467,
-0.024504559114575386,
-0.04724111407995224,
-0.08031941205263138,
-0.0748138576745987,
0.01671690121293068,
0.041341137140989304,
0.0554647296667099,
0.06820739060640335,
0.029979979619383812,
0.03756121173501015,
-0.06040450185537338,
0.10692306607961655,
0.0535583421587944,
0.01937592774629593,
-0.002295536920428276,
0.10435707122087479,
-0.023473994806408882,
0.0251340139657259,
-0.010862293653190136,
-0.06009636074304581,
-0.1046629548072815,
-0.023308169096708298,
-0.0762672945857048,
-0.019421180710196495,
0.0003296100185252726,
-0.010614125989377499,
0.0057439920492470264,
-0.04358365386724472,
0.01824209652841091,
0.05065393075346947,
-0.08381176739931107,
-0.03641512617468834,
-0.0027673267759382725,
0.09875981509685516,
-0.1789124608039856,
0.0068138353526592255,
0.06892146915197372,
-0.09099660813808441,
0.10383082181215286,
0.042710594832897186,
-0.012046423740684986,
0.0005618692375719547,
-0.11933375149965286,
0.017843125388026237,
-0.03549722582101822,
0.06360214948654175,
-0.004667368717491627,
-0.12658068537712097,
0.003102540038526058,
0.007815022021532059,
0.013228464871644974,
0.01956610567867756,
-0.011676725931465626,
-0.10444138944149017,
0.03253578022122383,
0.006768227554857731,
-0.09585538506507874,
-0.07137969136238098,
0.08602092415094376,
0.007951820269227028,
0.056227538734674454,
0.12794281542301178,
-0.08829368650913239,
0.03428056463599205,
-0.1473064124584198,
-0.04140175133943558,
0.04193275421857834,
0.03625922650098801,
-0.10673882067203522,
-0.025704074651002884,
0.03727880120277405,
-0.024945760145783424,
0.10212191194295883,
-0.0005037322407588363,
0.03125566989183426,
0.04720175266265869,
-0.04843756929039955,
-0.04729550704360008,
0.019200483337044716,
0.0631183534860611,
-0.04169108346104622,
0.002720012329518795,
-0.06317164748907089,
-0.0062036397866904736,
-0.0722779631614685,
0.025757264345884323,
0.1379365175962448,
0.22911755740642548,
0.027253655716776848,
0.025936314836144447,
0.0706772655248642,
-0.03548886254429817,
-0.10826034098863602,
0.08887680619955063,
0.012670818716287613,
0.0898505225777626,
-0.05306892469525337,
0.10585518181324005,
0.09469729661941528,
-0.20172643661499023,
0.12186878174543381,
-0.024856336414813995,
-0.0785379409790039,
-0.059588465839624405,
-0.09667699038982391,
-0.03887340426445007,
-0.061120133846998215,
0.03720187395811081,
-0.12514574825763702,
0.06521786749362946,
0.011710742488503456,
0.06833811104297638,
-0.06877510994672775,
0.20638354122638702,
-0.1440860778093338,
-0.05608302354812622,
0.11417284607887268,
-0.011906785890460014,
0.04441332072019577,
0.0027542393654584885,
0.020883433520793915,
0.010041878558695316,
0.02588825859129429,
0.0694580227136612,
0.05095015466213226,
0.05109211802482605,
-0.04982365295290947,
-0.017491944134235382,
-0.047733407467603683,
0.014112347736954689,
-0.04243002459406853,
0.011193281039595604,
0.13587887585163116,
0.06812673062086105,
-0.016847537830471992,
-0.02511139214038849,
0.19498644769191742,
-0.05971946939826012,
-0.1290428191423416,
-0.1621491014957428,
0.13179737329483032,
0.04948778823018074,
0.05923326313495636,
0.02418421395123005,
-0.11531219631433487,
0.015651412308216095,
0.09338156133890152,
0.05164649710059166,
-0.019836682826280594,
0.0026231964584439993,
0.0070231142453849316,
0.007598718628287315,
0.04915713891386986,
0.055472176522016525,
-0.015017162077128887,
0.1816876232624054,
-0.01917889155447483,
0.08882653713226318,
-0.04332611337304115,
-0.037667952477931976,
0.005109221674501896,
0.13916508853435516,
-0.025522824376821518,
-0.037619106471538544,
-0.1167181208729744,
0.0993085503578186,
-0.011243530549108982,
-0.27707570791244507,
0.009651356376707554,
-0.08184326440095901,
-0.12911398708820343,
-0.03525684028863907,
0.0640910267829895,
0.01579856313765049,
0.07933627814054489,
-0.002351246541365981,
-0.015433311462402344,
0.16479597985744476,
0.014097332954406738,
-0.018395965918898582,
-0.08518538624048233,
0.0773620754480362,
-0.09894417971372604,
0.18556958436965942,
0.02583440952003002,
0.14526133239269257,
0.09354256838560104,
0.0268259160220623,
-0.07609136402606964,
0.02024286612868309,
0.05398576706647873,
-0.10392299294471741,
-0.009448419325053692,
0.14275723695755005,
0.0014048186130821705,
0.14695914089679718,
0.0748455673456192,
-0.017319662496447563,
0.04425185173749924,
0.07838955521583557,
-0.05922972783446312,
-0.16902457177639008,
0.09091433137655258,
-0.046171434223651886,
0.14316526055335999,
0.19878505170345306,
-0.019316596910357475,
0.01757236011326313,
-0.08222320675849915,
0.06101043522357941,
0.02450731210410595,
0.08611664175987244,
0.006625915411859751,
-0.12749098241329193,
0.05478086695075035,
-0.08840746432542801,
0.023812253028154373,
-0.20424644649028778,
-0.10293872654438019,
0.07073462009429932,
0.0123169906437397,
0.0034204241819679737,
0.14885199069976807,
0.03041374683380127,
0.04356276988983154,
-0.030363820493221283,
-0.11448853462934494,
-0.01558514591306448,
0.10512465983629227,
-0.07810099422931671,
-0.05088153854012489
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# german-qg-t5-e2e-quad (Work in progress)
This model is a end-to-end question generation model in German. Given a text, it generates several questions about it. This model is a fine-tuned version of [valhalla/t5-base-e2e-qg](https://huggingface.co/valhalla/t5-base-e2e-qg) on the [GermanQuAD dataset from deepset](https://huggingface.co/datasets/deepset/germanquad).
## Model description
More information needed
## Training and evaluation data
Bleu_1: 0.196051
Bleu_2: 0.122380
Bleu_3: 0.079980
Bleu_4: 0.053672
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10.0
### Training results
### Framework versions
- Transformers 4.16.0.dev0
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"language": ["de"], "license": "mit", "tags": ["question generation"], "datasets": ["deepset/germanquad"], "widget": [{"text": "Naturschutzwarte haben auf der ostfriesischen Insel Wangerooge zwei seltene Kurzschn\u00e4uzige Seepferdchen entdeckt. Die Tiere seien vergangene Woche bei einer sogenannten Sp\u00fclsaumkontrolle entdeckt worden, bei der die Str\u00e4nde eigentlich nach M\u00fcll und toten V\u00f6geln abgesucht w\u00fcrden, sagte der Gesch\u00e4ftsf\u00fchrer der zust\u00e4ndigen Naturschutz- und Forschungsgemeinschaft Mellumrat, Mathias Heckroth. Dabei seien den Natursch\u00fctzern am Nordstrand kurz hintereinander die beiden leblosen, nur wenige Zentimeter gro\u00dfen Tiere aufgefallen. Experten der Nationalparkverwaltung bestimmten beide Tiere als Kurzschn\u00e4uzige Seepferdchen (Hippocampus hippocampus)."}], "inference": {"parameters": {"max_length": 128}}, "model-index": [{"name": "german-qg-t5-e2e-quad", "results": []}]}
|
text2text-generation
|
dehio/german-qg-t5-e2e-quad
|
[
"transformers",
"pytorch",
"tensorboard",
"t5",
"text2text-generation",
"question generation",
"de",
"dataset:deepset/germanquad",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#transformers #pytorch #tensorboard #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
|
# german-qg-t5-e2e-quad (Work in progress)
This model is a end-to-end question generation model in German. Given a text, it generates several questions about it. This model is a fine-tuned version of valhalla/t5-base-e2e-qg on the GermanQuAD dataset from deepset.
## Model description
More information needed
## Training and evaluation data
Bleu_1: 0.196051
Bleu_2: 0.122380
Bleu_3: 0.079980
Bleu_4: 0.053672
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 4
- eval_batch_size: 4
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 32
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10.0
### Training results
### Framework versions
- Transformers 4.16.0.dev0
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
[
"# german-qg-t5-e2e-quad (Work in progress)\n\nThis model is a end-to-end question generation model in German. Given a text, it generates several questions about it. This model is a fine-tuned version of valhalla/t5-base-e2e-qg on the GermanQuAD dataset from deepset.",
"## Model description \n\nMore information needed",
"## Training and evaluation data\n\nBleu_1: 0.196051 \nBleu_2: 0.122380 \nBleu_3: 0.079980 \nBleu_4: 0.053672",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n",
"# german-qg-t5-e2e-quad (Work in progress)\n\nThis model is a end-to-end question generation model in German. Given a text, it generates several questions about it. This model is a fine-tuned version of valhalla/t5-base-e2e-qg on the GermanQuAD dataset from deepset.",
"## Model description \n\nMore information needed",
"## Training and evaluation data\n\nBleu_1: 0.196051 \nBleu_2: 0.122380 \nBleu_3: 0.079980 \nBleu_4: 0.053672",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10.0",
"### Training results",
"### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
77,
80,
6,
37,
3,
113,
4,
38
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# german-qg-t5-e2e-quad (Work in progress)\n\nThis model is a end-to-end question generation model in German. Given a text, it generates several questions about it. This model is a fine-tuned version of valhalla/t5-base-e2e-qg on the GermanQuAD dataset from deepset.## Model description \n\nMore information needed## Training and evaluation data\n\nBleu_1: 0.196051 \nBleu_2: 0.122380 \nBleu_3: 0.079980 \nBleu_4: 0.053672## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 4\n- eval_batch_size: 4\n- seed: 42\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 32\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10.0### Training results### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
-0.10497993230819702,
0.08560960739850998,
-0.002143452875316143,
0.08463151752948761,
0.08199774473905563,
0.03155207261443138,
0.14950688183307648,
0.118480384349823,
-0.09575799852609634,
0.08728962391614914,
0.08153867721557617,
-0.009967215359210968,
0.03873475268483162,
0.12355028092861176,
-0.016207102686166763,
-0.2241460531949997,
-0.005267205182462931,
-0.020397165790200233,
-0.07386390119791031,
0.11116109788417816,
0.10938819497823715,
-0.12950007617473602,
0.04799521341919899,
-0.009369052946567535,
-0.13500213623046875,
0.07315998524427414,
-0.017748350277543068,
-0.02654922381043434,
0.12567272782325745,
0.047737427055835724,
0.10983318835496902,
-0.014296023175120354,
0.1133432537317276,
-0.18198180198669434,
-0.002295747399330139,
0.06354779750108719,
-0.008087698370218277,
0.06511219590902328,
0.03839471563696861,
-0.01781005784869194,
0.08649872988462448,
-0.10815682262182236,
0.0739562138915062,
-0.0004855177830904722,
-0.08752201497554779,
-0.16438452899456024,
-0.0819443091750145,
0.0895896852016449,
0.050629258155822754,
0.10208823531866074,
-0.02726617641746998,
0.07952295988798141,
-0.1305675357580185,
0.08828721940517426,
0.1804962158203125,
-0.29302236437797546,
-0.059789836406707764,
0.08594405651092529,
0.03217793628573418,
0.08378645032644272,
-0.11267974972724915,
-0.0017141440184786916,
0.06220146268606186,
0.024483678862452507,
0.06558734178543091,
-0.01664072833955288,
-0.012319691479206085,
0.012283830903470516,
-0.08619879931211472,
-0.08228374272584915,
0.1031925156712532,
0.03833697736263275,
-0.03483441472053528,
-0.13440369069576263,
-0.03210088983178139,
-0.14330190420150757,
0.005999576300382614,
-0.036541279405355453,
0.026690468192100525,
-0.0538400374352932,
-0.05545887351036072,
-0.018291722983121872,
-0.07992077618837357,
-0.05083649978041649,
0.0016508522676303983,
0.15869493782520294,
0.026846667751669884,
0.016980571672320366,
-0.05419101566076279,
0.1315135806798935,
-0.04590582102537155,
-0.16002558171749115,
-0.013353207148611546,
0.02383183315396309,
-0.11342699825763702,
-0.03269941732287407,
-0.06139502301812172,
-0.07057199627161026,
-0.03061828576028347,
0.15823455154895782,
-0.05033889412879944,
0.06395218521356583,
0.012224455364048481,
-0.002837723819538951,
-0.006189260631799698,
0.21097922325134277,
-0.02724243700504303,
-0.15023015439510345,
0.0016098055057227612,
0.0674310252070427,
0.016681265085935593,
-0.013333760201931,
-0.08896742016077042,
-0.037937719374895096,
0.08458404988050461,
0.044459931552410126,
-0.059992652386426926,
0.015393649227917194,
-0.02723051607608795,
-0.026841409504413605,
-0.03280172497034073,
-0.13054154813289642,
0.03292958065867424,
0.03161361441016197,
-0.0991855040192604,
-0.012708118185400963,
-0.011877307668328285,
-0.030886832624673843,
-0.059061724692583084,
0.13663843274116516,
-0.04220951721072197,
-0.03763150796294212,
-0.09842949360609055,
-0.12831515073776245,
-0.006467461585998535,
-0.05334872752428055,
-0.002977089025080204,
-0.056797489523887634,
-0.17220589518547058,
-0.046034134924411774,
0.07287034392356873,
-0.059701107442379,
-0.022880464792251587,
-0.051868099719285965,
-0.06814778596162796,
0.05267380177974701,
-0.005756753496825695,
0.11699668318033218,
-0.04190623760223389,
0.06873507797718048,
-0.03536797687411308,
0.031673308461904526,
0.01598348096013069,
0.04065721854567528,
-0.07615688443183899,
0.019120605662465096,
-0.16284796595573425,
0.10391180217266083,
-0.05684683844447136,
-0.0447198860347271,
-0.1013002023100853,
-0.1283612996339798,
0.006106061860918999,
0.006488677579909563,
0.09522044658660889,
0.09024678170681,
-0.19965305924415588,
-0.055238962173461914,
0.13829798996448517,
-0.06462564319372177,
-0.08020168542861938,
0.10976437479257584,
-0.08502752333879471,
-0.00931982696056366,
0.05646613985300064,
0.1507342904806137,
0.11972716450691223,
-0.1347152441740036,
-0.058263182640075684,
-0.005785187240689993,
0.08234243839979172,
0.06147836893796921,
0.05286742001771927,
-0.0020112006459385157,
0.07753033190965652,
-0.011607512831687927,
-0.06911305338144302,
0.009431561455130577,
-0.05129547417163849,
-0.07897309213876724,
-0.025933044031262398,
-0.0503321997821331,
0.07314477860927582,
0.01306562777608633,
0.04485168308019638,
-0.08594769239425659,
-0.1228867620229721,
0.10229401290416718,
0.12466932088136673,
-0.08862639963626862,
0.002948008244857192,
-0.04463743790984154,
0.05119270086288452,
-0.04499137029051781,
-0.018728995695710182,
-0.18664740025997162,
-0.10019515454769135,
0.038894426077604294,
-0.055826712399721146,
0.02639041282236576,
0.10243471711874008,
0.09263397753238678,
0.07096289843320847,
-0.08816246688365936,
-0.06084803491830826,
-0.14261095225811005,
-0.037741564214229584,
-0.061113711446523666,
-0.18457947671413422,
-0.06709201633930206,
-0.04165239632129669,
0.12434371560811996,
-0.15704818069934845,
-0.025976290926337242,
0.030356649309396744,
0.1372808963060379,
0.0472659170627594,
-0.03501075878739357,
-0.04578210040926933,
0.04011028632521629,
-0.023143354803323746,
-0.07255551218986511,
-0.005849557463079691,
-0.014886299148201942,
-0.08282431960105896,
-0.043040595948696136,
-0.11702137440443039,
0.055904120206832886,
0.08351575583219528,
0.07829922437667847,
-0.10202354937791824,
-0.04082159698009491,
-0.06852857768535614,
-0.0207191351801157,
-0.0721866637468338,
-0.04521998018026352,
0.2701987624168396,
0.014133796095848083,
0.11970692127943039,
-0.10719457268714905,
-0.09290824085474014,
-0.032707329839468,
0.038577139377593994,
0.006893378682434559,
0.12396756559610367,
0.006721096578985453,
-0.09139033406972885,
0.0920729860663414,
0.09774426370859146,
-0.06561220437288284,
0.2351296842098236,
-0.05975120887160301,
-0.0850834846496582,
-0.053185295313596725,
0.028751468285918236,
-0.015483562834560871,
0.19055764377117157,
-0.04078448563814163,
0.015009425580501556,
0.025294125080108643,
0.03625484183430672,
0.02987632527947426,
-0.1321430206298828,
0.0010180269600823522,
0.030497245490550995,
-0.0616542212665081,
0.00896800309419632,
-0.013789486140012741,
0.045520056039094925,
0.11719463765621185,
0.02016582153737545,
-0.02330685779452324,
-0.004769184160977602,
-0.04629041254520416,
-0.08457349985837936,
0.2096431404352188,
-0.07846303284168243,
-0.12357991933822632,
-0.09115605056285858,
0.04654451087117195,
-0.09203682094812393,
-0.018626002594828606,
0.006980929058045149,
-0.0937080904841423,
-0.04025041311979294,
-0.07094912230968475,
0.055579401552677155,
0.02214454859495163,
0.002090897411108017,
-0.02361677959561348,
-0.010911273770034313,
0.06737165153026581,
-0.17054779827594757,
0.007122923620045185,
-0.036571044474840164,
-0.021389618515968323,
-0.014243627898395061,
0.0759609192609787,
0.06699834764003754,
0.12579715251922607,
-0.05152377858757973,
0.032215941697359085,
-0.025982508435845375,
0.23717504739761353,
-0.1255682408809662,
0.007794592063874006,
0.08951840549707413,
-0.014818157069385052,
0.02911616675555706,
0.13380487263202667,
0.04313192889094353,
-0.08684191852807999,
0.024337293580174446,
0.0655052587389946,
-0.021116668358445168,
-0.2198641002178192,
-0.05378562584519386,
-0.04761091619729996,
-0.07596749067306519,
0.09702080488204956,
0.02229246124625206,
-0.03369443118572235,
0.007969308644533157,
-0.015382163226604462,
0.04902196303009987,
-0.009698384441435337,
0.06246008351445198,
0.04772819206118584,
0.04079696163535118,
0.09784962981939316,
-0.03614165261387825,
-0.043478332459926605,
0.07845481485128403,
0.01978001557290554,
0.1938202977180481,
-0.08155054599046707,
0.07255826145410538,
0.015253674238920212,
0.10564732551574707,
-0.01259908638894558,
0.06691240519285202,
-0.011806444264948368,
-0.03771362081170082,
-0.0023711430840194225,
-0.07948151230812073,
-0.027182981371879578,
0.06008589267730713,
0.020984720438718796,
0.07454060763120651,
-0.07711739093065262,
0.0024632420390844345,
0.04861639440059662,
0.2564713656902313,
0.08755470812320709,
-0.2363884001970291,
-0.09738786518573761,
0.012015940621495247,
-0.029489632695913315,
-0.041120048612356186,
0.011967583559453487,
0.1736811399459839,
-0.12862935662269592,
0.029503008350729942,
-0.04659000039100647,
0.08642885833978653,
-0.044168997555971146,
0.0023737233132123947,
0.04643949866294861,
0.10497581213712692,
-0.026993857696652412,
0.08074352890253067,
-0.24317005276679993,
0.21326981484889984,
0.018723562359809875,
0.09818826615810394,
-0.09391555935144424,
0.04549296200275421,
0.0020031810272485018,
-0.03244432806968689,
0.08214427530765533,
0.0005284919170662761,
-0.06875886768102646,
-0.09176025539636612,
-0.05848308280110359,
0.03654748573899269,
0.11975967884063721,
-0.05482177436351776,
0.09162124991416931,
-0.019985385239124298,
0.006910179741680622,
0.023404521867632866,
-0.020188115537166595,
-0.08962201327085495,
-0.10173553973436356,
-0.00907484907656908,
-0.022093188017606735,
-0.047845497727394104,
-0.06008441001176834,
-0.08010461926460266,
-0.03799229487776756,
0.1935051828622818,
-0.009240420535206795,
-0.034289903938770294,
-0.1322479397058487,
0.10007534921169281,
0.10748860985040665,
-0.08465100079774857,
0.041631944477558136,
0.042079776525497437,
0.1078161671757698,
-0.020705103874206543,
-0.07795088738203049,
0.05942431092262268,
-0.07327180355787277,
-0.1525615006685257,
-0.04398910328745842,
0.1419335901737213,
0.05860535055398941,
0.06364922970533371,
0.012886984273791313,
-0.002454019384458661,
0.02164270356297493,
-0.11972987651824951,
0.009679392911493778,
0.015260972082614899,
0.07367315143346786,
0.06548431515693665,
-0.07492613792419434,
-0.025373369455337524,
0.0031919116154313087,
0.03264066204428673,
0.15784941613674164,
0.23342713713645935,
-0.10787580162286758,
0.06719402223825455,
0.0359639897942543,
-0.0504285953938961,
-0.1860148161649704,
0.08561239391565323,
0.13724735379219055,
0.08810316771268845,
0.025838572531938553,
-0.16682153940200806,
0.07772573083639145,
0.05670638754963875,
-0.0049102287739515305,
-0.008638097904622555,
-0.3361828625202179,
-0.15325872600078583,
0.07584230601787567,
0.08946077525615692,
0.01168169267475605,
-0.07145500928163528,
-0.03451544791460037,
-0.04865105822682381,
-0.09245827049016953,
0.06456968933343887,
-0.06410469859838486,
0.101341113448143,
9.286469548897003e-7,
0.06745054572820663,
0.04581882432103157,
-0.05133899673819542,
0.1261550784111023,
0.032743655145168304,
0.10701680183410645,
-0.05668328329920769,
0.009690679609775543,
0.07817384600639343,
-0.056245870888233185,
0.07352560013532639,
-0.0932462066411972,
0.08181114494800568,
-0.1813768893480301,
-0.0038257876876741648,
-0.08561831712722778,
0.044668104499578476,
-0.08054015785455704,
-0.09112090617418289,
-0.08212383091449738,
0.07248485088348389,
0.08387041091918945,
-0.017476292327046394,
0.019538022577762604,
-0.0024864599108695984,
0.05694490671157837,
0.09463844448328018,
0.09032877534627914,
0.08025222271680832,
-0.11643704771995544,
0.025321602821350098,
0.019072769209742546,
0.06349952518939972,
-0.13371606171131134,
0.04719727858901024,
0.1518760472536087,
0.043985553085803986,
0.16260017454624176,
0.021275945007801056,
-0.06122308969497681,
0.015541632659733295,
0.04844028875231743,
-0.13496112823486328,
-0.12316741794347763,
-0.02749962918460369,
-0.06187373772263527,
-0.10844975709915161,
-0.0015145008219406009,
0.14694927632808685,
-0.04050946235656738,
0.006657132878899574,
-0.0074549950659275055,
-0.010791660286486149,
-0.053768403828144073,
0.17651358246803284,
0.023094680160284042,
0.07773470133543015,
-0.06551121175289154,
0.09913911670446396,
0.049165837466716766,
-0.06938399374485016,
0.06544162333011627,
0.03576879948377609,
-0.044973891228437424,
0.009740742854773998,
0.06791505962610245,
0.1543252170085907,
-0.04079611599445343,
-0.01822742260992527,
-0.1038106232881546,
-0.0998624861240387,
0.05952656641602516,
0.07846666872501373,
0.056173257529735565,
-0.010858044028282166,
-0.05571145564317703,
0.0029899962246418,
-0.09145872294902802,
0.11839334666728973,
0.07319683581590652,
0.03725795820355415,
-0.07887191325426102,
0.13909856975078583,
-0.014102337881922722,
0.012447194196283817,
-0.02087496966123581,
0.01541819330304861,
-0.11831538379192352,
-0.00559016689658165,
-0.16399385035037994,
-0.02601657249033451,
0.04626144468784332,
0.024386940523982048,
-0.039695918560028076,
-0.040911588817834854,
-0.04647498205304146,
0.032969966530799866,
-0.09867026656866074,
-0.02492832951247692,
0.017618726938962936,
0.03968019783496857,
-0.15560130774974823,
-0.012675628066062927,
0.021373972296714783,
-0.09123410284519196,
0.13490121066570282,
0.05170385167002678,
0.025737114250659943,
0.048152897506952286,
-0.05922015756368637,
-0.014300662092864513,
0.013597476296126842,
0.027713973075151443,
0.054814085364341736,
-0.09129801392555237,
0.01538816187530756,
-0.01959424465894699,
0.010475913062691689,
0.03881990909576416,
0.010883593000471592,
-0.10355883091688156,
0.04720142111182213,
-0.02643374167382717,
-0.06179714575409889,
-0.07879761606454849,
0.11087892949581146,
0.05096403881907463,
0.031297095119953156,
0.13000145554542542,
-0.07881259173154831,
0.06191007047891617,
-0.20443014800548553,
-0.043005988001823425,
0.027532340958714485,
-0.02691422775387764,
-0.09580013900995255,
-0.03445633128285408,
0.06588753312826157,
-0.06135144829750061,
0.09714457392692566,
0.009207004681229591,
0.058459024876356125,
0.04015979543328285,
-0.026037007570266724,
0.013885446824133396,
0.030070578679442406,
0.17742270231246948,
-0.01513037271797657,
-0.012540687806904316,
0.03557753190398216,
0.0073438831605017185,
0.037758730351924896,
0.13669055700302124,
0.2129673808813095,
0.1883961409330368,
0.02554694004356861,
0.07937344163656235,
0.060831934213638306,
-0.10325279831886292,
-0.1344580203294754,
0.09256304800510406,
-0.03232911229133606,
0.11353006213903427,
-0.05608705058693886,
0.1475449651479721,
0.08836904168128967,
-0.23196212947368622,
0.08016637712717056,
-0.026470789685845375,
-0.09282495826482773,
-0.07962611317634583,
-0.04496258124709129,
-0.06595529615879059,
-0.10426048934459686,
0.036362648010253906,
-0.14973297715187073,
0.03991614282131195,
0.06387636810541153,
0.0696810856461525,
-0.024114292114973068,
0.18878626823425293,
-0.09760671108961105,
-0.008913726545870304,
0.09970109909772873,
0.04763171076774597,
0.03245536610484123,
-0.038732122629880905,
-0.03655137121677399,
0.016055885702371597,
-0.024592559784650803,
0.059143535792827606,
-0.025165464729070663,
0.005858891643583775,
-0.007126007229089737,
0.013617714866995811,
-0.03733966872096062,
0.020208466798067093,
-0.01334824226796627,
0.08440008014440536,
0.10384363681077957,
0.04491179436445236,
0.010881190188229084,
-0.06447379291057587,
0.24464084208011627,
-0.07399550825357437,
-0.0839645192027092,
-0.14718382060527802,
0.12845347821712494,
0.027025891467928886,
0.04061155021190643,
0.05063572898507118,
-0.11845618486404419,
-0.03172486275434494,
0.11444860696792603,
0.16997994482517242,
-0.09993965178728104,
0.0005827039130963385,
0.013084245845675468,
-0.003731501055881381,
-0.02255646325647831,
0.09554115682840347,
0.043429259210824966,
0.14161312580108643,
-0.04032628610730171,
0.0187437254935503,
-0.024316010996699333,
-0.049243174493312836,
0.030451063066720963,
0.1661689579486847,
-0.0048830583691596985,
-0.052271343767642975,
-0.06450673192739487,
0.04620784893631935,
0.006938116159290075,
-0.19275599718093872,
0.03444104641675949,
-0.22657831013202667,
-0.20624080300331116,
-0.013012859970331192,
0.06599044799804688,
0.01174156740307808,
0.0748063325881958,
-0.015813900157809258,
-0.027824413031339645,
0.09935310482978821,
-0.022723158821463585,
-0.02428630366921425,
-0.15197895467281342,
0.08667699247598648,
-0.0719299241900444,
0.19915516674518585,
-0.0006756921648047864,
0.0905722826719284,
0.11940384656190872,
0.014200947247445583,
-0.06266330182552338,
0.01452484168112278,
0.09747644513845444,
-0.12631532549858093,
-0.001746581750921905,
0.15867184102535248,
-0.0558461993932724,
0.12408731877803802,
0.05309552699327469,
-0.09691140055656433,
0.027476608753204346,
-0.021944260224699974,
-0.042875636368989944,
-0.1215352863073349,
0.024511979892849922,
-0.05260293185710907,
0.15466155111789703,
0.23417463898658752,
-0.008381840772926807,
0.013921703211963177,
-0.05802478641271591,
0.07773815840482712,
0.04463426023721695,
0.08955240249633789,
-0.005123557057231665,
-0.19811411201953888,
0.05524835363030434,
-0.0024324399419128895,
-0.02393851801753044,
-0.20503854751586914,
-0.1125657930970192,
0.07028844952583313,
-0.04050857573747635,
-0.05363576114177704,
0.1350594013929367,
0.06419245153665543,
0.03699461743235588,
-0.029945921152830124,
-0.11437764763832092,
-0.05703834816813469,
0.09212995320558548,
-0.11345694959163666,
-0.05944158881902695
] |
null | null |
transformers
|
# german-qg-t5-quad
This model is fine-tuned in question generation in German. The expected answer must be highlighted with a
<hl> token.
## Task example
#### Input
generate question: Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des <hl> britischen Common Laws <hl> sind, setzt sich das amerikanische Recht bedeutend davon ab. Dies rührt größtenteils von dem langen Zeitraum her, [...]
#### Expected output
Von welchem Gesetzt stammt das Amerikanische ab?
## Model description
This model is a fine-tuned version of [valhalla/t5-base-qg-hl](https://huggingface.co/valhalla/t5-base-qg-hl) on the [GermanQUAD](https://www.deepset.ai/germanquad) dataset.
## Training and evaluation data
The training script can be accessed [here](https://github.com/d-e-h-i-o/german-qg).
### Evaluation
The model achieves a BLEU-4 score of **11.30** on the GermanQuAD test set (n=2204).
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 2
- seed: 100
- gradient_accumulation_steps: 8
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Framework versions
- Transformers 4.13.0.dev0
- Pytorch 1.10.0+cu102
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"language": ["de"], "license": "mit", "tags": ["question generation"], "datasets": ["deepset/germanquad"], "widget": [{"text": "Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des <hl>britischen Common Laws<hl> sind, setzt sich das amerikanische Recht bedeutend davon ab."}], "model-index": [{"name": "german-qg-t5-quad", "results": []}]}
|
text2text-generation
|
dehio/german-qg-t5-quad
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"question generation",
"de",
"dataset:deepset/germanquad",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# german-qg-t5-quad
This model is fine-tuned in question generation in German. The expected answer must be highlighted with a
<hl> token.
## Task example
#### Input
generate question: Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des <hl> britischen Common Laws <hl> sind, setzt sich das amerikanische Recht bedeutend davon ab. Dies rührt größtenteils von dem langen Zeitraum her, [...]
#### Expected output
Von welchem Gesetzt stammt das Amerikanische ab?
## Model description
This model is a fine-tuned version of valhalla/t5-base-qg-hl on the GermanQUAD dataset.
## Training and evaluation data
The training script can be accessed here.
### Evaluation
The model achieves a BLEU-4 score of 11.30 on the GermanQuAD test set (n=2204).
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 2
- eval_batch_size: 2
- seed: 100
- gradient_accumulation_steps: 8
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Framework versions
- Transformers 4.13.0.dev0
- Pytorch 1.10.0+cu102
- Datasets 1.16.1
- Tokenizers 0.10.3
|
[
"# german-qg-t5-quad\n\nThis model is fine-tuned in question generation in German. The expected answer must be highlighted with a\n<hl> token.",
"## Task example",
"#### Input\n\ngenerate question: Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des <hl> britischen Common Laws <hl> sind, setzt sich das amerikanische Recht bedeutend davon ab. Dies rührt größtenteils von dem langen Zeitraum her, [...]",
"#### Expected output\n\nVon welchem Gesetzt stammt das Amerikanische ab?",
"## Model description\n\nThis model is a fine-tuned version of valhalla/t5-base-qg-hl on the GermanQUAD dataset.",
"## Training and evaluation data\n\nThe training script can be accessed here.",
"### Evaluation\n\nThe model achieves a BLEU-4 score of 11.30 on the GermanQuAD test set (n=2204).",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 100\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10",
"### Framework versions\n\n- Transformers 4.13.0.dev0\n- Pytorch 1.10.0+cu102\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# german-qg-t5-quad\n\nThis model is fine-tuned in question generation in German. The expected answer must be highlighted with a\n<hl> token.",
"## Task example",
"#### Input\n\ngenerate question: Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des <hl> britischen Common Laws <hl> sind, setzt sich das amerikanische Recht bedeutend davon ab. Dies rührt größtenteils von dem langen Zeitraum her, [...]",
"#### Expected output\n\nVon welchem Gesetzt stammt das Amerikanische ab?",
"## Model description\n\nThis model is a fine-tuned version of valhalla/t5-base-qg-hl on the GermanQUAD dataset.",
"## Training and evaluation data\n\nThe training script can be accessed here.",
"### Evaluation\n\nThe model achieves a BLEU-4 score of 11.30 on the GermanQuAD test set (n=2204).",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 100\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10",
"### Framework versions\n\n- Transformers 4.13.0.dev0\n- Pytorch 1.10.0+cu102\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
69,
39,
4,
63,
17,
32,
14,
27,
112,
36
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #question generation #de #dataset-deepset/germanquad #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# german-qg-t5-quad\n\nThis model is fine-tuned in question generation in German. The expected answer must be highlighted with a\n<hl> token.## Task example#### Input\n\ngenerate question: Obwohl die Vereinigten Staaten wie auch viele Staaten des Commonwealth Erben des <hl> britischen Common Laws <hl> sind, setzt sich das amerikanische Recht bedeutend davon ab. Dies rührt größtenteils von dem langen Zeitraum her, [...]#### Expected output\n\nVon welchem Gesetzt stammt das Amerikanische ab?## Model description\n\nThis model is a fine-tuned version of valhalla/t5-base-qg-hl on the GermanQUAD dataset.## Training and evaluation data\n\nThe training script can be accessed here.### Evaluation\n\nThe model achieves a BLEU-4 score of 11.30 on the GermanQuAD test set (n=2204).### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 2\n- eval_batch_size: 2\n- seed: 100\n- gradient_accumulation_steps: 8\n- total_train_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 10### Framework versions\n\n- Transformers 4.13.0.dev0\n- Pytorch 1.10.0+cu102\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
-0.13540518283843994,
0.04941993206739426,
-0.003791832597926259,
0.1096971184015274,
0.09531113505363464,
0.03440289944410324,
0.15669289231300354,
0.1024634912610054,
-0.024486441165208817,
0.05605384707450867,
0.06793169677257538,
-0.00003968754390371032,
0.046959731727838516,
0.1088201105594635,
-0.017222201451659203,
-0.21820847690105438,
0.01738021895289421,
-0.0342995710670948,
-0.033001504838466644,
0.10878859460353851,
0.10568884760141373,
-0.11468763649463654,
0.040685247629880905,
-0.023338748142123222,
-0.11209359765052795,
0.05554249510169029,
-0.02245388738811016,
-0.022239506244659424,
0.11989793926477432,
0.02541893534362316,
0.09410680085420609,
-0.003916165325790644,
0.10912364721298218,
-0.18392525613307953,
-0.004945877939462662,
0.05054711550474167,
0.0016255484661087394,
0.0568067692220211,
0.050756487995386124,
-0.044485170394182205,
0.08237609267234802,
-0.11836450546979904,
0.019191188737750053,
0.010362142696976662,
-0.09282882511615753,
-0.13847185671329498,
-0.10196325182914734,
0.10705570131540298,
0.034338030964136124,
0.09154107421636581,
-0.04343890771269798,
0.08941739797592163,
-0.13598376512527466,
0.07547607272863388,
0.16592437028884888,
-0.2718025743961334,
-0.06256431341171265,
0.07562277466058731,
-0.005928338970988989,
0.04653996601700783,
-0.13057637214660645,
-0.021470611914992332,
0.05308133736252785,
0.05084982141852379,
0.039213065057992935,
-0.054029036313295364,
-0.018428433686494827,
0.017091017216444016,
-0.12188813835382462,
-0.03755839914083481,
0.12478280812501907,
0.03244650736451149,
-0.06304159760475159,
-0.0773015096783638,
-0.007811785209923983,
-0.03652309253811836,
0.03346510976552963,
-0.034551192075014114,
0.007302303798496723,
-0.04974493011832237,
-0.022245533764362335,
0.005491861142218113,
-0.08234944939613342,
-0.010878302156925201,
0.000781774113420397,
0.1600871980190277,
0.03376597911119461,
0.003866841085255146,
-0.038574423640966415,
0.10893987119197845,
-0.07385409623384476,
-0.14725308120250702,
-0.01807429827749729,
-0.0014604905154556036,
-0.08293306827545166,
-0.05302831158041954,
-0.05176271125674248,
-0.06727127730846405,
0.0030354843474924564,
0.1687678098678589,
-0.06104464828968048,
0.07581772655248642,
0.03479129821062088,
0.027444619685411453,
0.01409965194761753,
0.17077992856502533,
-0.06159260496497154,
-0.14224834740161896,
-0.0074707153253257275,
0.06106080859899521,
-0.010307910852134228,
-0.0017295246943831444,
-0.06570439040660858,
-0.026907969266176224,
0.11372792720794678,
0.02928953804075718,
-0.05149271711707115,
0.0391346700489521,
-0.06131116300821304,
-0.012206990271806717,
-0.011731409467756748,
-0.13290506601333618,
0.031415972858667374,
0.026098541915416718,
-0.1000283882021904,
0.045275550335645676,
-0.001709545380435884,
-0.0208683330565691,
-0.06868353486061096,
0.15489459037780762,
-0.03546523302793503,
0.008524884469807148,
-0.09364783763885498,
-0.10756631940603256,
-0.0004959100624546409,
-0.036477964371442795,
-0.02021247334778309,
-0.07324861735105515,
-0.14942961931228638,
-0.05131945013999939,
0.08575649559497833,
-0.04113087058067322,
0.014459253288805485,
-0.08591721951961517,
-0.08315896987915039,
0.008154314942657948,
-0.010369475930929184,
0.05209359899163246,
-0.034114107489585876,
0.054680388420820236,
-0.023522378876805305,
0.03580213338136673,
-0.004327587783336639,
0.039729587733745575,
-0.10827889293432236,
-0.0011885995045304298,
-0.2037453055381775,
0.10404162108898163,
-0.06951194256544113,
-0.03510119393467903,
-0.11343897879123688,
-0.1379062831401825,
0.011220666579902172,
0.0026525543071329594,
0.10976622998714447,
0.12141310423612595,
-0.21352674067020416,
-0.04648115485906601,
0.12886635959148407,
-0.06709826737642288,
-0.059574566781520844,
0.12612628936767578,
-0.0835140123963356,
0.014700312167406082,
0.06573712825775146,
0.18261854350566864,
0.1031700000166893,
-0.10693231225013733,
-0.06383330374956131,
-0.011690022423863411,
0.04161551967263222,
0.09511867165565491,
0.04249606654047966,
-0.007730064447969198,
0.03648338094353676,
0.010309736244380474,
-0.10466334223747253,
0.003119769971817732,
-0.06911540031433105,
-0.060115136206150055,
-0.008820409886538982,
-0.044795792549848557,
0.1106187254190445,
0.01612352579832077,
0.02049870602786541,
-0.09179812669754028,
-0.11052197962999344,
0.1094101294875145,
0.07882726937532425,
-0.06194377318024635,
0.01936212182044983,
-0.049210142344236374,
0.027717242017388344,
-0.006837382446974516,
-0.012794021517038345,
-0.1787073165178299,
-0.07892295718193054,
0.018687903881072998,
-0.033759474754333496,
0.07210960239171982,
0.11070041358470917,
0.08377610146999359,
0.08113408833742142,
-0.07690481841564178,
-0.046205684542655945,
-0.11021038889884949,
-0.016704393550753593,
-0.05253070220351219,
-0.19519829750061035,
-0.06379828602075577,
-0.021710475906729698,
0.13563448190689087,
-0.13198435306549072,
-0.005466855596750975,
0.009978977963328362,
0.11374645680189133,
-0.006421636324375868,
-0.025909464806318283,
-0.06891315430402756,
0.0494101457297802,
-0.022878719493746758,
-0.04418034106492996,
0.014175440184772015,
-0.0036144142504781485,
-0.04079427197575569,
-0.0657154992222786,
-0.05312586948275566,
-0.039139389991760254,
0.0741385743021965,
-0.01671847514808178,
-0.12372469902038574,
-0.021105263382196426,
-0.066365547478199,
-0.020722506567835808,
-0.07479491829872131,
-0.03796365112066269,
0.27669191360473633,
0.01846102811396122,
0.07883142679929733,
-0.07596684247255325,
-0.06402676552534103,
-0.0329134538769722,
-0.021186916157603264,
0.012018019333481789,
0.09202837198972702,
0.02409372478723526,
-0.0836947038769722,
0.08874855935573578,
0.08392315357923508,
-0.08887966722249985,
0.16524812579154968,
-0.025101548060774803,
-0.0534905269742012,
-0.05206994712352753,
0.03586943447589874,
-0.02223869040608406,
0.16263040900230408,
-0.05438107252120972,
0.009308907203376293,
0.01927764154970646,
0.0689033791422844,
0.02655975893139839,
-0.13408203423023224,
0.0016868849052116275,
-0.012159459292888641,
-0.037906404584646225,
-0.013014460913836956,
-0.025923795998096466,
0.04046808183193207,
0.10833972692489624,
0.00663388054817915,
-0.053428854793310165,
0.014365002512931824,
-0.03949650749564171,
-0.10273578017950058,
0.19392621517181396,
-0.07520570605993271,
-0.13356980681419373,
-0.07160820811986923,
0.052250973880290985,
-0.10476505011320114,
-0.01574992761015892,
0.022195713594555855,
-0.07672816514968872,
-0.05425523221492767,
-0.05382393300533295,
0.06588686257600784,
0.050511185079813004,
-0.024947822093963623,
-0.052556443959474564,
-0.027734210714697838,
0.06853950023651123,
-0.15340100228786469,
-0.019567837938666344,
-0.02841489389538765,
-0.05010998249053955,
0.031115666031837463,
0.0691889226436615,
0.061240874230861664,
0.10578424483537674,
-0.04978111386299133,
0.009154161438345909,
-0.031518783420324326,
0.2344505488872528,
-0.09537940472364426,
0.05338423699140549,
0.041018590331077576,
-0.022538024932146072,
0.05510499328374863,
0.13101093471050262,
0.04125390201807022,
-0.07991749793291092,
0.018253188580274582,
0.055127352476119995,
-0.01769077591598034,
-0.20626215636730194,
-0.06050149351358414,
-0.06884301453828812,
-0.021162046119570732,
0.08395848423242569,
0.014285985380411148,
-0.02280844934284687,
0.04648134112358093,
-0.016892340034246445,
0.033522527664899826,
0.016842592507600784,
0.06477243453264236,
0.025273585692048073,
0.054848648607730865,
0.10068134218454361,
-0.02490231581032276,
-0.04660005867481232,
0.08167510479688644,
0.01240924559533596,
0.20593373477458954,
-0.06568498909473419,
0.004839166067540646,
0.060344841331243515,
0.15124431252479553,
-0.04125436395406723,
0.09954103082418442,
-0.014048073440790176,
-0.025270262733101845,
-0.0057764179073274136,
-0.07646207511425018,
-0.005424512084573507,
0.0594407394528389,
-0.024415869265794754,
0.061979737132787704,
-0.0633552074432373,
0.010443381033837795,
0.10756547749042511,
0.20235289633274078,
0.10113678872585297,
-0.19329588115215302,
-0.09230361133813858,
0.013339992612600327,
-0.0004154749622102827,
-0.011158887296915054,
0.021047359332442284,
0.13235697150230408,
-0.15259656310081482,
0.05545816197991371,
-0.038782380521297455,
0.08813308924436569,
-0.02404266968369484,
0.016916383057832718,
0.03668839856982231,
0.13256697356700897,
-0.02445519156754017,
0.10313532501459122,
-0.3178495764732361,
0.18459124863147736,
-0.008114747703075409,
0.10183455795049667,
-0.10487186163663864,
0.027905533090233803,
0.014920410700142384,
-0.025687657296657562,
0.10018410533666611,
0.027207693085074425,
0.013127537444233894,
-0.03768647462129593,
-0.000733831780962646,
0.027423273772001266,
0.12524300813674927,
-0.007555294781923294,
0.10092199593782425,
-0.011205371469259262,
0.01972665637731552,
0.011199765838682652,
0.02265876717865467,
-0.14255644381046295,
-0.07244197279214859,
-0.0029149064794182777,
-0.023651426658034325,
-0.04032069072127342,
-0.04369182139635086,
-0.0771409347653389,
-0.14116761088371277,
0.19742384552955627,
-0.04679808020591736,
-0.03766957297921181,
-0.09279431402683258,
0.07518992573022842,
0.0922180637717247,
-0.0845828503370285,
-0.0076287891715765,
0.031338781118392944,
0.06850285828113556,
-0.025954212993383408,
-0.07290815562009811,
0.0741758793592453,
-0.08242172002792358,
-0.14043447375297546,
-0.03906366229057312,
0.11956171691417694,
0.0858374685049057,
0.05986712872982025,
0.006468515377491713,
0.01781507022678852,
0.006765267346054316,
-0.15300022065639496,
-0.001935527310706675,
-0.04855366423726082,
0.0460362508893013,
0.10051727294921875,
-0.08732280880212784,
-0.10806155204772949,
-0.04312962666153908,
0.0448618046939373,
0.11168371886014938,
0.20056892931461334,
-0.08807645738124847,
0.03723503649234772,
0.10125719755887985,
-0.054214075207710266,
-0.22541876137256622,
0.096121646463871,
0.10594413429498672,
0.047432791441679,
0.03526584804058075,
-0.15852363407611847,
0.1317903697490692,
0.07338161021471024,
0.013623417355120182,
-0.006309331860393286,
-0.29350435733795166,
-0.13143415749073029,
0.06046991050243378,
0.07855211943387985,
-0.028716126456856728,
-0.08175073564052582,
-0.021429821848869324,
-0.04235108569264412,
-0.12584494054317474,
0.05645892769098282,
-0.11729065328836441,
0.08106808364391327,
0.014416568912565708,
0.08604291081428528,
0.023990219458937645,
-0.05651674419641495,
0.11949244886636734,
0.12683476507663727,
0.0965491458773613,
-0.07907018810510635,
-0.026572715491056442,
0.11625790596008301,
-0.02192389778792858,
0.13131120800971985,
-0.09908165037631989,
0.043622937053442,
-0.12051423639059067,
0.009042768739163876,
-0.10678217560052872,
0.08599528670310974,
-0.07919254899024963,
-0.09959837794303894,
-0.06042185425758362,
0.09540111571550369,
0.0708380714058876,
-0.010633237659931183,
0.0056722043082118034,
-0.004459452349692583,
0.041983507573604584,
0.10067383199930191,
0.10491887480020523,
0.05771219730377197,
-0.06927948445081711,
0.02371779829263687,
0.015704870223999023,
0.030878908932209015,
-0.08696615695953369,
0.03880733624100685,
0.1626506894826889,
0.040547747164964676,
0.1850675791501999,
0.03404415771365166,
-0.06295063346624374,
0.012813517823815346,
0.05619964003562927,
-0.14698311686515808,
-0.12994128465652466,
-0.0007952831801958382,
-0.07356927543878555,
-0.07203936576843262,
0.04763771966099739,
0.11590500175952911,
-0.06790662556886673,
0.003852270310744643,
-0.013436454348266125,
0.001119303284212947,
-0.09549486637115479,
0.1575808972120285,
0.05595387890934944,
0.07210902869701385,
-0.0720519945025444,
0.13300453126430511,
0.0445159487426281,
-0.03810526803135872,
0.06494259834289551,
0.07214758545160294,
-0.08038491010665894,
-0.0009798192186281085,
0.039094120264053345,
0.18178172409534454,
-0.046926017850637436,
-0.026128778234124184,
-0.09831013530492783,
-0.10357905924320221,
0.03283790498971939,
0.03667445108294487,
0.06309346109628677,
-0.01245880126953125,
-0.041944220662117004,
0.011222549714148045,
-0.11760668456554413,
0.08303745836019516,
0.10671405494213104,
0.038551609963178635,
-0.030889421701431274,
0.16444209218025208,
-0.013477163389325142,
0.02607298269867897,
-0.023932838812470436,
-0.0010056786704808474,
-0.11854670196771622,
0.002175135537981987,
-0.17006172239780426,
-0.01786925457417965,
-0.01253587007522583,
0.013957787305116653,
-0.031572990119457245,
-0.07502982020378113,
-0.018277673050761223,
0.045879509299993515,
-0.08159259706735611,
-0.024204691872000694,
0.032451990991830826,
0.04733112454414368,
-0.17310501635074615,
-0.023604873567819595,
0.024709945544600487,
-0.07422337681055069,
0.10750129073858261,
0.05775650963187218,
-0.01821870170533657,
0.07416088134050369,
-0.10931646823883057,
-0.01867559924721718,
-0.005465775728225708,
0.023953992873430252,
0.07014122605323792,
-0.08179701864719391,
0.005339423660188913,
0.00857415422797203,
0.06262878328561783,
0.043735429644584656,
0.0029820268973708153,
-0.08538606017827988,
0.08423276245594025,
-0.020004190504550934,
-0.08762148022651672,
-0.07461405545473099,
0.10291791707277298,
0.04241497442126274,
0.039947040379047394,
0.16607198119163513,
-0.0769130066037178,
0.05965556204319,
-0.18539607524871826,
-0.03913971781730652,
0.03617342561483383,
-0.03952576592564583,
-0.09594607353210449,
-0.04532541707158089,
0.07982336729764938,
-0.05161656066775322,
0.07590077817440033,
0.01067743357270956,
0.05125780031085014,
0.02342688851058483,
-0.03315361961722374,
-0.04824226349592209,
0.041231390088796616,
0.10905315726995468,
-0.03408961370587349,
-0.027058761566877365,
-0.05703195556998253,
0.029339417815208435,
0.004820358939468861,
0.07627612352371216,
0.23926736414432526,
0.16864192485809326,
0.017546631395816803,
0.04848453402519226,
0.02668744884431362,
-0.08592621237039566,
-0.1169516071677208,
0.03987149894237518,
0.017263544723391533,
0.08147594332695007,
-0.05196686089038849,
0.15316368639469147,
0.05890069156885147,
-0.2217177301645279,
0.09627232700586319,
-0.06123654916882515,
-0.10167139768600464,
-0.0507049560546875,
-0.05469844862818718,
-0.04824807122349739,
-0.07021139562129974,
0.03385290876030922,
-0.14225243031978607,
0.03013233281672001,
0.07785220444202423,
0.07648472487926483,
-0.049733515828847885,
0.20753857493400574,
-0.10694926977157593,
-0.007517787162214518,
0.09723728895187378,
0.004956413526087999,
0.0390184223651886,
-0.004096443764865398,
-0.025122007355093956,
-0.02947024255990982,
-0.003558787750080228,
0.07621731609106064,
0.009974964894354343,
0.010936680249869823,
-0.040186114609241486,
-0.017136892303824425,
-0.05798725783824921,
0.0124904690310359,
-0.0012991668190807104,
0.055298831313848495,
0.10894478857517242,
0.048606935888528824,
0.007350838277488947,
-0.045749325305223465,
0.19227918982505798,
-0.07108215987682343,
-0.17013445496559143,
-0.12540967762470245,
0.18417677283287048,
0.03923621028661728,
0.03608741611242294,
0.03585698455572128,
-0.12044203281402588,
0.005943082273006439,
0.1174376979470253,
0.17377835512161255,
-0.1145191341638565,
0.004516772460192442,
-0.009792311117053032,
0.0016529407585039735,
0.020359119400382042,
0.09245569258928299,
0.0537247434258461,
0.14607280492782593,
-0.05528324842453003,
0.07367677241563797,
-0.049302004277706146,
-0.02831120602786541,
0.0059891352429986,
0.10611385107040405,
0.0007098583155311644,
-0.03801266849040985,
-0.09456658363342285,
0.07468349486589432,
-0.038217853754758835,
-0.21305876970291138,
0.00006865025352453813,
-0.16019634902477264,
-0.17082053422927856,
-0.02853325940668583,
0.0575391910970211,
0.02511836402118206,
0.09691641479730606,
-0.0102996826171875,
-0.0006249685538932681,
0.11496102809906006,
-0.011418784037232399,
0.003148184157907963,
-0.09876816719770432,
0.07275080680847168,
-0.10547789186239243,
0.17341873049736023,
-0.005658212583512068,
0.13488352298736572,
0.12233132869005203,
0.029116308316588402,
-0.05841917172074318,
0.04392535984516144,
0.07102607935667038,
-0.08331143110990524,
-0.02166043221950531,
0.10590718686580658,
-0.04282274842262268,
0.06989102065563202,
0.057675376534461975,
-0.07799059897661209,
0.02133602648973465,
0.03168218582868576,
-0.024892766028642654,
-0.13048537075519562,
0.04740126430988312,
-0.05042845010757446,
0.15617802739143372,
0.1990928202867508,
-0.010293005034327507,
0.019112663343548775,
-0.0840204730629921,
0.05892527475953102,
0.024076955392956734,
0.09728965163230896,
-0.021307330578565598,
-0.21744884550571442,
0.05117170140147209,
0.022260989993810654,
0.01884027011692524,
-0.22092022001743317,
-0.08351928740739822,
0.032170627266168594,
-0.035389386117458344,
-0.03876558318734169,
0.13706903159618378,
0.029112132266163826,
0.04592430219054222,
-0.02848862111568451,
-0.08896195888519287,
-0.06575796008110046,
0.12397107481956482,
-0.07823614776134491,
-0.06926707178354263
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-ner
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the conll2003 dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0602
- Precision: 0.9251
- Recall: 0.9370
- F1: 0.9310
- Accuracy: 0.9839
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.2435 | 1.0 | 878 | 0.0685 | 0.9182 | 0.9221 | 0.9202 | 0.9816 |
| 0.0515 | 2.0 | 1756 | 0.0602 | 0.9212 | 0.9368 | 0.9289 | 0.9834 |
| 0.0301 | 3.0 | 2634 | 0.0602 | 0.9251 | 0.9370 | 0.9310 | 0.9839 |
### Framework versions
- Transformers 4.11.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["conll2003"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "distilbert-base-uncased-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "conll2003", "type": "conll2003", "args": "conll2003"}, "metrics": [{"type": "precision", "value": 0.925115970841617, "name": "Precision"}, {"type": "recall", "value": 0.9370175634858485, "name": "Recall"}, {"type": "f1", "value": 0.9310287333963209, "name": "F1"}, {"type": "accuracy", "value": 0.9839388692074285, "name": "Accuracy"}]}]}]}
|
token-classification
|
delpart/distilbert-base-uncased-finetuned-ner
|
[
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"token-classification",
"generated_from_trainer",
"dataset:conll2003",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
distilbert-base-uncased-finetuned-ner
=====================================
This model is a fine-tuned version of distilbert-base-uncased on the conll2003 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0602
* Precision: 0.9251
* Recall: 0.9370
* F1: 0.9310
* Accuracy: 0.9839
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.11.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
69,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.11.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
-0.10729619115591049,
0.11170487105846405,
-0.0024668413680046797,
0.1329026073217392,
0.15346159040927887,
0.03060285374522209,
0.12314734607934952,
0.1130605936050415,
-0.08957827836275101,
0.025340307503938675,
0.13137179613113403,
0.1620047688484192,
0.014531397260725498,
0.11554904282093048,
-0.05175654962658882,
-0.24741321802139282,
-0.003778316080570221,
0.04328009858727455,
-0.04983799532055855,
0.13184985518455505,
0.09834866225719452,
-0.13277266919612885,
0.09422474354505539,
0.015143224969506264,
-0.19505620002746582,
-0.008028320968151093,
0.001975785242393613,
-0.05070829018950462,
0.14331135153770447,
0.016054434701800346,
0.1263270527124405,
-0.011164039373397827,
0.0948653295636177,
-0.17470844089984894,
0.0054352073930203915,
0.04361623898148537,
0.011913781054317951,
0.09577621519565582,
0.04120645299553871,
0.013048973865807056,
0.0949927419424057,
-0.06047135964035988,
0.05952925235033035,
0.012342672795057297,
-0.11941202729940414,
-0.2063056081533432,
-0.09214145690202713,
0.04734867811203003,
0.08381910622119904,
0.09775616228580475,
0.004027715418487787,
0.13780540227890015,
-0.08952639251947403,
0.08583863079547882,
0.20676785707473755,
-0.28458669781684875,
-0.06741346418857574,
0.04864529147744179,
0.010625599883496761,
0.04225083068013191,
-0.09846386313438416,
-0.044529687613248825,
0.04290497675538063,
0.05022360756993294,
0.13130420446395874,
-0.02805555798113346,
-0.11094546318054199,
0.012854020111262798,
-0.14317210018634796,
-0.04298195242881775,
0.16651012003421783,
0.04978102073073387,
-0.034793682396411896,
-0.04215110093355179,
-0.06355241686105728,
-0.16432198882102966,
-0.028256183490157127,
-0.013409837149083614,
0.04422791302204132,
-0.02658286690711975,
-0.055370766669511795,
0.002057285513728857,
-0.10065022855997086,
-0.06789647787809372,
-0.08063400536775589,
0.13343188166618347,
0.03530052304267883,
0.01535102166235447,
-0.020968226715922356,
0.11493100225925446,
0.0009730879100970924,
-0.1214853897690773,
0.021477220579981804,
0.02245979569852352,
0.002994735259562731,
-0.04603695869445801,
-0.05144331976771355,
-0.042642831802368164,
0.007270873989909887,
0.14587271213531494,
-0.0316849946975708,
0.03321457281708717,
0.054724209010601044,
0.04473569616675377,
-0.08711246401071548,
0.18306869268417358,
-0.04335803911089897,
-0.033172909170389175,
0.00688775023445487,
0.05382867529988289,
0.02479442209005356,
-0.004200859926640987,
-0.12466777116060257,
0.011393086053431034,
0.09770113974809647,
0.008696499280631542,
-0.06615270674228668,
0.06474265456199646,
-0.06295792013406754,
-0.02849176712334156,
0.019715456292033195,
-0.08536608517169952,
0.02831595577299595,
-0.010270304046571255,
-0.07990933209657669,
-0.022931229323148727,
0.017436621710658073,
0.0242549367249012,
-0.0035892874002456665,
0.1107342466711998,
-0.09624379873275757,
0.020481567829847336,
-0.0891905426979065,
-0.09974444657564163,
0.016905875876545906,
-0.10938262194395065,
0.03216332197189331,
-0.09532900154590607,
-0.19558550417423248,
-0.004062084015458822,
0.06435150653123856,
-0.023050464689731598,
-0.07049861550331116,
-0.04508310928940773,
-0.06704133749008179,
0.008156074211001396,
-0.00980397965759039,
0.12002138048410416,
-0.0654280036687851,
0.08958815038204193,
0.020265208557248116,
0.05810799449682236,
-0.052016593515872955,
0.05228670686483383,
-0.10866822302341461,
0.025015292689204216,
-0.15379709005355835,
0.030472084879875183,
-0.04799121245741844,
0.0643300861120224,
-0.08899099379777908,
-0.1001804918050766,
0.018022004514932632,
-0.018667452037334442,
0.06544941663742065,
0.08557108044624329,
-0.1849309355020523,
-0.06276530772447586,
0.13580062985420227,
-0.062083929777145386,
-0.121910959482193,
0.12359210848808289,
-0.06704094260931015,
0.042326420545578,
0.056975264102220535,
0.15301594138145447,
0.06862390041351318,
-0.07548070698976517,
0.00460656126961112,
0.010723439045250416,
0.05090536177158356,
-0.06200602278113365,
0.07691055536270142,
0.00595046067610383,
0.016870375722646713,
0.029190922155976295,
-0.03576023504137993,
0.05557539314031601,
-0.08896438032388687,
-0.10098562389612198,
-0.039299361407756805,
-0.09750738739967346,
0.046542707830667496,
0.06305907666683197,
0.06454874575138092,
-0.08857031166553497,
-0.07507462799549103,
0.05208577588200569,
0.09089350700378418,
-0.04467424377799034,
0.020153425633907318,
-0.06451145559549332,
0.07948905974626541,
-0.04533098265528679,
-0.03157876059412956,
-0.1743668019771576,
-0.03329304978251457,
0.01304448489099741,
0.0012421895517036319,
0.014247995801270008,
0.026272494345903397,
0.06222059205174446,
0.07379619032144547,
-0.04239821434020996,
-0.018828298896551132,
-0.03548905625939369,
0.005470484960824251,
-0.13122984766960144,
-0.19218336045742035,
-0.04484022408723831,
-0.019892705604434013,
0.15460105240345,
-0.2024933099746704,
0.0347650907933712,
-0.02564200758934021,
0.08691908419132233,
0.01578785479068756,
-0.015131190419197083,
-0.04285942390561104,
0.0687083899974823,
-0.049372173845767975,
-0.0537099651992321,
0.06535541266202927,
0.011375014670193195,
-0.09091871231794357,
-0.06632410734891891,
-0.08743926882743835,
0.16345930099487305,
0.12500053644180298,
-0.10102226585149765,
-0.07455809414386749,
-0.017632661387324333,
-0.06487873196601868,
-0.034873589873313904,
-0.04977201297879219,
0.02892371267080307,
0.17380523681640625,
-0.004972628317773342,
0.14213183522224426,
-0.06989419460296631,
-0.04512563347816467,
0.020711911842226982,
-0.03422049805521965,
0.02041296847164631,
0.11570698022842407,
0.1349744349718094,
-0.08259256184101105,
0.1521575003862381,
0.15385323762893677,
-0.0918005183339119,
0.11574441939592361,
-0.03957599028944969,
-0.062291089445352554,
-0.027094969525933266,
-0.02844499796628952,
-0.007790595293045044,
0.11545468866825104,
-0.1396806240081787,
0.008269189856946468,
0.035982780158519745,
0.021910402923822403,
0.00966420117765665,
-0.22049811482429504,
-0.041855111718177795,
0.03762253373861313,
-0.033527880907058716,
-0.006856957450509071,
-0.010856078006327152,
0.005852594040334225,
0.09893051534891129,
0.004399863071739674,
-0.10638973861932755,
0.04695836082100868,
0.008673450909554958,
-0.07249434292316437,
0.20483149588108063,
-0.08632659167051315,
-0.14148110151290894,
-0.1244591772556305,
-0.086214579641819,
-0.057209718972444534,
0.010847263038158417,
0.05373748391866684,
-0.07196866720914841,
-0.03568828105926514,
-0.07299581170082092,
0.0025222226977348328,
0.0010470461566001177,
0.028807690367102623,
0.016221003606915474,
-0.00829541776329279,
0.06641058623790741,
-0.10579152405261993,
-0.011925088241696358,
-0.05143450200557709,
-0.04870337247848511,
0.03715106472373009,
0.040987852960824966,
0.11477607488632202,
0.1507202833890915,
-0.011805986054241657,
0.006863725371658802,
-0.02104128710925579,
0.254359632730484,
-0.059132106602191925,
-0.019506633281707764,
0.13643833994865417,
-0.019915422424674034,
0.053541071712970734,
0.12202666699886322,
0.07633916288614273,
-0.08320034295320511,
-0.002191669074818492,
0.031527865678071976,
-0.03968091309070587,
-0.21098482608795166,
-0.05349694937467575,
-0.05528696998953819,
-0.005825080908834934,
0.09846484661102295,
0.023991234600543976,
0.03804696351289749,
0.0812789648771286,
0.03890794515609741,
0.09465791285037994,
-0.05167406424880028,
0.06304774433374405,
0.12080629169940948,
0.04518258199095726,
0.12332431226968765,
-0.03167833760380745,
-0.06199818477034569,
0.0462363064289093,
0.0065183453261852264,
0.22331787645816803,
0.013073262758553028,
0.12680551409721375,
0.06065293028950691,
0.18041524291038513,
-0.01043400913476944,
0.07700113952159882,
-0.00925402157008648,
-0.03181663528084755,
-0.021962087601423264,
-0.0370476208627224,
-0.03831987828016281,
0.028008071705698967,
-0.0566452257335186,
0.0726679340004921,
-0.10382122546434402,
0.02154398150742054,
0.05224856734275818,
0.25523263216018677,
0.03812659904360771,
-0.3385065793991089,
-0.09882564842700958,
0.0008044586866162717,
-0.03590685501694679,
-0.022538477554917336,
0.029411371797323227,
0.0826641097664833,
-0.09652464836835861,
0.020244376733899117,
-0.06554383784532547,
0.09141963720321655,
-0.05198970437049866,
0.04149988293647766,
0.08219318091869354,
0.0895611122250557,
0.012891626916825771,
0.0858054980635643,
-0.2705467939376831,
0.2704070508480072,
0.0014750811969861388,
0.05822845548391342,
-0.07839677482843399,
0.007010975852608681,
0.03458445891737938,
0.0632973164319992,
0.07320911437273026,
-0.0048237633891403675,
-0.02048802189528942,
-0.1971307247877121,
-0.06253495812416077,
0.021863088011741638,
0.05906146392226219,
-0.04209953173995018,
0.0884133130311966,
-0.03033582493662834,
0.008852189406752586,
0.06806191056966782,
0.007776471320539713,
-0.046157870441675186,
-0.09981688112020493,
-0.005984138697385788,
0.036186471581459045,
-0.04616936668753624,
-0.06199472025036812,
-0.10891903936862946,
-0.12179164588451385,
0.14247149229049683,
-0.032432470470666885,
-0.03729063272476196,
-0.10672535747289658,
0.07499585300683975,
0.08191166818141937,
-0.08270778506994247,
0.05107452720403671,
-0.005004670936614275,
0.07510897517204285,
0.031415387988090515,
-0.058666788041591644,
0.09963223338127136,
-0.08033344894647598,
-0.16872835159301758,
-0.07281523942947388,
0.1039537638425827,
0.037940386682748795,
0.06544226408004761,
-0.005945255979895592,
0.017806904390454292,
-0.05006497725844383,
-0.08899158239364624,
0.02381988614797592,
0.00008579312998335809,
0.08728152513504028,
0.015597648918628693,
-0.05004236474633217,
0.02744712494313717,
-0.05364499241113663,
-0.03272018954157829,
0.18946634232997894,
0.23378294706344604,
-0.10214816778898239,
0.019333843141794205,
0.028992829844355583,
-0.0618724599480629,
-0.17617851495742798,
0.02539622224867344,
0.053762804716825485,
0.0035556897055357695,
0.041777804493904114,
-0.1772610992193222,
0.14459235966205597,
0.11825281381607056,
-0.01866327039897442,
0.10524943470954895,
-0.3270835280418396,
-0.11941244453191757,
0.132179856300354,
0.13269273936748505,
0.10739506036043167,
-0.12331008166074753,
-0.02056019939482212,
-0.01745595596730709,
-0.14788946509361267,
0.11466430872678757,
-0.0726938396692276,
0.11370190232992172,
-0.0328618586063385,
0.09904667735099792,
0.0027053060475736856,
-0.057434432208538055,
0.12593679130077362,
0.03735506534576416,
0.09923868626356125,
-0.0572768971323967,
-0.04208608344197273,
0.032406412065029144,
-0.042014531791210175,
0.02296132594347,
-0.07974343746900558,
0.03738671541213989,
-0.10828760266304016,
-0.019727252423763275,
-0.06493612378835678,
0.04088207334280014,
-0.03484637662768364,
-0.07402781397104263,
-0.04357760399580002,
0.02989637665450573,
0.054297562688589096,
-0.011324381455779076,
0.1300516426563263,
0.048618484288454056,
0.13201722502708435,
0.09938982129096985,
0.06418218463659286,
-0.07642985880374908,
-0.08717020601034164,
-0.03050536848604679,
-0.01554948277771473,
0.058313868939876556,
-0.11793182790279388,
0.025276485830545425,
0.1447371244430542,
0.024876300245523453,
0.13764885067939758,
0.08141524344682693,
-0.01642945595085621,
0.006407063454389572,
0.05070642754435539,
-0.16926950216293335,
-0.07092011719942093,
-0.0018501817248761654,
-0.03667958080768585,
-0.11953166872262955,
0.05171406641602516,
0.09458079189062119,
-0.07017292082309723,
-0.008449258282780647,
-0.003385145217180252,
0.01324472576379776,
-0.04985802248120308,
0.19044886529445648,
0.05620422586798668,
0.04734309762716293,
-0.10155860334634781,
0.07125312089920044,
0.05546463280916214,
-0.05312572419643402,
-0.004817468114197254,
0.04763636738061905,
-0.0895695835351944,
-0.04248721897602081,
0.04928120598196983,
0.16901926696300507,
-0.07028555870056152,
-0.042853280901908875,
-0.13114990293979645,
-0.11482655256986618,
0.07960633933544159,
0.13612085580825806,
0.11740070581436157,
0.016423141583800316,
-0.06743282824754715,
0.0004952854942530394,
-0.10914558172225952,
0.09815873205661774,
0.048028554767370224,
0.07354472577571869,
-0.15660366415977478,
0.136787548661232,
0.0037395390681922436,
0.04061603918671608,
-0.015728900209069252,
0.02801499329507351,
-0.09362076222896576,
0.007902473211288452,
-0.11387563496828079,
-0.021407097578048706,
-0.03822539374232292,
0.013804258778691292,
-0.004424482583999634,
-0.05764398351311684,
-0.056014854460954666,
0.014140218496322632,
-0.10740665346384048,
-0.019148893654346466,
0.039114464074373245,
0.06230781227350235,
-0.11212554574012756,
-0.037895139306783676,
0.03012268617749214,
-0.06077441945672035,
0.07645225524902344,
0.04563285410404205,
0.02573278173804283,
0.041702594608068466,
-0.11940252780914307,
0.011790977790951729,
0.06615324318408966,
0.029435839504003525,
0.07743357867002487,
-0.1001812294125557,
-0.014399581588804722,
-0.003340624738484621,
0.03729750216007233,
0.013967427425086498,
0.07698642462491989,
-0.13875888288021088,
-0.009423704817891121,
-0.01075095683336258,
-0.0788838118314743,
-0.06491846591234207,
0.017123522236943245,
0.10400708764791489,
0.016586558893322945,
0.21097970008850098,
-0.06027191877365112,
0.04431015998125076,
-0.20626592636108398,
0.0022008027881383896,
-0.009128287434577942,
-0.1071648895740509,
-0.13156342506408691,
-0.06123747676610947,
0.05023034289479256,
-0.057487938553094864,
0.1542365849018097,
0.026944926008582115,
0.024933261796832085,
0.021405505016446114,
0.0033029967453330755,
0.021186215803027153,
0.00982475746423006,
0.19376583397388458,
0.041682109236717224,
-0.03585534542798996,
0.05682689696550369,
0.03945025056600571,
0.10432559251785278,
0.10283724218606949,
0.18677854537963867,
0.1381937861442566,
-0.0009193746955133975,
0.0860845148563385,
0.03772783651947975,
-0.06551966816186905,
-0.17457754909992218,
0.032378144562244415,
-0.036469414830207825,
0.10622807592153549,
-0.014869372360408306,
0.22802293300628662,
0.05521994084119797,
-0.16804608702659607,
0.0336894765496254,
-0.05021171271800995,
-0.08085586130619049,
-0.1010449007153511,
-0.06474733352661133,
-0.07695990800857544,
-0.12557187676429749,
-0.0007782362517900765,
-0.1108352541923523,
0.007293099537491798,
0.12870629131793976,
0.004985456820577383,
-0.024930808693170547,
0.14499539136886597,
0.0029434985481202602,
0.03848128020763397,
0.037979163229465485,
0.01326993852853775,
-0.0356140211224556,
-0.11110391467809677,
-0.07286140322685242,
-0.024722348898649216,
-0.015892574563622475,
0.036995403468608856,
-0.07153373956680298,
-0.03660186380147934,
0.026550576090812683,
-0.01143906731158495,
-0.09137442708015442,
0.006581400986760855,
0.0056204586289823055,
0.050336163491010666,
0.034516822546720505,
0.006004689261317253,
0.03598934784531593,
-0.008478443138301373,
0.19645535945892334,
-0.0734151229262352,
-0.0639379620552063,
-0.10732303559780121,
0.2284800261259079,
0.027836710214614868,
-0.021667765453457832,
0.04125048592686653,
-0.06545688956975937,
0.005906038451939821,
0.23232021927833557,
0.19998660683631897,
-0.09672721475362778,
-0.013669179752469063,
0.009719662368297577,
-0.013909703120589256,
-0.032967183738946915,
0.09354116767644882,
0.13035742938518524,
0.04509155452251434,
-0.09029502421617508,
-0.03952697664499283,
-0.07052641361951828,
-0.013187889941036701,
-0.034000858664512634,
0.05733012780547142,
0.041899245232343674,
0.005976204294711351,
-0.043792691081762314,
0.048430319875478745,
-0.06980261951684952,
-0.09074433147907257,
0.06209402531385422,
-0.1989745795726776,
-0.16924557089805603,
-0.010314646176993847,
0.09854589402675629,
0.0047895037569105625,
0.05862811952829361,
-0.034094102680683136,
-0.0015987419756129384,
0.08490732312202454,
-0.01967727579176426,
-0.09373173117637634,
-0.08199513703584671,
0.10624765604734421,
-0.07257609069347382,
0.22969235479831696,
-0.04523940756917,
0.07290603965520859,
0.12197079509496689,
0.06717532873153687,
-0.08179105073213577,
0.05630909651517868,
0.05498430132865906,
-0.05432926118373871,
0.02104112133383751,
0.0689741000533104,
-0.02684624120593071,
0.0823809802532196,
0.04461933299899101,
-0.13230682909488678,
0.011406057514250278,
-0.04306725040078163,
-0.054303012788295746,
-0.04578050598502159,
-0.033674053847789764,
-0.05521288141608238,
0.13873381912708282,
0.20776568353176117,
-0.035388726741075516,
-0.014461551792919636,
-0.06878078728914261,
0.02284581959247589,
0.06126018613576889,
0.008651421405375004,
-0.06242920830845833,
-0.2158544957637787,
0.016715748235583305,
0.04344794899225235,
-0.019845418632030487,
-0.2117632031440735,
-0.10337693244218826,
0.0007377088186331093,
-0.07528538256883621,
-0.08605077862739563,
0.07068557292222977,
0.08122121542692184,
0.0512462854385376,
-0.05990610644221306,
-0.026466837152838707,
-0.08433442562818527,
0.13527187705039978,
-0.13550008833408356,
-0.08976750075817108
] |
null | null |
transformers
|
#DialoGPT medium based model of Dwight Schrute, trained with 10 context lines of history for 20 epochs.
|
{"tags": ["conversational"]}
|
text-generation
|
delvan/DialoGPT-medium-DwightV1
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
#DialoGPT medium based model of Dwight Schrute, trained with 10 context lines of history for 20 epochs.
|
[] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
-0.009697278961539268,
0.03208012506365776,
-0.007204889785498381,
0.004809224978089333,
0.16726240515708923,
0.014898733235895634,
0.09765533357858658,
0.13672804832458496,
-0.007841327227652073,
-0.031050153076648712,
0.14490588009357452,
0.20411323010921478,
-0.006439372431486845,
0.0661218985915184,
-0.07572533935308456,
-0.2683109939098358,
0.05759621039032936,
0.046649303287267685,
0.016515716910362244,
0.1200079694390297,
0.08573378622531891,
-0.05473608896136284,
0.08714032918214798,
-0.014583407901227474,
-0.150366872549057,
0.017733458429574966,
0.043394338339567184,
-0.12260226160287857,
0.11910516023635864,
0.05462685227394104,
0.07063519209623337,
0.014929565601050854,
-0.07541623711585999,
-0.1631229966878891,
0.03031250834465027,
0.01425902172923088,
-0.0594632662832737,
0.04757995903491974,
0.059961482882499695,
-0.10165371745824814,
0.10819483548402786,
0.09530027210712433,
-0.013078106567263603,
0.06798283755779266,
-0.16849711537361145,
-0.020869607105851173,
-0.01446688175201416,
0.009899779222905636,
0.05550243332982063,
0.09964893013238907,
-0.03413357585668564,
0.10497362166643143,
-0.09214533120393753,
0.11017382889986038,
0.10932035744190216,
-0.32057443261146545,
-0.005767723545432091,
0.09167823940515518,
0.039358653128147125,
0.07352814823389053,
-0.04467793554067612,
0.06258884817361832,
0.018015462905168533,
0.017986174672842026,
-0.014015024527907372,
-0.07283061742782593,
-0.11612214148044586,
0.04717336222529411,
-0.08668071031570435,
-0.059868961572647095,
0.2244078367948532,
-0.05464440956711769,
0.06881742179393768,
-0.05281897634267807,
-0.10522868484258652,
-0.04308144748210907,
-0.029833965003490448,
0.00475557055324316,
-0.07660607248544693,
0.08692064881324768,
0.00869679357856512,
-0.09547875821590424,
-0.1376667022705078,
-0.02496783249080181,
-0.1776352822780609,
0.16140350699424744,
0.02465328387916088,
0.05232657864689827,
-0.2027255892753601,
0.09623090922832489,
0.017906051129102707,
-0.08045592904090881,
0.022091427817940712,
-0.10046248883008957,
0.029131146147847176,
0.013760408386588097,
-0.04754498973488808,
-0.061387211084365845,
0.0843690037727356,
0.11199145019054413,
-0.01731434464454651,
0.025486016646027565,
-0.039331406354904175,
0.08100687712430954,
0.03553595021367073,
0.09077847748994827,
0.007288969587534666,
-0.028338588774204254,
0.025842782109975815,
-0.13719046115875244,
-0.003647835226729512,
-0.07116208970546722,
-0.16572439670562744,
-0.021088803187012672,
0.02994808368384838,
0.08289173990488052,
0.015449047088623047,
0.11682453751564026,
-0.03272046521306038,
-0.025152435526251793,
0.03602350503206253,
-0.047656361013650894,
-0.012649794109165668,
0.016648368909955025,
0.013163427822291851,
0.12399329990148544,
-0.0022096503525972366,
0.03235051408410072,
-0.13653022050857544,
0.031423524022102356,
-0.06793295592069626,
-0.003740974934771657,
-0.03486552834510803,
-0.040637075901031494,
0.009043924510478973,
-0.06862333416938782,
0.003486064961180091,
-0.15030112862586975,
-0.15063877403736115,
0.007587034720927477,
-0.007836631499230862,
-0.04107699543237686,
-0.06370922178030014,
-0.06952770054340363,
-0.013550350442528725,
0.04251532256603241,
-0.07093454152345657,
-0.011352915316820145,
-0.06403283774852753,
0.11004766076803207,
-0.03197755664587021,
0.07921615242958069,
-0.11953279376029968,
0.08390819281339645,
-0.11260783672332764,
-0.02386913076043129,
-0.060801517218351364,
0.09317506104707718,
-0.0006014376995153725,
0.09549830108880997,
-0.006563255097717047,
-0.017931854352355003,
-0.07981178909540176,
0.06445012241601944,
-0.042872510850429535,
0.21701598167419434,
-0.0615808479487896,
-0.11181682348251343,
0.28781595826148987,
-0.052628401666879654,
-0.1370542049407959,
0.11647392809391022,
0.008682746440172195,
0.05777018144726753,
0.10703510791063309,
0.19733482599258423,
-0.015276194550096989,
0.004040541127324104,
0.09471915662288666,
0.11263324320316315,
-0.11276852339506149,
-0.033160366117954254,
0.013019153848290443,
-0.04081077128648758,
-0.10867965966463089,
0.04689536616206169,
0.09810488671064377,
0.07090286910533905,
-0.04786505550146103,
-0.03377414867281914,
-0.01366397924721241,
0.0052589005790650845,
0.08885077387094498,
-0.007157256826758385,
0.10962837189435959,
-0.05819983780384064,
-0.03796621412038803,
-0.029282379895448685,
-0.012126247398555279,
-0.03951939567923546,
0.03137664496898651,
-0.043376367539167404,
0.10821941494941711,
-0.011204327456653118,
0.06364280730485916,
-0.16185984015464783,
-0.07691477984189987,
-0.017002692446112633,
0.1581239402294159,
0.024538565427064896,
0.09859629720449448,
0.0552486926317215,
-0.040398042649030685,
-0.0012767292791977525,
0.012792680412530899,
0.15581141412258148,
-0.022091681137681007,
-0.065607450902462,
-0.052166227251291275,
0.08642971515655518,
-0.05641226842999458,
0.04504093527793884,
-0.05937713757157326,
0.012367865070700645,
0.05064384639263153,
0.10342344641685486,
-0.00018274025933351368,
0.03323284164071083,
-0.008164864964783192,
0.002145637758076191,
-0.058205123990774155,
0.007405933458358049,
0.10799351334571838,
0.00036868182360194623,
-0.07365862280130386,
0.22074243426322937,
-0.17796069383621216,
0.1765957772731781,
0.1893044263124466,
-0.299345999956131,
0.017949223518371582,
-0.10759581625461578,
-0.04561871662735939,
0.014407722279429436,
0.05567655712366104,
-0.0454222597181797,
0.1703362911939621,
-0.009871348738670349,
0.18874616920948029,
-0.04946064203977585,
-0.04464937001466751,
-0.0200483538210392,
-0.05118836089968681,
-0.0024189651012420654,
0.07781197130680084,
0.10685696452856064,
-0.13992026448249817,
0.1964332014322281,
0.1621224284172058,
0.048237916082143784,
0.19945049285888672,
0.015346456319093704,
-0.011589210480451584,
0.0909530371427536,
0.005220826715230942,
-0.058739423751831055,
-0.07409929484128952,
-0.2594851851463318,
-0.030033592134714127,
0.07992640137672424,
0.0422382652759552,
0.1212305948138237,
-0.11349532753229141,
-0.038956157863140106,
-0.01763172075152397,
-0.023146281018853188,
0.021672505885362625,
0.0914369598031044,
0.06075398623943329,
0.13201528787612915,
-0.001710098935291171,
-0.007300339173525572,
0.10524573177099228,
0.01783694699406624,
-0.09354141354560852,
0.18308524787425995,
-0.13652534782886505,
-0.37097251415252686,
-0.13911493122577667,
-0.18057456612586975,
-0.05449081212282181,
0.05712554603815079,
0.11679314076900482,
-0.12011238187551498,
-0.018752124160528183,
0.01578843593597412,
0.10931742936372757,
-0.08449502289295197,
0.0021454424131661654,
-0.06880278885364532,
0.0321490578353405,
-0.10310184955596924,
-0.09194442629814148,
-0.055416494607925415,
-0.031392451375722885,
-0.08001253753900528,
0.1423761546611786,
-0.10777941346168518,
0.04476889222860336,
0.20262959599494934,
0.04653622955083847,
0.05625178664922714,
-0.044105201959609985,
0.19377262890338898,
-0.11264272034168243,
-0.01661740615963936,
0.19215328991413116,
-0.048360925167798996,
0.07476246356964111,
0.1232115849852562,
-0.006348740309476852,
-0.08765771239995956,
0.03011748194694519,
-0.02085109055042267,
-0.07988511025905609,
-0.23219464719295502,
-0.13938382267951965,
-0.12429051846265793,
0.09477275609970093,
0.028005298227071762,
0.056365787982940674,
0.17219258844852448,
0.06577219814062119,
-0.038416244089603424,
0.006410336587578058,
0.02959546446800232,
0.08237514644861221,
0.23417828977108002,
-0.06035616248846054,
0.1364797055721283,
-0.03420931473374367,
-0.14982740581035614,
0.08169995993375778,
0.0713929831981659,
0.10213395953178406,
0.06678459793329239,
0.0804823637008667,
0.0149586396291852,
0.06188136339187622,
0.1311223804950714,
0.08191446959972382,
0.019586285576224327,
-0.02480296604335308,
-0.03388110175728798,
-0.025523077696561813,
-0.05937909707427025,
0.040128443390131,
0.06589099019765854,
-0.16763372719287872,
-0.039227183908224106,
-0.09338314831256866,
0.09657008945941925,
0.0873042419552803,
0.06609832495450974,
-0.1842060089111328,
-0.008006223477423191,
0.08488986641168594,
-0.03854905813932419,
-0.13727426528930664,
0.09535189718008041,
0.01523482333868742,
-0.15144726634025574,
0.03139317408204079,
-0.04061909019947052,
0.12188644707202911,
-0.07804752141237259,
0.09809603542089462,
-0.08108244836330414,
-0.07448557764291763,
0.02123199962079525,
0.1261177361011505,
-0.30527687072753906,
0.20240111649036407,
-0.0024993624538183212,
-0.06486981362104416,
-0.1243603527545929,
-0.0032166161108762026,
0.002410882618278265,
0.07357452809810638,
0.10519039630889893,
-0.007196315098553896,
0.001897757756523788,
-0.06300821900367737,
-0.01829923689365387,
0.032471053302288055,
0.13080233335494995,
-0.0401318334043026,
-0.021158374845981598,
-0.050194524228572845,
-0.001653497340157628,
-0.03173094615340233,
-0.06934895366430283,
0.02002747356891632,
-0.19509181380271912,
0.08751901984214783,
0.04166261479258537,
0.09648149460554123,
0.029994789510965347,
0.004265148192644119,
-0.09651939570903778,
0.24698667228221893,
-0.07148019969463348,
-0.10072879493236542,
-0.10919588059186935,
-0.046813901513814926,
0.03569883480668068,
-0.05628936365246773,
0.04309194162487984,
-0.0788632407784462,
0.028997479006648064,
-0.06352769583463669,
-0.19235502183437347,
0.12410202622413635,
-0.09027006477117538,
-0.04412810131907463,
-0.02371402643620968,
0.2110891044139862,
-0.05598580464720726,
0.010335659608244896,
0.02930437959730625,
0.01208863127976656,
-0.11645778268575668,
-0.09678568691015244,
0.031018631532788277,
-0.007351789623498917,
0.050603240728378296,
0.041841957718133926,
-0.05915454775094986,
-0.017138581722974777,
-0.052199993282556534,
-0.022926922887563705,
0.3496883809566498,
0.14231905341148376,
-0.043836336582899094,
0.19347235560417175,
0.12347975373268127,
-0.07452994585037231,
-0.3159443140029907,
-0.1066238060593605,
-0.10937739163637161,
-0.04680149629712105,
-0.07012093812227249,
-0.2002030611038208,
0.06474938243627548,
0.00662544509395957,
-0.013415241613984108,
0.12749312818050385,
-0.2561831772327423,
-0.07571036368608475,
0.15906259417533875,
-0.017980827018618584,
0.3745945692062378,
-0.1168576180934906,
-0.10926306992769241,
-0.03950892388820648,
-0.14175476133823395,
0.16968177258968353,
-0.01989765651524067,
0.11221715062856674,
-0.009765521623194218,
0.14388824999332428,
0.05548359826207161,
-0.023479344323277473,
0.08544106781482697,
0.004999885335564613,
-0.03290518373250961,
-0.10304180532693863,
-0.05676887184381485,
0.007092386484146118,
0.02477436140179634,
0.018026655539870262,
-0.041834570467472076,
0.02227151393890381,
-0.11731979995965958,
-0.04657655209302902,
-0.08982590585947037,
0.04431166127324104,
0.03899754583835602,
-0.07325074821710587,
-0.002380647463724017,
-0.07165111601352692,
-0.012272949330508709,
0.022334342822432518,
0.20356793701648712,
-0.08029330521821976,
0.16448934376239777,
0.09239562600851059,
0.12419285625219345,
-0.14376309514045715,
-0.00019283240544609725,
-0.0762530043721199,
-0.05611240118741989,
0.07737895101308823,
-0.09433035552501678,
0.058893077075481415,
0.10901971161365509,
-0.04567738622426987,
0.08828683942556381,
0.10377411544322968,
0.008936077356338501,
0.003213887568563223,
0.10916902124881744,
-0.2667325437068939,
-0.0296600554138422,
-0.07532413303852081,
0.000883326749317348,
0.09092561900615692,
0.08562852442264557,
0.18840822577476501,
0.025361526757478714,
-0.04293036088347435,
-0.002770674182102084,
0.028597986325621605,
-0.039021048694849014,
0.051667019724845886,
0.001123449532315135,
0.01947369985282421,
-0.1530752182006836,
0.072522833943367,
0.01490565575659275,
-0.15215420722961426,
0.021316176280379295,
0.16572684049606323,
-0.11656328290700912,
-0.1283872276544571,
-0.06520111113786697,
0.08313824236392975,
-0.11755692958831787,
-0.01578943058848381,
-0.03279297426342964,
-0.13145680725574493,
0.07992171496152878,
0.12629036605358124,
0.05557859688997269,
0.0972496047616005,
-0.06061713397502899,
-0.020469192415475845,
-0.018721895292401314,
-0.014099318534135818,
-0.012384648434817791,
-0.007667020428925753,
-0.055978111922740936,
0.0590752474963665,
-0.026677248999476433,
0.1425808072090149,
-0.09221141785383224,
-0.1037059873342514,
-0.16142144799232483,
0.0374140702188015,
-0.11013076454401016,
-0.08825794607400894,
-0.08821134269237518,
-0.050188567489385605,
0.002360827289521694,
-0.019856395199894905,
-0.04037635400891304,
-0.05829505994915962,
-0.12300454825162888,
0.0338277705013752,
-0.040771447122097015,
0.024727050215005875,
-0.07512269169092178,
0.015856385231018066,
0.08507686108350754,
-0.03285100311040878,
0.15655414760112762,
0.1450488418340683,
-0.1006515845656395,
0.10741901397705078,
-0.14806775748729706,
-0.09138492494821548,
0.11116421222686768,
0.015329592861235142,
0.0449691042304039,
0.09723787009716034,
0.013362943194806576,
0.0635865181684494,
0.032776717096567154,
0.05308786407113075,
0.027619892731308937,
-0.11959987878799438,
0.06483134627342224,
-0.03626115620136261,
-0.14700546860694885,
-0.049338050186634064,
-0.05282869189977646,
0.01647452637553215,
0.013054544106125832,
0.09622690081596375,
-0.05301849544048309,
0.10698331147432327,
-0.04055701196193695,
0.0346808135509491,
0.017554637044668198,
-0.1730053424835205,
-0.03816922754049301,
-0.08538098633289337,
0.03681723028421402,
0.014741539023816586,
0.25266793370246887,
0.030072299763560295,
0.012416383251547813,
0.032671261578798294,
0.08285367488861084,
0.03899408504366875,
0.010228337720036507,
0.17482228577136993,
0.1162426546216011,
-0.06621865928173065,
-0.10445023328065872,
0.0729617029428482,
0.016332454979419708,
0.01286179106682539,
0.13617953658103943,
0.008365051820874214,
0.005795429926365614,
0.08649782836437225,
-0.016865963116288185,
0.009968153201043606,
-0.10052056610584259,
-0.13426925241947174,
-0.022176474332809448,
0.05151832848787308,
-0.04655967652797699,
0.11727844923734665,
0.1406494379043579,
-0.01806013658642769,
0.03222079202532768,
-0.021771740168333054,
-0.05699979141354561,
-0.1683429479598999,
-0.1429590880870819,
-0.06883849948644638,
-0.13416796922683716,
0.00897989235818386,
-0.11180389672517776,
0.05395037308335304,
0.06001098081469536,
0.06750501692295074,
-0.06899319589138031,
0.10220931470394135,
0.04626858979463577,
-0.11440542340278625,
0.06264589726924896,
-0.0296088308095932,
0.09430401772260666,
-0.02759445086121559,
-0.019505485892295837,
-0.09039592742919922,
0.014574515633285046,
0.011419114656746387,
0.06245238706469536,
-0.04707273095846176,
0.007463190704584122,
-0.14696238934993744,
-0.08972041308879852,
-0.0523175448179245,
0.0718572810292244,
-0.050409089773893356,
0.14282815158367157,
0.00775480642914772,
-0.0170906875282526,
0.039554283022880554,
0.22787313163280487,
-0.07476283609867096,
-0.04778539761900902,
-0.05269690603017807,
0.20717895030975342,
0.02975541539490223,
0.1171872541308403,
-0.022938819602131844,
-0.006106364540755749,
-0.0919521227478981,
0.3764844834804535,
0.30030161142349243,
-0.09031439572572708,
0.011794124729931355,
0.02137952297925949,
0.04502861574292183,
0.1316293478012085,
0.1216534823179245,
0.10318691283464432,
0.3006802201271057,
-0.07452366501092911,
-0.04653361067175865,
-0.012629742734134197,
-0.023858042433857918,
-0.09059546142816544,
0.1021224707365036,
0.04839762672781944,
-0.06382183730602264,
-0.03313443064689636,
0.0954432487487793,
-0.25862133502960205,
0.1277991235256195,
-0.12311873584985733,
-0.17578600347042084,
-0.06654827296733856,
0.009760108776390553,
0.10465722531080246,
0.015642458572983742,
0.0946015790104866,
0.007128213066607714,
-0.11252258718013763,
0.06305865943431854,
0.03397420793771744,
-0.22762253880500793,
0.0006893770187161863,
0.06642123311758041,
-0.07006710022687912,
-0.0024247700348496437,
-0.026499588042497635,
0.05657242611050606,
0.0656052976846695,
0.054629553109407425,
-0.00971333310008049,
0.03816632181406021,
0.0034184439573436975,
-0.0585215799510479,
0.016623929142951965,
0.05121519789099693,
0.02472509816288948,
-0.09763528406620026,
0.06927435845136642,
-0.1574270874261856,
0.04766253009438515,
-0.0030655991286039352,
-0.04124255105853081,
0.006064958870410919,
0.008823691867291927,
-0.06491616368293762,
0.05165379121899605,
0.07916834205389023,
-0.0016257909592241049,
-0.0062433634884655476,
-0.057178743183612823,
-0.02632102556526661,
-0.027755750343203545,
-0.09291748702526093,
-0.10495562851428986,
-0.14682936668395996,
-0.11640441417694092,
0.09368976950645447,
-0.01011267676949501,
-0.1848134547472,
0.022154374048113823,
-0.08606051653623581,
0.08319322764873505,
-0.1670055389404297,
0.08040720224380493,
0.07041648775339127,
0.013038921169936657,
-0.0031511052511632442,
-0.02002427540719509,
0.054132770746946335,
0.086809903383255,
-0.10407156497240067,
-0.07400695979595184
] |
null | null |
transformers
|
This is finetune version of [SimCSE: Simple Contrastive Learning of Sentence Embeddings](https://arxiv.org/abs/2104.08821)
, train unsupervised on 570K stroke sentences from : stroke books, quora medical, quora's stroke and human annotates.
### Extract sentence representation
```
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("demdecuong/stroke_simcse")
model = AutoModel.from_pretrained("demdecuong/stroke_simcse")
text = "What are disease related to red stroke's causes?"
inputs = tokenizer(text, return_tensors='pt')
outputs = model(**inputs)[1]
```
### Build up embedding for database
```
database = [
'What is the daily checklist for stroke returning home',
'What are some tips for stroke adapt new life',
'What should I consider when using nursing-home care'
]
embedding = torch.zeros((len(database),768))
for i in range(len(database)):
inputs = tokenizer(database[i], return_tensors="pt")
outputs = model(**inputs)[1]
embedding[i] = outputs
print(embedding.shape)
```
### Result
On our Poc testset , which contains pairs of matching question related to stroke from human-generated.
| Model | Top-1 Accuracy |
| ------------- | ------------- |
| SimCSE (supervised) | 75.83 |
| SimCSE (ours) | 76.66 |
|
{}
|
feature-extraction
|
demdecuong/stroke_simcse
|
[
"transformers",
"pytorch",
"bert",
"feature-extraction",
"arxiv:2104.08821",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2104.08821"
] |
[] |
TAGS
#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #region-us
|
This is finetune version of SimCSE: Simple Contrastive Learning of Sentence Embeddings
, train unsupervised on 570K stroke sentences from : stroke books, quora medical, quora's stroke and human annotates.
### Extract sentence representation
### Build up embedding for database
### Result
On our Poc testset , which contains pairs of matching question related to stroke from human-generated.
|
[
"### Extract sentence representation",
"### Build up embedding for database",
"### Result\n\n\nOn our Poc testset , which contains pairs of matching question related to stroke from human-generated."
] |
[
"TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #region-us \n",
"### Extract sentence representation",
"### Build up embedding for database",
"### Result\n\n\nOn our Poc testset , which contains pairs of matching question related to stroke from human-generated."
] |
[
38,
7,
10,
29
] |
[
"passage: TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #region-us \n### Extract sentence representation### Build up embedding for database### Result\n\n\nOn our Poc testset , which contains pairs of matching question related to stroke from human-generated."
] |
[
-0.05554240942001343,
0.05574093759059906,
-0.004064109642058611,
0.02986733242869377,
0.06100252643227577,
0.03933796286582947,
0.06410261988639832,
0.13649000227451324,
-0.010296730324625969,
0.050693679600954056,
0.23594239354133606,
0.13071222603321075,
0.04225917160511017,
-0.0010421954793855548,
-0.008180477656424046,
-0.22604574263095856,
0.04921988397836685,
0.14086169004440308,
0.017938818782567978,
0.17841215431690216,
0.10180943459272385,
-0.06050674244761467,
0.0611439123749733,
0.07222970575094223,
-0.14235742390155792,
0.020076218992471695,
-0.010962078347802162,
-0.0586060993373394,
0.12412381917238235,
0.07544583082199097,
0.008397525176405907,
0.03478730469942093,
-0.04247848317027092,
-0.1846964806318283,
0.03496911749243736,
-0.05681052803993225,
-0.01974811963737011,
0.056515201926231384,
0.051043104380369186,
-0.12913623452186584,
-0.015120884403586388,
0.04678688943386078,
0.02018282189965248,
0.03482629731297493,
-0.09848776459693909,
-0.044789548963308334,
-0.03919424116611481,
0.07347863167524338,
0.029176313430070877,
0.13167409598827362,
-0.059128399938344955,
0.19447597861289978,
-0.14229126274585724,
0.07068690657615662,
0.25985047221183777,
-0.2594553828239441,
-0.0407562293112278,
-0.08511904627084732,
0.13985083997249603,
-0.09323132038116455,
-0.05968128889799118,
0.007794405333697796,
0.013779202476143837,
0.02377353236079216,
-0.03211504966020584,
-0.08829799294471741,
-0.06746776401996613,
0.06109187379479408,
-0.06471703946590424,
0.02655351720750332,
0.24268831312656403,
0.044511955231428146,
0.07085569202899933,
-0.032316941767930984,
0.01817159540951252,
-0.012681043706834316,
0.017603574320673943,
-0.032504402101039886,
-0.06451702862977982,
-0.036961376667022705,
-0.07330309599637985,
0.007933827117085457,
-0.08991090208292007,
-0.11452659219503403,
-0.09849558025598526,
-0.07693447917699814,
0.0104230847209692,
0.05217433348298073,
-0.12794229388237,
-0.023260872811079025,
-0.06788957864046097,
-0.03199559450149536,
0.013356111943721771,
-0.09108759462833405,
-0.059657517820596695,
0.018083354458212852,
-0.14779946208000183,
0.03976898267865181,
0.13355661928653717,
0.11177447438240051,
0.049504078924655914,
-0.004004481248557568,
-0.06076180189847946,
0.08967872709035873,
0.0753694400191307,
0.0005559158744290471,
-0.19964593648910522,
-0.020344216376543045,
-0.13889606297016144,
0.005006530787795782,
-0.15310300886631012,
0.018568865954875946,
-0.08470816910266876,
-0.013463391922414303,
0.1079145222902298,
0.04173275828361511,
-0.02845333144068718,
0.0013597877696156502,
-0.02645859122276306,
-0.021539993584156036,
-0.01555017288774252,
-0.035600777715444565,
-0.019385799765586853,
0.04084818810224533,
-0.036255475133657455,
0.03234425559639931,
-0.050076983869075775,
0.008357915095984936,
-0.10241467505693436,
0.0019428601954132318,
-0.11063457280397415,
0.07447056472301483,
-0.023181820288300514,
-0.08748003840446472,
0.01177095528692007,
-0.0007492153090424836,
0.006353008095175028,
-0.07224050909280777,
-0.031530383974313736,
-0.05849619209766388,
0.0025560096837580204,
-0.08615579456090927,
0.05903603881597519,
-0.07864933460950851,
0.04180479422211647,
-0.008128552697598934,
-0.07043450325727463,
-0.1181042492389679,
-0.026207514107227325,
0.09102106094360352,
-0.008616510778665543,
0.07785625755786896,
-0.1630111187696457,
0.06937719881534576,
-0.07558248937129974,
-0.00957483146339655,
-0.06897815316915512,
0.03891851752996445,
0.04872693493962288,
0.03180573508143425,
-0.09804237633943558,
-0.11032731086015701,
-0.05087202787399292,
-0.00788900163024664,
0.06447942554950714,
0.11565062403678894,
-0.1783406287431717,
-0.025373829528689384,
0.27656617760658264,
0.03364710882306099,
-0.1822516769170761,
0.12413131445646286,
-0.04845687001943588,
0.012555188499391079,
0.03022036701440811,
0.2878356873989105,
-0.019174588844180107,
0.06840294599533081,
0.019204268231987953,
-0.008835495449602604,
-0.05629655346274376,
0.028247151523828506,
0.13451209664344788,
-0.1014254242181778,
-0.048137638717889786,
0.021131223067641258,
0.003449423471465707,
-0.001065101707354188,
-0.1039239764213562,
-0.030190449208021164,
0.039410579949617386,
-0.05828113481402397,
0.08572739362716675,
0.022476905956864357,
0.031184546649456024,
-0.06385131925344467,
-0.008665437810122967,
0.08641307801008224,
0.03404756262898445,
-0.05106102675199509,
0.010870272293686867,
-0.1600487381219864,
0.03340320661664009,
-0.044269584119319916,
0.00047887852997519076,
-0.2272171825170517,
0.03125431388616562,
-0.049802131950855255,
0.14854909479618073,
0.02310672216117382,
-0.02038629725575447,
0.053734611719846725,
-0.009984306059777737,
0.04393813759088516,
0.017202239483594894,
0.1222982108592987,
-0.0239209346473217,
-0.043916814029216766,
-0.11659981310367584,
0.029144519940018654,
-0.03164444491267204,
0.0922786220908165,
0.03364422172307968,
-0.0370187908411026,
-0.10523775219917297,
-0.0007515622419305146,
-0.07819398492574692,
-0.0047743660397827625,
-0.022772617638111115,
-0.0008989700581878424,
-0.045309193432331085,
0.07721807807683945,
0.11022201925516129,
-0.04950003698468208,
-0.11089836061000824,
0.17741969227790833,
-0.12205639481544495,
0.024202995002269745,
0.1289754956960678,
-0.18414612114429474,
-0.01493711955845356,
-0.19763551652431488,
-0.07461098581552505,
0.02185620740056038,
-0.006227764766663313,
-0.0596749410033226,
0.07998881489038467,
0.0189872607588768,
0.10109353810548782,
-0.05830574408173561,
-0.03286540508270264,
-0.0065050930716097355,
0.019135015085339546,
-0.012840739451348782,
0.08562048524618149,
0.06767817586660385,
-0.19976262748241425,
0.08872869610786438,
0.013105766847729683,
0.011817348189651966,
0.06856859475374222,
0.04946647211909294,
-0.05885748565196991,
-0.07768990099430084,
0.02978760190308094,
-0.03807881847023964,
0.13153041899204254,
-0.21865849196910858,
0.008939048275351524,
0.08101589977741241,
-0.04187212511897087,
0.02267380803823471,
-0.10932212322950363,
-0.03447328507900238,
-0.007130755111575127,
-0.03800065070390701,
-0.10235300660133362,
0.06582725048065186,
0.043010760098695755,
0.06460730731487274,
0.01656183786690235,
-0.019919395446777344,
-0.03272823244333267,
0.0020465198904275894,
-0.1170319989323616,
0.17682239413261414,
-0.14330065250396729,
-0.2833958566188812,
-0.06931128352880478,
0.00027996161952614784,
0.0017118790419772267,
0.05084529146552086,
0.12891621887683868,
-0.07252392917871475,
0.05937846750020981,
-0.00669539999216795,
-0.16305476427078247,
-0.034177374094724655,
-0.001352845225483179,
-0.020236527547240257,
0.0018222751095890999,
-0.027687091380357742,
-0.06019747629761696,
-0.03645382821559906,
-0.11434005945920944,
0.06650552153587341,
0.07491199672222137,
-0.11470555514097214,
0.0839240700006485,
0.13447238504886627,
0.05568799376487732,
0.08081234246492386,
-0.08070788532495499,
0.15917520225048065,
-0.03129962459206581,
-0.08557175099849701,
0.03958224132657051,
-0.11151086539030075,
-0.0012139584869146347,
0.04408887028694153,
-0.02044561505317688,
-0.05041000619530678,
0.03378916531801224,
0.0269475057721138,
0.016138380393385887,
-0.2194301187992096,
-0.07985885441303253,
-0.06541190296411514,
0.09049603343009949,
0.06737075746059418,
0.013896836899220943,
0.05232568085193634,
0.06207982450723648,
0.011895046569406986,
-0.035617053508758545,
0.03811253234744072,
0.03350823372602463,
0.03589252382516861,
-0.07117459177970886,
0.14840306341648102,
0.07299495488405228,
-0.10110074281692505,
-0.020585406571626663,
-0.07711494714021683,
0.18041300773620605,
0.08725422620773315,
0.07111872732639313,
0.10626568645238876,
0.11745334416627884,
0.01059428509324789,
0.09867322444915771,
-0.022969326004385948,
0.005551367998123169,
-0.01533001009374857,
-0.060495078563690186,
-0.0759568139910698,
0.04650747403502464,
0.1838236153125763,
-0.011245505884289742,
-0.06664285808801651,
-0.026830215007066727,
0.10186386108398438,
0.09782345592975616,
0.1072501465678215,
-0.2358819991350174,
0.012253912165760994,
-0.025716107338666916,
-0.008828025311231613,
-0.09748676419258118,
0.0978129655122757,
-0.03284979239106178,
-0.07304318249225616,
-0.040543269366025925,
-0.01771383173763752,
0.13014031946659088,
0.051686834543943405,
0.061684928834438324,
-0.0778849869966507,
-0.18263250589370728,
-0.005393321625888348,
0.11354935169219971,
-0.27620765566825867,
0.2170896679162979,
0.02352479100227356,
-0.07014156877994537,
-0.06074579432606697,
-0.0646275207400322,
0.015315010212361813,
-0.005800568964332342,
0.23262764513492584,
0.009812998585402966,
0.06224591284990311,
0.010438736528158188,
-0.06633419543504715,
0.04443427547812462,
0.13437537848949432,
-0.05053235590457916,
0.047737181186676025,
-0.05775303766131401,
0.018820011988282204,
-0.028348810970783234,
0.14074982702732086,
-0.11582332104444504,
-0.0921962559223175,
0.029642503708600998,
0.008121712133288383,
0.2226037085056305,
0.015714270994067192,
0.012899457477033138,
0.030414247885346413,
0.10103464871644974,
-0.13909362256526947,
0.041818030178546906,
-0.09142283350229263,
0.11742839962244034,
0.12776409089565277,
-0.07099013775587082,
-0.05847384035587311,
-0.00038112280890345573,
0.05385361239314079,
-0.012065899558365345,
-0.04271062836050987,
0.08314637839794159,
-0.10349337756633759,
0.029771601781249046,
-0.08268460631370544,
0.1393864005804062,
0.04366312548518181,
0.11665280163288116,
0.09302294254302979,
-0.01895814947783947,
-0.10412034392356873,
-0.042001329362392426,
0.016315244138240814,
-0.01331180240958929,
0.07022544741630554,
0.09395594894886017,
-0.1547403633594513,
-0.026953425258398056,
-0.16040310263633728,
-0.015288247726857662,
0.33402884006500244,
0.11788676679134369,
-0.0352865569293499,
0.09037255495786667,
0.2437029331922531,
-0.04730121046304703,
-0.20347104966640472,
-0.03125883266329765,
0.014956970699131489,
-0.03114846721291542,
0.002739911200478673,
-0.05672212690114975,
0.013483243994414806,
0.03933236375451088,
0.010898718610405922,
-0.12269732356071472,
-0.18964549899101257,
-0.08742287009954453,
0.24820536375045776,
-0.012205115519464016,
0.29281097650527954,
-0.1892012655735016,
-0.07834643870592117,
-0.09041275829076767,
-0.19004595279693604,
0.16376064717769623,
0.003374979132786393,
0.07623366266489029,
-0.050564639270305634,
0.00416440749540925,
0.007429653778672218,
-0.03297984227538109,
0.08124556392431259,
0.10316891223192215,
0.008182530291378498,
0.009416449815034866,
-0.03341954946517944,
0.024703966453671455,
0.04273822903633118,
0.058321524411439896,
0.04992884397506714,
0.006312086246907711,
-0.2436245083808899,
0.015385828912258148,
-0.13227902352809906,
-0.06391867250204086,
0.033989619463682175,
-0.05018870159983635,
-0.03771331161260605,
0.018797283992171288,
0.029258226975798607,
0.07608956843614578,
-0.048307474702596664,
-0.13370990753173828,
0.10305121541023254,
-0.03994171321392059,
0.10502908378839493,
-0.08422228693962097,
-0.1848011016845703,
0.034756436944007874,
-0.007926537655293941,
0.11606481671333313,
0.014224734157323837,
0.055122729390859604,
0.16400647163391113,
0.00879497267305851,
0.17534418404102325,
0.1255694329738617,
-0.003717295592650771,
-0.020831165835261345,
0.08281669020652771,
-0.20325489342212677,
0.04349766671657562,
-0.02154988795518875,
0.010369202122092247,
-0.025743825361132622,
-0.10071595013141632,
0.09782174229621887,
0.012260721065104008,
-0.03649470582604408,
0.03455445542931557,
-0.01894708350300789,
-0.06251268088817596,
0.1361779421567917,
0.025446586310863495,
0.0531480573117733,
-0.10419750958681107,
-0.07522187381982803,
0.012846032157540321,
-0.19447943568229675,
0.040526892989873886,
0.06510892510414124,
-0.07991394400596619,
-0.1002151370048523,
-0.03952319175004959,
0.11571486294269562,
-0.11855098605155945,
-0.09381353110074997,
-0.09694720059633255,
-0.09146150946617126,
0.0978585034608841,
0.19480018317699432,
0.0729670450091362,
0.01133722998201847,
0.031504545360803604,
-0.04896964505314827,
-0.08649581670761108,
0.032839857041835785,
-0.03493988886475563,
-0.023824453353881836,
0.13353319466114044,
0.001778951263986528,
-0.04075273126363754,
0.1612101048231125,
-0.04117266833782196,
-0.06810691952705383,
-0.12793873250484467,
0.07626387476921082,
-0.16688977181911469,
-0.007190787699073553,
-0.10540244728326797,
-0.07140960544347763,
0.09120684117078781,
0.006130655761808157,
-0.023071005940437317,
-0.06184493750333786,
-0.05249482020735741,
0.0036810936871916056,
0.030370233580470085,
0.03213160112500191,
-0.15610504150390625,
-0.013464626856148243,
0.08791055530309677,
-0.006977782119065523,
0.08309102058410645,
0.11412477493286133,
-0.07438823580741882,
0.02107689343392849,
-0.05382876470685005,
-0.16873326897621155,
-0.03263446316123009,
0.031803395599126816,
0.02849959209561348,
-0.06622249633073807,
0.02103230357170105,
0.025004137307405472,
0.05835764482617378,
0.05418539419770241,
0.12282147258520126,
-0.037141162902116776,
-0.10300087183713913,
-0.07808119058609009,
-0.08778440207242966,
-0.051979050040245056,
-0.10742898285388947,
0.00645570270717144,
0.07414935529232025,
0.05146593973040581,
0.008015024475753307,
0.08625810593366623,
-0.0998893454670906,
0.03226879984140396,
-0.004103849641978741,
-0.021823156625032425,
0.0988282859325409,
-0.08967126160860062,
0.03577359765768051,
0.001720417058095336,
0.1773710399866104,
-0.03174716606736183,
0.028857622295618057,
0.021293219178915024,
0.006263116840273142,
-0.05613602697849274,
-0.025912165641784668,
0.1079074814915657,
0.09899979084730148,
-0.09964802116155624,
-0.07947269827127457,
0.0948350578546524,
0.02984721213579178,
0.09193291515111923,
0.13754890859127045,
0.03458353504538536,
0.1481439620256424,
0.06995784491300583,
0.08284006267786026,
0.029467031359672546,
0.08167657256126404,
-0.1158335953950882,
0.054873932152986526,
0.008078736253082752,
0.10757016390562057,
0.16466417908668518,
0.20747992396354675,
-0.08407412469387054,
0.04766039922833443,
-0.06826247274875641,
-0.09677376598119736,
-0.12536795437335968,
0.083671435713768,
-0.058413438498973846,
-0.11265021562576294,
0.013090337626636028,
-0.12384089827537537,
0.028320379555225372,
0.07452675700187683,
0.07663723081350327,
0.021929295733571053,
0.0892619788646698,
-0.037311676889657974,
-0.08515358716249466,
0.04967782273888588,
-0.04632914438843727,
0.06953691691160202,
-0.012393505312502384,
-0.008515416644513607,
0.0628301352262497,
-0.09367817640304565,
0.08260034769773483,
0.017731469124555588,
-0.021689623594284058,
-0.024114785715937614,
-0.12158695608377457,
-0.10155712068080902,
-0.023916572332382202,
-0.0037054361309856176,
0.013074005953967571,
0.1908482015132904,
0.03298136964440346,
0.042358458042144775,
0.03290510177612305,
0.11880557239055634,
-0.025771526619791985,
-0.005152759607881308,
-0.13092252612113953,
0.16538237035274506,
0.12344028800725937,
0.06201284006237984,
-0.023864764720201492,
0.019836384803056717,
-0.0058969431556761265,
0.3362446427345276,
0.20439141988754272,
-0.10471317172050476,
-0.029678190127015114,
0.06623104959726334,
0.032208967953920364,
0.08414234966039658,
0.03832991421222687,
0.12433735281229019,
0.2498953640460968,
-0.08882386237382889,
0.0020075880456715822,
-0.06451471149921417,
-0.04642048478126526,
-0.02228643372654915,
-0.029183728620409966,
0.1673765480518341,
0.0011846524430438876,
0.04848523437976837,
0.10762739926576614,
-0.18338912725448608,
0.09672637283802032,
0.03147709742188454,
-0.1088004782795906,
-0.07509255409240723,
-0.014160632155835629,
0.11361643671989441,
0.013131802901625633,
0.08789774775505066,
-0.01645040512084961,
-0.012265685014426708,
0.07097947597503662,
0.020880933851003647,
-0.22923529148101807,
-0.05158133804798126,
0.07239491492509842,
-0.04311617463827133,
0.050451211631298065,
-0.026922760531306267,
0.1850181668996811,
0.09491299837827682,
0.0307160597294569,
-0.003523731604218483,
0.017130421474575996,
-0.008392240852117538,
0.028944656252861023,
0.04850724712014198,
0.12438662350177765,
0.05108533799648285,
0.0012172668939456344,
0.04689870774745941,
-0.11037956178188324,
0.026620808988809586,
-0.11550834774971008,
0.04740495979785919,
-0.08850958943367004,
0.0983845442533493,
-0.04450491815805435,
0.05513279512524605,
0.06746188551187515,
-0.015629537403583527,
0.0374055951833725,
-0.029943404719233513,
-0.0016788783250376582,
0.006860390771180391,
-0.0060218567959964275,
-0.10740436613559723,
-0.14964857697486877,
-0.008300288580358028,
-0.05691844969987869,
-0.04425917565822601,
-0.3809347450733185,
-0.04717620462179184,
-0.03368808329105377,
-0.06287283450365067,
0.01967930980026722,
0.0579751692712307,
0.04942696541547775,
0.038066066801548004,
-0.05032945051789284,
-0.0601576492190361,
0.04682055860757828,
0.03266162425279617,
-0.17490366101264954,
-0.10622429102659225
] |
null | null |
transformers
|
This is finetune version of [SimCSE: Simple Contrastive Learning of Sentence Embeddings](https://arxiv.org/abs/2104.08821)
- Train supervised on 100K triplet samples samples related to stroke domain from : stroke books, quora medical, quora's stroke, quora's general and human annotates.
- Positive sentences are generated by paraphrasing and back-translate.
- Negative sentences are randomly selected in general domain.
### Extract sentence representation
```
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("demdecuong/stroke_sup_simcse")
model = AutoModel.from_pretrained("demdecuong/stroke_sup_simcse")
text = "What are disease related to red stroke's causes?"
inputs = tokenizer(text, return_tensors='pt')
outputs = model(**inputs)[1]
```
### Build up embedding for database
```
database = [
'What is the daily checklist for stroke returning home',
'What are some tips for stroke adapt new life',
'What should I consider when using nursing-home care'
]
embedding = torch.zeros((len(database),768))
for i in range(len(database)):
inputs = tokenizer(database[i], return_tensors="pt")
outputs = model(**inputs)[1]
embedding[i] = outputs
print(embedding.shape)
```
### Result
On our company's PoC project, the testset contains positive/negative pairs of matching question related to stroke from human-generation.
- SimCSE supervised + 100k : Train on 100K triplet samples contains : medical, stroke and general domain
- SimCSE supervised + 42k : Train on 42K triplet samples contains : medical, stroke domain
| Model | Top-1 Accuracy |
| ------------- | ------------- |
| SimCSE supervised (author) | 75.83 |
| SimCSE unsupervised (ours) | 76.66 |
| SimCSE supervised + 100k (ours) | 73.33 |
| SimCSE supervised + 42k (ours) | 75.83 |
|
{}
|
feature-extraction
|
demdecuong/stroke_sup_simcse
|
[
"transformers",
"pytorch",
"bert",
"feature-extraction",
"arxiv:2104.08821",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2104.08821"
] |
[] |
TAGS
#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #region-us
|
This is finetune version of SimCSE: Simple Contrastive Learning of Sentence Embeddings
* Train supervised on 100K triplet samples samples related to stroke domain from : stroke books, quora medical, quora's stroke, quora's general and human annotates.
* Positive sentences are generated by paraphrasing and back-translate.
* Negative sentences are randomly selected in general domain.
### Extract sentence representation
### Build up embedding for database
### Result
On our company's PoC project, the testset contains positive/negative pairs of matching question related to stroke from human-generation.
* SimCSE supervised + 100k : Train on 100K triplet samples contains : medical, stroke and general domain
* SimCSE supervised + 42k : Train on 42K triplet samples contains : medical, stroke domain
|
[
"### Extract sentence representation",
"### Build up embedding for database",
"### Result\n\n\nOn our company's PoC project, the testset contains positive/negative pairs of matching question related to stroke from human-generation.\n\n\n* SimCSE supervised + 100k : Train on 100K triplet samples contains : medical, stroke and general domain\n* SimCSE supervised + 42k : Train on 42K triplet samples contains : medical, stroke domain"
] |
[
"TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #region-us \n",
"### Extract sentence representation",
"### Build up embedding for database",
"### Result\n\n\nOn our company's PoC project, the testset contains positive/negative pairs of matching question related to stroke from human-generation.\n\n\n* SimCSE supervised + 100k : Train on 100K triplet samples contains : medical, stroke and general domain\n* SimCSE supervised + 42k : Train on 42K triplet samples contains : medical, stroke domain"
] |
[
38,
7,
10,
90
] |
[
"passage: TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #region-us \n### Extract sentence representation### Build up embedding for database### Result\n\n\nOn our company's PoC project, the testset contains positive/negative pairs of matching question related to stroke from human-generation.\n\n\n* SimCSE supervised + 100k : Train on 100K triplet samples contains : medical, stroke and general domain\n* SimCSE supervised + 42k : Train on 42K triplet samples contains : medical, stroke domain"
] |
[
-0.07777131348848343,
-0.0690411701798439,
-0.0007883163052611053,
0.016152512282133102,
0.030421143397688866,
0.081929512321949,
0.15589703619480133,
0.1257803738117218,
-0.033573102205991745,
0.02594810537993908,
0.21436913311481476,
0.0962965339422226,
0.04045126214623451,
0.011939788237214088,
-0.015664324164390564,
-0.20579101145267487,
0.049620985984802246,
0.0865812599658966,
-0.055935099720954895,
0.1669919341802597,
0.10616222769021988,
-0.04225386306643486,
0.07168689370155334,
0.05698152631521225,
-0.12946274876594543,
0.028275100514292717,
0.03527873381972313,
-0.05795549601316452,
0.12634453177452087,
0.05620188266038895,
0.007100390736013651,
0.0729060247540474,
-0.0004816583823412657,
-0.13897429406642914,
0.027211342006921768,
-0.062326185405254364,
-0.011023140512406826,
0.049535661935806274,
0.04622078686952591,
-0.08095531165599823,
0.0411856584250927,
0.0273198951035738,
0.09014423191547394,
0.028134005144238472,
-0.0899624153971672,
0.01673683524131775,
-0.15355059504508972,
0.11382243037223816,
0.07184145599603653,
0.08311651647090912,
-0.014995388686656952,
0.2396155148744583,
-0.19389033317565918,
0.06352301687002182,
0.1935451328754425,
-0.2704959809780121,
-0.04306415840983391,
0.02035304717719555,
0.15043926239013672,
-0.027056368067860603,
-0.0077988216653466225,
0.05273551866412163,
0.03348274528980255,
0.03134443238377571,
-0.018193909898400307,
-0.09476839751005173,
-0.013584057800471783,
0.04218398034572601,
-0.09884821623563766,
0.04446744546294212,
0.27326494455337524,
0.04896882176399231,
0.018367838114500046,
-0.06350789964199066,
0.031113576143980026,
-0.04778977483510971,
0.07516639679670334,
-0.10824072360992432,
-0.05027270317077637,
-0.09960372000932693,
-0.10621337592601776,
0.04463303089141846,
-0.06055772304534912,
-0.13633210957050323,
-0.06976200640201569,
-0.09304004907608032,
-0.011201678775250912,
0.06519868224859238,
-0.03828020021319389,
-0.0022339713759720325,
-0.06011355668306351,
0.01024296972900629,
0.027831807732582092,
-0.10208839923143387,
-0.1055714562535286,
0.000712812936399132,
-0.12117116153240204,
-0.06937844306230545,
0.08089986443519592,
0.1366734653711319,
0.053977593779563904,
0.018817877396941185,
-0.009522509761154652,
0.0827697291970253,
0.09205762296915054,
-0.03797036409378052,
-0.2572163939476013,
-0.05096938833594322,
-0.13241231441497803,
0.051072102040052414,
-0.04225825145840645,
0.0026290055830031633,
-0.054398149251937866,
0.05274568870663643,
0.07041572034358978,
0.03832416608929634,
-0.0924995020031929,
-0.02553861401975155,
-0.030606087297201157,
-0.03739965334534645,
0.11274199932813644,
-0.013432656414806843,
-0.0035778891760855913,
-0.008308690041303635,
-0.006136346608400345,
0.12289366126060486,
-0.06490860879421234,
0.006689661182463169,
-0.07833930850028992,
0.12182623893022537,
-0.0960354208946228,
0.0580657459795475,
-0.019216910004615784,
-0.08842490613460541,
0.005004425998777151,
0.054208870977163315,
-0.024136686697602272,
-0.05829685181379318,
-0.024695755913853645,
-0.04356285184621811,
0.01021024864166975,
-0.11002199351787567,
0.11105659604072571,
-0.015637200325727463,
0.03149372339248657,
-0.02461426705121994,
-0.04360842704772949,
-0.08702065795660019,
-0.0313703753054142,
0.04172974079847336,
-0.013904380612075329,
0.055819373577833176,
-0.20280179381370544,
0.01494090911000967,
-0.04394419491291046,
0.03580040857195854,
-0.12102905660867691,
0.02925649657845497,
-0.010305522009730339,
-0.01518663577735424,
-0.015116983093321323,
-0.11698955297470093,
0.009235680103302002,
-0.011286641471087933,
0.09540162235498428,
0.08040250837802887,
-0.18372949957847595,
0.012148255482316017,
0.16417409479618073,
-0.029189366847276688,
-0.19285011291503906,
0.15407536923885345,
-0.0450410470366478,
-0.034262318164110184,
0.018452690914273262,
0.2383328080177307,
-0.01004471629858017,
0.048713184893131256,
0.024737916886806488,
-0.061450809240341187,
-0.052772652357816696,
-0.0343935452401638,
0.16169235110282898,
-0.07481970638036728,
-0.064228355884552,
0.0060019404627382755,
0.05135776102542877,
-0.041165176779031754,
-0.11354097723960876,
-0.008438151329755783,
0.024209553375840187,
-0.0761762261390686,
0.07118616998195648,
-0.008089487440884113,
0.03275521099567413,
-0.1347501277923584,
-0.020238149911165237,
0.05648410692811012,
0.0759170800447464,
-0.012136129662394524,
-0.009490561671555042,
-0.1236184686422348,
0.03969400376081467,
-0.032916948199272156,
-0.04652198404073715,
-0.16418279707431793,
0.018604671582579613,
-0.0072028920985758305,
0.1455111801624298,
0.0038070979062467813,
0.04022957384586334,
0.024422865360975266,
0.07122757285833359,
-0.022183284163475037,
0.008065953850746155,
0.11112327128648758,
0.0030031984206289053,
-0.03892063349485397,
-0.12455128133296967,
-0.000042032628698507324,
-0.012984711676836014,
0.08274110406637192,
0.056114811450242996,
-0.004428189247846603,
-0.12988236546516418,
0.011142106726765633,
-0.06895308196544647,
-0.01827787235379219,
-0.020413316786289215,
0.038106776773929596,
-0.07381433248519897,
0.054584093391895294,
0.06472235918045044,
-0.026790738105773926,
-0.11305567622184753,
0.09242302179336548,
-0.11992577463388443,
0.004029925912618637,
0.07142803817987442,
-0.10073342174291611,
-0.07634645700454712,
-0.1263085901737213,
-0.05259092152118683,
0.021309060975909233,
-0.05569751188158989,
-0.07120174914598465,
-0.012151793576776981,
0.02445872500538826,
0.09077069163322449,
-0.051506705582141876,
-0.03298000246286392,
-0.01782284677028656,
0.004036037251353264,
-0.011236025020480156,
0.05313227325677872,
0.12362933158874512,
-0.1797010749578476,
0.04473613202571869,
0.03278198838233948,
0.003266132902354002,
0.035539548844099045,
0.06685391068458557,
-0.04938879236578941,
-0.11469492316246033,
-0.038110315799713135,
-0.01875237561762333,
0.19986151158809662,
-0.18839459121227264,
0.03769826889038086,
0.05910743772983551,
0.011181412264704704,
-0.016549676656723022,
-0.09255687892436981,
-0.03287820518016815,
-0.017398733645677567,
-0.01626341976225376,
-0.09112615883350372,
0.055941179394721985,
-0.006247473414987326,
0.08332575857639313,
-0.02674073725938797,
-0.010206056758761406,
0.0011268391972407699,
-0.007817549630999565,
-0.11119520664215088,
0.1744385063648224,
-0.11378300189971924,
-0.25254231691360474,
-0.08214956521987915,
0.04447658732533455,
-0.05370645225048065,
0.01646585948765278,
0.10296972841024399,
-0.04293234273791313,
0.03896414116024971,
-0.08468735963106155,
-0.1492164433002472,
-0.0663953423500061,
-0.02360406331717968,
-0.03000551089644432,
0.006712460424751043,
-0.017670372501015663,
-0.07328737527132034,
-0.039984602481126785,
-0.1294347196817398,
0.05273054912686348,
0.08993314206600189,
-0.054376162588596344,
0.062374211847782135,
0.16864128410816193,
0.03974028676748276,
0.0498238205909729,
-0.08055698871612549,
0.08385104686021805,
0.004120319616049528,
-0.031387362629175186,
0.0868760272860527,
-0.0814560204744339,
-0.010827847756445408,
0.04175393655896187,
0.035477105528116226,
-0.03699640557169914,
0.041070692241191864,
0.0205267034471035,
-0.008677856996655464,
-0.2811182141304016,
-0.058065999299287796,
-0.04177252948284149,
0.10953313857316971,
-0.053665537387132645,
-0.006554646883159876,
0.05253176763653755,
0.1235075294971466,
0.02558007277548313,
-0.037547048181295395,
-0.02670077420771122,
0.03597191721200943,
0.05856888368725777,
-0.06777013093233109,
0.18178965151309967,
0.04850013554096222,
-0.0713990107178688,
0.0363810732960701,
-0.10182353109121323,
0.14411890506744385,
0.07737130671739578,
0.05162622407078743,
0.12426932901144028,
0.11773867160081863,
-0.02565101720392704,
0.11097514629364014,
0.011008843779563904,
-0.02329757995903492,
-0.04351425915956497,
-0.03842656686902046,
-0.08259366452693939,
0.013155684806406498,
0.14639915525913239,
0.05330917611718178,
-0.07145083695650101,
0.023798370733857155,
0.09131810814142227,
0.07884301245212555,
0.11807811260223389,
-0.2229759842157364,
-0.009638666175305843,
-0.01447842363268137,
0.00856192596256733,
-0.04346027970314026,
0.0710063949227333,
0.03916935995221138,
-0.11802491545677185,
0.012148684822022915,
-0.010605166666209698,
0.09699483215808868,
-0.00789123959839344,
0.03680359572172165,
-0.00759280938655138,
-0.151939257979393,
-0.03554614260792732,
0.12136315554380417,
-0.2681652903556824,
0.16403602063655853,
-0.011484019458293915,
-0.07830680161714554,
-0.0941210612654686,
-0.045102301985025406,
0.0011478203814476728,
0.07647300511598587,
0.15176980197429657,
0.028167065232992172,
0.005150977987796068,
-0.0005473539349623024,
-0.06401166319847107,
-0.0031392534729093313,
0.11081520467996597,
-0.00081540149403736,
0.0214424766600132,
-0.042746227234601974,
0.007356334012001753,
-0.018551530316472054,
0.0790398046374321,
-0.18302021920681,
-0.0845564678311348,
0.06475820392370224,
-0.026528386399149895,
0.10035338252782822,
-0.036173414438962936,
-0.019016209989786148,
0.06424950063228607,
0.1679607629776001,
-0.048722341656684875,
0.031248364597558975,
-0.11957812309265137,
0.0940108671784401,
0.13533534109592438,
-0.03867453709244728,
-0.06525222957134247,
0.0019241654081270099,
0.019979307428002357,
0.03592723235487938,
-0.014650466851890087,
0.10544972866773605,
-0.14797981083393097,
-0.05006241798400879,
-0.16536583006381989,
0.19033274054527283,
-0.002654904732480645,
0.08498408645391464,
0.04328637197613716,
-0.04065677896142006,
-0.004504892975091934,
-0.10168889909982681,
0.0781739130616188,
-0.041854377835989,
0.008682015351951122,
0.1416611671447754,
-0.1624700129032135,
0.02170148864388466,
-0.10551893711090088,
-0.08402113616466522,
0.26848796010017395,
0.21062855422496796,
-0.022985609248280525,
0.07842165231704712,
0.22317185997962952,
-0.030761005356907845,
-0.16148579120635986,
-0.019412940368056297,
0.0015589471440762281,
-0.01724841073155403,
0.08222894370555878,
-0.005501158069819212,
0.06235116720199585,
0.11100558936595917,
0.007749450393021107,
-0.10190291702747345,
-0.28198572993278503,
-0.09739074856042862,
0.14245635271072388,
0.0302845798432827,
0.2779095470905304,
-0.19745087623596191,
-0.059839215129613876,
-0.11491256207227707,
-0.10844791680574417,
0.1459069848060608,
-0.0393822118639946,
0.08654545247554779,
-0.06425828486680984,
-0.03428748995065689,
-0.007567012216895819,
-0.029988383874297142,
0.07696869969367981,
0.07989677041769028,
0.04585966095328331,
-0.021363019943237305,
-0.026199953630566597,
0.18481077253818512,
0.022155852988362312,
0.0517982579767704,
0.037236955016851425,
0.019710205495357513,
-0.17139029502868652,
-0.0025532131548970938,
-0.08123081177473068,
-0.08833663910627365,
0.0027583714108914137,
-0.07320532202720642,
-0.09479141235351562,
0.03694499656558037,
-0.008367939852178097,
0.008915369398891926,
-0.11919422447681427,
-0.07463734596967697,
0.0424545556306839,
-0.02064276486635208,
0.1234535425901413,
-0.1231309249997139,
-0.17277559638023376,
0.0365770123898983,
-0.024287467822432518,
0.03923961520195007,
0.10828828066587448,
0.026483803987503052,
0.1198788434267044,
0.02356541156768799,
0.15265563130378723,
0.12384537607431412,
-0.02434736117720604,
-0.00587495556101203,
0.07826533913612366,
-0.18316932022571564,
0.04145153984427452,
0.011276005767285824,
0.03250029310584068,
-0.0521235391497612,
-0.03824803605675697,
0.12531302869319916,
0.003938645124435425,
-0.014488066546618938,
-0.007873883470892906,
-0.0025295375380665064,
-0.07646825909614563,
0.1917080134153366,
0.010640834458172321,
0.04669828712940216,
-0.06763172894716263,
-0.0693061500787735,
0.0359966978430748,
-0.08295150101184845,
-0.0028178340289741755,
0.08663877844810486,
-0.08965975046157837,
-0.09736143797636032,
-0.0569671206176281,
0.14965355396270752,
-0.12380311638116837,
-0.075498066842556,
-0.10300292074680328,
-0.12512339651584625,
0.05064300075173378,
0.1316070407629013,
0.056876759976148605,
-0.02079416997730732,
-0.011598027311265469,
-0.06456620246171951,
-0.0956362634897232,
0.06527238339185715,
-0.034032609313726425,
0.056403789669275284,
0.08287756145000458,
0.031771134585142136,
-0.00931099709123373,
0.10111871361732483,
-0.046247124671936035,
-0.005662571173161268,
-0.11436589807271957,
0.026416026055812836,
-0.19548918306827545,
-0.04477522894740105,
-0.07012639939785004,
-0.06683021038770676,
0.09475643932819366,
0.023440077900886536,
-0.027844101190567017,
-0.06518841534852982,
-0.013551708310842514,
-0.019680205732584,
0.054804567247629166,
0.03918621316552162,
-0.13431724905967712,
-0.011248434893786907,
0.06321649998426437,
-0.02444932982325554,
0.07854682952165604,
0.022067975252866745,
-0.05062446743249893,
-0.010968049056828022,
-0.06896407902240753,
-0.12016420811414719,
-0.04947711527347565,
0.06849507242441177,
0.03057270310819149,
-0.11779678612947464,
0.019285013899207115,
0.06474504619836807,
0.04451964795589447,
0.06460517644882202,
0.05975181609392166,
-0.03285428509116173,
-0.13387514650821686,
-0.09753275662660599,
-0.15077492594718933,
-0.021913442760705948,
-0.07092631608247757,
0.018125813454389572,
0.03443388268351555,
0.10065089911222458,
-0.011096585541963577,
0.062656931579113,
-0.13065794110298157,
-0.002018685219809413,
0.025848394259810448,
-0.007760601118206978,
-0.02442602440714836,
-0.06857745349407196,
0.04435162618756294,
0.017224226146936417,
0.25828826427459717,
-0.021978897973895073,
-0.046445198357105255,
0.023253237828612328,
-0.016788531094789505,
0.07010232657194138,
-0.02675057016313076,
0.110575370490551,
0.0489400215446949,
-0.08792036771774292,
-0.1332559585571289,
0.03690475970506668,
0.04548979923129082,
-0.03256838023662567,
0.23227037489414215,
-0.019715821370482445,
0.05023627728223801,
0.021584676578640938,
0.039609793573617935,
0.002652023918926716,
0.11595208197832108,
-0.15742042660713196,
0.04384515807032585,
0.0069634271785616875,
0.09109396487474442,
0.16295230388641357,
0.13058044016361237,
-0.087773397564888,
0.04857576638460159,
-0.07663729786872864,
-0.10878713428974152,
-0.09407196193933487,
0.10312656313180923,
-0.05517454817891121,
-0.16777625679969788,
-0.004455519374459982,
-0.13727815449237823,
0.022650819271802902,
0.2035827487707138,
0.1079297661781311,
0.004204846918582916,
0.054886430501937866,
-0.11069724708795547,
0.027394723147153854,
0.018468989059329033,
-0.007695311680436134,
0.04267455264925957,
0.003889975603669882,
0.00939712580293417,
0.029672710224986076,
-0.07363653182983398,
0.06408768147230148,
-0.010400518774986267,
-0.0411849170923233,
-0.0491812489926815,
-0.05162690207362175,
-0.07843545079231262,
-0.011641204357147217,
-0.07060897350311279,
0.04800122603774071,
0.1852695196866989,
0.061706509441137314,
0.050127703696489334,
0.020660674199461937,
0.048490010201931,
-0.03037288971245289,
-0.03917514160275459,
-0.12233313918113708,
0.15711413323879242,
0.10104074329137802,
0.06716970354318619,
-0.024707963690161705,
-0.03272544592618942,
0.003913074731826782,
0.26136064529418945,
0.25553223490715027,
-0.1231888085603714,
-0.07973547279834747,
0.05202396586537361,
0.012737257406115532,
0.05066296458244324,
0.10029541701078415,
0.07033048570156097,
0.1630920022726059,
-0.0465766042470932,
0.034357912838459015,
-0.06024780496954918,
-0.06826169788837433,
-0.036465298384428024,
0.012979748658835888,
0.16828887164592743,
0.020292682573199272,
0.06043354421854019,
0.11422459781169891,
-0.08902406692504883,
0.06763957440853119,
0.08731990307569504,
-0.0758403018116951,
-0.07146904617547989,
-0.0013881959021091461,
0.07524774968624115,
-0.009625629521906376,
0.07865510135889053,
-0.026120636612176895,
0.007035301066935062,
0.11318280547857285,
0.007906879298388958,
-0.2218993604183197,
-0.06388885527849197,
0.096119225025177,
-0.027984781190752983,
0.09108580648899078,
-0.005175372119992971,
0.2191426008939743,
0.1131584495306015,
0.007644259370863438,
0.013000065460801125,
0.07649444788694382,
-0.012737061828374863,
0.028110938146710396,
0.08512941002845764,
0.11547979712486267,
0.061987053602933884,
0.042609043419361115,
0.01779096946120262,
-0.09512311965227127,
0.03644198924303055,
-0.11803054809570312,
0.052085958421230316,
-0.10166356712579727,
0.12540099024772644,
-0.02792944945394993,
0.09048736095428467,
0.11584791541099548,
-0.04422741010785103,
0.06644712388515472,
-0.03703352436423302,
0.008746941573917866,
0.031833142042160034,
-0.009777024388313293,
-0.030800513923168182,
-0.21313579380512238,
0.047023434191942215,
-0.06826387345790863,
0.0061749606393277645,
-0.4115249514579773,
-0.06422459334135056,
-0.08400692790746689,
-0.04241359978914261,
-0.01090269722044468,
0.0389862060546875,
0.05189896747469902,
0.04872385784983635,
-0.06959362328052521,
0.045081932097673416,
-0.01793568767607212,
0.04813944175839424,
-0.1365898847579956,
-0.08447590470314026
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# iloko_model
This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0095
- Wer: 0.0840
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 30
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 3.2784 | 1.11 | 100 | 2.9875 | 1.0 |
| 2.6899 | 2.22 | 200 | 2.6741 | 1.0 |
| 2.6177 | 3.33 | 300 | 2.6516 | 1.0 |
| 2.5327 | 4.44 | 400 | 2.4530 | 1.0 |
| 0.8653 | 5.56 | 500 | 0.5227 | 0.6547 |
| 0.3414 | 6.67 | 600 | 0.1830 | 0.2487 |
| 0.2299 | 7.78 | 700 | 0.1212 | 0.1877 |
| 0.1739 | 8.89 | 800 | 0.0843 | 0.1441 |
| 0.1242 | 10.0 | 900 | 0.0766 | 0.1441 |
| 0.1116 | 11.11 | 1000 | 0.0530 | 0.1145 |
| 0.0861 | 12.22 | 1100 | 0.0442 | 0.1047 |
| 0.1007 | 13.33 | 1200 | 0.0379 | 0.1023 |
| 0.0613 | 14.44 | 1300 | 0.0291 | 0.1006 |
| 0.0629 | 15.56 | 1400 | 0.0264 | 0.0961 |
| 0.047 | 16.67 | 1500 | 0.0238 | 0.0935 |
| 0.0797 | 17.78 | 1600 | 0.0226 | 0.0913 |
| 0.034 | 18.89 | 1700 | 0.0197 | 0.0893 |
| 0.0485 | 20.0 | 1800 | 0.0173 | 0.0905 |
| 0.0402 | 21.11 | 1900 | 0.0148 | 0.0902 |
| 0.0231 | 22.22 | 2000 | 0.0135 | 0.0891 |
| 0.0512 | 23.33 | 2100 | 0.0134 | 0.0861 |
| 0.0181 | 24.44 | 2200 | 0.0118 | 0.0842 |
| 0.0371 | 25.56 | 2300 | 0.0116 | 0.0867 |
| 0.0342 | 26.67 | 2400 | 0.0104 | 0.0863 |
| 0.0344 | 27.78 | 2500 | 0.0100 | 0.0850 |
| 0.0182 | 28.89 | 2600 | 0.0096 | 0.0839 |
| 0.0171 | 30.0 | 2700 | 0.0095 | 0.0840 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu102
- Datasets 1.13.3
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "pipeline_tag": "automatic-speech-recognition"}
|
automatic-speech-recognition
|
denden/iloko_model
|
[
"transformers",
"pytorch",
"tensorboard",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
|
iloko\_model
============
This model is a fine-tuned version of facebook/wav2vec2-large-xlsr-53 on an unknown dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0095
* Wer: 0.0840
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0003
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* num\_epochs: 30
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.10.0+cu102
* Datasets 1.13.3
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu102\n* Datasets 1.13.3\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu102\n* Datasets 1.13.3\n* Tokenizers 0.10.3"
] |
[
56,
143,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu102\n* Datasets 1.13.3\n* Tokenizers 0.10.3"
] |
[
-0.1332528293132782,
0.10443782061338425,
-0.0019935211166739464,
0.09443783015012741,
0.14296869933605194,
0.018518490716814995,
0.12658271193504333,
0.12341658771038055,
-0.07698189467191696,
0.08201108872890472,
0.11455532163381577,
0.12721846997737885,
0.04257548972964287,
0.1306828409433365,
-0.046563826501369476,
-0.2579909861087799,
0.007292898837476969,
0.003313153050839901,
-0.09502141922712326,
0.13466624915599823,
0.06895965337753296,
-0.1244635060429573,
0.060040876269340515,
0.004352133721113205,
-0.16156138479709625,
-0.027147263288497925,
-0.014041034504771233,
-0.06265535950660706,
0.14229868352413177,
0.01709497533738613,
0.0982729122042656,
0.03880080580711365,
0.10937681049108505,
-0.21294865012168884,
0.004192087799310684,
0.049955639988183975,
0.02381843701004982,
0.07301374524831772,
0.06179524213075638,
-0.008323799818754196,
0.09252766519784927,
-0.099887914955616,
0.05882890522480011,
0.019598275423049927,
-0.1137276366353035,
-0.22968925535678864,
-0.08185897022485733,
0.02189798466861248,
0.07760418951511383,
0.10270296782255173,
-0.013176048174500465,
0.11171166598796844,
-0.06972900032997131,
0.0897110104560852,
0.264644980430603,
-0.3001701235771179,
-0.0627116709947586,
-0.007292950060218573,
0.032970380038022995,
0.06982457637786865,
-0.10224095731973648,
-0.009746965020895004,
0.02694093808531761,
0.03536912053823471,
0.12545089423656464,
-0.005420739762485027,
-0.05551505833864212,
0.012885142117738724,
-0.15386128425598145,
-0.04805523529648781,
0.1052234023809433,
0.045116543769836426,
-0.030835501849651337,
-0.0765465497970581,
-0.06561234593391418,
-0.21806970238685608,
-0.041296157985925674,
0.005335503723472357,
0.03891126811504364,
-0.0493648499250412,
-0.12150383740663528,
0.014292185194790363,
-0.0751882940530777,
-0.0880884975194931,
-0.022379925474524498,
0.14856001734733582,
0.03237376734614372,
0.006609921809285879,
-0.0016531981527805328,
0.09582782536745071,
0.041233111172914505,
-0.15285436809062958,
0.0077065229415893555,
0.04444488137960434,
-0.07953283190727234,
-0.013656712137162685,
-0.06170869618654251,
0.002383641665801406,
0.014142083935439587,
0.1436551809310913,
-0.012248489074409008,
0.056733135133981705,
0.03373130410909653,
0.023622576147317886,
-0.10284069925546646,
0.2083825320005417,
-0.08648516982793808,
-0.06036911532282829,
-0.02420649863779545,
0.10256669670343399,
0.026903176680207253,
-0.03209595009684563,
-0.09248431026935577,
0.012770837172865868,
0.11523731052875519,
0.03442966565489769,
-0.03759419173002243,
0.049708303064107895,
-0.045820072293281555,
-0.016865937039256096,
0.01377043779939413,
-0.10670552402734756,
0.02859022468328476,
0.027520274743437767,
-0.08234121650457382,
-0.006824846845120192,
-0.0011060484685003757,
0.015608594752848148,
-0.016788041219115257,
0.11561207473278046,
-0.07477601617574692,
0.01737356372177601,
-0.05558173358440399,
-0.11156938970088959,
0.016457049176096916,
-0.10401806235313416,
-0.001966387266293168,
-0.07713615894317627,
-0.11719824373722076,
-0.025431383401155472,
0.0402129665017128,
-0.055027689784765244,
-0.050773244351148605,
-0.07588779926300049,
-0.09262988716363907,
0.04199865087866783,
-0.02553205005824566,
0.12972187995910645,
-0.06846117228269577,
0.11043150722980499,
0.04533334821462631,
0.09182162582874298,
-0.002381477737799287,
0.05435749888420105,
-0.0644976869225502,
0.024695634841918945,
-0.13898520171642303,
0.08171254396438599,
-0.0687304437160492,
0.04175502061843872,
-0.12009947746992111,
-0.10518304258584976,
0.020495528355240822,
-0.013786647468805313,
0.09547190368175507,
0.09395642578601837,
-0.21646104753017426,
-0.08782345801591873,
0.17360326647758484,
-0.06297322362661362,
-0.1070345863699913,
0.13200335204601288,
-0.02689024619758129,
-0.03322821110486984,
0.0495319701731205,
0.16947604715824127,
0.048810526728630066,
-0.10680672526359558,
-0.009692936204373837,
-0.027734721079468727,
0.08673392236232758,
-0.04004829376935959,
0.09887069463729858,
-0.016711877658963203,
0.05526307225227356,
0.002922864630818367,
-0.05582922324538231,
0.05588023364543915,
-0.10288171470165253,
-0.09833554178476334,
-0.02995401993393898,
-0.10533572733402252,
0.0338410846889019,
0.050510164350271225,
0.049728937447071075,
-0.08370374143123627,
-0.10727161169052124,
0.008769294247031212,
0.09277118742465973,
-0.10215558856725693,
0.03306203708052635,
-0.07873827964067459,
0.07170403003692627,
-0.054198987782001495,
-0.01847240701317787,
-0.19640599191188812,
-0.011766422539949417,
0.023116329684853554,
-0.0153242452070117,
0.026019133627414703,
-0.016886690631508827,
0.07670444250106812,
0.06512609124183655,
-0.03455762192606926,
-0.039567578583955765,
-0.027300825342535973,
-0.007959951646625996,
-0.09682214260101318,
-0.21314892172813416,
-0.07326612621545792,
-0.02237704209983349,
0.16500946879386902,
-0.18715158104896545,
0.0050653680227696896,
0.007002643775194883,
0.08673600852489471,
0.010441170074045658,
-0.034399475902318954,
-0.02758421190083027,
0.08488782495260239,
-0.02421836368739605,
-0.06977076828479767,
0.05606778338551521,
0.001502664410509169,
-0.09928690642118454,
-0.010593661107122898,
-0.12783415615558624,
0.12203489989042282,
0.1396392583847046,
-0.02646845392882824,
-0.06616782397031784,
0.006034291815012693,
-0.05444895848631859,
-0.043527763336896896,
-0.029631735756993294,
0.018742434680461884,
0.18337810039520264,
0.024988923221826553,
0.12792161107063293,
-0.07215218245983124,
-0.04104897007346153,
0.03634396940469742,
0.009994719177484512,
0.004248656332492828,
0.12102190405130386,
0.09786074608564377,
-0.06688735634088516,
0.11324222385883331,
0.10027717053890228,
-0.09027568995952606,
0.09843035042285919,
-0.050542790442705154,
-0.09416072070598602,
-0.02742817811667919,
-0.010717568919062614,
0.029960013926029205,
0.11102540045976639,
-0.12521062791347504,
-0.029600225389003754,
0.025297490879893303,
0.004901566542685032,
0.008456053212285042,
-0.2188914716243744,
-0.019385717809200287,
0.04022201523184776,
-0.07174968719482422,
-0.038915202021598816,
-0.023648500442504883,
-0.0027940613217651844,
0.09942352026700974,
0.0027897232212126255,
-0.10761835426092148,
-0.0015898041892796755,
-0.01584947295486927,
-0.06898903846740723,
0.18747876584529877,
-0.068100705742836,
-0.1381545513868332,
-0.11303344368934631,
-0.07092338800430298,
-0.04372444376349449,
0.007402332033962011,
0.048221904784440994,
-0.08442813158035278,
-0.028707507997751236,
-0.0717800036072731,
0.009024540893733501,
-0.0032362958882004023,
0.05697150528430939,
0.04436077922582626,
-0.01254973653703928,
0.06580805778503418,
-0.10361859947443008,
0.009080232121050358,
-0.051285985857248306,
-0.03612548112869263,
0.03326844796538353,
0.05434248223900795,
0.1301448792219162,
0.1664956510066986,
0.008280975744128227,
0.025473611429333687,
-0.024114327505230904,
0.2202693521976471,
-0.08025462180376053,
-0.0389985665678978,
0.10272723436355591,
-0.01638379693031311,
0.05202878639101982,
0.13661865890026093,
0.06344859302043915,
-0.1002432256937027,
0.0017767642857506871,
0.019779972732067108,
-0.03939583897590637,
-0.20598214864730835,
-0.04727305471897125,
-0.05541712045669556,
-0.018903350457549095,
0.11557943373918533,
0.01985863409936428,
0.012497358955442905,
0.028039822354912758,
0.026303302496671677,
0.045422591269016266,
-0.018300360068678856,
0.04630905017256737,
0.0878337025642395,
0.06571214646100998,
0.135676771402359,
-0.022400522604584694,
-0.06027138978242874,
0.020608296617865562,
-0.016414742916822433,
0.23473066091537476,
-0.013095371425151825,
0.20137466490268707,
0.03061450459063053,
0.16906802356243134,
0.01872975565493107,
0.09675321727991104,
0.0046673971228301525,
-0.017704352736473083,
0.005051974672824144,
-0.05530989170074463,
-0.04945319890975952,
0.014122726395726204,
0.047104429453611374,
0.03516077622771263,
-0.10940807312726974,
0.012613285332918167,
0.03585733845829964,
0.3170495629310608,
0.0776170864701271,
-0.34839606285095215,
-0.07390937209129333,
-0.013016562908887863,
-0.036087241023778915,
-0.003267782973125577,
0.017177216708660126,
0.12457776814699173,
-0.07262729108333588,
0.04337482526898384,
-0.06426326185464859,
0.0813080370426178,
-0.0805990919470787,
0.02620769664645195,
0.09023211896419525,
0.07652002573013306,
0.007565135136246681,
0.041297268122434616,
-0.2587263584136963,
0.2838517725467682,
-0.004318278282880783,
0.070929616689682,
-0.07482105493545532,
0.001353977364487946,
0.022713853046298027,
-0.029823005199432373,
0.07360394299030304,
-0.022315090522170067,
-0.04555979371070862,
-0.20260806381702423,
-0.11688780039548874,
0.0290676262229681,
0.12648220360279083,
-0.05183584988117218,
0.12592625617980957,
-0.02922346629202366,
-0.03227366879582405,
0.062445268034935,
-0.0286552794277668,
-0.04415881261229515,
-0.1048024594783783,
0.010009999386966228,
0.042690493166446686,
-0.0029555903747677803,
-0.058409690856933594,
-0.12408589571714401,
-0.11299547553062439,
0.13086995482444763,
-0.04762505367398262,
-0.018438350409269333,
-0.12087701261043549,
0.0736624225974083,
0.14478622376918793,
-0.07976541668176651,
0.044788360595703125,
0.03250378742814064,
0.09116020053625107,
0.0313749685883522,
-0.06074545532464981,
0.10793709754943848,
-0.06689736247062683,
-0.19238173961639404,
-0.05130681395530701,
0.13856886327266693,
0.02958676777780056,
0.07222263514995575,
-0.023041825741529465,
0.03932351991534233,
-0.015764940530061722,
-0.0843951627612114,
0.048639558255672455,
0.0064085982739925385,
0.03589056059718132,
0.028153710067272186,
-0.001020034309476614,
-0.0014434574404731393,
-0.07885363698005676,
-0.016484862193465233,
0.1800914853811264,
0.2678245007991791,
-0.08971340209245682,
0.02392541989684105,
0.05084780603647232,
-0.022419143468141556,
-0.1774776130914688,
-0.006047061178833246,
0.0892411321401596,
0.0221013855189085,
0.0008178792195394635,
-0.1691150665283203,
0.08070521056652069,
0.0849737599492073,
-0.01959114335477352,
0.09106871485710144,
-0.33874836564064026,
-0.13899074494838715,
0.10581193119287491,
0.11596892029047012,
0.053304966539144516,
-0.15275275707244873,
-0.04811062663793564,
-0.020014092326164246,
-0.1304130256175995,
0.10383280366659164,
-0.05223904177546501,
0.1361638754606247,
-0.013433384709060192,
0.039002228528261185,
0.003746871603652835,
-0.05945664271712303,
0.12691622972488403,
0.039356086403131485,
0.058971427381038666,
-0.02408805675804615,
-0.006399394478648901,
0.033507030457258224,
-0.04616154357790947,
0.006908059120178223,
-0.08925729990005493,
0.03174324333667755,
-0.07138524949550629,
-0.03096630796790123,
-0.08720946311950684,
0.005129410419613123,
-0.022368695586919785,
-0.041853662580251694,
-0.013437256217002869,
0.017999302595853806,
0.059332914650440216,
-0.00799314770847559,
0.11010690778493881,
0.001965373521670699,
0.10903619229793549,
0.11134321987628937,
0.07414217293262482,
-0.03563152626156807,
-0.08307656645774841,
-0.016588136553764343,
-0.012996955774724483,
0.05061133950948715,
-0.09668236970901489,
0.024037646129727364,
0.1481938362121582,
0.056753043085336685,
0.1283615231513977,
0.0768548995256424,
-0.042894504964351654,
0.020637743175029755,
0.05204276740550995,
-0.12363353371620178,
-0.11646803468465805,
0.011746270582079887,
-0.050107892602682114,
-0.11625449359416962,
0.04610027000308037,
0.10457577556371689,
-0.04967151954770088,
-0.0058164759539067745,
-0.0062026078812778,
0.010494031012058258,
-0.04344742372632027,
0.2367316484451294,
0.0443745031952858,
0.06958796083927155,
-0.12970708310604095,
0.06730170547962189,
0.0530441552400589,
-0.10555243492126465,
0.023563243448734283,
0.06481710076332092,
-0.08600032329559326,
-0.013578763231635094,
0.05475283041596413,
0.1032770648598671,
-0.027477653697133064,
-0.03764942288398743,
-0.1265496015548706,
-0.12814106047153473,
0.09194999933242798,
0.15150848031044006,
0.07259228825569153,
0.03234374523162842,
-0.060742106288671494,
0.017939027398824692,
-0.12270718812942505,
0.0849352553486824,
0.06058381870388985,
0.07008872926235199,
-0.13519178330898285,
0.14397592842578888,
0.00496272137388587,
0.035630859434604645,
-0.014755996875464916,
0.017994169145822525,
-0.10412195324897766,
0.021360598504543304,
-0.1179341971874237,
-0.008286254480481148,
-0.04704565927386284,
0.00021156130242161453,
-0.0001413919817423448,
-0.0680614784359932,
-0.06431091576814651,
0.015312986448407173,
-0.10380477458238602,
-0.05067344754934311,
-0.0001936201733769849,
0.03036428987979889,
-0.1417863816022873,
-0.02250400371849537,
0.027500653639435768,
-0.10713373869657516,
0.09523878991603851,
0.0762501209974289,
0.013554534874856472,
0.05362871289253235,
-0.09519261121749878,
-0.029166514053940773,
0.06096276640892029,
0.008808131329715252,
0.07377056777477264,
-0.11417648941278458,
-0.012439193204045296,
-0.01197101827710867,
0.048004571348428726,
0.01434131059795618,
0.0826871320605278,
-0.1191350445151329,
0.0021236471366137266,
-0.04806188866496086,
-0.04227995127439499,
-0.05674881115555763,
0.03036189265549183,
0.10856911540031433,
0.03619423136115074,
0.17966106534004211,
-0.07503949850797653,
0.04227844253182411,
-0.22181573510169983,
-0.005052946507930756,
-0.027310647070407867,
-0.09386033564805984,
-0.09154418110847473,
-0.03884311765432358,
0.09631200134754181,
-0.06338021904230118,
0.10928799211978912,
-0.03844885528087616,
0.07256930321455002,
0.017495453357696533,
-0.022085146978497505,
-0.003777025965973735,
0.052841946482658386,
0.21867112815380096,
0.0628066286444664,
-0.037496842443943024,
0.0811912789940834,
0.027371589094400406,
0.09890337288379669,
0.14394418895244598,
0.19622710347175598,
0.141447052359581,
0.04351898655295372,
0.09680735319852829,
0.07180334627628326,
-0.10092625021934509,
-0.17296850681304932,
0.07387059926986694,
-0.07974476367235184,
0.11484849452972412,
-0.006950325332581997,
0.23503327369689941,
0.06437370181083679,
-0.17893406748771667,
0.03505922481417656,
-0.04763348773121834,
-0.09435434639453888,
-0.10086388140916824,
-0.03778677061200142,
-0.08260560035705566,
-0.14386063814163208,
0.0068368916399776936,
-0.10256924480199814,
0.03696717694401741,
0.1238536536693573,
0.026456814259290695,
0.021059952676296234,
0.13356822729110718,
0.04439624026417732,
0.0317283570766449,
0.07344765216112137,
0.01863616518676281,
-0.03553866595029831,
-0.04790304973721504,
-0.07998964935541153,
0.02077808789908886,
-0.013588842004537582,
0.06544902175664902,
-0.052868638187646866,
-0.09502889961004257,
0.07060978561639786,
-0.007886319421231747,
-0.11628642678260803,
0.024350028485059738,
-0.005492459051311016,
0.06598863750696182,
0.04939282685518265,
0.03625290468335152,
0.0019304718589410186,
-0.01119677722454071,
0.2192421853542328,
-0.0959109291434288,
-0.09220154583454132,
-0.11286187171936035,
0.2175886631011963,
0.01615220494568348,
-0.007896321825683117,
0.033421169966459274,
-0.0804862305521965,
-0.024278689175844193,
0.19920481741428375,
0.1836782842874527,
-0.088584303855896,
-0.009707811288535595,
0.0164477676153183,
-0.00728465523570776,
-0.03809462860226631,
0.08747480809688568,
0.14346349239349365,
0.086915984749794,
-0.08416326344013214,
-0.05516454204916954,
-0.04602649435400963,
-0.04841231182217598,
-0.01646982692182064,
0.0637408196926117,
0.021440500393509865,
-0.004825250711292028,
-0.04150371626019478,
0.06968195736408234,
-0.08096238225698471,
-0.1191185936331749,
0.05158674716949463,
-0.22191956639289856,
-0.19245262444019318,
-0.01652621105313301,
0.08106447756290436,
0.025600124150514603,
0.05544940382242203,
-0.006771062035113573,
-0.011821544729173183,
0.07584434747695923,
-0.008405872620642185,
-0.07951556891202927,
-0.061130590736866,
0.07315840572118759,
-0.08976893126964569,
0.16827833652496338,
-0.049956753849983215,
0.0704803317785263,
0.12101004272699356,
0.10006162524223328,
-0.08264777809381485,
0.0632331594824791,
0.06358042359352112,
-0.11292941123247147,
0.04537234827876091,
0.14795053005218506,
-0.04829733818769455,
0.09090689569711685,
0.046832937747240067,
-0.11174771934747696,
0.02450566552579403,
-0.05334358289837837,
-0.017603635787963867,
-0.0530296266078949,
-0.06328128278255463,
-0.03722033277153969,
0.14028745889663696,
0.1946919858455658,
-0.052109554409980774,
0.014621716924011707,
-0.037973225116729736,
-0.00507163442671299,
0.040148988366127014,
0.09003562480211258,
-0.07085806131362915,
-0.27041468024253845,
0.012157590128481388,
-0.0055399988777935505,
0.006381153129041195,
-0.23443563282489777,
-0.08782149851322174,
0.022942131385207176,
-0.06931576132774353,
-0.07337483018636703,
0.09570759534835815,
0.06668588519096375,
0.05350818485021591,
-0.05345505475997925,
-0.03835510462522507,
-0.0536501444876194,
0.1770176887512207,
-0.19843368232250214,
-0.08717186003923416
] |
null | null |
transformers
|
FINETUNED ILOKANO SPEECH RECOGNITION FROM WAV2VEC-XLSR-S3
|
{"language": ["en"], "license": "afl-3.0", "tags": ["audio", "automatic-speech-recognition", "speech"], "datasets": ["timit_asr"], "metrics": ["wer"], "pipeline_tag": "automatic-speech-recognition"}
|
automatic-speech-recognition
|
denden/new_iloko
|
[
"transformers",
"pytorch",
"tensorboard",
"wav2vec2",
"automatic-speech-recognition",
"audio",
"speech",
"en",
"dataset:timit_asr",
"license:afl-3.0",
"model-index",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #audio #speech #en #dataset-timit_asr #license-afl-3.0 #model-index #endpoints_compatible #region-us
|
FINETUNED ILOKANO SPEECH RECOGNITION FROM WAV2VEC-XLSR-S3
|
[] |
[
"TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #audio #speech #en #dataset-timit_asr #license-afl-3.0 #model-index #endpoints_compatible #region-us \n"
] |
[
69
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #audio #speech #en #dataset-timit_asr #license-afl-3.0 #model-index #endpoints_compatible #region-us \n"
] |
[
-0.13549625873565674,
0.08360369503498077,
-0.0043706889264285564,
-0.012407043017446995,
0.09940732270479202,
-0.04235539585351944,
0.13480183482170105,
0.09283918887376785,
0.06561984121799469,
-0.003991053439676762,
0.06014394015073776,
0.20712950825691223,
0.022189591079950333,
0.0006199197378009558,
-0.0686776414513588,
-0.2201278954744339,
0.04040423780679703,
0.04730897396802902,
0.08335338532924652,
0.08653709292411804,
0.09119757264852524,
-0.06303849816322327,
0.042352501302957535,
0.071365587413311,
-0.0920090302824974,
0.015422333963215351,
0.06510736793279648,
-0.16743482649326324,
0.08314935863018036,
0.0367581732571125,
0.03521411493420601,
0.07556039839982986,
0.042823247611522675,
-0.146124467253685,
0.02017354592680931,
-0.02021775208413601,
-0.015032707713544369,
0.041724059730768204,
0.07867497950792313,
-0.05363619327545166,
0.023802246898412704,
0.0947340726852417,
-0.0028362669982016087,
0.08185793459415436,
-0.05498197674751282,
-0.17168252170085907,
-0.01473484467715025,
0.07347101718187332,
0.036500997841358185,
0.05542294681072235,
-0.0251686442643404,
0.13319118320941925,
-0.06925851851701736,
0.10454121232032776,
0.07809668034315109,
-0.28788110613822937,
0.03834415227174759,
0.022538064047694206,
0.09347549080848694,
0.03267132490873337,
-0.017439357936382294,
0.058459825813770294,
0.021154850721359253,
0.023320883512496948,
-0.004474584478884935,
-0.0753093883395195,
-0.11071641743183136,
-0.009179039858281612,
-0.10763460397720337,
-0.01662987656891346,
0.21634754538536072,
-0.0028346667531877756,
0.024792583659291267,
-0.0879308208823204,
-0.01286745723336935,
-0.018977096304297447,
-0.05988733842968941,
-0.02557050995528698,
-0.018938841298222542,
0.04545826464891434,
-0.05122758820652962,
-0.029078323394060135,
-0.10612073540687561,
-0.04925152286887169,
-0.16439619660377502,
0.12592589855194092,
-0.0030657013412564993,
0.04506075382232666,
-0.1441737860441208,
-0.007335870061069727,
-0.024090707302093506,
-0.08425143361091614,
0.0005725830560550094,
-0.009142748080193996,
-0.01761866919696331,
-0.013619562610983849,
-0.015725884586572647,
-0.10687868297100067,
0.11793365329504013,
0.013430877588689327,
0.03364662453532219,
0.030017783865332603,
-0.06504029780626297,
0.10983699560165405,
0.007397054694592953,
0.10734952986240387,
-0.0669793114066124,
-0.032056599855422974,
0.017111921682953835,
-0.050593942403793335,
0.034586116671562195,
-0.04005853831768036,
-0.14819103479385376,
-0.02324441447854042,
0.03125687316060066,
0.04525146260857582,
0.04695838689804077,
0.024117162451148033,
-0.06003721430897713,
-0.04026680067181587,
0.03992938622832298,
-0.08453689515590668,
0.015417035669088364,
0.04905804991722107,
0.0660051703453064,
0.131379634141922,
0.010461051017045975,
0.05148673430085182,
-0.08968979120254517,
0.05769680440425873,
0.004581435117870569,
0.0204874649643898,
0.05432894825935364,
-0.04956791177392006,
0.06464708596467972,
-0.15109455585479736,
0.05258730798959732,
-0.146880105137825,
-0.06902972608804703,
0.0068725054152309895,
0.003945669624954462,
-0.01292718481272459,
-0.06192595884203911,
-0.03475647047162056,
-0.04158087819814682,
0.03238453343510628,
-0.11331330984830856,
-0.01263516116887331,
-0.07427626103162766,
0.07409268617630005,
0.007389073725789785,
0.08445385843515396,
-0.12427050620317459,
0.09658914059400558,
-0.05575410649180412,
-0.017550401389598846,
0.0048479377292096615,
0.08804500102996826,
-0.08559385687112808,
0.025177035480737686,
-0.08998765051364899,
-0.043711189180612564,
-0.13183780014514923,
0.055581942200660706,
-0.026769036427140236,
0.08545961230993271,
-0.17757917940616608,
-0.1477256864309311,
0.15260247886180878,
-0.09806959331035614,
-0.11402745544910431,
0.13654038310050964,
0.046072881668806076,
-0.002029797760769725,
0.11036445945501328,
0.2911289632320404,
0.018751829862594604,
-0.16706599295139313,
0.02091488428413868,
0.10443785786628723,
-0.09328580647706985,
-0.13292773067951202,
0.04824920371174812,
-0.07504253089427948,
0.01264759711921215,
0.015316488221287727,
0.06316658109426498,
0.07631177455186844,
-0.03509187698364258,
-0.07394593209028244,
-0.029369086027145386,
-0.08987962454557419,
0.016206013038754463,
0.010784140788018703,
0.05268368497490883,
-0.006670397240668535,
-0.009142079390585423,
0.0464569628238678,
0.023902451619505882,
-0.04851307347416878,
0.04430714249610901,
-0.15443669259548187,
0.17245209217071533,
-0.12576742470264435,
-0.015360964462161064,
-0.18342843651771545,
0.15776236355304718,
-0.04001631960272789,
0.02428242191672325,
0.10727527737617493,
0.06757842004299164,
0.08541589975357056,
-0.056078121066093445,
-0.0016013489803299308,
-0.010109839960932732,
0.1712413877248764,
0.07178958505392075,
-0.021313663572072983,
-0.173935666680336,
0.041428253054618835,
-0.0787975862622261,
-0.009073140099644661,
-0.03295684605836868,
-0.05105612054467201,
0.09208314120769501,
0.0942627564072609,
0.016834605485200882,
0.011499698273837566,
0.03409024700522423,
0.03408968821167946,
-0.006296144798398018,
0.011959158815443516,
0.0779418870806694,
0.013795674778521061,
-0.05699940770864487,
0.2393738478422165,
-0.17045776546001434,
0.2550928294658661,
0.2318093329668045,
-0.17233189940452576,
0.0519804023206234,
0.07004240155220032,
0.020847372710704803,
-0.010127482935786247,
0.03433708846569061,
-0.02165284752845764,
0.19506309926509857,
-0.01868952065706253,
0.15771935880184174,
-0.06318216025829315,
0.007490986958146095,
0.06586505472660065,
-0.040580447763204575,
-0.04572359472513199,
0.062210891395807266,
0.05115789920091629,
-0.13596603274345398,
0.10985822230577469,
0.13634435832500458,
-0.07479681074619293,
0.18662412464618683,
-0.04677288234233856,
-0.07106298953294754,
0.052564699202775955,
-0.025846516713500023,
-0.013396774418652058,
0.07910474389791489,
-0.26627978682518005,
-0.045301731675863266,
0.07952293008565903,
-0.023680217564105988,
0.0731632336974144,
-0.15538956224918365,
0.015092914924025536,
-0.008429840207099915,
-0.06408988684415817,
-0.12975192070007324,
0.08135440200567245,
-0.012240644544363022,
0.06984298676252365,
-0.07614192366600037,
-0.166228249669075,
0.0712142139673233,
-0.024259822443127632,
-0.11977732926607132,
0.06560184806585312,
-0.0737195834517479,
-0.25755295157432556,
-0.13702504336833954,
-0.12814241647720337,
0.007739553693681955,
0.044059012085199356,
0.10350622981786728,
-0.06940028816461563,
-0.029868649318814278,
0.0009594953735359013,
-0.007732667960226536,
-0.04387245327234268,
0.02009480446577072,
0.011991262435913086,
0.030215831473469734,
0.034968022257089615,
-0.11394145339727402,
-0.01042028609663248,
-0.053667616099119186,
0.0630316361784935,
0.03784768283367157,
-0.01846950687468052,
0.06149173155426979,
0.22456464171409607,
0.060377225279808044,
0.045644260942935944,
-0.028336992487311363,
0.15408557653427124,
-0.08747170120477676,
-0.04095581918954849,
0.1432211697101593,
-0.08879462629556656,
0.008388865739107132,
0.17521102726459503,
0.04657845199108124,
-0.025776615366339684,
-0.06709104031324387,
-0.054194219410419464,
-0.09234759956598282,
-0.23351089656352997,
-0.11272045969963074,
-0.11380650103092194,
-0.012197611853480339,
0.011049752123653889,
0.06112867593765259,
0.11612637341022491,
0.030769916251301765,
0.02884809300303459,
-0.029193930327892303,
0.04233783483505249,
0.008801708929240704,
0.2177366018295288,
-0.03607727214694023,
0.1274186670780182,
-0.08417370915412903,
-0.08640021085739136,
0.0688038021326065,
0.09684557467699051,
0.10914385318756104,
0.16923238337039948,
0.09962798655033112,
0.049982279539108276,
0.09363525360822678,
0.14252528548240662,
0.07361733913421631,
0.030868252739310265,
-0.008246704004704952,
0.017239514738321304,
-0.07055042684078217,
0.006741194520145655,
0.04783232882618904,
0.23899127542972565,
-0.08984679728746414,
-0.003313420107588172,
-0.09034429490566254,
0.004224391188472509,
0.12277261912822723,
0.086509108543396,
-0.2511967420578003,
0.009261716157197952,
0.0405179001390934,
-0.050748150795698166,
-0.02107560820877552,
0.0955590158700943,
0.01255137100815773,
-0.039713822305202484,
0.09152861684560776,
-0.003740505315363407,
0.07409925013780594,
-0.0686006024479866,
0.027802646160125732,
-0.0294120442122221,
-0.049208592623472214,
0.051084354519844055,
0.0320618562400341,
-0.2300817370414734,
0.25305336713790894,
0.02741946652531624,
0.07718540728092194,
-0.03243664652109146,
-0.002465157536789775,
0.026957334950566292,
0.0813678726553917,
0.17365676164627075,
-0.01469623576849699,
-0.042839620262384415,
-0.07496310025453568,
-0.07468492537736893,
0.02791535295546055,
0.0892632007598877,
0.1232747733592987,
-0.0522049143910408,
-0.02297995425760746,
-0.046376097947359085,
0.041600536555051804,
-0.11441162973642349,
-0.12271565198898315,
-0.1247379258275032,
0.031146373599767685,
0.24455255270004272,
0.009973435662686825,
0.013589518144726753,
-0.09662063419818878,
-0.23677395284175873,
0.08888331800699234,
-0.15586186945438385,
-0.017983589321374893,
-0.06959991902112961,
-0.15388700366020203,
0.11729525029659271,
-0.026326045393943787,
0.05412502959370613,
-0.00613178638741374,
0.006480544339865446,
-0.032117750495672226,
-0.16305242478847504,
0.12485634535551071,
-0.12236017733812332,
-0.020516563206911087,
-0.04165051877498627,
0.22219568490982056,
-0.01652330905199051,
0.06897369772195816,
0.05927624925971031,
0.03721662610769272,
-0.02741972729563713,
-0.028943320736289024,
0.0932512953877449,
0.07169579714536667,
-0.0661495253443718,
-0.006774176377803087,
0.010749083943665028,
-0.19389089941978455,
-0.021682264283299446,
0.00699691753834486,
0.21580910682678223,
0.09998921304941177,
-0.05400067940354347,
0.21341253817081451,
0.23708683252334595,
-0.05244521051645279,
-0.29050812125205994,
-0.08486324548721313,
-0.028284523636102676,
0.007168320938944817,
-0.02033318020403385,
-0.11914574354887009,
0.1044900119304657,
-0.08824335783720016,
-0.0892191082239151,
0.04974637180566788,
-0.21742604672908783,
-0.11236308515071869,
0.2840105891227722,
-0.049544334411621094,
0.2305469810962677,
-0.10498202592134476,
-0.12421122193336487,
-0.03468795493245125,
-0.13635849952697754,
0.1061081513762474,
-0.11701306700706482,
0.11314346641302109,
0.04101182892918587,
0.05362556874752045,
0.02317030355334282,
-0.05087834969162941,
0.06978579610586166,
0.07799869030714035,
-0.06528396159410477,
-0.03770894184708595,
0.0057250079698860645,
0.05592656508088112,
0.044973451644182205,
0.0689590722322464,
-0.04033688083291054,
0.015345252119004726,
-0.03527281805872917,
-0.0770154520869255,
-0.07462222874164581,
0.10190171748399734,
0.07720343768596649,
-0.022664019837975502,
0.030975086614489555,
-0.07251652330160141,
-0.005982062313705683,
-0.00985516794025898,
0.18100546300411224,
-0.08994046598672867,
0.022303035482764244,
0.19594524800777435,
0.2211301624774933,
-0.16735708713531494,
-0.10587461292743683,
-0.06745857745409012,
-0.10777518153190613,
0.09325125813484192,
-0.016195446252822876,
0.06576874852180481,
0.07617130130529404,
0.047768209129571915,
0.06051874905824661,
0.07902121543884277,
-0.057527847588062286,
0.03612038865685463,
0.1048332080245018,
-0.0755556970834732,
-0.11720147728919983,
-0.01771317981183529,
-0.047775961458683014,
0.06406097114086151,
0.10668355226516724,
0.12852831184864044,
-0.013078585267066956,
0.003992339130491018,
-0.002442652592435479,
-0.008625905960798264,
-0.13619142770767212,
0.1648930460214615,
0.08263710141181946,
0.006711372174322605,
-0.17568229138851166,
0.08649420738220215,
-0.023142391815781593,
-0.16849946975708008,
0.02312539704144001,
-0.026618115603923798,
-0.07773490995168686,
-0.10810575634241104,
-0.09219865500926971,
0.03423627093434334,
-0.0692325085401535,
-0.128292053937912,
-0.009848574176430702,
-0.13461405038833618,
0.05791603401303291,
0.15653783082962036,
0.04903043434023857,
0.06260035187005997,
-0.11426473408937454,
-0.056606195867061615,
-0.03638269379734993,
0.02323521487414837,
0.002878843806684017,
0.024615900591015816,
-0.18722499907016754,
-0.00939068105071783,
0.004753759130835533,
0.07031183689832687,
-0.09590084105730057,
-0.058940339833498,
-0.08634592592716217,
0.07392407953739166,
-0.11300572007894516,
0.015260727144777775,
-0.11080415546894073,
-0.0029246872290968895,
0.035500358790159225,
-0.08234845101833344,
-0.05768611282110214,
0.04900553822517395,
-0.10360513627529144,
0.04195258766412735,
0.02276856079697609,
0.0703430324792862,
-0.11666438728570938,
0.011903835460543633,
0.015861809253692627,
-0.010645565576851368,
0.09606423228979111,
0.15564627945423126,
-0.1589205116033554,
0.0916481465101242,
-0.2552160918712616,
-0.17139315605163574,
0.1665915697813034,
0.046627227216959,
0.02094328962266445,
-0.04459408298134804,
0.020144972950220108,
0.15071067214012146,
0.03850957006216049,
0.02398001216351986,
0.09591973572969437,
-0.06603536754846573,
-0.009421488270163536,
-0.14309445023536682,
-0.05898917838931084,
-0.028798943385481834,
-0.004268442280590534,
0.15646350383758545,
0.06594063341617584,
0.17983336746692657,
-0.05744825676083565,
-0.002828087192028761,
-0.05787104740738869,
0.050568658858537674,
-0.04977260157465935,
-0.11298460513353348,
-0.12517033517360687,
-0.02351921796798706,
0.02564983442425728,
-0.048153359442949295,
0.24428994953632355,
-0.003382281167432666,
-0.05010884255170822,
0.02343989908695221,
0.03736865520477295,
-0.04381848871707916,
0.021600201725959778,
0.3363698720932007,
0.06852886825799942,
-0.02928221970796585,
-0.08604293316602707,
0.012037944979965687,
0.023349788039922714,
0.09378482401371002,
-0.033529434353113174,
0.14561693370342255,
-0.04599663242697716,
0.11144206672906876,
0.1600867509841919,
-0.007878396660089493,
-0.09590311348438263,
0.036589935421943665,
-0.0669764056801796,
0.09298676997423172,
-0.007699829991906881,
0.14906732738018036,
0.14714469015598297,
0.003961531445384026,
0.04380554333329201,
-0.04500682279467583,
-0.04288739338517189,
-0.19326739013195038,
-0.17796018719673157,
-0.10501120984554291,
-0.10884387046098709,
0.04414762556552887,
-0.03141931816935539,
0.053008463233709335,
0.030399156734347343,
0.055422402918338776,
-0.028314976021647453,
0.05731359124183655,
-0.049218907952308655,
-0.04880337417125702,
0.08220558613538742,
-0.0418369397521019,
-0.03207944706082344,
-0.06727342307567596,
-0.006606273353099823,
0.0552370585501194,
-0.002479142975062132,
-0.0003171097778249532,
0.0017046728171408176,
-0.08257318288087845,
0.04583386704325676,
-0.12717881798744202,
-0.08503598719835281,
-0.012750714085996151,
0.0184719730168581,
0.03767387941479683,
0.11140426248311996,
0.07502266764640808,
-0.06098998710513115,
0.061532583087682724,
0.11877649277448654,
-0.08091478794813156,
-0.15110839903354645,
-0.05765498802065849,
0.10771114379167557,
-0.013474317267537117,
0.09226913005113602,
-0.0046342601999640465,
-0.048113882541656494,
-0.02790616638958454,
0.24846777319908142,
0.29253482818603516,
-0.03234269097447395,
0.043440524488687515,
-0.023993894457817078,
0.019681500270962715,
-0.039152778685092926,
0.018076876178383827,
0.09571801126003265,
0.2645002007484436,
-0.017976339906454086,
-0.06759243458509445,
-0.06433643400669098,
-0.02560669183731079,
-0.029092663899064064,
0.08139187842607498,
-0.01443407777696848,
-0.12476199865341187,
-0.02159402333199978,
0.13960543274879456,
-0.21342729032039642,
0.038906410336494446,
-0.11297053843736649,
-0.1114756166934967,
-0.050101108849048615,
-0.017566729336977005,
0.10341931879520416,
0.10894252359867096,
-0.013720747083425522,
-0.06663309037685394,
-0.05408855527639389,
0.01207411102950573,
0.003680514870211482,
-0.2475670874118805,
0.10241923481225967,
-0.008097765035927296,
-0.06413742899894714,
0.005297896917909384,
-0.005185208283364773,
0.16061437129974365,
0.018039962276816368,
0.15412281453609467,
-0.007380988448858261,
0.10227147489786148,
-0.004634958226233721,
-0.13080410659313202,
0.06818804889917374,
0.12312646210193634,
0.006803934928029776,
0.016599012538790703,
0.04741404950618744,
-0.06473908573389053,
0.06843111664056778,
-0.13313528895378113,
-0.08401574939489365,
-0.08396600186824799,
-0.0006879466236568987,
-0.04762842878699303,
0.07421748340129852,
-0.05381282418966293,
0.003752744058147073,
-0.04104170948266983,
-0.03555850312113762,
0.02631116658449173,
0.0868518278002739,
-0.11412766575813293,
-0.0655268058180809,
-0.13622330129146576,
-0.053469881415367126,
-0.08656245470046997,
0.004377757199108601,
-0.1849241852760315,
-0.02917812205851078,
-0.08508797734975815,
-0.016499320045113564,
-0.09406041353940964,
0.009476333856582642,
0.13245031237602234,
0.004016590770334005,
-0.002371553098782897,
0.016383089125156403,
0.0887615978717804,
0.08663437515497208,
-0.16421958804130554,
-0.07580972462892532
] |
null | null |
transformers
|
# BERT-Wiki-Paragraphs
Authors: Satya Almasian\*, Dennis Aumiller\*, Lucienne-Sophie Marmé, Michael Gertz
Contact us at `<lastname>@informatik.uni-heidelberg.de`
Details for the training method can be found in our work [Structural Text Segmentation of Legal Documents](https://arxiv.org/abs/2012.03619).
The training procedure follows the same setup, but we substitute legal documents for Wikipedia in this model.
Find the associated training data here: [wiki-paragraphs](https://huggingface.co/datasets/dennlinger/wiki-paragraphs)
Training is performed in a form of weakly-supervised fashion to determine whether paragraphs topically belong together or not.
We utilize automatically generated samples from Wikipedia for training, where paragraphs from within the same section are assumed to be topically coherent.
We use the same articles as ([Koshorek et al., 2018](https://arxiv.org/abs/1803.09337)),
albeit from a 2021 dump of Wikpeida, and split at paragraph boundaries instead of the sentence level.
## Usage
Preferred usage is through `transformers.pipeline`:
```python
from transformers import pipeline
pipe = pipeline("text-classification", model="dennlinger/bert-wiki-paragraphs")
pipe("{First paragraph} [SEP] {Second paragraph}")
```
A predicted "1" means that paragraphs belong to the same topic, a "0" indicates a disconnect.
## Training Setup
The model was trained for 3 epochs from `bert-base-uncased` on paragraph pairs (limited to 512 subwork with the `longest_first` truncation strategy).
We use a batch size of 24 wit 2 iterations gradient accumulation (effective batch size of 48), and a learning rate of 1e-4, with gradient clipping at 5.
Training was performed on a single Titan RTX GPU over the duration of 3 weeks.
|
{"language": ["en"], "license": "mit", "tags": ["sentence-similarity", "text-classification"], "datasets": ["dennlinger/wiki-paragraphs"], "metrics": ["f1"]}
|
text-classification
|
dennlinger/bert-wiki-paragraphs
|
[
"transformers",
"pytorch",
"safetensors",
"bert",
"text-classification",
"sentence-similarity",
"en",
"dataset:dennlinger/wiki-paragraphs",
"arxiv:2012.03619",
"arxiv:1803.09337",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2012.03619",
"1803.09337"
] |
[
"en"
] |
TAGS
#transformers #pytorch #safetensors #bert #text-classification #sentence-similarity #en #dataset-dennlinger/wiki-paragraphs #arxiv-2012.03619 #arxiv-1803.09337 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# BERT-Wiki-Paragraphs
Authors: Satya Almasian\*, Dennis Aumiller\*, Lucienne-Sophie Marmé, Michael Gertz
Contact us at '<lastname>@URL'
Details for the training method can be found in our work Structural Text Segmentation of Legal Documents.
The training procedure follows the same setup, but we substitute legal documents for Wikipedia in this model.
Find the associated training data here: wiki-paragraphs
Training is performed in a form of weakly-supervised fashion to determine whether paragraphs topically belong together or not.
We utilize automatically generated samples from Wikipedia for training, where paragraphs from within the same section are assumed to be topically coherent.
We use the same articles as (Koshorek et al., 2018),
albeit from a 2021 dump of Wikpeida, and split at paragraph boundaries instead of the sentence level.
## Usage
Preferred usage is through 'transformers.pipeline':
A predicted "1" means that paragraphs belong to the same topic, a "0" indicates a disconnect.
## Training Setup
The model was trained for 3 epochs from 'bert-base-uncased' on paragraph pairs (limited to 512 subwork with the 'longest_first' truncation strategy).
We use a batch size of 24 wit 2 iterations gradient accumulation (effective batch size of 48), and a learning rate of 1e-4, with gradient clipping at 5.
Training was performed on a single Titan RTX GPU over the duration of 3 weeks.
|
[
"# BERT-Wiki-Paragraphs\n\nAuthors: Satya Almasian\\*, Dennis Aumiller\\*, Lucienne-Sophie Marmé, Michael Gertz \nContact us at '<lastname>@URL' \nDetails for the training method can be found in our work Structural Text Segmentation of Legal Documents.\nThe training procedure follows the same setup, but we substitute legal documents for Wikipedia in this model.\nFind the associated training data here: wiki-paragraphs\n\nTraining is performed in a form of weakly-supervised fashion to determine whether paragraphs topically belong together or not.\nWe utilize automatically generated samples from Wikipedia for training, where paragraphs from within the same section are assumed to be topically coherent. \nWe use the same articles as (Koshorek et al., 2018), \nalbeit from a 2021 dump of Wikpeida, and split at paragraph boundaries instead of the sentence level.",
"## Usage\nPreferred usage is through 'transformers.pipeline':\n\n\nA predicted \"1\" means that paragraphs belong to the same topic, a \"0\" indicates a disconnect.",
"## Training Setup\nThe model was trained for 3 epochs from 'bert-base-uncased' on paragraph pairs (limited to 512 subwork with the 'longest_first' truncation strategy).\nWe use a batch size of 24 wit 2 iterations gradient accumulation (effective batch size of 48), and a learning rate of 1e-4, with gradient clipping at 5.\nTraining was performed on a single Titan RTX GPU over the duration of 3 weeks."
] |
[
"TAGS\n#transformers #pytorch #safetensors #bert #text-classification #sentence-similarity #en #dataset-dennlinger/wiki-paragraphs #arxiv-2012.03619 #arxiv-1803.09337 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# BERT-Wiki-Paragraphs\n\nAuthors: Satya Almasian\\*, Dennis Aumiller\\*, Lucienne-Sophie Marmé, Michael Gertz \nContact us at '<lastname>@URL' \nDetails for the training method can be found in our work Structural Text Segmentation of Legal Documents.\nThe training procedure follows the same setup, but we substitute legal documents for Wikipedia in this model.\nFind the associated training data here: wiki-paragraphs\n\nTraining is performed in a form of weakly-supervised fashion to determine whether paragraphs topically belong together or not.\nWe utilize automatically generated samples from Wikipedia for training, where paragraphs from within the same section are assumed to be topically coherent. \nWe use the same articles as (Koshorek et al., 2018), \nalbeit from a 2021 dump of Wikpeida, and split at paragraph boundaries instead of the sentence level.",
"## Usage\nPreferred usage is through 'transformers.pipeline':\n\n\nA predicted \"1\" means that paragraphs belong to the same topic, a \"0\" indicates a disconnect.",
"## Training Setup\nThe model was trained for 3 epochs from 'bert-base-uncased' on paragraph pairs (limited to 512 subwork with the 'longest_first' truncation strategy).\nWe use a batch size of 24 wit 2 iterations gradient accumulation (effective batch size of 48), and a learning rate of 1e-4, with gradient clipping at 5.\nTraining was performed on a single Titan RTX GPU over the duration of 3 weeks."
] |
[
85,
205,
44,
111
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #bert #text-classification #sentence-similarity #en #dataset-dennlinger/wiki-paragraphs #arxiv-2012.03619 #arxiv-1803.09337 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# BERT-Wiki-Paragraphs\n\nAuthors: Satya Almasian\\*, Dennis Aumiller\\*, Lucienne-Sophie Marmé, Michael Gertz \nContact us at '<lastname>@URL' \nDetails for the training method can be found in our work Structural Text Segmentation of Legal Documents.\nThe training procedure follows the same setup, but we substitute legal documents for Wikipedia in this model.\nFind the associated training data here: wiki-paragraphs\n\nTraining is performed in a form of weakly-supervised fashion to determine whether paragraphs topically belong together or not.\nWe utilize automatically generated samples from Wikipedia for training, where paragraphs from within the same section are assumed to be topically coherent. \nWe use the same articles as (Koshorek et al., 2018), \nalbeit from a 2021 dump of Wikpeida, and split at paragraph boundaries instead of the sentence level.## Usage\nPreferred usage is through 'transformers.pipeline':\n\n\nA predicted \"1\" means that paragraphs belong to the same topic, a \"0\" indicates a disconnect.## Training Setup\nThe model was trained for 3 epochs from 'bert-base-uncased' on paragraph pairs (limited to 512 subwork with the 'longest_first' truncation strategy).\nWe use a batch size of 24 wit 2 iterations gradient accumulation (effective batch size of 48), and a learning rate of 1e-4, with gradient clipping at 5.\nTraining was performed on a single Titan RTX GPU over the duration of 3 weeks."
] |
[
-0.04949145019054413,
0.04778731241822243,
-0.003253311151638627,
0.04117292910814285,
0.07969529926776886,
0.020363399758934975,
0.15439142286777496,
0.09532720595598221,
0.02338678017258644,
0.05085999518632889,
0.03147181123495102,
0.002499239519238472,
0.07789704203605652,
0.058180950582027435,
-0.0013631696347147226,
-0.2735196352005005,
0.045316386967897415,
-0.07556503266096115,
-0.07484546303749084,
0.07994433492422104,
0.10818462818861008,
-0.07049892842769623,
0.04818994179368019,
0.024052834138274193,
-0.08832080662250519,
-0.015739742666482925,
-0.034693799912929535,
-0.028698986396193504,
0.08829092979431152,
0.09439203143119812,
0.13616597652435303,
0.024797558784484863,
0.033323612064123154,
-0.15408626198768616,
0.023462040349841118,
0.09720330685377121,
0.06043456494808197,
0.054621368646621704,
0.06864233314990997,
-0.0382576547563076,
0.06234622001647949,
-0.019309304654598236,
0.030575208365917206,
0.02905302122235298,
-0.13620585203170776,
-0.05993502587080002,
-0.09787125885486603,
0.006732639390975237,
0.033041518181562424,
0.0035056015476584435,
-0.032953567802906036,
0.040421582758426666,
-0.044992949813604355,
0.017397748306393623,
0.05433798208832741,
-0.18835489451885223,
-0.028847329318523407,
-0.014268866740167141,
-0.03310906141996384,
0.035564154386520386,
-0.0990876778960228,
-0.055121056735515594,
0.012366615235805511,
-0.01662035658955574,
0.07790954411029816,
0.036010924726724625,
0.01319773681461811,
-0.04374590888619423,
-0.10602381825447083,
-0.022200601175427437,
0.15209873020648956,
0.01457285787910223,
-0.06919686496257782,
-0.11893940716981888,
-0.030120307579636574,
0.11805404722690582,
0.05418005213141441,
-0.01016983576118946,
-0.019483283162117004,
0.04574305936694145,
0.036318328231573105,
-0.04586486890912056,
-0.14227734506130219,
-0.02062508463859558,
-0.058943744748830795,
0.18689529597759247,
0.04162564501166344,
0.0517875999212265,
-0.006696093827486038,
0.09733545780181885,
0.011195125058293343,
-0.0760241374373436,
-0.0015848211478441954,
-0.08646736294031143,
-0.07078622281551361,
0.0021043934393674135,
-0.07605279982089996,
-0.1987149566411972,
-0.011980758048593998,
0.08928494155406952,
0.013507851399481297,
-0.0050821746699512005,
0.05156321078538895,
0.05432808771729469,
0.08780436217784882,
0.06947103887796402,
-0.04005463048815727,
0.010676062665879726,
-0.027144530788064003,
0.04469433054327965,
-0.03829370066523552,
-0.004501914605498314,
0.004094310570508242,
-0.044895924627780914,
0.004164198879152536,
-0.0026527075096964836,
0.014177090488374233,
0.07293149828910828,
-0.04560726881027222,
-0.08414310216903687,
0.15094439685344696,
-0.14117757976055145,
-0.03199876844882965,
0.03290935233235359,
-0.03935718908905983,
0.10598792135715485,
0.04405706748366356,
-0.013325270265340805,
-0.03853645175695419,
0.10599929094314575,
-0.03269176557660103,
-0.007500520907342434,
-0.09431839734315872,
-0.10702867060899734,
-0.0006855261162854731,
-0.030447032302618027,
-0.0668608695268631,
-0.10248963534832001,
-0.1832590252161026,
-0.08794834464788437,
0.07062126696109772,
-0.008274414576590061,
0.02202875353395939,
-0.0408359095454216,
0.056839216500520706,
-0.02737017720937729,
0.0032028236892074347,
-0.0025967333931475878,
-0.012912538833916187,
0.08388756960630417,
-0.09953325241804123,
0.07244590669870377,
-0.045562032610177994,
0.060814268887043,
-0.07866758108139038,
0.00804821215569973,
-0.15324676036834717,
0.14073310792446136,
-0.015917351469397545,
0.03853195533156395,
-0.08555621653795242,
0.00412357272580266,
-0.09143728017807007,
-0.0011053073685616255,
0.02529650554060936,
0.0853225588798523,
-0.1919722855091095,
-0.06388825178146362,
0.11958643049001694,
-0.14440324902534485,
0.029787203297019005,
0.12586356699466705,
-0.059009939432144165,
0.09457618743181229,
0.11892887949943542,
0.17357440292835236,
0.12780344486236572,
-0.060205500572919846,
-0.017477069050073624,
-0.06484165042638779,
-0.03347661346197128,
0.1058686226606369,
0.04205545037984848,
-0.024971840903162956,
-0.0020580049604177475,
0.04727397859096527,
-0.030472470447421074,
0.03820972144603729,
-0.0048998137935996056,
-0.05023536831140518,
0.009387318044900894,
-0.06081021577119827,
0.04665590077638626,
-0.028210259974002838,
-0.0008501215488649905,
-0.02190592885017395,
-0.09424279630184174,
0.04255636781454086,
0.08991851657629013,
-0.09283901751041412,
0.003980805166065693,
-0.03698793053627014,
0.042106274515390396,
0.030614927411079407,
-0.004389656241983175,
-0.13302890956401825,
-0.0783000960946083,
0.05445706844329834,
-0.09736227989196777,
0.07735159993171692,
0.00398238655179739,
0.019747449085116386,
0.037900641560554504,
-0.05658118799328804,
0.02136247046291828,
-0.047242309898138046,
-0.042683809995651245,
-0.03294357657432556,
-0.07370751351118088,
-0.028035108000040054,
-0.017625607550144196,
-0.018754027783870697,
-0.1297270506620407,
0.03069351054728031,
-0.014223819598555565,
0.0780542716383934,
0.030745254829525948,
-0.04329662770032883,
0.03134700283408165,
0.02360416203737259,
-0.028782542794942856,
-0.033860545605421066,
0.03249283507466316,
0.03565572574734688,
-0.018634416162967682,
0.045075252652168274,
-0.11194822192192078,
-0.23088498413562775,
0.03965805843472481,
0.057556863874197006,
-0.09482651203870773,
-0.0974106714129448,
-0.046876195818185806,
-0.03754754737019539,
-0.10520067065954208,
-0.04589463025331497,
0.13394464552402496,
0.03519921749830246,
0.06886466592550278,
-0.07307159900665283,
-0.008725106716156006,
-0.03933268412947655,
-0.01261451281607151,
-0.01145438477396965,
-0.016949910670518875,
0.033905718475580215,
-0.09462129324674606,
0.06697996705770493,
0.04206710681319237,
0.002496482105925679,
0.23056960105895996,
0.008965922519564629,
-0.11154768615961075,
0.004317048937082291,
0.0539972186088562,
-0.01902199350297451,
0.024757202714681625,
0.01461412850767374,
0.006086277775466442,
0.05046050250530243,
0.036927249282598495,
0.0037987015675753355,
-0.12315557897090912,
0.06792711466550827,
0.01908578909933567,
-0.025630544871091843,
-0.011306186206638813,
-0.012420513667166233,
-0.004917059559375048,
0.06558770686388016,
0.0345250740647316,
0.07877150923013687,
-0.029448334127664566,
-0.047835446894168854,
-0.04788153991103172,
0.13007937371730804,
-0.06902330368757248,
-0.26059943437576294,
-0.16097606718540192,
0.057661645114421844,
-0.06702259927988052,
-0.0010121348313987255,
0.016216592863202095,
-0.03930915147066116,
-0.06203524023294449,
-0.033548127859830856,
0.17727939784526825,
0.05583930015563965,
-0.025541698560118675,
-0.09054184705018997,
0.07749289274215698,
0.0033520315773785114,
-0.10320895910263062,
0.009534334763884544,
-0.027691436931490898,
-0.026035049930214882,
0.07398155331611633,
-0.02709890715777874,
0.013674695044755936,
0.05825291574001312,
-0.0188569575548172,
-0.05071841925382614,
-0.018616361543536186,
0.17005592584609985,
-0.07315950840711594,
0.10555152595043182,
0.13815641403198242,
-0.07944843918085098,
0.08733388781547546,
0.123905248939991,
0.010065286420285702,
-0.06812123954296112,
0.05761202797293663,
0.04642573744058609,
-0.019566843286156654,
-0.23236232995986938,
-0.0710163414478302,
-0.06073429435491562,
-0.04598742723464966,
0.13115017116069794,
0.014040147885680199,
-0.05442828685045242,
0.08465315401554108,
-0.08606593310832977,
0.035380925983190536,
0.026604492217302322,
0.06674734503030777,
0.09988272935152054,
0.027623821049928665,
0.10432764887809753,
-0.05246161296963692,
-0.022223161533474922,
0.07117485255002975,
0.1160091906785965,
0.3007010519504547,
-0.021918995305895805,
0.13451600074768066,
0.09002580493688583,
0.07613036036491394,
0.07285833358764648,
0.08442802727222443,
-0.015375562012195587,
0.0614825077354908,
-0.05853280797600746,
-0.05351248383522034,
-0.008599481545388699,
0.0881849005818367,
0.017350230365991592,
-0.06058415398001671,
0.011191974394023418,
-0.05588681250810623,
0.06629052013158798,
0.3145799934864044,
0.07770033925771713,
-0.19035302102565765,
-0.09849537163972855,
0.006099478807300329,
-0.06865059584379196,
-0.0735296905040741,
-0.02284787967801094,
0.08272865414619446,
-0.09469377249479294,
0.052900832146406174,
-0.01792684569954872,
0.07421116530895233,
-0.13297507166862488,
-0.024436311796307564,
-0.025656579062342644,
0.08671096712350845,
-0.052135780453681946,
0.07317577302455902,
-0.1273004412651062,
0.15580062568187714,
0.004222490359097719,
0.07941118627786636,
-0.050924576818943024,
0.0164805818349123,
0.05359648913145065,
-0.04452275112271309,
0.043423037976026535,
0.008891553618013859,
-0.0005376251647248864,
0.03872055560350418,
-0.10156764090061188,
0.03400710970163345,
0.07889723777770996,
-0.051971033215522766,
0.09147121012210846,
-0.03555896133184433,
0.044050268828868866,
-0.01968517154455185,
-0.004218325484544039,
-0.06976260244846344,
-0.15889590978622437,
0.01042462233453989,
-0.0534248948097229,
0.08230730146169662,
-0.009631126187741756,
-0.026151640340685844,
-0.008099132217466831,
0.17034912109375,
-0.0984342098236084,
-0.11750034242868423,
-0.08721227198839188,
0.02526051551103592,
0.10328540951013565,
-0.04103751480579376,
-0.025079848244786263,
-0.002259090542793274,
0.1789824217557907,
-0.05449488013982773,
-0.09674032032489777,
0.0008090754272416234,
-0.04635919630527496,
-0.14398528635501862,
-0.08420838415622711,
0.07308465987443924,
0.08484560251235962,
0.002158929593861103,
-0.005269858054816723,
0.022155605256557465,
-0.06926567107439041,
-0.06233600154519081,
-0.005988110322505236,
0.10947440564632416,
0.008982112631201744,
0.06611981987953186,
-0.11976113170385361,
-0.04433456435799599,
-0.1583935022354126,
-0.055270902812480927,
0.05565997585654259,
0.17127589881420135,
-0.03159709274768829,
0.1115243062376976,
0.20024345815181732,
-0.056916285306215286,
-0.19214388728141785,
-0.04733307659626007,
0.09703674912452698,
0.06844599545001984,
-0.04649906978011131,
-0.2633339762687683,
0.03433221951127052,
0.13471238315105438,
0.01242039818316698,
0.06629423797130585,
-0.1153012216091156,
-0.13095617294311523,
0.07639939337968826,
0.00847750622779131,
0.04798518866300583,
-0.06998904049396515,
0.004633009433746338,
-0.06186620891094208,
0.13340051472187042,
0.16320689022541046,
0.08087771385908127,
0.098151296377182,
-0.007467315997928381,
0.06981413066387177,
0.028527701273560524,
0.007359492126852274,
0.11741825938224792,
0.03094492480158806,
0.04626837000250816,
-0.02731447108089924,
0.05436360463500023,
0.012964264489710331,
-0.05649137869477272,
0.10413320362567902,
-0.02180919051170349,
0.029413221403956413,
-0.01058632880449295,
-0.08982454985380173,
-0.061201851814985275,
0.058550529181957245,
0.01328012254089117,
-0.0927216112613678,
-0.09279932081699371,
0.062498029321432114,
0.05283765122294426,
0.019709395244717598,
-0.06362169235944748,
-0.02626350149512291,
0.02493111416697502,
0.09501148760318756,
0.05208594724535942,
-0.005870317108929157,
-0.11990102380514145,
-0.015049058943986893,
-0.0036569724325090647,
0.1318967640399933,
-0.15536974370479584,
-0.0008061086991801858,
0.07837115228176117,
0.024726474657654762,
0.12918773293495178,
0.04191241040825844,
-0.1307530254125595,
0.01305653527379036,
0.07498442381620407,
-0.09779156744480133,
-0.10162752121686935,
0.024655383080244064,
-0.0047973995096981525,
-0.1136075034737587,
-0.08779583871364594,
0.12279908359050751,
-0.08127278834581375,
-0.0210096575319767,
0.011333822272717953,
0.037290822714567184,
-0.015625135973095894,
0.09444131702184677,
0.016077004373073578,
0.013675376772880554,
-0.06129702553153038,
0.15030162036418915,
0.11100843548774719,
-0.06199585273861885,
-0.005104449111968279,
0.02132990211248398,
-0.09338818490505219,
-0.03200845792889595,
-0.10332761704921722,
-0.0024592969566583633,
0.009791317395865917,
-0.06248423457145691,
-0.008004526607692242,
-0.05685414746403694,
0.01750779151916504,
0.011598322540521622,
0.026682766154408455,
0.09706830978393555,
-0.05227799713611603,
-0.0035120388492941856,
-0.07018331438302994,
0.05169927328824997,
0.05132738873362541,
0.021953076124191284,
-0.048719801008701324,
0.0725306048989296,
0.0073140403255820274,
-0.003472216660156846,
-0.0026665714103728533,
-0.06212785840034485,
-0.07525999844074249,
-0.0254192091524601,
-0.022166980430483818,
-0.013374986127018929,
-0.06495336443185806,
-0.007295419927686453,
0.04472947493195534,
-0.018878784030675888,
0.03735408931970596,
-0.0026576316449791193,
-0.06062454730272293,
-0.03241414576768875,
-0.027076654136180878,
0.07182388007640839,
-0.09535686671733856,
-0.012386063113808632,
0.07956766337156296,
-0.10214722901582718,
0.06366168707609177,
0.05873982608318329,
-0.03659149631857872,
0.00430888868868351,
-0.06346609443426132,
-0.017493126913905144,
0.0024957656860351562,
0.07187432795763016,
-0.03686670586466789,
-0.1251770257949829,
-0.015829553827643394,
-0.03409704193472862,
0.014296167530119419,
-0.025754932314157486,
0.0956420823931694,
-0.10704392194747925,
0.056662168353796005,
0.03377406671643257,
-0.024842606857419014,
-0.06546278297901154,
-0.0027347272261977196,
-0.00582378963008523,
0.0952177569270134,
0.12790943682193756,
-0.04515790566802025,
0.0106191486120224,
-0.13921058177947998,
0.008957578800618649,
0.008058879524469376,
-0.02160019241273403,
-0.012634042650461197,
-0.06264232844114304,
0.047614872455596924,
0.021882222965359688,
0.095740407705307,
0.019021091982722282,
-0.10810825228691101,
0.01666727103292942,
-0.046622443944215775,
-0.12653054296970367,
0.01724475435912609,
0.08274010568857193,
0.04867100343108177,
-0.031793877482414246,
-0.07113434374332428,
-0.02386072836816311,
-0.02310810051858425,
0.037147119641304016,
0.10642978549003601,
0.1679507941007614,
0.003981317393481731,
-0.01125642191618681,
0.05738753452897072,
-0.012899680063128471,
-0.0978865846991539,
0.032723937183618546,
-0.0002235634165117517,
0.043118372559547424,
-0.01680969074368477,
0.09816474467515945,
0.0997856855392456,
-0.14597547054290771,
0.12144294381141663,
0.002153394976630807,
-0.07354822754859924,
-0.06595592945814133,
-0.1568118780851364,
-0.02717970497906208,
-0.06409158557653427,
-0.009029069915413857,
-0.08247529715299606,
0.0017753223655745387,
0.0706082209944725,
0.02724899724125862,
-0.019205275923013687,
0.0628795474767685,
-0.03489701449871063,
-0.1328866332769394,
0.10044410824775696,
-0.009189629927277565,
0.007428196258842945,
0.0011064683785662055,
-0.03394695371389389,
0.054015085101127625,
0.06843889504671097,
0.03865443170070648,
0.07049878686666489,
0.06211840733885765,
0.01083396002650261,
-0.008665505796670914,
-0.07004254311323166,
-0.02556784264743328,
-0.007373676169663668,
-0.02783229760825634,
0.09706102311611176,
0.06364045292139053,
-0.01862998679280281,
0.011865289881825447,
0.13270190358161926,
0.0038471180014312267,
-0.08599244803190231,
-0.19056090712547302,
0.1530003845691681,
-0.03388868644833565,
-0.00710821570828557,
0.009657912887632847,
-0.08756588399410248,
-0.03400374576449394,
0.1615062952041626,
0.12899957597255707,
-0.00015131286636460572,
-0.007249491289258003,
-0.013697352260351181,
-0.0014868229627609253,
-0.018549766391515732,
0.13682180643081665,
0.009612550027668476,
0.17571179568767548,
-0.05865386500954628,
0.07143579423427582,
-0.02669353410601616,
-0.046291206032037735,
-0.04354776442050934,
0.0812482237815857,
-0.04948452115058899,
-0.0008380825165659189,
-0.11227135360240936,
0.10004474222660065,
-0.034801993519067764,
-0.20659013092517853,
0.011469175107777119,
-0.03139762207865715,
-0.1055210679769516,
-0.044582247734069824,
-0.06840085983276367,
-0.035238493233919144,
0.092164047062397,
-0.04302384704351425,
0.011393645778298378,
0.1630629003047943,
-0.020527740940451622,
-0.03721143677830696,
-0.019967373460531235,
0.09007842093706131,
0.04379956051707268,
0.04108452796936035,
0.032222483307123184,
0.05414750054478645,
0.031850554049015045,
0.048828933387994766,
-0.08626364171504974,
0.0390164852142334,
-0.006879203021526337,
0.04424136504530907,
0.03672356531023979,
0.08091283589601517,
0.021941279992461205,
0.03614871948957443,
0.0739658921957016,
-0.05691469460725784,
0.03509242460131645,
0.06011489778757095,
-0.047545015811920166,
-0.06694892793893814,
0.05617403984069824,
-0.06226692721247673,
0.12046775221824646,
0.2155269831418991,
-0.015996258705854416,
-0.009461529552936554,
-0.04705960303544998,
0.03146585449576378,
0.07349024713039398,
-0.007102471310645342,
-0.04355459660291672,
-0.14400629699230194,
-0.04021872207522392,
0.0627664178609848,
0.0417371541261673,
-0.1745276302099228,
-0.05297571420669556,
0.03249002620577812,
-0.027210645377635956,
0.03854523226618767,
0.07448194921016693,
-0.024519076570868492,
0.019616208970546722,
-0.04611283168196678,
-0.05766099691390991,
-0.0017627592897042632,
0.04358562454581261,
-0.0798141211271286,
-0.07459648698568344
] |
null | null |
transformers
|
# About this model: Topical Change Detection in Documents
This network has been fine-tuned for the task described in the paper *Topical Change Detection in Documents via Embeddings of Long Sequences* and is our best-performing base-transformer model. You can find more detailed information in our GitHub page for the paper [here](https://github.com/dennlinger/TopicalChange), or read the [paper itself](https://arxiv.org/abs/2012.03619). The weights are based on RoBERTa-base.
# Load the model
The preferred way is through pipelines
```python
from transformers import pipeline
pipe = pipeline("text-classification", model="dennlinger/roberta-cls-consec")
pipe("{First paragraph} [SEP] {Second paragraph}")
```
# Input Format
The model expects two segments that are separated with the `[SEP]` token. In our training setup, we had entire paragraphs as samples (or up to 512 tokens across two paragraphs), specifically trained on a Terms of Service data set. Note that this might lead to poor performance on "general" topics, such as news articles or Wikipedia.
# Training objective
The training task is to determine whether two text segments (paragraphs) belong to the same topical section or not. This can be utilized to create a topical segmentation of a document by consecutively predicting the "coherence" of two segments.
If you are experimenting via the Huggingface Model API, the following are interpretations of the `LABEL`s:
* `LABEL_0`: Two input segments separated by `[SEP]` do *not* belong to the same topic.
* `LABEL_1`: Two input segments separated by `[SEP]` do belong to the same topic.
# Performance
The results of this model can be found in the paper. We average over models from five different random seeds, which is why the specific results for this model might be different from the exact values in the paper.
Note that this model is *not* trained to work on classifying single texts, but only works with two (separated) inputs.
|
{}
|
text-classification
|
dennlinger/roberta-cls-consec
|
[
"transformers",
"pytorch",
"jax",
"safetensors",
"roberta",
"text-classification",
"arxiv:2012.03619",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2012.03619"
] |
[] |
TAGS
#transformers #pytorch #jax #safetensors #roberta #text-classification #arxiv-2012.03619 #autotrain_compatible #endpoints_compatible #region-us
|
# About this model: Topical Change Detection in Documents
This network has been fine-tuned for the task described in the paper *Topical Change Detection in Documents via Embeddings of Long Sequences* and is our best-performing base-transformer model. You can find more detailed information in our GitHub page for the paper here, or read the paper itself. The weights are based on RoBERTa-base.
# Load the model
The preferred way is through pipelines
# Input Format
The model expects two segments that are separated with the '[SEP]' token. In our training setup, we had entire paragraphs as samples (or up to 512 tokens across two paragraphs), specifically trained on a Terms of Service data set. Note that this might lead to poor performance on "general" topics, such as news articles or Wikipedia.
# Training objective
The training task is to determine whether two text segments (paragraphs) belong to the same topical section or not. This can be utilized to create a topical segmentation of a document by consecutively predicting the "coherence" of two segments.
If you are experimenting via the Huggingface Model API, the following are interpretations of the 'LABEL's:
* 'LABEL_0': Two input segments separated by '[SEP]' do *not* belong to the same topic.
* 'LABEL_1': Two input segments separated by '[SEP]' do belong to the same topic.
# Performance
The results of this model can be found in the paper. We average over models from five different random seeds, which is why the specific results for this model might be different from the exact values in the paper.
Note that this model is *not* trained to work on classifying single texts, but only works with two (separated) inputs.
|
[
"# About this model: Topical Change Detection in Documents\nThis network has been fine-tuned for the task described in the paper *Topical Change Detection in Documents via Embeddings of Long Sequences* and is our best-performing base-transformer model. You can find more detailed information in our GitHub page for the paper here, or read the paper itself. The weights are based on RoBERTa-base.",
"# Load the model\nThe preferred way is through pipelines",
"# Input Format\nThe model expects two segments that are separated with the '[SEP]' token. In our training setup, we had entire paragraphs as samples (or up to 512 tokens across two paragraphs), specifically trained on a Terms of Service data set. Note that this might lead to poor performance on \"general\" topics, such as news articles or Wikipedia.",
"# Training objective\nThe training task is to determine whether two text segments (paragraphs) belong to the same topical section or not. This can be utilized to create a topical segmentation of a document by consecutively predicting the \"coherence\" of two segments. \nIf you are experimenting via the Huggingface Model API, the following are interpretations of the 'LABEL's:\n* 'LABEL_0': Two input segments separated by '[SEP]' do *not* belong to the same topic.\n* 'LABEL_1': Two input segments separated by '[SEP]' do belong to the same topic.",
"# Performance\nThe results of this model can be found in the paper. We average over models from five different random seeds, which is why the specific results for this model might be different from the exact values in the paper.\n\nNote that this model is *not* trained to work on classifying single texts, but only works with two (separated) inputs."
] |
[
"TAGS\n#transformers #pytorch #jax #safetensors #roberta #text-classification #arxiv-2012.03619 #autotrain_compatible #endpoints_compatible #region-us \n",
"# About this model: Topical Change Detection in Documents\nThis network has been fine-tuned for the task described in the paper *Topical Change Detection in Documents via Embeddings of Long Sequences* and is our best-performing base-transformer model. You can find more detailed information in our GitHub page for the paper here, or read the paper itself. The weights are based on RoBERTa-base.",
"# Load the model\nThe preferred way is through pipelines",
"# Input Format\nThe model expects two segments that are separated with the '[SEP]' token. In our training setup, we had entire paragraphs as samples (or up to 512 tokens across two paragraphs), specifically trained on a Terms of Service data set. Note that this might lead to poor performance on \"general\" topics, such as news articles or Wikipedia.",
"# Training objective\nThe training task is to determine whether two text segments (paragraphs) belong to the same topical section or not. This can be utilized to create a topical segmentation of a document by consecutively predicting the \"coherence\" of two segments. \nIf you are experimenting via the Huggingface Model API, the following are interpretations of the 'LABEL's:\n* 'LABEL_0': Two input segments separated by '[SEP]' do *not* belong to the same topic.\n* 'LABEL_1': Two input segments separated by '[SEP]' do belong to the same topic.",
"# Performance\nThe results of this model can be found in the paper. We average over models from five different random seeds, which is why the specific results for this model might be different from the exact values in the paper.\n\nNote that this model is *not* trained to work on classifying single texts, but only works with two (separated) inputs."
] |
[
54,
97,
13,
84,
147,
78
] |
[
"passage: TAGS\n#transformers #pytorch #jax #safetensors #roberta #text-classification #arxiv-2012.03619 #autotrain_compatible #endpoints_compatible #region-us \n# About this model: Topical Change Detection in Documents\nThis network has been fine-tuned for the task described in the paper *Topical Change Detection in Documents via Embeddings of Long Sequences* and is our best-performing base-transformer model. You can find more detailed information in our GitHub page for the paper here, or read the paper itself. The weights are based on RoBERTa-base.# Load the model\nThe preferred way is through pipelines# Input Format\nThe model expects two segments that are separated with the '[SEP]' token. In our training setup, we had entire paragraphs as samples (or up to 512 tokens across two paragraphs), specifically trained on a Terms of Service data set. Note that this might lead to poor performance on \"general\" topics, such as news articles or Wikipedia.# Training objective\nThe training task is to determine whether two text segments (paragraphs) belong to the same topical section or not. This can be utilized to create a topical segmentation of a document by consecutively predicting the \"coherence\" of two segments. \nIf you are experimenting via the Huggingface Model API, the following are interpretations of the 'LABEL's:\n* 'LABEL_0': Two input segments separated by '[SEP]' do *not* belong to the same topic.\n* 'LABEL_1': Two input segments separated by '[SEP]' do belong to the same topic.# Performance\nThe results of this model can be found in the paper. We average over models from five different random seeds, which is why the specific results for this model might be different from the exact values in the paper.\n\nNote that this model is *not* trained to work on classifying single texts, but only works with two (separated) inputs."
] |
[
-0.0016745827160775661,
0.11451467126607895,
-0.007017507683485746,
0.029348863288760185,
0.06927727162837982,
-0.024721713736653328,
0.15653355419635773,
0.06039778143167496,
-0.09592932462692261,
0.07426249235868454,
0.008811515755951405,
0.0559367835521698,
0.034265972673892975,
0.07897483557462692,
0.03642823174595833,
-0.250629186630249,
0.08284693956375122,
-0.07030311226844788,
0.09680812060832977,
0.059685513377189636,
0.09703214466571808,
-0.0809677317738533,
0.06083177775144577,
0.02838103100657463,
-0.05148828774690628,
0.0034978538751602173,
0.011452462524175644,
-0.027951614931225777,
0.06733837723731995,
0.09484256058931351,
0.13120462000370026,
-0.0009481147862970829,
0.035682398825883865,
-0.16936777532100677,
0.015599200502038002,
0.1119953840970993,
0.005038589239120483,
0.06223270669579506,
0.13960905373096466,
-0.068273164331913,
0.10677514970302582,
-0.05031275376677513,
0.07630933076143265,
0.01897064596414566,
-0.13787822425365448,
-0.07264618575572968,
-0.12282086163759232,
0.09379571676254272,
0.12051410228013992,
0.041104815900325775,
-0.028381800279021263,
0.045138079673051834,
-0.046291302889585495,
0.05353507772088051,
0.16192051768302917,
-0.17228880524635315,
-0.04086846858263016,
0.030144724994897842,
-0.041787952184677124,
0.058523498475551605,
-0.034414805471897125,
-0.03819077089428902,
-0.02398095466196537,
0.0210846159607172,
0.04205261170864105,
-0.005873203743249178,
-0.06777547299861908,
-0.0047227768227458,
-0.1629599928855896,
-0.07405901700258255,
0.09386757016181946,
-0.0017612174851819873,
-0.07148919999599457,
-0.1530260145664215,
-0.052897367626428604,
0.0698561891913414,
-0.005947998724877834,
-0.012703986838459969,
0.027186812832951546,
0.03570707142353058,
0.0757913663983345,
-0.0530504435300827,
-0.14976702630519867,
-0.038025230169296265,
-0.09287743270397186,
0.0797131359577179,
0.017562266439199448,
0.028821244835853577,
-0.014390382915735245,
0.16062235832214355,
-0.041020408272743225,
-0.007222797255963087,
-0.01192752830684185,
-0.06145123019814491,
-0.14248481392860413,
-0.023232242092490196,
-0.0765286311507225,
-0.20107296109199524,
-0.021891897544264793,
0.11997102946043015,
-0.019226834177970886,
0.011195155791938305,
-0.028697500005364418,
0.040873706340789795,
0.09121204912662506,
0.1278744637966156,
-0.07247786968946457,
0.0702277421951294,
-0.025814030319452286,
0.02403833530843258,
0.01475178636610508,
-0.008071886375546455,
-0.02402912639081478,
0.02104972116649151,
0.02844819240272045,
0.022793881595134735,
0.07001347839832306,
0.07029690593481064,
-0.05604977160692215,
-0.021664321422576904,
0.08093845844268799,
-0.13011698424816132,
-0.0292892437428236,
-0.015558437444269657,
0.0016278979601338506,
0.008135727606713772,
0.07040058076381683,
0.022943740710616112,
-0.06881579756736755,
0.1046268418431282,
-0.05466286465525627,
-0.011682986281812191,
-0.10280115157365799,
-0.0760212242603302,
0.019880492240190506,
-0.04903623089194298,
-0.08805417269468307,
-0.04328608140349388,
-0.18814407289028168,
-0.09137038141489029,
0.03976700082421303,
-0.03609905019402504,
-0.010417260229587555,
-0.01898352988064289,
0.001539688790217042,
-0.026818694546818733,
0.0257195383310318,
0.08467972278594971,
-0.024658406153321266,
0.008204280398786068,
-0.08292724937200546,
0.016067324206233025,
0.05181904882192612,
0.007986257784068584,
-0.09565263241529465,
0.042656783014535904,
-0.18252332508563995,
0.1334574818611145,
-0.0430917963385582,
0.003978773020207882,
-0.060500361025333405,
0.04159383475780487,
-0.05769575014710426,
0.0424664169549942,
-0.038302090018987656,
0.07833380997180939,
-0.1345670521259308,
-0.03921026736497879,
0.15614859759807587,
-0.16109570860862732,
0.028500529006123543,
0.09769498556852341,
-0.06954490393400192,
0.06602029502391815,
0.11033111065626144,
0.13567502796649933,
0.051605481654405594,
-0.03594610095024109,
-0.047465983778238297,
0.020704049617052078,
-0.07636556029319763,
0.15593215823173523,
0.06575091183185577,
-0.054216645658016205,
0.03530368581414223,
0.03687915578484535,
-0.04043922945857048,
-0.05498742312192917,
0.002591879339888692,
-0.029134614393115044,
-0.0031823161989450455,
-0.005342110525816679,
-0.01271381787955761,
-0.041892241686582565,
-0.0070340512320399284,
0.04427463933825493,
-0.08622974902391434,
0.09128399938344955,
0.08445915579795837,
-0.030292976647615433,
0.03598083183169365,
-0.03794001787900925,
0.04191583767533302,
-0.054032765328884125,
-0.026371566578745842,
-0.21774736046791077,
-0.08665621280670166,
0.02662782184779644,
-0.07823154330253601,
0.10605202615261078,
0.12774991989135742,
0.000641878112219274,
0.043229326605796814,
-0.027989473193883896,
0.009273137897253036,
-0.016815224662423134,
-0.0054566264152526855,
-0.08079340308904648,
-0.146103635430336,
-0.04027802124619484,
-0.06200369447469711,
0.09451928734779358,
-0.10025647282600403,
0.019185783341526985,
0.020718064159154892,
0.08805490285158157,
0.0718381404876709,
-0.05559664964675903,
0.06410108506679535,
0.003824056126177311,
-0.0010482452344149351,
-0.025180062279105186,
0.010927090421319008,
0.028476938605308533,
-0.0304347462952137,
0.07341539114713669,
-0.15139172971248627,
-0.10742990672588348,
0.01924840174615383,
0.0034301576670259237,
-0.09325412660837173,
-0.06794366985559464,
-0.062023043632507324,
-0.031465016305446625,
-0.11364033818244934,
-0.06824911385774612,
0.1490149050951004,
0.07340555638074875,
0.06123872846364975,
-0.07765339314937592,
-0.03220575302839279,
-0.040323980152606964,
-0.07295980304479599,
-0.02484804205596447,
0.051472533494234085,
0.011521663516759872,
-0.10300613194704056,
0.05622932314872742,
-0.04735841602087021,
-0.021818585693836212,
0.10617158561944962,
0.0018922215094789863,
-0.11147116124629974,
-0.02167888544499874,
0.020365694537758827,
0.004334547556936741,
0.0019982242956757545,
-0.0009377397364005446,
0.025707468390464783,
0.06361137330532074,
0.04296452924609184,
0.04063578322529793,
-0.10411138087511063,
0.05599640682339668,
0.014864332042634487,
-0.023496676236391068,
-0.0005509693291969597,
-0.03117825835943222,
0.0026304093189537525,
0.09117569774389267,
0.015069498680531979,
0.06994620710611343,
-0.013048534281551838,
-0.06574373692274094,
-0.10413059592247009,
0.1437697410583496,
-0.05146579071879387,
-0.20215432345867157,
-0.2109754979610443,
0.0020319924224168062,
-0.05111780762672424,
0.05099949240684509,
0.030040966346859932,
-0.028792735189199448,
-0.0627620741724968,
-0.13079796731472015,
0.04798959568142891,
0.04282577335834503,
-0.0730569064617157,
-0.13911259174346924,
0.03132474794983864,
0.030197398737072945,
-0.1029101088643074,
-0.009215694852173328,
-0.0493457056581974,
-0.04603166505694389,
-0.0019166675629094243,
0.0017769199330359697,
0.03503873199224472,
0.11651783436536789,
0.06960869580507278,
-0.028304632753133774,
-0.031999412924051285,
0.16869106888771057,
-0.05438973754644394,
0.06072545424103737,
0.15663012862205505,
-0.052313558757305145,
0.10020429641008377,
0.0978950634598732,
0.0348520502448082,
-0.020157327875494957,
0.030978938564658165,
0.09469830244779587,
-0.025811178609728813,
-0.24893002212047577,
-0.06549592316150665,
-0.0025200084783136845,
-0.0021249025594443083,
0.039684396237134933,
0.013786920346319675,
-0.02853899821639061,
0.03633236512541771,
-0.08852728456258774,
0.02547726221382618,
0.011816360056400299,
0.06044311448931694,
0.07513672113418579,
-0.02639812044799328,
0.0970362201333046,
-0.0697312131524086,
-0.017694540321826935,
0.10066696256399155,
0.0035173383075743914,
0.1612343192100525,
-0.03324151039123535,
0.1174732968211174,
0.11517506092786789,
-0.008930324576795101,
0.04843292385339737,
0.035556018352508545,
-0.014497091993689537,
0.02563447691500187,
-0.05699997395277023,
-0.06627775728702545,
-0.02899412252008915,
0.0865098163485527,
-0.030542565509676933,
-0.046325188130140305,
-0.02467522583901882,
-0.02536940574645996,
0.05055301636457443,
0.279013454914093,
0.026167912408709526,
-0.15440525114536285,
-0.03602154180407524,
0.052625130861997604,
-0.07273824512958527,
-0.04953628405928612,
-0.05590837076306343,
0.056120458990335464,
-0.1583222895860672,
-0.006327130366116762,
-0.008973263204097748,
0.07852271944284439,
-0.1635136902332306,
0.004740400239825249,
-0.07690959423780441,
0.02380017377436161,
-0.04516396299004555,
0.11843333393335342,
-0.06430154293775558,
0.1312483698129654,
0.009875457733869553,
0.05297192558646202,
-0.04033416882157326,
0.04402473196387291,
0.018315475434064865,
0.03993073105812073,
0.10052285343408585,
0.010402539744973183,
-0.09488952159881592,
-0.04899469390511513,
-0.09547733515501022,
-0.0022208960726857185,
0.03707623854279518,
-0.08909472823143005,
0.09629305452108383,
-0.019965725019574165,
0.015556393191218376,
-0.018223902210593224,
-0.009529945440590382,
-0.031564146280288696,
-0.17119024693965912,
0.07096430659294128,
-0.10618907958269119,
0.0246272049844265,
-0.07519633322954178,
-0.0319170206785202,
0.012848692946135998,
0.16706743836402893,
-0.14385777711868286,
-0.08991698920726776,
-0.098550945520401,
0.013091640546917915,
0.08293807506561279,
-0.05670715868473053,
-0.015001581981778145,
-0.017534423619508743,
0.1658524125814438,
-0.006982407066971064,
-0.06662151962518692,
-0.007027371786534786,
-0.06535015255212784,
-0.17541621625423431,
-0.0691915974020958,
0.08850083500146866,
0.0818701833486557,
0.03511710464954376,
-0.00750085711479187,
0.04084925726056099,
-0.025329062715172768,
-0.08668583631515503,
0.02296280302107334,
0.19679921865463257,
0.1158064529299736,
0.10495548695325851,
-0.089899942278862,
-0.05108495429158211,
-0.08739892393350601,
-0.010967965237796307,
0.06237464025616646,
0.13815636932849884,
-0.04710471257567406,
0.13410058617591858,
0.1977311670780182,
-0.1334211230278015,
-0.1550557017326355,
-0.06513209640979767,
0.05908937379717827,
-0.009591544046998024,
0.025823956355452538,
-0.19838689267635345,
0.13891145586967468,
0.12080012261867523,
-0.020146768540143967,
-0.12058630585670471,
-0.2726302444934845,
-0.125142902135849,
0.0839368999004364,
0.00830087997019291,
0.0011098709655925632,
-0.07461252063512802,
-0.039986200630664825,
-0.041781846433877945,
0.04143244028091431,
0.08448610454797745,
-0.1302756816148758,
0.0692477598786354,
0.013000193983316422,
-0.005175557918846607,
0.04334501177072525,
-0.024043740704655647,
0.11742628365755081,
0.029170872643589973,
0.09001469612121582,
-0.04653637856245041,
-0.030554670840501785,
0.056220732629299164,
-0.07820543646812439,
0.1251583993434906,
0.04700513556599617,
0.04866623133420944,
-0.07541226595640182,
-0.06860776990652084,
-0.01833054982125759,
0.03391246497631073,
-0.021245094016194344,
-0.05712034925818443,
-0.059798162430524826,
0.047233887016773224,
0.04811052232980728,
0.002403514226898551,
0.01608811318874359,
-0.09331109374761581,
0.04334093630313873,
0.18828043341636658,
0.09432704001665115,
0.01430369820445776,
-0.1408081203699112,
-0.05125521123409271,
-0.017961444333195686,
0.03416779637336731,
-0.08634144812822342,
0.03295877203345299,
0.05594335496425629,
-0.00792715698480606,
0.14173223078250885,
0.015933794900774956,
-0.15246987342834473,
0.02274191379547119,
0.060490161180496216,
-0.17026707530021667,
-0.1277196854352951,
0.011808953247964382,
0.04900979995727539,
-0.17829521000385284,
-0.03467356786131859,
0.14661313593387604,
-0.0019633788615465164,
-0.02684580348432064,
0.026622779667377472,
0.06906352937221527,
-0.002920219674706459,
0.1020953431725502,
-0.0011832081945613027,
0.019018299877643585,
-0.043440062552690506,
0.11973603814840317,
0.14206120371818542,
-0.0838472992181778,
-0.022768452763557434,
0.02162538468837738,
-0.08346422016620636,
-0.05788128823041916,
-0.01448691263794899,
0.12461195886135101,
-0.012316022999584675,
-0.05966880917549133,
-0.04124251380562782,
-0.040810272097587585,
0.03547398000955582,
0.1582901030778885,
0.04249730333685875,
0.12665200233459473,
-0.06136873736977577,
-0.005645726807415485,
-0.00662256171926856,
0.06714049726724625,
0.08417893946170807,
0.054951537400484085,
-0.07941712439060211,
0.14507298171520233,
0.013073823414742947,
-0.0018899234710261226,
-0.014131943695247173,
-0.06766274571418762,
-0.07401055097579956,
-0.0077923997305333614,
-0.07489348948001862,
0.028786733746528625,
-0.05060049518942833,
-0.004849330056458712,
0.0268340352922678,
0.019857291132211685,
0.047024428844451904,
-0.00832049548625946,
-0.03750208392739296,
0.013860524632036686,
-0.04287826269865036,
0.11168239265680313,
-0.152967169880867,
0.031233053654432297,
0.09034733474254608,
-0.06775831431150436,
0.08183059841394424,
0.030448850244283676,
-0.003437360282987356,
0.0001352879189653322,
-0.07958994060754776,
0.04578879103064537,
-0.021705226972699165,
0.015735214576125145,
-0.025876102969050407,
-0.11466792225837708,
-0.018652983009815216,
-0.015472760424017906,
-0.022046364843845367,
0.00768979312852025,
0.055518604815006256,
-0.1173093169927597,
0.1193133145570755,
0.10076502710580826,
-0.06506600975990295,
-0.11761689186096191,
0.03595288097858429,
0.06303620338439941,
0.0676182359457016,
0.13118262588977814,
-0.03567451238632202,
0.018355010077357292,
-0.1319189965724945,
0.025773094967007637,
0.026094865053892136,
-0.008645329624414444,
0.034771114587783813,
-0.09412233531475067,
0.017598625272512436,
0.006342970300465822,
0.19333279132843018,
0.05804327502846718,
-0.028586873784661293,
0.01455605961382389,
0.010476349852979183,
0.06605485826730728,
0.03984713926911354,
-0.05890993773937225,
0.0134238600730896,
-0.030926765874028206,
-0.030705034732818604,
-0.04141227528452873,
-0.040313053876161575,
0.014717968180775642,
0.12008943408727646,
0.09943164139986038,
0.07438454777002335,
0.015075921081006527,
0.02890879102051258,
0.015176107175648212,
-0.07975814491510391,
-0.0020491662435233593,
-0.03005107305943966,
0.009127208963036537,
-0.06533586233854294,
0.09930934756994247,
0.12191882729530334,
-0.10870489478111267,
0.10708840191364288,
0.04091932252049446,
-0.030405784025788307,
-0.07439829409122467,
-0.20231890678405762,
-0.03389440104365349,
-0.012961589731276035,
-0.03372466191649437,
-0.10657552629709244,
0.03347788751125336,
0.14878010749816895,
-0.006476514041423798,
0.002126129576936364,
0.08913077414035797,
-0.07746818661689758,
-0.07629978656768799,
0.0002110631758114323,
-0.015498096123337746,
0.03245465084910393,
-0.023166663944721222,
0.00805054884403944,
0.061699818819761276,
0.07490891218185425,
0.03632067143917084,
0.06998465955257416,
0.11676707118749619,
-0.02739677205681801,
-0.049436166882514954,
-0.06493308395147324,
-0.00341446022503078,
-0.01650071144104004,
-0.020375581458210945,
0.1216103807091713,
0.0572500005364418,
-0.03430071473121643,
-0.0016052565770223737,
0.23423005640506744,
-0.04885070025920868,
-0.0447598397731781,
-0.1628791242837906,
0.2265409231185913,
-0.02475338988006115,
0.014897205866873264,
0.002125871367752552,
-0.11669006943702698,
0.02518283948302269,
0.14918434619903564,
0.10499239712953568,
0.002537396037951112,
0.00028744045994244516,
0.02318754605948925,
-0.0009370014304295182,
-0.006791003979742527,
0.05534829944372177,
-0.05037308111786842,
0.2618427574634552,
-0.07123568654060364,
0.12538160383701324,
-0.008619784377515316,
-0.007483480032533407,
-0.03535974770784378,
0.08069191873073578,
-0.03792266547679901,
0.0014606682816520333,
-0.0683860182762146,
0.10690267384052277,
-0.05431574210524559,
-0.22226108610630035,
-0.05728878825902939,
0.03098403848707676,
-0.09156562387943268,
0.009848587214946747,
-0.0015494925901293755,
-0.022795651108026505,
0.09740227460861206,
-0.019701719284057617,
-0.033205293118953705,
0.13670091331005096,
0.0009399899281561375,
-0.024691998958587646,
-0.09266091883182526,
0.1005348414182663,
-0.04202553629875183,
0.16409848630428314,
0.0158794317394495,
0.048025038093328476,
0.06620963662862778,
-0.016543639823794365,
-0.09042870253324509,
0.04636656120419502,
-0.01572011038661003,
-0.030938446521759033,
0.026857668533921242,
0.11836536973714828,
0.028186673298478127,
0.08239572495222092,
0.06350786983966827,
-0.07219289988279343,
0.0787503570318222,
-0.02091299369931221,
-0.05714254826307297,
-0.06670554727315903,
0.03769644722342491,
-0.0972808450460434,
0.12147437036037445,
0.18173368275165558,
-0.004653381649404764,
0.0269614327698946,
-0.02129095420241356,
0.01502246968448162,
0.00006882674642838538,
0.003711202647536993,
-0.00016858942399267107,
-0.11415945738554001,
0.03798791021108627,
0.06541701406240463,
0.029482269659638405,
-0.21376962959766388,
-0.08053188771009445,
0.0773753970861435,
-0.01760966144502163,
0.040919192135334015,
0.056616708636283875,
0.05294938012957573,
0.027268288657069206,
-0.04096471518278122,
-0.06664270162582397,
0.013188161887228489,
0.0875910297036171,
-0.06250003725290298,
-0.0631832629442215
] |
null | null |
transformers
|
# Bilingual English + German SQuAD2.0
We created German Squad 2.0 (**deQuAD 2.0**) and merged with [**SQuAD2.0**](https://rajpurkar.github.io/SQuAD-explorer/) into an English and German training data for question answering. The [**bert-base-multilingual-cased**](https://github.com/google-research/bert/blob/master/multilingual.md) is used to fine-tune bilingual QA downstream task.
## Details of deQuAD 2.0
[**SQuAD2.0**](https://rajpurkar.github.io/SQuAD-explorer/) was auto-translated into German. We hired professional editors to proofread the translated transcripts, correct mistakes and double check the answers to further polish the text and enhance annotation quality. The final German deQuAD dataset contains **130k** training and **11k** test samples.
## Overview
- **Language model:** bert-base-multilingual-cased
- **Language:** German, English
- **Training data:** deQuAD2.0 + SQuAD2.0 training set
- **Evaluation data:** SQuAD2.0 test set; deQuAD2.0 test set
- **Infrastructure:** 8xV100 GPU
- **Published**: July 9th, 2021
## Evaluation on English SQuAD2.0
```
HasAns_exact = 85.79622132253711
HasAns_f1 = 90.92004586077663
HasAns_total = 5928
NoAns_exact = 94.76871320437343
NoAns_f1 = 94.76871320437343
NoAns_total = 5945
exact = 90.28889076054915
f1 = 92.84713483219753
total = 11873
```
## Evaluation on German deQuAD2.0
```
HasAns_exact = 63.80526406330638
HasAns_f1 = 72.47269140789888
HasAns_total = 5813
NoAns_exact = 82.0291893792861
NoAns_f1 = 82.0291893792861
NoAns_total = 5687
exact = 72.81739130434782
f1 = 77.19858740470603
total = 11500
```
## Use Model in Pipeline
```python
from transformers import pipeline
qa_pipeline = pipeline(
"question-answering",
model="deutsche-telekom/bert-multi-english-german-squad2",
tokenizer="deutsche-telekom/bert-multi-english-german-squad2"
)
contexts = ["Die Allianz Arena ist ein Fußballstadion im Norden von München und bietet bei Bundesligaspielen 75.021 Plätze, zusammengesetzt aus 57.343 Sitzplätzen, 13.794 Stehplätzen, 1.374 Logenplätzen, 2.152 Business Seats und 966 Sponsorenplätzen. In der Allianz Arena bestreitet der FC Bayern München seit der Saison 2005/06 seine Heimspiele. Bis zum Saisonende 2017 war die Allianz Arena auch Spielstätte des TSV 1860 München.",
"Harvard is a large, highly residential research university. It operates several arts, cultural, and scientific museums, alongside the Harvard Library, which is the world's largest academic and private library system, comprising 79 individual libraries with over 18 million volumes. "]
questions = ["Wo befindet sich die Allianz Arena?",
"What is the worlds largest academic and private library system?"]
qa_pipeline(context=contexts, question=questions)
```
# Output:
```json
[{'score': 0.7290093898773193,
'start': 44,
'end': 62,
'answer': 'Norden von München'},
{'score': 0.7979822754859924,
'start': 134,
'end': 149,
'answer': 'Harvard Library'}]
```
## License - The MIT License
Copyright (c) 2021 Fang Xu, Deutsche Telekom AG
|
{"language": ["de", "en", "multilingual"], "license": "mit", "tags": ["english", "german"]}
|
question-answering
|
deutsche-telekom/bert-multi-english-german-squad2
|
[
"transformers",
"pytorch",
"safetensors",
"bert",
"question-answering",
"english",
"german",
"de",
"en",
"multilingual",
"license:mit",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de",
"en",
"multilingual"
] |
TAGS
#transformers #pytorch #safetensors #bert #question-answering #english #german #de #en #multilingual #license-mit #endpoints_compatible #has_space #region-us
|
# Bilingual English + German SQuAD2.0
We created German Squad 2.0 (deQuAD 2.0) and merged with SQuAD2.0 into an English and German training data for question answering. The bert-base-multilingual-cased is used to fine-tune bilingual QA downstream task.
## Details of deQuAD 2.0
SQuAD2.0 was auto-translated into German. We hired professional editors to proofread the translated transcripts, correct mistakes and double check the answers to further polish the text and enhance annotation quality. The final German deQuAD dataset contains 130k training and 11k test samples.
## Overview
- Language model: bert-base-multilingual-cased
- Language: German, English
- Training data: deQuAD2.0 + SQuAD2.0 training set
- Evaluation data: SQuAD2.0 test set; deQuAD2.0 test set
- Infrastructure: 8xV100 GPU
- Published: July 9th, 2021
## Evaluation on English SQuAD2.0
## Evaluation on German deQuAD2.0
## Use Model in Pipeline
# Output:
## License - The MIT License
Copyright (c) 2021 Fang Xu, Deutsche Telekom AG
|
[
"# Bilingual English + German SQuAD2.0\n\nWe created German Squad 2.0 (deQuAD 2.0) and merged with SQuAD2.0 into an English and German training data for question answering. The bert-base-multilingual-cased is used to fine-tune bilingual QA downstream task.",
"## Details of deQuAD 2.0\nSQuAD2.0 was auto-translated into German. We hired professional editors to proofread the translated transcripts, correct mistakes and double check the answers to further polish the text and enhance annotation quality. The final German deQuAD dataset contains 130k training and 11k test samples.",
"## Overview\n- Language model: bert-base-multilingual-cased \n- Language: German, English \n- Training data: deQuAD2.0 + SQuAD2.0 training set \n- Evaluation data: SQuAD2.0 test set; deQuAD2.0 test set\n- Infrastructure: 8xV100 GPU \n- Published: July 9th, 2021",
"## Evaluation on English SQuAD2.0",
"## Evaluation on German deQuAD2.0",
"## Use Model in Pipeline",
"# Output:",
"## License - The MIT License\nCopyright (c) 2021 Fang Xu, Deutsche Telekom AG"
] |
[
"TAGS\n#transformers #pytorch #safetensors #bert #question-answering #english #german #de #en #multilingual #license-mit #endpoints_compatible #has_space #region-us \n",
"# Bilingual English + German SQuAD2.0\n\nWe created German Squad 2.0 (deQuAD 2.0) and merged with SQuAD2.0 into an English and German training data for question answering. The bert-base-multilingual-cased is used to fine-tune bilingual QA downstream task.",
"## Details of deQuAD 2.0\nSQuAD2.0 was auto-translated into German. We hired professional editors to proofread the translated transcripts, correct mistakes and double check the answers to further polish the text and enhance annotation quality. The final German deQuAD dataset contains 130k training and 11k test samples.",
"## Overview\n- Language model: bert-base-multilingual-cased \n- Language: German, English \n- Training data: deQuAD2.0 + SQuAD2.0 training set \n- Evaluation data: SQuAD2.0 test set; deQuAD2.0 test set\n- Infrastructure: 8xV100 GPU \n- Published: July 9th, 2021",
"## Evaluation on English SQuAD2.0",
"## Evaluation on German deQuAD2.0",
"## Use Model in Pipeline",
"# Output:",
"## License - The MIT License\nCopyright (c) 2021 Fang Xu, Deutsche Telekom AG"
] |
[
57,
70,
79,
74,
9,
9,
6,
4,
18
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #bert #question-answering #english #german #de #en #multilingual #license-mit #endpoints_compatible #has_space #region-us \n# Bilingual English + German SQuAD2.0\n\nWe created German Squad 2.0 (deQuAD 2.0) and merged with SQuAD2.0 into an English and German training data for question answering. The bert-base-multilingual-cased is used to fine-tune bilingual QA downstream task.## Details of deQuAD 2.0\nSQuAD2.0 was auto-translated into German. We hired professional editors to proofread the translated transcripts, correct mistakes and double check the answers to further polish the text and enhance annotation quality. The final German deQuAD dataset contains 130k training and 11k test samples.## Overview\n- Language model: bert-base-multilingual-cased \n- Language: German, English \n- Training data: deQuAD2.0 + SQuAD2.0 training set \n- Evaluation data: SQuAD2.0 test set; deQuAD2.0 test set\n- Infrastructure: 8xV100 GPU \n- Published: July 9th, 2021## Evaluation on English SQuAD2.0## Evaluation on German deQuAD2.0## Use Model in Pipeline# Output:## License - The MIT License\nCopyright (c) 2021 Fang Xu, Deutsche Telekom AG"
] |
[
-0.09243157505989075,
0.14468905329704285,
-0.0026067516300827265,
0.01859891600906849,
0.08876033872365952,
0.05261389538645744,
0.18520276248455048,
0.09974652528762817,
0.02131708525121212,
0.005983793176710606,
0.032569121569395065,
-0.024132583290338516,
0.08873903006315231,
0.11918598413467407,
0.03385313227772713,
-0.1793467104434967,
0.06702887266874313,
-0.08404851704835892,
-0.0636550784111023,
0.08636743575334549,
0.1518707424402237,
-0.07227280735969543,
0.07374855130910873,
-0.01206954475492239,
-0.059752754867076874,
0.04081821069121361,
-0.0013578267535194755,
-0.03946351259946823,
0.0984511598944664,
0.06977185606956482,
0.04573175311088562,
0.01388435810804367,
0.06900865584611893,
-0.20932406187057495,
0.006135494913905859,
0.014364911243319511,
-0.02046205848455429,
0.029396267607808113,
0.07175792753696442,
0.03807617723941803,
0.04815420135855675,
-0.06835777312517166,
0.005329146981239319,
0.059322722256183624,
-0.08128474652767181,
-0.13793684542179108,
-0.1838463544845581,
0.11728934198617935,
0.08291774988174438,
0.04273245483636856,
-0.020866042003035545,
0.0377437099814415,
-0.048317570239305496,
0.029074745252728462,
-0.01662270724773407,
-0.3084257245063782,
-0.0587364137172699,
0.04011854529380798,
0.07391764968633652,
0.03857976943254471,
-0.08141990005970001,
0.02926529012620449,
0.031104186549782753,
0.033905915915966034,
-0.06299687922000885,
-0.032374463975429535,
0.10132709890604019,
-0.02455364540219307,
-0.09199754148721695,
-0.028537295758724213,
0.06839091330766678,
-0.02411830984055996,
-0.1234263926744461,
-0.1578899621963501,
0.029284857213497162,
0.01613348349928856,
0.011817583814263344,
-0.06542185693979263,
0.012106248177587986,
-0.029433446004986763,
0.0504799447953701,
-0.12140588462352753,
-0.0882798358798027,
-0.04676888883113861,
-0.037262968719005585,
0.23477427661418915,
0.07366624474525452,
0.031775109469890594,
0.020244142040610313,
0.10154608637094498,
-0.0530674122273922,
-0.08412283658981323,
-0.0569419302046299,
-0.09617660939693451,
-0.07397483289241791,
0.017366109415888786,
-0.04629247635602951,
-0.10494393110275269,
0.054448969662189484,
0.200065016746521,
-0.08195486664772034,
0.04717234894633293,
-0.03610176220536232,
0.00334936729632318,
0.024779455736279488,
0.1870177835226059,
-0.015583786182105541,
-0.13454625010490417,
-0.025192173197865486,
0.0319233275949955,
0.01175961084663868,
0.030176183208823204,
-0.08263276517391205,
-0.009375041350722313,
0.0321248434484005,
0.09417013078927994,
0.002956838347017765,
-0.013853300362825394,
-0.04701494053006172,
-0.03626514971256256,
0.12895841896533966,
-0.15468770265579224,
0.017310181632637978,
-0.011934085749089718,
0.015867343172430992,
0.06597701460123062,
-0.02646917663514614,
-0.008638991042971611,
-0.08876556158065796,
0.02107023261487484,
-0.003034224035218358,
-0.01526990532875061,
-0.09922939538955688,
-0.09820541739463806,
0.022800860926508904,
-0.026023538783192635,
-0.050592683255672455,
-0.06717747449874878,
-0.08698718994855881,
-0.06540600210428238,
0.07166294753551483,
-0.04236229509115219,
-0.017687443643808365,
-0.03818947821855545,
-0.032919954508543015,
0.061451252549886703,
-0.0440366193652153,
-0.025512685999274254,
-0.012313987128436565,
0.008839158341288567,
-0.000986873754300177,
0.013199584558606148,
-0.14700213074684143,
0.0271903108805418,
-0.062143001705408096,
0.009068144485354424,
-0.1473257690668106,
0.11177150160074234,
-0.11608845740556717,
-0.09913589060306549,
-0.11934816092252731,
-0.04333869740366936,
0.04932757094502449,
0.04686609283089638,
0.055649787187576294,
0.13795331120491028,
-0.20833154022693634,
-0.03204028308391571,
0.21037565171718597,
-0.14091932773590088,
-0.044761329889297485,
0.12099974602460861,
-0.02375202625989914,
0.01123406458646059,
0.07640286535024643,
0.027372729033231735,
0.12278509140014648,
-0.17991210520267487,
-0.05255439877510071,
-0.020955178886651993,
0.013666270300745964,
0.09887029230594635,
0.11331628262996674,
-0.047254569828510284,
0.0012066428316757083,
0.03795710206031799,
-0.17211094498634338,
0.0003326198202557862,
-0.03404970467090607,
-0.0749429389834404,
0.015299762599170208,
-0.026476779952645302,
0.12233764678239822,
-0.0334063358604908,
-0.0072634234093129635,
-0.06425260007381439,
-0.1453353464603424,
0.06554212421178818,
0.06368287652730942,
-0.030303938314318657,
0.019267931580543518,
-0.08407055586576462,
0.01625434309244156,
0.0032882345840334892,
0.02640889585018158,
-0.10932282358407974,
-0.12897861003875732,
0.043355606496334076,
-0.09216770529747009,
0.12926779687404633,
0.08227712661027908,
0.026350323110818863,
0.04212289676070213,
-0.09761719405651093,
-0.03405946120619774,
-0.14995034039020538,
-0.027656754478812218,
-0.001976920524612069,
-0.14349062740802765,
-0.026703210547566414,
-0.0617937371134758,
0.05874837562441826,
-0.053614355623722076,
-0.0011477598454803228,
0.03000553697347641,
0.07063019275665283,
0.011852703988552094,
-0.013284149579703808,
-0.11101561784744263,
0.05315534397959709,
-0.038712698966264725,
-0.0021383650600910187,
0.02206847444176674,
-0.024046555161476135,
-0.04750671982765198,
0.1656397581100464,
-0.01804116554558277,
-0.05381419137120247,
0.08611463010311127,
0.07528457045555115,
-0.05890285223722458,
-0.05213719978928566,
-0.04570848122239113,
-0.018980147317051888,
0.01372254453599453,
-0.11004438996315002,
0.189255028963089,
0.01374437939375639,
0.047466881573200226,
-0.050635792315006256,
-0.03797216713428497,
-0.02176530659198761,
0.008361238054931164,
-0.08367139101028442,
0.044668443500995636,
-0.006978233810514212,
-0.1152447983622551,
0.096394382417202,
0.1495395004749298,
0.045830078423023224,
0.18905773758888245,
-0.0423097088932991,
-0.07673314213752747,
-0.04477132484316826,
0.025675948709249496,
-0.025031443685293198,
0.16706430912017822,
0.010465082712471485,
-0.023322615772485733,
0.03981049731373787,
0.05101501941680908,
0.0442955382168293,
-0.09988854080438614,
0.06071411445736885,
-0.04572176933288574,
-0.06628535687923431,
0.0045290375128388405,
0.023702973499894142,
0.02266918681561947,
0.09068181365728378,
0.015494795516133308,
0.03489888459444046,
0.030661633238196373,
-0.044623807072639465,
-0.08400966227054596,
0.15197832882404327,
-0.14585895836353302,
-0.272041916847229,
-0.20325174927711487,
0.005556298419833183,
-0.1151188462972641,
0.0036507784388959408,
0.021524565294384956,
-0.05689162388443947,
-0.08182312548160553,
-0.035077352076768875,
0.1202881783246994,
0.02437785640358925,
-0.05060498043894768,
-0.051772866398096085,
0.010975940153002739,
0.045214392244815826,
-0.15691231191158295,
-0.004961439408361912,
0.015523575246334076,
-0.10684054344892502,
0.007917146198451519,
0.07586200535297394,
0.02699591964483261,
0.027272582054138184,
-0.0509251244366169,
-0.04693976417183876,
-0.031883761286735535,
0.1860007345676422,
-0.118025042116642,
0.053962524980306625,
0.1013176441192627,
-0.09422224760055542,
0.01695944368839264,
0.11855548620223999,
0.033774763345718384,
-0.06504632532596588,
-0.0031377002596855164,
0.03328608348965645,
-0.03194015473127365,
-0.3013095557689667,
-0.10591008514165878,
-0.05448178946971893,
-0.017898568883538246,
-0.005016998387873173,
0.03196191042661667,
-0.08460815995931625,
-0.006149930413812399,
-0.08709342032670975,
-0.029091011732816696,
0.05854632332921028,
0.043535761535167694,
0.13296982645988464,
-0.017759570851922035,
0.08792487531900406,
-0.04827287048101425,
0.004967194981873035,
0.1166013777256012,
0.10405631363391876,
0.13950541615486145,
0.05278072878718376,
-0.014112455770373344,
0.0875643864274025,
0.0559757761657238,
-0.006066907197237015,
0.04698532074689865,
0.040900904685258865,
0.021361926570534706,
-0.021003492176532745,
-0.04656665399670601,
-0.01533592864871025,
0.08145010471343994,
0.0936613380908966,
-0.04521949961781502,
-0.046295687556266785,
0.0005833414616063237,
0.08368240296840668,
0.21570536494255066,
0.07531799376010895,
-0.09324711561203003,
-0.08389238268136978,
0.02348944917321205,
-0.10265150666236877,
-0.08245207369327545,
0.03187621384859085,
0.06955179572105408,
-0.13309018313884735,
0.10243690758943558,
-0.04956924170255661,
0.06729324162006378,
-0.03895239159464836,
-0.0167393796145916,
0.05164394900202751,
0.008937385864555836,
0.024285634979605675,
0.08166898787021637,
-0.2845814526081085,
0.14096017181873322,
0.02549051307141781,
0.08531308174133301,
-0.04849901795387268,
0.07657439261674881,
0.01898796111345291,
-0.06500793993473053,
0.12758298218250275,
0.021982118487358093,
-0.08003374189138412,
0.005025883670896292,
-0.10616043955087662,
0.01442098617553711,
0.14875026047229767,
-0.018688129261136055,
0.09065571427345276,
-0.011756603606045246,
0.029392417520284653,
-0.06265565007925034,
0.0530979186296463,
-0.19172418117523193,
-0.13280728459358215,
0.042757440358400345,
-0.07148364931344986,
0.016448376700282097,
-0.07355403900146484,
-0.072112537920475,
-0.10244980454444885,
0.2251589596271515,
-0.19407017529010773,
-0.042577680200338364,
-0.100765161216259,
0.021017109975218773,
0.10039188712835312,
-0.06122816726565361,
0.079291433095932,
0.02515934407711029,
0.08275250345468521,
-0.04663597047328949,
-0.1037466898560524,
0.043536026030778885,
-0.09713450074195862,
-0.15596115589141846,
-0.021861769258975983,
0.18290722370147705,
0.09562887251377106,
0.052875012159347534,
0.02837149240076542,
0.011104569770395756,
0.0608469657599926,
-0.11622828245162964,
0.04726405069231987,
0.03229605406522751,
0.007202919572591782,
0.12024131417274475,
-0.1128823533654213,
-0.2178589254617691,
-0.04586998373270035,
-0.019875867292284966,
0.10224801301956177,
0.18338420987129211,
-0.04166119545698166,
0.16317307949066162,
0.11861483007669449,
-0.1171640083193779,
-0.3133432865142822,
-0.02089390717446804,
0.09473272413015366,
0.09878400713205338,
0.027354268357157707,
-0.1870301216840744,
0.029339149594306946,
0.05003449320793152,
-0.05896351486444473,
0.06658879667520523,
-0.1524583399295807,
-0.14247721433639526,
0.09397793561220169,
-0.00520089827477932,
0.0447997972369194,
-0.06859253346920013,
-0.05750425532460213,
-0.0485198087990284,
-0.12316876649856567,
0.10429974645376205,
-0.027914395555853844,
0.04655095562338829,
0.013702543452382088,
0.02153060957789421,
0.01370567362755537,
-0.022577540948987007,
0.14434008300304413,
0.04187699407339096,
0.0015722509706392884,
-0.029601244255900383,
-0.020413849502801895,
0.07522142678499222,
0.005193485412746668,
0.09795083105564117,
-0.045400623232126236,
0.0255485400557518,
-0.07644105702638626,
-0.01065464410930872,
-0.05631377175450325,
0.047775644809007645,
-0.07494381815195084,
-0.04111339896917343,
-0.06281048059463501,
0.12178971618413925,
-0.0026423363015055656,
0.008217592723667622,
0.07605931907892227,
-0.04168028384447098,
0.0012366761220619082,
0.16703131794929504,
0.1567622721195221,
0.10500387847423553,
-0.02439330331981182,
-0.030697528272867203,
-0.042903799563646317,
0.05929556488990784,
0.030188782140612602,
0.07851053774356842,
0.09937188774347305,
-0.023554421961307526,
0.10672183334827423,
-0.0046556005254387856,
-0.10393067449331284,
0.03802698105573654,
0.07977928221225739,
-0.10717305541038513,
-0.18659646809101105,
-0.02919928729534149,
-0.04739227518439293,
-0.027489636093378067,
0.06574398279190063,
0.21365582942962646,
-0.006630265153944492,
-0.04912718012928963,
-0.00012204015365568921,
0.07223918288946152,
-0.08474239706993103,
0.11749192327260971,
0.007485266774892807,
0.01554946694523096,
-0.0648842453956604,
0.14432643353939056,
0.04386919364333153,
-0.10292138159275055,
0.046623628586530685,
0.038299448788166046,
-0.037374455481767654,
-0.028957242146134377,
-0.04072725400328636,
0.020866932347416878,
-0.08838143199682236,
-0.09641773253679276,
-0.02152358554303646,
-0.11968190222978592,
0.00984471756964922,
-0.03094717115163803,
0.04396886005997658,
0.05682702735066414,
0.015166647732257843,
0.006391550879925489,
0.016594547778367996,
0.10148091614246368,
0.019552156329154968,
0.03536072000861168,
-0.010953470133244991,
0.02434493601322174,
-0.023321280255913734,
0.04194875434041023,
-0.03694266825914383,
-0.00018509938672650605,
-0.10010489821434021,
-0.0429871492087841,
-0.21880657970905304,
0.004550259094685316,
0.004135608673095703,
-0.01293326448649168,
0.009350469335913658,
-0.06862150877714157,
-0.000836575694847852,
0.03980422019958496,
-0.09096686542034149,
-0.03250481188297272,
-0.04567968472838402,
0.09341210126876831,
-0.114000104367733,
0.035248689353466034,
0.02673683501780033,
-0.055200718343257904,
0.11123780906200409,
0.031721025705337524,
0.001648670993745327,
0.10903443396091461,
-0.04120287671685219,
-0.01124147791415453,
-0.02061626687645912,
0.038265228271484375,
0.027711179107427597,
-0.12598846852779388,
-0.005057116504758596,
0.08140229433774948,
0.011012283153831959,
0.031124239787459373,
0.050137896090745926,
-0.04694352671504021,
0.02641734667122364,
-0.0021823628339916468,
-0.08680040389299393,
0.0019738252740353346,
0.07124488800764084,
0.08080154657363892,
0.04670285806059837,
0.14323127269744873,
-0.10834523290395737,
0.018247729167342186,
-0.10520756989717484,
0.030524225905537605,
0.016194919124245644,
-0.0004609344177879393,
-0.09901675581932068,
-0.02189726009964943,
0.040065184235572815,
0.0012265616096556187,
0.21849766373634338,
0.0006075096898712218,
0.025312628597021103,
0.05204690247774124,
-0.06422541290521622,
-0.10461241006851196,
0.08024166524410248,
0.09608713537454605,
-0.024169757962226868,
0.020070990547537804,
-0.11669406294822693,
-0.05195636302232742,
-0.09324372559785843,
-0.06145237758755684,
0.21221637725830078,
0.13455253839492798,
0.11857131868600845,
0.10392285883426666,
0.09542296826839447,
-0.032118409872055054,
-0.12349916994571686,
0.01452726311981678,
-0.04456121101975441,
0.06610887497663498,
-0.03947199508547783,
0.16128571331501007,
0.07960311323404312,
-0.2652442455291748,
0.10637419670820236,
-0.03311936557292938,
-0.04745406657457352,
-0.08046700060367584,
-0.07497010380029678,
-0.05484979972243309,
-0.083187997341156,
0.020023833960294724,
-0.1028057411313057,
0.04843124747276306,
0.0067308745346963406,
0.07496406137943268,
-0.06531546264886856,
0.11404484510421753,
-0.15552742779254913,
-0.07609862834215164,
0.10970783233642578,
0.03813362866640091,
0.09776896238327026,
0.04106571525335312,
-0.02441832795739174,
-0.024503111839294434,
0.09836401045322418,
0.07353463023900986,
0.04560700058937073,
0.004623616114258766,
0.004180015064775944,
-0.03103545680642128,
-0.031023433431982994,
0.005072899628430605,
0.021249575540423393,
0.06911759078502655,
0.19465306401252747,
0.03014770895242691,
-0.004601947031915188,
-0.005052513908594847,
0.08232198655605316,
-0.03783717751502991,
-0.08314233273267746,
-0.15363773703575134,
0.13574650883674622,
0.018753409385681152,
0.09077837318181992,
0.02556450106203556,
-0.11029750108718872,
-0.035228390246629715,
0.045263852924108505,
0.24794337153434753,
-0.049949828535318375,
-0.01739668846130371,
0.022575339302420616,
0.01362374983727932,
0.01358134113252163,
0.12547118961811066,
-0.012030926533043385,
0.2786366939544678,
-0.004474456422030926,
0.005160079337656498,
0.014757088385522366,
-0.010287520475685596,
0.02029978670179844,
0.16152271628379822,
-0.05410867556929588,
-0.0560842864215374,
-0.08295835554599762,
0.09925413131713867,
0.0298275388777256,
-0.21210257709026337,
-0.0076511637307703495,
-0.05492820590734482,
-0.14329330623149872,
-0.005354153458029032,
0.08582824468612671,
0.020024005323648453,
0.05690326169133186,
0.012875042855739594,
0.012902502901852131,
0.1934632509946823,
-0.02615009807050228,
0.010261602699756622,
-0.07065056264400482,
0.12445026636123657,
-0.03673745319247246,
0.21800968050956726,
0.027057116851210594,
0.13354656100273132,
0.06344901025295258,
0.010743087157607079,
-0.07319759577512741,
-0.013319754973053932,
0.013859366066753864,
-0.08338496834039688,
0.02511053904891014,
0.14557072520256042,
-0.020807361230254173,
0.09292834252119064,
0.0774705559015274,
0.0077879466116428375,
0.0592225156724453,
0.015355603769421577,
-0.147168830037117,
-0.09254947304725647,
0.13625049591064453,
-0.13574327528476715,
0.1358194500207901,
0.19247432053089142,
-0.006424103397876024,
0.023903630673885345,
-0.04904080554842949,
0.07578852772712708,
-0.005091620143502951,
0.12815500795841217,
0.01910128816962242,
-0.19133484363555908,
0.037729568779468536,
-0.025162555277347565,
0.012294379062950611,
-0.2316226065158844,
-0.056087676435709,
-0.038104914128780365,
0.0622076652944088,
-0.08236698061227798,
0.1206732839345932,
0.022234736010432243,
-0.02651108428835869,
-0.019621964544057846,
-0.1557619571685791,
-0.024568693712353706,
0.050955284386873245,
-0.06387258321046829,
-0.04590625315904617
] |
null | null |
transformers
|
We released the German Question Answering model fine-tuned with our own German Question Answering dataset (**deQuAD**) containing **130k** training and **11k** test QA pairs.
## Overview
- **Language model:** [electra-base-german-uncased](https://huggingface.co/german-nlp-group/electra-base-german-uncased)
- **Language:** German
- **Training data:** deQuAD2.0 training set (~42MB)
- **Evaluation data:** deQuAD2.0 test set (~4MB)
- **Infrastructure:** 8xV100 GPU
## Evaluation
We benchmarked the question answering performance on our deQuAD test data with some German language models. The fine-tuned electra-base-german-uncased model gives the best performance (Exact Match/F1).
| Model | All | HasAns | NoAns |
|-------|--------|--------|--------|
| electra-base-german-uncased | 70.97/76.18 | 67.73/78.02 | 74.29/74.29 |
| bert-base-german-cased |58.98/64.77| 49.19/60.63| 69.03/69.03|
|bert-base-german-dbmdz-uncased|63.70/68.00| 57.03/65.52| 70.51/70.51 |
|dbmdz/bert-base-german-europeana-uncased| 58.79/63.38| 52.14/61.22| 65.59/65.59|
## Use Model in Pipeline
```python
from transformers import pipeline
qa_pipeline = pipeline(
"question-answering",
model="deutsche-telekom/electra-base-de-squad2",
tokenizer="deutsche-telekom/electra-base-de-squad2"
)
contexts = ['''Die Robert Bosch GmbH ist ein im Jahr 1886 von Robert Bosch gegründetes multinationales deutsches Unternehmen.
Es ist tätig als Automobilzulieferer, Hersteller von Gebrauchsgütern und Industrie- und Gebäudetechnik und darüber hinaus
in der automatisierten Verpackungstechnik, wo Bosch den führenden Platz einnimmt. Die Robert Bosch GmbH und ihre rund 460
Tochter- und Regionalgesellschaften in mehr als 60 Ländern bilden die Bosch-Gruppe. Der Sitz der Geschäftsführung befindet
sich auf der Schillerhöhe in Gerlingen, der Firmensitz in Stuttgart. Seit dem 1. Juli 2012 ist Volkmar Denner Vorsitzender
der Geschäftsführung. Im Jahr 2015 konnte Bosch die Spitzenposition zurückgewinnen. Die Automobilsparte war im Jahr 2018
für 61 % des Konzernumsatzes von Bosch verantwortlich. Das Unternehmen hatte im Jahr 2018 in Deutschland an 85 Standorten
139.400 Mitarbeiter.''']*2
questions = ["Wer leitet die Robert Bosch GmbH?",
"Wer begründete die Robert Bosch GmbH?"]
qa_pipeline(context=contexts, question=questions)
```
## Output
```json
[{'score': 0.9537325501441956,
'start': 577,
'end': 591,
'answer': 'Volkmar Denner'},
{'score': 0.8804352879524231,
'start': 47,
'end': 59,
'answer': 'Robert Bosch'}]
```
## License - The MIT License
Copyright (c) 2021 Fang Xu, Deutsche Telekom AG
|
{"language": "de", "license": "mit", "tags": ["german"]}
|
question-answering
|
deutsche-telekom/electra-base-de-squad2
|
[
"transformers",
"pytorch",
"safetensors",
"electra",
"question-answering",
"german",
"de",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#transformers #pytorch #safetensors #electra #question-answering #german #de #license-mit #endpoints_compatible #region-us
|
We released the German Question Answering model fine-tuned with our own German Question Answering dataset (deQuAD) containing 130k training and 11k test QA pairs.
Overview
--------
* Language model: electra-base-german-uncased
* Language: German
* Training data: deQuAD2.0 training set (~42MB)
* Evaluation data: deQuAD2.0 test set (~4MB)
* Infrastructure: 8xV100 GPU
Evaluation
----------
We benchmarked the question answering performance on our deQuAD test data with some German language models. The fine-tuned electra-base-german-uncased model gives the best performance (Exact Match/F1).
Use Model in Pipeline
---------------------
Output
------
License - The MIT License
-------------------------
Copyright (c) 2021 Fang Xu, Deutsche Telekom AG
|
[] |
[
"TAGS\n#transformers #pytorch #safetensors #electra #question-answering #german #de #license-mit #endpoints_compatible #region-us \n"
] |
[
45
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #electra #question-answering #german #de #license-mit #endpoints_compatible #region-us \n"
] |
[
-0.08449117839336395,
0.03619786351919174,
-0.008217870257794857,
0.015316602773964405,
0.031202899292111397,
-0.029987851157784462,
0.13504624366760254,
0.026338232681155205,
0.17423726618289948,
0.028051884844899178,
0.16929280757904053,
0.16744767129421234,
-0.08234420418739319,
-0.027185408398509026,
-0.07236042618751526,
-0.1925390660762787,
0.09987369924783707,
0.031057382002472878,
-0.013458081521093845,
0.1035628616809845,
0.1340494006872177,
-0.11424410343170166,
0.039946433156728745,
0.01122412271797657,
-0.0016776443226262927,
0.08423969894647598,
-0.0034207936841994524,
-0.07836544513702393,
0.17893174290657043,
0.016791291534900665,
0.14835003018379211,
0.04653921723365784,
-0.037106920033693314,
-0.13529597222805023,
0.02099272422492504,
-0.07532862573862076,
-0.11186172813177109,
0.025242390111088753,
0.04198996350169182,
-0.10441122204065323,
0.07393445819616318,
0.04696216061711311,
-0.026771951466798782,
0.07330604642629623,
-0.12549002468585968,
-0.1879061609506607,
-0.09969668835401535,
0.10953617095947266,
0.06888800859451294,
0.10897140949964523,
-0.008191321976482868,
0.1591021865606308,
-0.16555871069431305,
0.059175968170166016,
0.1038971021771431,
-0.3271770477294922,
0.009439009241759777,
0.021360497921705246,
0.1096562072634697,
0.0011342300567775965,
-0.018475154414772987,
0.07661142945289612,
0.07498663663864136,
0.005879606120288372,
-0.06018770858645439,
-0.05215455964207649,
0.06609911471605301,
0.044500336050987244,
-0.08986527472734451,
-0.0976199209690094,
0.18935298919677734,
0.01704554446041584,
-0.005199788603931665,
-0.03666415810585022,
-0.04634173959493637,
0.02846437506377697,
0.04349401965737343,
-0.032312195748090744,
-0.041240137070417404,
0.033527273684740067,
0.05650767683982849,
0.027392413467168808,
-0.13504134118556976,
0.0638461783528328,
-0.1938898116350174,
0.3054439127445221,
0.035794831812381744,
0.06432893127202988,
-0.12725788354873657,
0.06476295739412308,
-0.04844662919640541,
-0.0906590074300766,
-0.02693568915128708,
-0.06216468662023544,
0.07682520151138306,
-0.004943970590829849,
-0.02561502531170845,
0.059442855417728424,
0.11333362758159637,
0.26780614256858826,
-0.037481967359781265,
-0.04095190763473511,
0.03230755776166916,
0.14803633093833923,
0.03733796626329422,
0.07658802717924118,
0.016930196434259415,
-0.018233072012662888,
-0.030932648107409477,
-0.14843033254146576,
0.0501246452331543,
0.035589490085840225,
-0.10123351961374283,
-0.04961348697543144,
-0.03500320389866829,
0.16571640968322754,
0.003498261794447899,
0.002240236382931471,
-0.06738719344139099,
0.08670944720506668,
0.06550854444503784,
-0.007396042346954346,
-0.004113343078643084,
-0.005383253563195467,
-0.0032665252219885588,
0.09622225165367126,
-0.08806221932172775,
0.02761903777718544,
-0.036454036831855774,
0.1874021738767624,
-0.08152247220277786,
-0.025584140792489052,
-0.014396908693015575,
-0.04112166166305542,
0.10686463117599487,
-0.13734745979309082,
0.08734915405511856,
-0.166870579123497,
-0.08792214840650558,
0.038447920233011246,
0.04645288735628128,
0.022572865709662437,
0.03589606657624245,
0.004663486033678055,
-0.04931071400642395,
-0.030319295823574066,
-0.06929218024015427,
-0.20284654200077057,
-0.0646907240152359,
0.09257283806800842,
-0.056061699986457825,
-0.037136103957891464,
-0.1535324603319168,
0.005668085999786854,
-0.16426996886730194,
0.012471091002225876,
-0.0817488580942154,
-0.09029132127761841,
-0.08043041825294495,
0.12306057661771774,
-0.03822065144777298,
-0.07565079629421234,
-0.018377985805273056,
0.03936528041958809,
-0.023512832820415497,
0.17439773678779602,
-0.023405998945236206,
-0.08761012554168701,
0.166891410946846,
-0.17793630063533783,
-0.17436282336711884,
0.10274496674537659,
0.0020411964505910873,
-0.0384027324616909,
0.1117967963218689,
0.1289079487323761,
-0.004774508532136679,
-0.22117629647254944,
-0.07048255950212479,
0.0683785006403923,
-0.18481197953224182,
-0.08919085562229156,
0.08881931006908417,
-0.05578437075018883,
-0.06964480131864548,
0.01609022356569767,
-0.03917849063873291,
0.0839284285902977,
-0.07181916385889053,
-0.05627477541565895,
-0.004021376837044954,
0.011990154162049294,
0.13808637857437134,
0.05389995500445366,
0.0015469040954485536,
-0.15174797177314758,
-0.04503226652741432,
0.027321714907884598,
-0.0006312865298241377,
0.0468716099858284,
-0.024236569181084633,
-0.08977081626653671,
0.060950424522161484,
-0.01962001994252205,
-0.03239764645695686,
-0.07840801030397415,
-0.12959904968738556,
-0.049443960189819336,
-0.030588895082473755,
0.04361148923635483,
0.24220120906829834,
0.08936715871095657,
-0.031103136017918587,
-0.03310704603791237,
-0.06113607436418533,
0.02071552351117134,
0.01638166420161724,
0.04586521536111832,
-0.11783791333436966,
0.030978592112660408,
-0.05690879374742508,
-0.04763154685497284,
-0.016458312049508095,
-0.02248493768274784,
0.14132730662822723,
0.144356369972229,
-0.05068005993962288,
0.09446538239717484,
-0.11415087431669235,
0.01837960258126259,
-0.02855038456618786,
0.05027719587087631,
0.07535611093044281,
-0.02093615010380745,
-0.057302314788103104,
0.13998718559741974,
-0.03599260747432709,
0.3526386320590973,
0.1743571162223816,
-0.14942024648189545,
0.01617996022105217,
-0.047440432012081146,
-0.09930670261383057,
0.04932459443807602,
-0.006348556838929653,
-0.0333174467086792,
0.05517401918768883,
0.0005895647336728871,
0.05685844272375107,
-0.05043957009911537,
-0.08051615953445435,
0.0031076334416866302,
-0.04018683731555939,
-0.08505743741989136,
0.10853615403175354,
0.09112123399972916,
-0.1777425855398178,
0.19812197983264923,
0.2848796248435974,
0.04297913983464241,
0.14785094559192657,
-0.08828800916671753,
0.010745187290012836,
-0.010334949940443039,
0.03348105773329735,
-0.04172614961862564,
0.1873289942741394,
-0.12497993558645248,
0.033145636320114136,
0.05873242765665054,
-0.0034162073861807585,
0.03145286813378334,
-0.13191285729408264,
-0.08781339973211288,
-0.025037413462996483,
-0.0029986887238919735,
-0.05334031954407692,
0.09820666909217834,
-0.024673406034708023,
0.11307808011770248,
-0.03880561515688896,
-0.1614062339067459,
0.11694308370351791,
-0.0012386173475533724,
-0.08120715618133545,
0.14668093621730804,
-0.07078756392002106,
-0.1890268474817276,
-0.0793919637799263,
0.01913013681769371,
0.001379908062517643,
0.009864247404038906,
0.10197102278470993,
-0.02342933602631092,
-0.08417192846536636,
0.07924197614192963,
0.07005417346954346,
-0.07447865605354309,
0.004227539524435997,
-0.07810327410697937,
0.03365408256649971,
-0.05203498154878616,
-0.12467575818300247,
-0.08587487041950226,
-0.051112741231918335,
-0.029033515602350235,
0.0734446793794632,
0.006523221731185913,
0.10302732139825821,
0.04194284975528717,
-0.020101461559534073,
0.021986618638038635,
-0.03933095559477806,
0.19647268950939178,
-0.1326965093612671,
0.02480299398303032,
0.1315508484840393,
-0.03652933984994888,
0.09257450699806213,
0.21690452098846436,
0.07664121687412262,
-0.038551803678274155,
-0.007501550484448671,
-0.05626419931650162,
-0.07182163000106812,
-0.23374740779399872,
-0.13680192828178406,
-0.06871707737445831,
0.10737509280443192,
-0.0029009822756052017,
0.05147821456193924,
0.11650421470403671,
0.11274152249097824,
-0.03034277632832527,
-0.2167426496744156,
-0.018172241747379303,
0.07191131263971329,
0.2735874354839325,
-0.028387166559696198,
0.07393184304237366,
-0.09720013290643692,
-0.16364651918411255,
0.09121078252792358,
0.02073257602751255,
0.0788751170039177,
0.15615597367286682,
-0.05316713824868202,
0.12717808783054352,
0.1169927567243576,
0.09034385532140732,
0.10658430308103561,
0.036918215453624725,
-0.08568599075078964,
0.006088737398386002,
-0.033836860209703445,
0.024956800043582916,
0.036162346601486206,
0.03197839483618736,
-0.0598074235022068,
-0.050005026161670685,
-0.23023362457752228,
0.1256221979856491,
0.06346922367811203,
0.08429889380931854,
-0.11092027276754379,
-0.03049689717590809,
0.028758667409420013,
-0.053011227399110794,
0.007254129741340876,
0.07291910797357559,
0.009861516766250134,
-0.12440475821495056,
0.06559494137763977,
-0.050231724977493286,
0.08062749356031418,
0.08496885001659393,
0.06048336625099182,
-0.07259849458932877,
-0.09892823547124863,
0.02468939870595932,
0.11368844658136368,
-0.25714942812919617,
0.32608914375305176,
0.005278809927403927,
-0.0007103790994733572,
-0.10447368770837784,
-0.03424438089132309,
-0.029032733291387558,
0.18045395612716675,
0.1999531239271164,
0.026870256289839745,
-0.12460807710886002,
-0.03851170837879181,
0.06410349160432816,
0.06098753213882446,
0.08128770440816879,
-0.00513752456754446,
-0.036796074360609055,
-0.014207971282303333,
0.014778693206608295,
-0.023484550416469574,
0.03133423626422882,
-0.04366413876414299,
-0.051624588668346405,
0.00876185018569231,
0.05053544044494629,
0.10367089509963989,
-0.04080453887581825,
-0.0317375622689724,
-0.11849340796470642,
0.0863066241145134,
-0.1778440624475479,
-0.056397512555122375,
-0.1083676740527153,
-0.17334194481372833,
0.0622115321457386,
-0.08583050221204758,
0.07698261737823486,
-0.010293276980519295,
-0.0405937023460865,
-0.13756857812404633,
-0.0932910367846489,
0.13488532602787018,
-0.16807229816913605,
-0.026376957073807716,
-0.030680643394589424,
0.21629434823989868,
0.024962004274129868,
-0.007845213636755943,
0.022751912474632263,
0.05192415043711662,
-0.07833370566368103,
-0.15082959830760956,
0.022295771166682243,
-0.17803031206130981,
0.09358684718608856,
0.023341061547398567,
-0.01751071773469448,
-0.08819493651390076,
0.04368993267416954,
-0.0008307584212161601,
0.1138378158211708,
0.32582077383995056,
-0.05647404491901398,
0.09060996770858765,
0.2479713410139084,
0.03654708340764046,
-0.28763633966445923,
-0.0789729580283165,
-0.17076478898525238,
-0.04934248700737953,
0.0455346442759037,
-0.04625057801604271,
0.132655069231987,
0.1199609637260437,
-0.11134327948093414,
0.08354498445987701,
-0.21695543825626373,
-0.06050818786025047,
0.1730404496192932,
-0.02240956574678421,
0.4011838436126709,
-0.04343419522047043,
-0.06156420335173607,
0.024713365361094475,
-0.2289353609085083,
0.09194625914096832,
-0.008630015887320042,
0.0005382683011703193,
0.012005574069917202,
0.03642973303794861,
0.012671094387769699,
-0.08212772011756897,
0.17300236225128174,
0.047409966588020325,
0.06128010153770447,
-0.07794661074876785,
-0.09257107228040695,
0.10191543400287628,
0.05958307161927223,
0.09437965601682663,
-0.10477197170257568,
0.05072292685508728,
-0.08382450044155121,
-0.028075112029910088,
-0.10711660236120224,
0.2054027020931244,
-0.011465385556221008,
-0.1204937994480133,
-0.07949641346931458,
0.024138640612363815,
-0.05057147145271301,
-0.011285929009318352,
0.22833527624607086,
-0.018408220261335373,
0.12805427610874176,
-0.010390238836407661,
0.13079734146595,
-0.11932920664548874,
-0.06495630741119385,
-0.04137829318642616,
-0.07269510626792908,
0.07104272395372391,
0.03851931169629097,
0.031024573370814323,
0.1910228729248047,
-0.018072571605443954,
0.056177154183387756,
0.04135001078248024,
0.011115928180515766,
-0.0033113439567387104,
0.09630070626735687,
-0.1962616741657257,
-0.10921469330787659,
-0.02173692174255848,
0.01620294526219368,
0.12559513747692108,
0.17865201830863953,
0.08396749198436737,
0.018187059089541435,
0.024375129491090775,
0.014602398499846458,
-0.0494111105799675,
-0.07283952087163925,
-0.0037146296817809343,
0.05399017035961151,
0.03394876793026924,
-0.08052225410938263,
0.05191466584801674,
-0.05507339909672737,
-0.24332678318023682,
-0.013509385287761688,
0.04204743355512619,
-0.09820996224880219,
-0.08785582333803177,
-0.12725751101970673,
0.028846461325883865,
-0.281197190284729,
-0.09034255892038345,
-0.08349770307540894,
-0.12753303349018097,
0.05077900364995003,
0.22350728511810303,
0.08888468146324158,
0.0829259604215622,
0.02668951451778412,
-0.02135659195482731,
0.012146703898906708,
0.02228313498198986,
-0.04444063454866409,
0.008753188885748386,
-0.06212209537625313,
-0.02589324116706848,
-0.01185398455709219,
0.10471295565366745,
-0.08558648079633713,
-0.030930940061807632,
-0.13636156916618347,
0.0515383742749691,
-0.13807909190654755,
-0.0929935872554779,
-0.13324233889579773,
-0.050486475229263306,
-0.013000587932765484,
-0.18714956939220428,
-0.031836435198783875,
-0.023657236248254776,
-0.12351967394351959,
0.09991874545812607,
0.09922535717487335,
0.06173375993967056,
-0.12787729501724243,
-0.033045537769794464,
0.1276441365480423,
-0.0012887563789263368,
0.13751190900802612,
0.09607908874750137,
-0.05763980746269226,
0.10542996227741241,
-0.1575189232826233,
-0.07657071202993393,
0.09294353425502777,
-0.020697249099612236,
0.02851545251905918,
0.04856489226222038,
-0.007673727814108133,
0.12078510969877243,
0.03714262321591377,
0.09153584390878677,
-0.17613297700881958,
-0.11840585619211197,
0.07019032537937164,
0.10197321325540543,
-0.173399418592453,
-0.038625895977020264,
-0.08958911150693893,
0.09133347868919373,
-0.08411964774131775,
0.19607001543045044,
-0.012022154405713081,
0.06519272923469543,
-0.0011044950224459171,
0.029714658856391907,
0.009974724613130093,
-0.15308868885040283,
-0.07115556299686432,
-0.04095477610826492,
-0.06728498637676239,
-0.057026054710149765,
0.2907242774963379,
-0.020690379664301872,
-0.03367001935839653,
0.1057986244559288,
0.07508925348520279,
-0.015535194426774979,
0.022376177832484245,
0.23760896921157837,
-0.0072365449741482735,
-0.002191342879086733,
-0.200851172208786,
0.021611997857689857,
-0.08115433901548386,
-0.22470423579216003,
0.1704956293106079,
0.15987753868103027,
0.03277089074254036,
0.003407515352591872,
0.06373753398656845,
0.028977464884519577,
-0.12055931240320206,
-0.2518075406551361,
0.07368951290845871,
-0.006908039562404156,
0.04319362714886665,
0.1711706668138504,
0.21236470341682434,
-0.04949649050831795,
0.019036205485463142,
0.003630060004070401,
0.04898155480623245,
-0.11970573663711548,
-0.10177087038755417,
-0.035712048411369324,
-0.0823884904384613,
0.08819780498743057,
-0.039155907928943634,
-0.002427390543743968,
0.08400917053222656,
0.08812201023101807,
-0.10226389765739441,
0.084551140666008,
-0.029749920591711998,
-0.0338025838136673,
0.00823525246232748,
0.028800155967473984,
0.09792204201221466,
-0.021002529188990593,
-0.03689350560307503,
-0.14844220876693726,
-0.06176169961690903,
-0.05161278694868088,
0.018167788162827492,
-0.0840829536318779,
-0.06856835633516312,
-0.1308242380619049,
-0.05519426614046097,
-0.05752367898821831,
0.1089816465973854,
0.025857195258140564,
0.11647361516952515,
-0.023034268990159035,
0.052650291472673416,
0.040040772408246994,
0.16237017512321472,
-0.02008880488574505,
-0.14702647924423218,
0.007042957004159689,
0.11729816347360611,
0.033764224499464035,
0.15634940564632416,
0.01797589659690857,
-0.00012346016592346132,
0.005743515677750111,
0.13719961047172546,
0.21991263329982758,
0.006515479646623135,
0.0606624074280262,
-0.02647886425256729,
0.04182923212647438,
0.09127090126276016,
-0.03305479511618614,
0.0024215979501605034,
0.35355067253112793,
-0.11438901722431183,
0.02125133015215397,
-0.04095030203461647,
0.04879133030772209,
-0.06164661794900894,
0.07482627779245377,
-0.03205372020602226,
-0.11307785660028458,
-0.08886747807264328,
0.13318616151809692,
-0.09945812076330185,
0.08100372552871704,
0.046215444803237915,
-0.07305940985679626,
-0.04500306770205498,
0.0027980420272797346,
0.1875554621219635,
0.08160264790058136,
0.08753904700279236,
-0.04895215481519699,
-0.09478376060724258,
0.012459388002753258,
0.0571553036570549,
-0.21037094295024872,
-0.11519044637680054,
0.083213672041893,
0.11871509253978729,
0.1297985315322876,
-0.0060134888626635075,
0.09221028536558151,
0.1166037991642952,
0.0015411688946187496,
-0.04730931669473648,
0.0516628623008728,
0.12459230422973633,
-0.12666091322898865,
-0.07950317859649658,
-0.12713319063186646,
0.04788367077708244,
-0.035283662378787994,
0.04793655127286911,
-0.18867701292037964,
0.07401373982429504,
0.017877331003546715,
-0.08072745054960251,
-0.09331395477056503,
0.17251019179821014,
-0.026053577661514282,
0.03766971454024315,
-0.00553891621530056,
0.0008273208513855934,
-0.05785874277353287,
-0.051068708300590515,
0.06535610556602478,
0.11406536400318146,
-0.12849341332912445,
-0.01744122803211212,
-0.020140239968895912,
0.003489695256575942,
0.08323796838521957,
-0.037192441523075104,
-0.0689101368188858,
-0.04931982606649399,
-0.07023435086011887,
0.07154727727174759,
-0.12782707810401917,
-0.04243786633014679,
0.07014364004135132,
0.00031758585828356445,
0.052680425345897675,
-0.08273996412754059,
0.013755901716649532,
-0.009362386539578438,
-0.03888668119907379,
-0.04068055376410484
] |
null | null |
transformers
|
# mT5-small-sum-de-en-v1
This is a bilingual summarization model for English and German. It is based on the multilingual T5 model [google/mt5-small](https://huggingface.co/google/mt5-small).
[](https://www.welove.ai/)
This model is provided by the [One Conversation](https://www.welove.ai/)
team of [Deutsche Telekom AG](https://www.telekom.com/).
## Training
The training was conducted with the following hyperparameters:
- base model: [google/mt5-small](https://huggingface.co/google/mt5-small)
- source_prefix: `"summarize: "`
- batch size: 3
- max_source_length: 800
- max_target_length: 96
- warmup_ratio: 0.3
- number of train epochs: 10
- gradient accumulation steps: 2
- learning rate: 5e-5
## Datasets and Preprocessing
The datasets were preprocessed as follows:
The summary was tokenized with the [google/mt5-small](https://huggingface.co/google/mt5-small) tokenizer. Then only the records with no more than 94 summary tokens were selected.
The MLSUM dataset has a special characteristic. In the text, the summary is often included completely as one or more sentences. These have been removed from the texts. The reason is that we do not want to train a model that ultimately extracts only sentences as a summary.
This model is trained on the following datasets:
| Name | Language | Size | License
|------|----------|------|--------
| [CNN Daily - Train](https://github.com/abisee/cnn-dailymail) | en | 218,223 | The license is unclear. The data comes from CNN and Daily Mail. We assume that it may only be used for research purposes and not commercially.
| [Extreme Summarization (XSum) - Train](https://github.com/EdinburghNLP/XSum) | en | 204,005 | The license is unclear. The data comes from BBC. We assume that it may only be used for research purposes and not commercially.
| [wiki_lingua English](https://github.com/esdurmus/Wikilingua) | en | 130,331 | [Creative Commons CC BY-NC-SA 3.0 License](https://www.wikihow.com/wikiHow:Terms-of-Use)
| [wiki_lingua German](https://github.com/esdurmus/Wikilingua) | de | 48,390 | [Creative Commons CC BY-NC-SA 3.0 License](https://www.wikihow.com/wikiHow:Terms-of-Use)
| [MLSUM German - Train](https://github.com/ThomasScialom/MLSUM) | de | 218,043 | Usage of dataset is restricted to non-commercial research purposes only. Copyright belongs to the original copyright holders (see [here](https://github.com/ThomasScialom/MLSUM#mlsum)).
| [SwissText 2019 - Train](https://www.swisstext.org/2019/shared-task/german-text-summarization-challenge.html) | de | 84,564 | The license is unclear. The data was published in the [German Text Summarization Challenge](https://www.swisstext.org/2019/shared-task/german-text-summarization-challenge.html). We assume that they may be used for research purposes and not commercially.
| Language | Size
|------|------
| German | 350,997
| English | 552,559
| Total | 903,556
## Evaluation on MLSUM German Test Set (no beams)
| Model | rouge1 | rouge2 | rougeL | rougeLsum
|-------|--------|--------|--------|----------
| [ml6team/mt5-small-german-finetune-mlsum](https://huggingface.co/ml6team/mt5-small-german-finetune-mlsum) | 18.3607 | 5.3604 | 14.5456 | 16.1946
| **deutsche-telekom/mT5-small-sum-de-en-01 (this)** | **21.7336** | **7.2614** | **17.1323** | **19.3977**
## Evaluation on CNN Daily English Test Set (no beams)
| Model | rouge1 | rouge2 | rougeL | rougeLsum
|-------|--------|--------|--------|----------
| [sshleifer/distilbart-xsum-12-6](https://huggingface.co/sshleifer/distilbart-xsum-12-6) | 26.7664 | 8.8243 | 18.3703 | 23.2614
| [facebook/bart-large-xsum](https://huggingface.co/facebook/bart-large-xsum) | 28.5374 | 9.8565 | 19.4829 | 24.7364
| [mrm8488/t5-base-finetuned-summarize-news](https://huggingface.co/mrm8488/t5-base-finetuned-summarize-news) | 37.576 | 14.7389 | 24.0254 | 34.4634
| **deutsche-telekom/mT5-small-sum-de-en-01 (this)** | **37.6339** | **16.5317** | **27.1418** | **34.9951**
## Evaluation on Extreme Summarization (XSum) English Test Set (no beams)
| Model | rouge1 | rouge2 | rougeL | rougeLsum
|-------|--------|--------|--------|----------
| [mrm8488/t5-base-finetuned-summarize-news](https://huggingface.co/mrm8488/t5-base-finetuned-summarize-news) | 18.6204 | 3.535 | 12.3997 | 15.2111
| [facebook/bart-large-xsum](https://huggingface.co/facebook/bart-large-xsum) | 28.5374 | 9.8565 | 19.4829 | 24.7364
| deutsche-telekom/mT5-small-sum-de-en-01 (this) | 32.3416 | 10.6191 | 25.3799 | 25.3908
| [sshleifer/distilbart-xsum-12-6](https://huggingface.co/sshleifer/distilbart-xsum-12-6) | 44.2553 ♣ | 21.4289 ♣ | 36.2639 ♣ | 36.2696 ♣
♣: These values seem to be unusually high. It could be that the test set was used in the training data.
## License
Copyright (c) 2021 Philip May, Deutsche Telekom AG
This work is licensed under the [Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0)](https://creativecommons.org/licenses/by-nc-sa/3.0/) license.
|
{"language": ["de", "en", "multilingual"], "license": "cc-by-nc-sa-4.0", "tags": ["summarization"], "datasets": ["cnn_dailymail", "xsum", "wiki_lingua", "mlsum", "swiss_text_2019"]}
|
summarization
|
deutsche-telekom/mt5-small-sum-de-en-v1
|
[
"transformers",
"pytorch",
"safetensors",
"mt5",
"text2text-generation",
"summarization",
"de",
"en",
"multilingual",
"dataset:cnn_dailymail",
"dataset:xsum",
"dataset:wiki_lingua",
"dataset:mlsum",
"dataset:swiss_text_2019",
"license:cc-by-nc-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de",
"en",
"multilingual"
] |
TAGS
#transformers #pytorch #safetensors #mt5 #text2text-generation #summarization #de #en #multilingual #dataset-cnn_dailymail #dataset-xsum #dataset-wiki_lingua #dataset-mlsum #dataset-swiss_text_2019 #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
mT5-small-sum-de-en-v1
======================
This is a bilingual summarization model for English and German. It is based on the multilingual T5 model google/mt5-small.

----------------------------------------------
Evaluation on CNN Daily English Test Set (no beams)
---------------------------------------------------
Evaluation on Extreme Summarization (XSum) English Test Set (no beams)
----------------------------------------------------------------------
♣: These values seem to be unusually high. It could be that the test set was used in the training data.
License
-------
Copyright (c) 2021 Philip May, Deutsche Telekom AG
This work is licensed under the Attribution-NonCommercial-ShareAlike 3.0 Unported (CC BY-NC-SA 3.0) license.
|
[] |
[
"TAGS\n#transformers #pytorch #safetensors #mt5 #text2text-generation #summarization #de #en #multilingual #dataset-cnn_dailymail #dataset-xsum #dataset-wiki_lingua #dataset-mlsum #dataset-swiss_text_2019 #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
119
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #mt5 #text2text-generation #summarization #de #en #multilingual #dataset-cnn_dailymail #dataset-xsum #dataset-wiki_lingua #dataset-mlsum #dataset-swiss_text_2019 #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
-0.10636130720376968,
0.11589985340833664,
-0.00610405532643199,
0.03729720786213875,
0.07756499201059341,
0.0033713136799633503,
0.16346332430839539,
0.05410437285900116,
-0.018856950104236603,
0.004264896735548973,
0.1445147544145584,
0.11049109697341919,
0.003419174812734127,
0.21456988155841827,
-0.09749846905469894,
-0.1512708216905594,
0.050273969769477844,
-0.010325231589376926,
-0.046589601784944534,
0.14142416417598724,
0.11532604694366455,
-0.07031048089265823,
0.08250012993812561,
-0.06874861568212509,
-0.019219815731048584,
0.011368307285010815,
0.02571016363799572,
-0.10089776664972305,
0.09033951908349991,
0.023601900786161423,
0.03928796574473381,
0.11743101477622986,
-0.0019962245132774115,
-0.1439061015844345,
0.036610908806324005,
0.008325262926518917,
-0.07137474417686462,
0.03495118021965027,
0.04596997797489166,
-0.08700823783874512,
0.10566367954015732,
-0.15462273359298706,
-0.055938124656677246,
0.050767384469509125,
-0.07093070447444916,
0.013409889303147793,
-0.09131032228469849,
0.004191539250314236,
-0.0022311739157885313,
0.07981057465076447,
-0.023712173104286194,
0.16250982880592346,
-0.07142870873212814,
0.10091841965913773,
0.04338082671165466,
-0.25560253858566284,
-0.02918507345020771,
0.08532853424549103,
0.008844575844705105,
0.12432776391506195,
-0.010060136206448078,
0.11257128417491913,
0.05943698808550835,
0.0024971701204776764,
0.028941631317138672,
-0.04737703502178192,
-0.07735060155391693,
0.011989389546215534,
-0.11657864600419998,
-0.023635637015104294,
0.31645500659942627,
-0.016619659960269928,
0.013172508217394352,
-0.06514209508895874,
-0.04699168726801872,
0.012625430710613728,
-0.031846173107624054,
0.029076995328068733,
-0.007489179726690054,
-0.012326838448643684,
-0.03276352211833,
-0.01852177083492279,
-0.10178614407777786,
0.03719724342226982,
-0.16301114857196808,
0.08551912754774094,
-0.0067000784911215305,
0.04508163407444954,
-0.1031019389629364,
0.013899672776460648,
0.0526859387755394,
-0.10250112414360046,
0.03154881298542023,
-0.09505607187747955,
0.043603770434856415,
0.014302002266049385,
-0.0061275833286345005,
-0.053553611040115356,
0.14533467590808868,
0.02040547877550125,
-0.08722729235887527,
-0.0020973291248083115,
-0.09001648426055908,
0.079890675842762,
0.06817225366830826,
-0.024666734039783478,
-0.08426415175199509,
-0.13081228733062744,
0.025398485362529755,
0.006284178700298071,
0.05965367332100868,
-0.00931144505739212,
-0.0467812642455101,
0.010900917463004589,
-0.006012612022459507,
0.10445999354124069,
0.044347092509269714,
0.1140281930565834,
-0.028073934838175774,
-0.0024600352626293898,
0.0663432851433754,
-0.08946769684553146,
-0.025133775547146797,
0.03921122848987579,
-0.03630070388317108,
0.15470463037490845,
-0.02917996048927307,
0.01174528431147337,
-0.1014515683054924,
0.08450927585363388,
-0.07003678381443024,
0.006514784414321184,
-0.003582332283258438,
-0.08943887799978256,
0.07062631100416183,
-0.065791554749012,
-0.04526901990175247,
-0.10475310683250427,
-0.14862053096294403,
-0.01470787264406681,
-0.03026418201625347,
0.01745476759970188,
-0.026267997920513153,
-0.10373593121767044,
-0.04430009424686432,
0.03215166553854942,
-0.06047865375876427,
0.035440437495708466,
-0.09456199407577515,
0.06540054082870483,
-0.11423026770353317,
0.06546933203935623,
-0.12346846610307693,
0.02259506843984127,
-0.10827744007110596,
-0.035565026104450226,
-0.025638055056333542,
0.11547203361988068,
-0.12210728228092194,
0.033638667315244675,
-0.06079469248652458,
-0.041843026876449585,
-0.003496611025184393,
0.03628439083695412,
-0.015356353484094143,
0.2701972424983978,
-0.25024157762527466,
-0.08458787947893143,
0.21886174380779266,
-0.1467123180627823,
-0.13768231868743896,
0.14717307686805725,
-0.014470459893345833,
-0.021614067256450653,
0.13221129775047302,
0.1672702431678772,
-0.025285325944423676,
-0.07433073967695236,
-0.08435501903295517,
0.0807977020740509,
-0.03419582545757294,
-0.06805092096328735,
0.07375232130289078,
0.03805353492498398,
-0.10934290289878845,
0.033587370067834854,
0.008487315848469734,
0.013597819954156876,
-0.10396981984376907,
-0.06874392181634903,
-0.005396184977144003,
-0.05086251720786095,
0.11491109430789948,
-0.008515761233866215,
0.00860003475099802,
-0.10357558727264404,
0.0017048458103090525,
-0.04482357203960419,
0.04226153343915939,
-0.01483883336186409,
0.017301738262176514,
-0.0634344145655632,
0.04943929240107536,
-0.02190123125910759,
0.037637725472450256,
-0.08817717432975769,
-0.10313501954078674,
-0.019049249589443207,
0.12852156162261963,
0.033304613083601,
0.005624392535537481,
0.043831825256347656,
0.004030407406389713,
-0.0615806020796299,
0.0012238365598022938,
0.12060080468654633,
-0.010653635486960411,
-0.02661607787013054,
-0.18397757411003113,
0.07904334366321564,
-0.06562277674674988,
0.1312568187713623,
-0.10940615087747574,
0.016526062041521072,
0.06111372262239456,
0.09896574914455414,
-0.014216240495443344,
0.050421927124261856,
0.03669245168566704,
0.05698598548769951,
-0.028790602460503578,
0.01637100614607334,
0.07919862866401672,
-0.009485634043812752,
-0.11476834118366241,
0.20529936254024506,
-0.05716303363442421,
0.20027977228164673,
0.1860555112361908,
-0.09732232242822647,
0.04748891666531563,
-0.07740169763565063,
-0.02678968571126461,
0.019385386258363724,
-0.027722755447030067,
0.014636939391493797,
0.0036276374012231827,
0.010544847697019577,
0.12671034038066864,
-0.09757769107818604,
-0.020493648946285248,
0.019065432250499725,
-0.02141125313937664,
-0.09312625974416733,
0.1401945799589157,
0.04427039995789528,
-0.1689123660326004,
0.15771684050559998,
0.17903932929039001,
0.06833088397979736,
0.21907088160514832,
-0.03636985272169113,
-0.000679517921525985,
0.02413814701139927,
-0.005191056523472071,
-0.04939357936382294,
0.010986891575157642,
-0.09559894353151321,
-0.0013439306057989597,
0.06960152834653854,
0.01980776898562908,
0.06896721571683884,
-0.07130758464336395,
-0.0612875372171402,
-0.06235816702246666,
-0.04993826895952225,
-0.08795229345560074,
0.06535772234201431,
0.01842350699007511,
0.1724674254655838,
-0.07156119495630264,
-0.032386671751737595,
0.0657932460308075,
-0.02784566767513752,
-0.08842208236455917,
0.19202680885791779,
-0.16470366716384888,
-0.3407112956047058,
-0.05796995386481285,
-0.07955162227153778,
-0.13360299170017242,
-0.01683926023542881,
0.06997492909431458,
-0.109965018928051,
-0.020335732027888298,
-0.08544564992189407,
0.04141681268811226,
-0.012516725808382034,
0.007907028310000896,
-0.025612516328692436,
0.03599828481674194,
-0.040028613060712814,
-0.1367727518081665,
-0.019305165857076645,
-0.04142862185835838,
-0.06566285341978073,
0.14056266844272614,
-0.04983941465616226,
0.0971110388636589,
0.08862900733947754,
-0.01585613377392292,
0.025200972333550453,
-0.03265608102083206,
0.12190055102109909,
-0.06801354140043259,
-0.003964426461607218,
0.14653588831424713,
0.020844457671046257,
0.052836623042821884,
0.0966317281126976,
0.02835783176124096,
-0.045409101992845535,
0.022533142939209938,
0.0172201469540596,
-0.04570972919464111,
-0.2917734682559967,
-0.2091819792985916,
-0.05330174043774605,
0.10059716552495956,
-0.020990939810872078,
0.028243692591786385,
-0.015297398902475834,
0.05317256227135658,
-0.05499991774559021,
-0.10418712347745895,
0.033064670860767365,
0.01437156368046999,
0.13695864379405975,
-0.043231748044490814,
0.12716704607009888,
-0.06854164600372314,
-0.06481758505105972,
0.11866341531276703,
0.028282247483730316,
0.06280522793531418,
-0.007889742963016033,
-0.003861860604956746,
0.08556727319955826,
0.044900476932525635,
0.0721256285905838,
0.11040892452001572,
0.03935300186276436,
-0.013849289156496525,
-0.01327244471758604,
-0.0763152465224266,
-0.04881969466805458,
-0.019834905862808228,
-0.015134323388338089,
-0.07917208969593048,
-0.0719369575381279,
0.04170874506235123,
0.15219047665596008,
0.09908831119537354,
0.062486421316862106,
-0.1964743286371231,
0.006033485755324364,
0.04534449055790901,
0.039548758417367935,
-0.047806140035390854,
0.08524122834205627,
0.10421409457921982,
-0.06269606947898865,
0.174247607588768,
-0.00033049710327759385,
0.10333351045846939,
-0.07986243814229965,
0.037059806287288666,
-0.0926785096526146,
-0.02647962048649788,
-0.022840384393930435,
0.08656638860702515,
-0.28466013073921204,
0.2585087716579437,
0.03179440274834633,
-0.010498033836483955,
-0.09136124700307846,
-0.04249033331871033,
0.01624491438269615,
0.1779704988002777,
0.16555488109588623,
0.04350521042943001,
0.0013715322129428387,
0.006369142327457666,
-0.10545627027750015,
0.042828962206840515,
0.08560658246278763,
0.04706212878227234,
-0.004108746536076069,
0.03940483555197716,
-0.023283367976546288,
-0.029073696583509445,
-0.06973907351493835,
-0.1354253888130188,
-0.1521165370941162,
0.04319193214178085,
0.1266283541917801,
0.007837804034352303,
-0.01613999903202057,
-0.09165562689304352,
-0.1564566195011139,
0.1727863848209381,
-0.024313313886523247,
-0.08608189970254898,
-0.11686228215694427,
-0.0216212198138237,
0.06977440416812897,
-0.04349854215979576,
-0.013543408364057541,
0.004071515519171953,
0.043695129454135895,
-0.09873495995998383,
-0.12373913824558258,
0.05652831867337227,
-0.11571725457906723,
-0.06539963185787201,
-0.033710796386003494,
0.19933106005191803,
0.014849300496280193,
0.006366895046085119,
0.003944593947380781,
-0.00205421494320035,
-0.0021465530153363943,
-0.11701840162277222,
0.03290748596191406,
0.0010930377757176757,
0.07319992780685425,
0.14294786751270294,
-0.07965957373380661,
-0.22967736423015594,
-0.037675246596336365,
-0.04199255257844925,
0.20591957867145538,
0.2464929223060608,
-0.04721830040216446,
0.15441389381885529,
0.21521833539009094,
-0.050504326820373535,
-0.35344135761260986,
-0.05541357770562172,
-0.09314267337322235,
-0.03257099166512489,
-0.06886100769042969,
-0.11549149453639984,
0.06436597555875778,
0.0582447350025177,
-0.015347626060247421,
0.03230718895792961,
-0.29205718636512756,
-0.0742131918668747,
0.046575289219617844,
-0.004738932475447655,
0.26114770770072937,
-0.08819451183080673,
-0.06674829870462418,
-0.06025886535644531,
-0.050899438560009,
0.20402273535728455,
-0.20154762268066406,
0.0398598276078701,
0.0005762876244261861,
0.023312343284487724,
0.004341920837759972,
-0.052575357258319855,
0.11799228191375732,
0.08172348886728287,
0.026798443868756294,
-0.04873800277709961,
-0.02281210757791996,
0.1825338751077652,
0.035387665033340454,
0.07398537546396255,
-0.13716790080070496,
0.05165993794798851,
-0.0741443857550621,
-0.026066284626722336,
-0.07869461923837662,
0.07577712833881378,
-0.03976084291934967,
-0.04893408343195915,
-0.0025021471083164215,
0.026552727445960045,
0.04090750962495804,
0.005633444990962744,
0.1419089436531067,
-0.05206281319260597,
0.06481246650218964,
0.1426209658384323,
0.17151908576488495,
-0.12023826688528061,
0.09924647957086563,
-0.0552738793194294,
-0.06227961182594299,
0.03531121462583542,
-0.10135150700807571,
0.028710942715406418,
0.12043439596891403,
-0.020032741129398346,
0.04373526945710182,
0.04909847676753998,
0.019384421408176422,
-0.00019966962281614542,
0.10949595272541046,
-0.21499869227409363,
-0.07460567355155945,
-0.03418778255581856,
-0.016004735603928566,
0.03440827876329422,
0.12812082469463348,
0.19168098270893097,
0.009189683943986893,
-0.006989543326199055,
-0.04666979983448982,
0.05059708282351494,
-0.033877719193696976,
0.14285558462142944,
-0.0010847313096746802,
0.03274984657764435,
-0.16496892273426056,
0.10718692094087601,
0.00259627983905375,
-0.16422277688980103,
-0.012194122187793255,
0.1398707628250122,
-0.1670800745487213,
-0.12153147161006927,
-0.02388451248407364,
0.09493757039308548,
-0.061020467430353165,
-0.07853424549102783,
-0.03458782285451889,
-0.12696172297000885,
0.07109507918357849,
0.16614438593387604,
0.04613107442855835,
0.042392171919345856,
-0.0025257812812924385,
-0.12988203763961792,
-0.01271496806293726,
0.07672388106584549,
0.07550517469644547,
-0.015180564485490322,
-0.033919453620910645,
0.05717293173074722,
-0.031013689935207367,
0.13225597143173218,
-0.06514671444892883,
-0.012776042334735394,
-0.05692462995648384,
0.015590667724609375,
-0.08368096500635147,
0.012395378202199936,
-0.06982467323541641,
-0.012858904898166656,
-0.06899288296699524,
-0.06769479811191559,
-0.05182156339287758,
-0.04054904356598854,
-0.07316615432500839,
0.03719508275389671,
-0.0031995640601962805,
0.09108541160821915,
-0.10526156425476074,
-0.027449481189250946,
0.019238125532865524,
-0.021673202514648438,
0.13929665088653564,
0.07932490110397339,
-0.10397785156965256,
0.08145643025636673,
-0.20320144295692444,
-0.05580216273665428,
0.08852780610322952,
0.040791403502225876,
0.025484906509518623,
0.056858621537685394,
0.028793925419449806,
0.10805033892393112,
-0.0049559795297682285,
0.07336801290512085,
0.03630745783448219,
-0.07307730615139008,
0.07465241849422455,
-0.05277104675769806,
-0.06390612572431564,
-0.032929811626672745,
0.002483370015397668,
0.08342262357473373,
0.020217986777424812,
0.14421908557415009,
-0.08754422515630722,
0.02533901110291481,
-0.08090715855360031,
0.027579108253121376,
0.00200470769777894,
-0.20053869485855103,
-0.1099621057510376,
-0.019746728241443634,
0.07580989599227905,
-0.03312036022543907,
0.2699242830276489,
0.018051959574222565,
-0.055450789630413055,
0.03197643533349037,
0.015376917086541653,
-0.05223994702100754,
0.030772527679800987,
0.18355798721313477,
0.016819162294268608,
-0.00472025852650404,
-0.07476133108139038,
-0.021814823150634766,
0.029839441180229187,
0.04533056542277336,
0.1443735510110855,
0.13187481462955475,
0.11056486517190933,
0.06176789477467537,
0.026263613253831863,
-0.020961234346032143,
-0.004794775042682886,
-0.06226411834359169,
-0.059245213866233826,
0.047145672142505646,
-0.008994266390800476,
0.06872550398111343,
0.10410840809345245,
-0.03563648462295532,
0.020017649978399277,
0.01153620332479477,
-0.042224690318107605,
-0.12696464359760284,
-0.1019899919629097,
-0.08918255567550659,
-0.06498578935861588,
0.0016156723722815514,
-0.1351214051246643,
0.023898962885141373,
0.04591309651732445,
0.12783941626548767,
-0.028843972831964493,
0.03275653347373009,
-0.013929137028753757,
-0.05775543302297592,
0.1171988919377327,
-0.015266992151737213,
0.032974522560834885,
-0.027331743389368057,
-0.045061737298965454,
-0.030562153086066246,
0.012818465940654278,
-0.02656169980764389,
0.06032450869679451,
0.030852628871798515,
0.04878208786249161,
-0.11009065806865692,
-0.0884060487151146,
-0.047186706215143204,
0.020500548183918,
0.06751905381679535,
0.13014356791973114,
0.04886789619922638,
-0.04785751551389694,
0.059525731950998306,
0.1611582636833191,
-0.007270554080605507,
-0.16858364641666412,
-0.054787032306194305,
0.13168905675411224,
-0.0006648413254879415,
0.10607650876045227,
0.01241021417081356,
-0.028585009276866913,
-0.019616106525063515,
0.20976029336452484,
0.40103358030319214,
-0.06206895411014557,
0.012586159631609917,
-0.06737560033798218,
0.025951573625206947,
0.08549486100673676,
0.0577852837741375,
0.05029473826289177,
0.2043374478816986,
-0.04930143058300018,
0.012596163898706436,
-0.08341644704341888,
0.011392614804208279,
-0.0674893707036972,
0.08790190517902374,
0.014135575853288174,
-0.06531954556703568,
-0.03090784139931202,
0.13074463605880737,
-0.1313926875591278,
0.020038871094584465,
-0.07828450202941895,
-0.14430679380893707,
-0.08626288175582886,
-0.03140230476856232,
0.055605676025152206,
0.07208652049303055,
0.025923265144228935,
-0.004254330415278673,
-0.044649139046669006,
-0.007693412713706493,
0.004564871545881033,
-0.15228572487831116,
-0.03204705938696861,
0.018421785905957222,
-0.06930103898048401,
0.08524661511182785,
0.010950491763651371,
0.09474970400333405,
0.08648838847875595,
-0.016683947294950485,
-0.04932112991809845,
0.0714695155620575,
0.01436723954975605,
0.029507650062441826,
0.04112615808844566,
0.015202873386442661,
-0.011113389395177364,
0.06038793921470642,
0.06849249452352524,
-0.05754341557621956,
0.04714782536029816,
0.008437374606728554,
-0.08662260323762894,
-0.03001244179904461,
0.10373080521821976,
-0.03106769360601902,
0.06854422390460968,
0.08404197543859482,
-0.0005248128436505795,
0.03454168885946274,
-0.06806278228759766,
0.017461955547332764,
-0.02282000333070755,
-0.0495942048728466,
0.0001910142309498042,
-0.14488337934017181,
-0.04996858909726143,
0.03283875063061714,
0.056419942528009415,
-0.21137666702270508,
0.05388525128364563,
-0.12542159855365753,
0.04687537997961044,
-0.11673128604888916,
0.05868500843644142,
0.08789742738008499,
-0.02951870672404766,
-0.010462646372616291,
-0.03088945895433426,
0.07109492272138596,
0.07856720685958862,
-0.04824569821357727,
-0.04419521242380142
] |
null | null |
transformers
|
# mT5-small-sum-de-mit-v1
This is a German summarization model. It is based on the multilingual T5 model [google/mt5-small](https://huggingface.co/google/mt5-small). The special characteristic of this model is that, unlike many other models, it is licensed under a permissive open source license (MIT). Among other things, this license allows commercial use.
[](https://www.welove.ai/)
This model is provided by the [One Conversation](https://www.welove.ai/)
team of [Deutsche Telekom AG](https://www.telekom.com/).
## Training
The training was conducted with the following hyperparameters:
- base model: [google/mt5-small](https://huggingface.co/google/mt5-small)
- source_prefix: `"summarize: "`
- batch size: 3 (6)
- max_source_length: 800
- max_target_length: 96
- warmup_ratio: 0.3
- number of train epochs: 10
- gradient accumulation steps: 2
- learning rate: 5e-5
## Datasets and Preprocessing
The datasets were preprocessed as follows:
The summary was tokenized with the [google/mt5-small](https://huggingface.co/google/mt5-small) tokenizer. Then only the records with no more than 94 summary tokens were selected.
This model is trained on the following dataset:
| Name | Language | Size | License
|------|----------|------|--------
| [SwissText 2019 - Train](https://www.swisstext.org/2019/shared-task/german-text-summarization-challenge.html) | de | 84,564 | Concrete license is unclear. The data was published in the [German Text Summarization Challenge](https://www.swisstext.org/2019/shared-task/german-text-summarization-challenge.html).
We have permission to use the Swisstext dataset and release the resulting summarization model under MIT license (see [permission-declaration-swisstext.pdf](https://huggingface.co/deutsche-telekom/mt5-small-sum-de-mit-v1/resolve/main/permission-declaration-swisstext.pdf)).
## Evaluation on MLSUM German Test Set (no beams)
| Model | rouge1 | rouge2 | rougeL | rougeLsum
|-------|--------|--------|--------|----------
| deutsche-telekom/mt5-small-sum-de-mit-v1 (this) | 16.8023 | 3.5531 | 12.6884 | 14.7624
| [ml6team/mt5-small-german-finetune-mlsum](https://huggingface.co/ml6team/mt5-small-german-finetune-mlsum) | 18.3607 | 5.3604 | 14.5456 | 16.1946
| **[deutsche-telekom/mt5-small-sum-de-en-01](https://huggingface.co/deutsche-telekom/mt5-small-sum-de-en-v1)** | **21.7336** | **7.2614** | **17.1323** | **19.3977**
## License
Copyright (c) 2021 Philip May, Deutsche Telekom AG
Licensed under the MIT License (the "License"); you may not use this work except in compliance with the License. You may obtain a copy of the License by reviewing the file [LICENSE](https://huggingface.co/deutsche-telekom/mt5-small-sum-de-mit-v1/blob/main/LICENSE) in the repository.
|
{"language": ["de"], "license": "mit", "tags": ["summarization"], "datasets": ["swiss_text_2019"]}
|
summarization
|
deutsche-telekom/mt5-small-sum-de-mit-v1
|
[
"transformers",
"pytorch",
"safetensors",
"mt5",
"text2text-generation",
"summarization",
"de",
"dataset:swiss_text_2019",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#transformers #pytorch #safetensors #mt5 #text2text-generation #summarization #de #dataset-swiss_text_2019 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
mT5-small-sum-de-mit-v1
=======================
This is a German summarization model. It is based on the multilingual T5 model google/mt5-small. The special characteristic of this model is that, unlike many other models, it is licensed under a permissive open source license (MIT). Among other things, this license allows commercial use.

* max\_source\_length: 800
* max\_target\_length: 96
* warmup\_ratio: 0.3
* number of train epochs: 10
* gradient accumulation steps: 2
* learning rate: 5e-5
Datasets and Preprocessing
--------------------------
The datasets were preprocessed as follows:
The summary was tokenized with the google/mt5-small tokenizer. Then only the records with no more than 94 summary tokens were selected.
This model is trained on the following dataset:
We have permission to use the Swisstext dataset and release the resulting summarization model under MIT license (see URL).
Evaluation on MLSUM German Test Set (no beams)
----------------------------------------------
License
-------
Copyright (c) 2021 Philip May, Deutsche Telekom AG
Licensed under the MIT License (the "License"); you may not use this work except in compliance with the License. You may obtain a copy of the License by reviewing the file LICENSE in the repository.
|
[] |
[
"TAGS\n#transformers #pytorch #safetensors #mt5 #text2text-generation #summarization #de #dataset-swiss_text_2019 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
76
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #mt5 #text2text-generation #summarization #de #dataset-swiss_text_2019 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
-0.07624642550945282,
0.09162940084934235,
-0.0043693725019693375,
0.026084382086992264,
0.12746384739875793,
-0.03691547363996506,
0.15284350514411926,
0.07757667452096939,
0.006687919609248638,
-0.0005486745503731072,
0.16404178738594055,
0.18524853885173798,
-0.0110512999817729,
0.19991305470466614,
-0.12584008276462555,
-0.1922912895679474,
0.0696270540356636,
0.005586906336247921,
0.023390181362628937,
0.12035886943340302,
0.13543255627155304,
-0.08844811469316483,
0.10019917786121368,
-0.0470099076628685,
-0.04603620991110802,
0.03658256679773331,
0.052279770374298096,
-0.11395637691020966,
0.10465333610773087,
0.04232633858919144,
0.05149545148015022,
0.11097395420074463,
-0.0156454648822546,
-0.15439888834953308,
0.022326385602355003,
0.006493805442005396,
-0.08589194715023041,
0.05333978682756424,
0.11682331562042236,
-0.06939192116260529,
0.09904398024082184,
-0.049957990646362305,
-0.03619806095957756,
0.06089618057012558,
-0.0831104964017868,
0.006623920984566212,
-0.08599056303501129,
0.013944672420620918,
0.046267252415418625,
0.08874168246984482,
0.0017624612664803863,
0.16025248169898987,
-0.08493568003177643,
0.11146295070648193,
0.08565724641084671,
-0.3353753387928009,
-0.009146010503172874,
0.061987437307834625,
0.04600977152585983,
0.08962543308734894,
-0.027577146887779236,
0.08550815284252167,
0.06327159702777863,
-0.0008390989969484508,
0.05482713505625725,
-0.051154524087905884,
-0.11852382123470306,
0.017918923869729042,
-0.1037793755531311,
-0.048641812056303024,
0.2833029329776764,
-0.024390846490859985,
0.02281234785914421,
-0.07922695577144623,
-0.08172126859426498,
-0.007111156824976206,
-0.01407593209296465,
0.016404565423727036,
0.003659656271338463,
0.039129145443439484,
0.005607732106000185,
0.008536087349057198,
-0.13253113627433777,
0.014192703180015087,
-0.17080561816692352,
0.03646860271692276,
-0.008036697283387184,
0.056446682661771774,
-0.14339253306388855,
0.035249900072813034,
0.0009436321561224759,
-0.11708922684192657,
0.023940905928611755,
-0.10220201313495636,
0.06789599359035492,
-0.013445223681628704,
-0.03148173540830612,
-0.11774946749210358,
0.12589162588119507,
0.12336039543151855,
-0.041058652102947235,
0.02964315377175808,
-0.10664309561252594,
0.11635494232177734,
0.022410843521356583,
0.06576426327228546,
-0.022465772926807404,
-0.10028038173913956,
0.03596463054418564,
-0.07064631581306458,
0.06311791390180588,
-0.020876821130514145,
-0.11215746402740479,
-0.006504423916339874,
0.02616543509066105,
0.11942005157470703,
0.0065354518592357635,
0.09300021827220917,
-0.03495131433010101,
0.024392083287239075,
0.10374214500188828,
-0.07819660007953644,
-0.016855647787451744,
0.014278697781264782,
-0.00805787742137909,
0.0727546215057373,
-0.018235931172966957,
0.014983566477894783,
-0.10601236671209335,
0.13427181541919708,
-0.06798388808965683,
-0.014542190358042717,
-0.02512410841882229,
-0.08470272272825241,
0.07145152240991592,
-0.08731851726770401,
-0.0005459252861328423,
-0.1515875905752182,
-0.16946321725845337,
0.006268623284995556,
0.016379233449697495,
-0.006232290528714657,
0.002325104083865881,
-0.054868005216121674,
-0.03357461839914322,
0.0380689911544323,
-0.08714012801647186,
0.003218985628336668,
-0.08929875493049622,
0.06911264359951019,
-0.11999344825744629,
0.04474320262670517,
-0.15854987502098083,
0.04057278856635094,
-0.16488365828990936,
-0.014061585068702698,
-0.039552219212055206,
0.061647821217775345,
-0.09397756308317184,
0.10728311538696289,
-0.04539681226015091,
-0.04332077503204346,
-0.03316846489906311,
0.0197274349629879,
0.0023520649410784245,
0.21636363863945007,
-0.183791846036911,
-0.06676267087459564,
0.21169033646583557,
-0.18104524910449982,
-0.1883622705936432,
0.12834545969963074,
-0.009174712002277374,
-0.00022965436801314354,
0.11634828895330429,
0.2199433296918869,
0.005276726093143225,
-0.06516934931278229,
-0.005796435289084911,
0.11194679886102676,
-0.09659664332866669,
-0.12230885028839111,
0.05540304630994797,
0.0007586773717775941,
-0.11992961168289185,
0.04439423605799675,
0.020609082654118538,
0.017968107014894485,
-0.07590050250291824,
-0.06523074954748154,
-0.025830361992120743,
-0.03550412505865097,
0.09967104345560074,
0.00800270400941372,
0.05374082177877426,
-0.12015707790851593,
0.014443224295973778,
-0.038974009454250336,
-0.00023118071840144694,
-0.01132509671151638,
0.001353442668914795,
-0.023871343582868576,
0.09504116326570511,
-0.0021184999495744705,
0.015209685079753399,
-0.11582473665475845,
-0.08355371654033661,
-0.009023587219417095,
0.09818460047245026,
-0.008284945040941238,
0.050038695335388184,
0.05901957303285599,
0.028073256835341454,
-0.02806101366877556,
-0.02030625380575657,
0.1704205423593521,
0.009688230231404305,
-0.046614501625299454,
-0.12898221611976624,
0.057578194886446,
-0.061209261417388916,
0.03683846443891525,
-0.08078040182590485,
0.005763854831457138,
0.03339453786611557,
0.09935283660888672,
-0.0020363829098641872,
0.07003383338451385,
-0.01586628518998623,
0.028809111565351486,
-0.0561818853020668,
0.016186639666557312,
0.10967607796192169,
0.014683142304420471,
-0.06634171307086945,
0.22422023117542267,
-0.11130774021148682,
0.2928282618522644,
0.19624283909797668,
-0.18717782199382782,
0.03704556077718735,
-0.11629243195056915,
-0.035328250378370285,
0.028161441907286644,
-0.0034789566416293383,
0.02167746052145958,
-0.017177622765302658,
0.0027944205794483423,
0.16933642327785492,
-0.10741277784109116,
-0.03701918572187424,
0.017775489017367363,
-0.04774139076471329,
-0.08127543330192566,
0.10092715919017792,
0.09618344157934189,
-0.2504684329032898,
0.19162794947624207,
0.24022717773914337,
0.07597395032644272,
0.20959632098674774,
-0.04118823632597923,
-0.0019291460048407316,
0.0666314885020256,
0.0361027829349041,
-0.04559087008237839,
-0.006207953207194805,
-0.11342285573482513,
0.005008844193071127,
0.08131857961416245,
0.026685213670134544,
0.07280361652374268,
-0.10431373119354248,
-0.04028011113405228,
-0.042047590017318726,
-0.032733529806137085,
-0.06768004596233368,
0.0773676186800003,
0.021069392561912537,
0.1542624682188034,
-0.049220651388168335,
-0.05963175371289253,
0.10556360334157944,
-0.015295222401618958,
-0.11385130882263184,
0.20839163661003113,
-0.1487307846546173,
-0.3064987063407898,
-0.12490332126617432,
-0.08507829904556274,
-0.09828044474124908,
0.033945657312870026,
0.11257221549749374,
-0.0576200895011425,
-0.05749554559588432,
-0.056032098829746246,
-0.023323746398091316,
-0.023059621453285217,
0.0009498083963990211,
-0.058921944350004196,
0.058704473078250885,
-0.0372585691511631,
-0.13989031314849854,
-0.05718091130256653,
-0.03263743966817856,
-0.04175257682800293,
0.11384475231170654,
-0.058908045291900635,
0.0961773693561554,
0.10282192379236221,
-0.05009061098098755,
0.02260035276412964,
-0.0671011209487915,
0.11858417838811874,
-0.03522542864084244,
-0.004353719763457775,
0.2058640420436859,
-0.022463759407401085,
0.08381157368421555,
0.14171740412712097,
0.006246519275009632,
-0.05910821631550789,
0.039837609976530075,
-0.04117589443922043,
-0.0719389095902443,
-0.27824389934539795,
-0.18614943325519562,
-0.0863722562789917,
0.1227983608841896,
0.04287685826420784,
0.04377490282058716,
0.06570880860090256,
0.09879125654697418,
-0.04552627354860306,
-0.08801104128360748,
0.05866985768079758,
0.06529456377029419,
0.15830698609352112,
-0.026065200567245483,
0.16017700731754303,
-0.08629652112722397,
-0.0825178250670433,
0.10867486149072647,
0.0008611291414126754,
0.09955206513404846,
0.04559928923845291,
-0.0018005907768383622,
0.0900571197271347,
0.0688176229596138,
0.09606156498193741,
0.16759543120861053,
0.0392221137881279,
-0.012933243997395039,
-0.03530935198068619,
-0.07652261108160019,
-0.06720660626888275,
0.020281093195080757,
-0.07336654514074326,
-0.10810427367687225,
-0.11036896705627441,
-0.010863837786018848,
0.13510656356811523,
0.08917947113513947,
0.05875973775982857,
-0.23246502876281738,
-0.0026318393647670746,
0.06015022471547127,
0.014534992165863514,
-0.06496725976467133,
0.11016926914453506,
0.06379355490207672,
-0.1059347540140152,
0.13459721207618713,
-0.033886902034282684,
0.12229509651660919,
0.007171644829213619,
0.06956720352172852,
-0.054604701697826385,
-0.05590818449854851,
0.0025093466974794865,
0.12396325916051865,
-0.31865012645721436,
0.26305466890335083,
0.006994869560003281,
-0.02169218845665455,
-0.112168088555336,
-0.014733001589775085,
0.015714317560195923,
0.16463491320610046,
0.1684403121471405,
0.02180004119873047,
-0.1354835331439972,
-0.024773817509412766,
-0.03496051952242851,
0.027422433719038963,
0.0946371778845787,
0.059007443487644196,
-0.03348369151353836,
-0.030360806733369827,
-0.028396230190992355,
-0.022001920267939568,
-0.03425181657075882,
-0.060384757816791534,
-0.15365508198738098,
0.02202434465289116,
0.0778837576508522,
0.07374672591686249,
-0.024198736995458603,
-0.040163248777389526,
-0.13689124584197998,
0.179198756814003,
-0.05377965420484543,
-0.08153656870126724,
-0.10976395756006241,
-0.08103081583976746,
0.026713579893112183,
-0.06014632061123848,
0.059477321803569794,
-0.047017622739076614,
0.046269383281469345,
-0.10219887644052505,
-0.19286374747753143,
0.10383409261703491,
-0.14041836559772491,
-0.05344776436686516,
-0.036657653748989105,
0.20044143497943878,
-0.07121779769659042,
-0.017809070646762848,
0.051186732947826385,
0.020726444199681282,
-0.05694209411740303,
-0.1263013631105423,
-0.007776910904794931,
-0.045614637434482574,
0.10269675403833389,
0.06680823862552643,
-0.10954004526138306,
-0.1948137730360031,
-0.005695798434317112,
-0.02807227149605751,
0.25159192085266113,
0.1959780752658844,
-0.07701626420021057,
0.19403888285160065,
0.21340353786945343,
-0.06994812190532684,
-0.3769930303096771,
-0.06335082650184631,
-0.11815574020147324,
-0.06257858872413635,
-0.011123806238174438,
-0.07306049764156342,
0.07297549396753311,
0.04641442745923996,
-0.049109406769275665,
0.06278856098651886,
-0.24679848551750183,
-0.08721958845853806,
0.11267121881246567,
0.02748880721628666,
0.3052624762058258,
-0.09806610643863678,
-0.08831115812063217,
-0.07996945083141327,
-0.12058822065591812,
0.1730773001909256,
-0.15686601400375366,
0.03984925150871277,
0.011453337967395782,
-0.010665930807590485,
0.006092838943004608,
-0.08646038919687271,
0.08886698633432388,
0.025351563468575478,
0.060812197625637054,
-0.09050987660884857,
0.0015482127200812101,
0.13926270604133606,
0.015033763833343983,
0.09226106107234955,
-0.14064650237560272,
0.07089392095804214,
-0.05499938875436783,
-0.026584327220916748,
-0.06568355858325958,
0.09201289713382721,
-0.008428391069173813,
-0.07840965688228607,
-0.009448192082345486,
-0.0012288230936974287,
0.04024916887283325,
0.007723755203187466,
0.18205201625823975,
-0.04265061765909195,
0.12005314975976944,
0.20895512402057648,
0.15520521998405457,
-0.1723887026309967,
0.09697502851486206,
-0.05579094961285591,
-0.08183646202087402,
0.04942021146416664,
-0.06239726394414902,
0.07199551910161972,
0.12925712764263153,
-0.0437745675444603,
0.05533774942159653,
0.07852175831794739,
0.055178750306367874,
-0.005451054312288761,
0.14061279594898224,
-0.23234155774116516,
-0.016584426164627075,
-0.05480999872088432,
0.02084365300834179,
0.0832996666431427,
0.11509348452091217,
0.1510498821735382,
-0.005586095619946718,
0.0014294997090473771,
-0.038307808339595795,
0.04182762652635574,
-0.05200131610035896,
0.1027836948633194,
-0.01747007854282856,
0.03260726481676102,
-0.15268242359161377,
0.10963910818099976,
0.004068184178322554,
-0.1406915932893753,
-0.006516613066196442,
0.12196234613656998,
-0.16871432960033417,
-0.1291317641735077,
-0.02310819923877716,
0.1195063367486,
-0.1151195615530014,
-0.09569939970970154,
-0.05518336221575737,
-0.14976626634597778,
0.061450887471437454,
0.1707635372877121,
0.06198539584875107,
0.05358939245343208,
-0.003280861768871546,
-0.10017070919275284,
-0.031054040417075157,
0.048933856189250946,
0.02108292654156685,
0.022768501192331314,
-0.08196936547756195,
0.06456277519464493,
-0.03886757418513298,
0.09685800969600677,
-0.07831232994794846,
-0.023113474249839783,
-0.10332479327917099,
0.011457277461886406,
-0.1198810487985611,
0.0031540084164589643,
-0.08149625360965729,
-0.02655138447880745,
-0.05938711017370224,
-0.02804923616349697,
-0.05975399166345596,
-0.024458397179841995,
-0.09589751064777374,
0.03495243936777115,
-0.022301465272903442,
0.06868795305490494,
-0.12995262444019318,
-0.02813684567809105,
0.04150394722819328,
-0.013837972655892372,
0.11688105016946793,
0.07880418002605438,
-0.09484732151031494,
0.09216835349798203,
-0.21163704991340637,
-0.11136148124933243,
0.10560113191604614,
0.027107303962111473,
0.014833373948931694,
0.06656591594219208,
0.03451098874211311,
0.1450267881155014,
-0.028559695929288864,
0.050830960273742676,
0.0022994026076048613,
-0.11408109962940216,
0.03520866110920906,
-0.02874966524541378,
-0.10692348331212997,
-0.03232945129275322,
-0.03212299942970276,
0.04868843033909798,
-0.008417691104114056,
0.17288248240947723,
-0.09440770745277405,
0.03687247261404991,
-0.08533891290426254,
0.015024038963019848,
0.010388423688709736,
-0.21207492053508759,
-0.1655595600605011,
-0.06297062337398529,
0.022557595744729042,
-0.030158216133713722,
0.2918246388435364,
0.07611124217510223,
-0.03981848806142807,
0.06063751131296158,
0.04981457069516182,
0.009692697785794735,
0.033893246203660965,
0.2517411410808563,
0.01008590403944254,
-0.018495453521609306,
-0.13012737035751343,
0.028879854828119278,
0.011434790678322315,
-0.024738451465964317,
0.17531472444534302,
0.14132818579673767,
-0.00916143599897623,
0.06198308616876602,
0.0303876344114542,
0.014018204994499683,
-0.08921496570110321,
-0.11953948438167572,
0.004801514092832804,
0.07831491529941559,
0.0044464850798249245,
0.04202749952673912,
0.15887710452079773,
-0.06858217716217041,
0.02407895401120186,
-0.016452837735414505,
-0.031596969813108444,
-0.16313882172107697,
-0.12238216400146484,
-0.09303107112646103,
-0.0956343561410904,
0.016704369336366653,
-0.10207537561655045,
0.0454053059220314,
0.04978732019662857,
0.10568146407604218,
-0.07177768647670746,
0.05830496549606323,
-0.05745352432131767,
-0.06017989665269852,
0.07615117728710175,
-0.018857400864362717,
0.056854378432035446,
-0.01619526371359825,
-0.06158449500799179,
-0.053868699818849564,
-0.03603782504796982,
-0.04181978106498718,
0.05130438879132271,
0.02656641975045204,
0.019886400550603867,
-0.10151916742324829,
-0.06973642110824585,
-0.04135884717106819,
0.0575542226433754,
0.030495982617139816,
0.14297159016132355,
0.015940334647893906,
-0.024561509490013123,
0.07530935108661652,
0.18732015788555145,
-0.048544712364673615,
-0.15524359047412872,
-0.03322405368089676,
0.24651971459388733,
0.04983223229646683,
0.1297672688961029,
0.013581971637904644,
-0.008597910404205322,
-0.017891502007842064,
0.25076597929000854,
0.3669176697731018,
-0.04784829542040825,
0.01370537281036377,
-0.0573812797665596,
0.02804391458630562,
0.1022467240691185,
0.12064208835363388,
0.05495356768369675,
0.24884262681007385,
-0.06841285526752472,
0.036986641585826874,
-0.04841790348291397,
0.022048020735383034,
-0.09469839930534363,
0.12827233970165253,
0.0031150609720498323,
-0.0835658460855484,
-0.026963001117110252,
0.12084414064884186,
-0.1319948136806488,
0.08030101656913757,
-0.0794529989361763,
-0.14915990829467773,
-0.05305064469575882,
-0.005672093480825424,
0.10723288357257843,
0.04315828159451485,
0.04538071155548096,
-0.010385595262050629,
-0.04294522479176521,
0.00789861660450697,
0.017525287345051765,
-0.1823565661907196,
-0.03425564989447594,
-0.0021203269716352224,
-0.05165214091539383,
0.10482943058013916,
0.003944898955523968,
0.062229763716459274,
0.08728572726249695,
0.02893630787730217,
-0.045836448669433594,
0.08891946077346802,
0.024046780541539192,
-0.024271631613373756,
0.05249622464179993,
-0.04640684276819229,
-0.013513846322894096,
0.018685828894376755,
0.038483113050460815,
-0.13192026317119598,
0.05478920787572861,
-0.00590663542971015,
-0.09564512968063354,
-0.0298590250313282,
0.09046836197376251,
-0.0319368951022625,
0.07594159245491028,
0.05093543604016304,
0.004102413542568684,
0.01930202543735504,
-0.08312255889177322,
0.035986028611660004,
0.0033124834299087524,
-0.13471649587154388,
-0.004237608518451452,
-0.11452628672122955,
-0.05198396369814873,
0.1046588197350502,
0.029806965962052345,
-0.2548910677433014,
0.04685474559664726,
-0.13810139894485474,
0.060558121651411057,
-0.16846242547035217,
0.04586046189069748,
0.09593286365270615,
-0.030186206102371216,
-0.005520305596292019,
-0.07483634352684021,
0.04979706555604935,
0.08083032816648483,
-0.02796066924929619,
-0.055169034749269485
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-NER-finetuned-ner
This model is a fine-tuned version of [dslim/bert-base-NER](https://huggingface.co/dslim/bert-base-NER) on the x_glue dataset.
It achieves the following results on the evaluation set:
- Loss: 1.4380
- Precision: 0.2274
- Recall: 0.1119
- F1: 0.1499
- Accuracy: 0.8485
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.0822 | 1.0 | 878 | 1.1648 | 0.2068 | 0.1101 | 0.1437 | 0.8471 |
| 0.0102 | 2.0 | 1756 | 1.2697 | 0.2073 | 0.1110 | 0.1445 | 0.8447 |
| 0.0049 | 3.0 | 2634 | 1.3945 | 0.2006 | 0.1073 | 0.1399 | 0.8368 |
| 0.0025 | 4.0 | 3512 | 1.3994 | 0.2243 | 0.1126 | 0.1499 | 0.8501 |
| 0.0011 | 5.0 | 4390 | 1.4380 | 0.2274 | 0.1119 | 0.1499 | 0.8485 |
### Framework versions
- Transformers 4.10.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
|
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["x_glue"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "bert-base-NER-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "x_glue", "type": "x_glue", "args": "ner"}, "metrics": [{"type": "precision", "value": 0.2273838630806846, "name": "Precision"}, {"type": "recall", "value": 0.11185727172496743, "name": "Recall"}, {"type": "f1", "value": 0.14994961370507223, "name": "F1"}, {"type": "accuracy", "value": 0.8485324947589099, "name": "Accuracy"}]}]}]}
|
token-classification
|
deval/bert-base-NER-finetuned-ner
|
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"generated_from_trainer",
"dataset:x_glue",
"license:mit",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-x_glue #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
|
bert-base-NER-finetuned-ner
===========================
This model is a fine-tuned version of dslim/bert-base-NER on the x\_glue dataset.
It achieves the following results on the evaluation set:
* Loss: 1.4380
* Precision: 0.2274
* Recall: 0.1119
* F1: 0.1499
* Accuracy: 0.8485
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.10.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-x_glue #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
65,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-x_glue #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
-0.11204000562429428,
0.08741585165262222,
-0.0020595684181898832,
0.11811661720275879,
0.17322340607643127,
0.04325484111905098,
0.12332811206579208,
0.11735055595636368,
-0.07542305439710617,
0.027176694944500923,
0.12320634722709656,
0.166658416390419,
0.010077103972434998,
0.11253514140844345,
-0.04312967136502266,
-0.27496448159217834,
-0.021340174600481987,
0.05992797017097473,
-0.0699741318821907,
0.12501519918441772,
0.08873410522937775,
-0.14246954023838043,
0.095629021525383,
0.019370798021554947,
-0.22466744482517242,
0.002933983225375414,
0.02149989642202854,
-0.051978472620248795,
0.14587728679180145,
0.02968757227063179,
0.14218702912330627,
0.005434670019894838,
0.10099728405475616,
-0.1704535335302353,
0.010929854586720467,
0.042324744164943695,
0.0095163993537426,
0.0974581241607666,
0.05768972635269165,
0.00916964840143919,
0.11074290424585342,
-0.07106031477451324,
0.06313815712928772,
0.024221304804086685,
-0.12562385201454163,
-0.22225385904312134,
-0.06879892945289612,
0.03504002466797829,
0.06561827659606934,
0.09237916022539139,
-0.006626896094530821,
0.13741393387317657,
-0.09470625221729279,
0.08552228659391403,
0.2217407375574112,
-0.29756098985671997,
-0.07055625319480896,
0.05496273189783096,
0.01638050563633442,
0.027506640180945396,
-0.11698400229215622,
-0.03163180500268936,
0.05508454889059067,
0.04434528201818466,
0.1309056580066681,
-0.03472936898469925,
-0.09133310616016388,
0.019788581877946854,
-0.13954780995845795,
-0.010757490061223507,
0.1388411521911621,
0.03741743415594101,
-0.03258535638451576,
-0.044040609151124954,
-0.042296573519706726,
-0.1489776223897934,
-0.035073421895504,
-0.017322923988103867,
0.04254864528775215,
-0.04025217890739441,
-0.08474785834550858,
0.005589840933680534,
-0.11028929799795151,
-0.07570082694292068,
-0.07269909232854843,
0.15202683210372925,
0.031044483184814453,
0.010134072043001652,
-0.026645131409168243,
0.11546070128679276,
-0.013275962322950363,
-0.12176945060491562,
0.02421063557267189,
0.02335241436958313,
-0.01027609407901764,
-0.06584003567695618,
-0.04814432933926582,
-0.055573586374521255,
0.00009634415619075298,
0.12596191465854645,
-0.042909640818834305,
0.035543572157621384,
0.06843787431716919,
0.04232855141162872,
-0.08094126731157303,
0.19342488050460815,
-0.048746269196271896,
-0.012459039688110352,
-0.008651169016957283,
0.03937222436070442,
-0.01047434937208891,
-0.007319428026676178,
-0.12009527534246445,
0.0011319223558530211,
0.08981388062238693,
-0.0012619653716683388,
-0.07453032582998276,
0.07674742490053177,
-0.05629601329565048,
-0.027278253808617592,
-0.010703152976930141,
-0.08215797692537308,
0.038308095186948776,
-0.00019206436991225928,
-0.07896386086940765,
-0.022939007729291916,
0.013300152495503426,
0.016613688319921494,
0.0015688994899392128,
0.14658576250076294,
-0.11178357154130936,
0.03028608113527298,
-0.09795162826776505,
-0.12476081401109695,
0.010253180749714375,
-0.1042865589261055,
0.033737897872924805,
-0.0902288556098938,
-0.15390358865261078,
-0.013948190025985241,
0.05204075947403908,
-0.03027362935245037,
-0.049656908959150314,
-0.044364966452121735,
-0.06569003313779831,
0.006330304313451052,
-0.005402009002864361,
0.12049879133701324,
-0.05763115733861923,
0.09717951714992523,
0.032515063881874084,
0.06076161190867424,
-0.039014849811792374,
0.05366943031549454,
-0.10250575095415115,
0.012902040965855122,
-0.15959607064723969,
0.020614314824342728,
-0.05325784906744957,
0.07070039957761765,
-0.07797512412071228,
-0.10675737261772156,
0.012341536581516266,
0.004233841318637133,
0.06561639904975891,
0.09025547653436661,
-0.1731782704591751,
-0.0916447639465332,
0.15240415930747986,
-0.07042522728443146,
-0.11928888410329819,
0.11511722952127457,
-0.06469964236021042,
0.06035134941339493,
0.06003936007618904,
0.1606815755367279,
0.07651521265506744,
-0.09296636283397675,
-0.005462347064167261,
0.01401356142014265,
0.04591801017522812,
-0.07006838172674179,
0.06573396921157837,
0.006815003696829081,
0.017073459923267365,
0.02791713923215866,
-0.023891035467386246,
0.05451766029000282,
-0.10980382561683655,
-0.09185892343521118,
-0.02395988628268242,
-0.0852014496922493,
0.05813274160027504,
0.08060292899608612,
0.07435882091522217,
-0.09474308788776398,
-0.08692137151956558,
0.08539774268865585,
0.07673681527376175,
-0.04881945252418518,
0.01179283857345581,
-0.06626024097204208,
0.07408999651670456,
-0.04629909247159958,
-0.03939053788781166,
-0.17299127578735352,
-0.05179859325289726,
0.003060265677049756,
0.010337304323911667,
0.020054401829838753,
0.04400157555937767,
0.07337621599435806,
0.07242677360773087,
-0.0513245053589344,
-0.01682126522064209,
-0.03457202389836311,
0.0036152375396341085,
-0.14046958088874817,
-0.20371808111667633,
-0.04924836754798889,
-0.02343982830643654,
0.1343969702720642,
-0.21951645612716675,
0.04040183126926422,
-0.0034607297275215387,
0.08690889179706573,
0.02345868945121765,
-0.016584189608693123,
-0.05047246068716049,
0.08212065696716309,
-0.03602677956223488,
-0.04887913167476654,
0.07213671505451202,
0.008103098720312119,
-0.08766170591115952,
-0.060191716998815536,
-0.11049782484769821,
0.18390639126300812,
0.12884840369224548,
-0.12252356857061386,
-0.0935615822672844,
-0.02794642746448517,
-0.06615444272756577,
-0.028551489114761353,
-0.05242878943681717,
0.02681465819478035,
0.16960974037647247,
-0.007744283881038427,
0.14470040798187256,
-0.05818357691168785,
-0.05039078742265701,
0.02648879587650299,
-0.03633788228034973,
0.01727432571351528,
0.11685776710510254,
0.147616907954216,
-0.07274049520492554,
0.15000779926776886,
0.13258157670497894,
-0.08804451674222946,
0.1544785499572754,
-0.02988484315574169,
-0.07397246360778809,
-0.030167536810040474,
-0.04155248403549194,
-0.0028116959147155285,
0.12544125318527222,
-0.14501355588436127,
-0.008746122941374779,
0.02553132176399231,
0.009955042041838169,
0.017406456172466278,
-0.223563089966774,
-0.049565140157938004,
0.03829386830329895,
-0.02249891124665737,
-0.012373538687825203,
-0.011375360190868378,
0.008465870283544064,
0.10536345839500427,
0.009486156515777111,
-0.07949641346931458,
0.032516274601221085,
0.008628933690488338,
-0.07991053909063339,
0.21934065222740173,
-0.06263770163059235,
-0.12563207745552063,
-0.12783470749855042,
-0.07735422253608704,
-0.03410201519727707,
0.004997862968593836,
0.054382581263780594,
-0.09067299216985703,
-0.030080344527959824,
-0.052639830857515335,
0.03676775470376015,
-0.009329481981694698,
0.03985472023487091,
-0.009970627725124359,
0.0009796163067221642,
0.05685562267899513,
-0.10745850205421448,
-0.012141426093876362,
-0.06877657026052475,
-0.05732744559645653,
0.045677412301301956,
0.038313526660203934,
0.11896954476833344,
0.15278944373130798,
-0.026852456852793694,
0.013312391005456448,
-0.036096446216106415,
0.2470267117023468,
-0.06501418352127075,
-0.03365315496921539,
0.12056136131286621,
-0.01615852303802967,
0.044036537408828735,
0.11530124396085739,
0.08901413530111313,
-0.07914798706769943,
-0.0004621722036972642,
0.04975951090455055,
-0.034008339047431946,
-0.2078361064195633,
-0.04918000102043152,
-0.0510464683175087,
0.004160833545029163,
0.09749405831098557,
0.020500248298048973,
0.03126939758658409,
0.07762105017900467,
0.04561779275536537,
0.07672718167304993,
-0.05597420409321785,
0.06240777671337128,
0.10668826848268509,
0.03924829140305519,
0.1312049776315689,
-0.03649533912539482,
-0.07692305743694305,
0.034867264330387115,
-0.019496653228998184,
0.22681976854801178,
0.011583910323679447,
0.11982287466526031,
0.0580395944416523,
0.16706451773643494,
0.004332789219915867,
0.08296976238489151,
-0.005241348408162594,
-0.0548035204410553,
-0.007007550913840532,
-0.0385330505669117,
-0.021327512338757515,
0.024280011653900146,
-0.042717549949884415,
0.061136964708566666,
-0.13444265723228455,
0.0018816499505192041,
0.049560509622097015,
0.22019094228744507,
0.04781236872076988,
-0.33229219913482666,
-0.10015825927257538,
-0.007164268288761377,
-0.0201300960034132,
-0.02173951268196106,
0.013810963369905949,
0.10413705557584763,
-0.09423418343067169,
0.01782890222966671,
-0.07720491290092468,
0.08909399062395096,
-0.049762748181819916,
0.048642028123140335,
0.09844114631414413,
0.10187191516160965,
-0.007908247411251068,
0.08270976692438126,
-0.2706356942653656,
0.28294339776039124,
0.010171866975724697,
0.06893447041511536,
-0.07125574350357056,
-0.001804016763344407,
0.029775090515613556,
0.061596497893333435,
0.07747609913349152,
-0.017661219462752342,
-0.02463940903544426,
-0.2359549105167389,
-0.044092632830142975,
0.026089921593666077,
0.08407420665025711,
-0.04458971321582794,
0.09908390790224075,
-0.0278281569480896,
0.009871512651443481,
0.07987138628959656,
0.0010006329976022243,
-0.04737371206283569,
-0.08949863910675049,
-0.007970327511429787,
0.010721907950937748,
-0.05992412939667702,
-0.059795595705509186,
-0.12713880836963654,
-0.12212279438972473,
0.159600168466568,
-0.03429308906197548,
-0.02836122363805771,
-0.12034851312637329,
0.10071143507957458,
0.06578363478183746,
-0.08958563208580017,
0.03970731049776077,
0.007488761097192764,
0.08214480429887772,
0.027722297236323357,
-0.06110389530658722,
0.1156373992562294,
-0.0665440559387207,
-0.15619045495986938,
-0.07204034179449081,
0.09705130755901337,
0.040416497737169266,
0.06616845726966858,
-0.011530529707670212,
0.015308178029954433,
-0.03724968433380127,
-0.08942385762929916,
0.034397777169942856,
-0.021306119859218597,
0.06996766477823257,
0.004420512821525335,
-0.07121090590953827,
0.036488112062215805,
-0.05047263950109482,
-0.01798987202346325,
0.18935874104499817,
0.23953554034233093,
-0.10911887884140015,
-0.00035485648550093174,
0.018415963277220726,
-0.06984210014343262,
-0.20222444832324982,
0.05633191019296646,
0.05354535952210426,
0.016739346086978912,
0.05175872892141342,
-0.17627139389514923,
0.15027716755867004,
0.10916288942098618,
-0.008818299509584904,
0.09978286176919937,
-0.29066357016563416,
-0.12675823271274567,
0.1350516378879547,
0.1452503502368927,
0.10781189799308777,
-0.12736523151397705,
-0.017396366223692894,
-0.003995136823505163,
-0.14602498710155487,
0.1202821135520935,
-0.08311554789543152,
0.11824169009923935,
-0.030948106199502945,
0.08150281757116318,
0.012565464712679386,
-0.0632171705365181,
0.10962051898241043,
0.028237544000148773,
0.0997294932603836,
-0.05363436043262482,
-0.071245476603508,
0.035358116030693054,
-0.03406720235943794,
0.017359957098960876,
-0.06546565145254135,
0.018563881516456604,
-0.09848086535930634,
-0.03187696263194084,
-0.0728384405374527,
0.055782780051231384,
-0.04357541352510452,
-0.08145420998334885,
-0.03618863597512245,
0.03892577812075615,
0.024357419461011887,
-0.01925303041934967,
0.12231243401765823,
0.017048928886651993,
0.16558587551116943,
0.09020088613033295,
0.08066675066947937,
-0.07069766521453857,
-0.07979471981525421,
-0.010947619564831257,
-0.01473010703921318,
0.06088880077004433,
-0.11577410995960236,
0.021444551646709442,
0.15733526647090912,
0.03256361559033394,
0.14067256450653076,
0.09423547238111496,
-0.025997716933488846,
0.003629566403105855,
0.06584978103637695,
-0.15707935392856598,
-0.08356378972530365,
-0.014492623507976532,
-0.08628503978252411,
-0.13087810575962067,
0.05182871222496033,
0.08963494002819061,
-0.07000753283500671,
-0.006477479822933674,
-0.005590666085481644,
-0.010563663206994534,
-0.057838842272758484,
0.19761089980602264,
0.06738030910491943,
0.048487015068531036,
-0.0943809375166893,
0.0452645979821682,
0.04167921841144562,
-0.058237653225660324,
-0.009071612730622292,
0.061291832476854324,
-0.077785924077034,
-0.04287891089916229,
0.06260087341070175,
0.20931467413902283,
-0.0750841423869133,
-0.028661852702498436,
-0.14637675881385803,
-0.12308105826377869,
0.07313314080238342,
0.15568894147872925,
0.11751992255449295,
0.016708487644791603,
-0.059521209448575974,
0.01914491131901741,
-0.1275540143251419,
0.09502004832029343,
0.04640274867415428,
0.07243067026138306,
-0.15296955406665802,
0.18694902956485748,
0.004916916601359844,
0.0466027595102787,
-0.024372246116399765,
0.026832254603505135,
-0.1225549727678299,
0.0020670967642217875,
-0.08946749567985535,
-0.03596251830458641,
-0.040036674588918686,
0.006032118573784828,
0.0000054766537687100936,
-0.060211990028619766,
-0.058527871966362,
0.0026994519867002964,
-0.11073881387710571,
-0.015433094464242458,
0.0456133633852005,
0.06157102808356285,
-0.10859968513250351,
-0.03232058882713318,
0.0219561830163002,
-0.05411498621106148,
0.06813053786754608,
0.03545650094747543,
0.03435594215989113,
0.056158799678087234,
-0.13280412554740906,
0.018231268972158432,
0.06512486189603806,
0.01245785690844059,
0.08009091019630432,
-0.07940094918012619,
-0.00833150465041399,
-0.0192489642649889,
0.058483269065618515,
0.020379681140184402,
0.058396391570568085,
-0.1396544873714447,
0.000060298320022411644,
-0.02048284001648426,
-0.08954589068889618,
-0.06983459740877151,
0.028542935848236084,
0.09014604985713959,
0.013288152404129505,
0.19487982988357544,
-0.07481036335229874,
0.048649102449417114,
-0.22285237908363342,
0.0012095845304429531,
-0.007359322626143694,
-0.09681596606969833,
-0.10041975229978561,
-0.0710211843252182,
0.06152372062206268,
-0.05759478360414505,
0.14245517551898956,
0.052504587918519974,
0.027844741940498352,
0.03177907317876816,
-0.018166549503803253,
0.022571789100766182,
0.01684950664639473,
0.2016894668340683,
0.03421207144856453,
-0.04100177064538002,
0.05099226534366608,
0.05810974910855293,
0.09496217966079712,
0.11398293077945709,
0.20440512895584106,
0.12969152629375458,
-0.01885337196290493,
0.07604627311229706,
0.05698307231068611,
-0.0663209855556488,
-0.17140735685825348,
0.034671902656555176,
-0.03835397586226463,
0.08715230226516724,
-0.02574721910059452,
0.21406283974647522,
0.07014364004135132,
-0.1689402014017105,
0.03971793130040169,
-0.05158233270049095,
-0.09231510758399963,
-0.11871647089719772,
-0.04581376165151596,
-0.07900939881801605,
-0.1248641237616539,
0.007664485368877649,
-0.11478465050458908,
-0.0038025379180908203,
0.12071015685796738,
0.005844728089869022,
-0.02050470933318138,
0.16806986927986145,
0.010934002697467804,
0.040012143552303314,
0.04660195857286453,
0.016113772988319397,
-0.027779238298535347,
-0.13378725945949554,
-0.049445703625679016,
-0.02545984834432602,
-0.016175271943211555,
0.023029986768960953,
-0.07969466596841812,
-0.05788135528564453,
0.028698114678263664,
-0.008162778802216053,
-0.10285516828298569,
0.00888791959732771,
0.01497230026870966,
0.058371223509311676,
0.018276793882250786,
-0.0025908013340085745,
0.021434711292386055,
-0.021489286795258522,
0.2009550929069519,
-0.07501047104597092,
-0.05095955729484558,
-0.10275547951459885,
0.2628997564315796,
0.03617445006966591,
-0.004136402625590563,
0.036569952964782715,
-0.06888997554779053,
0.01921793259680271,
0.24182619154453278,
0.21198058128356934,
-0.08020876348018646,
0.0012021336006000638,
0.019579879939556122,
-0.011616355739533901,
-0.02818806655704975,
0.10145531594753265,
0.11374456435441971,
0.06303666532039642,
-0.09503505378961563,
-0.039493072777986526,
-0.0571066290140152,
-0.01358674094080925,
-0.021992875263094902,
0.060943495482206345,
0.07029915601015091,
0.017283625900745392,
-0.04976920410990715,
0.05163959041237831,
-0.06205097585916519,
-0.08875861763954163,
0.08763118833303452,
-0.20825842022895813,
-0.1616288721561432,
-0.015176299959421158,
0.08770639449357986,
0.0011602737940847874,
0.07924093306064606,
-0.036878347396850586,
-0.008421741425991058,
0.07190011441707611,
-0.012841175310313702,
-0.10582175850868225,
-0.08187849074602127,
0.09846478700637817,
-0.08760308474302292,
0.22291876375675201,
-0.05199079588055611,
0.06149508059024811,
0.12780867516994476,
0.06508511304855347,
-0.06902015954256058,
0.049808014184236526,
0.04629822075366974,
-0.07700170576572418,
0.020813720300793648,
0.08062989264726639,
-0.03697039559483528,
0.07884789258241653,
0.03807433694601059,
-0.1717708259820938,
0.0287283044308424,
-0.07434290647506714,
-0.066924549639225,
-0.044670287519693375,
-0.026021212339401245,
-0.051829300820827484,
0.1352110654115677,
0.2228374183177948,
-0.02058221958577633,
-0.011777661740779877,
-0.06522335857152939,
0.018295541405677795,
0.07137282937765121,
0.034428760409355164,
-0.06617289781570435,
-0.22056996822357178,
0.018618864938616753,
0.04669734090566635,
-0.03172208368778229,
-0.24791713058948517,
-0.09664647281169891,
0.005800726357847452,
-0.07169618457555771,
-0.08681919425725937,
0.07435699552297592,
0.09238622337579727,
0.059017717838287354,
-0.0612550787627697,
-0.06453780084848404,
-0.07254186272621155,
0.14932255446910858,
-0.14599068462848663,
-0.0952487662434578
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-uncased-finetuned-ner
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the x_glue dataset.
It achieves the following results on the evaluation set:
- Loss: 2.7979
- Precision: 0.0919
- Recall: 0.1249
- F1: 0.1059
- Accuracy: 0.4927
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.1773 | 1.0 | 878 | 1.7953 | 0.1025 | 0.1352 | 0.1166 | 0.5058 |
| 0.0397 | 2.0 | 1756 | 2.0827 | 0.0906 | 0.1230 | 0.1043 | 0.4888 |
| 0.022 | 3.0 | 2634 | 2.8677 | 0.0864 | 0.1260 | 0.1025 | 0.4098 |
| 0.0126 | 4.0 | 3512 | 2.8584 | 0.0848 | 0.1201 | 0.0994 | 0.4424 |
| 0.0085 | 5.0 | 4390 | 2.7979 | 0.0919 | 0.1249 | 0.1059 | 0.4927 |
### Framework versions
- Transformers 4.10.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["x_glue"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "bert-base-uncased-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "x_glue", "type": "x_glue", "args": "ner"}, "metrics": [{"type": "precision", "value": 0.09187560910782316, "name": "Precision"}, {"type": "recall", "value": 0.1248795761078998, "name": "Recall"}, {"type": "f1", "value": 0.10586493798172632, "name": "F1"}, {"type": "accuracy", "value": 0.492660102891609, "name": "Accuracy"}]}]}]}
|
token-classification
|
deval/bert-base-uncased-finetuned-ner
|
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"token-classification",
"generated_from_trainer",
"dataset:x_glue",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-x_glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
bert-base-uncased-finetuned-ner
===============================
This model is a fine-tuned version of bert-base-uncased on the x\_glue dataset.
It achieves the following results on the evaluation set:
* Loss: 2.7979
* Precision: 0.0919
* Recall: 0.1249
* F1: 0.1059
* Accuracy: 0.4927
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 5
### Training results
### Framework versions
* Transformers 4.10.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-x_glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
68,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #bert #token-classification #generated_from_trainer #dataset-x_glue #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
-0.1073223426938057,
0.09431805461645126,
-0.002355789765715599,
0.1270780861377716,
0.16259904205799103,
0.038395293056964874,
0.12094757705926895,
0.1200953796505928,
-0.07317517697811127,
0.02243405021727085,
0.11943059414625168,
0.16132286190986633,
0.01724383421242237,
0.10913281887769699,
-0.04376200959086418,
-0.2590164542198181,
-0.014747118577361107,
0.053152021020650864,
-0.060697752982378006,
0.13016235828399658,
0.09058241546154022,
-0.12675786018371582,
0.0932106077671051,
0.022794367745518684,
-0.1999518722295761,
-0.0019191671162843704,
0.010533002205193043,
-0.054280705749988556,
0.1485012024641037,
0.017506815493106842,
0.13143247365951538,
-0.012126936577260494,
0.09531079232692719,
-0.18025359511375427,
0.00620498601347208,
0.04810600355267525,
0.014321583323180676,
0.09521390497684479,
0.049303919076919556,
0.016590021550655365,
0.09331396967172623,
-0.06921374797821045,
0.05910421535372734,
0.02253989316523075,
-0.11227984726428986,
-0.2320263385772705,
-0.07928729802370071,
0.03749632462859154,
0.08003577589988708,
0.0973755344748497,
-0.0001902107906062156,
0.13111332058906555,
-0.08358322083950043,
0.08059040457010269,
0.23082219064235687,
-0.28925082087516785,
-0.06780338287353516,
0.05094385892152786,
0.01053678523749113,
0.029781777411699295,
-0.10175963491201401,
-0.041728243231773376,
0.04649822413921356,
0.048332735896110535,
0.13127805292606354,
-0.030093075707554817,
-0.11363246291875839,
0.012508290819823742,
-0.1415204405784607,
-0.02701422944664955,
0.15465910732746124,
0.04530768096446991,
-0.036641500890254974,
-0.04304901510477066,
-0.05796530097723007,
-0.15087991952896118,
-0.035158757120370865,
-0.017869599163532257,
0.047112785279750824,
-0.02893294207751751,
-0.051986388862133026,
0.000984135433100164,
-0.1039285808801651,
-0.0763910710811615,
-0.07355404645204544,
0.13550274074077606,
0.035783857107162476,
0.014184646308422089,
-0.028035305440425873,
0.11726336181163788,
-0.014589209109544754,
-0.11766302585601807,
0.020175287500023842,
0.025591257959604263,
0.00143985275644809,
-0.04476320743560791,
-0.04902273043990135,
-0.04879586771130562,
0.006097792647778988,
0.1411391645669937,
-0.026432812213897705,
0.03505133092403412,
0.05996003746986389,
0.04434977099299431,
-0.09484397619962692,
0.19904644787311554,
-0.04091909900307655,
-0.027962839230895042,
0.003382266964763403,
0.043508466333150864,
0.012305272743105888,
-0.006793310400098562,
-0.12155789136886597,
0.0028948644176125526,
0.09114222228527069,
0.00572202168405056,
-0.069599449634552,
0.06907099485397339,
-0.05368790403008461,
-0.028140611946582794,
-0.00445096381008625,
-0.08785521984100342,
0.027861887589097023,
-0.005374443717300892,
-0.07643712311983109,
-0.012203628197312355,
0.024395711719989777,
0.019716454669833183,
-0.002469744300469756,
0.12278969585895538,
-0.09829694032669067,
0.024445123970508575,
-0.0963793620467186,
-0.10362318903207779,
0.014090124517679214,
-0.10416707396507263,
0.030709736049175262,
-0.09225983172655106,
-0.17997708916664124,
-0.004043454769998789,
0.06302487105131149,
-0.026305023580789566,
-0.05951308831572533,
-0.03887541964650154,
-0.06376408785581589,
0.00878581590950489,
-0.00972044002264738,
0.12412291020154953,
-0.05909852311015129,
0.08619197458028793,
0.027386263012886047,
0.059303149580955505,
-0.0325777493417263,
0.0559748150408268,
-0.10558875650167465,
0.022527925670146942,
-0.16114500164985657,
0.01994374953210354,
-0.05320257321000099,
0.07038766145706177,
-0.08794619888067245,
-0.10147985070943832,
0.01939505524933338,
-0.011246948502957821,
0.06013686582446098,
0.0810258612036705,
-0.17756080627441406,
-0.07542247325181961,
0.1463274210691452,
-0.06982468068599701,
-0.12409850209951401,
0.1195664182305336,
-0.059365060180425644,
0.050769876688718796,
0.049313984811306,
0.15373243391513824,
0.07140102982521057,
-0.08456867188215256,
-0.0013019080506637692,
0.020094288513064384,
0.04884641245007515,
-0.0677080973982811,
0.07597112655639648,
0.005691861268132925,
0.009166603907942772,
0.03316856175661087,
-0.036419354379177094,
0.06014993414282799,
-0.08787482231855392,
-0.09761562943458557,
-0.041095390915870667,
-0.08564845472574234,
0.04490729048848152,
0.06755557656288147,
0.06646256148815155,
-0.08409903198480606,
-0.0859094113111496,
0.059974100440740585,
0.08678780496120453,
-0.047287218272686005,
0.0165255106985569,
-0.05736799165606499,
0.07982968538999557,
-0.05209248140454292,
-0.03192949295043945,
-0.16853967308998108,
-0.050511009991168976,
0.012024913914501667,
-0.021349230781197548,
0.012442892417311668,
0.030747510492801666,
0.06585630029439926,
0.07083988934755325,
-0.043616652488708496,
-0.016519296914339066,
-0.04112211614847183,
0.003767461981624365,
-0.12914445996284485,
-0.18941694498062134,
-0.05275414139032364,
-0.019136136397719383,
0.14825741946697235,
-0.19703108072280884,
0.040926385670900345,
-0.021634837612509727,
0.085006944835186,
0.015156551264226437,
-0.009379925206303596,
-0.04792259633541107,
0.06968715786933899,
-0.03998367488384247,
-0.0558437779545784,
0.07186005264520645,
0.011013444513082504,
-0.08490883558988571,
-0.06045765057206154,
-0.08998966962099075,
0.16836939752101898,
0.1255364567041397,
-0.10393510013818741,
-0.07757038623094559,
-0.019671395421028137,
-0.07036719471216202,
-0.0371035598218441,
-0.05367215722799301,
0.027212610468268394,
0.1800520420074463,
-0.0037714040372520685,
0.14136925339698792,
-0.06576848030090332,
-0.049120672047138214,
0.022574422881007195,
-0.0348367765545845,
0.01886976882815361,
0.1143326684832573,
0.15087571740150452,
-0.06408754736185074,
0.15036819875240326,
0.15212668478488922,
-0.09505835920572281,
0.12842142581939697,
-0.03689089044928551,
-0.07062449306249619,
-0.02733122557401657,
-0.039538439363241196,
-0.007379848975688219,
0.1213550865650177,
-0.15023061633110046,
0.002060876926407218,
0.036185115575790405,
0.017295803874731064,
0.01715291664004326,
-0.221195250749588,
-0.04354974254965782,
0.03770047798752785,
-0.03168715909123421,
-0.002737661125138402,
-0.006357012316584587,
0.007462484762072563,
0.09698490798473358,
0.004890038166195154,
-0.09263008832931519,
0.040225815027952194,
0.007016918156296015,
-0.075718954205513,
0.21083468198776245,
-0.06911489367485046,
-0.14131991565227509,
-0.12833216786384583,
-0.08205227553844452,
-0.04811182618141174,
-0.0021224392112344503,
0.05747950077056885,
-0.08581463247537613,
-0.03519415482878685,
-0.06145524978637695,
0.029226042330265045,
0.0016870199469849467,
0.037685561925172806,
0.0023131421767175198,
-0.004187584854662418,
0.07130246609449387,
-0.10767600685358047,
-0.014587932266294956,
-0.05970567837357521,
-0.0518152117729187,
0.03428071737289429,
0.04709722846746445,
0.11302590370178223,
0.1476677507162094,
-0.016744527965784073,
0.01090992521494627,
-0.023137709125876427,
0.24160157144069672,
-0.05804247036576271,
-0.025003798305988312,
0.14348375797271729,
-0.016030987724661827,
0.04990333691239357,
0.11482026427984238,
0.08201468735933304,
-0.07366416603326797,
-0.0025918264873325825,
0.044059257954359055,
-0.03579413890838623,
-0.2161223441362381,
-0.04791320860385895,
-0.05249493941664696,
0.008320191875100136,
0.09651798754930496,
0.023505888879299164,
0.03504323214292526,
0.07686492055654526,
0.04624117165803909,
0.08715569972991943,
-0.0601109080016613,
0.059999365359544754,
0.11768489331007004,
0.041869912296533585,
0.12478247284889221,
-0.03201204910874367,
-0.06751242280006409,
0.03869917616248131,
0.0012997114099562168,
0.22741341590881348,
0.011979085393249989,
0.1217997670173645,
0.056865595281124115,
0.1776847541332245,
-0.005867352709174156,
0.08358912169933319,
-0.013503839261829853,
-0.03816740959882736,
-0.015208587981760502,
-0.03851133584976196,
-0.028488794341683388,
0.029481543228030205,
-0.06312017887830734,
0.07326989620923996,
-0.11304366588592529,
0.011423850432038307,
0.05205480754375458,
0.24465107917785645,
0.035798799246549606,
-0.32765090465545654,
-0.09906567633152008,
-0.004599445499479771,
-0.0319506898522377,
-0.019956188276410103,
0.02481093630194664,
0.08321531862020493,
-0.09899406880140305,
0.018383493646979332,
-0.08080081641674042,
0.09090752899646759,
-0.04706168547272682,
0.04241601377725601,
0.10453423857688904,
0.1037970632314682,
0.013640078715980053,
0.09014911204576492,
-0.2625143826007843,
0.27114078402519226,
0.0006202263757586479,
0.05884606018662453,
-0.07471327483654022,
0.011827378533780575,
0.03988160938024521,
0.05080451816320419,
0.07908263057470322,
-0.01299467496573925,
-0.029401330277323723,
-0.2080201953649521,
-0.0681426003575325,
0.026482969522476196,
0.06093015521764755,
-0.04673757776618004,
0.08980534970760345,
-0.03554641827940941,
0.0038404432125389576,
0.07580596208572388,
0.013747579418122768,
-0.051099468022584915,
-0.0935911163687706,
-0.0037104319781064987,
0.024303996935486794,
-0.05810495465993881,
-0.06330562382936478,
-0.12086440622806549,
-0.12035226076841354,
0.14835740625858307,
-0.01819501630961895,
-0.0341922827064991,
-0.11457189172506332,
0.08074093610048294,
0.07334825396537781,
-0.08633439242839813,
0.051506441086530685,
-0.001595062087289989,
0.08635848760604858,
0.03125184029340744,
-0.0682605430483818,
0.10657433420419693,
-0.08086822181940079,
-0.15838299691677094,
-0.07235150784254074,
0.09761571139097214,
0.034097764641046524,
0.06284487247467041,
-0.013197817839682102,
0.010511464439332485,
-0.04562978819012642,
-0.09125678241252899,
0.026897350326180458,
-0.003728273091837764,
0.07305838167667389,
0.00627840356901288,
-0.0630311369895935,
0.029592199251055717,
-0.05581233277916908,
-0.02776476927101612,
0.18723885715007782,
0.21815775334835052,
-0.10751862823963165,
0.010732034221291542,
0.0188328567892313,
-0.06505055725574493,
-0.18276908993721008,
0.03408278897404671,
0.056485190987586975,
0.005979710258543491,
0.041454825550317764,
-0.1733374297618866,
0.15951085090637207,
0.11422337591648102,
-0.017144937068223953,
0.10541599988937378,
-0.31188249588012695,
-0.12173184752464294,
0.1415783017873764,
0.13766363263130188,
0.10421857237815857,
-0.12109341472387314,
-0.019372770562767982,
-0.009814819321036339,
-0.14992022514343262,
0.11701096594333649,
-0.08302724361419678,
0.11346211284399033,
-0.041230589151382446,
0.09023362398147583,
0.003061956027522683,
-0.061849284917116165,
0.11884816735982895,
0.03490940481424332,
0.0938154086470604,
-0.05139606073498726,
-0.04750258848071098,
0.025469982996582985,
-0.0413878858089447,
0.030623171478509903,
-0.08288165181875229,
0.02879100665450096,
-0.10062489658594131,
-0.027139004319906235,
-0.06460586190223694,
0.052845992147922516,
-0.040432192385196686,
-0.07858289033174515,
-0.03607892245054245,
0.028871038928627968,
0.038133200258016586,
-0.01438910886645317,
0.1151198297739029,
0.04318395256996155,
0.14334529638290405,
0.09433087706565857,
0.055693622678518295,
-0.0731104388833046,
-0.10319751501083374,
-0.022891027852892876,
-0.012063108384609222,
0.06121528893709183,
-0.11747144162654877,
0.020930789411067963,
0.15004383027553558,
0.024458089843392372,
0.13910043239593506,
0.08613492548465729,
-0.023045700043439865,
0.0017728887032717466,
0.0508057177066803,
-0.16147179901599884,
-0.07842952758073807,
-0.011251363903284073,
-0.05896001681685448,
-0.1260777860879898,
0.05068107694387436,
0.09160789847373962,
-0.07487834244966507,
-0.006552824750542641,
-0.003223241539672017,
0.005429774988442659,
-0.05399176850914955,
0.18741637468338013,
0.059810493141412735,
0.049055181443691254,
-0.09326695650815964,
0.0622599720954895,
0.046460509300231934,
-0.058355070650577545,
-0.010552739724516869,
0.05298841372132301,
-0.08458556234836578,
-0.04349389299750328,
0.0597587451338768,
0.1927979290485382,
-0.0707741379737854,
-0.04033830761909485,
-0.1403915137052536,
-0.12026813626289368,
0.07592999190092087,
0.13463760912418365,
0.12042403966188431,
0.011931017972528934,
-0.0674620196223259,
0.01039811223745346,
-0.10870204865932465,
0.1040249615907669,
0.05034337565302849,
0.07244159281253815,
-0.15729555487632751,
0.13798418641090393,
0.015161748975515366,
0.047464318573474884,
-0.018020149320364,
0.02895607240498066,
-0.09779534488916397,
0.0023217860143631697,
-0.0988200455904007,
-0.023579666391015053,
-0.031405359506607056,
0.013029451481997967,
-0.0002039691898971796,
-0.05975349247455597,
-0.05747654289007187,
0.01349582988768816,
-0.10597492754459381,
-0.024445153772830963,
0.03970271348953247,
0.06419450044631958,
-0.10685982555150986,
-0.035744860768318176,
0.033662401139736176,
-0.06163023039698601,
0.07306139916181564,
0.04780104011297226,
0.03095853142440319,
0.04890086501836777,
-0.11981959640979767,
0.01843026839196682,
0.06617684662342072,
0.02916419319808483,
0.07558700442314148,
-0.1021575927734375,
-0.009787286631762981,
-0.014834136702120304,
0.038610830903053284,
0.012025712057948112,
0.06644468754529953,
-0.14489637315273285,
-0.009579910896718502,
-0.014175842516124249,
-0.08118375390768051,
-0.06630515307188034,
0.010616916231811047,
0.09950852394104004,
0.01720307394862175,
0.19843225181102753,
-0.06580585241317749,
0.05522497370839119,
-0.2190995216369629,
0.004732785280793905,
-0.00979935098439455,
-0.1031806692481041,
-0.10978438705205917,
-0.06435760855674744,
0.04953591153025627,
-0.05813833698630333,
0.15303808450698853,
0.04306768253445625,
0.027608536183834076,
0.02239495888352394,
-0.012951232492923737,
0.026714742183685303,
0.01095807459205389,
0.1933995932340622,
0.03723660856485367,
-0.030180910602211952,
0.06000940874218941,
0.04378868266940117,
0.10143513232469559,
0.10209115594625473,
0.19439011812210083,
0.13347429037094116,
-0.0018829182954505086,
0.08598434925079346,
0.04955016449093819,
-0.06698449701070786,
-0.1859288513660431,
0.04001697152853012,
-0.0373985692858696,
0.10383288562297821,
-0.018747305497527122,
0.22669534385204315,
0.06514142453670502,
-0.16731451451778412,
0.041135940700769424,
-0.04733568802475929,
-0.08182784169912338,
-0.11268752813339233,
-0.04757525771856308,
-0.07450632750988007,
-0.13415604829788208,
0.0013544035609811544,
-0.10729766637086868,
-0.0030102955643087626,
0.13178576529026031,
0.002855642233043909,
-0.02054465375840664,
0.16450263559818268,
0.01039770059287548,
0.03721750155091286,
0.042166028171777725,
0.017284806817770004,
-0.039450664073228836,
-0.12607039511203766,
-0.0573108047246933,
-0.026647290214896202,
-0.014105036854743958,
0.029411252588033676,
-0.0782112181186676,
-0.04452183097600937,
0.03139520063996315,
-0.005174761172384024,
-0.09410018473863602,
0.00580986263230443,
0.002168722450733185,
0.048563096672296524,
0.023334452882409096,
0.0009793656645342708,
0.0332832969725132,
-0.014330558478832245,
0.19112417101860046,
-0.07617418467998505,
-0.054992467164993286,
-0.10154232382774353,
0.2349359691143036,
0.026385920122265816,
-0.02027227357029915,
0.03723177686333656,
-0.0637427419424057,
0.0050416747108101845,
0.2405928522348404,
0.1971755474805832,
-0.07555343210697174,
-0.009204220958054066,
0.022182054817676544,
-0.016308654099702835,
-0.035574279725551605,
0.09430117160081863,
0.1353653073310852,
0.053800467401742935,
-0.09310603141784668,
-0.04364248737692833,
-0.06087196618318558,
-0.016485655680298805,
-0.02836809866130352,
0.06689197570085526,
0.05130705609917641,
0.013155482709407806,
-0.043460018932819366,
0.04437151551246643,
-0.05800279974937439,
-0.09491808712482452,
0.07429236173629761,
-0.20009660720825195,
-0.17198938131332397,
-0.012828825041651726,
0.10451550781726837,
0.0009165694355033338,
0.067192442715168,
-0.03583427518606186,
-0.008728145621716976,
0.09554238617420197,
-0.015439677983522415,
-0.10018971562385559,
-0.08559081703424454,
0.09785439074039459,
-0.08404438942670822,
0.24007926881313324,
-0.04880857467651367,
0.05913310497999191,
0.12404447048902512,
0.06605836749076843,
-0.07710861414670944,
0.04982343688607216,
0.05117475241422653,
-0.062333304435014725,
0.018670784309506416,
0.0758497342467308,
-0.03236943855881691,
0.08045320957899094,
0.0363239161670208,
-0.15003721415996552,
0.01567031256854534,
-0.04658665135502815,
-0.0640157163143158,
-0.043269287794828415,
-0.03229888156056404,
-0.057247817516326904,
0.1369854211807251,
0.21303260326385498,
-0.028445282950997353,
-0.017770564183592796,
-0.07156426459550858,
0.022515008226037025,
0.062003664672374725,
0.019826918840408325,
-0.05835948884487152,
-0.20483659207820892,
0.02232096903026104,
0.03935597091913223,
-0.027351416647434235,
-0.2228853404521942,
-0.10638045519590378,
0.013832073658704758,
-0.079557403922081,
-0.08797810226678848,
0.06394162774085999,
0.08708259463310242,
0.057210523635149,
-0.05883786454796791,
-0.04110092669725418,
-0.0814783051609993,
0.13946588337421417,
-0.14581292867660522,
-0.09103446453809738
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-ner
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the conll2003 dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0606
- Precision: 0.9277
- Recall: 0.9385
- F1: 0.9330
- Accuracy: 0.9844
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.2454 | 1.0 | 878 | 0.0692 | 0.9106 | 0.9212 | 0.9159 | 0.9809 |
| 0.0517 | 2.0 | 1756 | 0.0616 | 0.9203 | 0.9352 | 0.9277 | 0.9834 |
| 0.0314 | 3.0 | 2634 | 0.0606 | 0.9277 | 0.9385 | 0.9330 | 0.9844 |
### Framework versions
- Transformers 4.10.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.0
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["conll2003"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "distilbert-base-uncased-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "conll2003", "type": "conll2003", "args": "conll2003"}, "metrics": [{"type": "precision", "value": 0.9276788676324229, "name": "Precision"}, {"type": "recall", "value": 0.9384718648618414, "name": "Recall"}, {"type": "f1", "value": 0.9330441552663775, "name": "F1"}, {"type": "accuracy", "value": 0.9843836878643939, "name": "Accuracy"}]}]}]}
|
token-classification
|
deval/distilbert-base-uncased-finetuned-ner
|
[
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"token-classification",
"generated_from_trainer",
"dataset:conll2003",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
distilbert-base-uncased-finetuned-ner
=====================================
This model is a fine-tuned version of distilbert-base-uncased on the conll2003 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0606
* Precision: 0.9277
* Recall: 0.9385
* F1: 0.9330
* Accuracy: 0.9844
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.10.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3"
] |
[
69,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #token-classification #generated_from_trainer #dataset-conll2003 #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3"
] |
[
-0.10817645490169525,
0.11320975422859192,
-0.002463343320414424,
0.1325080841779709,
0.1525668501853943,
0.030717479065060616,
0.12243320047855377,
0.11324058473110199,
-0.09156515449285507,
0.025485731661319733,
0.13229374587535858,
0.16207194328308105,
0.015222606249153614,
0.11622914671897888,
-0.05286842957139015,
-0.24673697352409363,
-0.003482274943962693,
0.04362938925623894,
-0.04988817870616913,
0.13168871402740479,
0.09821856766939163,
-0.13306692242622375,
0.09448322653770447,
0.01561434380710125,
-0.1954789161682129,
-0.007503822445869446,
0.003234093077480793,
-0.05218910798430443,
0.1421356499195099,
0.015798868611454964,
0.12576164305210114,
-0.011038488708436489,
0.09547081589698792,
-0.17366495728492737,
0.00519853038713336,
0.04452250152826309,
0.012629679404199123,
0.0965389832854271,
0.041965991258621216,
0.01288489531725645,
0.09556606411933899,
-0.06101891025900841,
0.06007314845919609,
0.011879269033670425,
-0.11892972141504288,
-0.20520564913749695,
-0.09283290803432465,
0.04748104512691498,
0.08332131057977676,
0.09796521067619324,
0.004302435088902712,
0.1391502022743225,
-0.08987855166196823,
0.08536813408136368,
0.2053663432598114,
-0.2851874530315399,
-0.06802612543106079,
0.04824196919798851,
0.010730392299592495,
0.04269130527973175,
-0.09855502843856812,
-0.0438108891248703,
0.04361143335700035,
0.050126537680625916,
0.1312054693698883,
-0.027818024158477783,
-0.11129193007946014,
0.012868739664554596,
-0.1425374299287796,
-0.042615797370672226,
0.16715098917484283,
0.04990904778242111,
-0.03604314848780632,
-0.041193317621946335,
-0.06359817087650299,
-0.164606511592865,
-0.0288506131619215,
-0.013921602629125118,
0.04442150890827179,
-0.026521621271967888,
-0.0562010258436203,
0.0006074340781196952,
-0.10039065033197403,
-0.06885430961847305,
-0.08101525902748108,
0.1350230872631073,
0.035472165793180466,
0.015971506014466286,
-0.021399779245257378,
0.11426833271980286,
0.0004156090726610273,
-0.12140396237373352,
0.02053394354879856,
0.021590907126665115,
0.004075511358678341,
-0.04597953334450722,
-0.05174160748720169,
-0.04296202212572098,
0.007668188773095608,
0.14446380734443665,
-0.03264591470360756,
0.03329899162054062,
0.0537223219871521,
0.04418426379561424,
-0.08855423331260681,
0.18396814167499542,
-0.042952101677656174,
-0.03383726254105568,
0.008014483377337456,
0.054429519921541214,
0.02557644620537758,
-0.004421702120453119,
-0.12368816137313843,
0.009974930435419083,
0.0978240892291069,
0.00805945135653019,
-0.0656750500202179,
0.06479966640472412,
-0.06243341416120529,
-0.028439724817872047,
0.019054798409342766,
-0.08543899655342102,
0.028429865837097168,
-0.010553550906479359,
-0.08018594980239868,
-0.02219662256538868,
0.018496351316571236,
0.02436514012515545,
-0.003702285699546337,
0.11057643592357635,
-0.09640969336032867,
0.019585998728871346,
-0.09028411656618118,
-0.09864632040262222,
0.016174878925085068,
-0.10933615267276764,
0.031578194350004196,
-0.09535062313079834,
-0.1961366981267929,
-0.0036772103048861027,
0.06431546062231064,
-0.0231940820813179,
-0.07015454024076462,
-0.0454469658434391,
-0.06731566041707993,
0.009105165489017963,
-0.009637190029025078,
0.12045573443174362,
-0.06551243364810944,
0.0895988941192627,
0.020354699343442917,
0.05777054652571678,
-0.053436312824487686,
0.05229693278670311,
-0.10850567370653152,
0.02443447895348072,
-0.15330395102500916,
0.03093011863529682,
-0.04774477705359459,
0.062403757125139236,
-0.08881653100252151,
-0.10016635060310364,
0.01641538180410862,
-0.01878579519689083,
0.06655380129814148,
0.08686811476945877,
-0.18543602526187897,
-0.06253620982170105,
0.13510993123054504,
-0.06166541948914528,
-0.12211087346076965,
0.12273122370243073,
-0.06652545183897018,
0.04114757478237152,
0.05760481581091881,
0.151268869638443,
0.06827706098556519,
-0.07533963769674301,
0.0043062870390713215,
0.010957700200378895,
0.051408860832452774,
-0.06191996857523918,
0.07675977796316147,
0.005168503616005182,
0.015956176444888115,
0.0292753204703331,
-0.03666900470852852,
0.05517169088125229,
-0.08859530091285706,
-0.1008458063006401,
-0.04041777178645134,
-0.09800264984369278,
0.04665899649262428,
0.0626826137304306,
0.06456942111253738,
-0.08851597458124161,
-0.07604947686195374,
0.052120909094810486,
0.09142126888036728,
-0.044396091252565384,
0.020580459386110306,
-0.06487052142620087,
0.0796298161149025,
-0.04605669155716896,
-0.03179631754755974,
-0.17417313158512115,
-0.033221110701560974,
0.01266591064631939,
0.0014970781048759818,
0.014374371618032455,
0.025865375995635986,
0.06246580556035042,
0.07420652359724045,
-0.04267241433262825,
-0.01950160041451454,
-0.037407711148262024,
0.005499699618667364,
-0.13156111538410187,
-0.19282323122024536,
-0.04342617839574814,
-0.01942722126841545,
0.15563999116420746,
-0.20186986029148102,
0.034146133810281754,
-0.025317687541246414,
0.08778396993875504,
0.016044404357671738,
-0.015395388938486576,
-0.04330144077539444,
0.0685446634888649,
-0.049985047429800034,
-0.054318640381097794,
0.0652972161769867,
0.012191365472972393,
-0.09109502285718918,
-0.06641801446676254,
-0.08756392449140549,
0.1621599644422531,
0.12625901401042938,
-0.10113069415092468,
-0.0736972913146019,
-0.01609187014400959,
-0.06495258212089539,
-0.03419007360935211,
-0.0506848506629467,
0.03021325170993805,
0.17371228337287903,
-0.005152401514351368,
0.14251960813999176,
-0.07014402747154236,
-0.04465324804186821,
0.020611366257071495,
-0.03425287827849388,
0.01945994608104229,
0.1139461100101471,
0.13531753420829773,
-0.08171374350786209,
0.15208883583545685,
0.1549399048089981,
-0.09335020929574966,
0.1156088262796402,
-0.03903908282518387,
-0.06310130655765533,
-0.02674637921154499,
-0.028943706303834915,
-0.00767382001504302,
0.11595501750707626,
-0.13964858651161194,
0.008336789906024933,
0.03642470762133598,
0.021943621337413788,
0.009799150750041008,
-0.219939723610878,
-0.04100724309682846,
0.03671945631504059,
-0.0339110791683197,
-0.006095096468925476,
-0.01094911154359579,
0.006276230327785015,
0.09935463964939117,
0.0047756945714354515,
-0.10570273548364639,
0.04776104539632797,
0.008964328095316887,
-0.07248826324939728,
0.2040858119726181,
-0.08690149337053299,
-0.14122022688388824,
-0.12400861084461212,
-0.08494829386472702,
-0.05893726646900177,
0.01116553321480751,
0.0529136024415493,
-0.07159104198217392,
-0.03631899505853653,
-0.0723022073507309,
0.0022558560594916344,
0.0009501348831690848,
0.029797973111271858,
0.016634253785014153,
-0.008368164300918579,
0.06798326969146729,
-0.10520795732736588,
-0.012365929782390594,
-0.05160174146294594,
-0.048995353281497955,
0.03591780737042427,
0.042148180305957794,
0.11395005881786346,
0.14934197068214417,
-0.012972189113497734,
0.007610892411321402,
-0.02072436548769474,
0.2539505660533905,
-0.05908921733498573,
-0.02042444981634617,
0.13773727416992188,
-0.01955288276076317,
0.053054459393024445,
0.12119343876838684,
0.07568088173866272,
-0.08348016440868378,
-0.0023108357563614845,
0.031455181539058685,
-0.03918725624680519,
-0.21109984815120697,
-0.053314097225666046,
-0.05534370243549347,
-0.007145646493881941,
0.09790794551372528,
0.02383042499423027,
0.038233764469623566,
0.08166204392910004,
0.03835240751504898,
0.09565010666847229,
-0.052014078944921494,
0.06341741979122162,
0.12053665518760681,
0.04688894748687744,
0.12359534949064255,
-0.03213420882821083,
-0.06087375804781914,
0.0471905879676342,
0.005095073953270912,
0.2220466136932373,
0.012302805669605732,
0.12432891875505447,
0.06067948788404465,
0.18183033168315887,
-0.010069633834064007,
0.0759282186627388,
-0.00884784385561943,
-0.03117067739367485,
-0.02179272286593914,
-0.037693291902542114,
-0.040242258459329605,
0.027858445420861244,
-0.05666995048522949,
0.07338742911815643,
-0.10428783297538757,
0.02284305728971958,
0.051644206047058105,
0.2555640935897827,
0.038054946810007095,
-0.34092211723327637,
-0.09854152053594589,
0.00109822116792202,
-0.03457309305667877,
-0.024386251345276833,
0.02938893437385559,
0.0815875455737114,
-0.0963740348815918,
0.022230898961424828,
-0.06638722121715546,
0.09087297320365906,
-0.051630064845085144,
0.041319768875837326,
0.08099466562271118,
0.0915423035621643,
0.013591844588518143,
0.08538056910037994,
-0.27146899700164795,
0.2695838510990143,
0.0020335684530436993,
0.059616442769765854,
-0.07820868492126465,
0.007642426528036594,
0.03506791591644287,
0.06423255056142807,
0.07269669324159622,
-0.005352470558136702,
-0.02157154679298401,
-0.19570811092853546,
-0.06258122622966766,
0.021919243037700653,
0.05955266207456589,
-0.04061947762966156,
0.08851595968008041,
-0.031296901404857635,
0.008749022148549557,
0.06803934276103973,
0.0062082731164991856,
-0.048033151775598526,
-0.10001352429389954,
-0.005992088466882706,
0.037084951996803284,
-0.045628871768713,
-0.06241396814584732,
-0.10849673300981522,
-0.12127772718667984,
0.1436549872159958,
-0.03217245265841484,
-0.037102799862623215,
-0.10771254450082779,
0.07576044648885727,
0.08138390630483627,
-0.08232399821281433,
0.051328133791685104,
-0.005709149409085512,
0.07703285664319992,
0.03309512883424759,
-0.058879245072603226,
0.0984581857919693,
-0.08060692250728607,
-0.1688452810049057,
-0.07240866124629974,
0.10352367162704468,
0.03739522024989128,
0.06483674049377441,
-0.005450844764709473,
0.01844576932489872,
-0.049744948744773865,
-0.08899864554405212,
0.02327617071568966,
0.0023729519452899694,
0.08808863162994385,
0.01538530271500349,
-0.04917232692241669,
0.025563480332493782,
-0.05283571034669876,
-0.03200049325823784,
0.18853305280208588,
0.23288851976394653,
-0.10219664126634598,
0.019824828952550888,
0.028952129185199738,
-0.06224953010678291,
-0.17716413736343384,
0.02657325752079487,
0.05439988523721695,
0.004062270745635033,
0.040895476937294006,
-0.17603285610675812,
0.14484943449497223,
0.11823732405900955,
-0.018414992839097977,
0.10456965118646622,
-0.3253220319747925,
-0.11916499584913254,
0.13174839317798615,
0.13228487968444824,
0.1055133044719696,
-0.12217845022678375,
-0.020775066688656807,
-0.018731476739048958,
-0.1456051766872406,
0.11531335115432739,
-0.07345319539308548,
0.11319600045681,
-0.03303464874625206,
0.09853953868150711,
0.0030358540825545788,
-0.05728635564446449,
0.12637527287006378,
0.034807153046131134,
0.09825865924358368,
-0.0573401004076004,
-0.04132261499762535,
0.03259928897023201,
-0.04245501011610031,
0.02306911163032055,
-0.08006823062896729,
0.038028109818696976,
-0.1071585938334465,
-0.0195018257945776,
-0.06471694260835648,
0.04148260876536369,
-0.03478411212563515,
-0.07371383905410767,
-0.04367890581488609,
0.02920873649418354,
0.0542772002518177,
-0.011224198155105114,
0.1301102340221405,
0.049098215997219086,
0.1319703310728073,
0.09840372949838638,
0.06539186090230942,
-0.07598528265953064,
-0.0864650085568428,
-0.029904864728450775,
-0.016162382438778877,
0.05857016146183014,
-0.1186198815703392,
0.02580508589744568,
0.14418600499629974,
0.023877494037151337,
0.13740704953670502,
0.08120670169591904,
-0.016980841755867004,
0.006630040239542723,
0.05120522528886795,
-0.17020870745182037,
-0.06955239176750183,
-0.00207059015519917,
-0.03581907972693443,
-0.11978787183761597,
0.05151791498064995,
0.09553749114274979,
-0.07083508372306824,
-0.007862416096031666,
-0.00329012144356966,
0.013019414618611336,
-0.05001607909798622,
0.19012168049812317,
0.05593651905655861,
0.04586886614561081,
-0.10153039544820786,
0.07285546511411667,
0.054516613483428955,
-0.052903033792972565,
-0.00386789720505476,
0.047017667442560196,
-0.08960752189159393,
-0.041729580610990524,
0.05017653852701187,
0.16984573006629944,
-0.07022494822740555,
-0.04390936344861984,
-0.13073822855949402,
-0.11442448198795319,
0.08023715019226074,
0.13468000292778015,
0.11675281077623367,
0.016358956694602966,
-0.06769761443138123,
0.000990710104815662,
-0.1090991348028183,
0.09828060865402222,
0.047513313591480255,
0.07371482998132706,
-0.15733586251735687,
0.13783763349056244,
0.004347299225628376,
0.04014428332448006,
-0.01569807529449463,
0.02853160910308361,
-0.0922224372625351,
0.007776952814310789,
-0.11451610922813416,
-0.020412571728229523,
-0.037293434143066406,
0.013678283430635929,
-0.004293015226721764,
-0.05913880467414856,
-0.05720795691013336,
0.01445155218243599,
-0.10744114220142365,
-0.019912000745534897,
0.03903994336724281,
0.06187763437628746,
-0.11213311553001404,
-0.03753271326422691,
0.029462359845638275,
-0.06141684204339981,
0.0757351815700531,
0.04615128040313721,
0.025802932679653168,
0.042793937027454376,
-0.11831625550985336,
0.011860872618854046,
0.06629679352045059,
0.029679937288165092,
0.07761236280202866,
-0.10005507618188858,
-0.013756772503256798,
-0.002232906175777316,
0.037544623017311096,
0.014099475927650928,
0.0775010734796524,
-0.1385369747877121,
-0.010671062394976616,
-0.010623650625348091,
-0.07799592614173889,
-0.06507866829633713,
0.017960112541913986,
0.10501421242952347,
0.017397819086909294,
0.2117377072572708,
-0.060620490461587906,
0.04301229119300842,
-0.20622839033603668,
0.0028291158378124237,
-0.009644659236073494,
-0.10717295855283737,
-0.13312563300132751,
-0.06061554327607155,
0.05153422802686691,
-0.057664208114147186,
0.15196025371551514,
0.02733137086033821,
0.026553161442279816,
0.022098751738667488,
0.0024877474643290043,
0.02094321697950363,
0.009928583167493343,
0.19421900808811188,
0.04273122921586037,
-0.03530365973711014,
0.057670336216688156,
0.03872951120138168,
0.10562846064567566,
0.10216277837753296,
0.18794463574886322,
0.13751569390296936,
-0.0006151287234388292,
0.08835700154304504,
0.037604618817567825,
-0.06515274196863174,
-0.1754055619239807,
0.03262369707226753,
-0.03752991184592247,
0.10625074803829193,
-0.015471761114895344,
0.2279808521270752,
0.05541122704744339,
-0.16879211366176605,
0.0331403985619545,
-0.05034196004271507,
-0.08079177886247635,
-0.10232765227556229,
-0.0639391541481018,
-0.07805169373750687,
-0.1259469985961914,
-0.0004644935834221542,
-0.11075980961322784,
0.008620215579867363,
0.12805825471878052,
0.00445753987878561,
-0.02505398541688919,
0.14523586630821228,
0.0017235492123290896,
0.038010526448488235,
0.03869561105966568,
0.01353471539914608,
-0.03502430394291878,
-0.1110701784491539,
-0.07435449212789536,
-0.024660782888531685,
-0.014861605130136013,
0.037946898490190506,
-0.07131223380565643,
-0.036500174552202225,
0.02696118876338005,
-0.010986940935254097,
-0.09129567444324493,
0.007318954914808273,
0.004903777968138456,
0.05119321867823601,
0.03402009233832359,
0.006444526370614767,
0.03641402721405029,
-0.008416485972702503,
0.19665338099002838,
-0.0748487114906311,
-0.06379985064268112,
-0.10773223638534546,
0.2278040498495102,
0.028930043801665306,
-0.021108688786625862,
0.04188592731952667,
-0.06633566319942474,
0.0050066313706338406,
0.23243562877178192,
0.19834978878498077,
-0.09656862169504166,
-0.01400521956384182,
0.009468605741858482,
-0.014175452291965485,
-0.034356601536273956,
0.09437727928161621,
0.13050082325935364,
0.04275999218225479,
-0.08997633308172226,
-0.0398663729429245,
-0.07035182416439056,
-0.012330877594649792,
-0.033772196620702744,
0.058181677013635635,
0.03981747478246689,
0.005713402759283781,
-0.04382667317986488,
0.047551676630973816,
-0.06796412169933319,
-0.09200788289308548,
0.061600495129823685,
-0.199005588889122,
-0.17041145265102386,
-0.009912470355629921,
0.09833572804927826,
0.004328522831201553,
0.05889531224966049,
-0.03365052863955498,
-0.0014563101576641202,
0.08455890417098999,
-0.019923580810427666,
-0.09293641149997711,
-0.08293168246746063,
0.10648664087057114,
-0.0718170553445816,
0.23169483244419098,
-0.04480297118425369,
0.07209361344575882,
0.12203482538461685,
0.06703657656908035,
-0.0822090283036232,
0.056391771882772446,
0.05491861328482628,
-0.05440649762749672,
0.02168349176645279,
0.0693649873137474,
-0.02778610773384571,
0.08277266472578049,
0.044878680258989334,
-0.13061197102069855,
0.010545195080339909,
-0.043900374323129654,
-0.0553804486989975,
-0.04542292654514313,
-0.033240627497434616,
-0.055771492421627045,
0.13874585926532745,
0.20725072920322418,
-0.03533142805099487,
-0.014447090215981007,
-0.06954600661993027,
0.023784196004271507,
0.06087842211127281,
0.00841572880744934,
-0.061394672840833664,
-0.21567288041114807,
0.016550946980714798,
0.04382448270916939,
-0.019952859729528427,
-0.2107541561126709,
-0.1027153804898262,
0.001159833511337638,
-0.07493405789136887,
-0.08686104416847229,
0.0711950734257698,
0.08035538345575333,
0.05000495910644531,
-0.05983276665210724,
-0.025278707966208458,
-0.08488844335079193,
0.13487741351127625,
-0.1352304220199585,
-0.08917593955993652
] |
null | null |
transformers
|
# Fintuned Wav2Vec of Timit - 4001 checkpoint
|
{}
|
automatic-speech-recognition
|
devin132/w2v-timit-ft-4001
|
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #wav2vec2 #automatic-speech-recognition #endpoints_compatible #region-us
|
# Fintuned Wav2Vec of Timit - 4001 checkpoint
|
[
"# Fintuned Wav2Vec of Timit - 4001 checkpoint"
] |
[
"TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #endpoints_compatible #region-us \n",
"# Fintuned Wav2Vec of Timit - 4001 checkpoint"
] |
[
37,
17
] |
[
"passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #endpoints_compatible #region-us \n# Fintuned Wav2Vec of Timit - 4001 checkpoint"
] |
[
-0.05309497192502022,
-0.1442043036222458,
-0.007042224518954754,
-0.03038093075156212,
0.07965343445539474,
-0.02979755587875843,
0.0821806862950325,
0.08774377405643463,
-0.020471950992941856,
0.008100052364170551,
0.1090833991765976,
0.1654638946056366,
0.055881068110466,
0.05458346754312515,
-0.0549735389649868,
-0.18069714307785034,
0.00915868952870369,
0.003421335481107235,
0.0175471268594265,
0.13520117104053497,
0.09379167854785919,
-0.04679679870605469,
0.04747065156698227,
0.07277613133192062,
-0.08095793426036835,
0.044426169246435165,
0.08126817643642426,
-0.16052234172821045,
0.13549773395061493,
0.02815888077020645,
0.08674181997776031,
0.05348679795861244,
0.026865480467677116,
-0.11546026915311813,
-0.0005501749110408127,
0.014209829270839691,
-0.018420400097966194,
-0.0006559207104146481,
0.018112407997250557,
0.023828361183404922,
0.0014611114747822285,
-0.026798129081726074,
-0.003563180798664689,
0.04780321195721626,
-0.06881944835186005,
-0.07237550616264343,
-0.037337370216846466,
0.11582854390144348,
0.07417219877243042,
0.08555255830287933,
0.005826129578053951,
0.17927585542201996,
-0.12853193283081055,
0.10523739457130432,
0.09902424365282059,
-0.38141128420829773,
0.027900610119104385,
0.003794363234192133,
0.022916190326213837,
-0.030344154685735703,
-0.018580731004476547,
0.09135125577449799,
0.008418915793299675,
0.0006362146232277155,
-0.019135242328047752,
-0.04417882114648819,
-0.2010965794324875,
0.033498670905828476,
-0.13831184804439545,
-0.035923656076192856,
0.11226364970207214,
-0.003127331379801035,
0.0007716581458225846,
-0.008298935368657112,
-0.09712065011262894,
-0.049263861030340195,
-0.01923273876309395,
-0.10377372056245804,
-0.05824608355760574,
0.04224368557333946,
-0.12487620115280151,
-0.031351394951343536,
-0.09515804797410965,
-0.0688810646533966,
-0.19892871379852295,
0.12446774542331696,
0.010341161862015724,
0.07838907092809677,
-0.1954411119222641,
0.06289240717887878,
-0.01609334908425808,
-0.023837054148316383,
0.004709313157945871,
-0.0014090167824178934,
-0.014188635163009167,
-0.006367601919919252,
-0.031003441661596298,
-0.010305830277502537,
0.049343857914209366,
0.10835950076580048,
0.14946863055229187,
0.02568492852151394,
-0.0762728676199913,
0.06579727679491043,
-0.13166552782058716,
0.07198463380336761,
-0.01798274926841259,
-0.038834162056446075,
0.06881887465715408,
-0.12736117839813232,
0.0746944397687912,
-0.05182505026459694,
-0.07586111128330231,
-0.11061889678239822,
0.06870581209659576,
0.11346734315156937,
0.0016063832445070148,
0.044443197548389435,
-0.003818274475634098,
-0.002262970432639122,
0.09204032272100449,
-0.06568099558353424,
0.012427741661667824,
0.014894159510731697,
0.05460437759757042,
0.12592169642448425,
0.010513709858059883,
0.05351531133055687,
-0.09195492416620255,
0.058188546448946,
0.02892414666712284,
0.008435023948550224,
0.08041319251060486,
-0.04227612912654877,
0.023190263658761978,
-0.058564119040966034,
0.0011081171687692404,
-0.2392376810312271,
-0.09815990179777145,
0.06089720129966736,
-0.019012629985809326,
0.00860859826207161,
-0.009567013941705227,
-0.023682720959186554,
-0.0017804937670007348,
0.054659292101860046,
-0.048926204442977905,
0.04697493463754654,
-0.020034857094287872,
0.030806373804807663,
0.05222519487142563,
0.13720989227294922,
-0.019646884873509407,
-0.011324010789394379,
-0.04735215753316879,
0.009450741112232208,
0.05805087462067604,
0.0015284912660717964,
-0.034636300057172775,
0.07653091102838516,
-0.06556257605552673,
0.004910824820399284,
-0.10722451657056808,
0.0036023962311446667,
0.007597628515213728,
0.145980566740036,
-0.143860325217247,
-0.1037110686302185,
0.21530529856681824,
-0.10463850945234299,
-0.05866464599967003,
0.08582814782857895,
0.08931015431880951,
-0.007164135109633207,
0.07530893385410309,
0.2071051150560379,
-0.016656886786222458,
-0.2719457149505615,
0.08182480931282043,
0.0902906283736229,
-0.1919519603252411,
-0.07551629096269608,
0.045339543372392654,
-0.0778084322810173,
-0.11083724349737167,
-0.01132587157189846,
0.07389377802610397,
0.035849422216415405,
-0.07359057664871216,
-0.07880449295043945,
-0.04964360222220421,
-0.05246090888977051,
0.020694321021437645,
-0.07971212267875671,
0.09815209358930588,
-0.10632310807704926,
0.009451944380998611,
-0.10507266968488693,
0.04806686192750931,
0.035688407719135284,
0.10235028713941574,
-0.17392608523368835,
0.0684712827205658,
-0.06188026815652847,
0.00530580896884203,
-0.1066208928823471,
0.21477511525154114,
-0.02239830419421196,
0.10926476866006851,
0.0071665700525045395,
0.10547997802495956,
0.10773342847824097,
-0.05488968640565872,
-0.02837623469531536,
0.004114435985684395,
0.10723616927862167,
0.07748112082481384,
-0.025120142847299576,
-0.06572495400905609,
0.01390104927122593,
-0.028402339667081833,
0.05202246457338333,
-0.05893595516681671,
-0.0008206636994145811,
0.19287095963954926,
0.09419003874063492,
0.043621424585580826,
0.022548865526914597,
-0.04025673866271973,
0.01797386445105076,
-0.01238334272056818,
-0.004663808736950159,
0.06320331245660782,
0.022813422605395317,
-0.0013887903187423944,
0.1573079228401184,
-0.028626227751374245,
0.130240797996521,
0.214710995554924,
-0.2064262330532074,
-0.022889001294970512,
0.10797303169965744,
-0.025087421759963036,
-0.008775157853960991,
0.03935307636857033,
-0.05785079672932625,
0.04862573742866516,
-0.042234450578689575,
0.15153537690639496,
0.022185882553458214,
0.012775443494319916,
0.06198503077030182,
-0.036855585873126984,
-0.026248687878251076,
0.039796654134988785,
0.04149475693702698,
-0.11841050535440445,
0.008536247536540031,
0.1753634363412857,
-0.0036874248180538416,
0.004849021788686514,
0.0034928738605231047,
-0.02117706462740898,
0.025084182620048523,
-0.001679484499618411,
0.0422179251909256,
-0.028185192495584488,
-0.2200649529695511,
-0.09180101752281189,
0.06203943118453026,
0.02073582448065281,
0.06516288965940475,
-0.15445365011692047,
0.021152649074792862,
0.057497765868902206,
-0.04730594530701637,
-0.14861714839935303,
0.09514108300209045,
-0.06863832473754883,
0.06290280818939209,
0.002433598507195711,
-0.09821705520153046,
0.06054415553808212,
-0.030495882034301758,
-0.1039193794131279,
0.019799960777163506,
-0.07895158231258392,
-0.2933849096298218,
-0.11643481999635696,
-0.13041619956493378,
0.040979426354169846,
0.05479172617197037,
0.084192655980587,
-0.0999295637011528,
0.01594051532447338,
0.014485005289316177,
0.0601041205227375,
-0.05103333294391632,
0.12784339487552643,
0.10115833580493927,
0.045326799154281616,
0.004460927564650774,
-0.09108967334032059,
-0.01767781563103199,
-0.042683351784944534,
0.003384880954399705,
0.07311279326677322,
-0.04365428909659386,
0.08805790543556213,
0.12616942822933197,
0.014558621682226658,
0.0650659129023552,
-0.029250383377075195,
0.2285095453262329,
-0.026960333809256554,
-0.005067955702543259,
0.1399381309747696,
0.020451311022043228,
0.008990625850856304,
0.16604849696159363,
-0.012165137566626072,
-0.05218472331762314,
-0.027494313195347786,
-0.06103760376572609,
-0.09604983776807785,
-0.17314785718917847,
-0.05983477458357811,
-0.07670290768146515,
-0.0019371468806639314,
0.02715885266661644,
0.04168372601270676,
0.10572290420532227,
0.022715410217642784,
0.04195265844464302,
0.04406820610165596,
0.010063189081847668,
0.04043696075677872,
0.1490774005651474,
-0.02687721885740757,
0.10232869535684586,
-0.07274802774190903,
-0.12213040888309479,
0.07358915358781815,
0.03945526108145714,
0.11851368099451065,
0.15565747022628784,
0.06783492863178253,
0.004269652534276247,
0.17120487987995148,
0.1498037874698639,
0.1474151611328125,
0.05984330549836159,
-0.04707598313689232,
0.03566703200340271,
-0.038418516516685486,
-0.0008281998452730477,
0.03929990157485008,
0.15465249121189117,
-0.042184870690107346,
-0.027890024706721306,
-0.13122433423995972,
0.05451807752251625,
0.16461648046970367,
0.10860729217529297,
-0.12504634261131287,
0.003972613252699375,
0.036155249923467636,
-0.07767137140035629,
0.00699563417583704,
0.09532541036605835,
0.010049781762063503,
-0.10265658050775528,
0.10044967383146286,
-0.013215531595051289,
0.07883694767951965,
-0.027816765010356903,
0.0636686310172081,
-0.11356069147586823,
-0.15362808108329773,
0.02675131894648075,
0.03106110915541649,
-0.2543798089027405,
0.20336075127124786,
0.012788703665137291,
0.0716472640633583,
-0.08839083462953568,
-0.03408803790807724,
-0.008748775348067284,
0.13246022164821625,
0.1191362664103508,
0.021225009113550186,
-0.03444904461503029,
-0.07062120735645294,
-0.028923679143190384,
0.01693679578602314,
0.14903731644153595,
0.06186225637793541,
0.017273465171456337,
-0.040479034185409546,
-0.06636962294578552,
0.00291343848221004,
0.04192570969462395,
-0.18984641134738922,
-0.09674748033285141,
0.031005918979644775,
0.24464187026023865,
0.08394679427146912,
-0.015890618786215782,
-0.052749086171388626,
-0.11614582687616348,
0.20164144039154053,
-0.1623702198266983,
0.026917723938822746,
-0.05837385728955269,
-0.16178227961063385,
0.04769954830408096,
-0.042364541441202164,
0.03809450566768646,
0.01541940774768591,
0.03554648533463478,
-0.03534761816263199,
-0.12238551676273346,
0.12411876022815704,
-0.12569840252399445,
-0.06024465709924698,
-0.03887564688920975,
0.2129315733909607,
-0.09443243592977524,
0.02443297952413559,
0.04732668772339821,
-0.007655661553144455,
-0.04401812329888344,
-0.02526211366057396,
0.09194882214069366,
0.03669082000851631,
-0.11797372996807098,
-0.05877286568284035,
0.10571353137493134,
-0.058577463030815125,
-0.0762351006269455,
-0.06879763305187225,
0.24258233606815338,
0.17948375642299652,
-0.036064498126506805,
0.1908971518278122,
0.09612005949020386,
0.01699063740670681,
-0.23514698445796967,
-0.09628757834434509,
-0.10959674417972565,
0.021793575957417488,
-0.06746365875005722,
0.024319741874933243,
-0.03678646683692932,
-0.05087978020310402,
-0.08101296424865723,
0.15351524949073792,
-0.18946923315525055,
-0.13815507292747498,
0.1183537021279335,
-0.02705574221909046,
0.3532489836215973,
-0.06234735623002052,
-0.1311674267053604,
-0.06063860282301903,
-0.18865840137004852,
0.011279277503490448,
-0.12916596233844757,
0.08216924965381622,
0.03255581110715866,
0.002992374124005437,
0.010664446279406548,
-0.050351582467556,
0.0863187238574028,
0.03287411853671074,
-0.040834691375494,
-0.059264663606882095,
-0.03981652855873108,
0.059967074543237686,
0.02856488898396492,
-0.008482599630951881,
0.04965204373002052,
0.0438963882625103,
-0.03983968123793602,
-0.051830969750881195,
-0.08888791501522064,
0.08638016879558563,
0.040059059858322144,
-0.054656125605106354,
0.020444296300411224,
-0.1412011831998825,
-0.018324049189686775,
-0.015521587803959846,
0.12730808556079865,
-0.03109416551887989,
0.02382504753768444,
0.2182871252298355,
0.010617190040647984,
-0.15845032036304474,
0.06502372026443481,
-0.004437574185431004,
-0.13227495551109314,
0.09903476387262344,
0.05184166505932808,
0.06394434720277786,
0.044055260717868805,
0.055603593587875366,
0.01701192744076252,
0.07772500067949295,
-0.02154584974050522,
0.0027922498993575573,
0.09476833790540695,
-0.14801332354545593,
-0.0013464648509398103,
-0.024752892553806305,
-0.008631488308310509,
0.15600299835205078,
0.16197215020656586,
0.1573517769575119,
-0.006708279252052307,
-0.014249468222260475,
-0.023158526048064232,
-0.009276270866394043,
-0.1265951246023178,
0.11825981736183167,
0.0956524908542633,
0.0389687679708004,
-0.13986894488334656,
0.020130954682826996,
-0.008631776086986065,
-0.17101244628429413,
-0.009463425725698471,
-0.0077597107738256454,
-0.1557278335094452,
-0.06991669535636902,
-0.030283743515610695,
0.049302030354738235,
-0.005328909028321505,
-0.11522912979125977,
0.008961228653788567,
-0.1225472092628479,
0.09588591754436493,
0.07574214786291122,
0.05787477269768715,
0.10072749853134155,
-0.06362608075141907,
-0.04186059534549713,
-0.0667426809668541,
-0.03825962170958519,
-0.027320964261889458,
0.003546450985595584,
-0.17024646699428558,
0.10417095571756363,
-0.004943125415593386,
0.1237269788980484,
-0.08529255539178848,
-0.04222310706973076,
-0.07761086523532867,
0.0742403119802475,
-0.09488638490438461,
-0.04332490265369415,
-0.08303442597389221,
-0.04670196399092674,
0.0699005126953125,
-0.07588370144367218,
-0.07731606811285019,
0.03755878284573555,
-0.07518059015274048,
0.03923177346587181,
-0.031650058925151825,
-0.053271226584911346,
-0.06891096383333206,
0.007067802827805281,
0.06095216050744057,
-0.05641316995024681,
0.06343387812376022,
0.17861124873161316,
-0.09470511227846146,
0.10393349826335907,
-0.0946071445941925,
-0.12099462002515793,
0.10222821682691574,
-0.0010143837425857782,
0.025612886995077133,
0.021018942818045616,
0.0034099649637937546,
0.07846560329198837,
0.044833727180957794,
0.026422882452607155,
0.13796184957027435,
-0.05074941739439964,
0.006125097628682852,
-0.11029182374477386,
-0.10213842988014221,
-0.024855684489011765,
-0.027571948245167732,
0.09304790198802948,
0.006595725659281015,
0.10470904409885406,
-0.05638554319739342,
0.02768724597990513,
-0.018186336383223534,
0.022076882421970367,
-0.07264265418052673,
-0.03246007114648819,
-0.07181143015623093,
-0.03741448372602463,
0.03030933067202568,
-0.04156625270843506,
0.1529887318611145,
-0.11976223438978195,
0.052967917174100876,
0.028855424374341965,
-0.11926350742578506,
-0.1067095398902893,
-0.010717589408159256,
0.3347967267036438,
0.15354637801647186,
-0.045578837394714355,
-0.0921899676322937,
-0.006563973613083363,
0.032198335975408554,
0.016804290935397148,
-0.0034682846162468195,
0.10041040927171707,
-0.07527852803468704,
0.1510501205921173,
0.01663403958082199,
-0.010714979842305183,
-0.17856743931770325,
-0.13251419365406036,
-0.1425049901008606,
0.023641573265194893,
-0.03089797869324684,
0.05211934447288513,
0.1798042207956314,
0.013762537389993668,
-0.03820650279521942,
-0.03066188097000122,
-0.027514668181538582,
-0.15024274587631226,
-0.15109217166900635,
-0.08662659674882889,
-0.13907024264335632,
0.043270133435726166,
0.038815148174762726,
0.0685439333319664,
0.05160591006278992,
0.08513341844081879,
0.02658797986805439,
0.07209151983261108,
-0.0376618392765522,
-0.051269326359033585,
0.05597978085279465,
-0.05846719071269035,
-0.042333755642175674,
-0.016294827684760094,
-0.015435624867677689,
-0.014781313017010689,
-0.055878713726997375,
-0.005281397141516209,
0.037607163190841675,
-0.11998385190963745,
0.027011021971702576,
-0.050003331154584885,
-0.0660788044333458,
-0.05059905722737312,
0.018882857635617256,
0.05912244692444801,
0.11183495074510574,
0.07174594700336456,
-0.04081982001662254,
0.028447480872273445,
0.1360890120267868,
-0.11474685370922089,
-0.2142026126384735,
-0.03483181446790695,
0.11327292770147324,
0.08954755961894989,
0.17525464296340942,
-0.039370346814394,
-0.005696743726730347,
-0.10197798162698746,
0.41419196128845215,
0.21594220399856567,
0.03099815547466278,
0.0703822523355484,
0.029685592278838158,
0.0324765220284462,
0.03200565651059151,
0.02339191362261772,
0.1086992397904396,
0.1743854433298111,
-0.03554132208228111,
-0.10981254279613495,
-0.048457901924848557,
-0.06599400192499161,
-0.11597346514463425,
0.14232099056243896,
-0.10146307200193405,
-0.09553690254688263,
-0.02546701952815056,
0.06555454432964325,
-0.05751904845237732,
-0.031026745215058327,
0.035076990723609924,
-0.12568749487400055,
-0.027161741629242897,
0.062398090958595276,
0.10063549876213074,
0.0988563671708107,
0.05922743305563927,
-0.05009949952363968,
-0.02449052967131138,
0.060467783361673355,
0.06835665553808212,
-0.18336743116378784,
0.03201155737042427,
0.03169209882616997,
0.014042713679373264,
-0.07932557910680771,
-0.0461474284529686,
0.165387824177742,
0.06663019955158234,
0.10615822672843933,
-0.04968419671058655,
0.13900581002235413,
-0.008527060970664024,
-0.07138457894325256,
0.002095498377457261,
0.06944486498832703,
-0.0121987359598279,
-0.08150611817836761,
0.03388321027159691,
-0.12782306969165802,
0.021388456225395203,
0.0801382064819336,
0.0608011856675148,
-0.08711493760347366,
-0.010867445729672909,
-0.021719489246606827,
0.05646192654967308,
-0.012164020910859108,
-0.03873822093009949,
0.06489668786525726,
-0.02163674309849739,
-0.010808495804667473,
0.052589163184165955,
-0.13237226009368896,
-0.08198685944080353,
-0.14629846811294556,
-0.034654296934604645,
-0.052401382476091385,
0.010819482617080212,
-0.11753968894481659,
-0.051673904061317444,
-0.12854990363121033,
-0.029517309740185738,
-0.08366873860359192,
0.04417303949594498,
0.09000878036022186,
0.01568813994526863,
0.024260388687253,
-0.03125281259417534,
0.0803215354681015,
0.10099511593580246,
-0.15019544959068298,
-0.10737254470586777
] |
null | null |
transformers
|
# Dummy Model
This be a dummmmmy
|
{}
|
fill-mask
|
devtrent/dummy-model
|
[
"transformers",
"pytorch",
"camembert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
|
# Dummy Model
This be a dummmmmy
|
[
"# Dummy Model\n\nThis be a dummmmmy"
] |
[
"TAGS\n#transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"# Dummy Model\n\nThis be a dummmmmy"
] |
[
38,
10
] |
[
"passage: TAGS\n#transformers #pytorch #camembert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# Dummy Model\n\nThis be a dummmmmy"
] |
[
-0.023303888738155365,
-0.009363984689116478,
-0.008085619658231735,
0.061085399240255356,
0.12596508860588074,
0.05969986692070961,
0.1818746030330658,
0.03848603740334511,
0.09215144068002701,
0.009374219924211502,
0.16620922088623047,
0.19869080185890198,
-0.007112557999789715,
0.13458587229251862,
-0.001957059372216463,
-0.33907395601272583,
0.09402205049991608,
0.01640002802014351,
-0.05403877794742584,
0.0747612863779068,
0.053326565772295,
-0.07976214587688446,
0.07219637185335159,
-0.007934603840112686,
-0.10783993452787399,
0.029125336557626724,
0.029643068090081215,
-0.07653975486755371,
0.08896521478891373,
0.024054236710071564,
0.18411177396774292,
0.030214112251996994,
-0.03973047062754631,
-0.0727335587143898,
0.048112787306308746,
-0.004395116586238146,
-0.04369718208909035,
0.03817043825984001,
-0.02754698321223259,
-0.0030872465576976538,
0.08248524367809296,
0.11446903645992279,
0.026347218081355095,
0.05777931585907936,
-0.1377258598804474,
-0.04645105078816414,
-0.023639339953660965,
0.06507541984319687,
0.02611382119357586,
0.06161630153656006,
0.004732150584459305,
0.25212010741233826,
-0.14372317492961884,
0.06785470992326736,
0.17650721967220306,
-0.2080456167459488,
-0.019938386976718903,
0.16675856709480286,
0.07643230259418488,
-0.0826495960354805,
-0.028859244659543037,
0.06491175293922424,
0.03328294679522514,
0.06482083350419998,
-0.013662856072187424,
-0.04344938322901726,
0.07265592366456985,
-0.01482623815536499,
-0.09447546303272247,
-0.04337197542190552,
0.05755149945616722,
-0.0057138679549098015,
0.007557623088359833,
0.006172167602926493,
-0.1296118199825287,
-0.032126594334840775,
-0.04648655652999878,
-0.055905260145664215,
-0.0168883316218853,
0.029583003371953964,
-0.032015956938266754,
0.0031295327935367823,
-0.06574369221925735,
0.02248653769493103,
-0.22595994174480438,
0.22167284786701202,
0.013156861998140812,
0.057523760944604874,
-0.1568077653646469,
0.06389820575714111,
-0.06692145019769669,
-0.1261081099510193,
0.053673405200242996,
-0.12264474481344223,
0.030080081894993782,
0.01372369471937418,
-0.034358832985162735,
-0.021466989070177078,
0.08215154707431793,
0.09068480879068375,
0.030491553246974945,
0.04504018276929855,
0.057326413691043854,
0.10004830360412598,
0.05779077112674713,
0.08310728520154953,
-0.0827958956360817,
0.020360242575407028,
0.026705993339419365,
-0.02046465501189232,
0.09254365414381027,
-0.05153755471110344,
-0.10756842792034149,
-0.014162903651595116,
0.017776964232325554,
0.048566512763500214,
0.05733742564916611,
0.030562622472643852,
-0.11092999577522278,
-0.03316608443856239,
0.1016169935464859,
0.007674457971006632,
-0.010537908412516117,
-0.034976594150066376,
-0.010098001919686794,
0.17026996612548828,
0.05039626359939575,
0.04104173183441162,
0.00503772497177124,
0.10924984514713287,
-0.07505499571561813,
-0.030749337747693062,
-0.07223396748304367,
-0.0764254480600357,
0.034873101860284805,
-0.10347520560026169,
0.06467153131961823,
-0.15074510872364044,
-0.16068688035011292,
0.0693507194519043,
0.07207891345024109,
-0.01945926621556282,
-0.055030517280101776,
-0.02601361647248268,
-0.04580829292535782,
0.004231027327477932,
0.007549862377345562,
-0.030833669006824493,
-0.028894852846860886,
0.059869587421417236,
-0.05956292897462845,
0.11988797783851624,
-0.10991691052913666,
0.027757929638028145,
-0.06154928356409073,
0.05677441507577896,
-0.20460103452205658,
-0.0026233086828142405,
-0.02508741430938244,
0.10324333608150482,
-0.031013084575533867,
-0.05891887843608856,
-0.11743304133415222,
0.027552960440516472,
0.02605665661394596,
0.24197076261043549,
-0.13173240423202515,
-0.10969635844230652,
0.12679523229599,
-0.10215432196855545,
-0.12303022295236588,
0.0837186723947525,
-0.04711504653096199,
0.07137320190668106,
0.02848660945892334,
0.011515953578054905,
-0.0387285016477108,
-0.13164843618869781,
0.07236536592245102,
0.06839016824960709,
-0.12413877993822098,
0.007297032047063112,
-0.006301861722022295,
0.024709898978471756,
-0.183444082736969,
0.04099469631910324,
0.06663379818201065,
0.08878818899393082,
-0.08819802850484848,
-0.06753913313150406,
-0.011788768693804741,
-0.07253598421812057,
0.15631486475467682,
0.05915560573339462,
0.11201854795217514,
-0.02717057801783085,
-0.0669567883014679,
-0.025792529806494713,
0.020547587424516678,
0.08464913070201874,
-0.017573021352291107,
-0.04905225336551666,
0.0751623660326004,
-0.0842057317495346,
0.011624281294643879,
-0.1432516872882843,
-0.14128202199935913,
-0.023654628545045853,
0.07006414234638214,
0.021985767409205437,
0.13878771662712097,
0.12530535459518433,
-0.01208333857357502,
0.00799527857452631,
-0.013889666646718979,
0.04397786036133766,
0.04959917813539505,
-0.04152819141745567,
-0.1395334154367447,
-0.0013227821327745914,
-0.06717100739479065,
0.03202307969331741,
0.04545116424560547,
0.011185073293745518,
-0.14478956162929535,
0.16266658902168274,
-0.030946992337703705,
0.02078762836754322,
-0.01432196144014597,
0.005423382855951786,
-0.029596714302897453,
-0.012349752709269524,
0.06892946362495422,
-0.01166937779635191,
-0.04312461242079735,
0.08146068453788757,
-0.10389537364244461,
0.20825287699699402,
0.1858181655406952,
-0.25438758730888367,
-0.05164794251322746,
0.08183054625988007,
-0.000005112817689223448,
-0.02393471449613571,
0.06536941975355148,
0.009977702982723713,
0.061361994594335556,
-0.00261955289170146,
0.09680946916341782,
-0.012523929588496685,
0.029856937006115913,
0.0780390053987503,
-0.04471313953399658,
-0.0963350161910057,
0.07406120747327805,
0.16888776421546936,
-0.0552486814558506,
0.13689325749874115,
0.19187425076961517,
-0.01700926572084427,
0.0769134908914566,
0.04214766249060631,
0.011713319458067417,
-0.039570704102516174,
-0.0664260983467102,
-0.01749749667942524,
0.07008492946624756,
-0.20630335807800293,
0.0035840049386024475,
0.06504224985837936,
-0.05606653168797493,
0.0330275222659111,
-0.09560731798410416,
-0.0413484200835228,
0.001811192836612463,
0.05426456779241562,
-0.0791056901216507,
0.1339488923549652,
-0.00025799806462600827,
0.07246602326631546,
0.0004516495973803103,
-0.08873199671506882,
0.08309639990329742,
0.025384463369846344,
-0.006824865937232971,
0.1377742886543274,
-0.10346555709838867,
-0.23169440031051636,
-0.09382312744855881,
-0.12555554509162903,
-0.024895016103982925,
0.03641626983880997,
0.022055609151721,
-0.07086841762065887,
-0.07799391448497772,
0.06988461315631866,
0.06936901062726974,
-0.0525984987616539,
0.0765564814209938,
-0.037472985684871674,
-0.03108656033873558,
-0.059770457446575165,
-0.07777664065361023,
-0.055145032703876495,
-0.04195476323366165,
0.007077386137098074,
0.10387866199016571,
-0.16514697670936584,
0.04907781630754471,
0.16787603497505188,
0.01728634163737297,
0.06780293583869934,
0.03226136416196823,
0.2722597122192383,
-0.061709679663181305,
0.04411567747592926,
0.15906646847724915,
-0.008482913486659527,
0.07896418124437332,
0.17297832667827606,
0.043084438890218735,
-0.025027375668287277,
0.0197417251765728,
-0.03632853925228119,
-0.12155190855264664,
-0.09607941657304764,
-0.06940697878599167,
-0.06973031163215637,
-0.04503108933568001,
0.04218635708093643,
0.030808158218860626,
0.13094733655452728,
0.13396868109703064,
-0.004128890577703714,
0.04441390186548233,
-0.11576589941978455,
0.029259679839015007,
0.10871027410030365,
-0.04239475354552269,
0.11486838757991791,
-0.01911877654492855,
-0.17341294884681702,
0.033289723098278046,
-0.03938191384077072,
0.08850248157978058,
0.08660602569580078,
-0.05636902153491974,
0.06617967039346695,
0.07073896378278732,
0.12101387232542038,
0.15172985196113586,
0.04203354939818382,
-0.06511013954877853,
-0.033758409321308136,
-0.025511641055345535,
-0.0333833247423172,
0.05799996852874756,
0.14369657635688782,
-0.021283792331814766,
-0.07361593842506409,
-0.018990572541952133,
0.011539136059582233,
0.0938962996006012,
0.06790732592344284,
-0.23884552717208862,
-0.01916545443236828,
0.05091261491179466,
0.04583213850855827,
-0.06934668868780136,
0.02024131268262863,
-0.10806673020124435,
-0.1376434862613678,
0.03727934509515762,
-0.029592564329504967,
0.07357016205787659,
0.07406152039766312,
0.07796263694763184,
-0.04485973343253136,
0.0024334026966243982,
0.004828227683901787,
0.02529827132821083,
-0.22107471525669098,
0.18874183297157288,
-0.018223563209176064,
-0.012933928519487381,
-0.0916738212108612,
-0.020046859979629517,
0.10188345611095428,
0.0889156311750412,
0.12278103083372116,
0.013563440181314945,
-0.020013244822621346,
-0.05817205458879471,
-0.006933159194886684,
0.013576739467680454,
0.04985581338405609,
-0.002821695990860462,
0.017866823822259903,
0.016305284574627876,
-0.048074282705783844,
0.024506468325853348,
0.0674055814743042,
-0.05822423845529556,
-0.059711162000894547,
0.09312085062265396,
0.048978373408317566,
-0.07900404185056686,
-0.030016429722309113,
-0.10092294216156006,
-0.11571966856718063,
0.16762889921665192,
-0.032019153237342834,
-0.06426220387220383,
-0.09330524504184723,
-0.04734570533037186,
0.1485389918088913,
-0.07243040949106216,
0.1037633940577507,
-0.11722832173109055,
0.03189217299222946,
-0.08609236776828766,
-0.17988909780979156,
0.14810051023960114,
-0.1283453106880188,
-0.054128699004650116,
-0.06395440548658371,
0.05795685946941376,
-0.017469258978962898,
0.05361754447221756,
-0.00988706760108471,
0.04021196812391281,
-0.1170169785618782,
-0.0812324732542038,
0.005882361903786659,
-0.022543665021657944,
-0.060459692031145096,
0.09297847002744675,
-0.0353146493434906,
-0.028337057679891586,
-0.0007546652923338115,
0.051824942231178284,
0.1849585920572281,
0.25805017352104187,
-0.08949314802885056,
0.09664483368396759,
0.15533481538295746,
-0.02607879415154457,
-0.31846603751182556,
-0.06669551879167557,
-0.06262052804231644,
0.030174730345606804,
-0.0044248756021261215,
-0.08281554281711578,
0.11381188035011292,
-0.031764283776283264,
-0.03744542971253395,
0.11198664456605911,
-0.23965677618980408,
-0.09307488054037094,
0.2774597406387329,
0.05378209799528122,
0.4747520089149475,
-0.10065429657697678,
-0.0468580462038517,
-0.022620927542448044,
-0.09520755708217621,
0.055128373205661774,
0.011405943892896175,
0.09394478797912598,
-0.03576767444610596,
0.1296641230583191,
0.03722972050309181,
-0.07876568287611008,
0.10782323777675629,
-0.026978515088558197,
-0.0025691629853099585,
-0.07715468108654022,
-0.2027837187051773,
0.09564302861690521,
-0.004476421047002077,
0.020522305741906166,
0.021529197692871094,
0.018129782751202583,
0.04635341838002205,
-0.027028461918234825,
-0.10305499285459518,
0.09142283350229263,
-0.02100241556763649,
-0.06277719140052795,
-0.018069617450237274,
0.006534196902066469,
0.009834728203713894,
0.006687732879072428,
0.08233554661273956,
-0.0008425620617344975,
0.1047135517001152,
0.14209505915641785,
-0.002388952998444438,
-0.09659235179424286,
-0.05673744156956673,
-0.005640077404677868,
-0.08225767314434052,
0.06574109196662903,
-0.0040709273889660835,
-0.0016715575475245714,
0.11222083866596222,
0.016477247700095177,
0.036000512540340424,
0.12164780497550964,
-0.02960047870874405,
-0.013983353972434998,
0.10171424597501755,
-0.17824891209602356,
-0.0510270930826664,
-0.033716414123773575,
0.090386763215065,
-0.005012856796383858,
0.02132534794509411,
0.07945647835731506,
-0.03437960520386696,
-0.007937442511320114,
0.0005238960147835314,
0.011315478011965752,
-0.04315487667918205,
0.02378714270889759,
0.05373052880167961,
0.045620739459991455,
-0.12921881675720215,
-0.015995969995856285,
-0.018534736707806587,
-0.12232745438814163,
-0.02113231271505356,
0.044726625084877014,
-0.05950305238366127,
-0.10597918927669525,
-0.02254565805196762,
0.11222159117460251,
-0.12350580096244812,
0.002599213272333145,
-0.09589525312185287,
-0.16678749024868011,
0.02434501238167286,
0.12732578814029694,
0.11308057606220245,
0.07317417114973068,
-0.06392127275466919,
-0.02538236975669861,
-0.051380448043346405,
0.01126366201788187,
0.09086832404136658,
0.05704314261674881,
-0.08074147999286652,
-0.016876159235835075,
-0.017432989552617073,
0.14221476018428802,
-0.12458078563213348,
-0.059873826801776886,
-0.15745584666728973,
0.0203547365963459,
-0.1459956020116806,
-0.06947854161262512,
-0.11365029215812683,
-0.09434226900339127,
0.03623035550117493,
-0.059930697083473206,
-0.0553860105574131,
-0.022442378103733063,
-0.0904330462217331,
0.006503646727651358,
0.02853524312376976,
-0.015841485932469368,
-0.015897046774625778,
-0.036452386528253555,
0.1104283481836319,
-0.06542547047138214,
0.031093351542949677,
0.14027737081050873,
-0.05491066351532936,
0.07421565800905228,
-0.10506030917167664,
-0.08145912736654282,
0.07210994511842728,
-0.0012627753894776106,
0.09074991196393967,
-0.0005635669222101569,
0.004316226579248905,
-0.00578829413279891,
0.03378031402826309,
0.04220375791192055,
0.10222319513559341,
-0.09568106383085251,
0.026636647060513496,
0.04396352544426918,
-0.147064670920372,
-0.03814718872308731,
-0.05590466782450676,
0.04299948364496231,
0.015031162649393082,
0.14293131232261658,
-0.04051335155963898,
0.10399393737316132,
-0.05618751794099808,
0.004630338866263628,
-0.041141945868730545,
-0.10852374136447906,
0.055838219821453094,
-0.043345965445041656,
0.020121214911341667,
-0.011664511635899544,
0.18355733156204224,
-0.02300058677792549,
0.03940863162279129,
0.016666995361447334,
0.056012313812971115,
0.0170342605561018,
0.012071160599589348,
0.14815746247768402,
0.06186318397521973,
-0.031379908323287964,
-0.15743638575077057,
0.10092970728874207,
0.02651234343647957,
-0.06970164179801941,
0.16966751217842102,
0.026349233463406563,
-0.0031230712775141,
0.12869878113269806,
0.0388091616332531,
0.05086732655763626,
-0.03400060161948204,
-0.09347975254058838,
-0.0523797832429409,
0.05473916977643967,
0.028387777507305145,
-0.014803428202867508,
0.1325543224811554,
-0.034029893577098846,
0.022808214649558067,
-0.028902973979711533,
-0.04188108444213867,
-0.16351787745952606,
-0.04779505729675293,
-0.07096868753433228,
-0.06908117234706879,
0.04431961104273796,
-0.012220671400427818,
-0.06901421397924423,
0.1016310378909111,
0.02343088760972023,
-0.055976543575525284,
0.15147636830806732,
-0.018724219873547554,
0.009327346459031105,
0.04178355261683464,
0.006026104092597961,
-0.014069229364395142,
-0.00394617673009634,
-0.029710806906223297,
-0.15373316407203674,
0.01932544820010662,
-0.07847579568624496,
-0.020033815875649452,
-0.11772358417510986,
0.009781335480511189,
-0.03270702809095383,
-0.1214534342288971,
-0.04438183084130287,
-0.001547470223158598,
-0.04525167495012283,
0.07666272670030594,
0.014080180786550045,
0.028349997475743294,
-0.01730850525200367,
-0.00792565755546093,
-0.05318561568856239,
-0.1318037509918213,
-0.08435554802417755,
0.2659154534339905,
-0.051309432834386826,
0.06911676377058029,
0.0014561379794031382,
0.03257821500301361,
-0.12896159291267395,
0.2178359031677246,
0.34029585123062134,
-0.15077659487724304,
0.04568163678050041,
0.04341759905219078,
0.01909308321774006,
0.01641029492020607,
0.11694994568824768,
0.05938687175512314,
0.25903090834617615,
-0.1283854991197586,
0.033948179334402084,
-0.0754624456167221,
-0.059354979544878006,
-0.06832566857337952,
-0.04666280001401901,
0.09167035669088364,
-0.03679922595620155,
-0.056431058794260025,
0.11337943375110626,
-0.25450652837753296,
0.1323258876800537,
0.05745059251785278,
-0.14951923489570618,
-0.063968725502491,
-0.014436231926083565,
0.015870604664087296,
0.04419267922639847,
0.13207796216011047,
-0.09383336454629898,
-0.039527568966150284,
0.0811118334531784,
-0.00427812896668911,
-0.1613801270723343,
-0.024413615465164185,
0.1054244413971901,
0.05582626163959503,
0.10785121470689774,
-0.0159479808062315,
0.03675178810954094,
0.10424141585826874,
0.06314414739608765,
-0.013588344678282738,
0.027204381301999092,
0.06100098043680191,
-0.0760880634188652,
-0.057943448424339294,
-0.01760799065232277,
-0.0007751446682959795,
-0.13201597332954407,
0.015167810954153538,
-0.14852243661880493,
0.060294244438409805,
-0.06168009713292122,
-0.06916344165802002,
-0.00044998154044151306,
0.09705628454685211,
-0.02476087398827076,
0.06813414394855499,
0.09983348101377487,
0.021114913746714592,
-0.035617534071207047,
-0.0029434484895318747,
0.027736423537135124,
0.06894242763519287,
-0.055366698652505875,
-0.13229990005493164,
-0.08513187617063522,
-0.04516023397445679,
0.06851546466350555,
0.0009388424805365503,
-0.21925997734069824,
-0.079050213098526,
-0.14717957377433777,
0.02528442069888115,
-0.08204727619886398,
0.021867819130420685,
0.047355588525533676,
0.023167429491877556,
0.012797089293599129,
0.09106758236885071,
-0.013487575575709343,
0.06382793933153152,
-0.12145821005105972,
-0.11686690151691437
] |
null | null |
transformers
|
DistilBERT model trained on OSCAR nepali corpus from huggingface datasets.
We trained the DitilBERT language model on OSCAR nepali corpus and then for downstream sentiment analysis task. The dataset we used for sentiment analysis was first extracted from twitter filtering for devenagari text then labelled it as postive,negative and neutral. However, since neutral labels exceeded the positive and negative tweets we decided to use only positive and negative tweets for ease of training.
LABEL_1 = negative
LABEL_0 = positive
|
{}
|
text-classification
|
dexhrestha/Nepali-DistilBERT
|
[
"transformers",
"pytorch",
"distilbert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
DistilBERT model trained on OSCAR nepali corpus from huggingface datasets.
We trained the DitilBERT language model on OSCAR nepali corpus and then for downstream sentiment analysis task. The dataset we used for sentiment analysis was first extracted from twitter filtering for devenagari text then labelled it as postive,negative and neutral. However, since neutral labels exceeded the positive and negative tweets we decided to use only positive and negative tweets for ease of training.
LABEL_1 = negative
LABEL_0 = positive
|
[] |
[
"TAGS\n#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
38
] |
[
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.03533577919006348,
0.06443645805120468,
-0.007824759930372238,
0.02963758073747158,
0.21134145557880402,
0.0368538573384285,
0.06359195709228516,
0.10786357522010803,
0.047846585512161255,
-0.029699334874749184,
0.09624463319778442,
0.2456933856010437,
-0.04527274891734123,
0.11506538093090057,
-0.1315092295408249,
-0.2995516061782837,
0.0646483302116394,
0.06820031255483627,
0.01968790777027607,
0.11027561128139496,
0.08955937623977661,
-0.08577843010425568,
0.06416945904493332,
-0.03987749293446541,
-0.13028311729431152,
0.036934368312358856,
0.037670549005270004,
-0.12557227909564972,
0.08850666880607605,
0.03936105594038963,
0.16363440454006195,
0.029493317008018494,
-0.0571451373398304,
-0.13760130107402802,
0.03542056307196617,
0.003107793163508177,
-0.08173839002847672,
0.035451244562864304,
0.07971785217523575,
-0.13606007397174835,
0.03269175812602043,
0.01657985709607601,
0.028779901564121246,
0.05034712329506874,
-0.13549968600273132,
-0.06766978651285172,
-0.009825913235545158,
0.02846479043364525,
0.08123840391635895,
0.06563035398721695,
-0.00027321543893776834,
0.11571130156517029,
-0.14468228816986084,
0.13729768991470337,
0.08681581169366837,
-0.26667332649230957,
-0.01513616368174553,
0.09300960600376129,
0.014211298897862434,
0.03189397603273392,
-0.05005642771720886,
0.03387840837240219,
0.021587392315268517,
0.012041964568197727,
-0.005505601409822702,
-0.06911619752645493,
-0.12172640115022659,
0.01909228041768074,
-0.0760328620672226,
-0.039914727210998535,
0.2024218738079071,
-0.06752687692642212,
0.06574457883834839,
-0.03853347897529602,
-0.09920144081115723,
-0.04725521057844162,
-0.028420861810445786,
0.03284634277224541,
-0.05052020400762558,
0.06803859770298004,
0.04873250797390938,
0.02093963511288166,
-0.10541380196809769,
0.027895580977201462,
-0.2198127955198288,
0.21804359555244446,
0.00917235016822815,
0.04113364964723587,
-0.17035873234272003,
0.06059039384126663,
0.043774571269750595,
-0.10760118812322617,
0.049048252403736115,
-0.10497406870126724,
0.019541887566447258,
-0.04680290073156357,
-0.07833123207092285,
-0.044003088027238846,
0.0761561468243599,
0.15131190419197083,
0.024525625631213188,
0.0676354393362999,
-0.023907558992505074,
0.08125972747802734,
0.03615585342049599,
0.12704050540924072,
0.04965166375041008,
-0.030767392367124557,
0.03752761334180832,
-0.13245059549808502,
-0.00002132852932845708,
-0.07070981711149216,
-0.1520344465970993,
-0.028104213997721672,
0.058518148958683014,
0.07771685719490051,
0.007545619271695614,
0.09117837250232697,
-0.07305282354354858,
-0.03670652583241463,
0.09205243736505508,
-0.09038619697093964,
0.022389709949493408,
0.0189626757055521,
0.024910688400268555,
0.11437109857797623,
-0.01640472002327442,
-0.004441923461854458,
-0.08554866164922714,
0.15481221675872803,
-0.05412428826093674,
0.01906411163508892,
-0.027951309457421303,
-0.07562480866909027,
0.023844171315431595,
-0.16517141461372375,
0.024268588051199913,
-0.16968505084514618,
-0.12177367508411407,
0.0011497566010802984,
0.01497613824903965,
0.0003558929602149874,
-0.029599502682685852,
-0.034584347158670425,
0.0028823118191212416,
0.05339471623301506,
-0.05009040981531143,
-0.08925710618495941,
-0.0734119787812233,
0.09545788168907166,
-0.03665677830576897,
0.07958490401506424,
-0.12844105064868927,
0.0784672200679779,
-0.0987219363451004,
-0.0187049712985754,
-0.14024826884269714,
0.05743253231048584,
-0.04765705391764641,
0.18340644240379333,
0.01636499911546707,
-0.05442013591527939,
-0.05629796162247658,
0.05081459879875183,
-0.06792773306369781,
0.17081454396247864,
-0.10482346266508102,
-0.11688733100891113,
0.18975088000297546,
-0.09539731591939926,
-0.11199936270713806,
0.08214274048805237,
-0.012322766706347466,
-0.002544441493228078,
0.10592521727085114,
0.18774141371250153,
0.11772145330905914,
0.015394842252135277,
0.071439228951931,
0.1266816407442093,
-0.09738999605178833,
-0.10514426231384277,
-0.016195401549339294,
-0.010998358018696308,
-0.11682542413473129,
0.06311710923910141,
0.08283041417598724,
0.0693083181977272,
-0.04381299018859863,
-0.038738906383514404,
-0.015374792739748955,
-0.0029897931963205338,
0.14953550696372986,
0.06494788080453873,
0.11409911513328552,
-0.07472079247236252,
0.010434641502797604,
0.010832404717803001,
-0.008651630952954292,
0.016917014494538307,
0.02875317819416523,
-0.061046965420246124,
0.11194391548633575,
0.03876045346260071,
0.02736404910683632,
-0.24566538631916046,
-0.06682449579238892,
-0.011323003098368645,
0.1456235647201538,
-0.02446315996348858,
0.10121438652276993,
0.045561324805021286,
-0.0504569448530674,
-0.010978372767567635,
-0.029581138864159584,
0.17828664183616638,
0.022655870765447617,
-0.06422974169254303,
-0.0612877793610096,
0.0651540756225586,
-0.07150227576494217,
0.012235969305038452,
-0.07036937773227692,
0.020627280697226524,
0.08606486022472382,
0.12204300612211227,
0.010734139941632748,
0.06475073099136353,
-0.02579765021800995,
0.07209211587905884,
-0.07104320824146271,
0.019227510318160057,
0.11117701232433319,
-0.010595849715173244,
-0.07011682540178299,
0.13524381816387177,
-0.1373681277036667,
0.2673107087612152,
0.19483336806297302,
-0.2967563271522522,
0.0005786092369817197,
-0.04439404606819153,
-0.007282515522092581,
0.030610160902142525,
0.030042126774787903,
0.014859852381050587,
0.08437592536211014,
0.0014727829257026315,
0.20341786742210388,
-0.021047484129667282,
-0.03919289633631706,
-0.018922755494713783,
-0.04877391830086708,
-0.03148360177874565,
0.08788784593343735,
0.06451795995235443,
-0.192406564950943,
0.19050060212612152,
0.21731194853782654,
0.010114802047610283,
0.16024211049079895,
-0.010486523620784283,
0.043989237397909164,
0.09252246469259262,
-0.03757351264357567,
-0.024272754788398743,
-0.08932791650295258,
-0.1848243772983551,
-0.03918878361582756,
0.07472185045480728,
0.03010893426835537,
0.06895712018013,
-0.10219920426607132,
-0.027038687840104103,
0.0004840063920710236,
0.021132981404662132,
-0.01947878859937191,
0.08704918622970581,
0.08203180879354477,
0.1052171140909195,
-0.017219819128513336,
-0.07267280668020248,
0.11330383270978928,
-0.0011106154415756464,
-0.07149384170770645,
0.18412140011787415,
-0.15954560041427612,
-0.36233094334602356,
-0.1530739665031433,
-0.20456592738628387,
-0.02883506752550602,
0.06615062057971954,
0.10685895383358002,
-0.12165717035531998,
-0.048558108508586884,
0.0375000461935997,
-0.013693227432668209,
-0.04041895270347595,
0.03981194645166397,
-0.05303730443120003,
0.07329315692186356,
-0.05222955346107483,
-0.06364883482456207,
-0.06660815328359604,
-0.03131863474845886,
-0.004695216193795204,
0.16393853724002838,
-0.12483653426170349,
0.06658802926540375,
0.1819998174905777,
0.0010995424818247557,
0.06644674390554428,
-0.032483141869306564,
0.1697184294462204,
-0.08651559799909592,
-0.02343188226222992,
0.1893177032470703,
-0.07345744967460632,
0.07808925211429596,
0.15666639804840088,
0.020104380324482918,
-0.0712679922580719,
0.0352557972073555,
-0.035343270748853683,
-0.08934015780687332,
-0.2058166265487671,
-0.1703205555677414,
-0.12546730041503906,
0.05237005278468132,
0.0663270428776741,
0.07582127302885056,
0.12632738053798676,
0.06528977304697037,
0.00627241050824523,
0.010700550861656666,
0.006936580874025822,
0.07483439892530441,
0.24698598682880402,
-0.0010819705203175545,
0.14767786860466003,
-0.057353224605321884,
-0.13245494663715363,
0.08233633637428284,
0.000922833161894232,
0.1185675784945488,
0.08539658784866333,
0.017674902454018593,
0.005295653361827135,
0.05462205410003662,
0.164198637008667,
0.1299368292093277,
0.04298880323767662,
-0.013622048310935497,
-0.01172587089240551,
0.0032578855752944946,
-0.0797785148024559,
0.006457295268774033,
0.07906489074230194,
-0.14195358753204346,
-0.08270972222089767,
-0.11039547622203827,
0.10006770491600037,
0.08380265533924103,
0.042938295751810074,
-0.2052999883890152,
0.005745685659348965,
0.09206069260835648,
-0.027502331882715225,
-0.09957162290811539,
0.06463603675365448,
-0.04812092334032059,
-0.13455109298229218,
0.10769277811050415,
-0.029609164223074913,
0.13354617357254028,
-0.0870715081691742,
0.08272852748632431,
-0.0378170944750309,
-0.11202792823314667,
0.03467349335551262,
0.10786684602499008,
-0.27751585841178894,
0.2031957507133484,
0.007435075007379055,
-0.06144534796476364,
-0.07824365049600601,
-0.015199865214526653,
0.039944443851709366,
0.22591036558151245,
0.06934285908937454,
0.004277070518583059,
-0.05739999935030937,
-0.1865520477294922,
-0.009981787763535976,
-0.008337096311151981,
0.12231403589248657,
-0.03427664935588837,
-0.01814279891550541,
-0.036011241376399994,
-0.030255382880568504,
-0.03578435257077217,
-0.06897740066051483,
0.02666986919939518,
-0.17997102439403534,
0.056329283863306046,
0.034454237669706345,
0.05416429787874222,
0.01469043642282486,
-0.04343695193529129,
-0.11887014657258987,
0.19838201999664307,
-0.10767136514186859,
-0.09184177964925766,
-0.11828504502773285,
-0.07852382957935333,
0.02535579912364483,
-0.08476060628890991,
0.06807194650173187,
-0.08172672241926193,
0.018900277093052864,
-0.06600436568260193,
-0.20524995028972626,
0.11596046388149261,
-0.10182060301303864,
-0.03258875012397766,
-0.058350928127765656,
0.1526644378900528,
-0.07479622215032578,
0.010474151000380516,
0.03318091109395027,
0.02239469438791275,
-0.08559903502464294,
-0.08446884155273438,
-0.018381169065833092,
0.03129338473081589,
0.06142119690775871,
0.08739607781171799,
-0.09792511910200119,
-0.07674866914749146,
-0.03134889155626297,
0.02817792072892189,
0.2929084002971649,
0.1401015967130661,
-0.06586769968271255,
0.1629326492547989,
0.10387758165597916,
-0.06942285597324371,
-0.3373493552207947,
-0.09150945395231247,
-0.09645266830921173,
-0.03972399979829788,
-0.042589932680130005,
-0.16358928382396698,
0.13413257896900177,
-0.004249863792210817,
-0.010055972263216972,
0.08473600447177887,
-0.16361457109451294,
-0.08480892330408096,
0.19654500484466553,
-0.0355062410235405,
0.36373743414878845,
-0.09189414978027344,
-0.09806639701128006,
-0.07035496085882187,
-0.1232207641005516,
0.12262474000453949,
0.007738110609352589,
0.08150525391101837,
-0.02050303854048252,
0.04451111704111099,
0.04815887659788132,
-0.03690929710865021,
0.10097026824951172,
0.036669690161943436,
0.025901002809405327,
-0.11938466131687164,
-0.09219347685575485,
0.023168733343482018,
-0.019243339076638222,
-0.007111898623406887,
-0.01547485776245594,
0.01685570739209652,
-0.17164339125156403,
-0.04131095111370087,
-0.07032524049282074,
0.05912882834672928,
0.04161927476525307,
-0.029813537374138832,
0.012351144105196,
-0.020498499274253845,
-0.000361355283530429,
0.006620287895202637,
0.251852810382843,
-0.03737054020166397,
0.1604781597852707,
0.08527542650699615,
0.141584113240242,
-0.15723979473114014,
0.01194052491337061,
-0.07652142643928528,
-0.05061504244804382,
0.06191904842853546,
-0.06635212153196335,
0.07575498521327972,
0.13591395318508148,
-0.05730273202061653,
0.07247055321931839,
0.11612356454133987,
0.07706465572118759,
-0.034392137080430984,
0.16330119967460632,
-0.2292891889810562,
0.04589579999446869,
-0.050483379513025284,
-0.033954232931137085,
0.06465915590524673,
0.0655360221862793,
0.1258573830127716,
0.06694923341274261,
-0.04017629101872444,
0.005630772560834885,
0.00028037314768880606,
0.005372054409235716,
0.07443340867757797,
0.04748379439115524,
0.04316747188568115,
-0.14709694683551788,
0.05031560733914375,
0.05119774490594864,
-0.15819577872753143,
-0.022534551098942757,
0.1376158893108368,
-0.1704932600259781,
-0.1271103173494339,
-0.021827740594744682,
0.12368015199899673,
-0.09311434626579285,
-0.046253565698862076,
-0.07048245519399643,
-0.13402129709720612,
0.07112511247396469,
0.18836617469787598,
0.12805050611495972,
0.09663103520870209,
-0.06118634715676308,
-0.04969988390803337,
0.0036050756461918354,
-0.004089095629751682,
0.017009761184453964,
0.03120747022330761,
-0.12284451723098755,
0.046005018055438995,
-0.02090919390320778,
0.15390309691429138,
-0.09199176728725433,
-0.07624588906764984,
-0.1582917422056198,
0.04238278418779373,
-0.09195777773857117,
-0.023019742220640182,
-0.09330286085605621,
-0.01648246869444847,
0.0030273916199803352,
-0.030272169038653374,
-0.026145517826080322,
-0.06213071197271347,
-0.11623096466064453,
0.04011767357587814,
-0.028817979618906975,
0.04146858677268028,
-0.06920336186885834,
-0.04603973776102066,
0.09102679789066315,
-0.03833403438329697,
0.10358903557062149,
0.10654495656490326,
-0.0914529487490654,
0.0934479758143425,
-0.14121071994304657,
-0.1319282501935959,
0.1433861404657364,
0.030263781547546387,
0.07207431644201279,
0.07694290578365326,
0.03595962002873421,
0.07349478453397751,
0.004535248037427664,
0.06631990522146225,
0.06761990487575531,
-0.12337382882833481,
0.061452679336071014,
-0.046973392367362976,
-0.17189696431159973,
-0.05778007209300995,
-0.04047338292002678,
0.10660306364297867,
0.010234192945063114,
0.1720496565103531,
-0.05692226439714432,
0.1017514169216156,
-0.03180769085884094,
0.0038062711246311665,
-0.01604292169213295,
-0.20698778331279755,
-0.06364472210407257,
-0.08077114075422287,
0.026275143027305603,
0.005083381198346615,
0.23303534090518951,
0.061751753091812134,
0.033835094422101974,
0.04869496077299118,
0.09752455353736877,
-0.0014774927403777838,
0.023545393720269203,
0.17794077098369598,
0.10133370757102966,
-0.05567975342273712,
-0.05575546622276306,
0.05616139620542526,
0.029215605929493904,
0.006353367585688829,
0.14132826030254364,
0.07252193242311478,
-0.041009921580553055,
0.07551323622465134,
-0.03376345708966255,
0.04427867755293846,
-0.1321653574705124,
-0.16054923832416534,
-0.05143791809678078,
0.07023841142654419,
0.01740087941288948,
0.03448288515210152,
0.07088012248277664,
-0.028410857543349266,
0.05220868065953255,
-0.033101536333560944,
-0.05869230628013611,
-0.18244294822216034,
-0.09428907185792923,
-0.09423913061618805,
-0.09753676503896713,
0.0058974651619791985,
-0.07943454384803772,
-0.01026046834886074,
0.06547573953866959,
0.037508975714445114,
-0.05198773369193077,
0.07752657681703568,
0.003285798244178295,
-0.05593571066856384,
0.08687124401330948,
-0.045962750911712646,
0.02649652026593685,
0.00841206219047308,
-0.029707664623856544,
-0.1380927860736847,
-0.013390704058110714,
-0.04401649907231331,
0.040850814431905746,
-0.058590736240148544,
0.007230483461171389,
-0.1483704298734665,
-0.12039808928966522,
-0.019934508949518204,
0.0580129399895668,
-0.06074916571378708,
0.14175079762935638,
0.015395265072584152,
0.00611070916056633,
0.047287240624427795,
0.17810532450675964,
-0.0544942207634449,
-0.06539076566696167,
-0.04489162564277649,
0.24079638719558716,
0.09303659200668335,
0.10803006589412689,
0.0026883413083851337,
-0.013426939956843853,
-0.07931891828775406,
0.28847232460975647,
0.27526742219924927,
-0.04996299743652344,
0.054827310144901276,
0.007495596073567867,
0.03283945098519325,
0.15242771804332733,
0.1401364952325821,
0.09061526507139206,
0.24117816984653473,
-0.0521743968129158,
-0.05017128586769104,
-0.026741530746221542,
-0.03419290855526924,
-0.13402216136455536,
0.0581725612282753,
0.05382576957345009,
-0.0488528348505497,
-0.06285785138607025,
0.10921014845371246,
-0.21934591233730316,
0.16537490487098694,
0.019078493118286133,
-0.20565392076969147,
-0.06819386035203934,
-0.03284084051847458,
0.1365688294172287,
-0.0016830840613692999,
0.07499389350414276,
-0.00323955318890512,
-0.11883772164583206,
0.042848069220781326,
0.01306091621518135,
-0.20812170207500458,
-0.0041817850433290005,
0.06021128222346306,
-0.05781300365924835,
-0.0120098190382123,
-0.02640264853835106,
0.03757385164499283,
0.06560133397579193,
0.07958315312862396,
-0.0117155397310853,
0.04959989711642265,
-0.012948726303875446,
-0.030828366056084633,
0.029231732711195946,
0.02946082502603531,
0.0038178605027496815,
-0.09871038049459457,
0.06783884763717651,
-0.16667571663856506,
0.0549757145345211,
-0.05383889377117157,
-0.05352160334587097,
-0.019258368760347366,
0.04339629411697388,
-0.05456918105483055,
0.04438189044594765,
0.10450860112905502,
0.011940731666982174,
-0.025312455371022224,
-0.04523419588804245,
-0.04262804985046387,
-0.012295196764171124,
-0.1369558572769165,
-0.14967197179794312,
-0.09997987747192383,
-0.08965370059013367,
0.09313849359750748,
0.0034958450123667717,
-0.12975360453128815,
-0.006513827480375767,
-0.11122267693281174,
0.05365913361310959,
-0.16868756711483002,
0.09322161972522736,
0.0323028489947319,
0.015595607459545135,
-0.011563225649297237,
-0.040581803768873215,
0.04532773047685623,
0.07905946671962738,
-0.1267605572938919,
-0.08728102594614029
] |
null | null |
transformers
|
#Aerith GPT model
|
{"tags": ["conversational"]}
|
text-generation
|
df4rfrrf/DialoGPT-medium-Aerith
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
#Aerith GPT model
|
[] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
-0.009697278961539268,
0.03208012506365776,
-0.007204889785498381,
0.004809224978089333,
0.16726240515708923,
0.014898733235895634,
0.09765533357858658,
0.13672804832458496,
-0.007841327227652073,
-0.031050153076648712,
0.14490588009357452,
0.20411323010921478,
-0.006439372431486845,
0.0661218985915184,
-0.07572533935308456,
-0.2683109939098358,
0.05759621039032936,
0.046649303287267685,
0.016515716910362244,
0.1200079694390297,
0.08573378622531891,
-0.05473608896136284,
0.08714032918214798,
-0.014583407901227474,
-0.150366872549057,
0.017733458429574966,
0.043394338339567184,
-0.12260226160287857,
0.11910516023635864,
0.05462685227394104,
0.07063519209623337,
0.014929565601050854,
-0.07541623711585999,
-0.1631229966878891,
0.03031250834465027,
0.01425902172923088,
-0.0594632662832737,
0.04757995903491974,
0.059961482882499695,
-0.10165371745824814,
0.10819483548402786,
0.09530027210712433,
-0.013078106567263603,
0.06798283755779266,
-0.16849711537361145,
-0.020869607105851173,
-0.01446688175201416,
0.009899779222905636,
0.05550243332982063,
0.09964893013238907,
-0.03413357585668564,
0.10497362166643143,
-0.09214533120393753,
0.11017382889986038,
0.10932035744190216,
-0.32057443261146545,
-0.005767723545432091,
0.09167823940515518,
0.039358653128147125,
0.07352814823389053,
-0.04467793554067612,
0.06258884817361832,
0.018015462905168533,
0.017986174672842026,
-0.014015024527907372,
-0.07283061742782593,
-0.11612214148044586,
0.04717336222529411,
-0.08668071031570435,
-0.059868961572647095,
0.2244078367948532,
-0.05464440956711769,
0.06881742179393768,
-0.05281897634267807,
-0.10522868484258652,
-0.04308144748210907,
-0.029833965003490448,
0.00475557055324316,
-0.07660607248544693,
0.08692064881324768,
0.00869679357856512,
-0.09547875821590424,
-0.1376667022705078,
-0.02496783249080181,
-0.1776352822780609,
0.16140350699424744,
0.02465328387916088,
0.05232657864689827,
-0.2027255892753601,
0.09623090922832489,
0.017906051129102707,
-0.08045592904090881,
0.022091427817940712,
-0.10046248883008957,
0.029131146147847176,
0.013760408386588097,
-0.04754498973488808,
-0.061387211084365845,
0.0843690037727356,
0.11199145019054413,
-0.01731434464454651,
0.025486016646027565,
-0.039331406354904175,
0.08100687712430954,
0.03553595021367073,
0.09077847748994827,
0.007288969587534666,
-0.028338588774204254,
0.025842782109975815,
-0.13719046115875244,
-0.003647835226729512,
-0.07116208970546722,
-0.16572439670562744,
-0.021088803187012672,
0.02994808368384838,
0.08289173990488052,
0.015449047088623047,
0.11682453751564026,
-0.03272046521306038,
-0.025152435526251793,
0.03602350503206253,
-0.047656361013650894,
-0.012649794109165668,
0.016648368909955025,
0.013163427822291851,
0.12399329990148544,
-0.0022096503525972366,
0.03235051408410072,
-0.13653022050857544,
0.031423524022102356,
-0.06793295592069626,
-0.003740974934771657,
-0.03486552834510803,
-0.040637075901031494,
0.009043924510478973,
-0.06862333416938782,
0.003486064961180091,
-0.15030112862586975,
-0.15063877403736115,
0.007587034720927477,
-0.007836631499230862,
-0.04107699543237686,
-0.06370922178030014,
-0.06952770054340363,
-0.013550350442528725,
0.04251532256603241,
-0.07093454152345657,
-0.011352915316820145,
-0.06403283774852753,
0.11004766076803207,
-0.03197755664587021,
0.07921615242958069,
-0.11953279376029968,
0.08390819281339645,
-0.11260783672332764,
-0.02386913076043129,
-0.060801517218351364,
0.09317506104707718,
-0.0006014376995153725,
0.09549830108880997,
-0.006563255097717047,
-0.017931854352355003,
-0.07981178909540176,
0.06445012241601944,
-0.042872510850429535,
0.21701598167419434,
-0.0615808479487896,
-0.11181682348251343,
0.28781595826148987,
-0.052628401666879654,
-0.1370542049407959,
0.11647392809391022,
0.008682746440172195,
0.05777018144726753,
0.10703510791063309,
0.19733482599258423,
-0.015276194550096989,
0.004040541127324104,
0.09471915662288666,
0.11263324320316315,
-0.11276852339506149,
-0.033160366117954254,
0.013019153848290443,
-0.04081077128648758,
-0.10867965966463089,
0.04689536616206169,
0.09810488671064377,
0.07090286910533905,
-0.04786505550146103,
-0.03377414867281914,
-0.01366397924721241,
0.0052589005790650845,
0.08885077387094498,
-0.007157256826758385,
0.10962837189435959,
-0.05819983780384064,
-0.03796621412038803,
-0.029282379895448685,
-0.012126247398555279,
-0.03951939567923546,
0.03137664496898651,
-0.043376367539167404,
0.10821941494941711,
-0.011204327456653118,
0.06364280730485916,
-0.16185984015464783,
-0.07691477984189987,
-0.017002692446112633,
0.1581239402294159,
0.024538565427064896,
0.09859629720449448,
0.0552486926317215,
-0.040398042649030685,
-0.0012767292791977525,
0.012792680412530899,
0.15581141412258148,
-0.022091681137681007,
-0.065607450902462,
-0.052166227251291275,
0.08642971515655518,
-0.05641226842999458,
0.04504093527793884,
-0.05937713757157326,
0.012367865070700645,
0.05064384639263153,
0.10342344641685486,
-0.00018274025933351368,
0.03323284164071083,
-0.008164864964783192,
0.002145637758076191,
-0.058205123990774155,
0.007405933458358049,
0.10799351334571838,
0.00036868182360194623,
-0.07365862280130386,
0.22074243426322937,
-0.17796069383621216,
0.1765957772731781,
0.1893044263124466,
-0.299345999956131,
0.017949223518371582,
-0.10759581625461578,
-0.04561871662735939,
0.014407722279429436,
0.05567655712366104,
-0.0454222597181797,
0.1703362911939621,
-0.009871348738670349,
0.18874616920948029,
-0.04946064203977585,
-0.04464937001466751,
-0.0200483538210392,
-0.05118836089968681,
-0.0024189651012420654,
0.07781197130680084,
0.10685696452856064,
-0.13992026448249817,
0.1964332014322281,
0.1621224284172058,
0.048237916082143784,
0.19945049285888672,
0.015346456319093704,
-0.011589210480451584,
0.0909530371427536,
0.005220826715230942,
-0.058739423751831055,
-0.07409929484128952,
-0.2594851851463318,
-0.030033592134714127,
0.07992640137672424,
0.0422382652759552,
0.1212305948138237,
-0.11349532753229141,
-0.038956157863140106,
-0.01763172075152397,
-0.023146281018853188,
0.021672505885362625,
0.0914369598031044,
0.06075398623943329,
0.13201528787612915,
-0.001710098935291171,
-0.007300339173525572,
0.10524573177099228,
0.01783694699406624,
-0.09354141354560852,
0.18308524787425995,
-0.13652534782886505,
-0.37097251415252686,
-0.13911493122577667,
-0.18057456612586975,
-0.05449081212282181,
0.05712554603815079,
0.11679314076900482,
-0.12011238187551498,
-0.018752124160528183,
0.01578843593597412,
0.10931742936372757,
-0.08449502289295197,
0.0021454424131661654,
-0.06880278885364532,
0.0321490578353405,
-0.10310184955596924,
-0.09194442629814148,
-0.055416494607925415,
-0.031392451375722885,
-0.08001253753900528,
0.1423761546611786,
-0.10777941346168518,
0.04476889222860336,
0.20262959599494934,
0.04653622955083847,
0.05625178664922714,
-0.044105201959609985,
0.19377262890338898,
-0.11264272034168243,
-0.01661740615963936,
0.19215328991413116,
-0.048360925167798996,
0.07476246356964111,
0.1232115849852562,
-0.006348740309476852,
-0.08765771239995956,
0.03011748194694519,
-0.02085109055042267,
-0.07988511025905609,
-0.23219464719295502,
-0.13938382267951965,
-0.12429051846265793,
0.09477275609970093,
0.028005298227071762,
0.056365787982940674,
0.17219258844852448,
0.06577219814062119,
-0.038416244089603424,
0.006410336587578058,
0.02959546446800232,
0.08237514644861221,
0.23417828977108002,
-0.06035616248846054,
0.1364797055721283,
-0.03420931473374367,
-0.14982740581035614,
0.08169995993375778,
0.0713929831981659,
0.10213395953178406,
0.06678459793329239,
0.0804823637008667,
0.0149586396291852,
0.06188136339187622,
0.1311223804950714,
0.08191446959972382,
0.019586285576224327,
-0.02480296604335308,
-0.03388110175728798,
-0.025523077696561813,
-0.05937909707427025,
0.040128443390131,
0.06589099019765854,
-0.16763372719287872,
-0.039227183908224106,
-0.09338314831256866,
0.09657008945941925,
0.0873042419552803,
0.06609832495450974,
-0.1842060089111328,
-0.008006223477423191,
0.08488986641168594,
-0.03854905813932419,
-0.13727426528930664,
0.09535189718008041,
0.01523482333868742,
-0.15144726634025574,
0.03139317408204079,
-0.04061909019947052,
0.12188644707202911,
-0.07804752141237259,
0.09809603542089462,
-0.08108244836330414,
-0.07448557764291763,
0.02123199962079525,
0.1261177361011505,
-0.30527687072753906,
0.20240111649036407,
-0.0024993624538183212,
-0.06486981362104416,
-0.1243603527545929,
-0.0032166161108762026,
0.002410882618278265,
0.07357452809810638,
0.10519039630889893,
-0.007196315098553896,
0.001897757756523788,
-0.06300821900367737,
-0.01829923689365387,
0.032471053302288055,
0.13080233335494995,
-0.0401318334043026,
-0.021158374845981598,
-0.050194524228572845,
-0.001653497340157628,
-0.03173094615340233,
-0.06934895366430283,
0.02002747356891632,
-0.19509181380271912,
0.08751901984214783,
0.04166261479258537,
0.09648149460554123,
0.029994789510965347,
0.004265148192644119,
-0.09651939570903778,
0.24698667228221893,
-0.07148019969463348,
-0.10072879493236542,
-0.10919588059186935,
-0.046813901513814926,
0.03569883480668068,
-0.05628936365246773,
0.04309194162487984,
-0.0788632407784462,
0.028997479006648064,
-0.06352769583463669,
-0.19235502183437347,
0.12410202622413635,
-0.09027006477117538,
-0.04412810131907463,
-0.02371402643620968,
0.2110891044139862,
-0.05598580464720726,
0.010335659608244896,
0.02930437959730625,
0.01208863127976656,
-0.11645778268575668,
-0.09678568691015244,
0.031018631532788277,
-0.007351789623498917,
0.050603240728378296,
0.041841957718133926,
-0.05915454775094986,
-0.017138581722974777,
-0.052199993282556534,
-0.022926922887563705,
0.3496883809566498,
0.14231905341148376,
-0.043836336582899094,
0.19347235560417175,
0.12347975373268127,
-0.07452994585037231,
-0.3159443140029907,
-0.1066238060593605,
-0.10937739163637161,
-0.04680149629712105,
-0.07012093812227249,
-0.2002030611038208,
0.06474938243627548,
0.00662544509395957,
-0.013415241613984108,
0.12749312818050385,
-0.2561831772327423,
-0.07571036368608475,
0.15906259417533875,
-0.017980827018618584,
0.3745945692062378,
-0.1168576180934906,
-0.10926306992769241,
-0.03950892388820648,
-0.14175476133823395,
0.16968177258968353,
-0.01989765651524067,
0.11221715062856674,
-0.009765521623194218,
0.14388824999332428,
0.05548359826207161,
-0.023479344323277473,
0.08544106781482697,
0.004999885335564613,
-0.03290518373250961,
-0.10304180532693863,
-0.05676887184381485,
0.007092386484146118,
0.02477436140179634,
0.018026655539870262,
-0.041834570467472076,
0.02227151393890381,
-0.11731979995965958,
-0.04657655209302902,
-0.08982590585947037,
0.04431166127324104,
0.03899754583835602,
-0.07325074821710587,
-0.002380647463724017,
-0.07165111601352692,
-0.012272949330508709,
0.022334342822432518,
0.20356793701648712,
-0.08029330521821976,
0.16448934376239777,
0.09239562600851059,
0.12419285625219345,
-0.14376309514045715,
-0.00019283240544609725,
-0.0762530043721199,
-0.05611240118741989,
0.07737895101308823,
-0.09433035552501678,
0.058893077075481415,
0.10901971161365509,
-0.04567738622426987,
0.08828683942556381,
0.10377411544322968,
0.008936077356338501,
0.003213887568563223,
0.10916902124881744,
-0.2667325437068939,
-0.0296600554138422,
-0.07532413303852081,
0.000883326749317348,
0.09092561900615692,
0.08562852442264557,
0.18840822577476501,
0.025361526757478714,
-0.04293036088347435,
-0.002770674182102084,
0.028597986325621605,
-0.039021048694849014,
0.051667019724845886,
0.001123449532315135,
0.01947369985282421,
-0.1530752182006836,
0.072522833943367,
0.01490565575659275,
-0.15215420722961426,
0.021316176280379295,
0.16572684049606323,
-0.11656328290700912,
-0.1283872276544571,
-0.06520111113786697,
0.08313824236392975,
-0.11755692958831787,
-0.01578943058848381,
-0.03279297426342964,
-0.13145680725574493,
0.07992171496152878,
0.12629036605358124,
0.05557859688997269,
0.0972496047616005,
-0.06061713397502899,
-0.020469192415475845,
-0.018721895292401314,
-0.014099318534135818,
-0.012384648434817791,
-0.007667020428925753,
-0.055978111922740936,
0.0590752474963665,
-0.026677248999476433,
0.1425808072090149,
-0.09221141785383224,
-0.1037059873342514,
-0.16142144799232483,
0.0374140702188015,
-0.11013076454401016,
-0.08825794607400894,
-0.08821134269237518,
-0.050188567489385605,
0.002360827289521694,
-0.019856395199894905,
-0.04037635400891304,
-0.05829505994915962,
-0.12300454825162888,
0.0338277705013752,
-0.040771447122097015,
0.024727050215005875,
-0.07512269169092178,
0.015856385231018066,
0.08507686108350754,
-0.03285100311040878,
0.15655414760112762,
0.1450488418340683,
-0.1006515845656395,
0.10741901397705078,
-0.14806775748729706,
-0.09138492494821548,
0.11116421222686768,
0.015329592861235142,
0.0449691042304039,
0.09723787009716034,
0.013362943194806576,
0.0635865181684494,
0.032776717096567154,
0.05308786407113075,
0.027619892731308937,
-0.11959987878799438,
0.06483134627342224,
-0.03626115620136261,
-0.14700546860694885,
-0.049338050186634064,
-0.05282869189977646,
0.01647452637553215,
0.013054544106125832,
0.09622690081596375,
-0.05301849544048309,
0.10698331147432327,
-0.04055701196193695,
0.0346808135509491,
0.017554637044668198,
-0.1730053424835205,
-0.03816922754049301,
-0.08538098633289337,
0.03681723028421402,
0.014741539023816586,
0.25266793370246887,
0.030072299763560295,
0.012416383251547813,
0.032671261578798294,
0.08285367488861084,
0.03899408504366875,
0.010228337720036507,
0.17482228577136993,
0.1162426546216011,
-0.06621865928173065,
-0.10445023328065872,
0.0729617029428482,
0.016332454979419708,
0.01286179106682539,
0.13617953658103943,
0.008365051820874214,
0.005795429926365614,
0.08649782836437225,
-0.016865963116288185,
0.009968153201043606,
-0.10052056610584259,
-0.13426925241947174,
-0.022176474332809448,
0.05151832848787308,
-0.04655967652797699,
0.11727844923734665,
0.1406494379043579,
-0.01806013658642769,
0.03222079202532768,
-0.021771740168333054,
-0.05699979141354561,
-0.1683429479598999,
-0.1429590880870819,
-0.06883849948644638,
-0.13416796922683716,
0.00897989235818386,
-0.11180389672517776,
0.05395037308335304,
0.06001098081469536,
0.06750501692295074,
-0.06899319589138031,
0.10220931470394135,
0.04626858979463577,
-0.11440542340278625,
0.06264589726924896,
-0.0296088308095932,
0.09430401772260666,
-0.02759445086121559,
-0.019505485892295837,
-0.09039592742919922,
0.014574515633285046,
0.011419114656746387,
0.06245238706469536,
-0.04707273095846176,
0.007463190704584122,
-0.14696238934993744,
-0.08972041308879852,
-0.0523175448179245,
0.0718572810292244,
-0.050409089773893356,
0.14282815158367157,
0.00775480642914772,
-0.0170906875282526,
0.039554283022880554,
0.22787313163280487,
-0.07476283609867096,
-0.04778539761900902,
-0.05269690603017807,
0.20717895030975342,
0.02975541539490223,
0.1171872541308403,
-0.022938819602131844,
-0.006106364540755749,
-0.0919521227478981,
0.3764844834804535,
0.30030161142349243,
-0.09031439572572708,
0.011794124729931355,
0.02137952297925949,
0.04502861574292183,
0.1316293478012085,
0.1216534823179245,
0.10318691283464432,
0.3006802201271057,
-0.07452366501092911,
-0.04653361067175865,
-0.012629742734134197,
-0.023858042433857918,
-0.09059546142816544,
0.1021224707365036,
0.04839762672781944,
-0.06382183730602264,
-0.03313443064689636,
0.0954432487487793,
-0.25862133502960205,
0.1277991235256195,
-0.12311873584985733,
-0.17578600347042084,
-0.06654827296733856,
0.009760108776390553,
0.10465722531080246,
0.015642458572983742,
0.0946015790104866,
0.007128213066607714,
-0.11252258718013763,
0.06305865943431854,
0.03397420793771744,
-0.22762253880500793,
0.0006893770187161863,
0.06642123311758041,
-0.07006710022687912,
-0.0024247700348496437,
-0.026499588042497635,
0.05657242611050606,
0.0656052976846695,
0.054629553109407425,
-0.00971333310008049,
0.03816632181406021,
0.0034184439573436975,
-0.0585215799510479,
0.016623929142951965,
0.05121519789099693,
0.02472509816288948,
-0.09763528406620026,
0.06927435845136642,
-0.1574270874261856,
0.04766253009438515,
-0.0030655991286039352,
-0.04124255105853081,
0.006064958870410919,
0.008823691867291927,
-0.06491616368293762,
0.05165379121899605,
0.07916834205389023,
-0.0016257909592241049,
-0.0062433634884655476,
-0.057178743183612823,
-0.02632102556526661,
-0.027755750343203545,
-0.09291748702526093,
-0.10495562851428986,
-0.14682936668395996,
-0.11640441417694092,
0.09368976950645447,
-0.01011267676949501,
-0.1848134547472,
0.022154374048113823,
-0.08606051653623581,
0.08319322764873505,
-0.1670055389404297,
0.08040720224380493,
0.07041648775339127,
0.013038921169936657,
-0.0031511052511632442,
-0.02002427540719509,
0.054132770746946335,
0.086809903383255,
-0.10407156497240067,
-0.07400695979595184
] |
null | null |
transformers
|
This the repo for the final project
|
{}
|
text-classification
|
dhairya2303/bert-base-uncased-emotion-AD
|
[
"transformers",
"tf",
"distilbert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #tf #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
This the repo for the final project
|
[] |
[
"TAGS\n#transformers #tf #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
37
] |
[
"passage: TAGS\n#transformers #tf #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.013508454896509647,
0.01568102464079857,
-0.00609795656055212,
0.03628997132182121,
0.1953449249267578,
0.04532367363572121,
0.057361066341400146,
0.10917055606842041,
0.02398945949971676,
-0.044445574283599854,
0.08713234961032867,
0.20914500951766968,
-0.040614817291498184,
0.1613861471414566,
-0.15467429161071777,
-0.2831328809261322,
0.051150787621736526,
0.05726104974746704,
-0.04057672992348671,
0.08529701828956604,
0.10851805657148361,
-0.03678816184401512,
0.09065315872430801,
-0.0628233551979065,
-0.17547662556171417,
0.06781362742185593,
0.0796811431646347,
-0.13586737215518951,
0.08911304175853729,
0.11890596151351929,
0.13617494702339172,
0.05271866172552109,
-0.06458207219839096,
-0.12120760977268219,
0.03316290304064751,
0.023270748555660248,
-0.11652618646621704,
0.014068183489143848,
0.05908108875155449,
-0.10618927329778671,
0.04767508804798126,
0.020873302593827248,
0.026440855115652084,
0.08891870081424713,
-0.16232652962207794,
-0.052880119532346725,
-0.014139981009066105,
0.006462214980274439,
0.07680805027484894,
0.040071625262498856,
0.009075193665921688,
0.11527775228023529,
-0.1019601970911026,
0.15828937292099,
0.05298613756895065,
-0.289473295211792,
-0.024283578619360924,
0.10820731520652771,
-0.004882104694843292,
0.03612781688570976,
-0.04864448308944702,
0.061358578503131866,
0.05920802801847458,
0.005465460009872913,
0.013350129127502441,
-0.08234354108572006,
-0.14921057224273682,
0.03746598958969116,
-0.07268105447292328,
-0.002507627010345459,
0.24175207316875458,
-0.027337027713656425,
0.03962808474898338,
-0.009774531237781048,
-0.10474999248981476,
-0.04305870831012726,
-0.030359676107764244,
-0.007942020893096924,
-0.05029444396495819,
0.09680100530385971,
0.02944670058786869,
0.00439244881272316,
-0.09574981033802032,
0.025957435369491577,
-0.2515367269515991,
0.2098863124847412,
-0.025372127071022987,
0.040052130818367004,
-0.18087999522686005,
0.021922048181295395,
-0.02753911353647709,
-0.0994853526353836,
0.034247275441884995,
-0.10477811843156815,
-0.0554308146238327,
-0.08534710854291916,
-0.07143861800432205,
-0.09669230133295059,
0.07856198400259018,
0.168414905667305,
0.05522601306438446,
0.0797770693898201,
-0.09113283455371857,
0.04874400049448013,
0.006306092720478773,
0.11841475963592529,
0.025315528735518456,
-0.06898961216211319,
0.0387662835419178,
-0.19771695137023926,
-0.04732300341129303,
-0.07910472899675369,
-0.1652696132659912,
-0.030910253524780273,
0.02597956545650959,
0.06652552634477615,
-0.01463855430483818,
0.12003534287214279,
-0.0476749986410141,
-0.02694217674434185,
0.08320073038339615,
-0.08629371970891953,
0.007508325856178999,
-0.014767996966838837,
0.05017826706171036,
0.10010749846696854,
0.0066520944237709045,
-0.005964980460703373,
-0.04961960390210152,
0.10882893949747086,
-0.05787031352519989,
-0.03086760826408863,
-0.003183632390573621,
-0.09529116749763489,
0.02792106568813324,
-0.14246568083763123,
0.038099732249975204,
-0.21604011952877045,
-0.11064969003200531,
0.01951693929731846,
0.01742374338209629,
-0.024007802829146385,
0.0004822379269171506,
-0.023570753633975983,
-0.022180018946528435,
0.07859465479850769,
-0.035012587904930115,
-0.08296900242567062,
-0.07271513342857361,
0.06613212823867798,
-0.031488850712776184,
0.08630204945802689,
-0.18387259542942047,
0.062023237347602844,
-0.053042855113744736,
-0.017423711717128754,
-0.16101795434951782,
0.08657034486532211,
-0.04926241934299469,
0.20680339634418488,
-0.011387261562049389,
-0.015159567818045616,
-0.09665347635746002,
0.05723804235458374,
-0.08194683492183685,
0.16121000051498413,
-0.16026687622070312,
-0.08520128577947617,
0.198213130235672,
-0.08919282257556915,
-0.15341277420520782,
0.08544854074716568,
-0.005501038394868374,
0.05180487409234047,
0.10100556910037994,
0.18184180557727814,
0.14576639235019684,
-0.0018489942885935307,
0.11589036136865616,
0.15596067905426025,
-0.1237921342253685,
-0.05719515308737755,
-0.028102435171604156,
-0.0012073073303326964,
-0.15218110382556915,
0.0374184213578701,
0.12089990824460983,
0.10220547765493393,
-0.03312598913908005,
-0.0220076534897089,
-0.026965966448187828,
-0.013442243449389935,
0.12496168911457062,
0.02706170082092285,
0.10599066317081451,
-0.08803289383649826,
0.02733398601412773,
0.014947229996323586,
-0.03584874048829079,
0.029909558594226837,
0.02128968946635723,
-0.09099870920181274,
0.05802462622523308,
0.04277497157454491,
0.04197821393609047,
-0.21246756613254547,
-0.12722401320934296,
-0.008452661335468292,
0.20811936259269714,
-0.005059750284999609,
0.15012092888355255,
0.05428575724363327,
-0.07239655405282974,
-0.03332066535949707,
0.004370018839836121,
0.21119606494903564,
0.06059196591377258,
-0.04272458702325821,
-0.07458359748125076,
0.11109868437051773,
-0.07892254739999771,
-0.012811440043151379,
-0.0899607315659523,
0.005261971615254879,
0.1559956669807434,
0.13767196238040924,
0.06369031965732574,
0.07607471942901611,
-0.03256765753030777,
0.04073195531964302,
-0.08026492595672607,
-0.020297419279813766,
0.0683383122086525,
0.0026302821934223175,
-0.0981755405664444,
0.17747555673122406,
-0.1418410837650299,
0.29875147342681885,
0.18524940311908722,
-0.2499033361673355,
-0.06694182008504868,
-0.010180915705859661,
0.003639934351667762,
0.03089190647006035,
0.056414928287267685,
-0.029718779027462006,
0.03129078075289726,
-0.035941317677497864,
0.19720226526260376,
-0.020803719758987427,
-0.045986659824848175,
-0.0003005715843755752,
-0.03134095296263695,
-0.05296753719449043,
0.05186139792203903,
0.021421363577246666,
-0.28361421823501587,
0.1837073713541031,
0.21367208659648895,
0.06747549027204514,
0.19717411696910858,
-0.023349767550826073,
0.032182417809963226,
0.07614444196224213,
-0.00721743842586875,
0.014120814390480518,
-0.07831474393606186,
-0.16655080020427704,
-0.03330739587545395,
0.06418252736330032,
0.041391521692276,
0.05133110284805298,
-0.06914188712835312,
-0.02246452495455742,
0.033258106559515,
0.00031968732946552336,
0.009242691099643707,
0.11255364120006561,
0.05137859284877777,
0.1214650347828865,
-0.002437328454107046,
-0.07930290699005127,
0.13356061279773712,
0.00392551813274622,
-0.11512082815170288,
0.1692608743906021,
-0.1842028796672821,
-0.3308919668197632,
-0.1526893824338913,
-0.22108326852321625,
-0.019868751987814903,
0.08083143830299377,
0.10041883587837219,
-0.1170768067240715,
-0.07321483641862869,
0.014932911843061447,
-0.014058683067560196,
-0.03817500174045563,
0.07387414574623108,
-0.029820261523127556,
0.07477542012929916,
-0.026865674182772636,
-0.06370266526937485,
-0.06365495175123215,
0.024282041937112808,
0.018165770918130875,
0.1508738398551941,
-0.17709611356258392,
0.1002909243106842,
0.18546098470687866,
-0.022319326177239418,
0.07522179931402206,
-0.05805611610412598,
0.1956574022769928,
-0.11384128034114838,
0.024700522422790527,
0.14905671775341034,
-0.08426613360643387,
0.037691060453653336,
0.16327105462551117,
0.0007392764091491699,
-0.10385708510875702,
0.062117937952280045,
-0.017467541620135307,
-0.10519344359636307,
-0.19071146845817566,
-0.13467933237552643,
-0.14553125202655792,
0.06752705574035645,
0.04316609352827072,
0.08374269306659698,
0.1757919192314148,
0.04966075345873833,
0.027808915823698044,
0.06840041279792786,
0.02496904879808426,
0.0721738263964653,
0.22385257482528687,
0.009884152561426163,
0.12436345964670181,
-0.08571165055036545,
-0.1115255206823349,
0.12313172966241837,
-0.013214384205639362,
0.08387532830238342,
0.11431549489498138,
0.04639727994799614,
-0.007122491020709276,
0.0009732297039590776,
0.14959679543972015,
0.14206723868846893,
0.050233665853738785,
-0.04126787930727005,
-0.014598359353840351,
-0.0007573743932880461,
-0.02801721915602684,
0.038556694984436035,
0.03322141617536545,
-0.1740269511938095,
-0.05518098548054695,
-0.08605959266424179,
0.10463382303714752,
0.0795227512717247,
0.06589583307504654,
-0.20518141984939575,
0.002359851496294141,
0.08499189466238022,
-0.038736362010240555,
-0.10779839009046555,
0.06420190632343292,
0.008829314261674881,
-0.10321372002363205,
0.14975306391716003,
-0.012890004552900791,
0.11967772990465164,
-0.010838047601282597,
0.07147050648927689,
0.00751267047598958,
-0.12258655577898026,
-0.0017241573659703135,
0.0839301198720932,
-0.33603790402412415,
0.1943800151348114,
0.02909170091152191,
-0.07050973922014236,
-0.0717535987496376,
-0.01630263775587082,
0.03718230873346329,
0.2610548436641693,
0.09166548401117325,
0.0030416660010814667,
-0.08844857662916183,
-0.16589483618736267,
0.0369446724653244,
-0.016368385404348373,
0.15208597481250763,
-0.017282336950302124,
-0.025128155946731567,
-0.04848826676607132,
-0.03570812940597534,
0.010659679770469666,
-0.016330577433109283,
-0.010610956698656082,
-0.17699725925922394,
0.059734705835580826,
0.037856653332710266,
0.04736451059579849,
0.01057372149080038,
0.006709715351462364,
-0.12058372050523758,
0.19446995854377747,
-0.07625894248485565,
-0.05350392684340477,
-0.17218469083309174,
-0.06769312918186188,
0.009960283525288105,
-0.06504183262586594,
0.07818499207496643,
-0.06261719763278961,
0.03552405908703804,
-0.045271046459674835,
-0.24275410175323486,
0.14483092725276947,
-0.10551480203866959,
-0.008752617053687572,
-0.06438622623682022,
0.10718917846679688,
-0.09611298888921738,
0.004043095279484987,
0.047150805592536926,
-0.0020015020854771137,
-0.003136041574180126,
-0.0753227025270462,
-0.02360936626791954,
-0.023382848128676414,
0.010927858762443066,
0.05627825856208801,
-0.09057381749153137,
-0.058981698006391525,
-0.011360466480255127,
0.011085661128163338,
0.26996174454689026,
0.13272280991077423,
-0.08098062127828598,
0.1497524082660675,
0.04212837293744087,
-0.055468618869781494,
-0.32886067032814026,
-0.032638851553201675,
-0.10844086110591888,
-0.03573088347911835,
-0.0315665677189827,
-0.11909491568803787,
0.11200772225856781,
-0.037008412182331085,
0.017459701746702194,
0.09829957038164139,
-0.09895294159650803,
-0.09962201863527298,
0.1813213974237442,
-0.03681996464729309,
0.3421141803264618,
-0.10857312381267548,
-0.09756923466920853,
-0.09042752534151077,
-0.13268981873989105,
0.13568423688411713,
-0.05275978520512581,
0.0630483627319336,
0.02627634070813656,
0.006145235151052475,
0.0446162186563015,
-0.0005159254651516676,
0.08266224712133408,
0.034801676869392395,
0.0479268953204155,
-0.13571010529994965,
-0.06266872584819794,
0.05135650187730789,
-0.018533457070589066,
0.00578203285112977,
0.010653017088770866,
0.022098930552601814,
-0.14515259861946106,
-0.03575956076383591,
-0.043898191303014755,
0.0467614009976387,
0.04690644145011902,
-0.05771426856517792,
-0.00885720830410719,
-0.010712144896388054,
0.027221351861953735,
-0.041209906339645386,
0.21780996024608612,
-0.0649494156241417,
0.16645391285419464,
0.1081744134426117,
0.17357978224754333,
-0.18081913888454437,
0.07732399553060532,
-0.04428189620375633,
-0.06327948719263077,
0.05086290091276169,
-0.11531749367713928,
0.1052265614271164,
0.09898768365383148,
-0.06300456076860428,
0.10326434671878815,
0.11595330387353897,
0.06910184770822525,
-0.04703761264681816,
0.18756987154483795,
-0.2050054669380188,
0.02067248709499836,
-0.08906055986881256,
-0.09570451080799103,
0.08420304208993912,
0.038182929158210754,
0.15054930746555328,
0.05989672616124153,
0.008429978042840958,
0.013632940128445625,
-0.03493218123912811,
-0.02772756852209568,
0.0648454800248146,
0.06808432936668396,
0.024564260616898537,
-0.1321037858724594,
0.049559302628040314,
0.03240500018000603,
-0.13265642523765564,
0.0041791340336203575,
0.11882128566503525,
-0.17804211378097534,
-0.11308864504098892,
0.01529060211032629,
0.1529897153377533,
-0.1332138031721115,
-0.03977452963590622,
-0.04502803087234497,
-0.135286346077919,
0.07461816817522049,
0.2978265881538391,
0.08529166132211685,
0.12402704358100891,
-0.03854496777057648,
-0.041670091450214386,
0.010444069281220436,
-0.014487391337752342,
-0.0010834430577233434,
0.049138616770505905,
-0.15437109768390656,
0.07335909456014633,
-0.07622494548559189,
0.14631399512290955,
-0.11269695311784744,
-0.03881227225065231,
-0.17946462333202362,
0.008294780738651752,
-0.11636082828044891,
-0.02548850141465664,
-0.0639713779091835,
-0.004650413990020752,
0.028838710859417915,
-0.03475451469421387,
-0.040130529552698135,
-0.06374537944793701,
-0.11489926278591156,
0.03965337574481964,
-0.008941588923335075,
0.04428498446941376,
-0.038529932498931885,
-0.05404394865036011,
0.06395529210567474,
-0.03798069432377815,
0.10097184032201767,
0.08092007040977478,
-0.09754543751478195,
0.1077558621764183,
-0.15613099932670593,
-0.11616120487451553,
0.14566670358181,
0.017509503290057182,
0.09954837709665298,
0.09593215584754944,
0.03367144241929054,
0.05662086606025696,
-0.0008124340092763305,
0.07030697911977768,
0.03926016017794609,
-0.08998897671699524,
0.04330950602889061,
-0.09044003486633301,
-0.15327788889408112,
-0.05554389953613281,
-0.024580324068665504,
0.1306833177804947,
-0.0024173702113330364,
0.13051512837409973,
-0.06636594235897064,
0.06144222989678383,
-0.029070906341075897,
-0.01134984940290451,
-0.003905582008883357,
-0.1881074756383896,
-0.031628355383872986,
-0.07348880171775818,
0.022182034328579903,
-0.002208020770922303,
0.2120659500360489,
0.04402490332722664,
0.04191233962774277,
0.06991370022296906,
0.03089935891330242,
0.02009156346321106,
0.015833299607038498,
0.21341410279273987,
0.09542709589004517,
-0.05260615050792694,
-0.11565204709768295,
0.05816751718521118,
0.03327472135424614,
-0.014356084167957306,
0.13757623732089996,
0.03764916583895683,
-0.1644933968782425,
0.0860944613814354,
-0.0358855314552784,
0.007054685149341822,
-0.07318374514579773,
-0.12752597033977509,
-0.07328153401613235,
0.07694418728351593,
0.011888375505805016,
-0.049794506281614304,
0.11003571003675461,
-0.0604579858481884,
0.05341075733304024,
-0.03574401140213013,
-0.05464261770248413,
-0.18404990434646606,
-0.09398619085550308,
-0.11046914011240005,
-0.1234632134437561,
0.012381363660097122,
-0.09565775096416473,
0.026443319395184517,
0.014370107091963291,
0.06383217871189117,
-0.026314040645956993,
0.11228405684232712,
-0.0700870230793953,
-0.07115274667739868,
0.10943242907524109,
-0.04176071658730507,
0.021765414625406265,
0.005539920646697283,
-0.032465945929288864,
-0.12855559587478638,
-0.008983265608549118,
-0.08348371088504791,
0.031226472929120064,
-0.009573068469762802,
0.009853784926235676,
-0.15051385760307312,
-0.0967787504196167,
-0.010985838249325752,
0.06228780746459961,
-0.09249438345432281,
0.10734081268310547,
0.02957982011139393,
-0.002639044774696231,
0.05634211748838425,
0.16844052076339722,
-0.06863506883382797,
-0.06949624419212341,
-0.11427353322505951,
0.19616612792015076,
0.07560871541500092,
0.1447131335735321,
-0.02807728759944439,
-0.032596442848443985,
-0.07579296082258224,
0.29457905888557434,
0.23213930428028107,
-0.05629273131489754,
0.05847287178039551,
0.0012319441884756088,
0.03630107641220093,
0.15130659937858582,
0.17220762372016907,
0.051916591823101044,
0.18380406498908997,
-0.012772094458341599,
-0.08656375855207443,
-0.0013967609265819192,
-0.02520974911749363,
-0.09315089881420135,
0.11861637979745865,
0.07236126065254211,
-0.0418969951570034,
-0.038926269859075546,
0.12678371369838715,
-0.18100608885288239,
0.13916277885437012,
0.01579192839562893,
-0.17710335552692413,
-0.051634080708026886,
-0.053932636976242065,
0.046283528208732605,
-0.014388833194971085,
0.06573020666837692,
-0.0140520129352808,
-0.1115402802824974,
0.022399915382266045,
0.024362994357943535,
-0.22000715136528015,
0.013517675921320915,
0.04566873982548714,
-0.0733935683965683,
-0.02704431861639023,
-0.03492184355854988,
-0.005227707792073488,
0.07885763049125671,
0.07570509612560272,
-0.0036116221453994513,
0.06791701167821884,
-0.017412476241588593,
-0.005500115919858217,
0.025286540389060974,
0.04461054503917694,
-0.008078323677182198,
-0.06506017595529556,
0.07267521321773529,
-0.2035190761089325,
0.05389382690191269,
-0.01073064561933279,
-0.06632687896490097,
-0.02661988139152527,
0.030778739601373672,
-0.08384375274181366,
0.07572232186794281,
0.12706787884235382,
0.01046997494995594,
0.016810817644000053,
-0.044822704046964645,
-0.02993215247988701,
0.015644192695617676,
-0.12733115255832672,
-0.13652819395065308,
-0.07370613515377045,
-0.09697027504444122,
0.08596285432577133,
-0.009802427142858505,
-0.13083021342754364,
-0.011835134588181973,
-0.11618591845035553,
0.07142731547355652,
-0.15336956083774567,
0.144621342420578,
0.083065465092659,
0.021282320842146873,
-0.017649799585342407,
-0.08638981729745865,
0.05515220761299133,
0.08350114524364471,
-0.10982725024223328,
-0.07850786298513412
] |
null | null |
transformers
|
{'sadness':0,'joy':1,'love':2,'anger':3,'fear':4,'surprise':5}
|
{}
|
text-classification
|
dhairya2303/bert-base-uncased-emotion_holler
|
[
"transformers",
"tf",
"distilbert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #tf #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
{'sadness':0,'joy':1,'love':2,'anger':3,'fear':4,'surprise':5}
|
[] |
[
"TAGS\n#transformers #tf #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
37
] |
[
"passage: TAGS\n#transformers #tf #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.013508454896509647,
0.01568102464079857,
-0.00609795656055212,
0.03628997132182121,
0.1953449249267578,
0.04532367363572121,
0.057361066341400146,
0.10917055606842041,
0.02398945949971676,
-0.044445574283599854,
0.08713234961032867,
0.20914500951766968,
-0.040614817291498184,
0.1613861471414566,
-0.15467429161071777,
-0.2831328809261322,
0.051150787621736526,
0.05726104974746704,
-0.04057672992348671,
0.08529701828956604,
0.10851805657148361,
-0.03678816184401512,
0.09065315872430801,
-0.0628233551979065,
-0.17547662556171417,
0.06781362742185593,
0.0796811431646347,
-0.13586737215518951,
0.08911304175853729,
0.11890596151351929,
0.13617494702339172,
0.05271866172552109,
-0.06458207219839096,
-0.12120760977268219,
0.03316290304064751,
0.023270748555660248,
-0.11652618646621704,
0.014068183489143848,
0.05908108875155449,
-0.10618927329778671,
0.04767508804798126,
0.020873302593827248,
0.026440855115652084,
0.08891870081424713,
-0.16232652962207794,
-0.052880119532346725,
-0.014139981009066105,
0.006462214980274439,
0.07680805027484894,
0.040071625262498856,
0.009075193665921688,
0.11527775228023529,
-0.1019601970911026,
0.15828937292099,
0.05298613756895065,
-0.289473295211792,
-0.024283578619360924,
0.10820731520652771,
-0.004882104694843292,
0.03612781688570976,
-0.04864448308944702,
0.061358578503131866,
0.05920802801847458,
0.005465460009872913,
0.013350129127502441,
-0.08234354108572006,
-0.14921057224273682,
0.03746598958969116,
-0.07268105447292328,
-0.002507627010345459,
0.24175207316875458,
-0.027337027713656425,
0.03962808474898338,
-0.009774531237781048,
-0.10474999248981476,
-0.04305870831012726,
-0.030359676107764244,
-0.007942020893096924,
-0.05029444396495819,
0.09680100530385971,
0.02944670058786869,
0.00439244881272316,
-0.09574981033802032,
0.025957435369491577,
-0.2515367269515991,
0.2098863124847412,
-0.025372127071022987,
0.040052130818367004,
-0.18087999522686005,
0.021922048181295395,
-0.02753911353647709,
-0.0994853526353836,
0.034247275441884995,
-0.10477811843156815,
-0.0554308146238327,
-0.08534710854291916,
-0.07143861800432205,
-0.09669230133295059,
0.07856198400259018,
0.168414905667305,
0.05522601306438446,
0.0797770693898201,
-0.09113283455371857,
0.04874400049448013,
0.006306092720478773,
0.11841475963592529,
0.025315528735518456,
-0.06898961216211319,
0.0387662835419178,
-0.19771695137023926,
-0.04732300341129303,
-0.07910472899675369,
-0.1652696132659912,
-0.030910253524780273,
0.02597956545650959,
0.06652552634477615,
-0.01463855430483818,
0.12003534287214279,
-0.0476749986410141,
-0.02694217674434185,
0.08320073038339615,
-0.08629371970891953,
0.007508325856178999,
-0.014767996966838837,
0.05017826706171036,
0.10010749846696854,
0.0066520944237709045,
-0.005964980460703373,
-0.04961960390210152,
0.10882893949747086,
-0.05787031352519989,
-0.03086760826408863,
-0.003183632390573621,
-0.09529116749763489,
0.02792106568813324,
-0.14246568083763123,
0.038099732249975204,
-0.21604011952877045,
-0.11064969003200531,
0.01951693929731846,
0.01742374338209629,
-0.024007802829146385,
0.0004822379269171506,
-0.023570753633975983,
-0.022180018946528435,
0.07859465479850769,
-0.035012587904930115,
-0.08296900242567062,
-0.07271513342857361,
0.06613212823867798,
-0.031488850712776184,
0.08630204945802689,
-0.18387259542942047,
0.062023237347602844,
-0.053042855113744736,
-0.017423711717128754,
-0.16101795434951782,
0.08657034486532211,
-0.04926241934299469,
0.20680339634418488,
-0.011387261562049389,
-0.015159567818045616,
-0.09665347635746002,
0.05723804235458374,
-0.08194683492183685,
0.16121000051498413,
-0.16026687622070312,
-0.08520128577947617,
0.198213130235672,
-0.08919282257556915,
-0.15341277420520782,
0.08544854074716568,
-0.005501038394868374,
0.05180487409234047,
0.10100556910037994,
0.18184180557727814,
0.14576639235019684,
-0.0018489942885935307,
0.11589036136865616,
0.15596067905426025,
-0.1237921342253685,
-0.05719515308737755,
-0.028102435171604156,
-0.0012073073303326964,
-0.15218110382556915,
0.0374184213578701,
0.12089990824460983,
0.10220547765493393,
-0.03312598913908005,
-0.0220076534897089,
-0.026965966448187828,
-0.013442243449389935,
0.12496168911457062,
0.02706170082092285,
0.10599066317081451,
-0.08803289383649826,
0.02733398601412773,
0.014947229996323586,
-0.03584874048829079,
0.029909558594226837,
0.02128968946635723,
-0.09099870920181274,
0.05802462622523308,
0.04277497157454491,
0.04197821393609047,
-0.21246756613254547,
-0.12722401320934296,
-0.008452661335468292,
0.20811936259269714,
-0.005059750284999609,
0.15012092888355255,
0.05428575724363327,
-0.07239655405282974,
-0.03332066535949707,
0.004370018839836121,
0.21119606494903564,
0.06059196591377258,
-0.04272458702325821,
-0.07458359748125076,
0.11109868437051773,
-0.07892254739999771,
-0.012811440043151379,
-0.0899607315659523,
0.005261971615254879,
0.1559956669807434,
0.13767196238040924,
0.06369031965732574,
0.07607471942901611,
-0.03256765753030777,
0.04073195531964302,
-0.08026492595672607,
-0.020297419279813766,
0.0683383122086525,
0.0026302821934223175,
-0.0981755405664444,
0.17747555673122406,
-0.1418410837650299,
0.29875147342681885,
0.18524940311908722,
-0.2499033361673355,
-0.06694182008504868,
-0.010180915705859661,
0.003639934351667762,
0.03089190647006035,
0.056414928287267685,
-0.029718779027462006,
0.03129078075289726,
-0.035941317677497864,
0.19720226526260376,
-0.020803719758987427,
-0.045986659824848175,
-0.0003005715843755752,
-0.03134095296263695,
-0.05296753719449043,
0.05186139792203903,
0.021421363577246666,
-0.28361421823501587,
0.1837073713541031,
0.21367208659648895,
0.06747549027204514,
0.19717411696910858,
-0.023349767550826073,
0.032182417809963226,
0.07614444196224213,
-0.00721743842586875,
0.014120814390480518,
-0.07831474393606186,
-0.16655080020427704,
-0.03330739587545395,
0.06418252736330032,
0.041391521692276,
0.05133110284805298,
-0.06914188712835312,
-0.02246452495455742,
0.033258106559515,
0.00031968732946552336,
0.009242691099643707,
0.11255364120006561,
0.05137859284877777,
0.1214650347828865,
-0.002437328454107046,
-0.07930290699005127,
0.13356061279773712,
0.00392551813274622,
-0.11512082815170288,
0.1692608743906021,
-0.1842028796672821,
-0.3308919668197632,
-0.1526893824338913,
-0.22108326852321625,
-0.019868751987814903,
0.08083143830299377,
0.10041883587837219,
-0.1170768067240715,
-0.07321483641862869,
0.014932911843061447,
-0.014058683067560196,
-0.03817500174045563,
0.07387414574623108,
-0.029820261523127556,
0.07477542012929916,
-0.026865674182772636,
-0.06370266526937485,
-0.06365495175123215,
0.024282041937112808,
0.018165770918130875,
0.1508738398551941,
-0.17709611356258392,
0.1002909243106842,
0.18546098470687866,
-0.022319326177239418,
0.07522179931402206,
-0.05805611610412598,
0.1956574022769928,
-0.11384128034114838,
0.024700522422790527,
0.14905671775341034,
-0.08426613360643387,
0.037691060453653336,
0.16327105462551117,
0.0007392764091491699,
-0.10385708510875702,
0.062117937952280045,
-0.017467541620135307,
-0.10519344359636307,
-0.19071146845817566,
-0.13467933237552643,
-0.14553125202655792,
0.06752705574035645,
0.04316609352827072,
0.08374269306659698,
0.1757919192314148,
0.04966075345873833,
0.027808915823698044,
0.06840041279792786,
0.02496904879808426,
0.0721738263964653,
0.22385257482528687,
0.009884152561426163,
0.12436345964670181,
-0.08571165055036545,
-0.1115255206823349,
0.12313172966241837,
-0.013214384205639362,
0.08387532830238342,
0.11431549489498138,
0.04639727994799614,
-0.007122491020709276,
0.0009732297039590776,
0.14959679543972015,
0.14206723868846893,
0.050233665853738785,
-0.04126787930727005,
-0.014598359353840351,
-0.0007573743932880461,
-0.02801721915602684,
0.038556694984436035,
0.03322141617536545,
-0.1740269511938095,
-0.05518098548054695,
-0.08605959266424179,
0.10463382303714752,
0.0795227512717247,
0.06589583307504654,
-0.20518141984939575,
0.002359851496294141,
0.08499189466238022,
-0.038736362010240555,
-0.10779839009046555,
0.06420190632343292,
0.008829314261674881,
-0.10321372002363205,
0.14975306391716003,
-0.012890004552900791,
0.11967772990465164,
-0.010838047601282597,
0.07147050648927689,
0.00751267047598958,
-0.12258655577898026,
-0.0017241573659703135,
0.0839301198720932,
-0.33603790402412415,
0.1943800151348114,
0.02909170091152191,
-0.07050973922014236,
-0.0717535987496376,
-0.01630263775587082,
0.03718230873346329,
0.2610548436641693,
0.09166548401117325,
0.0030416660010814667,
-0.08844857662916183,
-0.16589483618736267,
0.0369446724653244,
-0.016368385404348373,
0.15208597481250763,
-0.017282336950302124,
-0.025128155946731567,
-0.04848826676607132,
-0.03570812940597534,
0.010659679770469666,
-0.016330577433109283,
-0.010610956698656082,
-0.17699725925922394,
0.059734705835580826,
0.037856653332710266,
0.04736451059579849,
0.01057372149080038,
0.006709715351462364,
-0.12058372050523758,
0.19446995854377747,
-0.07625894248485565,
-0.05350392684340477,
-0.17218469083309174,
-0.06769312918186188,
0.009960283525288105,
-0.06504183262586594,
0.07818499207496643,
-0.06261719763278961,
0.03552405908703804,
-0.045271046459674835,
-0.24275410175323486,
0.14483092725276947,
-0.10551480203866959,
-0.008752617053687572,
-0.06438622623682022,
0.10718917846679688,
-0.09611298888921738,
0.004043095279484987,
0.047150805592536926,
-0.0020015020854771137,
-0.003136041574180126,
-0.0753227025270462,
-0.02360936626791954,
-0.023382848128676414,
0.010927858762443066,
0.05627825856208801,
-0.09057381749153137,
-0.058981698006391525,
-0.011360466480255127,
0.011085661128163338,
0.26996174454689026,
0.13272280991077423,
-0.08098062127828598,
0.1497524082660675,
0.04212837293744087,
-0.055468618869781494,
-0.32886067032814026,
-0.032638851553201675,
-0.10844086110591888,
-0.03573088347911835,
-0.0315665677189827,
-0.11909491568803787,
0.11200772225856781,
-0.037008412182331085,
0.017459701746702194,
0.09829957038164139,
-0.09895294159650803,
-0.09962201863527298,
0.1813213974237442,
-0.03681996464729309,
0.3421141803264618,
-0.10857312381267548,
-0.09756923466920853,
-0.09042752534151077,
-0.13268981873989105,
0.13568423688411713,
-0.05275978520512581,
0.0630483627319336,
0.02627634070813656,
0.006145235151052475,
0.0446162186563015,
-0.0005159254651516676,
0.08266224712133408,
0.034801676869392395,
0.0479268953204155,
-0.13571010529994965,
-0.06266872584819794,
0.05135650187730789,
-0.018533457070589066,
0.00578203285112977,
0.010653017088770866,
0.022098930552601814,
-0.14515259861946106,
-0.03575956076383591,
-0.043898191303014755,
0.0467614009976387,
0.04690644145011902,
-0.05771426856517792,
-0.00885720830410719,
-0.010712144896388054,
0.027221351861953735,
-0.041209906339645386,
0.21780996024608612,
-0.0649494156241417,
0.16645391285419464,
0.1081744134426117,
0.17357978224754333,
-0.18081913888454437,
0.07732399553060532,
-0.04428189620375633,
-0.06327948719263077,
0.05086290091276169,
-0.11531749367713928,
0.1052265614271164,
0.09898768365383148,
-0.06300456076860428,
0.10326434671878815,
0.11595330387353897,
0.06910184770822525,
-0.04703761264681816,
0.18756987154483795,
-0.2050054669380188,
0.02067248709499836,
-0.08906055986881256,
-0.09570451080799103,
0.08420304208993912,
0.038182929158210754,
0.15054930746555328,
0.05989672616124153,
0.008429978042840958,
0.013632940128445625,
-0.03493218123912811,
-0.02772756852209568,
0.0648454800248146,
0.06808432936668396,
0.024564260616898537,
-0.1321037858724594,
0.049559302628040314,
0.03240500018000603,
-0.13265642523765564,
0.0041791340336203575,
0.11882128566503525,
-0.17804211378097534,
-0.11308864504098892,
0.01529060211032629,
0.1529897153377533,
-0.1332138031721115,
-0.03977452963590622,
-0.04502803087234497,
-0.135286346077919,
0.07461816817522049,
0.2978265881538391,
0.08529166132211685,
0.12402704358100891,
-0.03854496777057648,
-0.041670091450214386,
0.010444069281220436,
-0.014487391337752342,
-0.0010834430577233434,
0.049138616770505905,
-0.15437109768390656,
0.07335909456014633,
-0.07622494548559189,
0.14631399512290955,
-0.11269695311784744,
-0.03881227225065231,
-0.17946462333202362,
0.008294780738651752,
-0.11636082828044891,
-0.02548850141465664,
-0.0639713779091835,
-0.004650413990020752,
0.028838710859417915,
-0.03475451469421387,
-0.040130529552698135,
-0.06374537944793701,
-0.11489926278591156,
0.03965337574481964,
-0.008941588923335075,
0.04428498446941376,
-0.038529932498931885,
-0.05404394865036011,
0.06395529210567474,
-0.03798069432377815,
0.10097184032201767,
0.08092007040977478,
-0.09754543751478195,
0.1077558621764183,
-0.15613099932670593,
-0.11616120487451553,
0.14566670358181,
0.017509503290057182,
0.09954837709665298,
0.09593215584754944,
0.03367144241929054,
0.05662086606025696,
-0.0008124340092763305,
0.07030697911977768,
0.03926016017794609,
-0.08998897671699524,
0.04330950602889061,
-0.09044003486633301,
-0.15327788889408112,
-0.05554389953613281,
-0.024580324068665504,
0.1306833177804947,
-0.0024173702113330364,
0.13051512837409973,
-0.06636594235897064,
0.06144222989678383,
-0.029070906341075897,
-0.01134984940290451,
-0.003905582008883357,
-0.1881074756383896,
-0.031628355383872986,
-0.07348880171775818,
0.022182034328579903,
-0.002208020770922303,
0.2120659500360489,
0.04402490332722664,
0.04191233962774277,
0.06991370022296906,
0.03089935891330242,
0.02009156346321106,
0.015833299607038498,
0.21341410279273987,
0.09542709589004517,
-0.05260615050792694,
-0.11565204709768295,
0.05816751718521118,
0.03327472135424614,
-0.014356084167957306,
0.13757623732089996,
0.03764916583895683,
-0.1644933968782425,
0.0860944613814354,
-0.0358855314552784,
0.007054685149341822,
-0.07318374514579773,
-0.12752597033977509,
-0.07328153401613235,
0.07694418728351593,
0.011888375505805016,
-0.049794506281614304,
0.11003571003675461,
-0.0604579858481884,
0.05341075733304024,
-0.03574401140213013,
-0.05464261770248413,
-0.18404990434646606,
-0.09398619085550308,
-0.11046914011240005,
-0.1234632134437561,
0.012381363660097122,
-0.09565775096416473,
0.026443319395184517,
0.014370107091963291,
0.06383217871189117,
-0.026314040645956993,
0.11228405684232712,
-0.0700870230793953,
-0.07115274667739868,
0.10943242907524109,
-0.04176071658730507,
0.021765414625406265,
0.005539920646697283,
-0.032465945929288864,
-0.12855559587478638,
-0.008983265608549118,
-0.08348371088504791,
0.031226472929120064,
-0.009573068469762802,
0.009853784926235676,
-0.15051385760307312,
-0.0967787504196167,
-0.010985838249325752,
0.06228780746459961,
-0.09249438345432281,
0.10734081268310547,
0.02957982011139393,
-0.002639044774696231,
0.05634211748838425,
0.16844052076339722,
-0.06863506883382797,
-0.06949624419212341,
-0.11427353322505951,
0.19616612792015076,
0.07560871541500092,
0.1447131335735321,
-0.02807728759944439,
-0.032596442848443985,
-0.07579296082258224,
0.29457905888557434,
0.23213930428028107,
-0.05629273131489754,
0.05847287178039551,
0.0012319441884756088,
0.03630107641220093,
0.15130659937858582,
0.17220762372016907,
0.051916591823101044,
0.18380406498908997,
-0.012772094458341599,
-0.08656375855207443,
-0.0013967609265819192,
-0.02520974911749363,
-0.09315089881420135,
0.11861637979745865,
0.07236126065254211,
-0.0418969951570034,
-0.038926269859075546,
0.12678371369838715,
-0.18100608885288239,
0.13916277885437012,
0.01579192839562893,
-0.17710335552692413,
-0.051634080708026886,
-0.053932636976242065,
0.046283528208732605,
-0.014388833194971085,
0.06573020666837692,
-0.0140520129352808,
-0.1115402802824974,
0.022399915382266045,
0.024362994357943535,
-0.22000715136528015,
0.013517675921320915,
0.04566873982548714,
-0.0733935683965683,
-0.02704431861639023,
-0.03492184355854988,
-0.005227707792073488,
0.07885763049125671,
0.07570509612560272,
-0.0036116221453994513,
0.06791701167821884,
-0.017412476241588593,
-0.005500115919858217,
0.025286540389060974,
0.04461054503917694,
-0.008078323677182198,
-0.06506017595529556,
0.07267521321773529,
-0.2035190761089325,
0.05389382690191269,
-0.01073064561933279,
-0.06632687896490097,
-0.02661988139152527,
0.030778739601373672,
-0.08384375274181366,
0.07572232186794281,
0.12706787884235382,
0.01046997494995594,
0.016810817644000053,
-0.044822704046964645,
-0.02993215247988701,
0.015644192695617676,
-0.12733115255832672,
-0.13652819395065308,
-0.07370613515377045,
-0.09697027504444122,
0.08596285432577133,
-0.009802427142858505,
-0.13083021342754364,
-0.011835134588181973,
-0.11618591845035553,
0.07142731547355652,
-0.15336956083774567,
0.144621342420578,
0.083065465092659,
0.021282320842146873,
-0.017649799585342407,
-0.08638981729745865,
0.05515220761299133,
0.08350114524364471,
-0.10982725024223328,
-0.07850786298513412
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# layoutlmv2-finetuned-funsd-test
This model is a fine-tuned version of [microsoft/layoutlmv2-base-uncased](https://huggingface.co/microsoft/layoutlmv2-base-uncased) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 1000
### Training results
### Framework versions
- Transformers 4.16.0.dev0
- Pytorch 1.10.1
- Datasets 1.18.0
- Tokenizers 0.11.0
|
{"license": "cc-by-nc-sa-4.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "layoutlmv2-finetuned-funsd-test", "results": []}]}
|
token-classification
|
dhanesh123in/layoutlmv2-finetuned-funsd-test
|
[
"transformers",
"pytorch",
"tensorboard",
"layoutlmv2",
"token-classification",
"generated_from_trainer",
"license:cc-by-nc-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #layoutlmv2 #token-classification #generated_from_trainer #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #region-us
|
# layoutlmv2-finetuned-funsd-test
This model is a fine-tuned version of microsoft/layoutlmv2-base-uncased on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_ratio: 0.1
- training_steps: 1000
### Training results
### Framework versions
- Transformers 4.16.0.dev0
- Pytorch 1.10.1
- Datasets 1.18.0
- Tokenizers 0.11.0
|
[
"# layoutlmv2-finetuned-funsd-test\n\nThis model is a fine-tuned version of microsoft/layoutlmv2-base-uncased on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- training_steps: 1000",
"### Training results",
"### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.1\n- Datasets 1.18.0\n- Tokenizers 0.11.0"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #layoutlmv2 #token-classification #generated_from_trainer #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# layoutlmv2-finetuned-funsd-test\n\nThis model is a fine-tuned version of microsoft/layoutlmv2-base-uncased on an unknown dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- training_steps: 1000",
"### Training results",
"### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.1\n- Datasets 1.18.0\n- Tokenizers 0.11.0"
] |
[
65,
45,
6,
12,
8,
3,
104,
4,
37
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #layoutlmv2 #token-classification #generated_from_trainer #license-cc-by-nc-sa-4.0 #autotrain_compatible #endpoints_compatible #region-us \n# layoutlmv2-finetuned-funsd-test\n\nThis model is a fine-tuned version of microsoft/layoutlmv2-base-uncased on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- training_steps: 1000### Training results### Framework versions\n\n- Transformers 4.16.0.dev0\n- Pytorch 1.10.1\n- Datasets 1.18.0\n- Tokenizers 0.11.0"
] |
[
-0.12939299643039703,
0.1575356423854828,
-0.0008371934527531266,
0.11533793807029724,
0.1316145956516266,
0.031371332705020905,
0.06549704819917679,
0.16673792898654938,
-0.06874769181013107,
0.05339241027832031,
0.08972364664077759,
0.061528466641902924,
0.052387725561857224,
0.14107495546340942,
-0.01731722056865692,
-0.21742090582847595,
-0.02241181954741478,
-0.007584086153656244,
-0.006037536077201366,
0.1069665402173996,
0.10036634653806686,
-0.0893363505601883,
0.07372357696294785,
-0.0067273275926709175,
-0.13106265664100647,
0.00392192555591464,
-0.05015648528933525,
-0.049135226756334305,
0.0919104814529419,
-0.022059593349695206,
0.07364112138748169,
-0.003255438292399049,
0.14979074895381927,
-0.19841700792312622,
-0.005856142845004797,
0.057269614189863205,
0.048690564930438995,
0.08848342299461365,
0.05932025983929634,
-0.01685389317572117,
0.05770810693502426,
-0.16881534457206726,
0.10183598846197128,
0.02115250565111637,
-0.0538848415017128,
-0.14988459646701813,
-0.08616939932107925,
0.06880583614110947,
0.062277745455503464,
0.11130992323160172,
0.02434743382036686,
0.14979083836078644,
-0.06159372627735138,
0.08507713675498962,
0.18823884427547455,
-0.21912851929664612,
-0.045672640204429626,
0.08248795568943024,
0.07131271064281464,
0.04970621317625046,
-0.11923985928297043,
-0.009629289619624615,
0.036564186215400696,
0.03369196131825447,
0.07567079365253448,
-0.016099069267511368,
-0.12902067601680756,
0.00955523643642664,
-0.11424846947193146,
-0.034829769283533096,
0.16559375822544098,
0.025645961984992027,
-0.042611218988895416,
-0.12262944877147675,
-0.010450228117406368,
-0.11264877766370773,
-0.004305563401430845,
-0.016925418749451637,
0.02812868356704712,
-0.06467985361814499,
-0.05317233130335808,
-0.07387511432170868,
-0.08452844619750977,
-0.08087935298681259,
0.0305375587195158,
0.07424328476190567,
0.019382260739803314,
0.023912113159894943,
-0.02769801765680313,
0.13148990273475647,
0.00396007252857089,
-0.120609812438488,
-0.03621948882937431,
-0.006588221061974764,
-0.08143056184053421,
-0.0596768856048584,
-0.029397178441286087,
-0.003053333144634962,
-0.012421393766999245,
0.11675746738910675,
-0.03608483821153641,
0.06797269731760025,
0.040365032851696014,
-0.0013864259235560894,
-0.00393235357478261,
0.14356406033039093,
-0.0199501421302557,
-0.03717250004410744,
0.002555436920374632,
0.09383994340896606,
0.006033350247889757,
-0.02080547995865345,
-0.10196320712566376,
-0.04144662246108055,
0.10959089547395706,
0.07495386898517609,
-0.010979040525853634,
0.02920874021947384,
-0.03845008835196495,
-0.032215509563684464,
0.033307455480098724,
-0.149316668510437,
0.05907799303531647,
-0.010987665504217148,
-0.10319013148546219,
-0.08907660096883774,
0.061528731137514114,
-0.00918909627944231,
-0.03733383119106293,
0.051979776471853256,
-0.06580771505832672,
0.016498100012540817,
-0.08263096213340759,
-0.04798656702041626,
0.029396818950772285,
-0.12109731137752533,
0.0029031666927039623,
-0.05501644313335419,
-0.28309041261672974,
-0.06101866438984871,
0.04425784945487976,
-0.07247073948383331,
-0.02918047085404396,
-0.045987289398908615,
-0.0696694627404213,
0.004074576310813427,
-0.0062528327107429504,
0.15904535353183746,
-0.05236189812421799,
0.08398742228746414,
-0.0009636467439122498,
0.01949933171272278,
0.03370199725031853,
0.058460962027311325,
-0.09391278773546219,
0.01375618577003479,
-0.0876379981637001,
0.0954611524939537,
-0.1012168899178505,
0.030804337933659554,
-0.10199339687824249,
-0.10016226023435593,
0.015924250707030296,
-0.031812168657779694,
0.055889472365379333,
0.14006590843200684,
-0.1894666850566864,
0.009425231255590916,
0.09833752363920212,
-0.06475997716188431,
-0.07522659748792648,
0.06085137277841568,
-0.04756258800625801,
0.07649538666009903,
0.05805083364248276,
0.1622248739004135,
0.14083099365234375,
-0.1691412329673767,
-0.011999238282442093,
0.04020261391997337,
0.04430092126131058,
-0.020413076505064964,
0.05465017259120941,
-0.0011102940188720822,
0.05037952587008476,
0.0234586913138628,
-0.07111961394548416,
-0.026603123173117638,
-0.08248724788427353,
-0.08865822851657867,
-0.0725342184305191,
-0.09871932119131088,
0.0584295280277729,
0.049586713314056396,
0.026263855397701263,
-0.057771947234869,
-0.10853186249732971,
0.10312487930059433,
0.13948790729045868,
-0.04262711480259895,
0.00009643495286582038,
-0.06554698944091797,
0.018423985689878464,
-0.020883042365312576,
-0.05018465593457222,
-0.21471169590950012,
-0.11555662751197815,
0.03646582365036011,
-0.050735633820295334,
0.025837523862719536,
0.005262298043817282,
0.0709928646683693,
0.0558047853410244,
-0.03195890039205551,
-0.01971183344721794,
-0.08074785023927689,
0.01132049597799778,
-0.10870189219713211,
-0.1711205393075943,
-0.08128903061151505,
-0.04137100279331207,
0.16187384724617004,
-0.23600567877292633,
0.012767746113240719,
-0.01977449655532837,
0.12020131200551987,
0.020065223798155785,
-0.07128611952066422,
0.041487134993076324,
0.027845686301589012,
-0.0033396263606846333,
-0.10757793486118317,
0.03870905190706253,
0.011419342830777168,
-0.08549784123897552,
-0.07025958597660065,
-0.10585127770900726,
-0.0037657953798770905,
0.05726910009980202,
0.08705582469701767,
-0.11725250631570816,
-0.0313127338886261,
-0.057474032044410706,
-0.048518966883420944,
-0.09019745141267776,
0.008090773597359657,
0.18972013890743256,
0.015641985461115837,
0.10271663218736649,
-0.05321625620126724,
-0.09493141621351242,
-0.0012268135324120522,
0.0070521775633096695,
-0.01496418658643961,
0.08487345278263092,
0.040998298674821854,
-0.08367575705051422,
0.08723287284374237,
0.04569792002439499,
-0.04502886161208153,
0.14795203506946564,
-0.04335576295852661,
-0.0883985087275505,
-0.023401768878102303,
0.01170116662979126,
-0.02641597017645836,
0.15161143243312836,
-0.09026963263750076,
0.0023029856383800507,
0.04287653788924217,
0.009279169142246246,
0.032066699117422104,
-0.17665080726146698,
-0.01297433115541935,
0.026353666558861732,
-0.04913992062211037,
-0.019068440422415733,
-0.03623003512620926,
0.03079306147992611,
0.06887762248516083,
0.01581759937107563,
-0.006885693408548832,
0.03772728145122528,
-0.012234535999596119,
-0.10231117159128189,
0.1712239533662796,
-0.10850478708744049,
-0.17229746282100677,
-0.12124525755643845,
0.06995275616645813,
-0.04360293969511986,
-0.031045131385326385,
0.004369784612208605,
-0.05471450090408325,
-0.040555667132139206,
-0.09358299523591995,
-0.0646066814661026,
-0.03427529335021973,
-0.024138180539011955,
0.016857899725437164,
0.03185458853840828,
0.06633520871400833,
-0.12328526377677917,
0.012356769293546677,
-0.011887315660715103,
-0.06296682357788086,
0.0042466577142477036,
0.05393025651574135,
0.10122600197792053,
0.10626626759767532,
-0.027890894562005997,
0.032392989844083786,
-0.03545549511909485,
0.19121868908405304,
-0.0734284445643425,
0.007799208164215088,
0.13594992458820343,
0.004342028871178627,
0.04521629214286804,
0.08922342211008072,
0.019682234153151512,
-0.07923505455255508,
0.021839294582605362,
0.03583594039082527,
-0.0233453456312418,
-0.23807208240032196,
-0.04591453820466995,
-0.016276516020298004,
-0.05991070345044136,
0.09377844631671906,
0.05379447713494301,
0.013983845710754395,
0.048194628208875656,
0.017412280663847923,
-0.001773718511685729,
-0.023998063057661057,
0.06919100135564804,
0.06464161723852158,
0.016055842861533165,
0.07692815363407135,
-0.03555774688720703,
-0.00025831456878222525,
0.05816120281815529,
0.04553331807255745,
0.2523238956928253,
-0.05037101358175278,
0.1101704090833664,
0.007649913430213928,
0.16908055543899536,
-0.03985349461436272,
0.042329974472522736,
0.03747762367129326,
0.01120032835751772,
0.03424714505672455,
-0.06662753224372864,
-0.03913966566324234,
0.034202273935079575,
-0.010742993094027042,
0.034462232142686844,
-0.08999045193195343,
0.016693593934178352,
0.004545214120298624,
0.21758821606636047,
0.04510254040360451,
-0.3049599528312683,
-0.09102493524551392,
-0.0029936498031020164,
0.0039246175438165665,
-0.0890863686800003,
-0.01650717668235302,
0.11582688242197037,
-0.1640883982181549,
0.07207094877958298,
-0.050430215895175934,
0.08713342994451523,
-0.06282340735197067,
-0.01428751926869154,
0.05702907592058182,
0.086785227060318,
0.004919316619634628,
0.11105846613645554,
-0.19020608067512512,
0.19899779558181763,
0.032276347279548645,
0.12171805649995804,
-0.1028611809015274,
0.028273331001400948,
0.006413045804947615,
0.07060014456510544,
0.13523021340370178,
-0.01017756573855877,
-0.04373420402407646,
-0.17290543019771576,
-0.11130515486001968,
0.019284764304757118,
0.09282328933477402,
-0.02103896625339985,
0.06607323884963989,
-0.03894248977303505,
0.001976065104827285,
0.03524616360664368,
-0.12528257071971893,
-0.15494227409362793,
-0.10876639187335968,
0.03336547315120697,
0.007106666453182697,
-0.024497969076037407,
-0.06399434804916382,
-0.10820256918668747,
0.007428579963743687,
0.18302887678146362,
0.0033822187688201666,
-0.07156718522310257,
-0.15603262186050415,
0.08063308149576187,
0.14721497893333435,
-0.04990420863032341,
0.015915624797344208,
0.017804795876145363,
0.13057668507099152,
0.039835043251514435,
-0.07923146337270737,
0.03072371333837509,
-0.06765860319137573,
-0.1594899445772171,
-0.048798952251672745,
0.12242525070905685,
0.06443459540605545,
0.05712278187274933,
0.0031949698459357023,
0.031264565885066986,
-0.02040962502360344,
-0.08276476711034775,
0.031074896454811096,
0.07960289716720581,
0.07750927656888962,
0.05663248151540756,
-0.08061404526233673,
0.02877287194132805,
-0.023004015907645226,
-0.025980325415730476,
0.14563141763210297,
0.14845126867294312,
-0.10997140407562256,
0.11232282221317291,
0.04556744173169136,
-0.08935988694429398,
-0.20581157505512238,
0.07317309081554413,
0.11729029566049576,
0.021672364324331284,
0.06337989866733551,
-0.18494285643100739,
0.09199993312358856,
0.09292842447757721,
-0.037454355508089066,
0.09760470688343048,
-0.31080183386802673,
-0.13111281394958496,
0.07964663207530975,
0.08718634396791458,
-0.04766271635890007,
-0.11981914192438126,
-0.04866243526339531,
-0.0015554676065221429,
-0.12789417803287506,
0.10339047759771347,
-0.03135896474123001,
0.1145767942070961,
-0.012219583615660667,
0.1171257421374321,
0.03784890100359917,
-0.045987363904714584,
0.1385539174079895,
0.03801758214831352,
0.0742398276925087,
-0.06819094717502594,
0.02993231825530529,
0.08012990653514862,
-0.08630736917257309,
0.1153021976351738,
-0.02008446678519249,
0.08319730311632156,
-0.20318342745304108,
-0.015665823593735695,
-0.07277575880289078,
0.08950537443161011,
-0.047934383153915405,
-0.05765025317668915,
-0.024073351174592972,
0.06383964419364929,
0.03208957985043526,
-0.04420669004321098,
0.0500650517642498,
0.02082945592701435,
0.08213936537504196,
0.12868571281433105,
0.09888062626123428,
-0.0032046856358647346,
-0.12232762575149536,
-0.005346791818737984,
-0.018482264131307602,
0.034931156784296036,
-0.10544437915086746,
0.0014579104026779532,
0.13709476590156555,
0.0741216242313385,
0.12203620374202728,
0.015020451508462429,
-0.0479494109749794,
-0.015204940922558308,
0.026290902867913246,
-0.10419895499944687,
-0.08365964144468307,
-0.016250956803560257,
-0.05363776907324791,
-0.15315307676792145,
0.008125345222651958,
0.08210553973913193,
-0.06644699722528458,
-0.011627817526459694,
-0.03542962297797203,
0.03950762003660202,
-0.01283010933548212,
0.19562272727489471,
0.062244970351457596,
0.06955559551715851,
-0.08238489925861359,
0.13970524072647095,
0.07964414358139038,
-0.06062934547662735,
0.06551554799079895,
0.07197340577840805,
-0.08604728430509567,
-0.03825271502137184,
0.09388438612222672,
0.15653176605701447,
-0.051968980580568314,
-0.05263407900929451,
-0.06929019093513489,
-0.06534460932016373,
0.046158209443092346,
0.09564672410488129,
0.04545328766107559,
-0.004279364366084337,
-0.01654740795493126,
0.004169587045907974,
-0.14432623982429504,
0.11451023072004318,
0.0528482161462307,
0.054160069674253464,
-0.15542061626911163,
0.1114792451262474,
0.010662478394806385,
0.0776214525103569,
-0.01964089833199978,
0.018588604405522346,
-0.05873763933777809,
-0.001625991309992969,
-0.10439858585596085,
0.01396574079990387,
-0.0217350535094738,
0.014288280159235,
-0.011847656220197678,
-0.05552314221858978,
-0.01052154041826725,
0.05514116957783699,
-0.06245501711964607,
-0.06988111883401871,
-0.000026013200113084167,
0.04856118559837341,
-0.14282925426959991,
-0.032717905938625336,
0.016238631680607796,
-0.08077221363782883,
0.0651901587843895,
0.06523260474205017,
0.01401549018919468,
0.006997478194534779,
-0.07928834855556488,
0.0023027402348816395,
0.01874922774732113,
0.028453774750232697,
0.05956859514117241,
-0.11420505493879318,
-0.018994508311152458,
-0.018269749358296394,
0.024821484461426735,
0.0288689024746418,
0.09618481993675232,
-0.13560329377651215,
-0.03229998052120209,
-0.0520947128534317,
-0.02796630747616291,
-0.06031254306435585,
0.06704878062009811,
0.10924359411001205,
0.03286980092525482,
0.13882753252983093,
-0.06711547076702118,
0.06773237884044647,
-0.1955292522907257,
-0.04426615685224533,
-0.014711946249008179,
-0.013489773496985435,
-0.0709853321313858,
-0.020139843225479126,
0.1022149920463562,
-0.04365230351686478,
0.10976777225732803,
0.028226656839251518,
0.09879055619239807,
0.021501343697309494,
0.0013506535906344652,
-0.01582120545208454,
-0.0009109565289691091,
0.13760793209075928,
0.037466395646333694,
-0.008528640493750572,
0.13483783602714539,
0.00712980842217803,
0.04274694249033928,
0.022781159728765488,
0.18433834612369537,
0.12702476978302002,
-0.017410198226571083,
0.06441616266965866,
0.08119409531354904,
-0.11824830621480942,
-0.14666754007339478,
0.09201085567474365,
-0.006954318378120661,
0.12150601297616959,
-0.07580123096704483,
0.14277851581573486,
0.05349342152476311,
-0.17771929502487183,
0.06049371883273125,
-0.06205184385180473,
-0.1225130558013916,
-0.09413734078407288,
-0.0734499916434288,
-0.06794987618923187,
-0.08951496332883835,
0.014132303185760975,
-0.10636276006698608,
0.03637300431728363,
0.08050846308469772,
-0.001388090313412249,
-0.001565847429446876,
0.1530587375164032,
-0.04727982357144356,
0.018634159117937088,
0.054330434650182724,
0.01707538589835167,
0.006960129830986261,
-0.09500893950462341,
-0.03744908422231674,
0.06159553304314613,
0.037510402500629425,
0.07939226180315018,
-0.047635767608881,
0.009777726605534554,
0.03001527488231659,
-0.0022419344168156385,
-0.07418700307607651,
0.014638552442193031,
0.030417434871196747,
0.05373956263065338,
0.05446969345211983,
0.05943483114242554,
0.007231456693261862,
-0.04773145914077759,
0.2894349694252014,
-0.06486456096172333,
-0.09507351368665695,
-0.11453033238649368,
0.1885373443365097,
0.010330267250537872,
-0.0447404570877552,
0.07495032250881195,
-0.13216441869735718,
0.006848725490272045,
0.16765400767326355,
0.148485004901886,
-0.04477768391370773,
-0.02131475880742073,
-0.014899351634085178,
-0.012260522693395615,
-0.0294172465801239,
0.09371715784072876,
0.09202253818511963,
0.04340265691280365,
-0.07964577525854111,
-0.013233445584774017,
-0.0026471528690308332,
-0.05115861818194389,
-0.09641697257757187,
0.0609746016561985,
-0.006110343150794506,
0.011872686445713043,
-0.03129298612475395,
0.0856252908706665,
-0.010613374412059784,
-0.19551916420459747,
0.06165355443954468,
-0.16979223489761353,
-0.1754380464553833,
-0.023174501955509186,
0.10087478160858154,
-0.024403024464845657,
0.033487528562545776,
-0.002802349394187331,
0.02294827811419964,
0.11006884276866913,
-0.028125828132033348,
-0.07039503008127213,
-0.10090699791908264,
0.07376852631568909,
-0.06716066598892212,
0.22692182660102844,
-0.005435547325760126,
0.09402067214250565,
0.08910728245973587,
0.015650970861315727,
-0.1737547665834427,
0.03079827129840851,
0.07956214994192123,
-0.035564251244068146,
0.05187419801950455,
0.16304895281791687,
-0.045798201113939285,
0.07781769335269928,
0.029742442071437836,
-0.10389295220375061,
-0.025821318849921227,
-0.013605249114334583,
-0.001413363846950233,
-0.08234294503927231,
-0.007736407220363617,
-0.0412246398627758,
0.16746960580348969,
0.1779661923646927,
-0.04780750721693039,
-0.001045025885105133,
-0.09386097639799118,
0.023906493559479713,
0.03837434574961662,
0.038320839405059814,
-0.02661401405930519,
-0.18377305567264557,
0.02892969734966755,
0.019585955888032913,
0.029801318421959877,
-0.20406650006771088,
-0.08947855234146118,
0.06227293238043785,
-0.068752720952034,
-0.03058244101703167,
0.09987154603004456,
0.04640905559062958,
0.02981945499777794,
-0.02993587590754032,
-0.06661785393953323,
-0.02808310091495514,
0.1259656399488449,
-0.17817841470241547,
-0.04869915917515755
] |
null | null |
transformers
|
# AMy San
|
{"tags": ["conversational"]}
|
text-generation
|
dhanushlnaik/amySan
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# AMy San
|
[
"# AMy San"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# AMy San"
] |
[
51,
4
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# AMy San"
] |
[
-0.034503597766160965,
0.05791027098894119,
-0.006130848079919815,
0.011839490383863449,
0.15175068378448486,
0.018564075231552124,
0.12553340196609497,
0.13319537043571472,
0.02341933734714985,
-0.02243289165198803,
0.11069229990243912,
0.15768446028232574,
0.013916644267737865,
0.07604608684778214,
-0.08493801951408386,
-0.23301102221012115,
0.06775275617837906,
0.053848713636398315,
0.012264534831047058,
0.10192547738552094,
0.09065353870391846,
-0.06479903310537338,
0.09428836405277252,
-0.03679957613348961,
-0.1126900389790535,
0.03728736937046051,
0.02901248261332512,
-0.14871996641159058,
0.11429822444915771,
0.05450974404811859,
0.06155392527580261,
0.030204081907868385,
-0.06223631650209427,
-0.12711618840694427,
0.036692604422569275,
-0.006317433435469866,
-0.06325369328260422,
0.040487248450517654,
0.06151629611849785,
-0.07805519551038742,
0.08049053698778152,
0.1168392226099968,
-0.03118942677974701,
0.06842998415231705,
-0.17478004097938538,
-0.06362064182758331,
0.0039273942820727825,
0.04949631541967392,
0.06784673780202866,
0.09917443245649338,
-0.020093822851777077,
0.05238761380314827,
-0.06744785606861115,
0.08008002489805222,
0.06048249080777168,
-0.2241532802581787,
-0.000583960092626512,
0.0994722917675972,
0.04346431419253349,
0.10259459912776947,
-0.0277554951608181,
0.05829557403922081,
0.019424282014369965,
-0.0025650623720139265,
-0.02632829360663891,
-0.09456374496221542,
-0.05357062444090843,
0.05376877263188362,
-0.058959346264600754,
-0.02940904162824154,
0.24599120020866394,
-0.0502227284014225,
0.05934837833046913,
-0.028666852042078972,
-0.08658040314912796,
-0.016078680753707886,
-0.03962216153740883,
-0.015050001442432404,
-0.07898438721895218,
0.07885488122701645,
0.020460084080696106,
-0.08570882678031921,
-0.1295834183692932,
-0.029384126886725426,
-0.15524356067180634,
0.14666029810905457,
0.03449846804141998,
0.04132954403758049,
-0.19195549190044403,
0.06478624790906906,
-0.03219274803996086,
-0.09606828540563583,
0.021715398877859116,
-0.08188413828611374,
0.04452001675963402,
0.023323681205511093,
-0.022399844601750374,
-0.08209113031625748,
0.07390753924846649,
0.11417675018310547,
-0.05913829058408737,
0.04516439512372017,
-0.01737787015736103,
0.06608621031045914,
0.04755320027470589,
0.05233478546142578,
0.012087864801287651,
-0.03834446147084236,
0.020589547231793404,
-0.09969794750213623,
0.004129775334149599,
-0.07063548266887665,
-0.1736479550600052,
0.0007402616320177913,
0.01282219123095274,
0.06928330659866333,
0.018476063385605812,
0.10188546776771545,
-0.033658355474472046,
-0.0304133091121912,
0.06975767016410828,
-0.04813167825341225,
-0.012473898008465767,
0.020826101303100586,
0.011812273412942886,
0.09196173399686813,
-0.030287036672234535,
0.045016270130872726,
-0.09731055796146393,
0.06506960093975067,
-0.06808749586343765,
-0.02184952050447464,
-0.05081033706665039,
-0.012993939220905304,
0.022166380658745766,
-0.07488831877708435,
0.0016426050569862127,
-0.1686902791261673,
-0.15086615085601807,
0.012040948495268822,
-0.01035540271550417,
-0.07063645869493484,
-0.06593597680330276,
-0.061372578144073486,
0.001788355060853064,
0.034085649996995926,
-0.07697397470474243,
-0.02027811110019684,
-0.06795120984315872,
0.10717147588729858,
-0.005107425153255463,
0.08485275506973267,
-0.13260209560394287,
0.08067521452903748,
-0.09114328771829605,
-0.0033242511563003063,
-0.0725526511669159,
0.06334098428487778,
0.0016113484743982553,
0.09289704263210297,
-0.010633835569024086,
-0.008148711174726486,
-0.08246453106403351,
0.07032380998134613,
-0.041938383132219315,
0.23313383758068085,
-0.06688468903303146,
-0.09766252338886261,
0.2709226906299591,
-0.072092704474926,
-0.14711585640907288,
0.12642012536525726,
0.013637599535286427,
0.06457658857107162,
0.09943649917840958,
0.22542425990104675,
-0.08651869744062424,
-0.036073293536901474,
0.08995763212442398,
0.0775609090924263,
-0.07482761144638062,
-0.09548136591911316,
0.05460142344236374,
-0.0017428041901439428,
-0.10566357523202896,
0.039435047656297684,
0.0961378887295723,
0.0722755491733551,
-0.030572345480322838,
-0.05533842369914055,
0.006443878635764122,
-0.00994260422885418,
0.01431290153414011,
0.0012562547344714403,
0.11037395149469376,
-0.05455419421195984,
-0.011861885897815228,
-0.04347106069326401,
0.005915557034313679,
-0.028436174616217613,
0.03296252712607384,
-0.08222856372594833,
0.11455083638429642,
-0.013076258823275566,
0.06634348630905151,
-0.13313597440719604,
-0.006408945191651583,
-0.01723700948059559,
0.14114373922348022,
0.025396231561899185,
0.07148723304271698,
0.045840486884117126,
-0.025385631248354912,
-0.031395044177770615,
0.0035600264091044664,
0.16539357602596283,
-0.00497045274823904,
-0.07310260832309723,
-0.06794573366641998,
0.07681340724229813,
-0.04649722948670387,
0.03692067787051201,
-0.08835076540708542,
0.02869839407503605,
0.056605733931064606,
0.07498320192098618,
-0.023777827620506287,
0.029971526935696602,
-0.006965316366404295,
0.021710548549890518,
-0.07357007265090942,
0.01939641870558262,
0.11540242284536362,
0.014539851807057858,
-0.12072569131851196,
0.25278958678245544,
-0.19190821051597595,
0.14634345471858978,
0.16507138311862946,
-0.2684532403945923,
0.015495391562581062,
-0.07756452262401581,
-0.03714727237820625,
-0.00319844507612288,
0.040411729365587234,
-0.06947300583124161,
0.1713167428970337,
-0.017383398488163948,
0.214442640542984,
-0.05019012466073036,
-0.01176861859858036,
-0.0062072002328932285,
-0.05420646816492081,
-0.04106232896447182,
0.0978865921497345,
0.12643706798553467,
-0.11668699979782104,
0.20550376176834106,
0.15674792230129242,
0.04252654314041138,
0.20334205031394958,
0.02272057719528675,
-0.005406200420111418,
0.07163656502962112,
-0.014483843930065632,
-0.031773362308740616,
-0.03307066485285759,
-0.23822708427906036,
-0.018769899383187294,
0.06914038211107254,
0.043388377875089645,
0.10831968486309052,
-0.10676556825637817,
-0.048741664737463,
-0.007020852528512478,
-0.010233753360807896,
-0.0062072016298770905,
0.07694323360919952,
0.02496817521750927,
0.13168196380138397,
-0.00489236181601882,
-0.0006336478982120752,
0.08581164479255676,
0.015317480079829693,
-0.08502757549285889,
0.1778576523065567,
-0.11393330246210098,
-0.3630668818950653,
-0.12590786814689636,
-0.1840478777885437,
-0.007897612638771534,
0.07457596808671951,
0.10700579732656479,
-0.14280259609222412,
-0.007767796516418457,
-0.004555910360068083,
0.04924564063549042,
-0.09729479253292084,
-0.0009881329024210572,
-0.042182281613349915,
0.039100635796785355,
-0.08297105878591537,
-0.06988170742988586,
-0.06245442479848862,
-0.0031756646931171417,
-0.06625521928071976,
0.14852078258991241,
-0.13650688529014587,
0.08079452812671661,
0.18460161983966827,
0.07321491837501526,
0.06015081703662872,
-0.04435430094599724,
0.15612590312957764,
-0.11157484352588654,
0.01477632112801075,
0.24629367887973785,
-0.03402290865778923,
0.0705198347568512,
0.12976111471652985,
-0.009392702020704746,
-0.08122546225786209,
0.03361900523304939,
-0.03342496603727341,
-0.09856335818767548,
-0.21554894745349884,
-0.10989131778478622,
-0.12342378497123718,
0.12290970981121063,
0.015469296835362911,
0.05668918415904045,
0.20512273907661438,
0.08524180203676224,
-0.03147950768470764,
0.0029982661362737417,
0.0036004651337862015,
0.07920229434967041,
0.21028542518615723,
-0.0467088408768177,
0.1361730843782425,
-0.031968481838703156,
-0.13635601103305817,
0.11119319498538971,
0.07995143532752991,
0.05519958585500717,
0.06967643648386002,
0.09032703936100006,
0.02682489901781082,
0.06706785410642624,
0.12192315608263016,
0.03287861868739128,
0.0063356636092066765,
-0.026102660223841667,
-0.06626502424478531,
-0.040240827947854996,
-0.03993576765060425,
0.07060938328504562,
0.04748355224728584,
-0.15767990052700043,
-0.02412777580320835,
-0.07379548996686935,
0.06841205805540085,
0.06275638192892075,
0.032260265201330185,
-0.18089333176612854,
0.00031170627335086465,
0.11402637511491776,
0.003180157858878374,
-0.1083453893661499,
0.08624409884214401,
0.06698409467935562,
-0.13012640178203583,
0.04085153341293335,
-0.02544941008090973,
0.12948456406593323,
-0.06945140659809113,
0.09916239231824875,
-0.04770258814096451,
-0.06122667342424393,
0.028141029179096222,
0.11856961995363235,
-0.3366113305091858,
0.19823580980300903,
-0.0041079032234847546,
-0.04441896453499794,
-0.13152547180652618,
-0.010521185584366322,
-0.004085357766598463,
0.06423842161893845,
0.11732949316501617,
-0.004085042513906956,
-0.048602182418107986,
-0.11752688884735107,
-0.050164975225925446,
0.030650462955236435,
0.1141660064458847,
-0.03742427006363869,
-0.027862943708896637,
-0.03025038354098797,
-0.0008930927724577487,
-0.02947222627699375,
-0.02708463743329048,
0.0017357761971652508,
-0.1872776448726654,
0.0938023179769516,
0.11862604320049286,
0.10410252213478088,
0.011202886700630188,
-0.013652723282575607,
-0.07547388225793839,
0.18406479060649872,
-0.03247378021478653,
-0.09005269408226013,
-0.14040859043598175,
-0.054341092705726624,
0.018096940591931343,
-0.06141328811645508,
0.010229131206870079,
-0.07126523554325104,
0.004903138615190983,
-0.05804945155978203,
-0.16170144081115723,
0.13566328585147858,
-0.11130725592374802,
-0.044057607650756836,
-0.0563783198595047,
0.18729408085346222,
-0.032116588205099106,
0.025897784158587456,
0.011847928166389465,
0.002295003505423665,
-0.08268125355243683,
-0.09215182065963745,
0.043784093111753464,
-0.05617628991603851,
0.10072947293519974,
-0.0022227768786251545,
-0.0860438421368599,
-0.04270874708890915,
-0.05246559530496597,
-0.09027466177940369,
0.26798298954963684,
0.14715224504470825,
-0.02884705550968647,
0.19408418238162994,
0.15523114800453186,
-0.073142871260643,
-0.31961098313331604,
-0.11709347367286682,
-0.10047858208417892,
-0.02634456567466259,
-0.06252287328243256,
-0.22138158977031708,
0.04752001911401749,
0.027890028432011604,
-0.021211175248026848,
0.08590163290500641,
-0.27079537510871887,
-0.10064990818500519,
0.10221453756093979,
0.014943348243832588,
0.41970130801200867,
-0.14618785679340363,
-0.11615955084562302,
-0.03254932910203934,
-0.1364319920539856,
0.13292677700519562,
0.023578781634569168,
0.11059404164552689,
0.0021957922726869583,
0.11884858459234238,
0.04279136657714844,
-0.03713797777891159,
0.08319099992513657,
0.010754833929240704,
-0.03538413345813751,
-0.0983971431851387,
-0.10040385276079178,
-0.00020148491603322327,
0.026026450097560883,
0.005908457096666098,
-0.04044819623231888,
0.012626233510673046,
-0.12458387762308121,
-0.047790564596652985,
-0.08176164329051971,
0.014381681568920612,
0.03440950810909271,
-0.08836765587329865,
0.015307464636862278,
-0.022633273154497147,
-0.0018121397588402033,
0.0005845458945259452,
0.25334078073501587,
-0.09683237969875336,
0.1744239628314972,
0.09013206511735916,
0.10859128832817078,
-0.15109513700008392,
0.02552468702197075,
-0.0906495451927185,
-0.05626637116074562,
0.0642663910984993,
-0.10113456100225449,
0.06896421313285828,
0.09386683255434036,
-0.06290482729673386,
0.09915239363908768,
0.08490782976150513,
0.005205215886235237,
0.027157755568623543,
0.12364442646503448,
-0.2501172423362732,
-0.04700401425361633,
-0.06984096020460129,
0.06078099459409714,
0.11494912952184677,
0.0927421897649765,
0.1605692356824875,
0.008964338339865208,
-0.03886980563402176,
-0.008214788511395454,
0.029132908210158348,
-0.01918761432170868,
0.06771068274974823,
0.010654987767338753,
0.01635289192199707,
-0.12487317621707916,
0.07627534121274948,
0.032446347177028656,
-0.19181722402572632,
0.028879914432764053,
0.15974603593349457,
-0.11731182783842087,
-0.10268423706293106,
-0.08270814269781113,
0.07363291084766388,
-0.10806555300951004,
-0.0632317066192627,
-0.07101478427648544,
-0.13526666164398193,
0.08169455081224442,
0.10158473998308182,
0.04908238351345062,
0.07242603600025177,
-0.06696213781833649,
-0.029070062562823296,
-0.011184301227331161,
-0.005113371182233095,
-0.033298593014478683,
-0.007492642384022474,
-0.053936999291181564,
0.02330336906015873,
-0.019869066774845123,
0.14242377877235413,
-0.09428968280553818,
-0.08733408153057098,
-0.18478865921497345,
0.046540528535842896,
-0.12163010239601135,
-0.09029277414083481,
-0.10711920261383057,
-0.08329404890537262,
-0.02362775430083275,
-0.057919904589653015,
-0.04777891933917999,
-0.04638642072677612,
-0.11116855591535568,
0.03625694662332535,
-0.048160165548324585,
0.043771252036094666,
-0.04643435403704643,
-0.004465772770345211,
0.08550962060689926,
-0.016784118488430977,
0.12513367831707,
0.1265159696340561,
-0.10822022706270218,
0.030550824478268623,
-0.12576092779636383,
-0.044850245118141174,
0.09353233128786087,
0.03811391815543175,
0.0364527590572834,
0.08260220289230347,
0.010421582497656345,
0.07939644902944565,
0.024460887536406517,
0.04326748102903366,
0.03963814675807953,
-0.1136714369058609,
0.08700776845216751,
-0.025222793221473694,
-0.1517987996339798,
-0.05425124987959862,
-0.032847385853528976,
0.03278736025094986,
0.0004325160407461226,
0.09043148159980774,
-0.07244554162025452,
0.09713112562894821,
-0.03261864557862282,
0.042283978313207626,
0.02680591680109501,
-0.1637836992740631,
-0.0844251960515976,
-0.09872637689113617,
0.043320778757333755,
0.023583702743053436,
0.22839781641960144,
-0.007891588844358921,
-0.018734993413090706,
0.05184979736804962,
0.05915558710694313,
0.08538543432950974,
-0.01390082761645317,
0.1953580230474472,
0.09997651726007462,
-0.07176085561513901,
-0.043223969638347626,
0.08783303201198578,
0.016900116577744484,
0.048390891402959824,
0.11372548341751099,
0.07161981612443924,
0.04953893646597862,
0.07995540648698807,
-0.01796010509133339,
0.013909808360040188,
-0.07137705385684967,
-0.12605327367782593,
-0.08000683039426804,
0.025135330855846405,
-0.04663263261318207,
0.1528535634279251,
0.1412050724029541,
-0.03267458453774452,
0.042578112334012985,
-0.03983870521187782,
-0.04348186403512955,
-0.153091162443161,
-0.16766124963760376,
-0.07044954597949982,
-0.09856898337602615,
0.002520801266655326,
-0.09591630846261978,
0.05254006013274193,
0.07258371263742447,
0.07409831136465073,
-0.0727149173617363,
0.10209966450929642,
0.04901834577322006,
-0.09429551661014557,
0.06293875724077225,
-0.020395928993821144,
0.10215606540441513,
-0.0345512218773365,
-0.00968143716454506,
-0.0821819007396698,
0.017197662964463234,
0.0133675467222929,
0.055656999349594116,
-0.022983936592936516,
0.0009320663521066308,
-0.11052758246660233,
-0.10004579275846481,
-0.05597703903913498,
0.04170294851064682,
-0.009444739669561386,
0.11491471529006958,
-0.025441359728574753,
0.007919104769825935,
0.04603798687458038,
0.24679575860500336,
-0.0647275522351265,
-0.02432415820658207,
-0.03446296229958534,
0.16776308417320251,
0.02167212963104248,
0.0927865207195282,
-0.026850806549191475,
-0.0024479266721755266,
-0.07337986677885056,
0.3976871073246002,
0.2962881922721863,
-0.11541914194822311,
-0.004818781744688749,
0.019469939172267914,
0.052864447236061096,
0.12440398335456848,
0.11900145560503006,
0.0876491591334343,
0.2740756571292877,
-0.07328770309686661,
-0.03487854450941086,
-0.019186606630682945,
-0.0002752451691776514,
-0.0943528562784195,
0.14846430718898773,
0.02326691709458828,
-0.0553177073597908,
-0.022043611854314804,
0.07871093600988388,
-0.20903682708740234,
0.11013882607221603,
-0.11400587111711502,
-0.18668018281459808,
-0.07631507515907288,
0.012238680385053158,
0.14190904796123505,
0.0304570235311985,
0.07977765798568726,
0.008864812552928925,
-0.06448257714509964,
0.025378407910466194,
0.04240885004401207,
-0.19010989367961884,
-0.0017061004182323813,
0.06570187211036682,
-0.04448280856013298,
0.018031682819128036,
-0.039431821554899216,
0.0361529141664505,
0.07758104056119919,
0.025971349328756332,
-0.008459226228296757,
0.07027579843997955,
0.025912679731845856,
-0.06502635776996613,
-0.00629518274217844,
0.018447455018758774,
0.016163617372512817,
-0.0502772256731987,
0.06442396342754364,
-0.15554489195346832,
0.04917682334780693,
-0.019935591146349907,
-0.07038287818431854,
-0.01370813138782978,
0.029052313417196274,
-0.07097256183624268,
0.07020667940378189,
0.05563155934214592,
-0.01377400103956461,
-0.04880702123045921,
-0.04618578776717186,
-0.01772887445986271,
-0.021223459392786026,
-0.08341095596551895,
-0.08752051740884781,
-0.18387679755687714,
-0.1066039651632309,
0.03103257529437542,
0.02934052050113678,
-0.18237951397895813,
0.026865236461162567,
-0.10723299533128738,
0.08279033750295639,
-0.16845887899398804,
0.0578363873064518,
0.08328942954540253,
0.03007352538406849,
-0.003901307936757803,
-0.07291600108146667,
0.06683693826198578,
0.07828386127948761,
-0.08846165984869003,
-0.06896743923425674
] |
null | null |
transformers
|
"hello"
|
{}
|
text-classification
|
dhikri/question_answering_glue
|
[
"transformers",
"pytorch",
"distilbert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
"hello"
|
[] |
[
"TAGS\n#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
38
] |
[
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.03533577919006348,
0.06443645805120468,
-0.007824759930372238,
0.02963758073747158,
0.21134145557880402,
0.0368538573384285,
0.06359195709228516,
0.10786357522010803,
0.047846585512161255,
-0.029699334874749184,
0.09624463319778442,
0.2456933856010437,
-0.04527274891734123,
0.11506538093090057,
-0.1315092295408249,
-0.2995516061782837,
0.0646483302116394,
0.06820031255483627,
0.01968790777027607,
0.11027561128139496,
0.08955937623977661,
-0.08577843010425568,
0.06416945904493332,
-0.03987749293446541,
-0.13028311729431152,
0.036934368312358856,
0.037670549005270004,
-0.12557227909564972,
0.08850666880607605,
0.03936105594038963,
0.16363440454006195,
0.029493317008018494,
-0.0571451373398304,
-0.13760130107402802,
0.03542056307196617,
0.003107793163508177,
-0.08173839002847672,
0.035451244562864304,
0.07971785217523575,
-0.13606007397174835,
0.03269175812602043,
0.01657985709607601,
0.028779901564121246,
0.05034712329506874,
-0.13549968600273132,
-0.06766978651285172,
-0.009825913235545158,
0.02846479043364525,
0.08123840391635895,
0.06563035398721695,
-0.00027321543893776834,
0.11571130156517029,
-0.14468228816986084,
0.13729768991470337,
0.08681581169366837,
-0.26667332649230957,
-0.01513616368174553,
0.09300960600376129,
0.014211298897862434,
0.03189397603273392,
-0.05005642771720886,
0.03387840837240219,
0.021587392315268517,
0.012041964568197727,
-0.005505601409822702,
-0.06911619752645493,
-0.12172640115022659,
0.01909228041768074,
-0.0760328620672226,
-0.039914727210998535,
0.2024218738079071,
-0.06752687692642212,
0.06574457883834839,
-0.03853347897529602,
-0.09920144081115723,
-0.04725521057844162,
-0.028420861810445786,
0.03284634277224541,
-0.05052020400762558,
0.06803859770298004,
0.04873250797390938,
0.02093963511288166,
-0.10541380196809769,
0.027895580977201462,
-0.2198127955198288,
0.21804359555244446,
0.00917235016822815,
0.04113364964723587,
-0.17035873234272003,
0.06059039384126663,
0.043774571269750595,
-0.10760118812322617,
0.049048252403736115,
-0.10497406870126724,
0.019541887566447258,
-0.04680290073156357,
-0.07833123207092285,
-0.044003088027238846,
0.0761561468243599,
0.15131190419197083,
0.024525625631213188,
0.0676354393362999,
-0.023907558992505074,
0.08125972747802734,
0.03615585342049599,
0.12704050540924072,
0.04965166375041008,
-0.030767392367124557,
0.03752761334180832,
-0.13245059549808502,
-0.00002132852932845708,
-0.07070981711149216,
-0.1520344465970993,
-0.028104213997721672,
0.058518148958683014,
0.07771685719490051,
0.007545619271695614,
0.09117837250232697,
-0.07305282354354858,
-0.03670652583241463,
0.09205243736505508,
-0.09038619697093964,
0.022389709949493408,
0.0189626757055521,
0.024910688400268555,
0.11437109857797623,
-0.01640472002327442,
-0.004441923461854458,
-0.08554866164922714,
0.15481221675872803,
-0.05412428826093674,
0.01906411163508892,
-0.027951309457421303,
-0.07562480866909027,
0.023844171315431595,
-0.16517141461372375,
0.024268588051199913,
-0.16968505084514618,
-0.12177367508411407,
0.0011497566010802984,
0.01497613824903965,
0.0003558929602149874,
-0.029599502682685852,
-0.034584347158670425,
0.0028823118191212416,
0.05339471623301506,
-0.05009040981531143,
-0.08925710618495941,
-0.0734119787812233,
0.09545788168907166,
-0.03665677830576897,
0.07958490401506424,
-0.12844105064868927,
0.0784672200679779,
-0.0987219363451004,
-0.0187049712985754,
-0.14024826884269714,
0.05743253231048584,
-0.04765705391764641,
0.18340644240379333,
0.01636499911546707,
-0.05442013591527939,
-0.05629796162247658,
0.05081459879875183,
-0.06792773306369781,
0.17081454396247864,
-0.10482346266508102,
-0.11688733100891113,
0.18975088000297546,
-0.09539731591939926,
-0.11199936270713806,
0.08214274048805237,
-0.012322766706347466,
-0.002544441493228078,
0.10592521727085114,
0.18774141371250153,
0.11772145330905914,
0.015394842252135277,
0.071439228951931,
0.1266816407442093,
-0.09738999605178833,
-0.10514426231384277,
-0.016195401549339294,
-0.010998358018696308,
-0.11682542413473129,
0.06311710923910141,
0.08283041417598724,
0.0693083181977272,
-0.04381299018859863,
-0.038738906383514404,
-0.015374792739748955,
-0.0029897931963205338,
0.14953550696372986,
0.06494788080453873,
0.11409911513328552,
-0.07472079247236252,
0.010434641502797604,
0.010832404717803001,
-0.008651630952954292,
0.016917014494538307,
0.02875317819416523,
-0.061046965420246124,
0.11194391548633575,
0.03876045346260071,
0.02736404910683632,
-0.24566538631916046,
-0.06682449579238892,
-0.011323003098368645,
0.1456235647201538,
-0.02446315996348858,
0.10121438652276993,
0.045561324805021286,
-0.0504569448530674,
-0.010978372767567635,
-0.029581138864159584,
0.17828664183616638,
0.022655870765447617,
-0.06422974169254303,
-0.0612877793610096,
0.0651540756225586,
-0.07150227576494217,
0.012235969305038452,
-0.07036937773227692,
0.020627280697226524,
0.08606486022472382,
0.12204300612211227,
0.010734139941632748,
0.06475073099136353,
-0.02579765021800995,
0.07209211587905884,
-0.07104320824146271,
0.019227510318160057,
0.11117701232433319,
-0.010595849715173244,
-0.07011682540178299,
0.13524381816387177,
-0.1373681277036667,
0.2673107087612152,
0.19483336806297302,
-0.2967563271522522,
0.0005786092369817197,
-0.04439404606819153,
-0.007282515522092581,
0.030610160902142525,
0.030042126774787903,
0.014859852381050587,
0.08437592536211014,
0.0014727829257026315,
0.20341786742210388,
-0.021047484129667282,
-0.03919289633631706,
-0.018922755494713783,
-0.04877391830086708,
-0.03148360177874565,
0.08788784593343735,
0.06451795995235443,
-0.192406564950943,
0.19050060212612152,
0.21731194853782654,
0.010114802047610283,
0.16024211049079895,
-0.010486523620784283,
0.043989237397909164,
0.09252246469259262,
-0.03757351264357567,
-0.024272754788398743,
-0.08932791650295258,
-0.1848243772983551,
-0.03918878361582756,
0.07472185045480728,
0.03010893426835537,
0.06895712018013,
-0.10219920426607132,
-0.027038687840104103,
0.0004840063920710236,
0.021132981404662132,
-0.01947878859937191,
0.08704918622970581,
0.08203180879354477,
0.1052171140909195,
-0.017219819128513336,
-0.07267280668020248,
0.11330383270978928,
-0.0011106154415756464,
-0.07149384170770645,
0.18412140011787415,
-0.15954560041427612,
-0.36233094334602356,
-0.1530739665031433,
-0.20456592738628387,
-0.02883506752550602,
0.06615062057971954,
0.10685895383358002,
-0.12165717035531998,
-0.048558108508586884,
0.0375000461935997,
-0.013693227432668209,
-0.04041895270347595,
0.03981194645166397,
-0.05303730443120003,
0.07329315692186356,
-0.05222955346107483,
-0.06364883482456207,
-0.06660815328359604,
-0.03131863474845886,
-0.004695216193795204,
0.16393853724002838,
-0.12483653426170349,
0.06658802926540375,
0.1819998174905777,
0.0010995424818247557,
0.06644674390554428,
-0.032483141869306564,
0.1697184294462204,
-0.08651559799909592,
-0.02343188226222992,
0.1893177032470703,
-0.07345744967460632,
0.07808925211429596,
0.15666639804840088,
0.020104380324482918,
-0.0712679922580719,
0.0352557972073555,
-0.035343270748853683,
-0.08934015780687332,
-0.2058166265487671,
-0.1703205555677414,
-0.12546730041503906,
0.05237005278468132,
0.0663270428776741,
0.07582127302885056,
0.12632738053798676,
0.06528977304697037,
0.00627241050824523,
0.010700550861656666,
0.006936580874025822,
0.07483439892530441,
0.24698598682880402,
-0.0010819705203175545,
0.14767786860466003,
-0.057353224605321884,
-0.13245494663715363,
0.08233633637428284,
0.000922833161894232,
0.1185675784945488,
0.08539658784866333,
0.017674902454018593,
0.005295653361827135,
0.05462205410003662,
0.164198637008667,
0.1299368292093277,
0.04298880323767662,
-0.013622048310935497,
-0.01172587089240551,
0.0032578855752944946,
-0.0797785148024559,
0.006457295268774033,
0.07906489074230194,
-0.14195358753204346,
-0.08270972222089767,
-0.11039547622203827,
0.10006770491600037,
0.08380265533924103,
0.042938295751810074,
-0.2052999883890152,
0.005745685659348965,
0.09206069260835648,
-0.027502331882715225,
-0.09957162290811539,
0.06463603675365448,
-0.04812092334032059,
-0.13455109298229218,
0.10769277811050415,
-0.029609164223074913,
0.13354617357254028,
-0.0870715081691742,
0.08272852748632431,
-0.0378170944750309,
-0.11202792823314667,
0.03467349335551262,
0.10786684602499008,
-0.27751585841178894,
0.2031957507133484,
0.007435075007379055,
-0.06144534796476364,
-0.07824365049600601,
-0.015199865214526653,
0.039944443851709366,
0.22591036558151245,
0.06934285908937454,
0.004277070518583059,
-0.05739999935030937,
-0.1865520477294922,
-0.009981787763535976,
-0.008337096311151981,
0.12231403589248657,
-0.03427664935588837,
-0.01814279891550541,
-0.036011241376399994,
-0.030255382880568504,
-0.03578435257077217,
-0.06897740066051483,
0.02666986919939518,
-0.17997102439403534,
0.056329283863306046,
0.034454237669706345,
0.05416429787874222,
0.01469043642282486,
-0.04343695193529129,
-0.11887014657258987,
0.19838201999664307,
-0.10767136514186859,
-0.09184177964925766,
-0.11828504502773285,
-0.07852382957935333,
0.02535579912364483,
-0.08476060628890991,
0.06807194650173187,
-0.08172672241926193,
0.018900277093052864,
-0.06600436568260193,
-0.20524995028972626,
0.11596046388149261,
-0.10182060301303864,
-0.03258875012397766,
-0.058350928127765656,
0.1526644378900528,
-0.07479622215032578,
0.010474151000380516,
0.03318091109395027,
0.02239469438791275,
-0.08559903502464294,
-0.08446884155273438,
-0.018381169065833092,
0.03129338473081589,
0.06142119690775871,
0.08739607781171799,
-0.09792511910200119,
-0.07674866914749146,
-0.03134889155626297,
0.02817792072892189,
0.2929084002971649,
0.1401015967130661,
-0.06586769968271255,
0.1629326492547989,
0.10387758165597916,
-0.06942285597324371,
-0.3373493552207947,
-0.09150945395231247,
-0.09645266830921173,
-0.03972399979829788,
-0.042589932680130005,
-0.16358928382396698,
0.13413257896900177,
-0.004249863792210817,
-0.010055972263216972,
0.08473600447177887,
-0.16361457109451294,
-0.08480892330408096,
0.19654500484466553,
-0.0355062410235405,
0.36373743414878845,
-0.09189414978027344,
-0.09806639701128006,
-0.07035496085882187,
-0.1232207641005516,
0.12262474000453949,
0.007738110609352589,
0.08150525391101837,
-0.02050303854048252,
0.04451111704111099,
0.04815887659788132,
-0.03690929710865021,
0.10097026824951172,
0.036669690161943436,
0.025901002809405327,
-0.11938466131687164,
-0.09219347685575485,
0.023168733343482018,
-0.019243339076638222,
-0.007111898623406887,
-0.01547485776245594,
0.01685570739209652,
-0.17164339125156403,
-0.04131095111370087,
-0.07032524049282074,
0.05912882834672928,
0.04161927476525307,
-0.029813537374138832,
0.012351144105196,
-0.020498499274253845,
-0.000361355283530429,
0.006620287895202637,
0.251852810382843,
-0.03737054020166397,
0.1604781597852707,
0.08527542650699615,
0.141584113240242,
-0.15723979473114014,
0.01194052491337061,
-0.07652142643928528,
-0.05061504244804382,
0.06191904842853546,
-0.06635212153196335,
0.07575498521327972,
0.13591395318508148,
-0.05730273202061653,
0.07247055321931839,
0.11612356454133987,
0.07706465572118759,
-0.034392137080430984,
0.16330119967460632,
-0.2292891889810562,
0.04589579999446869,
-0.050483379513025284,
-0.033954232931137085,
0.06465915590524673,
0.0655360221862793,
0.1258573830127716,
0.06694923341274261,
-0.04017629101872444,
0.005630772560834885,
0.00028037314768880606,
0.005372054409235716,
0.07443340867757797,
0.04748379439115524,
0.04316747188568115,
-0.14709694683551788,
0.05031560733914375,
0.05119774490594864,
-0.15819577872753143,
-0.022534551098942757,
0.1376158893108368,
-0.1704932600259781,
-0.1271103173494339,
-0.021827740594744682,
0.12368015199899673,
-0.09311434626579285,
-0.046253565698862076,
-0.07048245519399643,
-0.13402129709720612,
0.07112511247396469,
0.18836617469787598,
0.12805050611495972,
0.09663103520870209,
-0.06118634715676308,
-0.04969988390803337,
0.0036050756461918354,
-0.004089095629751682,
0.017009761184453964,
0.03120747022330761,
-0.12284451723098755,
0.046005018055438995,
-0.02090919390320778,
0.15390309691429138,
-0.09199176728725433,
-0.07624588906764984,
-0.1582917422056198,
0.04238278418779373,
-0.09195777773857117,
-0.023019742220640182,
-0.09330286085605621,
-0.01648246869444847,
0.0030273916199803352,
-0.030272169038653374,
-0.026145517826080322,
-0.06213071197271347,
-0.11623096466064453,
0.04011767357587814,
-0.028817979618906975,
0.04146858677268028,
-0.06920336186885834,
-0.04603973776102066,
0.09102679789066315,
-0.03833403438329697,
0.10358903557062149,
0.10654495656490326,
-0.0914529487490654,
0.0934479758143425,
-0.14121071994304657,
-0.1319282501935959,
0.1433861404657364,
0.030263781547546387,
0.07207431644201279,
0.07694290578365326,
0.03595962002873421,
0.07349478453397751,
0.004535248037427664,
0.06631990522146225,
0.06761990487575531,
-0.12337382882833481,
0.061452679336071014,
-0.046973392367362976,
-0.17189696431159973,
-0.05778007209300995,
-0.04047338292002678,
0.10660306364297867,
0.010234192945063114,
0.1720496565103531,
-0.05692226439714432,
0.1017514169216156,
-0.03180769085884094,
0.0038062711246311665,
-0.01604292169213295,
-0.20698778331279755,
-0.06364472210407257,
-0.08077114075422287,
0.026275143027305603,
0.005083381198346615,
0.23303534090518951,
0.061751753091812134,
0.033835094422101974,
0.04869496077299118,
0.09752455353736877,
-0.0014774927403777838,
0.023545393720269203,
0.17794077098369598,
0.10133370757102966,
-0.05567975342273712,
-0.05575546622276306,
0.05616139620542526,
0.029215605929493904,
0.006353367585688829,
0.14132826030254364,
0.07252193242311478,
-0.041009921580553055,
0.07551323622465134,
-0.03376345708966255,
0.04427867755293846,
-0.1321653574705124,
-0.16054923832416534,
-0.05143791809678078,
0.07023841142654419,
0.01740087941288948,
0.03448288515210152,
0.07088012248277664,
-0.028410857543349266,
0.05220868065953255,
-0.033101536333560944,
-0.05869230628013611,
-0.18244294822216034,
-0.09428907185792923,
-0.09423913061618805,
-0.09753676503896713,
0.0058974651619791985,
-0.07943454384803772,
-0.01026046834886074,
0.06547573953866959,
0.037508975714445114,
-0.05198773369193077,
0.07752657681703568,
0.003285798244178295,
-0.05593571066856384,
0.08687124401330948,
-0.045962750911712646,
0.02649652026593685,
0.00841206219047308,
-0.029707664623856544,
-0.1380927860736847,
-0.013390704058110714,
-0.04401649907231331,
0.040850814431905746,
-0.058590736240148544,
0.007230483461171389,
-0.1483704298734665,
-0.12039808928966522,
-0.019934508949518204,
0.0580129399895668,
-0.06074916571378708,
0.14175079762935638,
0.015395265072584152,
0.00611070916056633,
0.047287240624427795,
0.17810532450675964,
-0.0544942207634449,
-0.06539076566696167,
-0.04489162564277649,
0.24079638719558716,
0.09303659200668335,
0.10803006589412689,
0.0026883413083851337,
-0.013426939956843853,
-0.07931891828775406,
0.28847232460975647,
0.27526742219924927,
-0.04996299743652344,
0.054827310144901276,
0.007495596073567867,
0.03283945098519325,
0.15242771804332733,
0.1401364952325821,
0.09061526507139206,
0.24117816984653473,
-0.0521743968129158,
-0.05017128586769104,
-0.026741530746221542,
-0.03419290855526924,
-0.13402216136455536,
0.0581725612282753,
0.05382576957345009,
-0.0488528348505497,
-0.06285785138607025,
0.10921014845371246,
-0.21934591233730316,
0.16537490487098694,
0.019078493118286133,
-0.20565392076969147,
-0.06819386035203934,
-0.03284084051847458,
0.1365688294172287,
-0.0016830840613692999,
0.07499389350414276,
-0.00323955318890512,
-0.11883772164583206,
0.042848069220781326,
0.01306091621518135,
-0.20812170207500458,
-0.0041817850433290005,
0.06021128222346306,
-0.05781300365924835,
-0.0120098190382123,
-0.02640264853835106,
0.03757385164499283,
0.06560133397579193,
0.07958315312862396,
-0.0117155397310853,
0.04959989711642265,
-0.012948726303875446,
-0.030828366056084633,
0.029231732711195946,
0.02946082502603531,
0.0038178605027496815,
-0.09871038049459457,
0.06783884763717651,
-0.16667571663856506,
0.0549757145345211,
-0.05383889377117157,
-0.05352160334587097,
-0.019258368760347366,
0.04339629411697388,
-0.05456918105483055,
0.04438189044594765,
0.10450860112905502,
0.011940731666982174,
-0.025312455371022224,
-0.04523419588804245,
-0.04262804985046387,
-0.012295196764171124,
-0.1369558572769165,
-0.14967197179794312,
-0.09997987747192383,
-0.08965370059013367,
0.09313849359750748,
0.0034958450123667717,
-0.12975360453128815,
-0.006513827480375767,
-0.11122267693281174,
0.05365913361310959,
-0.16868756711483002,
0.09322161972522736,
0.0323028489947319,
0.015595607459545135,
-0.011563225649297237,
-0.040581803768873215,
0.04532773047685623,
0.07905946671962738,
-0.1267605572938919,
-0.08728102594614029
] |
null | null |
transformers
|
# DistilBert Dummy Sentiment Model
## Purpose
This is a dummy model that can be used for testing the transformers `pipeline` with the task `sentiment-analysis`. It should always give random results (i.e. `{"label": "negative", "score": 0.5}`).
## How to use
```python
classifier = pipeline("sentiment-analysis", "dhpollack/distilbert-dummy-sentiment")
results = classifier(["this is a test", "another test"])
```
## Notes
This was created as follows:
1. Create a vocab.txt file (in /tmp/vocab.txt in this example).
```
[UNK]
[SEP]
[PAD]
[CLS]
[MASK]
```
2. Open a python shell:
```python
import transformers
config = transformers.DistilBertConfig(vocab_size=5, n_layers=1, n_heads=1, dim=1, hidden_dim=4 * 1, num_labels=2, id2label={0: "negative", 1: "positive"}, label2id={"negative": 0, "positive": 1})
model = transformers.DistilBertForSequenceClassification(config)
tokenizer = transformers.DistilBertTokenizer("/tmp/vocab.txt", model_max_length=512)
config.save_pretrained(".")
model.save_pretrained(".")
tokenizer.save_pretrained(".")
```
|
{"language": ["multilingual", "en"], "tags": ["sentiment-analysis", "testing", "unit tests"]}
|
text-classification
|
dhpollack/distilbert-dummy-sentiment
|
[
"transformers",
"pytorch",
"distilbert",
"text-classification",
"sentiment-analysis",
"testing",
"unit tests",
"multilingual",
"en",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"multilingual",
"en"
] |
TAGS
#transformers #pytorch #distilbert #text-classification #sentiment-analysis #testing #unit tests #multilingual #en #autotrain_compatible #endpoints_compatible #region-us
|
# DistilBert Dummy Sentiment Model
## Purpose
This is a dummy model that can be used for testing the transformers 'pipeline' with the task 'sentiment-analysis'. It should always give random results (i.e. '{"label": "negative", "score": 0.5}').
## How to use
## Notes
This was created as follows:
1. Create a URL file (in /tmp/URL in this example).
2. Open a python shell:
|
[
"# DistilBert Dummy Sentiment Model",
"## Purpose\n\nThis is a dummy model that can be used for testing the transformers 'pipeline' with the task 'sentiment-analysis'. It should always give random results (i.e. '{\"label\": \"negative\", \"score\": 0.5}').",
"## How to use",
"## Notes\n\nThis was created as follows:\n\n1. Create a URL file (in /tmp/URL in this example).\n\n\n\n2. Open a python shell:"
] |
[
"TAGS\n#transformers #pytorch #distilbert #text-classification #sentiment-analysis #testing #unit tests #multilingual #en #autotrain_compatible #endpoints_compatible #region-us \n",
"# DistilBert Dummy Sentiment Model",
"## Purpose\n\nThis is a dummy model that can be used for testing the transformers 'pipeline' with the task 'sentiment-analysis'. It should always give random results (i.e. '{\"label\": \"negative\", \"score\": 0.5}').",
"## How to use",
"## Notes\n\nThis was created as follows:\n\n1. Create a URL file (in /tmp/URL in this example).\n\n\n\n2. Open a python shell:"
] |
[
57,
10,
64,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #sentiment-analysis #testing #unit tests #multilingual #en #autotrain_compatible #endpoints_compatible #region-us \n# DistilBert Dummy Sentiment Model## Purpose\n\nThis is a dummy model that can be used for testing the transformers 'pipeline' with the task 'sentiment-analysis'. It should always give random results (i.e. '{\"label\": \"negative\", \"score\": 0.5}').## How to use## Notes\n\nThis was created as follows:\n\n1. Create a URL file (in /tmp/URL in this example).\n\n\n\n2. Open a python shell:"
] |
[
-0.05684439092874527,
-0.03230179846286774,
-0.006328945979475975,
0.05696459859609604,
0.11259806901216507,
0.02298140339553356,
0.10383247584104538,
0.03543538227677345,
0.05520784854888916,
0.0018696869956329465,
0.19897252321243286,
0.1586330085992813,
-0.061602380126714706,
0.02831428498029709,
-0.037470608949661255,
-0.13001178205013275,
0.06379128247499466,
0.027059393003582954,
0.010227653197944164,
0.18582473695278168,
0.030362717807292938,
-0.05505184829235077,
0.06157214939594269,
0.01782911829650402,
-0.1696074903011322,
0.00540568633005023,
0.009584074839949608,
-0.033423565328121185,
0.06676008552312851,
0.0005958249676041305,
0.12068939954042435,
0.05302930995821953,
-0.028814172372221947,
-0.08984646201133728,
0.02543201856315136,
0.006673918571323156,
0.043047092854976654,
0.032186802476644516,
-0.0027114497497677803,
-0.07381334155797958,
0.1424257606267929,
0.09001592546701431,
0.057373132556676865,
0.07008732110261917,
-0.05133940279483795,
-0.20595389604568481,
0.005177277140319347,
0.023930691182613373,
-0.017191557213664055,
0.1188960075378418,
-0.024104153737425804,
0.25991833209991455,
-0.16213379800319672,
0.12220367789268494,
0.25103145837783813,
-0.027191700413823128,
-0.045111361891031265,
0.11231070011854172,
0.04812587425112724,
-0.020115332677960396,
0.0329677015542984,
-0.048041652888059616,
0.05916154757142067,
0.08061113953590393,
0.0019843378104269505,
-0.09421270340681076,
0.013290083035826683,
-0.02476447820663452,
-0.15769734978675842,
-0.09994906932115555,
0.14677534997463226,
-0.016196904703974724,
-0.0868987962603569,
0.016847841441631317,
-0.05651956424117088,
-0.017835043370723724,
0.020665286108851433,
-0.022449716925621033,
-0.051343534141778946,
-0.014608300291001797,
-0.008517053909599781,
0.05991802364587784,
-0.055678073316812515,
-0.12474297732114792,
-0.048549506813287735,
0.2485717386007309,
0.02790018916130066,
0.054593149572610855,
-0.08544167876243591,
0.10426325350999832,
0.012557800859212875,
-0.07961325347423553,
-0.06305448710918427,
-0.08672410249710083,
-0.021227803081274033,
0.020881287753582,
-0.1020965576171875,
-0.12527112662792206,
0.10231104493141174,
0.06709016859531403,
0.01767323724925518,
0.06602030992507935,
0.022434143349528313,
0.05364951118826866,
0.05698900297284126,
0.1458464115858078,
0.020849190652370453,
0.18674761056900024,
-0.00428420165553689,
0.03250909224152565,
-0.04968008026480675,
-0.016533799469470978,
0.0085511589422822,
-0.02050492912530899,
0.10737519711256027,
0.06576163321733475,
0.06337873637676239,
0.11134304851293564,
-0.09619665890932083,
-0.0364769846200943,
-0.0678812563419342,
-0.1120147630572319,
-0.004457463510334492,
0.06999508291482925,
-0.044208381325006485,
0.15296538174152374,
0.10816124081611633,
-0.07884358614683151,
-0.0703013464808464,
0.052078165113925934,
-0.015954675152897835,
0.027535216882824898,
-0.13681522011756897,
-0.15342825651168823,
0.0344204306602478,
0.03704715892672539,
-0.015745382755994797,
-0.1477489024400711,
-0.211137592792511,
-0.05392839014530182,
0.10252454876899719,
0.026828380301594734,
0.02624361403286457,
-0.046324606984853745,
-0.08219249546527863,
-0.025285255163908005,
-0.0008631193777546287,
-0.07402556389570236,
-0.029355166479945183,
0.11081378161907196,
0.0377202071249485,
0.07325592637062073,
-0.0315069705247879,
0.02543015591800213,
-0.11692681908607483,
0.05812700092792511,
-0.13977931439876556,
0.11936111748218536,
-0.11771126091480255,
0.08470568060874939,
-0.09234678745269775,
-0.11450550705194473,
0.008613638579845428,
0.0008070773910731077,
0.04753607138991356,
0.22304850816726685,
-0.11777541041374207,
-0.03365088999271393,
0.04266407713294029,
-0.13124217092990875,
-0.11847928166389465,
0.168802410364151,
-0.05609000846743584,
0.1773219108581543,
0.056052424013614655,
0.019269831478595734,
0.09356991946697235,
-0.1385750025510788,
0.08353804796934128,
0.059708371758461,
-0.1500786393880844,
0.05119038745760918,
0.008393537253141403,
0.028714340180158615,
-0.2253979593515396,
0.031045524403452873,
-0.1038750633597374,
0.0757727175951004,
-0.08312130719423294,
-0.03400522097945213,
-0.05477385222911835,
-0.03552388399839401,
0.14980585873126984,
0.06090761721134186,
0.023262793198227882,
-0.010214433073997498,
-0.10176422446966171,
0.06998676806688309,
0.042691607028245926,
-0.07408738881349564,
0.018459344282746315,
-0.10772652924060822,
0.09442348778247833,
0.006356751546263695,
0.02478538639843464,
-0.20031265914440155,
-0.012461025267839432,
0.020466715097427368,
0.026338618248701096,
0.0520433746278286,
0.07897435873746872,
0.03685412555932999,
-0.04857306554913521,
0.0678996741771698,
-0.050135884433984756,
0.004690933972597122,
0.009478925727307796,
-0.04517669603228569,
-0.10623712837696075,
-0.03799069672822952,
-0.030450742691755295,
-0.044953856617212296,
-0.0479760505259037,
0.05501708760857582,
0.031134692952036858,
-0.020541010424494743,
-0.03414931520819664,
0.01316547766327858,
-0.015169179998338223,
0.05200028046965599,
-0.05238496884703636,
0.00032066533458419144,
0.028424430638551712,
-0.010878527536988258,
-0.08305615931749344,
0.008280009962618351,
-0.042307764291763306,
0.03452802449464798,
0.11124242097139359,
-0.16263268887996674,
-0.1366773247718811,
0.13483983278274536,
-0.030468441545963287,
0.015402067452669144,
-0.011424285359680653,
-0.029865628108382225,
0.0530465692281723,
0.01580520160496235,
0.05394284427165985,
-0.06214186176657677,
0.020718639716506004,
0.009016228839755058,
-0.0834137499332428,
0.061095889657735825,
0.10612142086029053,
0.018696269020438194,
-0.009302672930061817,
0.03477770462632179,
0.02376902848482132,
-0.19138437509536743,
-0.01364090759307146,
-0.027934862300753593,
-0.016374023631215096,
-0.0285524670034647,
0.103112131357193,
0.023951373994350433,
0.0006921928143128753,
-0.17548304796218872,
0.019990993663668633,
0.04665104299783707,
-0.006266776937991381,
-0.0158863365650177,
-0.1050717756152153,
0.0074781146831810474,
0.03703785687685013,
-0.012246654368937016,
-0.10325980186462402,
0.06392029672861099,
0.006795994471758604,
0.006143955979496241,
-0.010556901805102825,
-0.07411552220582962,
0.08690517395734787,
0.006785562261939049,
-0.1485021859407425,
0.21926091611385345,
-0.008661499246954918,
-0.16640450060367584,
-0.1160116195678711,
-0.09802522510290146,
-0.011586997658014297,
0.01895664632320404,
0.1049385741353035,
-0.09043411910533905,
-0.04467812925577164,
-0.033474233001470566,
0.15201152861118317,
0.007382549811154604,
0.060715533792972565,
-0.0008524617296643555,
-0.031686931848526,
0.06001349911093712,
-0.07443702220916748,
-0.000999463489279151,
-0.00737354438751936,
0.01770835928618908,
0.11995775252580643,
-0.15310387313365936,
0.06156096234917641,
0.15930962562561035,
-0.02131684683263302,
0.04778718575835228,
0.004122518468648195,
0.2967338562011719,
-0.011064443737268448,
-0.03938442841172218,
0.17015840113162994,
-0.09547953307628632,
0.06179485097527504,
0.04948355257511139,
0.027963345870375633,
-0.060289643704891205,
0.0802154392004013,
-0.03223448619246483,
-0.09320692718029022,
-0.18086901307106018,
-0.05257966369390488,
-0.03810356557369232,
0.09444797784090042,
0.04477175697684288,
-0.0021106002386659384,
0.009264986030757427,
0.14409838616847992,
0.0698123648762703,
0.013922875747084618,
-0.03260347247123718,
0.11553532630205154,
0.21279555559158325,
-0.0041602118872106075,
0.060687027871608734,
-0.03572579100728035,
-0.04177619144320488,
-0.009108630008995533,
-0.13156601786613464,
0.14603735506534576,
-0.03507637605071068,
0.01070130430161953,
0.11378249526023865,
0.09524579346179962,
0.05298112332820892,
0.14509770274162292,
-0.00402245344594121,
0.0024531548842787743,
0.033052876591682434,
-0.06312888115644455,
-0.13408386707305908,
0.0287222508341074,
-0.03613919019699097,
0.06067923083901405,
-0.06328465044498444,
0.020863695070147514,
0.02202196605503559,
0.12614668905735016,
0.03623568266630173,
-0.3025561571121216,
-0.14137083292007446,
-0.08593648672103882,
-0.025173896923661232,
-0.028539611026644707,
0.007338923402130604,
-0.03780260682106018,
-0.10214361548423767,
-0.015623778104782104,
-0.07476473599672318,
0.1070503294467926,
0.03081987425684929,
0.07804632186889648,
-0.03482378274202347,
0.006868706550449133,
-0.006791411433368921,
0.07297239452600479,
-0.2580818831920624,
0.16339081525802612,
-0.029607610777020454,
0.02213660627603531,
-0.055147185921669006,
-0.02474479004740715,
0.018056415021419525,
0.10419469326734543,
0.07114019244909286,
-0.04093347489833832,
0.16339685022830963,
-0.11260797828435898,
0.022146709263324738,
0.050021614879369736,
0.0322827473282814,
0.09222879260778427,
0.11005902290344238,
0.018759440630674362,
0.036844875663518906,
0.04129216447472572,
0.08039242774248123,
-0.14312157034873962,
0.010805999860167503,
0.002723844489082694,
0.010505933314561844,
0.029950547963380814,
-0.041505128145217896,
-0.07926058024168015,
-0.022791428491473198,
0.16793958842754364,
-0.035705920308828354,
-0.08957674354314804,
-0.08581643551588058,
0.0495525486767292,
0.10375253856182098,
-0.08005044609308243,
-0.009678533300757408,
-0.07154987007379532,
-0.022704215720295906,
0.02534201554954052,
-0.12023015320301056,
0.07281691581010818,
-0.09682346135377884,
-0.07699534296989441,
-0.018165305256843567,
0.007623191457241774,
0.1098875030875206,
0.05372817441821098,
-0.00872027687728405,
-0.020611586049199104,
-0.07592438161373138,
-0.13504672050476074,
-0.0384095273911953,
0.026576124131679535,
-0.13246378302574158,
0.14165565371513367,
0.02977588027715683,
-0.029864458367228508,
-0.05333209037780762,
0.07796665281057358,
0.08188125491142273,
0.0997677743434906,
-0.037822041660547256,
0.05981561914086342,
0.07769262045621872,
-0.03377082571387291,
-0.2786305546760559,
0.014695283956825733,
0.11249415576457977,
0.038548603653907776,
-0.0041213249787688255,
-0.10791777819395065,
0.13034456968307495,
-0.01578398235142231,
-0.001670515164732933,
0.010173001326620579,
-0.314492404460907,
-0.08116543292999268,
0.21295490860939026,
0.08341675251722336,
0.18481844663619995,
-0.12376971542835236,
-0.04018727317452431,
-0.021354926750063896,
-0.23807044327259064,
0.06354869902133942,
-0.03388848900794983,
0.06081963703036308,
-0.034760233014822006,
0.1524629443883896,
0.037206705659627914,
-0.06200790777802467,
0.11526758968830109,
0.011576482094824314,
0.030668461695313454,
-0.03840532526373863,
-0.044778332114219666,
0.06466864794492722,
-0.01912728138267994,
0.1450919210910797,
-0.032516613602638245,
0.05604180321097374,
-0.0328049473464489,
-0.07998399436473846,
-0.10362155735492706,
0.08946732431650162,
-0.020068198442459106,
-0.04062044247984886,
-0.09579607844352722,
-0.05699312314391136,
0.06468967348337173,
0.0020919954404234886,
-0.002835921011865139,
-0.0567447766661644,
0.06409447640180588,
0.25681719183921814,
0.10457763075828552,
-0.04464706405997276,
-0.1333223432302475,
0.07337645441293716,
-0.021868905052542686,
0.11651773750782013,
-0.13055568933486938,
0.007899712771177292,
0.11669807136058807,
0.05445081368088722,
0.08649873733520508,
0.09698504954576492,
-0.08780576288700104,
0.007488231640309095,
-0.005172755103558302,
-0.09359478950500488,
-0.10345370322465897,
0.025864971801638603,
-0.00936390832066536,
-0.13316741585731506,
0.032380178570747375,
0.0807344913482666,
-0.11904338002204895,
0.0280756838619709,
0.029779285192489624,
-0.049856070429086685,
-0.06314319372177124,
0.11446093767881393,
0.025740064680576324,
0.08803447335958481,
-0.08396978676319122,
-0.06306847929954529,
-0.032681990414857864,
-0.07478578388690948,
-0.016425881534814835,
0.05117422714829445,
-0.11054158210754395,
-0.05237481743097305,
0.03997337445616722,
0.10184028744697571,
-0.21219299733638763,
-0.021341316401958466,
-0.12222445756196976,
-0.13236047327518463,
0.054026197642087936,
0.1672435700893402,
0.16232633590698242,
0.0984690710902214,
-0.07136072218418121,
0.024448473006486893,
-0.05004442110657692,
0.0436599925160408,
0.13243906199932098,
0.025708450004458427,
-0.13181552290916443,
0.09316527843475342,
0.03324452042579651,
0.113068126142025,
-0.08315709233283997,
-0.05906250327825546,
-0.14317947626113892,
0.08644228428602219,
-0.2396233081817627,
0.04326509311795235,
-0.08419155329465866,
-0.014816773124039173,
0.09716422855854034,
-0.03145092725753784,
0.001081690308637917,
0.046855807304382324,
-0.09020960330963135,
0.011335722170770168,
0.03591551631689072,
-0.0025704947765916586,
-0.07955458760261536,
0.0016815676353871822,
0.06733863800764084,
-0.08255653828382492,
0.04134785756468773,
0.1043868362903595,
-0.1902363896369934,
0.013462516479194164,
-0.17466787993907928,
0.0019090314162895083,
0.10684461146593094,
-0.014937340281903744,
0.027127820998430252,
-0.09064820408821106,
0.014056956395506859,
0.04441200941801071,
0.05156446620821953,
0.06700456142425537,
0.1227216050028801,
-0.09141728281974792,
0.09156038612127304,
0.041533611714839935,
-0.031755708158016205,
-0.06963939219713211,
0.018034886568784714,
-0.02835720032453537,
0.13639971613883972,
0.16562162339687347,
-0.07912853360176086,
0.09971155971288681,
-0.06422467529773712,
-0.0255997646600008,
0.036317501217126846,
0.031221769750118256,
-0.08927168697118759,
-0.012766712345182896,
0.042649634182453156,
-0.01846948452293873,
0.07579238712787628,
-0.03071509674191475,
-0.016805075109004974,
-0.0013009562389925122,
0.010579902678728104,
0.12395399063825607,
-0.01785174198448658,
0.13605931401252747,
-0.026772331446409225,
-0.014165178872644901,
-0.007687851786613464,
-0.010021934285759926,
0.03306097164750099,
-0.07068303227424622,
0.19598065316677094,
0.04909782484173775,
-0.007432274986058474,
0.10795417428016663,
-0.04592122137546539,
0.14929932355880737,
0.006481179967522621,
-0.1155196949839592,
-0.028619535267353058,
0.09858224540948868,
0.02780509740114212,
0.14307835698127747,
0.06862065941095352,
0.04051530733704567,
0.051160167902708054,
-0.033094372600317,
-0.08695061504840851,
-0.11212590336799622,
-0.11392629146575928,
-0.055316925048828125,
-0.22117048501968384,
-0.01422304566949606,
-0.032091785222291946,
-0.046748366206884384,
-0.013181518763303757,
0.038431935012340546,
-0.020185567438602448,
0.18022316694259644,
0.04546748474240303,
-0.03847464546561241,
0.03489302098751068,
-0.1056719571352005,
-0.0972062274813652,
0.047919388860464096,
0.03559506684541702,
0.030079634860157967,
0.03123101405799389,
0.01723269745707512,
0.06929852068424225,
-0.06181272491812706,
0.024979179725050926,
-0.09823358803987503,
-0.07532557100057602,
0.00937835406512022,
0.09593923389911652,
-0.04311126098036766,
0.11488368362188339,
0.00824450422078371,
0.0018816647352650762,
-0.05165977403521538,
0.058958351612091064,
-0.048690471798181534,
-0.13619612157344818,
-0.13234005868434906,
0.39077186584472656,
-0.07980772852897644,
0.027043141424655914,
-0.006265996489673853,
-0.03630981221795082,
-0.07579731196165085,
0.2105707973241806,
0.13489584624767303,
-0.034329403191804886,
-0.002706026192754507,
-0.011653322726488113,
0.002013236517086625,
-0.03703710064291954,
0.08793549239635468,
0.054000090807676315,
0.217129647731781,
-0.061618220061063766,
0.01177199836820364,
-0.09371547400951385,
-0.0460796095430851,
-0.07032410800457001,
-0.15929964184761047,
0.056052908301353455,
-0.07312271744012833,
-0.09460625052452087,
0.13083647191524506,
-0.22261399030685425,
-0.015830211341381073,
0.06568171828985214,
0.06667181104421616,
-0.028317682445049286,
-0.03873763233423233,
0.006864537019282579,
0.009030045010149479,
0.088426373898983,
-0.098968006670475,
-0.019202766939997673,
0.09424226731061935,
-0.0581812858581543,
-0.10192161053419113,
-0.03703920543193817,
0.12700046598911285,
-0.055880703032016754,
0.024356333538889885,
0.021620625630021095,
0.190713033080101,
0.0642431229352951,
0.0033989876974374056,
-0.08902154117822647,
0.07221966236829758,
0.0700535848736763,
-0.07882021367549896,
-0.0160513985902071,
0.04256073012948036,
0.0480843223631382,
0.055873773992061615,
-0.03774735704064369,
-0.13927727937698364,
-0.0041328477673232555,
0.01619039848446846,
0.029284542426466942,
-0.13828925788402557,
-0.04325824975967407,
-0.06206975132226944,
0.0295210350304842,
0.16316527128219604,
-0.04312356933951378,
0.00433407723903656,
-0.024135978892445564,
0.031747255474328995,
-0.0069108158349990845,
-0.1271907091140747,
-0.049023572355508804,
-0.16347341239452362,
0.02241295389831066,
0.03893476352095604,
-0.030355848371982574,
-0.17777754366397858,
-0.044408880174160004,
-0.02168784849345684,
-0.0067890118807554245,
0.06699977815151215,
0.03639213740825653,
-0.05571041256189346,
0.04725421592593193,
0.004768766462802887,
-0.10622401535511017,
-0.020735250785946846,
0.17807744443416595,
-0.08394647389650345,
-0.11467685550451279
] |
null | null |
transformers
|
### TUNiB-Electra Stereotype Detector
Finetuned TUNiB-Electra base with K-StereoSet.
Original Code: https://github.com/newfull5/Stereotype-Detector
|
{}
|
text-classification
|
dhtocks/tunib-electra-stereotype-classifier
|
[
"transformers",
"pytorch",
"electra",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #electra #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
### TUNiB-Electra Stereotype Detector
Finetuned TUNiB-Electra base with K-StereoSet.
Original Code: URL
|
[
"### TUNiB-Electra Stereotype Detector\n\nFinetuned TUNiB-Electra base with K-StereoSet.\n\nOriginal Code: URL"
] |
[
"TAGS\n#transformers #pytorch #electra #text-classification #autotrain_compatible #endpoints_compatible #region-us \n",
"### TUNiB-Electra Stereotype Detector\n\nFinetuned TUNiB-Electra base with K-StereoSet.\n\nOriginal Code: URL"
] |
[
37,
39
] |
[
"passage: TAGS\n#transformers #pytorch #electra #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### TUNiB-Electra Stereotype Detector\n\nFinetuned TUNiB-Electra base with K-StereoSet.\n\nOriginal Code: URL"
] |
[
-0.03850841149687767,
0.030270999297499657,
-0.005522241350263357,
0.025306399911642075,
0.14704973995685577,
0.000690627028234303,
0.10828276723623276,
0.07555967569351196,
0.015202726237475872,
-0.012094279751181602,
0.07717158645391464,
0.17315149307250977,
0.009876166470348835,
0.14836075901985168,
-0.04614526405930519,
-0.24554356932640076,
0.014703435823321342,
0.1316354125738144,
0.0173940509557724,
0.0931360051035881,
0.10508871078491211,
-0.0866813063621521,
0.04543071240186691,
0.00794524047523737,
-0.15248805284500122,
0.07895483076572418,
0.04015423357486725,
-0.16841702163219452,
0.08337362855672836,
-0.0005128532648086548,
0.17643414437770844,
0.054630935192108154,
-0.008201572112739086,
-0.06924039870500565,
0.037121571600437164,
-0.042397864162921906,
-0.03383679687976837,
0.038452040404081345,
0.03134933114051819,
-0.07128914445638657,
0.05538376420736313,
-0.0024042236618697643,
0.07614557445049286,
-0.009980599395930767,
-0.018646355718374252,
-0.04731445387005806,
0.012997058220207691,
0.09352797269821167,
0.06833086907863617,
0.05499599128961563,
-0.008699620142579079,
0.07482092082500458,
-0.09652192890644073,
0.11566852033138275,
0.10953263193368912,
-0.11488237231969833,
0.027386009693145752,
-0.09977611899375916,
-0.01898854412138462,
0.014285348355770111,
-0.060534823685884476,
-0.009271156042814255,
0.03156592696905136,
0.038821231573820114,
0.017654798924922943,
-0.08140873908996582,
-0.18454428017139435,
0.010709086433053017,
-0.11636067926883698,
0.041878338903188705,
0.1657661646604538,
-0.02001689001917839,
0.02849874459207058,
0.03924574702978134,
-0.07563480734825134,
-0.11358406394720078,
-0.018686562776565552,
0.023611128330230713,
-0.08703166246414185,
-0.0447995588183403,
-0.08628140389919281,
0.033927228301763535,
-0.07771400362253189,
-0.0369163416326046,
-0.21630802750587463,
0.2426266223192215,
0.05374254658818245,
0.032149363309144974,
-0.12002138793468475,
0.018292497843503952,
-0.030888933688402176,
-0.09742780774831772,
0.03977871686220169,
-0.034292928874492645,
0.05377938225865364,
-0.049741923809051514,
-0.0489649698138237,
-0.1209312453866005,
0.10295070707798004,
0.20226867496967316,
0.05737842246890068,
0.042006708681583405,
-0.015460516326129436,
0.023888850584626198,
0.04873745143413544,
0.09389494359493256,
-0.012918749824166298,
0.029623011127114296,
0.10947667062282562,
-0.01122226845473051,
0.06786973774433136,
-0.062068235129117966,
-0.1200394406914711,
-0.06979845464229584,
0.22240860760211945,
0.03708263486623764,
-0.011523997411131859,
0.08890952914953232,
-0.04624781012535095,
-0.05293702334165573,
0.03157839924097061,
-0.13604432344436646,
0.0011914793867617846,
0.023913558572530746,
-0.012870904989540577,
0.02238098904490471,
0.007344829849898815,
0.006083829794079065,
-0.026944464072585106,
0.08736860752105713,
-0.03848351910710335,
0.014699535444378853,
-0.010006913915276527,
-0.12023115158081055,
0.01776370219886303,
-0.1241341382265091,
0.07930532842874527,
-0.22050602734088898,
-0.07283473014831543,
0.024534102529287338,
-0.02111966907978058,
0.044696319848299026,
-0.03115485981106758,
-0.04095207899808884,
-0.054230619221925735,
0.0875687375664711,
-0.02893778681755066,
-0.07955288141965866,
-0.045771390199661255,
0.09275174140930176,
0.07019689679145813,
0.1247166246175766,
-0.13217438757419586,
0.03993605822324753,
-0.11986009776592255,
-0.030406158417463303,
-0.2021925002336502,
0.0523550808429718,
-0.08301433175802231,
0.027768392115831375,
-0.031993769109249115,
-0.022420918568968773,
-0.07672197371721268,
0.030808139592409134,
0.027018696069717407,
0.07954666018486023,
-0.11398682743310928,
-0.11650995910167694,
0.08535949885845184,
-0.14162476360797882,
-0.07817122340202332,
0.06967456638813019,
-0.006571060512214899,
-0.011603301391005516,
0.0675259381532669,
0.24527469277381897,
0.08453696966171265,
-0.06632915884256363,
-0.019348997622728348,
0.007087807171046734,
-0.10598182678222656,
-0.0207041185349226,
-0.0014825814869254827,
0.0021278141066432,
-0.06882771104574203,
0.009511427953839302,
-0.0037158974446356297,
0.10369590669870377,
-0.034670598804950714,
-0.036696549504995346,
0.0011392614105716348,
-0.03278867155313492,
0.1500544250011444,
0.00017989985644817352,
0.07512958347797394,
-0.06572878360748291,
-0.02553711272776127,
0.15052351355552673,
0.05988100916147232,
-0.044053398072719574,
0.015593277290463448,
-0.13917860388755798,
0.09569619596004486,
-0.040165577083826065,
-0.007577145006507635,
-0.16536970436573029,
-0.0010371941607445478,
-0.02729172073304653,
0.0674629658460617,
0.07837836444377899,
-0.05030597373843193,
0.029392411932349205,
-0.01656690426170826,
-0.022291626781225204,
-0.07678954303264618,
0.04117962718009949,
0.04780968278646469,
-0.07310118526220322,
-0.08949320763349533,
0.06996568292379379,
-0.009381575509905815,
-0.013842841610312462,
-0.0018247633706778288,
0.03659871965646744,
0.10895967483520508,
0.0933629497885704,
-0.01727980375289917,
0.10278628021478653,
-0.0050123510882258415,
0.0765388160943985,
-0.07327444851398468,
0.005641493014991283,
0.052523963153362274,
-0.018849406391382217,
-0.0413128063082695,
0.10159765183925629,
-0.13116642832756042,
0.08086083084344864,
0.19200041890144348,
-0.1795518398284912,
-0.04038789123296738,
0.069469153881073,
0.025919854640960693,
0.016833830624818802,
0.013100674375891685,
-0.0639229267835617,
0.16395504772663116,
-0.0643366202712059,
0.08133969455957413,
-0.0123903201892972,
0.014664053916931152,
-0.002371231559664011,
-0.061745285987854004,
-0.018659818917512894,
0.08109299838542938,
0.12087944895029068,
-0.21221423149108887,
0.09540476649999619,
0.06484178453683853,
-0.16941115260124207,
0.1322420835494995,
-0.04314048960804939,
0.013270741328597069,
0.05999961495399475,
-0.08346840739250183,
0.0031948876567184925,
0.05080961436033249,
-0.09308256208896637,
-0.033957287669181824,
0.050576042383909225,
-0.07665722817182541,
0.031314827501773834,
-0.11979015171527863,
0.012315705418586731,
-0.017312929034233093,
0.020123306661844254,
-0.0379541739821434,
0.06848306953907013,
0.034358005970716476,
0.09067527949810028,
-0.023254593834280968,
-0.11514364182949066,
0.08466234058141708,
0.02095971442759037,
-0.10702407360076904,
0.15906254947185516,
-0.15202857553958893,
-0.2172616422176361,
-0.12150794267654419,
-0.03754981979727745,
0.018176058307290077,
0.05249866843223572,
0.09720920771360397,
-0.11380515247583389,
-0.010097543708980083,
-0.03103892132639885,
0.09684336185455322,
0.03680001571774483,
0.12924064695835114,
-0.007112258113920689,
-0.012369275093078613,
-0.04392765089869499,
-0.05639093369245529,
-0.024733245372772217,
-0.027772223576903343,
0.03616886958479881,
0.14483051002025604,
-0.08531749248504639,
0.10837356746196747,
0.22267118096351624,
-0.08343086391687393,
-0.006273905746638775,
-0.03333095461130142,
0.13162286579608917,
-0.046239711344242096,
0.04337083175778389,
0.1369163990020752,
-0.053840260952711105,
0.03219776600599289,
0.17340728640556335,
-0.02428627386689186,
0.024563923478126526,
0.03893858566880226,
0.011430950835347176,
-0.0723995491862297,
-0.12468509376049042,
-0.058366090059280396,
-0.012315219268202782,
0.018931766971945763,
0.07045777142047882,
0.046148765832185745,
0.11494356393814087,
0.10007481276988983,
0.026641810312867165,
-0.021419672295451164,
0.05241115018725395,
0.05078534409403801,
0.19016329944133759,
0.07091972231864929,
0.1375068724155426,
-0.059755824506282806,
-0.08697560429573059,
0.03965800255537033,
-0.002305947709828615,
0.12951484322547913,
-0.0314875952899456,
-0.02061443217098713,
0.07577822357416153,
0.03278277441859245,
0.19048725068569183,
0.06296513229608536,
-0.0053176553919911385,
-0.0007140442612580955,
0.03832343965768814,
-0.012950289063155651,
-0.0031182256061583757,
-0.009709659032523632,
0.01972595788538456,
-0.0012235089670866728,
-0.06448359787464142,
-0.048180293291807175,
0.03462823107838631,
-0.02776246704161167,
0.04419821873307228,
-0.2846021056175232,
-0.03461442142724991,
0.03057440184056759,
-0.016845520585775375,
-0.02734461799263954,
0.020886698737740517,
-0.0408790186047554,
-0.046644456684589386,
0.1144399419426918,
-0.015666326507925987,
0.06296069175004959,
-0.03680061176419258,
0.020609717816114426,
-0.018246086314320564,
-0.025731146335601807,
-0.004493247717618942,
0.019433369860053062,
-0.2548390328884125,
0.07098561525344849,
0.009674696251749992,
0.05840678885579109,
-0.041665274649858475,
-0.05303872376680374,
0.0722184106707573,
0.12944616377353668,
-0.04254712909460068,
-0.02344297058880329,
-0.0987035259604454,
-0.19198960065841675,
0.0025275268126279116,
0.05174063518643379,
0.04977063089609146,
-0.0024890073109418154,
0.043467238545417786,
0.003739504609256983,
0.005392079241573811,
0.009215191937983036,
-0.07193821668624878,
-0.12807370722293854,
-0.0902218446135521,
-0.02096022292971611,
0.10826225578784943,
0.051769353449344635,
-0.007984875701367855,
-0.09598713368177414,
-0.0349227599799633,
0.021500276401638985,
-0.12005175650119781,
-0.10327117145061493,
-0.07178609818220139,
0.02536403201520443,
0.06237737461924553,
-0.13257667422294617,
0.0887250304222107,
-0.031573645770549774,
0.05029879882931709,
-0.005107458680868149,
-0.20294420421123505,
0.09215071052312851,
-0.058479588478803635,
-0.03604452684521675,
-0.03238176926970482,
0.06421677768230438,
0.03368505463004112,
0.02353648468852043,
0.005883510690182447,
0.013134598731994629,
-0.017708439379930496,
-0.03714954853057861,
0.028254812583327293,
0.02556542493402958,
-0.009431560523808002,
0.03939638286828995,
-0.073165163397789,
-0.19195619225502014,
-0.11484818160533905,
0.015648430213332176,
0.19418945908546448,
0.10253544896841049,
-0.019757887348532677,
0.08912399411201477,
0.06394875049591064,
-0.028880853205919266,
-0.3559117913246155,
-0.06128743290901184,
0.026923593133687973,
0.024541962891817093,
-0.02849803864955902,
-0.14707355201244354,
0.12907466292381287,
-0.013792356476187706,
-0.006634030491113663,
0.1016387939453125,
-0.1913745105266571,
-0.11812569946050644,
0.23674122989177704,
0.05248124152421951,
0.14405915141105652,
-0.11085726320743561,
-0.07687516510486603,
-0.04025464877486229,
-0.0676259845495224,
0.13866941630840302,
-0.038326919078826904,
0.08761730045080185,
-0.013460981659591198,
0.11112793534994125,
-0.0018323648255318403,
0.02036602422595024,
0.0827181339263916,
0.012915792874991894,
0.0012682841625064611,
-0.08154097199440002,
-0.12964853644371033,
0.020360300317406654,
-0.010930254124104977,
0.007492148317396641,
-0.009151813574135303,
0.05000938102602959,
-0.07469254732131958,
-0.030235791578888893,
-0.0816749557852745,
0.1493128091096878,
0.058665357530117035,
-0.032910726964473724,
-0.08230599761009216,
-0.08094731718301773,
-0.025824010372161865,
-0.026050496846437454,
0.1470024585723877,
-0.02817387506365776,
0.03685770183801651,
0.01423178892582655,
0.09915963560342789,
-0.008054269477725029,
-0.037711817771196365,
-0.008005659095942974,
-0.07775330543518066,
0.11598582565784454,
-0.0071367486380040646,
0.06169988960027695,
0.12092296779155731,
0.007919415831565857,
0.02922879159450531,
0.10273882001638412,
0.004197985865175724,
0.05891794711351395,
0.11024050414562225,
-0.10912485420703888,
0.14277926087379456,
-0.028841901570558548,
0.008820104412734509,
0.0249722171574831,
0.10472013801336288,
0.11880357563495636,
0.019809706136584282,
-0.0786355659365654,
-0.019876355305314064,
0.011900100857019424,
-0.01892852783203125,
0.1252453774213791,
0.16245469450950623,
0.039898838847875595,
-0.0813579112291336,
0.05853094905614853,
0.005633876658976078,
-0.13884980976581573,
0.07150186598300934,
0.06050978973507881,
-0.15669557452201843,
-0.09392805397510529,
0.07936882972717285,
0.18754884600639343,
-0.19982680678367615,
-0.07472848892211914,
-0.07874955236911774,
-0.08970683813095093,
0.030369481071829796,
0.29442542791366577,
0.07772985845804214,
0.023472117260098457,
-0.024742845445871353,
-0.0001406118244631216,
-0.06793750077486038,
0.0866805762052536,
-0.01353191863745451,
0.08023660629987717,
-0.11576244980096817,
0.005909434519708157,
-0.0008304174989461899,
0.11764378845691681,
-0.10163839161396027,
-0.01932070590555668,
-0.09563274681568146,
0.006504903547465801,
-0.04879052937030792,
0.02835739776492119,
-0.0468028225004673,
-0.027226248756051064,
0.03502092882990837,
-0.04186980053782463,
-0.05252514407038689,
0.022087769582867622,
-0.05520934984087944,
0.021735861897468567,
-0.027120834216475487,
-0.003875789698213339,
-0.03521271422505379,
-0.06065649911761284,
0.0613633394241333,
-0.035296883434057236,
0.0347711518406868,
0.1273936927318573,
-0.06321108341217041,
0.08461654931306839,
-0.09613684564828873,
-0.10239905118942261,
0.11457307636737823,
0.0025582637172192335,
-0.026903575286269188,
-0.011270619928836823,
0.015220673754811287,
0.03697029873728752,
0.04425140470266342,
0.036602653563022614,
0.14568990468978882,
-0.03810172528028488,
0.02704627439379692,
-0.08779679238796234,
-0.09411585330963135,
-0.09165150672197342,
0.046167049556970596,
0.1264035999774933,
0.04336850717663765,
0.15673546493053436,
-0.1167430654168129,
0.003477341728284955,
0.0009702086681500077,
0.03386583551764488,
-0.05546801537275314,
-0.0642438605427742,
-0.08032475411891937,
-0.011917823925614357,
0.022269323468208313,
-0.04341460019350052,
0.13098543882369995,
-0.010575192980468273,
-0.09265746921300888,
0.04719505086541176,
0.06529103219509125,
0.04129455238580704,
0.07752469927072525,
0.34128516912460327,
0.0317898690700531,
-0.05632457137107849,
-0.015493214130401611,
0.06350588798522949,
0.05146421864628792,
0.02379259094595909,
0.06897915154695511,
0.10111894458532333,
-0.04176574572920799,
0.09265786409378052,
-0.018542706966400146,
0.059660933911800385,
-0.029450809583067894,
-0.07330621778964996,
-0.05332616716623306,
-0.003322102827951312,
0.07799737900495529,
0.236978679895401,
0.16348585486412048,
0.007412999868392944,
-0.007628603372722864,
-0.014165188185870647,
-0.10928473621606827,
-0.11865868419408798,
-0.16510634124279022,
-0.11415286362171173,
-0.11695446074008942,
0.032243840396404266,
-0.05836676433682442,
-0.07333912700414658,
0.11054746806621552,
0.01002421136945486,
-0.044823192059993744,
0.12622852623462677,
0.008993017487227917,
-0.09371966868638992,
0.05148769170045853,
-0.027578990906476974,
-0.09889544546604156,
0.05989549309015274,
-0.028445210307836533,
0.03749629855155945,
0.05135829374194145,
-0.009360690601170063,
-0.01319299079477787,
-0.07724156975746155,
0.017062705010175705,
-0.09945232421159744,
-0.09645021706819534,
0.02350512705743313,
0.008035746403038502,
0.023775536566972733,
0.0331549197435379,
0.05331198126077652,
-0.04615877568721771,
0.03236456960439682,
0.16000626981258392,
0.023237256333231926,
-0.1533171683549881,
-0.05694227293133736,
0.1527210772037506,
0.035329267382621765,
0.11194387823343277,
-0.02390584722161293,
-0.12647368013858795,
-0.05539972707629204,
0.1320749819278717,
0.1169673427939415,
-0.0014051623875275254,
0.028980081900954247,
-0.07191199064254761,
0.011902653612196445,
0.02661798521876335,
0.024274056777358055,
0.07602348178625107,
0.13482744991779327,
-0.062028564512729645,
-0.08276762813329697,
-0.07744809985160828,
0.026052886620163918,
-0.1212952733039856,
-0.01572086103260517,
0.016026180237531662,
-0.05045776441693306,
-0.09698770195245743,
0.05754704028367996,
-0.13116039335727692,
0.02129950560629368,
0.04466857388615608,
-0.08805804699659348,
-0.046736426651477814,
0.006061479449272156,
0.12517878413200378,
0.0030321697704494,
0.018288854509592056,
-0.02921302057802677,
0.0008912741905078292,
-0.09138815104961395,
0.039390791207551956,
-0.16063857078552246,
0.06760913133621216,
0.02321552112698555,
-0.11010093986988068,
0.028107237070798874,
-0.05986245349049568,
0.1956903636455536,
0.10671372711658478,
0.07254216820001602,
-0.00969199649989605,
0.10520388931035995,
-0.01431574858725071,
0.005295607261359692,
0.07605107128620148,
0.106767438352108,
0.03397500514984131,
-0.03384888172149658,
0.08136777579784393,
-0.076835498213768,
0.008405344560742378,
-0.1005650982260704,
-0.043199289590120316,
-0.07417479902505875,
0.00628212234005332,
-0.11065945774316788,
0.07585899531841278,
-0.05382814258337021,
-0.04287896305322647,
-0.02166409231722355,
-0.031880203634500504,
-0.04719125106930733,
0.044601328670978546,
-0.05210144445300102,
-0.027261963114142418,
-0.15013432502746582,
-0.05304086208343506,
-0.006943778600543737,
0.008045091293752193,
-0.2002602070569992,
0.007887788116931915,
-0.18358001112937927,
-0.004256500396877527,
-0.1000848263502121,
0.07383419573307037,
0.04333076626062393,
-0.010521325282752514,
-0.03694658726453781,
-0.0676766186952591,
0.015914496034383774,
0.0943983644247055,
-0.1847722977399826,
-0.0916072279214859
] |
null | null |
transformers
|
Language Model 2
For Language agnostic Dense Passage Retrieval
|
{}
|
feature-extraction
|
diarsabri/LaDPR-context-encoder
|
[
"transformers",
"pytorch",
"dpr",
"feature-extraction",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #dpr #feature-extraction #endpoints_compatible #region-us
|
Language Model 2
For Language agnostic Dense Passage Retrieval
|
[] |
[
"TAGS\n#transformers #pytorch #dpr #feature-extraction #endpoints_compatible #region-us \n"
] |
[
30
] |
[
"passage: TAGS\n#transformers #pytorch #dpr #feature-extraction #endpoints_compatible #region-us \n"
] |
[
-0.0771576464176178,
-0.011293538846075535,
-0.008893195539712906,
-0.0029936896171420813,
0.15081144869327545,
0.03713664412498474,
-0.009117815643548965,
0.08786246180534363,
0.00884011760354042,
-0.016918599605560303,
0.11535348743200302,
0.2754141092300415,
-0.029518838971853256,
0.04749128222465515,
-0.0667886734008789,
-0.2843414843082428,
0.10356485843658447,
0.09573152661323547,
-0.054970353841781616,
0.08667478710412979,
0.06971856951713562,
-0.0849144458770752,
0.05574442446231842,
-0.034097034484148026,
-0.16442269086837769,
0.0291686300188303,
0.0066535393707454205,
-0.08196933567523956,
0.1070394292473793,
0.00406640162691474,
0.1780167520046234,
0.0194949172437191,
-0.09128759056329727,
-0.17535772919654846,
0.021508917212486267,
0.00810398068279028,
-0.04078548401594162,
0.028031304478645325,
0.070148766040802,
-0.06318888068199158,
0.038901910185813904,
0.07854779064655304,
0.029407963156700134,
0.009312859736382961,
-0.14044368267059326,
-0.22536885738372803,
-0.06520035117864609,
0.055914465337991714,
0.009929531253874302,
0.11065451800823212,
0.03456724062561989,
0.13426202535629272,
-0.1601000428199768,
0.08349344879388809,
0.12370701134204865,
-0.23546431958675385,
0.03261985257267952,
0.09824281185865402,
0.08011092990636826,
-0.005897137802094221,
0.021381795406341553,
0.012515866197645664,
-0.02407817170023918,
0.03179292380809784,
-0.02172640524804592,
-0.07730061560869217,
-0.02404259331524372,
0.051261626183986664,
-0.09005977213382721,
-0.09812229126691818,
0.22244198620319366,
-0.022331899031996727,
0.06451907008886337,
0.0030178262386471033,
-0.10590830445289612,
-0.07290996611118317,
-0.03296685218811035,
-0.03098144568502903,
-0.012817480601370335,
0.06499318033456802,
0.028993545100092888,
0.007819942198693752,
-0.0991925522685051,
0.0037996734026819468,
-0.21662995219230652,
0.23994344472885132,
0.027378831058740616,
0.08603030443191528,
-0.21901105344295502,
0.07319362461566925,
-0.05809526890516281,
-0.0945400521159172,
0.004450338892638683,
-0.11313807219266891,
0.05781668424606323,
0.007172194309532642,
-0.08655758947134018,
0.00293210637755692,
0.07274742424488068,
0.12979653477668762,
-0.031103793531656265,
0.01575881615281105,
0.015605123713612556,
0.10138440132141113,
0.0338687002658844,
0.1240370124578476,
-0.016041675582528114,
0.006660472601652145,
0.029961666092276573,
-0.13705113530158997,
-0.017762040719389915,
-0.03147087246179581,
-0.09600777179002762,
-0.043064434081315994,
0.05176663398742676,
0.08674394339323044,
0.03872986510396004,
0.0369756855070591,
-0.07951563596725464,
-0.03221004083752632,
0.10411225259304047,
-0.08400167524814606,
-0.032884832471609116,
-0.009907873347401619,
0.018575269728899002,
0.20586106181144714,
-0.015155013650655746,
-0.028702588751912117,
-0.04571284353733063,
0.10365565121173859,
-0.06473726034164429,
0.01283703651279211,
-0.060718949884176254,
-0.06401096284389496,
0.03756453096866608,
-0.1464945524930954,
0.06524793803691864,
-0.16935338079929352,
-0.1325785368680954,
0.02533368021249771,
0.037524789571762085,
0.013445237651467323,
0.018039360642433167,
-0.0032444039825350046,
-0.021643640473484993,
-0.020006388425827026,
-0.07342390716075897,
-0.10293912887573242,
-0.05803784728050232,
0.10866802930831909,
0.006455678027123213,
0.04413048177957535,
-0.14772838354110718,
0.10207537561655045,
-0.10166215896606445,
0.02669479325413704,
-0.19395917654037476,
0.03278869017958641,
-0.03204043582081795,
0.17186911404132843,
-0.0011990026105195284,
-0.07787400484085083,
-0.12745113670825958,
0.03499797359108925,
-0.04436706006526947,
0.14732803404331207,
-0.07827179878950119,
-0.11213232576847076,
0.21086016297340393,
-0.13408932089805603,
-0.16676723957061768,
0.03212177753448486,
-0.004075905308127403,
-0.0180201418697834,
0.06149792671203613,
0.17824704945087433,
0.1367417424917221,
-0.05950278416275978,
0.07108042389154434,
0.1423618346452713,
-0.16533714532852173,
-0.14511768519878387,
0.013231072574853897,
-0.04738805070519447,
-0.03790054842829704,
0.05990088731050491,
-0.00013170260353945196,
0.10647415369749069,
-0.06054762005805969,
-0.03188963606953621,
-0.02680780366063118,
-0.015218580141663551,
0.06179942935705185,
0.06766457855701447,
0.0942283645272255,
-0.014076316729187965,
0.02696545608341694,
0.021010806784033775,
0.02131623588502407,
0.0007353039109148085,
0.03286464884877205,
-0.05946100503206253,
0.14868520200252533,
-0.07830331474542618,
0.022020231932401657,
-0.2623787522315979,
-0.0955447182059288,
-0.0033999949228018522,
0.04796777665615082,
-0.07262763381004333,
0.17288662493228912,
0.08602514117956161,
-0.08080308884382248,
0.033688608556985855,
-0.034288518130779266,
0.047081608325242996,
0.032703474164009094,
-0.005820977035909891,
-0.01510295644402504,
0.013858399353921413,
-0.08235502988100052,
-0.06219638139009476,
0.03491079807281494,
-0.011341252364218235,
0.07574722915887833,
0.10614197701215744,
0.010256623849272728,
0.03823024034500122,
-0.061550576239824295,
0.07145240902900696,
-0.019423669204115868,
0.0032544005662202835,
0.08965693414211273,
-0.021074337884783745,
-0.034531254321336746,
0.13329492509365082,
-0.09468328207731247,
0.32915472984313965,
0.1998574584722519,
-0.30229148268699646,
-0.0010246927849948406,
-0.026521479710936546,
-0.015395472757518291,
0.029904410243034363,
0.08503261208534241,
-0.024927476420998573,
0.09023560583591461,
0.03456827998161316,
0.13088765740394592,
-0.014477264136075974,
-0.018823698163032532,
0.0009506511269137263,
-0.01087108813226223,
-0.04965429753065109,
0.039593406021595,
0.07814808189868927,
-0.11861027032136917,
0.1790565848350525,
0.21732142567634583,
0.04586665332317352,
0.09192486107349396,
-0.09303727000951767,
-0.04207499697804451,
0.041489966213703156,
0.025277314707636833,
-0.059949833899736404,
0.03981201723217964,
-0.22375264763832092,
-0.03302648290991783,
0.08054045587778091,
0.04121636226773262,
0.09861346334218979,
-0.12431726604700089,
-0.044744525104761124,
0.03984569013118744,
-0.002545339986681938,
-0.10235144942998886,
0.07257664203643799,
0.068331778049469,
0.061762016266584396,
0.02251657471060753,
-0.067115917801857,
0.1252036839723587,
0.009761491790413857,
-0.032491110265254974,
0.16988682746887207,
-0.1360948085784912,
-0.2907842695713043,
-0.08841365575790405,
-0.09324696660041809,
0.0028301978018134832,
0.0291859433054924,
0.08205289393663406,
-0.08322185277938843,
-0.02952837571501732,
0.0852353572845459,
0.024966320022940636,
-0.1422043889760971,
0.0041829682886600494,
-0.06277982890605927,
0.04868394508957863,
-0.06482505798339844,
-0.10089325904846191,
-0.061034195125103,
-0.0669785588979721,
0.04088655486702919,
0.10600893199443817,
-0.08941914141178131,
0.09391910582780838,
0.1461140215396881,
0.03808148205280304,
0.0777505412697792,
0.007632375694811344,
0.14254970848560333,
-0.06825922429561615,
-0.10250192880630493,
0.2388693243265152,
-0.01791151612997055,
0.08385621011257172,
0.1125468835234642,
0.027100255712866783,
-0.07295090705156326,
-0.019307907670736313,
-0.08013535290956497,
-0.10020557790994644,
-0.16539764404296875,
-0.1356297731399536,
-0.14478740096092224,
0.020626552402973175,
0.0027235951274633408,
0.048579588532447815,
0.06545101851224899,
0.08081704378128052,
0.04853370785713196,
-0.034957513213157654,
-0.03293607011437416,
0.029078802093863487,
0.252279669046402,
-0.029261326417326927,
0.08921851962804794,
-0.0663212388753891,
-0.12329292297363281,
0.05829456448554993,
0.05001640319824219,
0.2581522762775421,
0.07627600431442261,
-0.010291879996657372,
0.04578480124473572,
0.16563363373279572,
0.12264170497655869,
0.16660550236701965,
0.028107773512601852,
-0.024365942925214767,
-0.0008751833229325712,
0.01126205176115036,
-0.09102796763181686,
0.011197999119758606,
0.15726660192012787,
-0.15419559180736542,
-0.08713242411613464,
-0.18948383629322052,
0.06528633832931519,
0.07812059670686722,
0.01919584535062313,
-0.21937312185764313,
0.004709960427135229,
0.03469812124967575,
0.01866038702428341,
-0.043996721506118774,
0.06826886534690857,
-0.0670149028301239,
-0.1183512806892395,
0.047355666756629944,
-0.05754063278436661,
0.09349112212657928,
-0.0645756646990776,
0.07083898782730103,
-0.011760256253182888,
-0.11689464002847672,
0.08463416993618011,
0.0764663890004158,
-0.24340397119522095,
0.24872039258480072,
-0.022256478667259216,
-0.04756268486380577,
-0.05736717954277992,
0.022660071030259132,
0.0034099917393177748,
0.14966729283332825,
0.1060599684715271,
0.016693023964762688,
-0.040531694889068604,
-0.14275166392326355,
0.02551034279167652,
0.03561164066195488,
0.13583005964756012,
-0.03059639222919941,
-0.033940453082323074,
-0.005631269421428442,
-0.0004697570693679154,
-0.022613689303398132,
0.02722436934709549,
0.06650249660015106,
-0.129581019282341,
0.06410782784223557,
-0.06444472074508667,
0.07818739861249924,
-0.0015784663846716285,
0.01048368401825428,
-0.07452952116727829,
0.15155048668384552,
-0.04620887339115143,
-0.07883980125188828,
-0.10227945446968079,
-0.12484326213598251,
0.09702008962631226,
-0.07681452482938766,
0.09714347124099731,
-0.0612587071955204,
-0.048495396971702576,
-0.045634448528289795,
-0.22512544691562653,
0.12113524973392487,
-0.10486915707588196,
0.05322318524122238,
-0.03403407707810402,
0.15408580005168915,
-0.0566464364528656,
-0.02536352537572384,
0.021352341398596764,
0.0010804099729284644,
-0.11338434368371964,
-0.09454425424337387,
-0.028241928666830063,
0.03797867149114609,
0.06931943446397781,
0.09351684898138046,
-0.04437108710408211,
0.059218987822532654,
0.010763518512248993,
0.06137523055076599,
0.23291432857513428,
0.12172061204910278,
-0.054572463035583496,
0.10782315582036972,
0.1097603291273117,
-0.002162437653169036,
-0.2970094680786133,
-0.07678167521953583,
-0.11589315533638,
-0.058791905641555786,
-0.02327010966837406,
-0.14232297241687775,
0.11557788401842117,
0.06967196613550186,
-0.012950683943927288,
0.1288210153579712,
-0.2600191533565521,
-0.0452185720205307,
0.18707002699375153,
0.01638496294617653,
0.4032147526741028,
-0.10144077241420746,
-0.07852423191070557,
-0.004975946620106697,
-0.2818116843700409,
0.11822978407144547,
0.025682790204882622,
0.05601249635219574,
-0.03451127931475639,
0.07345592230558395,
0.045787617564201355,
-0.07920095324516296,
0.11209888011217117,
0.05505137890577316,
0.054896675050258636,
-0.08072216808795929,
-0.05159546062350273,
0.03475319966673851,
-0.013634257018566132,
0.049173448234796524,
0.10643810033798218,
0.037725552916526794,
-0.14398570358753204,
-0.024383481591939926,
-0.12512867152690887,
0.054293543100357056,
0.05504593253135681,
-0.04876774922013283,
-0.01691318117082119,
-0.040568578988313675,
0.0030931225046515465,
0.021032381802797318,
0.24438899755477905,
-0.0007239680853672326,
0.09918426722288132,
0.010561276227235794,
0.03948104381561279,
-0.1806289553642273,
-0.17780561745166779,
-0.07077882438898087,
-0.03840376064181328,
0.10576653480529785,
-0.04933864250779152,
0.052111729979515076,
0.14613358676433563,
-0.020590890198946,
0.009519942104816437,
0.1356181651353836,
0.012213428504765034,
-0.023874331265687943,
0.11577675491571426,
-0.1853639930486679,
-0.03374156728386879,
-0.021911391988396645,
-0.10395221412181854,
0.09993544965982437,
0.10269584506750107,
0.0875738263130188,
0.041046567261219025,
-0.0013339685974642634,
-0.024068187922239304,
-0.02320464514195919,
-0.08606594800949097,
0.07273035496473312,
0.04470119625329971,
0.03290369734168053,
-0.1521778553724289,
0.07556816190481186,
-0.030587995424866676,
-0.2899177670478821,
-0.059222325682640076,
0.0417257696390152,
-0.1429247111082077,
-0.11031993478536606,
-0.10541680455207825,
0.062391553074121475,
-0.1629520058631897,
-0.04232624173164368,
-0.0421537384390831,
-0.13332974910736084,
0.0729721412062645,
0.12342998385429382,
0.09937863796949387,
0.14029204845428467,
-0.031109964475035667,
-0.010362326167523861,
-0.028799405321478844,
-0.07336407154798508,
-0.002587374532595277,
0.030408669263124466,
-0.12607474625110626,
-0.0019771764054894447,
-0.021572168916463852,
0.16592061519622803,
-0.07962234318256378,
-0.055081672966480255,
-0.12502525746822357,
0.07285080850124359,
-0.0719267800450325,
-0.10987258702516556,
-0.12971244752407074,
-0.04985805228352547,
0.03030523844063282,
-0.018094217404723167,
-0.006056319456547499,
-0.010556134395301342,
-0.13977400958538055,
0.03212582692503929,
0.004566697403788567,
-0.026997210457921028,
-0.05586300790309906,
-0.033186741173267365,
0.07953429967164993,
-0.05568533018231392,
0.09002365171909332,
0.2131437510251999,
-0.06744331121444702,
0.10348276793956757,
-0.08725912123918533,
-0.1646062582731247,
0.13500051200389862,
0.0337543785572052,
0.07426470518112183,
0.06613021343946457,
0.03512054681777954,
0.08105332404375076,
-0.03650834411382675,
0.035143036395311356,
-0.004462578799575567,
-0.1476754993200302,
-0.019473792985081673,
-0.013380797579884529,
-0.12439816445112228,
-0.05798431858420372,
-0.0884857028722763,
0.1479969471693039,
0.05932215228676796,
0.12610341608524323,
0.01924048364162445,
0.1092076376080513,
-0.029964160174131393,
-0.02214142307639122,
-0.0011265284847468138,
-0.165082186460495,
0.07994086295366287,
-0.04243525117635727,
0.01381773129105568,
-0.005377802532166243,
0.28020521998405457,
-0.02308158576488495,
0.03674847260117531,
0.017203141003847122,
0.03224368765950203,
0.054660871624946594,
0.018418367952108383,
0.2485504150390625,
0.12140955775976181,
-0.045491334050893784,
-0.08971124142408371,
0.1203898936510086,
0.005706361494958401,
-0.03194354474544525,
0.11862444877624512,
0.13286738097667694,
0.0016485553933307528,
0.08688269555568695,
0.04214884713292122,
0.02319319173693657,
-0.09937527775764465,
-0.23869918286800385,
0.0044044130481779575,
0.060131121426820755,
0.032101862132549286,
0.06318297982215881,
0.11096379905939102,
-0.015672191977500916,
0.09111893177032471,
-0.0063836341723799706,
-0.029071340337395668,
-0.11646446585655212,
-0.025223134085536003,
-0.04255421459674835,
-0.1599310040473938,
0.015905221924185753,
-0.06897944211959839,
-0.01467001810669899,
0.18589547276496887,
0.018507737666368484,
-0.04584968835115433,
0.13110458850860596,
0.03058410808444023,
-0.07448843866586685,
0.05627581104636192,
-0.011277900077402592,
-0.011352160014212132,
0.058755695819854736,
-0.021515129134058952,
-0.10789714753627777,
-0.07428885996341705,
-0.0222402922809124,
0.024521082639694214,
-0.1031985655426979,
0.01650399900972843,
-0.11171942949295044,
-0.09159082919359207,
-0.04442979022860527,
0.07792434096336365,
-0.08539313077926636,
0.156040221452713,
-0.009158983826637268,
-0.004093414172530174,
0.0173709187656641,
0.14773523807525635,
-0.06885794550180435,
-0.048014577478170395,
-0.02434067241847515,
0.2232249677181244,
0.1297008991241455,
0.1107347160577774,
0.01645386964082718,
-0.0070083835162222385,
-0.06972646713256836,
0.25435757637023926,
0.2595534324645996,
-0.020387442782521248,
0.04778193682432175,
0.06959507614374161,
0.03458646312355995,
0.06981329619884491,
0.11061305552721024,
0.13888351619243622,
0.3230479061603546,
-0.06457775831222534,
-0.040262725204229355,
-0.045022476464509964,
0.009440749883651733,
-0.11041921377182007,
0.017502648755908012,
0.06614047288894653,
-0.07889125496149063,
-0.05350761488080025,
0.11732836067676544,
-0.19201576709747314,
0.14887875318527222,
0.1097492128610611,
-0.17156335711479187,
-0.04053778201341629,
-0.07498209178447723,
0.12883324921131134,
-0.01604624278843403,
0.10572823137044907,
-0.04506263509392738,
-0.10860075056552887,
0.1091226190328598,
0.03648069500923157,
-0.27506494522094727,
-0.08638770878314972,
0.08410345762968063,
0.06054694578051567,
-0.03482033684849739,
-0.025489995256066322,
0.03414231538772583,
0.05500299111008644,
0.11694014072418213,
-0.012200409546494484,
0.07832269370555878,
0.02175845578312874,
-0.10501791536808014,
-0.05801232531666756,
-0.007649905513972044,
-0.0018009295454248786,
-0.10757627338171005,
0.03222259506583214,
-0.1845710575580597,
0.028056129813194275,
0.00592972943559289,
-0.02200135961174965,
-0.02127671055495739,
-0.02465643920004368,
-0.04999049752950668,
0.033981435000896454,
0.07177714258432388,
0.02560494653880596,
0.00029618380358442664,
-0.06664629280567169,
0.005004276055842638,
0.06591204553842545,
-0.08213956654071808,
-0.17344744503498077,
-0.04242769256234169,
-0.08809846639633179,
0.10480636358261108,
-0.05044872686266899,
-0.025413550436496735,
-0.03622115030884743,
-0.03745803236961365,
0.03806062787771225,
-0.08795610070228577,
0.027324259281158447,
-0.0028321470599621534,
0.04344157502055168,
0.0025351871736347675,
0.013853252865374088,
0.03083946369588375,
0.08085013926029205,
-0.1408834606409073,
-0.09294464439153671
] |
null | null |
transformers
|
Language Model 1
For Language agnostic Dense Passage Retrieval
|
{}
|
feature-extraction
|
diarsabri/LaDPR-query-encoder
|
[
"transformers",
"pytorch",
"dpr",
"feature-extraction",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #dpr #feature-extraction #endpoints_compatible #region-us
|
Language Model 1
For Language agnostic Dense Passage Retrieval
|
[] |
[
"TAGS\n#transformers #pytorch #dpr #feature-extraction #endpoints_compatible #region-us \n"
] |
[
30
] |
[
"passage: TAGS\n#transformers #pytorch #dpr #feature-extraction #endpoints_compatible #region-us \n"
] |
[
-0.0771576464176178,
-0.011293538846075535,
-0.008893195539712906,
-0.0029936896171420813,
0.15081144869327545,
0.03713664412498474,
-0.009117815643548965,
0.08786246180534363,
0.00884011760354042,
-0.016918599605560303,
0.11535348743200302,
0.2754141092300415,
-0.029518838971853256,
0.04749128222465515,
-0.0667886734008789,
-0.2843414843082428,
0.10356485843658447,
0.09573152661323547,
-0.054970353841781616,
0.08667478710412979,
0.06971856951713562,
-0.0849144458770752,
0.05574442446231842,
-0.034097034484148026,
-0.16442269086837769,
0.0291686300188303,
0.0066535393707454205,
-0.08196933567523956,
0.1070394292473793,
0.00406640162691474,
0.1780167520046234,
0.0194949172437191,
-0.09128759056329727,
-0.17535772919654846,
0.021508917212486267,
0.00810398068279028,
-0.04078548401594162,
0.028031304478645325,
0.070148766040802,
-0.06318888068199158,
0.038901910185813904,
0.07854779064655304,
0.029407963156700134,
0.009312859736382961,
-0.14044368267059326,
-0.22536885738372803,
-0.06520035117864609,
0.055914465337991714,
0.009929531253874302,
0.11065451800823212,
0.03456724062561989,
0.13426202535629272,
-0.1601000428199768,
0.08349344879388809,
0.12370701134204865,
-0.23546431958675385,
0.03261985257267952,
0.09824281185865402,
0.08011092990636826,
-0.005897137802094221,
0.021381795406341553,
0.012515866197645664,
-0.02407817170023918,
0.03179292380809784,
-0.02172640524804592,
-0.07730061560869217,
-0.02404259331524372,
0.051261626183986664,
-0.09005977213382721,
-0.09812229126691818,
0.22244198620319366,
-0.022331899031996727,
0.06451907008886337,
0.0030178262386471033,
-0.10590830445289612,
-0.07290996611118317,
-0.03296685218811035,
-0.03098144568502903,
-0.012817480601370335,
0.06499318033456802,
0.028993545100092888,
0.007819942198693752,
-0.0991925522685051,
0.0037996734026819468,
-0.21662995219230652,
0.23994344472885132,
0.027378831058740616,
0.08603030443191528,
-0.21901105344295502,
0.07319362461566925,
-0.05809526890516281,
-0.0945400521159172,
0.004450338892638683,
-0.11313807219266891,
0.05781668424606323,
0.007172194309532642,
-0.08655758947134018,
0.00293210637755692,
0.07274742424488068,
0.12979653477668762,
-0.031103793531656265,
0.01575881615281105,
0.015605123713612556,
0.10138440132141113,
0.0338687002658844,
0.1240370124578476,
-0.016041675582528114,
0.006660472601652145,
0.029961666092276573,
-0.13705113530158997,
-0.017762040719389915,
-0.03147087246179581,
-0.09600777179002762,
-0.043064434081315994,
0.05176663398742676,
0.08674394339323044,
0.03872986510396004,
0.0369756855070591,
-0.07951563596725464,
-0.03221004083752632,
0.10411225259304047,
-0.08400167524814606,
-0.032884832471609116,
-0.009907873347401619,
0.018575269728899002,
0.20586106181144714,
-0.015155013650655746,
-0.028702588751912117,
-0.04571284353733063,
0.10365565121173859,
-0.06473726034164429,
0.01283703651279211,
-0.060718949884176254,
-0.06401096284389496,
0.03756453096866608,
-0.1464945524930954,
0.06524793803691864,
-0.16935338079929352,
-0.1325785368680954,
0.02533368021249771,
0.037524789571762085,
0.013445237651467323,
0.018039360642433167,
-0.0032444039825350046,
-0.021643640473484993,
-0.020006388425827026,
-0.07342390716075897,
-0.10293912887573242,
-0.05803784728050232,
0.10866802930831909,
0.006455678027123213,
0.04413048177957535,
-0.14772838354110718,
0.10207537561655045,
-0.10166215896606445,
0.02669479325413704,
-0.19395917654037476,
0.03278869017958641,
-0.03204043582081795,
0.17186911404132843,
-0.0011990026105195284,
-0.07787400484085083,
-0.12745113670825958,
0.03499797359108925,
-0.04436706006526947,
0.14732803404331207,
-0.07827179878950119,
-0.11213232576847076,
0.21086016297340393,
-0.13408932089805603,
-0.16676723957061768,
0.03212177753448486,
-0.004075905308127403,
-0.0180201418697834,
0.06149792671203613,
0.17824704945087433,
0.1367417424917221,
-0.05950278416275978,
0.07108042389154434,
0.1423618346452713,
-0.16533714532852173,
-0.14511768519878387,
0.013231072574853897,
-0.04738805070519447,
-0.03790054842829704,
0.05990088731050491,
-0.00013170260353945196,
0.10647415369749069,
-0.06054762005805969,
-0.03188963606953621,
-0.02680780366063118,
-0.015218580141663551,
0.06179942935705185,
0.06766457855701447,
0.0942283645272255,
-0.014076316729187965,
0.02696545608341694,
0.021010806784033775,
0.02131623588502407,
0.0007353039109148085,
0.03286464884877205,
-0.05946100503206253,
0.14868520200252533,
-0.07830331474542618,
0.022020231932401657,
-0.2623787522315979,
-0.0955447182059288,
-0.0033999949228018522,
0.04796777665615082,
-0.07262763381004333,
0.17288662493228912,
0.08602514117956161,
-0.08080308884382248,
0.033688608556985855,
-0.034288518130779266,
0.047081608325242996,
0.032703474164009094,
-0.005820977035909891,
-0.01510295644402504,
0.013858399353921413,
-0.08235502988100052,
-0.06219638139009476,
0.03491079807281494,
-0.011341252364218235,
0.07574722915887833,
0.10614197701215744,
0.010256623849272728,
0.03823024034500122,
-0.061550576239824295,
0.07145240902900696,
-0.019423669204115868,
0.0032544005662202835,
0.08965693414211273,
-0.021074337884783745,
-0.034531254321336746,
0.13329492509365082,
-0.09468328207731247,
0.32915472984313965,
0.1998574584722519,
-0.30229148268699646,
-0.0010246927849948406,
-0.026521479710936546,
-0.015395472757518291,
0.029904410243034363,
0.08503261208534241,
-0.024927476420998573,
0.09023560583591461,
0.03456827998161316,
0.13088765740394592,
-0.014477264136075974,
-0.018823698163032532,
0.0009506511269137263,
-0.01087108813226223,
-0.04965429753065109,
0.039593406021595,
0.07814808189868927,
-0.11861027032136917,
0.1790565848350525,
0.21732142567634583,
0.04586665332317352,
0.09192486107349396,
-0.09303727000951767,
-0.04207499697804451,
0.041489966213703156,
0.025277314707636833,
-0.059949833899736404,
0.03981201723217964,
-0.22375264763832092,
-0.03302648290991783,
0.08054045587778091,
0.04121636226773262,
0.09861346334218979,
-0.12431726604700089,
-0.044744525104761124,
0.03984569013118744,
-0.002545339986681938,
-0.10235144942998886,
0.07257664203643799,
0.068331778049469,
0.061762016266584396,
0.02251657471060753,
-0.067115917801857,
0.1252036839723587,
0.009761491790413857,
-0.032491110265254974,
0.16988682746887207,
-0.1360948085784912,
-0.2907842695713043,
-0.08841365575790405,
-0.09324696660041809,
0.0028301978018134832,
0.0291859433054924,
0.08205289393663406,
-0.08322185277938843,
-0.02952837571501732,
0.0852353572845459,
0.024966320022940636,
-0.1422043889760971,
0.0041829682886600494,
-0.06277982890605927,
0.04868394508957863,
-0.06482505798339844,
-0.10089325904846191,
-0.061034195125103,
-0.0669785588979721,
0.04088655486702919,
0.10600893199443817,
-0.08941914141178131,
0.09391910582780838,
0.1461140215396881,
0.03808148205280304,
0.0777505412697792,
0.007632375694811344,
0.14254970848560333,
-0.06825922429561615,
-0.10250192880630493,
0.2388693243265152,
-0.01791151612997055,
0.08385621011257172,
0.1125468835234642,
0.027100255712866783,
-0.07295090705156326,
-0.019307907670736313,
-0.08013535290956497,
-0.10020557790994644,
-0.16539764404296875,
-0.1356297731399536,
-0.14478740096092224,
0.020626552402973175,
0.0027235951274633408,
0.048579588532447815,
0.06545101851224899,
0.08081704378128052,
0.04853370785713196,
-0.034957513213157654,
-0.03293607011437416,
0.029078802093863487,
0.252279669046402,
-0.029261326417326927,
0.08921851962804794,
-0.0663212388753891,
-0.12329292297363281,
0.05829456448554993,
0.05001640319824219,
0.2581522762775421,
0.07627600431442261,
-0.010291879996657372,
0.04578480124473572,
0.16563363373279572,
0.12264170497655869,
0.16660550236701965,
0.028107773512601852,
-0.024365942925214767,
-0.0008751833229325712,
0.01126205176115036,
-0.09102796763181686,
0.011197999119758606,
0.15726660192012787,
-0.15419559180736542,
-0.08713242411613464,
-0.18948383629322052,
0.06528633832931519,
0.07812059670686722,
0.01919584535062313,
-0.21937312185764313,
0.004709960427135229,
0.03469812124967575,
0.01866038702428341,
-0.043996721506118774,
0.06826886534690857,
-0.0670149028301239,
-0.1183512806892395,
0.047355666756629944,
-0.05754063278436661,
0.09349112212657928,
-0.0645756646990776,
0.07083898782730103,
-0.011760256253182888,
-0.11689464002847672,
0.08463416993618011,
0.0764663890004158,
-0.24340397119522095,
0.24872039258480072,
-0.022256478667259216,
-0.04756268486380577,
-0.05736717954277992,
0.022660071030259132,
0.0034099917393177748,
0.14966729283332825,
0.1060599684715271,
0.016693023964762688,
-0.040531694889068604,
-0.14275166392326355,
0.02551034279167652,
0.03561164066195488,
0.13583005964756012,
-0.03059639222919941,
-0.033940453082323074,
-0.005631269421428442,
-0.0004697570693679154,
-0.022613689303398132,
0.02722436934709549,
0.06650249660015106,
-0.129581019282341,
0.06410782784223557,
-0.06444472074508667,
0.07818739861249924,
-0.0015784663846716285,
0.01048368401825428,
-0.07452952116727829,
0.15155048668384552,
-0.04620887339115143,
-0.07883980125188828,
-0.10227945446968079,
-0.12484326213598251,
0.09702008962631226,
-0.07681452482938766,
0.09714347124099731,
-0.0612587071955204,
-0.048495396971702576,
-0.045634448528289795,
-0.22512544691562653,
0.12113524973392487,
-0.10486915707588196,
0.05322318524122238,
-0.03403407707810402,
0.15408580005168915,
-0.0566464364528656,
-0.02536352537572384,
0.021352341398596764,
0.0010804099729284644,
-0.11338434368371964,
-0.09454425424337387,
-0.028241928666830063,
0.03797867149114609,
0.06931943446397781,
0.09351684898138046,
-0.04437108710408211,
0.059218987822532654,
0.010763518512248993,
0.06137523055076599,
0.23291432857513428,
0.12172061204910278,
-0.054572463035583496,
0.10782315582036972,
0.1097603291273117,
-0.002162437653169036,
-0.2970094680786133,
-0.07678167521953583,
-0.11589315533638,
-0.058791905641555786,
-0.02327010966837406,
-0.14232297241687775,
0.11557788401842117,
0.06967196613550186,
-0.012950683943927288,
0.1288210153579712,
-0.2600191533565521,
-0.0452185720205307,
0.18707002699375153,
0.01638496294617653,
0.4032147526741028,
-0.10144077241420746,
-0.07852423191070557,
-0.004975946620106697,
-0.2818116843700409,
0.11822978407144547,
0.025682790204882622,
0.05601249635219574,
-0.03451127931475639,
0.07345592230558395,
0.045787617564201355,
-0.07920095324516296,
0.11209888011217117,
0.05505137890577316,
0.054896675050258636,
-0.08072216808795929,
-0.05159546062350273,
0.03475319966673851,
-0.013634257018566132,
0.049173448234796524,
0.10643810033798218,
0.037725552916526794,
-0.14398570358753204,
-0.024383481591939926,
-0.12512867152690887,
0.054293543100357056,
0.05504593253135681,
-0.04876774922013283,
-0.01691318117082119,
-0.040568578988313675,
0.0030931225046515465,
0.021032381802797318,
0.24438899755477905,
-0.0007239680853672326,
0.09918426722288132,
0.010561276227235794,
0.03948104381561279,
-0.1806289553642273,
-0.17780561745166779,
-0.07077882438898087,
-0.03840376064181328,
0.10576653480529785,
-0.04933864250779152,
0.052111729979515076,
0.14613358676433563,
-0.020590890198946,
0.009519942104816437,
0.1356181651353836,
0.012213428504765034,
-0.023874331265687943,
0.11577675491571426,
-0.1853639930486679,
-0.03374156728386879,
-0.021911391988396645,
-0.10395221412181854,
0.09993544965982437,
0.10269584506750107,
0.0875738263130188,
0.041046567261219025,
-0.0013339685974642634,
-0.024068187922239304,
-0.02320464514195919,
-0.08606594800949097,
0.07273035496473312,
0.04470119625329971,
0.03290369734168053,
-0.1521778553724289,
0.07556816190481186,
-0.030587995424866676,
-0.2899177670478821,
-0.059222325682640076,
0.0417257696390152,
-0.1429247111082077,
-0.11031993478536606,
-0.10541680455207825,
0.062391553074121475,
-0.1629520058631897,
-0.04232624173164368,
-0.0421537384390831,
-0.13332974910736084,
0.0729721412062645,
0.12342998385429382,
0.09937863796949387,
0.14029204845428467,
-0.031109964475035667,
-0.010362326167523861,
-0.028799405321478844,
-0.07336407154798508,
-0.002587374532595277,
0.030408669263124466,
-0.12607474625110626,
-0.0019771764054894447,
-0.021572168916463852,
0.16592061519622803,
-0.07962234318256378,
-0.055081672966480255,
-0.12502525746822357,
0.07285080850124359,
-0.0719267800450325,
-0.10987258702516556,
-0.12971244752407074,
-0.04985805228352547,
0.03030523844063282,
-0.018094217404723167,
-0.006056319456547499,
-0.010556134395301342,
-0.13977400958538055,
0.03212582692503929,
0.004566697403788567,
-0.026997210457921028,
-0.05586300790309906,
-0.033186741173267365,
0.07953429967164993,
-0.05568533018231392,
0.09002365171909332,
0.2131437510251999,
-0.06744331121444702,
0.10348276793956757,
-0.08725912123918533,
-0.1646062582731247,
0.13500051200389862,
0.0337543785572052,
0.07426470518112183,
0.06613021343946457,
0.03512054681777954,
0.08105332404375076,
-0.03650834411382675,
0.035143036395311356,
-0.004462578799575567,
-0.1476754993200302,
-0.019473792985081673,
-0.013380797579884529,
-0.12439816445112228,
-0.05798431858420372,
-0.0884857028722763,
0.1479969471693039,
0.05932215228676796,
0.12610341608524323,
0.01924048364162445,
0.1092076376080513,
-0.029964160174131393,
-0.02214142307639122,
-0.0011265284847468138,
-0.165082186460495,
0.07994086295366287,
-0.04243525117635727,
0.01381773129105568,
-0.005377802532166243,
0.28020521998405457,
-0.02308158576488495,
0.03674847260117531,
0.017203141003847122,
0.03224368765950203,
0.054660871624946594,
0.018418367952108383,
0.2485504150390625,
0.12140955775976181,
-0.045491334050893784,
-0.08971124142408371,
0.1203898936510086,
0.005706361494958401,
-0.03194354474544525,
0.11862444877624512,
0.13286738097667694,
0.0016485553933307528,
0.08688269555568695,
0.04214884713292122,
0.02319319173693657,
-0.09937527775764465,
-0.23869918286800385,
0.0044044130481779575,
0.060131121426820755,
0.032101862132549286,
0.06318297982215881,
0.11096379905939102,
-0.015672191977500916,
0.09111893177032471,
-0.0063836341723799706,
-0.029071340337395668,
-0.11646446585655212,
-0.025223134085536003,
-0.04255421459674835,
-0.1599310040473938,
0.015905221924185753,
-0.06897944211959839,
-0.01467001810669899,
0.18589547276496887,
0.018507737666368484,
-0.04584968835115433,
0.13110458850860596,
0.03058410808444023,
-0.07448843866586685,
0.05627581104636192,
-0.011277900077402592,
-0.011352160014212132,
0.058755695819854736,
-0.021515129134058952,
-0.10789714753627777,
-0.07428885996341705,
-0.0222402922809124,
0.024521082639694214,
-0.1031985655426979,
0.01650399900972843,
-0.11171942949295044,
-0.09159082919359207,
-0.04442979022860527,
0.07792434096336365,
-0.08539313077926636,
0.156040221452713,
-0.009158983826637268,
-0.004093414172530174,
0.0173709187656641,
0.14773523807525635,
-0.06885794550180435,
-0.048014577478170395,
-0.02434067241847515,
0.2232249677181244,
0.1297008991241455,
0.1107347160577774,
0.01645386964082718,
-0.0070083835162222385,
-0.06972646713256836,
0.25435757637023926,
0.2595534324645996,
-0.020387442782521248,
0.04778193682432175,
0.06959507614374161,
0.03458646312355995,
0.06981329619884491,
0.11061305552721024,
0.13888351619243622,
0.3230479061603546,
-0.06457775831222534,
-0.040262725204229355,
-0.045022476464509964,
0.009440749883651733,
-0.11041921377182007,
0.017502648755908012,
0.06614047288894653,
-0.07889125496149063,
-0.05350761488080025,
0.11732836067676544,
-0.19201576709747314,
0.14887875318527222,
0.1097492128610611,
-0.17156335711479187,
-0.04053778201341629,
-0.07498209178447723,
0.12883324921131134,
-0.01604624278843403,
0.10572823137044907,
-0.04506263509392738,
-0.10860075056552887,
0.1091226190328598,
0.03648069500923157,
-0.27506494522094727,
-0.08638770878314972,
0.08410345762968063,
0.06054694578051567,
-0.03482033684849739,
-0.025489995256066322,
0.03414231538772583,
0.05500299111008644,
0.11694014072418213,
-0.012200409546494484,
0.07832269370555878,
0.02175845578312874,
-0.10501791536808014,
-0.05801232531666756,
-0.007649905513972044,
-0.0018009295454248786,
-0.10757627338171005,
0.03222259506583214,
-0.1845710575580597,
0.028056129813194275,
0.00592972943559289,
-0.02200135961174965,
-0.02127671055495739,
-0.02465643920004368,
-0.04999049752950668,
0.033981435000896454,
0.07177714258432388,
0.02560494653880596,
0.00029618380358442664,
-0.06664629280567169,
0.005004276055842638,
0.06591204553842545,
-0.08213956654071808,
-0.17344744503498077,
-0.04242769256234169,
-0.08809846639633179,
0.10480636358261108,
-0.05044872686266899,
-0.025413550436496735,
-0.03622115030884743,
-0.03745803236961365,
0.03806062787771225,
-0.08795610070228577,
0.027324259281158447,
-0.0028321470599621534,
0.04344157502055168,
0.0025351871736347675,
0.013853252865374088,
0.03083946369588375,
0.08085013926029205,
-0.1408834606409073,
-0.09294464439153671
] |
null | null |
transformers
|
# Wav2Vec2-Large-XLSR-53
---
language: gl
datasets:
- OpenSLR 77
metrics:
- wer
tags:
- audio
- automatic-speech-recognition
- speech
- xlsr-fine-tuning-week
license: apache-2.0
model-index:
- name: Galician Wav2Vec2-Large-XLSR-53
results:
- task:
name: Speech Recognition
type: automatic-speech-recognition
dataset:
name: OpenSLR
type: openslr
args: gl
metrics:
- name: Test WER
type: wer
value: 16.79
---
Wav2Vec2-Large-XLSR-53-galician
Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on galician using the [OpenSLR](https://huggingface.co/datasets/common_voice) dataset
When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
```python
import torch
import torchaudio
from datasets import load_dataset
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
test_dataset = load_dataset("common_voice", "gl", split="test[:2%]") # This is not available yet, load OpenSLR or your dataset instead
processor = Wav2Vec2Processor.from_pretrained("diego-fustes/wav2vec2-large-xlsr-gl")
model = Wav2Vec2ForCTC.from_pretrained("diego-fustes/wav2vec2-large-xlsr-gl")
resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = resampler(speech_array).squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits
predicted_ids = torch.argmax(logits, dim=-1)
print("Prediction:", processor.batch_decode(predicted_ids))
print("Reference:", test_dataset["sentence"][:2])
```
## Evaluation
The model can be evaluated as follows on the Galician test data of Common Voice (when it is released).
```python
import torch
import torchaudio
from datasets import load_dataset, load_metric
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
import re
test_dataset = load_dataset("common_voice", "gl", split="test") # This is not available yet, load OpenSLR or your dataset instead
wer = load_metric("wer")
processor = Wav2Vec2Processor.from_pretrained("diego-fustes/wav2vec2-large-xlsr-gl")
model = Wav2Vec2ForCTC.from_pretrained("diego-fustes/wav2vec2-large-xlsr-gl")
model.to("cuda")
chars_to_ignore_regex = '[^a-záéíóúñ ]'
resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower()
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = resampler(speech_array).squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def evaluate(batch):
inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits
pred_ids = torch.argmax(logits, dim=-1)
batch["pred_strings"] = processor.batch_decode(pred_ids)
return batch
result = test_dataset.map(evaluate, batched=True, batch_size=8)
print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"])))
```
**Test Result**: 16.79 % on OpenSLR split
## Training
The OpenSLR [SLR77](https://openslr.org/77/) dataset was used for training and validation. The dataset was split as 70% for training, 15% for validation and 15% for testing
The script used for training can be found [here](https://github.com/diego-fustes/xlsr-fine-tuning-gl)
|
{}
|
automatic-speech-recognition
|
diego-fustes/wav2vec2-large-xlsr-gl
|
[
"transformers",
"pytorch",
"jax",
"safetensors",
"wav2vec2",
"automatic-speech-recognition",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #safetensors #wav2vec2 #automatic-speech-recognition #endpoints_compatible #region-us
|
# Wav2Vec2-Large-XLSR-53
---
language: gl
datasets:
- OpenSLR 77
metrics:
- wer
tags:
- audio
- automatic-speech-recognition
- speech
- xlsr-fine-tuning-week
license: apache-2.0
model-index:
- name: Galician Wav2Vec2-Large-XLSR-53
results:
- task:
name: Speech Recognition
type: automatic-speech-recognition
dataset:
name: OpenSLR
type: openslr
args: gl
metrics:
- name: Test WER
type: wer
value: 16.79
---
Wav2Vec2-Large-XLSR-53-galician
Fine-tuned facebook/wav2vec2-large-xlsr-53 on galician using the OpenSLR dataset
When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
## Evaluation
The model can be evaluated as follows on the Galician test data of Common Voice (when it is released).
Test Result: 16.79 % on OpenSLR split
## Training
The OpenSLR SLR77 dataset was used for training and validation. The dataset was split as 70% for training, 15% for validation and 15% for testing
The script used for training can be found here
|
[
"# Wav2Vec2-Large-XLSR-53\n\n---\nlanguage: gl\ndatasets:\n- OpenSLR 77\nmetrics:\n- wer\ntags:\n- audio\n- automatic-speech-recognition\n- speech\n- xlsr-fine-tuning-week\nlicense: apache-2.0\nmodel-index:\n- name: Galician Wav2Vec2-Large-XLSR-53\n results:\n - task: \n name: Speech Recognition\n type: automatic-speech-recognition\n dataset:\n name: OpenSLR\n type: openslr\n args: gl\n metrics:\n - name: Test WER\n type: wer\n value: 16.79\n---\n\nWav2Vec2-Large-XLSR-53-galician\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on galician using the OpenSLR dataset\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Galician test data of Common Voice (when it is released).\n\n\n\nTest Result: 16.79 % on OpenSLR split",
"## Training\n\nThe OpenSLR SLR77 dataset was used for training and validation. The dataset was split as 70% for training, 15% for validation and 15% for testing \n\nThe script used for training can be found here"
] |
[
"TAGS\n#transformers #pytorch #jax #safetensors #wav2vec2 #automatic-speech-recognition #endpoints_compatible #region-us \n",
"# Wav2Vec2-Large-XLSR-53\n\n---\nlanguage: gl\ndatasets:\n- OpenSLR 77\nmetrics:\n- wer\ntags:\n- audio\n- automatic-speech-recognition\n- speech\n- xlsr-fine-tuning-week\nlicense: apache-2.0\nmodel-index:\n- name: Galician Wav2Vec2-Large-XLSR-53\n results:\n - task: \n name: Speech Recognition\n type: automatic-speech-recognition\n dataset:\n name: OpenSLR\n type: openslr\n args: gl\n metrics:\n - name: Test WER\n type: wer\n value: 16.79\n---\n\nWav2Vec2-Large-XLSR-53-galician\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on galician using the OpenSLR dataset\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Galician test data of Common Voice (when it is released).\n\n\n\nTest Result: 16.79 % on OpenSLR split",
"## Training\n\nThe OpenSLR SLR77 dataset was used for training and validation. The dataset was split as 70% for training, 15% for validation and 15% for testing \n\nThe script used for training can be found here"
] |
[
45,
205,
20,
39,
46
] |
[
"passage: TAGS\n#transformers #pytorch #jax #safetensors #wav2vec2 #automatic-speech-recognition #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53\n\n---\nlanguage: gl\ndatasets:\n- OpenSLR 77\nmetrics:\n- wer\ntags:\n- audio\n- automatic-speech-recognition\n- speech\n- xlsr-fine-tuning-week\nlicense: apache-2.0\nmodel-index:\n- name: Galician Wav2Vec2-Large-XLSR-53\n results:\n - task: \n name: Speech Recognition\n type: automatic-speech-recognition\n dataset:\n name: OpenSLR\n type: openslr\n args: gl\n metrics:\n - name: Test WER\n type: wer\n value: 16.79\n---\n\nWav2Vec2-Large-XLSR-53-galician\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on galician using the OpenSLR dataset\n\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:## Evaluation\n\nThe model can be evaluated as follows on the Galician test data of Common Voice (when it is released).\n\n\n\nTest Result: 16.79 % on OpenSLR split## Training\n\nThe OpenSLR SLR77 dataset was used for training and validation. The dataset was split as 70% for training, 15% for validation and 15% for testing \n\nThe script used for training can be found here"
] |
[
-0.16852854192256927,
0.07457727193832397,
-0.005109880119562149,
0.01657489314675331,
0.04355485364794731,
0.003311007982119918,
0.09592228382825851,
0.1208687350153923,
0.06647731363773346,
0.08396410197019577,
0.029308250173926353,
0.05018753185868263,
0.042383648455142975,
0.00701461685821414,
-0.04264957085251808,
-0.12848207354545593,
0.0612918995320797,
-0.057185545563697815,
0.030630124732851982,
0.06406662613153458,
0.1058449000120163,
-0.044170595705509186,
0.003666651900857687,
0.06639081984758377,
0.009701271541416645,
0.013800876215100288,
0.029100392013788223,
-0.13304144144058228,
0.1162639707326889,
0.0915442630648613,
0.03718985617160797,
0.030210182070732117,
0.016617320477962494,
-0.1893390566110611,
0.0018412121571600437,
0.011328569613397121,
0.029507599771022797,
0.002741952193900943,
0.1361527293920517,
-0.016463078558444977,
0.09198836982250214,
-0.05503552407026291,
0.03706352412700653,
0.062371473759412766,
-0.05983829125761986,
-0.2984055280685425,
-0.07779236137866974,
-0.08592849224805832,
0.09149841219186783,
0.0783030167222023,
-0.03807307407259941,
0.10887214541435242,
-0.0936804935336113,
0.056664399802684784,
0.06639397144317627,
-0.17969107627868652,
-0.021786410361528397,
0.015571903437376022,
0.011425289325416088,
0.06669891625642776,
-0.06932123750448227,
0.029325537383556366,
0.05769186466932297,
-0.007561453152447939,
-0.07485300302505493,
-0.02710137888789177,
-0.11407113820314407,
-0.015440349467098713,
-0.13218006491661072,
-0.05476847290992737,
0.2292538583278656,
-0.035506848245859146,
-0.080482117831707,
-0.20188406109809875,
0.006574929691851139,
0.13301719725131989,
0.02295033074915409,
-0.013899090699851513,
-0.013801543973386288,
-0.020237307995557785,
0.0659325122833252,
-0.060768939554691315,
-0.10144083201885223,
-0.07694143801927567,
-0.03509599342942238,
0.18543760478496552,
0.043924685567617416,
0.02240290492773056,
-0.005148295313119888,
0.09465250372886658,
-0.1283990889787674,
-0.07463139295578003,
-0.11667884886264801,
0.02069367654621601,
-0.1176278367638588,
-0.017546718940138817,
-0.04023124650120735,
-0.15699808299541473,
0.12806186079978943,
0.106521375477314,
-0.07218155264854431,
0.029266593977808952,
-0.017512207850813866,
0.010765003971755505,
0.06506512314081192,
0.20969796180725098,
-0.054140474647283554,
-0.018162282183766365,
-0.005796935874968767,
-0.04953436180949211,
0.028328541666269302,
-0.013289759866893291,
-0.03252819925546646,
-0.11039414256811142,
0.050414685159921646,
0.10862426459789276,
0.05670563131570816,
-0.004782807547599077,
-0.07334698736667633,
-0.027668442577123642,
0.09464970231056213,
-0.13140779733657837,
0.03869081288576126,
0.07745880633592606,
-0.009952250868082047,
0.07153061777353287,
0.030536307021975517,
0.0094611132517457,
-0.06849554181098938,
0.03715382516384125,
0.029352784156799316,
-0.0011964230798184872,
-0.028377728536725044,
-0.06919073313474655,
0.05541461706161499,
0.030380787327885628,
-0.02440391480922699,
-0.05068611353635788,
-0.1301611065864563,
-0.09150011092424393,
0.00021901249419897795,
-0.06143954023718834,
0.023503214120864868,
-0.07091835886240005,
-0.00019972350855823606,
0.025358345359563828,
-0.04403678700327873,
-0.019800256937742233,
-0.057089705020189285,
0.06270531564950943,
0.044248610734939575,
-0.00010340318840462714,
-0.017224818468093872,
0.030454128980636597,
-0.07363036274909973,
-0.019153663888573647,
-0.08216853439807892,
0.16939756274223328,
-0.13754241168498993,
-0.019944362342357635,
-0.09165075421333313,
-0.047641389071941376,
0.004177856724709272,
0.02724761888384819,
0.050143763422966,
0.11297783255577087,
-0.18489757180213928,
-0.03802553564310074,
0.23964644968509674,
-0.14944986999034882,
-0.030383801087737083,
0.19666141271591187,
-0.0006424442399293184,
0.012120860628783703,
0.08014890551567078,
0.18498820066452026,
0.22161681950092316,
-0.21710027754306793,
0.03610694408416748,
-0.002316785044968128,
-0.006142770871520042,
0.04659109562635422,
0.13609620928764343,
-0.16393299400806427,
0.033815063536167145,
0.047057777643203735,
-0.17083899676799774,
0.005494894925504923,
-0.009399399161338806,
-0.04865390434861183,
-0.028934504836797714,
-0.017559228464961052,
0.060015104711055756,
0.0007005900260992348,
-0.046803854405879974,
-0.09131523221731186,
-0.05597925931215286,
0.04074835032224655,
0.11820327490568161,
-0.07467565685510635,
0.030823538079857826,
-0.129349023103714,
0.163286954164505,
-0.11138372123241425,
0.0024580699391663074,
-0.10720331966876984,
-0.03624948486685753,
0.024042002856731415,
-0.029522748664021492,
0.021050648763775826,
0.0719822570681572,
0.01497964933514595,
-0.012597696855664253,
-0.015150166116654873,
-0.0030929779168218374,
0.03965350612998009,
-0.044646456837654114,
-0.011796616949141026,
-0.1157013401389122,
0.004058615304529667,
-0.047593217343091965,
0.09225574880838394,
-0.20264174044132233,
-0.02445979230105877,
0.1223418191075325,
-0.012585891410708427,
-0.005900338292121887,
-0.027067258954048157,
0.04200805723667145,
0.05674104019999504,
0.005768763367086649,
-0.031964804977178574,
0.03400236740708351,
0.022298112511634827,
-0.05621281638741493,
0.12973591685295105,
-0.09937471151351929,
-0.03043866530060768,
0.04504687339067459,
0.03565001115202904,
-0.05567752942442894,
0.009826088324189186,
-0.045868437737226486,
0.008096625097095966,
-0.13557952642440796,
-0.061858076602220535,
0.24985221028327942,
0.004652994219213724,
0.0883745551109314,
-0.12041676044464111,
-0.05517929047346115,
-0.004787676967680454,
-0.08266795426607132,
-0.02804054319858551,
0.15400473773479462,
-0.09029728919267654,
0.021800408139824867,
0.016384299844503403,
-0.06759174913167953,
-0.06618616729974747,
0.24601978063583374,
-0.060990337282419205,
-0.0976678729057312,
0.001354239764623344,
0.09716926515102386,
0.031105155125260353,
0.10171964019536972,
-0.16752290725708008,
-0.00006290899909799919,
0.03426744043827057,
0.05884206295013428,
0.054170381277799606,
-0.14606642723083496,
0.03959762677550316,
0.03531140089035034,
-0.1257704496383667,
-0.05548947677016258,
0.09859497100114822,
0.005226416513323784,
0.07226242870092392,
-0.060174550861120224,
0.03689977899193764,
-0.015385660342872143,
-0.06905319541692734,
-0.1612703800201416,
0.16058070957660675,
-0.09911870211362839,
-0.08583515882492065,
-0.18509480357170105,
0.05419984459877014,
-0.05914635583758354,
0.001436514430679381,
0.07678963243961334,
-0.07898035645484924,
-0.054093699902296066,
-0.04293564707040787,
0.05436088889837265,
0.019190602004528046,
-0.041948117315769196,
-0.008814632892608643,
0.0017515479121357203,
0.0479518286883831,
-0.14880767464637756,
0.015995493158698082,
-0.021060185506939888,
-0.08866430073976517,
0.025003844872117043,
-0.0017305335495620966,
0.09096415340900421,
0.11551415175199509,
-0.006761204916983843,
0.001822984078899026,
-0.026701761409640312,
0.1484318971633911,
-0.11582700908184052,
-0.03692159801721573,
0.1385166049003601,
-0.006272351369261742,
-0.014433149248361588,
0.051941096782684326,
0.010718622244894505,
-0.08463452756404877,
-0.04281540587544441,
-0.0017907421570271254,
-0.04853944480419159,
-0.3204115033149719,
-0.14142128825187683,
-0.06829199194908142,
0.013526776805520058,
-0.005240481812506914,
0.031409457325935364,
-0.02187914401292801,
0.022533271461725235,
-0.03239067271351814,
-0.11839709430932999,
0.053419049829244614,
0.028555989265441895,
0.23744653165340424,
0.022231195122003555,
0.04422213137149811,
-0.0704573392868042,
0.020419936627149582,
0.10054797679185867,
0.04397144541144371,
0.1234736293554306,
-0.006861346773803234,
0.12894770503044128,
0.09179334342479706,
0.10213033854961395,
-0.012173631228506565,
0.030550846830010414,
-0.04843125864863396,
0.020397743210196495,
0.04641750082373619,
-0.089043527841568,
-0.01967538148164749,
0.025000182911753654,
0.06963969022035599,
-0.06303714960813522,
-0.044835902750492096,
-0.07872558385133743,
0.07198397815227509,
0.13979539275169373,
0.054792266339063644,
-0.11450521647930145,
-0.09629063308238983,
-0.0003810166090261191,
-0.04341604560613632,
-0.05044914409518242,
0.017076676711440086,
0.1223158985376358,
-0.12253079563379288,
0.09376021474599838,
-0.004741220735013485,
0.060400284826755524,
-0.06065312772989273,
-0.03663179650902748,
-0.024268457666039467,
0.07114455848932266,
-0.014759838581085205,
0.11333269625902176,
-0.20370261371135712,
0.17166662216186523,
0.005432323087006807,
0.09696425497531891,
-0.05447714775800705,
0.02342773601412773,
-0.005211034789681435,
0.006105876062065363,
0.15877355635166168,
0.01140857394784689,
-0.03508797660470009,
-0.06807022541761398,
-0.08717739582061768,
0.061357155442237854,
0.01288739126175642,
0.011145041324198246,
0.0849672183394432,
-0.015607306733727455,
0.023516323417425156,
-0.01865386962890625,
-0.01874285563826561,
-0.0948638990521431,
-0.1284465789794922,
0.04256388545036316,
0.12197738140821457,
0.0059652868658304214,
-0.038930680602788925,
-0.07800372689962387,
-0.19868600368499756,
0.10210277140140533,
-0.193124458193779,
-0.11567211896181107,
-0.12779173254966736,
-0.051824238151311874,
0.16137008368968964,
-0.0585886649787426,
-0.0034205198753625154,
0.01598528027534485,
0.11670532822608948,
-0.014496778137981892,
0.0009467093041166663,
0.0521957129240036,
-0.11279310286045074,
-0.1695447713136673,
-0.01099771074950695,
0.16268014907836914,
0.04716261103749275,
0.04342953488230705,
0.06470546126365662,
0.027231736108660698,
0.03364330157637596,
-0.11857032030820847,
-0.0011188766220584512,
0.11711568385362625,
-0.11807881295681,
0.046728719025850296,
-0.00026412002625875175,
-0.20803773403167725,
-0.0865236222743988,
-0.11407119035720825,
0.12637676298618317,
0.12622632086277008,
-0.060055557638406754,
0.14585529267787933,
0.10702750831842422,
-0.1299772560596466,
-0.20143763720989227,
0.009141293354332447,
0.10841862112283707,
0.0776011198759079,
-0.012602836824953556,
-0.15286673605442047,
0.02665615640580654,
0.06076008826494217,
-0.03670527786016464,
0.0289626382291317,
-0.27971869707107544,
-0.16376839578151703,
0.10680584609508514,
0.012534325942397118,
0.11337187886238098,
-0.0790228471159935,
-0.08766303211450577,
0.007267228327691555,
-0.011065135709941387,
-0.011539077386260033,
-0.02215164713561535,
0.06426873803138733,
0.02061355486512184,
0.10817010700702667,
0.07823126018047333,
-0.018613647669553757,
0.17326940596103668,
0.020959697663784027,
-0.012653736397624016,
-0.01371010672301054,
0.1239035576581955,
-0.008772602304816246,
-0.00901529286056757,
0.1552055925130844,
0.012164550833404064,
0.030674265697598457,
-0.158665731549263,
-0.06785713881254196,
-0.05970223993062973,
0.08593779802322388,
-0.015676012262701988,
-0.0018315572524443269,
0.010880779474973679,
-0.002706526080146432,
0.04065936803817749,
-0.01472962461411953,
-0.013111680746078491,
-0.11091026663780212,
0.00494695408269763,
0.052589017897844315,
0.19025155901908875,
0.027565130963921547,
-0.020088953897356987,
0.014736728742718697,
0.006498374976217747,
0.06528867781162262,
-0.04106957092881203,
0.052890341728925705,
0.0721467062830925,
0.06188591197133064,
0.057007014751434326,
-0.025925815105438232,
-0.12104049324989319,
0.06705929338932037,
0.041471708565950394,
-0.053963467478752136,
-0.12620821595191956,
0.020896416157484055,
0.022867826744914055,
-0.03514062985777855,
0.0407402329146862,
0.11851458251476288,
-0.004078276921063662,
-0.013874183408915997,
-0.03228161111474037,
-0.006641640793532133,
-0.02880512736737728,
0.21424338221549988,
-0.01060839556157589,
0.059968188405036926,
-0.08070768415927887,
0.04044326767325401,
0.028320787474513054,
-0.08018167316913605,
0.05688747763633728,
-0.018365835770964622,
-0.04641834646463394,
-0.036295875906944275,
-0.06194247305393219,
0.08230604231357574,
-0.07557740062475204,
-0.16132040321826935,
-0.02918652817606926,
-0.10120493173599243,
-0.011696373112499714,
0.15007492899894714,
0.05085931345820427,
0.04875204712152481,
0.0073616295121610165,
-0.04343663156032562,
-0.07377706468105316,
0.09659209847450256,
0.048305533826351166,
-0.01974857971072197,
-0.06945081055164337,
0.03177005425095558,
-0.008205997757613659,
-0.02199963852763176,
-0.025018922984600067,
-0.042898643761873245,
-0.057485539466142654,
0.0573287233710289,
-0.10839259624481201,
0.05384667590260506,
-0.012440468184649944,
0.029043955728411674,
0.009032283909618855,
-0.06674055010080338,
0.010994677431881428,
0.0688013881444931,
-0.05591554194688797,
0.030625920742750168,
-0.020454898476600647,
0.10930384695529938,
-0.10776069760322571,
0.006570328492671251,
0.03400982916355133,
-0.05988188460469246,
0.09999580681324005,
0.058013252913951874,
-0.0684932991862297,
0.05015524476766586,
-0.2730239927768707,
0.016173267737030983,
0.0034355721436440945,
0.048980385065078735,
-0.0015524940099567175,
-0.1564698964357376,
0.01905526965856552,
0.09470365196466446,
0.026953037828207016,
0.037086453288793564,
0.07884728163480759,
-0.07733004540205002,
0.06797818094491959,
-0.009112386964261532,
-0.09247393161058426,
0.009425083175301552,
0.061707753688097,
0.031608108431100845,
0.04359656944870949,
0.12639591097831726,
-0.1489650458097458,
0.057328931987285614,
-0.10275916755199432,
-0.00494989613071084,
0.01016148179769516,
0.055939435958862305,
-0.0825418159365654,
0.002569616539403796,
0.07285693287849426,
0.0004970607114955783,
0.20126602053642273,
0.007368301972746849,
0.10567810386419296,
0.02094888873398304,
-0.04247940331697464,
-0.0019000426400452852,
-0.005794209893792868,
0.18689772486686707,
0.012832386419177055,
0.058711931109428406,
-0.0021451259963214397,
0.020530931651592255,
-0.02382401004433632,
-0.040779098868370056,
-0.013553539291024208,
0.14804403483867645,
0.03532844036817551,
0.0372658297419548,
0.065127432346344,
-0.07680443674325943,
-0.016946563497185707,
-0.03255370259284973,
-0.07702560722827911,
0.04445130378007889,
-0.029743831604719162,
0.09808193892240524,
0.1286800056695938,
-0.14788751304149628,
0.09486813098192215,
0.02508455701172352,
-0.0717732161283493,
-0.14764946699142456,
-0.1694239228963852,
-0.06376104056835175,
-0.11975701153278351,
0.032085925340652466,
-0.09588439762592316,
0.044271357357501984,
-0.038634054362773895,
0.07195819169282913,
-0.026866860687732697,
0.15485991537570953,
-0.06409212201833725,
-0.09655997902154922,
0.0947074368596077,
-0.030697625130414963,
0.08972999453544617,
0.030048063024878502,
0.024602796882390976,
0.10504567623138428,
0.04463931918144226,
0.0700836107134819,
0.07741361856460571,
-0.002521033165976405,
0.010183733887970448,
-0.050964582711458206,
-0.03226453438401222,
-0.012392151169478893,
0.009180230088531971,
0.06769950687885284,
0.12208211421966553,
0.09010037034749985,
-0.03582295402884483,
0.003391626523807645,
0.15170688927173615,
-0.014051727019250393,
-0.10578140616416931,
-0.13909420371055603,
0.05649175867438316,
-0.012842492200434208,
0.03248002007603645,
-0.01574825309216976,
-0.09021927416324615,
0.05423492193222046,
0.1934429258108139,
0.0895962193608284,
0.06918343156576157,
0.0017571555217728019,
-0.04663822054862976,
0.0017050581518560648,
-0.0020306946244090796,
0.034919869154691696,
-0.02880672551691532,
0.28844016790390015,
0.01385467778891325,
-0.014556138776242733,
-0.02977372333407402,
0.03245392069220543,
-0.04644298180937767,
0.05882618948817253,
-0.07143562287092209,
-0.06539875268936157,
-0.03434126079082489,
0.1414569765329361,
-0.010341309942305088,
-0.14624784886837006,
-0.010275164619088173,
-0.06524009257555008,
-0.11239112168550491,
-0.015154517255723476,
0.0705811157822609,
0.04479611665010452,
0.0317777656018734,
-0.014435583725571632,
-0.03413153067231178,
0.1323484182357788,
-0.011502527631819248,
-0.02386597916483879,
-0.0307972002774477,
-0.004327110014855862,
-0.017203055322170258,
0.16177481412887573,
0.031909361481666565,
0.17329701781272888,
0.0853361189365387,
0.023153144866228104,
-0.09551995992660522,
0.09898629039525986,
0.011089337058365345,
-0.08024810999631882,
0.08102139830589294,
0.03685890510678291,
0.01297071110457182,
0.12670861184597015,
0.07587142288684845,
-0.031167276203632355,
0.05752405896782875,
0.02030976489186287,
-0.016978401690721512,
-0.12620332837104797,
0.09020966291427612,
-0.08406565338373184,
0.08963274210691452,
0.10345397889614105,
-0.05420694127678871,
-0.026152294129133224,
-0.06483414769172668,
0.06583583354949951,
-0.00022699292458128184,
0.0682663768529892,
-0.03283846005797386,
-0.21850666403770447,
0.06396479904651642,
-0.02130032889544964,
0.013195358216762543,
-0.2065049111843109,
0.0010187617735937238,
0.03670193627476692,
-0.04522739350795746,
0.03885338827967644,
0.048142533749341965,
0.04546116292476654,
0.04041464626789093,
0.00982193648815155,
-0.09032813459634781,
0.05652453005313873,
0.10713579505681992,
-0.08173280954360962,
-0.04872133955359459
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1
This model is a fine-tuned version of [patrickvonplaten/t5-tiny-random](https://huggingface.co/patrickvonplaten/t5-tiny-random) on the wmt16_en_ro_pre_processed dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.12.5
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"tags": ["generated_from_trainer"], "datasets": ["wmt16_en_ro_pre_processed"], "model-index": [{"name": "t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1", "results": []}]}
|
text2text-generation
|
diegor2/t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetu-truncated-d22eed
|
[
"transformers",
"pytorch",
"tensorboard",
"t5",
"text2text-generation",
"generated_from_trainer",
"dataset:wmt16_en_ro_pre_processed",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1
This model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.12.5
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
[
"# t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1\n\nThis model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1\n\nThis model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
75,
84,
6,
12,
8,
3,
103,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# t5-tiny-random-length-128-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1\n\nThis model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
-0.05709008872509003,
0.11927098035812378,
-0.004617511760443449,
0.06730909645557404,
0.11596396565437317,
0.04187273979187012,
0.09426993131637573,
0.13758471608161926,
-0.09400095790624619,
0.1102629080414772,
0.1022653877735138,
0.030502766370773315,
0.07178987562656403,
0.10967738926410675,
-0.01583126187324524,
-0.2655247449874878,
0.013467744924128056,
-0.02880304493010044,
-0.03755617514252663,
0.08817952126264572,
0.1055125817656517,
-0.08442045003175735,
0.05540211871266365,
0.014971167780458927,
-0.12490026652812958,
0.030219754204154015,
-0.03315334767103195,
-0.06569980829954147,
0.06485748291015625,
0.054822564125061035,
0.08891119807958603,
0.017102252691984177,
0.09641842544078827,
-0.24234046041965485,
-0.0009137435117736459,
0.07753430306911469,
0.029467258602380753,
0.08735611289739609,
0.13437306880950928,
-0.0025978884659707546,
0.07850024849176407,
-0.11353504657745361,
0.07403537631034851,
0.0390380397439003,
-0.08526816219091415,
-0.1336285024881363,
-0.06875461339950562,
0.07884585857391357,
0.03524702787399292,
0.09164813905954361,
-0.021815024316310883,
0.13367924094200134,
-0.10390229523181915,
0.052494172006845474,
0.20757223665714264,
-0.22793172299861908,
-0.06766355037689209,
0.020855218172073364,
0.04813803359866142,
0.056328773498535156,
-0.10531149059534073,
-0.05438414588570595,
0.056417860090732574,
0.026012269780039787,
0.07402629405260086,
0.009760713204741478,
-0.05985049158334732,
0.014560997486114502,
-0.13292866945266724,
-0.024457793682813644,
0.12232107669115067,
0.063077911734581,
-0.01553588081151247,
-0.12866626679897308,
-0.03588642179965973,
-0.14009803533554077,
-0.014136727899312973,
0.013946224004030228,
0.008439670316874981,
-0.029581492766737938,
-0.06502312421798706,
0.005077533423900604,
-0.053952187299728394,
-0.04304852336645126,
0.015358676202595234,
0.08560247719287872,
0.05319783091545105,
-0.010025322437286377,
-0.022430790588259697,
0.10848801583051682,
0.017360717058181763,
-0.11692287772893906,
-0.022288544103503227,
-0.00906387995928526,
-0.11790267378091812,
-0.060758817940950394,
-0.04634535685181618,
-0.01382813137024641,
0.0015939846634864807,
0.16950981318950653,
-0.022083144634962082,
0.0780637189745903,
0.01897623762488365,
-0.0005750865675508976,
0.0017053531482815742,
0.09643680602312088,
-0.08076963573694229,
-0.015843046829104424,
-0.04041556268930435,
0.11839237809181213,
-0.03367592766880989,
0.0005571143701672554,
-0.053997501730918884,
-0.015276853926479816,
0.0681133046746254,
0.04760294407606125,
-0.01386561430990696,
0.04408302158117294,
-0.03899303078651428,
-0.048677992075681686,
0.016483085229992867,
-0.12397368252277374,
0.028970792889595032,
0.024625003337860107,
-0.09143900126218796,
0.019975846633315086,
-0.018517067655920982,
-0.013799057342112064,
-0.0665363147854805,
0.10040169954299927,
-0.08234543353319168,
-0.023740055039525032,
-0.08345599472522736,
-0.0866333618760109,
0.0054688043892383575,
-0.011504164896905422,
-0.01641460880637169,
-0.07807048410177231,
-0.14782381057739258,
-0.052121490240097046,
0.025514638051390648,
-0.07912075519561768,
-0.08290647715330124,
-0.04948442056775093,
-0.06036050245165825,
0.04498196020722389,
-0.006928268354386091,
0.17450565099716187,
-0.046581558883190155,
0.05555315315723419,
-0.01793009229004383,
0.02117736265063286,
0.09292164444923401,
0.03931750729680061,
-0.07847906649112701,
0.044812873005867004,
-0.04482172057032585,
0.10268163681030273,
-0.08210349082946777,
0.028985973447561264,
-0.12687726318836212,
-0.1052994504570961,
-0.02476135827600956,
-0.006042106077075005,
0.056643567979335785,
0.09919621050357819,
-0.13161787390708923,
-0.03831424564123154,
0.16157886385917664,
-0.021191906183958054,
-0.04432755708694458,
0.09600377827882767,
-0.045722972601652145,
-0.011845790781080723,
0.024734728038311005,
0.13059037923812866,
0.13443294167518616,
-0.07224331796169281,
-0.0237550288438797,
0.00609408225864172,
0.054948482662439346,
0.06951309740543365,
0.05408020690083504,
-0.005628805607557297,
-0.01831846497952938,
0.010840967297554016,
-0.030934780836105347,
0.033490024507045746,
-0.06333128362894058,
-0.0759165957570076,
-0.040580857545137405,
-0.07205416262149811,
0.03413453325629234,
0.013518076390028,
0.04522153362631798,
-0.06393541395664215,
-0.13259480893611908,
0.07440316677093506,
0.14417755603790283,
-0.05691500008106232,
0.00043630634900182486,
-0.07954121381044388,
0.007993186824023724,
-0.005001484416425228,
-0.008876832202076912,
-0.18510702252388,
-0.09054969251155853,
0.06000090017914772,
-0.09694083034992218,
0.03516639769077301,
0.0312630869448185,
0.05998824164271355,
0.05392290651798248,
-0.037595391273498535,
-0.0623365119099617,
-0.05575813353061676,
-0.026951966807246208,
-0.07765889912843704,
-0.16710561513900757,
-0.06398356705904007,
-0.015473054721951485,
0.1545647829771042,
-0.2519260346889496,
0.007008031010627747,
-0.007338352035731077,
0.13545158505439758,
0.024835120886564255,
-0.048080697655677795,
-0.0025866241194307804,
0.03255079314112663,
-0.035318247973918915,
-0.08682288229465485,
0.017810799181461334,
-0.006205420009791851,
-0.07858365029096603,
-0.01177245657891035,
-0.13762745261192322,
-0.013952630572021008,
0.07215963304042816,
0.050190042704343796,
-0.12495835870504379,
-0.032705992460250854,
-0.07952211052179337,
-0.05117698013782501,
-0.07571934163570404,
-0.027867496013641357,
0.16760115325450897,
0.02045644260942936,
0.11859956383705139,
-0.05496976524591446,
-0.06682688742876053,
-0.0033748308196663857,
0.025377536192536354,
-0.025673629716038704,
0.12137100100517273,
0.07430274039506912,
-0.08177538961172104,
0.07021753489971161,
0.025084706023335457,
-0.04476862773299217,
0.13735154271125793,
-0.043067462742328644,
-0.12189779430627823,
-0.022487211972475052,
0.032418206334114075,
0.025932926684617996,
0.04168476164340973,
-0.04831862822175026,
0.015226820483803749,
0.05780674144625664,
0.02304139919579029,
0.03229954093694687,
-0.13433551788330078,
-0.005256619770079851,
0.036398034542798996,
-0.017337799072265625,
0.002974467119202018,
-0.03809984773397446,
0.0456559918820858,
0.09024114906787872,
0.026915688067674637,
-0.018576567992568016,
0.006753949914127588,
-0.029442913830280304,
-0.08449901640415192,
0.17246785759925842,
-0.08424850553274155,
-0.1448841691017151,
-0.14636294543743134,
-0.011763087473809719,
-0.03224742040038109,
-0.029242247343063354,
0.006120714358985424,
-0.08424986898899078,
-0.07160301506519318,
-0.10868678241968155,
0.02444208227097988,
-0.06766071915626526,
0.0057189008221030235,
0.00634659081697464,
0.03275982663035393,
0.09305331856012344,
-0.10919512808322906,
0.03868443891406059,
0.01325088832527399,
-0.04247420281171799,
0.029822947457432747,
0.030723996460437775,
0.0890113115310669,
0.1459866315126419,
0.017178960144519806,
0.006505148950964212,
-0.03366561606526375,
0.20388366281986237,
-0.060909006744623184,
0.0037312349304556847,
0.08624570071697235,
0.019673988223075867,
0.06497128307819366,
0.12264832854270935,
0.038241222500801086,
-0.06313157826662064,
0.06635396927595139,
0.07762094587087631,
0.005777440033853054,
-0.2585159242153168,
-0.03459053859114647,
-0.03256247565150261,
-0.023268144577741623,
0.1824992597103119,
0.04045366868376732,
-0.012412985786795616,
0.062316153198480606,
-0.05064404755830765,
0.07340095192193985,
-0.01644926145672798,
0.0850316733121872,
0.0485713966190815,
0.06252707540988922,
0.10091058909893036,
-0.015122608281672001,
-0.03933166339993477,
0.06960323452949524,
-0.012671895325183868,
0.23885618150234222,
-0.05456520989537239,
0.20089717209339142,
0.001007733982987702,
0.14882639050483704,
0.010524709708988667,
0.022891897708177567,
0.030372951179742813,
-0.0024937442503869534,
0.02147054672241211,
-0.06484334170818329,
0.012579464353621006,
0.05794992670416832,
0.03867074102163315,
0.04554455354809761,
-0.0692601203918457,
0.005673780106008053,
0.0009302058024331927,
0.27629250288009644,
0.05647474527359009,
-0.2953653633594513,
-0.050731729716062546,
0.015137633308768272,
-0.046250466257333755,
-0.07275049388408661,
0.02696944959461689,
0.09772275388240814,
-0.13355296850204468,
0.05034641921520233,
-0.07117211073637009,
0.1013120710849762,
-0.04522736743092537,
-0.007943421602249146,
0.08091134577989578,
0.14600251615047455,
-0.00886393804103136,
0.10362859815359116,
-0.17328795790672302,
0.1650707721710205,
0.006790526676923037,
0.06710229068994522,
-0.07079331576824188,
0.041147615760564804,
0.005633010063320398,
-0.013163259252905846,
0.08473812788724899,
-0.003854673122987151,
-0.07992150634527206,
-0.1648842990398407,
-0.0835828185081482,
-0.0034188509453088045,
0.12398013472557068,
-0.04076618701219559,
0.11472509801387787,
-0.057803213596343994,
0.0006089733215048909,
0.03838690370321274,
-0.08942840993404388,
-0.14229117333889008,
-0.16252931952476501,
0.04026423767209053,
0.019277891144156456,
-0.011194752529263496,
-0.10167691856622696,
-0.09394966065883636,
-0.061265043914318085,
0.2038552314043045,
-0.029703203588724136,
-0.021731097251176834,
-0.13611790537834167,
0.10619734227657318,
0.14310890436172485,
-0.059399958699941635,
-0.005057309754192829,
0.012853640131652355,
0.1632123440504074,
0.015148495323956013,
-0.07976452261209488,
0.07449556887149811,
-0.05049682781100273,
-0.1721445769071579,
-0.06824840605258942,
0.12052479386329651,
0.05176100879907608,
0.059684451669454575,
-0.047471676021814346,
0.010085323825478554,
0.00468733673915267,
-0.08411352336406708,
0.009894711896777153,
0.059184323996305466,
0.007042417302727699,
0.08445189148187637,
-0.05374868959188461,
0.020024467259645462,
-0.06077590212225914,
-0.0193781815469265,
0.11409690231084824,
0.2199353277683258,
-0.0862446129322052,
0.058841507881879807,
0.023794109001755714,
-0.05620453506708145,
-0.1634812355041504,
0.053917884826660156,
0.1321038156747818,
0.014802013523876667,
0.03369368240237236,
-0.2025781273841858,
0.09078411012887955,
0.1085917055606842,
0.008704539388418198,
0.06315187364816666,
-0.3633438050746918,
-0.13199499249458313,
0.06603781133890152,
0.06713613122701645,
-0.040337614715099335,
-0.11721407622098923,
-0.06352689117193222,
-0.04868341609835625,
-0.10371764004230499,
0.08244218677282333,
-0.027800817042589188,
0.10414854437112808,
-0.024239594116806984,
0.08043103665113449,
0.0430714450776577,
-0.04779433831572533,
0.14336436986923218,
0.019343586638569832,
0.0598711296916008,
-0.05052170157432556,
0.036997195333242416,
0.05837104097008705,
-0.06684009730815887,
0.03338437154889107,
-0.057975005358457565,
0.05419343709945679,
-0.1707206666469574,
-0.04604673013091087,
-0.0619371123611927,
0.05129668489098549,
-0.04731737822294235,
-0.053442202508449554,
-0.05756572261452675,
0.05319301784038544,
0.08932993561029434,
-0.01356052327901125,
0.04606381058692932,
0.006878816522657871,
0.04206390306353569,
0.11192388832569122,
0.06980683654546738,
0.0345255471765995,
-0.14146095514297485,
-0.006353897042572498,
0.00021959976584184915,
0.05514724180102348,
-0.09255316853523254,
0.010520709678530693,
0.13390618562698364,
0.043330807238817215,
0.13808660209178925,
0.040304455906152725,
-0.08382896333932877,
-0.010314974933862686,
0.028646159917116165,
-0.08297496289014816,
-0.142769917845726,
-0.03626970201730728,
0.05270570516586304,
-0.17475765943527222,
-0.0764206275343895,
0.10324189066886902,
-0.04564383253455162,
-0.011413595639169216,
0.003950127866119146,
0.020015591755509377,
0.0005931070190854371,
0.1841384470462799,
0.023525243625044823,
0.07561533153057098,
-0.07213105261325836,
0.103927381336689,
0.1159791424870491,
-0.09048119932413101,
0.024285929277539253,
0.09710124880075455,
-0.06548291444778442,
-0.017885437235236168,
0.06613560020923615,
0.0849311575293541,
-0.03913997486233711,
-0.022134168073534966,
-0.05069796368479729,
-0.11555653065443039,
0.0976078137755394,
0.005203305743634701,
0.02516285330057144,
-0.013193902559578419,
-0.030838564038276672,
-0.0005429740995168686,
-0.12613171339035034,
0.08537930250167847,
0.07408834248781204,
0.09636132419109344,
-0.11388944834470749,
0.12637273967266083,
-0.00624658539891243,
0.01788399927318096,
0.0016044883523136377,
0.015298943035304546,
-0.09673194587230682,
-0.020257098600268364,
-0.135057270526886,
0.027041122317314148,
-0.05081659555435181,
0.009457051753997803,
-0.0030391011387109756,
-0.002310622250661254,
-0.036153361201286316,
0.020261581987142563,
-0.06691563874483109,
-0.08848579227924347,
-0.0068577309139072895,
0.08274582028388977,
-0.10478191822767258,
0.0027475913520902395,
0.028106804937124252,
-0.11801239848136902,
0.06826949119567871,
0.03848472610116005,
0.02064460702240467,
0.0036798426881432533,
-0.058319639414548874,
-0.017477931454777718,
0.00687683280557394,
0.05170488730072975,
0.04437217861413956,
-0.10002297163009644,
0.0017525514122098684,
-0.057484451681375504,
0.011934751644730568,
0.008699591271579266,
-0.017951687797904015,
-0.14711196720600128,
-0.016414804384112358,
-0.043959032744169235,
-0.053876638412475586,
-0.0894460380077362,
0.04675034061074257,
0.04768460988998413,
-0.0018745723646134138,
0.16039752960205078,
-0.07835201919078827,
0.05687815323472023,
-0.20940223336219788,
-0.01422239188104868,
0.0006544528296217322,
-0.012438288889825344,
-0.005940085276961327,
-0.0065717026591300964,
0.08417513221502304,
-0.053943634033203125,
0.11366812884807587,
-0.04973527789115906,
0.00504977535456419,
0.038370486348867416,
-0.06721724569797516,
0.014537922106683254,
0.03559356555342674,
0.1890319287776947,
0.07932295650243759,
-0.030224114656448364,
0.055341627448797226,
-0.04052233323454857,
0.07139552384614944,
0.10610698163509369,
0.15159872174263,
0.1718352735042572,
0.01055725384503603,
0.06266667693853378,
0.03944489359855652,
-0.13114456832408905,
-0.08985161036252975,
0.17821577191352844,
-0.05706694722175598,
0.1261405199766159,
-0.02481703832745552,
0.15792441368103027,
0.10720954835414886,
-0.16544681787490845,
0.028498444706201553,
-0.061763063073158264,
-0.09944475442171097,
-0.07900401949882507,
-0.11028233170509338,
-0.08848723024129868,
-0.102122463285923,
0.024835528805851936,
-0.1078030914068222,
0.04513483867049217,
0.06216081976890564,
0.02561766840517521,
0.024977920576930046,
0.11614435166120529,
-0.008316089399158955,
-0.023043937981128693,
0.0746687725186348,
0.009329576045274734,
-0.00886544119566679,
-0.04854857176542282,
-0.06577029079198837,
0.06680873036384583,
-0.009233837015926838,
0.06157410517334938,
-0.03722113370895386,
0.020778408274054527,
0.03872276470065117,
0.006193731911480427,
-0.07187896221876144,
0.031178832054138184,
-0.0021083769388496876,
0.0017360369674861431,
0.05148332193493843,
0.06400139629840851,
0.013037547469139099,
-0.061269018799066544,
0.23764514923095703,
-0.06519699096679688,
-0.054395657032728195,
-0.14597104489803314,
0.15723949670791626,
0.02127237059175968,
0.0048693022690713406,
0.06480362266302109,
-0.09624418616294861,
-0.030628571286797523,
0.16674017906188965,
0.15134011209011078,
-0.055578868836164474,
-0.029650317505002022,
0.014255682937800884,
-0.019581560045480728,
-0.040904391556978226,
0.1124299019575119,
0.09127647429704666,
-0.010574989020824432,
-0.04946616291999817,
-0.007904192432761192,
0.0027625274378806353,
-0.019417215138673782,
-0.06375588476657867,
0.09248808771371841,
0.035064201802015305,
0.012989609502255917,
-0.041373360902071,
0.04976058006286621,
-0.026354148983955383,
-0.12668874859809875,
0.035562023520469666,
-0.09655831009149551,
-0.18326342105865479,
-0.03097566030919552,
0.04751069098711014,
-0.011093631386756897,
0.06684841960668564,
-0.023846320807933807,
-0.02434762939810753,
0.10534418374300003,
-0.019795753061771393,
-0.070963054895401,
-0.08583595603704453,
0.06451946496963501,
-0.08097852766513824,
0.210165873169899,
-0.00022020141477696598,
0.03995175659656525,
0.11573604494333267,
0.02295568212866783,
-0.15617144107818604,
0.020632049068808556,
0.08887045085430145,
-0.06252176314592361,
0.04624367505311966,
0.1871391236782074,
-0.02936403639614582,
0.13221527636051178,
0.05266094207763672,
-0.13510937988758087,
-0.00714847119525075,
-0.008422582410275936,
0.021950559690594673,
-0.09840311110019684,
-0.021098481491208076,
-0.03452279791235924,
0.14449813961982727,
0.23738183081150055,
-0.04632890224456787,
-0.024451863020658493,
-0.0474223792552948,
0.006082580890506506,
0.020504269748926163,
0.12704047560691833,
-0.014174818061292171,
-0.19472260773181915,
0.0005976452957838774,
-0.009906074963510036,
0.0638137087225914,
-0.2582281529903412,
-0.08741550892591476,
0.026998383924365044,
-0.054264653474092484,
-0.043067507445812225,
0.15567581355571747,
0.04286764934659004,
0.045400746166706085,
-0.039456434547901154,
-0.10389768332242966,
-0.04092639684677124,
0.14418795704841614,
-0.14629872143268585,
-0.05903198942542076
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.005-finetuned-en-to-ro-TRAIN_EPOCHS-1
This model is a fine-tuned version of [patrickvonplaten/t5-tiny-random](https://huggingface.co/patrickvonplaten/t5-tiny-random) on the wmt16_en_ro_pre_processed dataset.
It achieves the following results on the evaluation set:
- Loss: 6.4897
- Bleu: 0.0002
- Gen Len: 9.0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|:-------------:|:-----:|:-----:|:---------------:|:------:|:-------:|
| 6.2585 | 1.0 | 76290 | 6.4897 | 0.0002 | 9.0 |
### Framework versions
- Transformers 4.12.5
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"tags": ["generated_from_trainer"], "datasets": ["wmt16_en_ro_pre_processed"], "metrics": ["bleu"], "model-index": [{"name": "t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.005-finetuned-en-to-ro-TRAIN_EPOCHS-1", "results": [{"task": {"type": "text2text-generation", "name": "Sequence-to-sequence Language Modeling"}, "dataset": {"name": "wmt16_en_ro_pre_processed", "type": "wmt16_en_ro_pre_processed", "args": "enro"}, "metrics": [{"type": "bleu", "value": 0.0002, "name": "Bleu"}]}]}]}
|
text2text-generation
|
diegor2/t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.005-finetu-truncated-41f800
|
[
"transformers",
"pytorch",
"tensorboard",
"t5",
"text2text-generation",
"generated_from_trainer",
"dataset:wmt16_en_ro_pre_processed",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
t5-tiny-random-length-96-learning\_rate-2e-05-weight\_decay-0.005-finetuned-en-to-ro-TRAIN\_EPOCHS-1
====================================================================================================
This model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16\_en\_ro\_pre\_processed dataset.
It achieves the following results on the evaluation set:
* Loss: 6.4897
* Bleu: 0.0002
* Gen Len: 9.0
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 1
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.12.5
* Pytorch 1.10.0+cu111
* Datasets 1.16.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
79,
113,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
-0.11762234568595886,
0.11869093775749207,
-0.0023572053760290146,
0.10143560916185379,
0.11423034965991974,
0.007299970835447311,
0.1640792191028595,
0.15716683864593506,
-0.09518809616565704,
0.06109800562262535,
0.15002241730690002,
0.12956224381923676,
0.05164729803800583,
0.17859943211078644,
-0.06912942975759506,
-0.2431810200214386,
0.04195452481508255,
0.0539972260594368,
0.007511356845498085,
0.13410258293151855,
0.08610542118549347,
-0.13231797516345978,
0.09619329124689102,
0.021521935239434242,
-0.19718606770038605,
-0.02784913405776024,
-0.003947020974010229,
-0.06812836229801178,
0.11123358458280563,
0.015217429958283901,
0.09847404807806015,
0.041100531816482544,
0.055412933230400085,
-0.15698154270648956,
0.009932467713952065,
0.049472834914922714,
0.0124352453276515,
0.10668007284402847,
0.05784331262111664,
-0.021936479955911636,
0.06157234311103821,
-0.07268621027469635,
0.06504002213478088,
0.012571565806865692,
-0.13983500003814697,
-0.2507651150226593,
-0.09298445284366608,
0.03559785336256027,
0.059914764016866684,
0.08210179209709167,
-0.02023705281317234,
0.18771420419216156,
-0.02166573703289032,
0.10994362831115723,
0.23456719517707825,
-0.2903193235397339,
-0.05957802012562752,
-0.010040086694061756,
0.04002989083528519,
0.08527655154466629,
-0.08424162119626999,
-0.0370381698012352,
0.03462142497301102,
0.046580422669649124,
0.13432441651821136,
-0.013534643687307835,
-0.02138367109000683,
-0.016424786299467087,
-0.13325262069702148,
-0.06664206832647324,
0.1798846572637558,
0.03395282104611397,
-0.041032228618860245,
-0.07462378591299057,
-0.07721948623657227,
-0.19573287665843964,
-0.03870333731174469,
0.007711920887231827,
0.025154095143079758,
-0.043525438755750656,
-0.10628996044397354,
-0.0072980886325240135,
-0.08416944742202759,
-0.04048290476202965,
-0.0267204400151968,
0.14149685204029083,
0.037552036345005035,
0.019222743809223175,
-0.06756513565778732,
0.09006769210100174,
-0.01791190728545189,
-0.1746922880411148,
-0.0007983801187947392,
0.007149295415729284,
-0.000633381016086787,
-0.04704642668366432,
-0.03918517380952835,
-0.12796062231063843,
-0.008950360119342804,
0.147288516163826,
-0.08190587162971497,
0.07580725848674774,
-0.016965776681900024,
0.030520621687173843,
-0.04649312421679497,
0.17738474905490875,
-0.024898674339056015,
0.00316299544647336,
-0.003948863595724106,
0.08859537541866302,
0.03802084922790527,
-0.04134548828005791,
-0.11161340773105621,
0.034481458365917206,
0.13216163218021393,
0.006964390631765127,
-0.03576691821217537,
0.06078061833977699,
-0.056172117590904236,
-0.040987640619277954,
0.03670203313231468,
-0.09965868294239044,
0.025492802262306213,
-0.010050278156995773,
-0.06910853087902069,
-0.02303549088537693,
-0.0000017711739701553597,
0.012435363605618477,
-0.03250144422054291,
0.08010172843933105,
-0.10580108314752579,
0.01832175813615322,
-0.076022669672966,
-0.14152325689792633,
0.01799888163805008,
-0.10185114294290543,
0.00423926068469882,
-0.08324088156223297,
-0.15379047393798828,
-0.019202036783099174,
0.05792422220110893,
-0.04899311438202858,
-0.04491494596004486,
-0.0637725219130516,
-0.07458434998989105,
0.047107722610235214,
-0.009324663318693638,
0.09004973620176315,
-0.070281021296978,
0.09329541027545929,
0.05171763524413109,
0.0783977210521698,
-0.03991055488586426,
0.04533771425485611,
-0.09523950517177582,
0.04493928700685501,
-0.19492344558238983,
0.07152874767780304,
-0.04124080389738083,
0.0762602910399437,
-0.10788986086845398,
-0.10201715677976608,
0.023198245093226433,
-0.02878539264202118,
0.10517372936010361,
0.10667894035577774,
-0.17774547636508942,
-0.07108975946903229,
0.19496609270572662,
-0.08464464545249939,
-0.13379864394664764,
0.13355891406536102,
-0.057939156889915466,
0.03927577659487724,
0.05111553519964218,
0.23771286010742188,
0.03987357020378113,
-0.08045145869255066,
-0.021587498486042023,
-0.0518181286752224,
0.08131302893161774,
-0.05660994350910187,
0.08115928620100021,
0.01714698225259781,
0.0576365701854229,
0.00576950004324317,
0.016080597415566444,
0.023042194545269012,
-0.0954742506146431,
-0.08584235608577728,
-0.03927142173051834,
-0.07727786153554916,
0.032665006816387177,
0.04684663191437721,
0.07177557796239853,
-0.127925306558609,
-0.10143636167049408,
0.041007235646247864,
0.07521962374448776,
-0.08764663338661194,
0.04201657697558403,
-0.09963101893663406,
0.10520141571760178,
-0.09124429523944855,
-0.015144651755690575,
-0.1870497316122055,
-0.03037218190729618,
0.03510212525725365,
0.01993461325764656,
0.008497117087244987,
-0.03686816617846489,
0.0723988488316536,
0.0770735889673233,
-0.031603146344423294,
-0.04165862128138542,
-0.025696169584989548,
-0.0007530493894591928,
-0.12459482252597809,
-0.1929691731929779,
-0.05122978985309601,
-0.04687740281224251,
0.13179874420166016,
-0.1709364503622055,
0.0370689295232296,
0.05641726776957512,
0.10986407101154327,
0.04858900606632233,
-0.04383236542344093,
0.001462422194890678,
0.07109331339597702,
-0.052814874798059464,
-0.06524305790662766,
0.05974530801177025,
0.02495834231376648,
-0.0936293751001358,
-0.008928978815674782,
-0.16906827688217163,
0.14259333908557892,
0.12805375456809998,
-0.009996016509830952,
-0.06925080716609955,
-0.018337372690439224,
-0.054077088832855225,
-0.027013421058654785,
-0.031310465186834335,
0.01662742905318737,
0.16214929521083832,
0.028958074748516083,
0.15343286097049713,
-0.0956718772649765,
-0.05259385332465172,
0.05243600159883499,
-0.03279923275113106,
-0.01587907038629055,
0.119529590010643,
0.02768615446984768,
-0.1405152976512909,
0.1385892629623413,
0.12361547350883484,
-0.03835688903927803,
0.14982730150222778,
-0.04833187162876129,
-0.07119720429182053,
-0.04328945279121399,
-0.016208985820412636,
0.024593528360128403,
0.11244405061006546,
-0.12417406588792801,
-0.024463942274451256,
0.04019009321928024,
0.02645927667617798,
0.01019915472716093,
-0.18609707057476044,
-0.005488212686032057,
0.042220715433359146,
-0.03325406461954117,
-0.052020709961652756,
-0.012072951532900333,
0.008352301083505154,
0.09980355203151703,
0.019048640504479408,
-0.04987288638949394,
0.031916189938783646,
0.009868484921753407,
-0.0731106549501419,
0.19412536919116974,
-0.09231875091791153,
-0.161652073264122,
-0.11874702572822571,
-0.10801995545625687,
-0.0611783005297184,
0.00048161778249777853,
0.06863609701395035,
-0.0839429497718811,
-0.03945045545697212,
-0.10220512002706528,
-0.02266642451286316,
-0.01633346639573574,
0.02256220206618309,
0.04726168513298035,
-0.02418660931289196,
0.0557505302131176,
-0.10342051833868027,
-0.028171850368380547,
-0.023017926141619682,
0.02813233993947506,
0.08183634281158447,
0.0005623606848530471,
0.11474134773015976,
0.13263055682182312,
-0.02173576131463051,
0.05058987811207771,
-0.033019013702869415,
0.2605779767036438,
-0.07205478101968765,
-0.02111814357340336,
0.11403048038482666,
-0.030385207384824753,
0.08755756914615631,
0.12383577972650528,
0.04894635081291199,
-0.09778329730033875,
0.0010194523492828012,
0.0018163762288168073,
-0.04755605384707451,
-0.21128712594509125,
-0.015195955522358418,
-0.05065782740712166,
-0.00004652851930586621,
0.11654745042324066,
0.015641285106539726,
0.019468318670988083,
0.06567083299160004,
0.02396034449338913,
0.05690566822886467,
-0.030373839661478996,
0.10636243224143982,
0.09223578870296478,
0.06107299402356148,
0.1403898447751999,
-0.04286037012934685,
-0.03938502073287964,
0.03971216827630997,
0.01967073604464531,
0.2115047723054886,
-0.015042317099869251,
0.20725736021995544,
0.035722766071558,
0.15516093373298645,
0.029075054451823235,
0.07914332300424576,
-0.019137857481837273,
-0.009022805839776993,
-0.017626997083425522,
-0.04622987285256386,
-0.053369611501693726,
0.022175567224621773,
-0.048831816762685776,
0.033397894352674484,
-0.12055544555187225,
0.029367191717028618,
0.04557763785123825,
0.2898017168045044,
0.06295573711395264,
-0.39455604553222656,
-0.10984548926353455,
0.011520154774188995,
-0.025694023817777634,
-0.050617095082998276,
-0.003189696464687586,
0.11023873835802078,
-0.08039010316133499,
0.062483400106430054,
-0.0841727927327156,
0.1047835573554039,
-0.07993128150701523,
0.03933681547641754,
0.040984656661748886,
0.07828709483146667,
-0.022479912266135216,
0.051881108433008194,
-0.26288700103759766,
0.27646929025650024,
0.03154194355010986,
0.06033966690301895,
-0.08005660027265549,
0.0014565272722393274,
-0.0012533552944660187,
0.02290274016559124,
0.06715632230043411,
-0.0039542182348668575,
-0.10716034471988678,
-0.1828795224428177,
-0.10539639741182327,
0.01487481314688921,
0.09432888776063919,
-0.007215550635010004,
0.12975841760635376,
-0.007039083167910576,
0.008705227635800838,
0.03998817875981331,
-0.016330434009432793,
-0.04160052910447121,
-0.10907219350337982,
0.03420240432024002,
0.03598343953490257,
-0.027357246726751328,
-0.06391917169094086,
-0.10998329520225525,
-0.046154625713825226,
0.17030061781406403,
0.035780105739831924,
-0.0666741356253624,
-0.13231754302978516,
0.05470671132206917,
0.08324086666107178,
-0.09211993217468262,
0.007834542542696,
-0.00884043425321579,
0.1283423900604248,
-0.013367167674005032,
-0.06381797045469284,
0.1153179332613945,
-0.04600128158926964,
-0.16931836307048798,
-0.05117921531200409,
0.11103887856006622,
0.031738847494125366,
0.0604439415037632,
-0.008430639281868935,
0.04281913861632347,
-0.04082879424095154,
-0.06357698142528534,
0.04837494343519211,
-0.014585636556148529,
0.08765307068824768,
-0.03493117541074753,
0.0034805526956915855,
0.02771511673927307,
-0.07381080836057663,
-0.017923492938280106,
0.1903163492679596,
0.2600342333316803,
-0.07962862402200699,
0.04317159950733185,
0.053412601351737976,
-0.04991797357797623,
-0.14828744530677795,
0.03187364339828491,
0.0640186071395874,
0.013400946743786335,
-0.0022386410273611546,
-0.17947594821453094,
0.03314914554357529,
0.08393094688653946,
-0.008953769691288471,
0.07577149569988251,
-0.30214136838912964,
-0.12157702445983887,
0.08671846985816956,
0.13416126370429993,
0.09818107634782791,
-0.1420474499464035,
-0.04704982414841652,
-0.020354393869638443,
-0.15329058468341827,
0.13377605378627777,
-0.09118147194385529,
0.11876503378152847,
-0.025628000497817993,
0.11327492445707321,
0.02060660906136036,
-0.06692808121442795,
0.12289921194314957,
0.01670100726187229,
0.07618942856788635,
-0.06049386039376259,
-0.002190167084336281,
0.09919718652963638,
-0.08607960492372513,
0.04010126739740372,
-0.07555827498435974,
0.03554534912109375,
-0.13414476811885834,
-0.01853867992758751,
-0.06887606531381607,
-0.0025901468470692635,
-0.03180156648159027,
-0.04927222430706024,
-0.028588663786649704,
0.022650739178061485,
0.0678052082657814,
-0.01880151964724064,
0.1844644397497177,
0.012790876440703869,
0.14812974631786346,
0.15727336704730988,
0.09984400123357773,
-0.11213783174753189,
-0.03562299162149429,
0.00147440645378083,
-0.027875427156686783,
0.043726980686187744,
-0.14856931567192078,
0.03226180747151375,
0.14352664351463318,
0.010534687899053097,
0.13310667872428894,
0.06983016431331635,
-0.06292122602462769,
0.032707955688238144,
0.05462546646595001,
-0.18549145758152008,
-0.10051966458559036,
-0.00607857434079051,
0.030064338818192482,
-0.14013607800006866,
0.03623500466346741,
0.13691271841526031,
-0.05048292502760887,
-0.022949567064642906,
-0.006856056395918131,
0.016647638753056526,
-0.013323841616511345,
0.18410931527614594,
0.03419708460569382,
0.07162408530712128,
-0.11024314910173416,
0.07131536304950714,
0.0585392527282238,
-0.09447906166315079,
0.05566360801458359,
0.09642662853002548,
-0.09855270385742188,
-0.027651235461235046,
0.02275582030415535,
0.18432186543941498,
-0.056081533432006836,
-0.036507099866867065,
-0.16177736222743988,
-0.10799726843833923,
0.08446124196052551,
0.16069437563419342,
0.06766130030155182,
0.011565064080059528,
-0.03849225491285324,
-0.017097854986786842,
-0.13178011775016785,
0.09879805147647858,
0.06368597596883774,
0.08098889142274857,
-0.12154489010572433,
0.13035956025123596,
-0.02147497609257698,
0.05066727101802826,
-0.010339966975152493,
0.0016182753024622798,
-0.11688407510519028,
0.010812913998961449,
-0.11991787701845169,
0.0027616361621767282,
-0.05109553784132004,
-0.004306616261601448,
-0.023346152156591415,
-0.040800172835588455,
-0.059159934520721436,
0.005371925421059132,
-0.11750318855047226,
-0.031729988753795624,
0.020698199048638344,
0.0195570457726717,
-0.11663636565208435,
-0.021492941305041313,
0.013048199005424976,
-0.08973711729049683,
0.08592503517866135,
0.050610002130270004,
-0.008153469301760197,
0.013652211986482143,
-0.021136576309800148,
0.005034693516790867,
0.04347935691475868,
0.0007709894562140107,
0.08154866099357605,
-0.10392054170370102,
-0.017111673951148987,
-0.004580767825245857,
0.02385145239531994,
0.03275898098945618,
0.12673428654670715,
-0.11407846957445145,
0.01763104274868965,
0.0015344946878030896,
-0.059555843472480774,
-0.07802026718854904,
0.07592623680830002,
0.08971939980983734,
0.02620142512023449,
0.17428965866565704,
-0.06667599827051163,
0.0389961302280426,
-0.19890591502189636,
-0.006840529385954142,
0.017310786992311478,
-0.1390291005373001,
-0.05306535214185715,
-0.04737800359725952,
0.07318258285522461,
-0.07958079874515533,
0.10994521528482437,
0.003948519472032785,
0.026142938062548637,
0.03524551913142204,
-0.027420584112405777,
-0.03895360603928566,
0.014092511497437954,
0.1660822629928589,
0.02306404709815979,
-0.04847608879208565,
0.0829785168170929,
0.030796896666288376,
0.08912364393472672,
0.1585051566362381,
0.20633696019649506,
0.12082184851169586,
0.06597544997930527,
0.09287811070680618,
0.020017962902784348,
-0.027647288516163826,
-0.18998582661151886,
0.0569041445851326,
-0.03378736972808838,
0.13026787340641022,
0.0016653402708470821,
0.18070603907108307,
0.11361140757799149,
-0.15571962296962738,
0.036161258816719055,
-0.04233860224485397,
-0.09633176773786545,
-0.10275135934352875,
-0.10164400190114975,
-0.09229782968759537,
-0.11971606314182281,
-0.0032918492797762156,
-0.1305738240480423,
0.03702491149306297,
0.07152602076530457,
0.020437048748135567,
-0.0006849247147329152,
0.1300862729549408,
0.03912721946835518,
0.0186474546790123,
0.062306176871061325,
0.002283879090100527,
-0.032215140759944916,
-0.05778397619724274,
-0.07438582181930542,
0.01736138015985489,
0.012684022076427937,
0.0587305873632431,
-0.018226783722639084,
-0.014013045467436314,
0.03653502091765404,
-0.02111346647143364,
-0.13109880685806274,
0.0017689357046037912,
0.028933217748999596,
0.06767170876264572,
0.0239518191665411,
0.013628571294248104,
-0.0010616782819852233,
-0.021282747387886047,
0.19437652826309204,
-0.0683252289891243,
-0.03944990783929825,
-0.1229279413819313,
0.22812527418136597,
0.00614332826808095,
-0.03871673718094826,
0.04051067680120468,
-0.07701557874679565,
0.0008712558192200959,
0.18546900153160095,
0.1815793216228485,
-0.050347838550806046,
-0.008239129558205605,
-0.011359622702002525,
-0.005814899690449238,
-0.02015700563788414,
0.09554057568311691,
0.10188080370426178,
0.013480775989592075,
-0.08455420285463333,
-0.02091820165514946,
-0.06885595619678497,
-0.02176433987915516,
-0.03144792094826698,
0.07389221340417862,
0.03639128804206848,
0.011252909898757935,
-0.05517052859067917,
0.05670837685465813,
-0.0698084905743599,
-0.0569416843354702,
0.014838695526123047,
-0.21993525326251984,
-0.16920848190784454,
0.0009378113900311291,
0.059745945036411285,
-0.012583688832819462,
0.07014261931180954,
-0.011559108272194862,
0.00407826853916049,
0.07544203847646713,
-0.010787494480609894,
-0.0671285018324852,
-0.07419276237487793,
0.10376418381929398,
-0.15143460035324097,
0.18049044907093048,
-0.03833585977554321,
0.03159596398472786,
0.1409854143857956,
0.041617877781391144,
-0.10967668890953064,
0.0543251670897007,
0.04749006778001785,
-0.022837087512016296,
0.005799754522740841,
0.14313364028930664,
-0.03233359009027481,
0.0897451639175415,
0.05039709433913231,
-0.1327231377363205,
-0.010367914102971554,
-0.07829894870519638,
-0.009230230934917927,
-0.02326442115008831,
-0.054556041955947876,
-0.0356777086853981,
0.1344176083803177,
0.19883574545383453,
-0.04459614306688309,
-0.007842350751161575,
-0.05562353879213333,
0.016196677461266518,
0.0649767741560936,
-0.0013992063468322158,
-0.06464970856904984,
-0.2755783796310425,
-0.005317751318216324,
0.08297022432088852,
-0.004270834382623434,
-0.2808447778224945,
-0.08667538315057755,
0.001977662555873394,
-0.04584384709596634,
-0.08973360806703568,
0.09893858432769775,
0.08138792961835861,
0.051962096244096756,
-0.07165896147489548,
0.007909631356596947,
-0.06990186870098114,
0.16112998127937317,
-0.1338447779417038,
-0.07359246164560318
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1
This model is a fine-tuned version of [patrickvonplaten/t5-tiny-random](https://huggingface.co/patrickvonplaten/t5-tiny-random) on the wmt16_en_ro_pre_processed dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.12.5
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"tags": ["generated_from_trainer"], "datasets": ["wmt16_en_ro_pre_processed"], "model-index": [{"name": "t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1", "results": []}]}
|
text2text-generation
|
diegor2/t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1
|
[
"transformers",
"pytorch",
"tensorboard",
"t5",
"text2text-generation",
"generated_from_trainer",
"dataset:wmt16_en_ro_pre_processed",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1
This model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.12.5
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
[
"# t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1\n\nThis model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1\n\nThis model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
75,
84,
6,
12,
8,
3,
103,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #t5 #text2text-generation #generated_from_trainer #dataset-wmt16_en_ro_pre_processed #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# t5-tiny-random-length-96-learning_rate-2e-05-weight_decay-0.01-finetuned-en-to-ro-TRAIN_EPOCHS-1\n\nThis model is a fine-tuned version of patrickvonplaten/t5-tiny-random on the wmt16_en_ro_pre_processed dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 1\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.12.5\n- Pytorch 1.10.0+cu111\n- Datasets 1.16.1\n- Tokenizers 0.10.3"
] |
[
-0.056631602346897125,
0.11787871271371841,
-0.0046043493784964085,
0.0665258914232254,
0.11544112116098404,
0.04180483520030975,
0.09397821873426437,
0.13814657926559448,
-0.09320984780788422,
0.11069393903017044,
0.10204347223043442,
0.03175485506653786,
0.07148104906082153,
0.1108565554022789,
-0.016524478793144226,
-0.26442015171051025,
0.013733095489442348,
-0.02832566201686859,
-0.03647289797663689,
0.088606096804142,
0.10586108267307281,
-0.08495404571294785,
0.05506204068660736,
0.015005960129201412,
-0.12671828269958496,
0.030138947069644928,
-0.033784572035074234,
-0.06529321521520615,
0.06425530463457108,
0.053793855011463165,
0.08883589506149292,
0.01692148484289646,
0.09546376764774323,
-0.24125109612941742,
-0.0009367919992655516,
0.07791248708963394,
0.029386993497610092,
0.08751307427883148,
0.1343144327402115,
-0.003459260566160083,
0.07966095954179764,
-0.11363314837217331,
0.07463254034519196,
0.039064303040504456,
-0.08536045998334885,
-0.1358259916305542,
-0.06860033422708511,
0.07989724725484848,
0.037368062883615494,
0.09192100912332535,
-0.022287510335445404,
0.13449133932590485,
-0.10404834151268005,
0.05227380245923996,
0.207101508975029,
-0.2285856306552887,
-0.06768717616796494,
0.01989390328526497,
0.04813970625400543,
0.055098723620176315,
-0.10625729709863663,
-0.053938254714012146,
0.05557578057050705,
0.02472921460866928,
0.07372946292161942,
0.009310418739914894,
-0.05986955761909485,
0.0137112932279706,
-0.13289520144462585,
-0.023523304611444473,
0.12364499270915985,
0.06263020634651184,
-0.01507954578846693,
-0.12966999411582947,
-0.0355944000184536,
-0.13825826346874237,
-0.013861612416803837,
0.013696298003196716,
0.008590338751673698,
-0.02977505326271057,
-0.06408720463514328,
0.005108275916427374,
-0.054610103368759155,
-0.043532200157642365,
0.014644274488091469,
0.08574002981185913,
0.05282832682132721,
-0.010613739490509033,
-0.022808078676462173,
0.10874951630830765,
0.01764606684446335,
-0.11718785017728806,
-0.022361015900969505,
-0.00889099296182394,
-0.11846166849136353,
-0.06121009960770607,
-0.04658886417746544,
-0.014187110587954521,
0.001605028985068202,
0.16910941898822784,
-0.020220011472702026,
0.07802926748991013,
0.017803648486733437,
-0.0002775304310489446,
0.0022975821048021317,
0.09664186835289001,
-0.08048957586288452,
-0.015510479919612408,
-0.04091772064566612,
0.11807940900325775,
-0.03438655659556389,
0.0008084089495241642,
-0.05422388017177582,
-0.014541549608111382,
0.06900898367166519,
0.04797826707363129,
-0.01316283829510212,
0.043659936636686325,
-0.039724916219711304,
-0.048930395394563675,
0.014407698065042496,
-0.12367306649684906,
0.02959810383617878,
0.024886997416615486,
-0.09041398763656616,
0.019973959773778915,
-0.018842622637748718,
-0.013130302540957928,
-0.06607718020677567,
0.10064446181058884,
-0.08283235132694244,
-0.023682622238993645,
-0.08326602727174759,
-0.0856652781367302,
0.005090038292109966,
-0.012650558724999428,
-0.01637852191925049,
-0.07707411050796509,
-0.14689551293849945,
-0.05287438631057739,
0.02593042701482773,
-0.07921356707811356,
-0.08202292770147324,
-0.049471721053123474,
-0.060180842876434326,
0.04568793624639511,
-0.006941589992493391,
0.17705394327640533,
-0.04626404866576195,
0.055375657975673676,
-0.018455836921930313,
0.020974552258849144,
0.09406048059463501,
0.04049371927976608,
-0.07795688509941101,
0.044730376452207565,
-0.045585747808218,
0.10239806026220322,
-0.08196776360273361,
0.028196273371577263,
-0.12795232236385345,
-0.10455362498760223,
-0.02352260611951351,
-0.0064420499838888645,
0.05592567101120949,
0.0991600751876831,
-0.13108892738819122,
-0.038802627474069595,
0.16154059767723083,
-0.021158764138817787,
-0.04417623579502106,
0.094035804271698,
-0.04587312042713165,
-0.010322966612875462,
0.024032019078731537,
0.13204558193683624,
0.13266748189926147,
-0.07334680110216141,
-0.023509491235017776,
0.004876741673797369,
0.0551152378320694,
0.06977300345897675,
0.0545833595097065,
-0.006664247252047062,
-0.017849529162049294,
0.01006584707647562,
-0.03383105993270874,
0.0328330360352993,
-0.0632636770606041,
-0.0755874365568161,
-0.04012148454785347,
-0.07181794941425323,
0.03348807618021965,
0.013894940726459026,
0.04523428529500961,
-0.06347018480300903,
-0.13235674798488617,
0.07384628057479858,
0.1441386193037033,
-0.0565897561609745,
0.0000069161255851213355,
-0.0791994035243988,
0.006915595382452011,
-0.005266773514449596,
-0.008622581139206886,
-0.185341015458107,
-0.09074008464813232,
0.0601409375667572,
-0.09797332435846329,
0.035222023725509644,
0.030883656814694405,
0.05989597737789154,
0.05276988074183464,
-0.03703644499182701,
-0.06305816769599915,
-0.05591119825839996,
-0.0267522931098938,
-0.07797984033823013,
-0.16861113905906677,
-0.06323079764842987,
-0.016167163848876953,
0.15661758184432983,
-0.2504924237728119,
0.006171052344143391,
-0.006256165448576212,
0.13650764524936676,
0.025263434275984764,
-0.048194725066423416,
-0.001928650657646358,
0.03189942240715027,
-0.03484984114766121,
-0.08696409314870834,
0.01860521174967289,
-0.006830961909145117,
-0.07839842885732651,
-0.011145244352519512,
-0.13778150081634521,
-0.013853974640369415,
0.07227424532175064,
0.051600173115730286,
-0.1253069043159485,
-0.033642981201410294,
-0.07937423139810562,
-0.05093386024236679,
-0.07558973878622055,
-0.02790992334485054,
0.16964726150035858,
0.02134382352232933,
0.11822474747896194,
-0.05455530434846878,
-0.06834044307470322,
-0.003623714204877615,
0.025937959551811218,
-0.026282712817192078,
0.12108714133501053,
0.07381772249937057,
-0.08537333458662033,
0.0704842060804367,
0.028139574453234673,
-0.043369825929403305,
0.1379416137933731,
-0.04296037554740906,
-0.1221066415309906,
-0.02283366024494171,
0.03264867141842842,
0.026054810732603073,
0.04140324890613556,
-0.04924526810646057,
0.014312156476080418,
0.05815020576119423,
0.022711779922246933,
0.031952425837516785,
-0.13389714062213898,
-0.00498973298817873,
0.03545086085796356,
-0.016403770074248314,
0.00466839550063014,
-0.03727847710251808,
0.04544800892472267,
0.09036378562450409,
0.02767266146838665,
-0.01889953389763832,
0.006227758247405291,
-0.029360083863139153,
-0.08398020267486572,
0.17272911965847015,
-0.08434724807739258,
-0.1444651335477829,
-0.14536839723587036,
-0.01208073366433382,
-0.03254128247499466,
-0.029700981453061104,
0.006147012580186129,
-0.08441782742738724,
-0.07225579023361206,
-0.10740978270769119,
0.02463182434439659,
-0.06742885708808899,
0.005653770174831152,
0.00706555787473917,
0.03250586614012718,
0.09190350025892258,
-0.10929565131664276,
0.038486793637275696,
0.012547419406473637,
-0.04313826188445091,
0.029604637995362282,
0.03088773600757122,
0.08872383832931519,
0.1457955688238144,
0.01881452091038227,
0.007475847844034433,
-0.03331190347671509,
0.2016729861497879,
-0.06211965158581734,
0.0035650120116770267,
0.08585849404335022,
0.019664455205202103,
0.0643996074795723,
0.1226562038064003,
0.037492990493774414,
-0.0635996162891388,
0.06578364223241806,
0.07730058580636978,
0.006533101201057434,
-0.258542001247406,
-0.03481921926140785,
-0.03296151012182236,
-0.02648928388953209,
0.1832960695028305,
0.040504731237888336,
-0.012480144388973713,
0.062064073979854584,
-0.050494447350502014,
0.07185807079076767,
-0.015972815454006195,
0.08575927466154099,
0.04862399771809578,
0.06165400519967079,
0.10148055106401443,
-0.014638036489486694,
-0.039737164974212646,
0.06900030374526978,
-0.013341930694878101,
0.24091456830501556,
-0.05329604819417,
0.19876261055469513,
0.0019108577398583293,
0.14647826552391052,
0.009457368403673172,
0.023342393338680267,
0.029810888692736626,
-0.0024746721610426903,
0.021178817376494408,
-0.06493061035871506,
0.013506438583135605,
0.058133963495492935,
0.039684537798166275,
0.04464063048362732,
-0.06940717250108719,
0.003308260114863515,
0.0011667190119624138,
0.27665433287620544,
0.05692668259143829,
-0.29442232847213745,
-0.04950625076889992,
0.01527989748865366,
-0.04592115059494972,
-0.0734536200761795,
0.025804542005062103,
0.09658721834421158,
-0.1338421255350113,
0.051553770899772644,
-0.07077246904373169,
0.10140788555145264,
-0.0452389232814312,
-0.008336643688380718,
0.07998733222484589,
0.1450849175453186,
-0.00826860312372446,
0.10297052562236786,
-0.17277993261814117,
0.16685305535793304,
0.007455681450664997,
0.06726912409067154,
-0.07112179696559906,
0.04124021902680397,
0.005821870174258947,
-0.013338168151676655,
0.08578284084796906,
-0.0040087648667395115,
-0.07770410925149918,
-0.16354532539844513,
-0.0838150829076767,
-0.003710482269525528,
0.12493912875652313,
-0.042984917759895325,
0.11457573622465134,
-0.05784943327307701,
0.00023458185023628175,
0.03828202560544014,
-0.08809861540794373,
-0.1414528340101242,
-0.16227878630161285,
0.041296739131212234,
0.019436882808804512,
-0.011218874715268612,
-0.1019541397690773,
-0.09392792731523514,
-0.05955357477068901,
0.20443202555179596,
-0.031493522226810455,
-0.02169983834028244,
-0.13607589900493622,
0.10545118898153305,
0.1434822380542755,
-0.05945984646677971,
-0.005347317550331354,
0.012256396934390068,
0.16366486251354218,
0.014983062632381916,
-0.07951177656650543,
0.07480040192604065,
-0.050647735595703125,
-0.17217031121253967,
-0.06832585483789444,
0.1209762915968895,
0.051527224481105804,
0.05904305353760719,
-0.04634649679064751,
0.011450720019638538,
0.0032524766866117716,
-0.08413973450660706,
0.011194942519068718,
0.060068730264902115,
0.008235855959355831,
0.08427159488201141,
-0.05445070564746857,
0.018720068037509918,
-0.06082438677549362,
-0.019743096083402634,
0.11472735553979874,
0.2196866124868393,
-0.08572110533714294,
0.059946879744529724,
0.024263931438326836,
-0.05637560784816742,
-0.16361089050769806,
0.05444932356476784,
0.13278500735759735,
0.014316564425826073,
0.03222798556089401,
-0.20341865718364716,
0.09085401892662048,
0.10853783786296844,
0.008251668885350227,
0.06298128515481949,
-0.3610743582248688,
-0.131814107298851,
0.06728412955999374,
0.06695558875799179,
-0.0403561145067215,
-0.11762755364179611,
-0.0636293813586235,
-0.04783136025071144,
-0.102586530148983,
0.0835040956735611,
-0.03064260259270668,
0.1047607883810997,
-0.02471918985247612,
0.08013560622930527,
0.042967285960912704,
-0.047386329621076584,
0.1426991969347,
0.019789189100265503,
0.05976302549242973,
-0.05006599426269531,
0.036692894995212555,
0.05890519171953201,
-0.06720495223999023,
0.033019330352544785,
-0.058308862149715424,
0.05361495912075043,
-0.17039810121059418,
-0.04562047868967056,
-0.06227567791938782,
0.05222625657916069,
-0.04712488129734993,
-0.05350765958428383,
-0.05746276676654816,
0.053348079323768616,
0.08909228444099426,
-0.013826797716319561,
0.045900557190179825,
0.006841877941042185,
0.04299059137701988,
0.11133641004562378,
0.06988245248794556,
0.034484799951314926,
-0.14184683561325073,
-0.006688105873763561,
-0.0002480304683558643,
0.05544275790452957,
-0.09283553808927536,
0.009894086048007011,
0.13318830728530884,
0.04343603923916817,
0.13860097527503967,
0.04062367230653763,
-0.08403001725673676,
-0.010945811867713928,
0.029048118740320206,
-0.08285050839185715,
-0.14216996729373932,
-0.0357305109500885,
0.054818570613861084,
-0.17437496781349182,
-0.07653685659170151,
0.10320449620485306,
-0.044202666729688644,
-0.011794478632509708,
0.003786307293921709,
0.020225966349244118,
0.0009550524409860373,
0.18401935696601868,
0.02277955412864685,
0.07562919706106186,
-0.07266271859407425,
0.10289701819419861,
0.11610280722379684,
-0.09044143557548523,
0.025210218504071236,
0.09688473492860794,
-0.06520599871873856,
-0.018408391624689102,
0.06657665967941284,
0.0879712849855423,
-0.038714975118637085,
-0.021188905462622643,
-0.04920013248920441,
-0.11570737510919571,
0.09763138741254807,
0.005411895923316479,
0.02489103190600872,
-0.013461613096296787,
-0.03062729723751545,
-0.0011068463791161776,
-0.12549804151058197,
0.08512897044420242,
0.07286351174116135,
0.09605209529399872,
-0.11321480572223663,
0.12572899460792542,
-0.006097028497606516,
0.01729481853544712,
0.0015926484484225512,
0.015078602358698845,
-0.09671282768249512,
-0.020240608602762222,
-0.13152334094047546,
0.026581423357129097,
-0.05008687078952789,
0.009176368825137615,
-0.0029416829347610474,
-0.0029112910851836205,
-0.03625399246811867,
0.02046896703541279,
-0.06759505718946457,
-0.08891621232032776,
-0.007474572863429785,
0.08314427733421326,
-0.10554684698581696,
0.0022893091663718224,
0.02820564992725849,
-0.11867024749517441,
0.0689779743552208,
0.03870166838169098,
0.021237943321466446,
0.0036658886820077896,
-0.05932406336069107,
-0.018162280321121216,
0.006632550619542599,
0.051589637994766235,
0.04473999887704849,
-0.10220351815223694,
0.0019193864427506924,
-0.05780304595828056,
0.012366876937448978,
0.008966704830527306,
-0.016829578205943108,
-0.14687588810920715,
-0.017386039718985558,
-0.044708821922540665,
-0.054383520036935806,
-0.08913388103246689,
0.04746411368250847,
0.04837307706475258,
-0.0010121418163180351,
0.16045449674129486,
-0.07791109383106232,
0.05720148980617523,
-0.20868326723575592,
-0.013924580067396164,
0.00022596784401685,
-0.012732267379760742,
-0.004645845387130976,
-0.007018127478659153,
0.08376644551753998,
-0.05477449670433998,
0.11342956870794296,
-0.049610696732997894,
0.004895092453807592,
0.038540031760931015,
-0.06613683700561523,
0.013519113883376122,
0.035135142505168915,
0.18851150572299957,
0.07961241155862808,
-0.029862722381949425,
0.05553029477596283,
-0.04006442427635193,
0.07094796746969223,
0.10606145113706589,
0.15209262073040009,
0.17197860777378082,
0.010497191920876503,
0.06174090877175331,
0.03997907042503357,
-0.13273777067661285,
-0.09059038758277893,
0.17867276072502136,
-0.05635261535644531,
0.1259535551071167,
-0.024758098646998405,
0.15818536281585693,
0.10723002254962921,
-0.16611865162849426,
0.02842201478779316,
-0.06187174841761589,
-0.09864171594381332,
-0.078596331179142,
-0.1079121008515358,
-0.08914778381586075,
-0.10134769976139069,
0.024959851056337357,
-0.10745418071746826,
0.045310065150260925,
0.05816548690199852,
0.02571956254541874,
0.025282833725214005,
0.11694740504026413,
-0.008331909775733948,
-0.023679496720433235,
0.07576245069503784,
0.009517555125057697,
-0.010285258293151855,
-0.049372389912605286,
-0.0651988610625267,
0.06646481156349182,
-0.009860965423285961,
0.06181446835398674,
-0.03677859902381897,
0.019601725041866302,
0.03800074756145477,
0.005360580049455166,
-0.07232724130153656,
0.03143922984600067,
-0.0021593007259070873,
0.001880782307125628,
0.05058691278100014,
0.06385179609060287,
0.012900550849735737,
-0.06159372627735138,
0.23690465092658997,
-0.06500072777271271,
-0.05376458168029785,
-0.14477860927581787,
0.15814808011054993,
0.02130182459950447,
0.004568759351968765,
0.06440545618534088,
-0.09559925645589828,
-0.031199097633361816,
0.16594623029232025,
0.14995665848255157,
-0.05494021624326706,
-0.02932974509894848,
0.014608723111450672,
-0.019364697858691216,
-0.040254149585962296,
0.11211951076984406,
0.09183353930711746,
-0.009816955775022507,
-0.04943426698446274,
-0.006117813289165497,
0.002848985604941845,
-0.019963867962360382,
-0.0633416697382927,
0.09221811592578888,
0.03600791469216347,
0.0124776940792799,
-0.041006263345479965,
0.0501977875828743,
-0.025056466460227966,
-0.1257382035255432,
0.036712564527988434,
-0.09802277386188507,
-0.18408849835395813,
-0.03127814829349518,
0.04793616384267807,
-0.011905920691788197,
0.06729704141616821,
-0.02344444952905178,
-0.025012535974383354,
0.10412199795246124,
-0.01926272176206112,
-0.07065167278051376,
-0.08695255219936371,
0.06498489528894424,
-0.08277948200702667,
0.20977884531021118,
-0.00046672203461639583,
0.03964032232761383,
0.11598142236471176,
0.022097308188676834,
-0.15688343346118927,
0.021023457869887352,
0.08846724778413773,
-0.06307440996170044,
0.04722738265991211,
0.18822714686393738,
-0.0299412589520216,
0.1305421143770218,
0.052611421793699265,
-0.13534091413021088,
-0.006687596905976534,
-0.009113206528127193,
0.021665554493665695,
-0.09849169105291367,
-0.020312203094363213,
-0.03494182601571083,
0.14518627524375916,
0.2372259497642517,
-0.04589560627937317,
-0.025082694366574287,
-0.04798869043588638,
0.006730161141604185,
0.0212616715580225,
0.1281527876853943,
-0.014641189947724342,
-0.19420166313648224,
0.0009018320124596357,
-0.008928404189646244,
0.0645056739449501,
-0.2583758533000946,
-0.087093286216259,
0.026385504752397537,
-0.05421748757362366,
-0.04382237046957016,
0.1547597348690033,
0.04255213961005211,
0.04500274732708931,
-0.03956538811326027,
-0.10570084303617477,
-0.040368497371673584,
0.14349961280822754,
-0.1466028094291687,
-0.05812155827879906
] |
null | null |
transformers
|
# Twitter4SSE
This model maps texts to 768 dimensional dense embeddings that encode semantic similarity.
It was trained with Multiple Negatives Ranking Loss (MNRL) on a Twitter dataset.
It was initialized from [BERTweet](https://huggingface.co/vinai/bertweet-base) and trained with [Sentence-transformers](https://www.sbert.net/).
## Usage
The model is easier to use with sentence-trainsformers library
```
pip install -U sentence-transformers
```
```
from sentence_transformers import SentenceTransformer
sentences = ["This is the first tweet", "This is the second tweet"]
model = SentenceTransformer('digio/Twitter4SSE')
embeddings = model.encode(sentences)
print(embeddings)
```
Without sentence-transfomer library, please refer to [this repository](https://huggingface.co/sentence-transformers) for detailed instructions on how to use Sentence Transformers on Huggingface.
## Citing & Authors
The official paper [Exploiting Twitter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings](https://arxiv.org/abs/2110.02030) will be presented at EMNLP 2021. Further details will be available soon.
```
@inproceedings{di-giovanni-brambilla-2021-exploiting,
title = "Exploiting {T}witter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings",
author = "Di Giovanni, Marco and
Brambilla, Marco",
booktitle = "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2021",
address = "Online and Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.emnlp-main.780",
pages = "9902--9910",
}
```
The official code is available on [GitHub](https://github.com/marco-digio/Twitter4SSE)
|
{"language": ["en"], "license": "apache-2.0", "tags": ["Pytorch", "Sentence Transformers", "Transformers"], "pipeline_tag": "sentence-similarity"}
|
sentence-similarity
|
digio/Twitter4SSE
|
[
"transformers",
"pytorch",
"roberta",
"feature-extraction",
"Pytorch",
"Sentence Transformers",
"Transformers",
"sentence-similarity",
"en",
"arxiv:2110.02030",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2110.02030"
] |
[
"en"
] |
TAGS
#transformers #pytorch #roberta #feature-extraction #Pytorch #Sentence Transformers #Transformers #sentence-similarity #en #arxiv-2110.02030 #license-apache-2.0 #endpoints_compatible #region-us
|
# Twitter4SSE
This model maps texts to 768 dimensional dense embeddings that encode semantic similarity.
It was trained with Multiple Negatives Ranking Loss (MNRL) on a Twitter dataset.
It was initialized from BERTweet and trained with Sentence-transformers.
## Usage
The model is easier to use with sentence-trainsformers library
Without sentence-transfomer library, please refer to this repository for detailed instructions on how to use Sentence Transformers on Huggingface.
## Citing & Authors
The official paper Exploiting Twitter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings will be presented at EMNLP 2021. Further details will be available soon.
The official code is available on GitHub
|
[
"# Twitter4SSE\n\nThis model maps texts to 768 dimensional dense embeddings that encode semantic similarity. \nIt was trained with Multiple Negatives Ranking Loss (MNRL) on a Twitter dataset. \nIt was initialized from BERTweet and trained with Sentence-transformers.",
"## Usage\n\nThe model is easier to use with sentence-trainsformers library\n\n\n\n\n\n\nWithout sentence-transfomer library, please refer to this repository for detailed instructions on how to use Sentence Transformers on Huggingface.",
"## Citing & Authors\n\nThe official paper Exploiting Twitter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings will be presented at EMNLP 2021. Further details will be available soon. \n\n\n\nThe official code is available on GitHub"
] |
[
"TAGS\n#transformers #pytorch #roberta #feature-extraction #Pytorch #Sentence Transformers #Transformers #sentence-similarity #en #arxiv-2110.02030 #license-apache-2.0 #endpoints_compatible #region-us \n",
"# Twitter4SSE\n\nThis model maps texts to 768 dimensional dense embeddings that encode semantic similarity. \nIt was trained with Multiple Negatives Ranking Loss (MNRL) on a Twitter dataset. \nIt was initialized from BERTweet and trained with Sentence-transformers.",
"## Usage\n\nThe model is easier to use with sentence-trainsformers library\n\n\n\n\n\n\nWithout sentence-transfomer library, please refer to this repository for detailed instructions on how to use Sentence Transformers on Huggingface.",
"## Citing & Authors\n\nThe official paper Exploiting Twitter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings will be presented at EMNLP 2021. Further details will be available soon. \n\n\n\nThe official code is available on GitHub"
] |
[
72,
70,
51,
61
] |
[
"passage: TAGS\n#transformers #pytorch #roberta #feature-extraction #Pytorch #Sentence Transformers #Transformers #sentence-similarity #en #arxiv-2110.02030 #license-apache-2.0 #endpoints_compatible #region-us \n# Twitter4SSE\n\nThis model maps texts to 768 dimensional dense embeddings that encode semantic similarity. \nIt was trained with Multiple Negatives Ranking Loss (MNRL) on a Twitter dataset. \nIt was initialized from BERTweet and trained with Sentence-transformers.## Usage\n\nThe model is easier to use with sentence-trainsformers library\n\n\n\n\n\n\nWithout sentence-transfomer library, please refer to this repository for detailed instructions on how to use Sentence Transformers on Huggingface.## Citing & Authors\n\nThe official paper Exploiting Twitter as Source of Large Corpora of Weakly Similar Pairs for Semantic Sentence Embeddings will be presented at EMNLP 2021. Further details will be available soon. \n\n\n\nThe official code is available on GitHub"
] |
[
0.008478150703012943,
-0.042290762066841125,
-0.004393023904412985,
0.009472976438701153,
0.058122679591178894,
0.020847588777542114,
0.07105812430381775,
0.11150268465280533,
-0.07971373200416565,
0.04160982742905617,
0.03296057879924774,
0.09218712151050568,
0.025628797709941864,
-0.04775990545749664,
0.05235494300723076,
-0.334568589925766,
0.006366208661347628,
0.05491786450147629,
-0.023027298972010612,
0.05471854656934738,
0.16744345426559448,
-0.03820237144827843,
0.07345152646303177,
0.0661819651722908,
-0.03743600472807884,
0.07102365046739578,
-0.06349126249551773,
-0.03947700187563896,
0.09344188123941422,
0.015732990577816963,
0.041925761848688126,
0.028188403695821762,
-0.028767993673682213,
-0.15830978751182556,
0.0236753448843956,
0.03758334368467331,
0.00527972774580121,
0.0449470691382885,
0.021650278940796852,
-0.09845302999019623,
0.06793634593486786,
-0.17226488888263702,
0.09450472891330719,
0.060091931372880936,
-0.03645143285393715,
-0.16921214759349823,
-0.004367473069578409,
-0.0019978925120085478,
0.0019891676492989063,
0.028438465669751167,
-0.026829106733202934,
0.15305091440677643,
-0.11112532764673233,
0.032952725887298584,
0.1758602410554886,
-0.2501082718372345,
-0.006421730387955904,
-0.03317033126950264,
-0.007880575023591518,
0.02864457294344902,
-0.026511233299970627,
-0.01796857826411724,
0.04435998201370239,
0.053349386900663376,
0.05962427705526352,
0.01970795728266239,
-0.024060875177383423,
0.04175266996026039,
-0.12160675972700119,
0.05700957030057907,
0.14599081873893738,
-0.028357410803437233,
-0.0063812327571213245,
-0.1987624317407608,
-0.06943459808826447,
0.1210068091750145,
-0.023017842322587967,
-0.08714919537305832,
0.08242282271385193,
0.06879934668540955,
-0.036673691123723984,
-0.07088857889175415,
-0.13867394626140594,
0.05781881511211395,
-0.07724756002426147,
0.06583650410175323,
0.07327422499656677,
-0.021242177113890648,
-0.0646246150135994,
-0.016765572130680084,
-0.12782147526741028,
-0.10328346490859985,
-0.04102813079953194,
-0.09730782359838486,
0.03901407867670059,
0.05365247279405594,
-0.0972951352596283,
-0.14183922111988068,
0.054512858390808105,
0.0810457244515419,
0.036847252398729324,
0.016662951558828354,
0.06077433377504349,
0.05785951018333435,
0.007475422695279121,
0.10768956691026688,
-0.0648031085729599,
0.00593218207359314,
0.06116604432463646,
-0.030154665932059288,
0.07806416600942612,
0.016742516309022903,
-0.12588082253932953,
-0.04219202697277069,
0.0577390193939209,
0.022654732689261436,
-0.006423140875995159,
0.04308043047785759,
-0.06030011549592018,
-0.06590968370437622,
0.10445645451545715,
-0.052907440811395645,
-0.020054783672094345,
-0.024474823847413063,
-0.0636742115020752,
0.1966048926115036,
0.013166270218789577,
0.02183411456644535,
-0.07755221426486969,
0.05846012383699417,
-0.07149723172187805,
-0.02801457978785038,
-0.03285416215658188,
-0.14944350719451904,
0.02401771955192089,
-0.030569415539503098,
-0.012228661216795444,
-0.16716822981834412,
-0.21236197650432587,
0.027513379231095314,
-0.00941143836826086,
0.020839784294366837,
-0.009992200881242752,
-0.07063071429729462,
-0.01239736471325159,
0.008676460012793541,
-0.016799582168459892,
-0.08228053897619247,
-0.0029477563221007586,
-0.005855079740285873,
-0.024614641442894936,
0.0674203559756279,
-0.054995037615299225,
0.03447559103369713,
-0.18213984370231628,
-0.06007926166057587,
-0.13478808104991913,
0.0848735123872757,
-0.041095852851867676,
0.018755627796053886,
-0.05500170961022377,
-0.04997207224369049,
-0.091493621468544,
0.030314480885863304,
-0.004896089434623718,
0.09494291245937347,
-0.1665884405374527,
-0.13823522627353668,
0.14832936227321625,
-0.08263669908046722,
-0.06649048626422882,
0.19077275693416595,
-0.01387833058834076,
-0.0117476936429739,
0.15928584337234497,
0.169611394405365,
0.0778798907995224,
-0.057644277811050415,
-0.04269036278128624,
0.033187657594680786,
-0.08190462738275528,
0.12314664572477341,
0.03991527855396271,
-0.011227147653698921,
0.10238130390644073,
0.014492221176624298,
0.042941801249980927,
0.035246167331933975,
-0.022331558167934418,
0.00963333249092102,
-0.05717898905277252,
-0.02814529836177826,
0.06532829254865646,
-0.04738954082131386,
-0.006790288724005222,
-0.08012904226779938,
-0.13977770507335663,
0.12658542394638062,
0.09253884106874466,
-0.08169529587030411,
-0.004613189026713371,
-0.0752120241522789,
0.004002532921731472,
-0.03927689045667648,
-0.03383331000804901,
-0.09830905497074127,
-0.05244666337966919,
0.06480865925550461,
-0.01883699744939804,
0.12686900794506073,
0.010393400676548481,
0.06223997101187706,
0.025540893897414207,
-0.015061268582940102,
0.022364649921655655,
0.02218651957809925,
0.005498373415321112,
-0.03197672218084335,
-0.09416376054286957,
-0.001561742159537971,
-0.009576964192092419,
0.05179931968450546,
0.04224717989563942,
0.03741564601659775,
0.04516788199543953,
0.13898466527462006,
-0.007542332634329796,
0.03872258961200714,
0.016311418265104294,
-0.011036980897188187,
-0.021395167335867882,
-0.030455410480499268,
0.04813035950064659,
0.005176083650439978,
-0.12082039564847946,
0.22287802398204803,
-0.1924598067998886,
-0.07301748543977737,
0.14329589903354645,
0.010261209681630135,
0.03165107220411301,
-0.008172466419637203,
-0.06897835433483124,
0.019066615030169487,
-0.032963741570711136,
-0.12227675318717957,
0.3008379638195038,
-0.023253118619322777,
0.1490551084280014,
-0.15479807555675507,
-0.11584049463272095,
-0.020086579024791718,
-0.020290743559598923,
-0.03788848966360092,
0.10493529587984085,
-0.028495822101831436,
-0.16689041256904602,
0.08396780490875244,
0.053829729557037354,
-0.0097722839564085,
0.11715464293956757,
0.019314425066113472,
-0.055203408002853394,
0.008657439611852169,
-0.04325813800096512,
0.0009538919548504055,
-0.02051899954676628,
-0.03672819212079048,
-0.018998706713318825,
0.06757190078496933,
0.024836720898747444,
0.12105482071638107,
-0.0766734704375267,
0.03174047917127609,
0.08732426166534424,
-0.025247277691960335,
0.11290176212787628,
0.05278049036860466,
0.0717448890209198,
0.08334764838218689,
-0.010119548998773098,
0.043451230973005295,
-0.05655815824866295,
0.0072783417999744415,
-0.1255662590265274,
0.1420058161020279,
-0.13788579404354095,
-0.26308441162109375,
-0.09197928756475449,
0.1415766477584839,
0.012585590593516827,
-0.020593734458088875,
0.1118551567196846,
-0.13517338037490845,
-0.06519389152526855,
-0.08917611837387085,
0.08306161314249039,
-0.0734371542930603,
-0.017069563269615173,
-0.04100700095295906,
0.014722038991749287,
-0.023420508950948715,
-0.19813662767410278,
-0.043633997440338135,
-0.021239472553133965,
-0.04191550984978676,
0.012396994046866894,
-0.1315791755914688,
0.003886610735207796,
0.08155635744333267,
-0.04413815960288048,
-0.010414188727736473,
-0.05650204047560692,
0.09556913375854492,
0.010258463211357594,
0.09354653209447861,
0.2401382327079773,
0.04106631502509117,
0.083157017827034,
-0.05742380768060684,
-0.020675599575042725,
-0.019194496795535088,
0.06407928466796875,
0.03162427246570587,
0.01085819210857153,
-0.17310087382793427,
-0.1352849006652832,
-0.06426216661930084,
0.04432523623108864,
0.0671713575720787,
0.01408805139362812,
0.03722340986132622,
0.06452982127666473,
-0.13125000894069672,
-0.005280844867229462,
0.060002896934747696,
0.1214752197265625,
0.15663780272006989,
0.04273126646876335,
0.09909701347351074,
-0.053679272532463074,
-0.14092698693275452,
0.11725567281246185,
-0.09674903005361557,
0.1196313425898552,
-0.021203547716140747,
0.1494651734828949,
0.06941785663366318,
0.004939474165439606,
0.08548515290021896,
0.05638941004872322,
-0.09469673037528992,
0.045813243836164474,
-0.08042352646589279,
-0.0923524796962738,
0.019079212099313736,
0.07221006602048874,
0.08303491771221161,
0.0017153857043012977,
-0.04669705033302307,
-0.004982770420610905,
0.1284245401620865,
0.10951601713895798,
0.044277776032686234,
-0.25639569759368896,
-0.09680113941431046,
0.010542022064328194,
-0.04443995654582977,
0.02695428766310215,
0.05033428594470024,
0.0486900769174099,
-0.0672193318605423,
0.11279033124446869,
0.009452610276639462,
0.08674116432666779,
-0.014321395196020603,
0.05145430192351341,
-0.09047872573137283,
0.15246990323066711,
0.043894246220588684,
0.10194916278123856,
-0.17940033972263336,
0.0033514040987938643,
0.035289958119392395,
0.07950778305530548,
-0.07996447384357452,
-0.021080702543258667,
0.0857817605137825,
-0.09589097648859024,
0.016717087477445602,
-0.02765279822051525,
0.1201409325003624,
0.058113712817430496,
-0.1563628911972046,
0.052387453615665436,
0.10235966742038727,
-0.022609341889619827,
0.09009486436843872,
-0.07868582010269165,
-0.025133220478892326,
0.05173898860812187,
0.07579325139522552,
-0.11644582450389862,
-0.20392873883247375,
0.00422947620972991,
0.07434996217489243,
-0.03214350715279579,
-0.02093595638871193,
-0.05042029917240143,
-0.006975189317017794,
0.15222960710525513,
-0.08953919261693954,
-0.1407928168773651,
-0.09011021256446838,
-0.006219314411282539,
0.16398507356643677,
-0.036221157759428024,
-0.009961017407476902,
0.007770530879497528,
0.20730221271514893,
-0.09644263982772827,
-0.16065296530723572,
0.014038830995559692,
-0.0710955560207367,
-0.09977822750806808,
-0.011964390054345131,
0.10024374723434448,
0.02990131266415119,
0.07831183820962906,
0.05722500756382942,
0.027941958978772163,
-0.12152345478534698,
-0.08577421307563782,
-0.021035034209489822,
0.1563105583190918,
0.07152239233255386,
0.05448964610695839,
-0.15264101326465607,
0.03507086634635925,
-0.07370619475841522,
0.0401303768157959,
0.15234915912151337,
0.23701442778110504,
-0.06637279689311981,
0.09286095947027206,
0.15038694441318512,
-0.06720532476902008,
-0.15704184770584106,
-0.07484321296215057,
0.005079078488051891,
0.0611729770898819,
-0.02262832038104534,
-0.09724688529968262,
0.1065102219581604,
0.11867949366569519,
0.03251495212316513,
-0.12967267632484436,
-0.2648484408855438,
-0.11730677634477615,
0.05449515953660011,
-0.0606876015663147,
0.11359986662864685,
-0.08909119665622711,
0.006812752224504948,
-0.07598592340946198,
-0.03265837952494621,
0.2086416780948639,
-0.1721341460943222,
0.12552429735660553,
-0.03076886385679245,
0.027219166979193687,
0.00701288552954793,
-0.010461941361427307,
0.06896849721670151,
-0.028326118364930153,
0.06533285230398178,
0.022549744695425034,
0.026712359860539436,
0.12740540504455566,
-0.0017176142428070307,
0.19429150223731995,
-0.07190582901239395,
0.06931300461292267,
-0.10776420682668686,
-0.0692945346236229,
-0.06077749654650688,
0.09083724021911621,
0.015930738300085068,
-0.07206280529499054,
-0.10556580871343613,
0.0045368061400949955,
0.13292314112186432,
0.00026158979744650424,
-0.014184698462486267,
-0.029774319380521774,
0.029438648372888565,
-0.040320977568626404,
0.18695376813411713,
-0.08689746260643005,
-0.10849107801914215,
-0.006248049903661013,
-0.019926130771636963,
0.04489261284470558,
-0.06304492801427841,
0.017322435975074768,
0.10201162844896317,
-0.006652079988270998,
0.06357487291097641,
0.07124882936477661,
-0.08210179954767227,
-0.0013355985283851624,
0.022324593737721443,
-0.07466553896665573,
-0.04049412161111832,
-0.058977190405130386,
0.04163755849003792,
-0.005350008606910706,
-0.035660840570926666,
0.139754980802536,
-0.12341788411140442,
-0.08216864615678787,
0.02337225154042244,
0.0049081528559327126,
0.007747120223939419,
0.08641035854816437,
0.08580169826745987,
0.0004091047157999128,
-0.04249023646116257,
0.162286639213562,
0.0479508601129055,
-0.16836632788181305,
0.07384496182203293,
0.11451206356287003,
-0.09878557920455933,
-0.0017815126338973641,
-0.04881042242050171,
0.09612331539392471,
-0.094220370054245,
-0.02124069631099701,
-0.13653668761253357,
-0.05031454935669899,
0.024603847414255142,
0.18760085105895996,
0.030860427767038345,
0.04702823981642723,
-0.0715571939945221,
0.05808330327272415,
-0.12079865485429764,
0.06679429113864899,
0.07823094725608826,
0.04926954209804535,
-0.029050083830952644,
0.14700117707252502,
-0.01769503392279148,
0.001304118661209941,
-0.05383341386914253,
0.026330312713980675,
-0.03484155982732773,
-0.08388444036245346,
-0.10184501111507416,
0.03027810901403427,
-0.05131182447075844,
-0.05588775873184204,
0.023082606494426727,
0.0003719384840223938,
-0.03505690023303032,
0.037302110344171524,
0.0059592705219984055,
-0.03874463215470314,
-0.0270164143294096,
0.020045287907123566,
-0.0778268352150917,
0.0044295224361121655,
0.04879646375775337,
-0.06076887622475624,
0.0651654601097107,
0.027077525854110718,
-0.07941319048404694,
-0.04282703250646591,
-0.14970126748085022,
-0.13531717658042908,
0.04684484004974365,
0.07498032599687576,
0.010401381179690361,
-0.07875467836856842,
0.03694647178053856,
0.023320063948631287,
-0.013468964956700802,
-0.02948244847357273,
0.11506412923336029,
-0.0689002275466919,
0.016800744459033012,
-0.00600945670157671,
-0.00981832668185234,
-0.09703200310468674,
-0.07751476019620895,
0.018296824768185616,
0.05554155260324478,
0.12161525338888168,
-0.07265867292881012,
0.019122233614325523,
-0.14942970871925354,
0.0038074487820267677,
-0.04905876889824867,
-0.09033194929361343,
0.03454923629760742,
0.012112158350646496,
0.0395338200032711,
-0.007489667274057865,
0.23311226069927216,
-0.06131727248430252,
-0.03251471370458603,
0.06478823721408844,
0.032465241849422455,
-0.004601554945111275,
0.006557660177350044,
0.08548203110694885,
0.04136514291167259,
-0.009000145830214024,
-0.0958908423781395,
0.035367656499147415,
0.07348135858774185,
-0.054860927164554596,
0.09001956135034561,
0.12991809844970703,
0.06598690897226334,
0.09200777858495712,
0.04099034518003464,
0.04452728480100632,
-0.05814208462834358,
0.0020824437960982323,
0.07640741020441055,
0.02394477091729641,
0.040841810405254364,
0.14452794194221497,
0.23284892737865448,
-0.06272950768470764,
0.09292880445718765,
0.013433023355901241,
-0.03545662760734558,
-0.1237340047955513,
-0.03734573349356651,
-0.08039169758558273,
-0.10066467523574829,
-0.03397991135716438,
-0.16564799845218658,
-0.024260500445961952,
0.04789382219314575,
0.04837561026215553,
0.006496890913695097,
0.0319809690117836,
-0.04510561004281044,
-0.14420731365680695,
0.1327979564666748,
-0.055364981293678284,
0.04331414774060249,
0.09365292638540268,
-0.05191677436232567,
0.046291571110486984,
0.06328989565372467,
0.08779660612344742,
0.04572534188628197,
0.060280632227659225,
-0.032152753323316574,
-0.08082602918148041,
-0.08331413567066193,
0.004846102558076382,
-0.04077104851603508,
-0.01596633903682232,
-0.03511284664273262,
0.023846721276640892,
-0.07832156866788864,
0.022911211475729942,
0.16166622936725616,
-0.02003566361963749,
-0.170126810669899,
-0.07951299846172333,
0.1168278306722641,
0.05235925316810608,
0.04649193584918976,
0.05590696632862091,
-0.09933698922395706,
-0.056946761906147,
0.20135478675365448,
0.16117747128009796,
-0.03449210524559021,
-0.001972764730453491,
0.015050764195621014,
0.006127202417701483,
0.03944189473986626,
-0.011900357902050018,
0.04973456263542175,
0.27592211961746216,
-0.08024422824382782,
0.04925413057208061,
-0.03833196684718132,
0.022400131449103355,
-0.17800946533679962,
0.06652489304542542,
-0.016802646219730377,
-0.02906537801027298,
0.00735008530318737,
0.04938222095370293,
-0.026066910475492477,
-0.008512814529240131,
-0.11088063567876816,
-0.1502099186182022,
-0.10481620579957962,
-0.020484957844018936,
0.01617489941418171,
0.02623424306511879,
0.11678575724363327,
0.020751766860485077,
-0.045488595962524414,
0.024176446720957756,
-0.0031771326903253794,
-0.16163478791713715,
0.06721770018339157,
0.04061266407370567,
-0.0504312701523304,
0.031496718525886536,
-0.011628154665231705,
0.0767216831445694,
0.110519640147686,
-0.016623729839920998,
-0.029880395159125328,
0.004669971764087677,
0.018641207367181778,
-0.01696392148733139,
0.11074477434158325,
0.045299697667360306,
-0.0009359242976643145,
0.08061365783214569,
0.06526689976453781,
-0.2251419872045517,
0.046063680201768875,
0.15605951845645905,
-0.05064774304628372,
-0.03452042117714882,
0.06223537027835846,
-0.08633633702993393,
0.07313449680805206,
0.09845173358917236,
-0.05047391355037689,
-0.07263993471860886,
-0.053732555359601974,
-0.025578536093235016,
0.053258609026670456,
0.0848555862903595,
0.0025488692335784435,
-0.1393878161907196,
-0.039285317063331604,
0.08317187428474426,
0.06314078718423843,
-0.2671966850757599,
-0.03427208214998245,
-0.023453857749700546,
-0.03174387291073799,
-0.04944890737533569,
0.0829547718167305,
0.07431331276893616,
0.026077335700392723,
-0.043648164719343185,
-0.0592784620821476,
0.03502640873193741,
0.0667843222618103,
-0.13049566745758057,
-0.08067207783460617
] |
null | null |
transformers
|
# COVID-Twitter-BERT v2 MNLI
## Model description
This model provides a zero-shot classifier to be used in cases where it is not possible to finetune CT-BERT on a specific task, due to lack of labelled data.
The technique is based on [Yin et al.](https://arxiv.org/abs/1909.00161).
The article describes a very clever way of using pre-trained MNLI models as zero-shot sequence classifiers.
The model is already finetuned on 400'000 generaic logical tasks.
We can then use it as a zero-shot classifier by reformulating the classification task as a question.
Let's say we want to classify COVID-tweets as vaccine-related and not vaccine-related.
The typical way would be to collect a few hunder pre-annotated tweets and organise them in two classes.
Then you would finetune the model on this.
With the zero-shot mnli-classifier, you can instead reformulate your question as "This text is about vaccines", and use this directly on inference - without any training.
Find more info about the model on our [GitHub page](https://github.com/digitalepidemiologylab/covid-twitter-bert).
## Usage
Please note that how you formulate the question can give slightly different results.
Collecting a training set and finetuning on this, will most likely give you better accuracy.
The easiest way to try this out is by using the Hugging Face pipeline.
This uses the default Enlish template where it puts the text "This example is " in front of the text.
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model="digitalepidemiologylab/covid-twitter-bert-v2-mnli")
```
You can then use this pipeline to classify sequences into any of the class names you specify.
```python
sequence_to_classify = 'To stop the pandemic it is important that everyone turns up for their shots.'
candidate_labels = ['health', 'sport', 'vaccine','guns']
hypothesis_template = 'This example is {}.'
classifier(sequence_to_classify, candidate_labels, hypothesis_template=hypothesis_template, multi_class=True)
```
## Training procedure
The model is finetuned on the 400k large [MNLI-task](https://cims.nyu.edu/~sbowman/multinli/).
## References
```bibtex
@article{muller2020covid,
title={COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter},
author={M{\"u}ller, Martin and Salath{\'e}, Marcel and Kummervold, Per E},
journal={arXiv preprint arXiv:2005.07503},
year={2020}
}
```
or
```
Martin Müller, Marcel Salathé, and Per E. Kummervold.
COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter.
arXiv preprint arXiv:2005.07503 (2020).
```
|
{"language": ["en"], "license": "mit", "tags": ["Twitter", "COVID-19", "text-classification", "pytorch", "tensorflow", "bert"], "datasets": ["mnli"], "thumbnail": "https://raw.githubusercontent.com/digitalepidemiologylab/covid-twitter-bert/master/images/COVID-Twitter-BERT_small.png", "pipeline_tag": "zero-shot-classification", "widget": [{"text": "To stop the pandemic it is important that everyone turns up for their shots.", "candidate_labels": "health, sport, vaccine, guns"}]}
|
zero-shot-classification
|
digitalepidemiologylab/covid-twitter-bert-v2-mnli
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"Twitter",
"COVID-19",
"tensorflow",
"zero-shot-classification",
"en",
"dataset:mnli",
"arxiv:1909.00161",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1909.00161"
] |
[
"en"
] |
TAGS
#transformers #pytorch #jax #bert #text-classification #Twitter #COVID-19 #tensorflow #zero-shot-classification #en #dataset-mnli #arxiv-1909.00161 #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# COVID-Twitter-BERT v2 MNLI
## Model description
This model provides a zero-shot classifier to be used in cases where it is not possible to finetune CT-BERT on a specific task, due to lack of labelled data.
The technique is based on Yin et al..
The article describes a very clever way of using pre-trained MNLI models as zero-shot sequence classifiers.
The model is already finetuned on 400'000 generaic logical tasks.
We can then use it as a zero-shot classifier by reformulating the classification task as a question.
Let's say we want to classify COVID-tweets as vaccine-related and not vaccine-related.
The typical way would be to collect a few hunder pre-annotated tweets and organise them in two classes.
Then you would finetune the model on this.
With the zero-shot mnli-classifier, you can instead reformulate your question as "This text is about vaccines", and use this directly on inference - without any training.
Find more info about the model on our GitHub page.
## Usage
Please note that how you formulate the question can give slightly different results.
Collecting a training set and finetuning on this, will most likely give you better accuracy.
The easiest way to try this out is by using the Hugging Face pipeline.
This uses the default Enlish template where it puts the text "This example is " in front of the text.
You can then use this pipeline to classify sequences into any of the class names you specify.
## Training procedure
The model is finetuned on the 400k large MNLI-task.
## References
or
|
[
"# COVID-Twitter-BERT v2 MNLI",
"## Model description\nThis model provides a zero-shot classifier to be used in cases where it is not possible to finetune CT-BERT on a specific task, due to lack of labelled data.\n\nThe technique is based on Yin et al..\nThe article describes a very clever way of using pre-trained MNLI models as zero-shot sequence classifiers.\nThe model is already finetuned on 400'000 generaic logical tasks.\nWe can then use it as a zero-shot classifier by reformulating the classification task as a question.\n\nLet's say we want to classify COVID-tweets as vaccine-related and not vaccine-related.\nThe typical way would be to collect a few hunder pre-annotated tweets and organise them in two classes.\nThen you would finetune the model on this.\n\nWith the zero-shot mnli-classifier, you can instead reformulate your question as \"This text is about vaccines\", and use this directly on inference - without any training.\n\nFind more info about the model on our GitHub page.",
"## Usage\nPlease note that how you formulate the question can give slightly different results.\nCollecting a training set and finetuning on this, will most likely give you better accuracy.\n\nThe easiest way to try this out is by using the Hugging Face pipeline.\nThis uses the default Enlish template where it puts the text \"This example is \" in front of the text.\n\n\nYou can then use this pipeline to classify sequences into any of the class names you specify.",
"## Training procedure\nThe model is finetuned on the 400k large MNLI-task.",
"## References\n\nor"
] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #Twitter #COVID-19 #tensorflow #zero-shot-classification #en #dataset-mnli #arxiv-1909.00161 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# COVID-Twitter-BERT v2 MNLI",
"## Model description\nThis model provides a zero-shot classifier to be used in cases where it is not possible to finetune CT-BERT on a specific task, due to lack of labelled data.\n\nThe technique is based on Yin et al..\nThe article describes a very clever way of using pre-trained MNLI models as zero-shot sequence classifiers.\nThe model is already finetuned on 400'000 generaic logical tasks.\nWe can then use it as a zero-shot classifier by reformulating the classification task as a question.\n\nLet's say we want to classify COVID-tweets as vaccine-related and not vaccine-related.\nThe typical way would be to collect a few hunder pre-annotated tweets and organise them in two classes.\nThen you would finetune the model on this.\n\nWith the zero-shot mnli-classifier, you can instead reformulate your question as \"This text is about vaccines\", and use this directly on inference - without any training.\n\nFind more info about the model on our GitHub page.",
"## Usage\nPlease note that how you formulate the question can give slightly different results.\nCollecting a training set and finetuning on this, will most likely give you better accuracy.\n\nThe easiest way to try this out is by using the Hugging Face pipeline.\nThis uses the default Enlish template where it puts the text \"This example is \" in front of the text.\n\n\nYou can then use this pipeline to classify sequences into any of the class names you specify.",
"## Training procedure\nThe model is finetuned on the 400k large MNLI-task.",
"## References\n\nor"
] |
[
77,
13,
239,
107,
21,
4
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #Twitter #COVID-19 #tensorflow #zero-shot-classification #en #dataset-mnli #arxiv-1909.00161 #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# COVID-Twitter-BERT v2 MNLI## Model description\nThis model provides a zero-shot classifier to be used in cases where it is not possible to finetune CT-BERT on a specific task, due to lack of labelled data.\n\nThe technique is based on Yin et al..\nThe article describes a very clever way of using pre-trained MNLI models as zero-shot sequence classifiers.\nThe model is already finetuned on 400'000 generaic logical tasks.\nWe can then use it as a zero-shot classifier by reformulating the classification task as a question.\n\nLet's say we want to classify COVID-tweets as vaccine-related and not vaccine-related.\nThe typical way would be to collect a few hunder pre-annotated tweets and organise them in two classes.\nThen you would finetune the model on this.\n\nWith the zero-shot mnli-classifier, you can instead reformulate your question as \"This text is about vaccines\", and use this directly on inference - without any training.\n\nFind more info about the model on our GitHub page.## Usage\nPlease note that how you formulate the question can give slightly different results.\nCollecting a training set and finetuning on this, will most likely give you better accuracy.\n\nThe easiest way to try this out is by using the Hugging Face pipeline.\nThis uses the default Enlish template where it puts the text \"This example is \" in front of the text.\n\n\nYou can then use this pipeline to classify sequences into any of the class names you specify.## Training procedure\nThe model is finetuned on the 400k large MNLI-task.## References\n\nor"
] |
[
0.013733896426856518,
0.09244390577077866,
-0.00670159887522459,
0.008383787237107754,
0.12143030762672424,
0.06324376910924911,
0.15370948612689972,
0.09869381785392761,
-0.07643888890743256,
0.10087990760803223,
0.03121373802423477,
0.05676734074950218,
0.08183290809392929,
0.0943368449807167,
0.038590267300605774,
-0.27426955103874207,
0.022883949801325798,
-0.028706980869174004,
0.04290331155061722,
0.07525064051151276,
0.10328028351068497,
-0.08855579793453217,
0.04444039985537529,
0.002998941345140338,
0.0026298987213522196,
-0.01983901672065258,
0.0034292598720639944,
0.0031644997652620077,
0.08282530307769775,
0.02802198939025402,
0.13107500970363617,
0.025222770869731903,
0.09095396846532822,
-0.17586924135684967,
0.04003845155239105,
0.10498686134815216,
0.02234387770295143,
0.0863080769777298,
0.08817929774522781,
0.02543819323182106,
0.11926653236150742,
-0.16273429989814758,
0.042013391852378845,
0.07739388197660446,
-0.10725570470094681,
-0.07043562829494476,
-0.08581681549549103,
0.08117879182100296,
0.15430223941802979,
0.06157093122601509,
-0.026746049523353577,
0.09638344496488571,
0.05709383264183998,
0.05413274094462395,
0.11009936779737473,
-0.2048334926366806,
-0.058932848274707794,
0.09623999893665314,
0.018932867795228958,
0.1267542988061905,
-0.053136445581912994,
0.06626971065998077,
-0.04106580838561058,
-0.04197397455573082,
0.05737512186169624,
-0.02631644904613495,
-0.07011620700359344,
-0.03175332769751549,
-0.1363082230091095,
-0.024733318015933037,
0.09108319878578186,
-0.00788719393312931,
-0.034886594861745834,
-0.12063473463058472,
-0.008796711452305317,
-0.01534203439950943,
-0.03248349204659462,
-0.17013388872146606,
0.021365949884057045,
-0.013708344660699368,
0.07800735533237457,
-0.134121373295784,
-0.11142838001251221,
-0.0009611623827368021,
-0.034036971628665924,
-0.046863239258527756,
0.01777389831840992,
0.02386554703116417,
-0.049238722771406174,
0.037844184786081314,
-0.017201481387019157,
-0.014177916571497917,
-0.014434817247092724,
-0.008786124177277088,
-0.2211056798696518,
-0.11138090491294861,
-0.04407421499490738,
-0.16663719713687897,
0.04720749333500862,
0.06381077319383621,
-0.06101361662149429,
0.008228748105466366,
0.022721659392118454,
0.02232721447944641,
0.021578442305326462,
0.08574338257312775,
-0.12337931990623474,
0.011411006562411785,
0.011929989792406559,
-0.024088576436042786,
-0.017376607283949852,
-0.04136423021554947,
-0.07841099798679352,
-0.05573759227991104,
0.006686775479465723,
0.08379211276769638,
-0.03548409417271614,
0.050820935517549515,
-0.030857209116220474,
-0.09005535393953323,
0.041106414049863815,
-0.07340990751981735,
0.05451161786913872,
0.04643134027719498,
0.01932314597070217,
0.15089315176010132,
-0.007772862911224365,
-0.04278649762272835,
-0.07058458775281906,
-0.013872368261218071,
-0.08901727199554443,
-0.008385887369513512,
-0.0859074518084526,
-0.10672067850828171,
0.05121264606714249,
-0.06397203356027603,
-0.0855148434638977,
-0.1083691343665123,
0.049158886075019836,
-0.02239564061164856,
0.01607922464609146,
-0.026009900495409966,
-0.01567365601658821,
-0.05341361463069916,
0.0034847406204789877,
-0.016328081488609314,
0.035963159054517746,
0.04415779560804367,
-0.025251680985093117,
-0.02856716327369213,
-0.1996423900127411,
0.06755519658327103,
0.014603225514292717,
-0.042166173458099365,
-0.010786866769194603,
0.03691459819674492,
-0.09465671330690384,
0.0728258416056633,
-0.03433613479137421,
0.007655576802790165,
-0.17001931369304657,
-0.006587731186300516,
0.05150587856769562,
-0.09891284257173538,
0.0012297426583245397,
0.09238317608833313,
-0.06564170122146606,
-0.011546758003532887,
0.12334419041872025,
-0.04488801956176758,
-0.004176892340183258,
0.09098149091005325,
-0.0494358167052269,
0.023448044434189796,
0.14017003774642944,
0.10595263540744781,
0.18524950742721558,
-0.23663346469402313,
-0.00944754108786583,
0.022668391466140747,
-0.14425525069236755,
0.04947490245103836,
0.07083162665367126,
0.014552466571331024,
-0.07026686519384384,
-0.006486286874860525,
-0.041383128613233566,
-0.021863391622900963,
-0.06577706336975098,
-0.03020070120692253,
0.08778591454029083,
-0.04941307380795479,
-0.016776325181126595,
-0.004800469148904085,
-0.021920474246144295,
0.0223860926926136,
-0.07394075393676758,
-0.020622380077838898,
0.08694331347942352,
-0.058969881385564804,
0.030631665140390396,
-0.11963438987731934,
0.054773252457380295,
-0.02162768319249153,
0.008907251060009003,
-0.10854917764663696,
-0.24303662776947021,
0.036817390471696854,
-0.07836480438709259,
0.11807755380868912,
0.10429255664348602,
-0.0075334119610488415,
0.07861622422933578,
-0.06126222014427185,
0.08058078587055206,
0.04263899102807045,
-0.009850822389125824,
-0.1320112645626068,
-0.13342638313770294,
-0.08188014477491379,
-0.040327172726392746,
0.2630020081996918,
-0.04926709085702896,
0.03233329951763153,
0.0274182315915823,
0.05563608184456825,
0.03190914914011955,
-0.036157261580228806,
0.06353094428777695,
0.030919821932911873,
0.0700078159570694,
-0.03808088228106499,
0.06545088440179825,
-0.009594929404556751,
-0.1065826565027237,
0.00003998848842456937,
-0.1732417494058609,
-0.3030669391155243,
0.010481692850589752,
-0.03905750438570976,
-0.11214160919189453,
-0.10987977683544159,
-0.03010195679962635,
0.035766974091529846,
-0.07435831427574158,
-0.020325878635048866,
0.1441374123096466,
0.03930501267313957,
0.07035142183303833,
-0.04819232225418091,
-0.0647054985165596,
-0.024929940700531006,
-0.03657311946153641,
-0.050229061394929886,
0.056889891624450684,
0.07908464968204498,
-0.26132479310035706,
-0.019324764609336853,
-0.05149686709046364,
-0.030954694375395775,
0.10773435235023499,
0.0959542989730835,
-0.11945474147796631,
-0.05205267295241356,
-0.026593538001179695,
0.0878835991024971,
-0.028051383793354034,
0.03815920278429985,
0.0717567577958107,
0.07657795399427414,
-0.03913889452815056,
0.01415534969419241,
-0.04477018490433693,
0.0631440207362175,
0.014136701822280884,
-0.03364146873354912,
0.041669439524412155,
0.006142932455986738,
-0.0008582334266975522,
0.12164946645498276,
0.04087211936712265,
0.046661484986543655,
-0.016018830239772797,
-0.0055148103274405,
-0.11267206817865372,
0.11066734790802002,
-0.10796761512756348,
-0.24118320643901825,
-0.15984094142913818,
-0.05485854297876358,
-0.014449176378548145,
0.01395715493708849,
-0.0043305628933012486,
-0.07739987224340439,
-0.05026189982891083,
-0.06734838336706161,
-0.016376562416553497,
0.0029393124859780073,
-0.04814884811639786,
-0.08755005896091461,
0.04508494585752487,
0.037909135222435,
-0.06201593950390816,
0.030948910862207413,
0.0056932163424789906,
-0.12688086926937103,
0.07488016784191132,
-0.06138572469353676,
0.06224796175956726,
0.1947144716978073,
-0.04449894651770592,
0.007287828717380762,
-0.03493505343794823,
0.1982874572277069,
-0.09412349760532379,
0.1404809206724167,
0.09507206827402115,
0.02225508540868759,
0.04241803288459778,
0.0818047896027565,
0.023361986503005028,
-0.015247895382344723,
0.00016640746616758406,
0.10963129252195358,
-0.049391381442546844,
-0.24915963411331177,
-0.057428453117609024,
-0.06280115991830826,
-0.13496758043766022,
0.05923963710665703,
0.0016931976424530149,
0.11380447447299957,
0.10401274263858795,
-0.07525601238012314,
-0.0131378173828125,
0.06337140500545502,
0.11587321013212204,
0.06824847310781479,
0.05286913365125656,
0.06234654039144516,
-0.01646381989121437,
-0.055957503616809845,
0.08001543581485748,
0.016589734703302383,
0.12682852149009705,
0.02806117758154869,
0.17804498970508575,
0.05372369661927223,
0.003941098693758249,
0.05928686261177063,
0.04653298854827881,
0.050750333815813065,
0.04456568509340286,
-0.03987614065408707,
-0.0509018711745739,
0.005655972752720118,
0.012529348023235798,
0.043918199837207794,
-0.09966138005256653,
0.004258951172232628,
0.04203604534268379,
0.06247885897755623,
0.15861807763576508,
0.10226468741893768,
-0.19986571371555328,
-0.03412846848368645,
-0.0031457149889320135,
-0.052565015852451324,
-0.0839240550994873,
-0.038827210664749146,
0.10529787838459015,
-0.09394065290689468,
0.10506120324134827,
-0.019985105842351913,
0.13634729385375977,
-0.011564620770514011,
0.025611791759729385,
-0.003915255423635244,
0.05729585886001587,
-0.056458838284015656,
0.1472989320755005,
-0.10852119326591492,
0.05446439981460571,
0.03332476690411568,
0.07533315569162369,
-0.06425191462039948,
-0.03337198495864868,
0.018146490678191185,
0.1785535365343094,
0.13563646376132965,
0.023337334394454956,
-0.020257463678717613,
-0.07311820238828659,
-0.08845355361700058,
-0.026032177731394768,
0.0847388207912445,
-0.08082982152700424,
0.13769899308681488,
-0.009654603898525238,
-0.028807979077100754,
-0.09262390434741974,
0.07843869179487228,
-0.05503973737359047,
-0.18147006630897522,
0.0980241671204567,
0.0047889756970107555,
0.08482153713703156,
0.032652728259563446,
0.02897600457072258,
-0.05475509166717529,
0.11370225250720978,
-0.09903663396835327,
-0.09089211374521255,
-0.17474541068077087,
0.08884968608617783,
0.11482071131467819,
-0.07734966278076172,
-0.04102900251746178,
-0.006540937349200249,
0.17195720970630646,
-0.06718076020479202,
-0.09691999852657318,
-0.06495177000761032,
-0.020148267969489098,
-0.14375081658363342,
-0.011129172518849373,
0.109782874584198,
0.08526303619146347,
0.09331353008747101,
0.08627837151288986,
0.06210774555802345,
-0.0036899270489811897,
-0.06576555222272873,
0.054050836712121964,
0.04062996804714203,
0.11840305477380753,
0.0011488923337310553,
-0.06684444099664688,
-0.03194020688533783,
-0.19764858484268188,
0.038973867893218994,
0.11046164482831955,
0.22949029505252838,
-0.07814539968967438,
0.04297444224357605,
0.17772233486175537,
-0.14603795111179352,
-0.1516687572002411,
-0.10284776240587234,
0.10905122756958008,
0.001476426376029849,
-0.04486655816435814,
-0.21425896883010864,
0.07793030887842178,
0.13095910847187042,
0.01345638558268547,
-0.06387616693973541,
-0.19552670419216156,
-0.1100120022892952,
0.005552361719310284,
0.02204129472374916,
-0.03652658313512802,
-0.12249767035245895,
-0.004162960220128298,
-0.015980640426278114,
0.0934910699725151,
0.19975247979164124,
-0.11167678236961365,
0.09009397774934769,
0.00859842449426651,
0.038195621222257614,
0.06414668262004852,
-0.018018480390310287,
0.16752754151821136,
0.011854033917188644,
0.04339342191815376,
-0.016811897978186607,
-0.0584070049226284,
0.20315782725811005,
-0.04394284263253212,
0.030250919982790947,
0.09237625449895859,
-0.028172802180051804,
-0.07648077607154846,
-0.09680918604135513,
-0.04335285350680351,
0.030301934108138084,
-0.02844437211751938,
-0.059153031557798386,
-0.1163831576704979,
0.07136663794517517,
0.07656165957450867,
-0.021947236731648445,
0.05659773573279381,
-0.152737095952034,
0.07593458145856857,
0.07187829911708832,
0.19875551760196686,
0.060482773929834366,
0.02114052139222622,
0.0015867053298279643,
-0.047347575426101685,
0.08493432402610779,
0.053949013352394104,
0.017635762691497803,
0.08015706390142441,
-0.03944776579737663,
0.10642760246992111,
0.008564257994294167,
-0.07620923966169357,
0.009207000955939293,
0.03786511719226837,
-0.13814279437065125,
-0.16382691264152527,
-0.023037387058138847,
0.08904881030321121,
-0.11884574592113495,
-0.008156119845807552,
0.17372551560401917,
-0.002560296095907688,
-0.07696175575256348,
0.003355835797265172,
0.06718507409095764,
-0.038565654307603836,
0.0790187418460846,
0.0529622957110405,
0.04328130930662155,
-0.05274108424782753,
-0.039035871624946594,
0.14756202697753906,
-0.11603063344955444,
0.026116589084267616,
0.1550171971321106,
-0.12041357904672623,
-0.11845598369836807,
0.0171979833394289,
0.10673309117555618,
-0.00896365474909544,
-0.06615974009037018,
0.08800270408391953,
-0.04185740277171135,
0.019970400258898735,
0.07850850373506546,
-0.05840322747826576,
0.03521350398659706,
-0.040637221187353134,
-0.04669754207134247,
-0.1801253855228424,
0.07680999487638474,
-0.07500290125608444,
0.002529498189687729,
-0.004914098419249058,
0.2377534657716751,
-0.047492615878582,
-0.008000469766557217,
-0.043804142624139786,
-0.005023851525038481,
-0.08129613101482391,
-0.03644553944468498,
-0.028566237539052963,
0.05909600853919983,
-0.06276442110538483,
-0.028666457161307335,
0.037739403545856476,
-0.010270061902701855,
0.0003181681386195123,
-0.027457114309072495,
-0.04140927270054817,
-0.02352665737271309,
-0.027052687481045723,
0.06337650120258331,
-0.07770217955112457,
-0.01327413972467184,
0.040237247943878174,
-0.024061216041445732,
0.11679141223430634,
-0.012673829682171345,
-0.029930762946605682,
0.03826740011572838,
-0.08717053383588791,
0.017111150547862053,
-0.03458627685904503,
0.10459800064563751,
0.010174715891480446,
-0.1237519234418869,
0.017289839684963226,
-0.022635567933321,
-0.10265360027551651,
0.026469703763723373,
-0.00044761295430362225,
-0.03342406824231148,
0.023960772901773453,
-0.012970764189958572,
-0.03569892793893814,
-0.1340477615594864,
0.04909512773156166,
0.015099265612661839,
-0.02648244984447956,
0.0760352835059166,
-0.030472977086901665,
0.09425833821296692,
-0.10280683636665344,
-0.016870718449354172,
0.034943342208862305,
0.007421704009175301,
-0.03740403428673744,
-0.1293467879295349,
0.10886124521493912,
-0.032882194966077805,
-0.0003417553671170026,
-0.010075923055410385,
0.03998437523841858,
0.009541640058159828,
0.008316120132803917,
0.01695849746465683,
0.017482157796621323,
0.08626258373260498,
0.06844429671764374,
-0.011286655440926552,
0.06451170891523361,
-0.04039592668414116,
-0.051873452961444855,
0.03877038136124611,
0.17211484909057617,
0.05735893175005913,
0.05517450347542763,
-0.007013846188783646,
-0.056082166731357574,
-0.05045466497540474,
-0.02391888201236725,
0.08260428160429001,
0.00888585951179266,
-0.010804593563079834,
-0.009125459007918835,
-0.02644011378288269,
0.2125491201877594,
-0.18314526975154877,
0.13563016057014465,
0.002041924512013793,
-0.08430218696594238,
-0.10689318925142288,
-0.23136243224143982,
-0.061509788036346436,
0.01615031808614731,
-0.041214242577552795,
-0.13580222427845,
0.0003646411350928247,
0.037097662687301636,
0.04064483195543289,
0.0025886385701596737,
0.08612167835235596,
-0.09297642111778259,
-0.037542253732681274,
0.014123686589300632,
-0.004617894534021616,
0.10066771507263184,
-0.006634087767452002,
0.014103252440690994,
0.0844423845410347,
0.012474766001105309,
0.06906594336032867,
0.04319169372320175,
0.1607632040977478,
-0.0005929373437538743,
-0.0709897130727768,
-0.0945514589548111,
0.021318752318620682,
-0.03339952602982521,
-0.012190359644591808,
0.08638088405132294,
0.010233767330646515,
-0.058298323303461075,
-0.014665577560663223,
0.1186058521270752,
0.009871231392025948,
-0.11567744612693787,
-0.15307097136974335,
0.2045326679944992,
0.0023237233981490135,
0.0031123303342610598,
0.00908680446445942,
-0.03974854201078415,
0.05368684232234955,
0.20667071640491486,
0.12361452728509903,
-0.06353417038917542,
-0.039714690297842026,
0.020642327144742012,
0.02088143117725849,
0.07221274077892303,
0.12122445553541183,
-0.050989553332328796,
0.1649254858493805,
-0.08318907767534256,
0.14600305259227753,
-0.03561687096953392,
0.023524729534983635,
-0.11583153158426285,
0.04767587035894394,
0.026688361540436745,
-0.011859611608088017,
-0.0004508635029196739,
0.1251177340745926,
-0.08693144470453262,
-0.153893381357193,
0.07483407855033875,
0.04634973406791687,
-0.07660974562168121,
-0.007887249812483788,
-0.07441222667694092,
-0.044116370379924774,
0.0678224265575409,
-0.021044237539172173,
-0.005969290155917406,
0.18480274081230164,
0.008937947452068329,
-0.05675477162003517,
-0.03850998729467392,
0.09740384668111801,
-0.03661932796239853,
0.1352112740278244,
0.013760075904428959,
0.06402844190597534,
0.0885978564620018,
-0.12786826491355896,
-0.11450278759002686,
0.10225607454776764,
-0.005846566520631313,
-0.03157733380794525,
0.1294112503528595,
0.09516490250825882,
0.008390353992581367,
0.02832501381635666,
0.03325653076171875,
-0.11562381684780121,
0.011458936147391796,
-0.009766635484993458,
-0.004393401555716991,
0.022570010274648666,
0.10862117260694504,
-0.07515155524015427,
0.08823009580373764,
0.19645847380161285,
-0.029452426359057426,
0.02481316216289997,
-0.021456092596054077,
0.00965822208672762,
0.03776717931032181,
0.010631549172103405,
-0.05516081675887108,
-0.07417396456003189,
-0.032060641795396805,
-0.14143596589565277,
0.05864651873707771,
-0.1349381059408188,
0.012870509177446365,
-0.0107822110876441,
0.008508208207786083,
-0.04005075991153717,
0.07246498763561249,
0.021589495241642,
-0.052543941885232925,
-0.00751560740172863,
-0.04046626761555672,
0.017582843080163002,
0.08450009673833847,
-0.13665041327476501,
-0.0026093260385096073
] |
null | null |
transformers
|
# COVID-Twitter-BERT v2
## Model description
BERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. This model is identical to [covid-twitter-bert](https://huggingface.co/digitalepidemiologylab/covid-twitter-bert) - but trained on more data, resulting in higher downstream performance.
Find more info on our [GitHub page](https://github.com/digitalepidemiologylab/covid-twitter-bert).
## Intended uses & limitations
The model can e.g. be used in the `fill-mask` task (see below). You can also use the model without the MLM/NSP heads and train a classifier with it.
#### How to use
```python
from transformers import pipeline
import json
pipe = pipeline(task='fill-mask', model='digitalepidemiologylab/covid-twitter-bert-v2')
out = pipe(f"In places with a lot of people, it's a good idea to wear a {pipe.tokenizer.mask_token}")
print(json.dumps(out, indent=4))
[
{
"sequence": "[CLS] in places with a lot of people, it's a good idea to wear a mask [SEP]",
"score": 0.9998226761817932,
"token": 7308,
"token_str": "mask"
},
...
]
```
## Training procedure
This model was trained on 97M unique tweets (1.2B training examples) collected between January 12 and July 5, 2020 containing at least one of the keywords "wuhan", "ncov", "coronavirus", "covid", or "sars-cov-2". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.
## Eval results
The model was evaluated based on downstream Twitter text classification tasks from previous SemEval challenges.
### BibTeX entry and citation info
```bibtex
@article{muller2020covid,
title={COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter},
author={M{\"u}ller, Martin and Salath{\'e}, Marcel and Kummervold, Per E},
journal={arXiv preprint arXiv:2005.07503},
year={2020}
}
```
or
```Martin Müller, Marcel Salathé, and Per E. Kummervold.
COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter.
arXiv preprint arXiv:2005.07503 (2020).
```
|
{"language": "en", "license": "mit", "tags": ["Twitter", "COVID-19"], "thumbnail": "https://raw.githubusercontent.com/digitalepidemiologylab/covid-twitter-bert/master/images/COVID-Twitter-BERT_small.png"}
| null |
digitalepidemiologylab/covid-twitter-bert-v2
|
[
"transformers",
"pytorch",
"tf",
"jax",
"bert",
"Twitter",
"COVID-19",
"en",
"license:mit",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #tf #jax #bert #Twitter #COVID-19 #en #license-mit #endpoints_compatible #has_space #region-us
|
# COVID-Twitter-BERT v2
## Model description
BERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. This model is identical to covid-twitter-bert - but trained on more data, resulting in higher downstream performance.
Find more info on our GitHub page.
## Intended uses & limitations
The model can e.g. be used in the 'fill-mask' task (see below). You can also use the model without the MLM/NSP heads and train a classifier with it.
#### How to use
## Training procedure
This model was trained on 97M unique tweets (1.2B training examples) collected between January 12 and July 5, 2020 containing at least one of the keywords "wuhan", "ncov", "coronavirus", "covid", or "sars-cov-2". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.
## Eval results
The model was evaluated based on downstream Twitter text classification tasks from previous SemEval challenges.
### BibTeX entry and citation info
or
|
[
"# COVID-Twitter-BERT v2",
"## Model description\n\nBERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. This model is identical to covid-twitter-bert - but trained on more data, resulting in higher downstream performance.\n\nFind more info on our GitHub page.",
"## Intended uses & limitations\n\nThe model can e.g. be used in the 'fill-mask' task (see below). You can also use the model without the MLM/NSP heads and train a classifier with it.",
"#### How to use",
"## Training procedure\nThis model was trained on 97M unique tweets (1.2B training examples) collected between January 12 and July 5, 2020 containing at least one of the keywords \"wuhan\", \"ncov\", \"coronavirus\", \"covid\", or \"sars-cov-2\". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.",
"## Eval results\nThe model was evaluated based on downstream Twitter text classification tasks from previous SemEval challenges.",
"### BibTeX entry and citation info\n\n\n\nor"
] |
[
"TAGS\n#transformers #pytorch #tf #jax #bert #Twitter #COVID-19 #en #license-mit #endpoints_compatible #has_space #region-us \n",
"# COVID-Twitter-BERT v2",
"## Model description\n\nBERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. This model is identical to covid-twitter-bert - but trained on more data, resulting in higher downstream performance.\n\nFind more info on our GitHub page.",
"## Intended uses & limitations\n\nThe model can e.g. be used in the 'fill-mask' task (see below). You can also use the model without the MLM/NSP heads and train a classifier with it.",
"#### How to use",
"## Training procedure\nThis model was trained on 97M unique tweets (1.2B training examples) collected between January 12 and July 5, 2020 containing at least one of the keywords \"wuhan\", \"ncov\", \"coronavirus\", \"covid\", or \"sars-cov-2\". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.",
"## Eval results\nThe model was evaluated based on downstream Twitter text classification tasks from previous SemEval challenges.",
"### BibTeX entry and citation info\n\n\n\nor"
] |
[
46,
10,
67,
54,
5,
107,
26,
12
] |
[
"passage: TAGS\n#transformers #pytorch #tf #jax #bert #Twitter #COVID-19 #en #license-mit #endpoints_compatible #has_space #region-us \n# COVID-Twitter-BERT v2## Model description\n\nBERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. This model is identical to covid-twitter-bert - but trained on more data, resulting in higher downstream performance.\n\nFind more info on our GitHub page.## Intended uses & limitations\n\nThe model can e.g. be used in the 'fill-mask' task (see below). You can also use the model without the MLM/NSP heads and train a classifier with it.#### How to use## Training procedure\nThis model was trained on 97M unique tweets (1.2B training examples) collected between January 12 and July 5, 2020 containing at least one of the keywords \"wuhan\", \"ncov\", \"coronavirus\", \"covid\", or \"sars-cov-2\". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.## Eval results\nThe model was evaluated based on downstream Twitter text classification tasks from previous SemEval challenges.### BibTeX entry and citation info\n\n\n\nor"
] |
[
0.001053317915648222,
-0.002200333634391427,
-0.0031009328085929155,
0.007920284755527973,
0.030330320820212364,
0.02683950960636139,
0.15458643436431885,
0.09390321373939514,
0.029434965923428535,
0.06978664547204971,
0.10964024811983109,
-0.011172897182404995,
0.02462046965956688,
0.16623255610466003,
0.04859472066164017,
-0.17909522354602814,
0.02238307148218155,
-0.08253918588161469,
-0.06830915808677673,
0.11375504732131958,
0.09282059222459793,
-0.074486143887043,
0.07547148317098618,
0.01618446595966816,
-0.09113488346338272,
0.017247704789042473,
-0.015044189058244228,
-0.06043251231312752,
0.1118537187576294,
0.0639604926109314,
0.11204608529806137,
0.05612150579690933,
0.09110040962696075,
-0.1934879571199417,
0.028053775429725647,
0.08846144378185272,
-0.05918273702263832,
0.029182283207774162,
0.0013962170341983438,
-0.03332434967160225,
0.07106250524520874,
-0.05758845806121826,
0.10984344035387039,
0.04137226566672325,
-0.12335922569036484,
0.03952058404684067,
-0.13195328414440155,
0.03942573070526123,
0.11429180204868317,
0.10099995881319046,
-0.01940038800239563,
0.1447416990995407,
-0.07280238717794418,
0.04119638353586197,
0.13687454164028168,
-0.23232822120189667,
-0.04279198870062828,
0.08897506445646286,
0.07016334682703018,
0.049817223101854324,
-0.038111548870801926,
0.050560105592012405,
0.06413158029317856,
-0.0022769407369196415,
0.07894374430179596,
-0.016456760466098785,
0.04033172130584717,
-0.039032239466905594,
-0.14111819863319397,
-0.07366105914115906,
0.10570620000362396,
0.04846949875354767,
-0.05240804702043533,
-0.11056583374738693,
-0.033036667853593826,
-0.07618064433336258,
-0.020036401227116585,
-0.0725075975060463,
0.01994268037378788,
0.006223868578672409,
-0.007782325614243746,
-0.05380156636238098,
-0.162165567278862,
-0.033264607191085815,
-0.040585074573755264,
0.04404240474104881,
0.014636916108429432,
0.01063738577067852,
-0.07851419597864151,
0.06376703083515167,
-0.07925654947757721,
-0.08684451878070831,
-0.011413284577429295,
-0.016133703291416168,
-0.14131709933280945,
-0.05227386951446533,
-0.09698246419429779,
-0.11380644142627716,
0.033218372613191605,
0.13371050357818604,
-0.012642760761082172,
-0.027393849566578865,
0.0269616786390543,
0.018949564546346664,
0.06836771219968796,
0.046208519488573074,
-0.08802193403244019,
0.03659540042281151,
0.07705571502447128,
-0.013666383922100067,
-0.034078724682331085,
-0.044203195720911026,
-0.056010130792856216,
0.03589634969830513,
-0.01725657843053341,
0.06102238595485687,
0.0033220688346773386,
0.0593847930431366,
-0.01986854523420334,
-0.04732806608080864,
0.12081169337034225,
-0.08665785938501358,
-0.020217081531882286,
0.04111501947045326,
-0.014975456520915031,
-0.059174858033657074,
0.05005766823887825,
0.057330530136823654,
0.013368112966418266,
-0.012974578887224197,
-0.11516059190034866,
-0.04057997092604637,
-0.033566463738679886,
-0.12646134197711945,
0.06564197689294815,
-0.0775376483798027,
-0.05809866636991501,
-0.13804978132247925,
-0.1354983001947403,
-0.062087591737508774,
-0.008920632302761078,
0.009010056965053082,
-0.010857250541448593,
-0.06051022559404373,
-0.041995711624622345,
0.07243037968873978,
0.00581697141751647,
0.05377880856394768,
-0.03276367858052254,
0.031211435794830322,
-0.13725963234901428,
0.0295870378613472,
0.02495950646698475,
-0.0022890930995345116,
-0.09142420440912247,
0.0033363820984959602,
-0.0783664658665657,
0.0840500071644783,
-0.0643215999007225,
0.13465335965156555,
-0.12276583164930344,
0.0022778548300266266,
-0.024546891450881958,
-0.013729449361562729,
-0.0015144251519814134,
0.13125380873680115,
-0.1878775656223297,
-0.053783874958753586,
0.1765023171901703,
-0.08362355828285217,
-0.06116028130054474,
0.11306636780500412,
-0.05610167235136032,
0.05750755965709686,
0.08778614550828934,
0.10871497541666031,
0.08535508811473846,
-0.1515040397644043,
0.016697708517313004,
-0.052995242178440094,
-0.014056648127734661,
0.09848646074533463,
0.059358417987823486,
-0.013458284549415112,
0.05107775703072548,
-0.03789784386754036,
-0.0012881099246442318,
0.0069334679283201694,
-0.06882070004940033,
-0.060468029230833054,
0.022323615849018097,
-0.03191608190536499,
0.038074955344200134,
-0.045061223208904266,
0.03922935202717781,
-0.05484205111861229,
-0.133584663271904,
-0.06132487207651138,
0.09210364520549774,
-0.06866341084241867,
0.03303035721182823,
-0.11017930507659912,
-0.016899269074201584,
0.05215640366077423,
0.0009202006040140986,
-0.08420463651418686,
-0.12578506767749786,
0.05580046772956848,
0.03852306306362152,
0.11841743439435959,
0.08960040658712387,
0.020310912281274796,
0.04336058348417282,
0.0038718420546501875,
0.03182043507695198,
-0.05996967479586601,
-0.02639871835708618,
-0.106430783867836,
-0.16684800386428833,
-0.024523304775357246,
-0.07836394757032394,
0.192807137966156,
-0.13807006180286407,
0.02286107838153839,
0.048999153077602386,
0.10797347873449326,
0.003694135695695877,
-0.019295312464237213,
0.02669208124279976,
-0.0025399066507816315,
-0.009970199316740036,
-0.04160656780004501,
-0.017637059092521667,
-0.011159992776811123,
-0.09893739223480225,
0.11455647647380829,
-0.168206587433815,
-0.13141636550426483,
0.09227470308542252,
0.04359864443540573,
-0.098150834441185,
0.05604025721549988,
-0.05078199878334999,
0.00774596119299531,
-0.012404530309140682,
-0.040167372673749924,
0.1773352473974228,
0.01676330715417862,
0.1089339330792427,
-0.10268951207399368,
-0.10147688537836075,
0.0306658037006855,
-0.03520124405622482,
-0.031404249370098114,
0.027046557515859604,
0.028829308226704597,
-0.26246944069862366,
0.04320601373910904,
0.01377625111490488,
0.0157371424138546,
0.18563535809516907,
0.03725919499993324,
-0.07772588729858398,
-0.007398630958050489,
-0.03476980701088905,
0.010318956337869167,
0.018721571192145348,
-0.01126910001039505,
0.004823349416255951,
0.031357940286397934,
0.0076129380613565445,
0.0193492379039526,
-0.09890525788068771,
-0.01078911405056715,
0.029244517907500267,
-0.0580374076962471,
-0.052058227360248566,
0.02362678200006485,
-0.016710596159100533,
0.08857082575559616,
0.019626300781965256,
0.0027952941600233316,
-0.028950165957212448,
-0.046542152762413025,
-0.169207364320755,
0.1936953216791153,
-0.16235999763011932,
-0.31392744183540344,
-0.07895305007696152,
0.0561952143907547,
0.03656189888715744,
0.03523711487650871,
0.0842941626906395,
-0.10603330284357071,
-0.04980805516242981,
-0.09119170904159546,
0.036359965801239014,
-0.033446941524744034,
0.02260945737361908,
-0.007602309808135033,
0.006531930062919855,
-0.013460561633110046,
-0.11863770335912704,
0.018747398629784584,
-0.04891537129878998,
-0.13731233775615692,
0.061010874807834625,
-0.05466507002711296,
0.03504429757595062,
0.15672701597213745,
0.007810324430465698,
-0.012507001869380474,
-0.06916207075119019,
0.1841881275177002,
-0.10212951898574829,
0.028658777475357056,
0.06270207464694977,
0.02224375121295452,
0.019652925431728363,
0.04645461589097977,
0.0011944674188271165,
-0.09250440448522568,
0.050934601575136185,
0.0928829237818718,
-0.0368787944316864,
-0.2599617540836334,
-0.07859300822019577,
-0.006746040191501379,
-0.0632038339972496,
0.09128129482269287,
0.035101186484098434,
0.16802819073200226,
0.01622856967151165,
-0.09527434408664703,
-0.005797137040644884,
0.04370224103331566,
0.05645851418375969,
0.021059807389974594,
0.042212601751089096,
0.08333925902843475,
-0.022080423310399055,
0.03494026139378548,
0.08584383130073547,
-0.03386082500219345,
0.1781310886144638,
0.08291947096586227,
0.1552954614162445,
0.10987470299005508,
0.03684798255562782,
0.04358959197998047,
-0.0028648111037909985,
0.01202229131013155,
-0.016277780756354332,
-0.03841840848326683,
-0.06892572343349457,
-0.039162527769804,
0.08998984098434448,
0.06846198439598083,
-0.09885038435459137,
-0.07254131883382797,
0.00600136024877429,
0.05311747267842293,
0.26053354144096375,
0.05327441915869713,
-0.20256157219409943,
-0.07858218997716904,
-0.020466800779104233,
-0.159463033080101,
0.018083231523633003,
-0.009891591034829617,
0.12311369925737381,
-0.11959255486726761,
0.0788775309920311,
0.01736193522810936,
0.09079208970069885,
0.045005396008491516,
0.0007545499829575419,
-0.021222086623311043,
-0.03007611818611622,
-0.030445298179984093,
0.07373505085706711,
-0.19630660116672516,
0.23822703957557678,
0.006127347704023123,
0.042626164853572845,
-0.05349833518266678,
-0.009093327447772026,
-0.007359407842159271,
0.06254816800355911,
0.11839178949594498,
0.015978747978806496,
0.09845871478319168,
-0.04521692171692848,
-0.13879483938217163,
-0.001574261230416596,
0.07595960795879364,
-0.15879377722740173,
0.111860491335392,
0.03206110745668411,
-0.011276866309344769,
-0.0008385141263715923,
0.08807330578565598,
-0.14872556924819946,
-0.09561523795127869,
0.06687285006046295,
-0.06853906810283661,
-0.04318603500723839,
-0.04509630426764488,
-0.06655555963516235,
-0.09562674909830093,
0.12736420333385468,
-0.11927074939012527,
-0.021658344194293022,
-0.13403376936912537,
0.08823957294225693,
0.09095889329910278,
-0.05751022323966026,
0.04302390664815903,
0.027927612885832787,
0.11884085088968277,
-0.0985385999083519,
-0.11053577810525894,
0.047557227313518524,
-0.09411043673753738,
-0.16346263885498047,
-0.0694529116153717,
0.1642034947872162,
0.11562146246433258,
0.08210060000419617,
0.07391554117202759,
0.04086105152964592,
0.0020087878219783306,
-0.08159024268388748,
0.02168305404484272,
0.09549295157194138,
0.06958937644958496,
-0.03029266558587551,
-0.004453300964087248,
0.023856673389673233,
-0.12136680632829666,
-0.001817335607483983,
0.12381596118211746,
0.25151127576828003,
-0.07835409045219421,
0.1479891985654831,
0.05816470459103584,
-0.0961468368768692,
-0.1762365847826004,
-0.04454047977924347,
0.1262771040201187,
0.036693982779979706,
-0.07003741711378098,
-0.16160784661769867,
0.008572014980018139,
0.0852966159582138,
-0.013267525471746922,
-0.029504062607884407,
-0.23414620757102966,
-0.17223556339740753,
0.028926020488142967,
-0.02928626537322998,
0.07885101437568665,
-0.08791302144527435,
-0.02845027856528759,
-0.00981295108795166,
0.00009396243694936857,
0.2267751395702362,
-0.11427245289087296,
0.07908013463020325,
0.0510234534740448,
0.0510965958237648,
0.00045378171489574015,
-0.021009530872106552,
0.10408584028482437,
0.030611760914325714,
0.056829772889614105,
-0.02116851881146431,
0.019749192520976067,
0.16585080325603485,
-0.03778452053666115,
0.0498352125287056,
0.03819793462753296,
0.03682268410921097,
-0.14126667380332947,
-0.048439864069223404,
-0.057324428111314774,
0.07768096774816513,
-0.03839205950498581,
-0.04750892147421837,
-0.0558723583817482,
0.028828948736190796,
0.08247850090265274,
-0.013480215333402157,
0.08064331859350204,
-0.08220882713794708,
0.10072188824415207,
0.08843944221735,
0.1486140936613083,
-0.028412900865077972,
0.028361806645989418,
-0.016307564452290535,
-0.048271551728248596,
0.0867820754647255,
-0.17764858901500702,
0.013450231403112411,
0.05087144300341606,
0.03473050147294998,
0.07705968618392944,
-0.00491140503436327,
-0.09943141043186188,
0.015018559992313385,
0.08332415670156479,
-0.1665516197681427,
-0.0541159063577652,
0.007621524389833212,
0.027704162523150444,
-0.020586110651493073,
-0.015132537111639977,
0.20866750180721283,
-0.04919877275824547,
-0.05810777470469475,
-0.014191586524248123,
0.04523087292909622,
-0.03881464898586273,
0.13323095440864563,
-0.0007077427580952644,
0.05486716702580452,
-0.09727741777896881,
0.04580767825245857,
0.09804946929216385,
-0.050501033663749695,
0.07186548411846161,
0.10027392208576202,
-0.11648484319448471,
-0.0679059848189354,
-0.060613296926021576,
0.038118332624435425,
-0.08682969212532043,
0.004149456042796373,
0.009148898534476757,
0.0018227073596790433,
0.03832418844103813,
0.26842010021209717,
-0.005958690773695707,
0.0631386786699295,
-0.01612226665019989,
0.004539611283689737,
-0.029280073940753937,
0.005574042443186045,
-0.015077443793416023,
0.0037503906060010195,
-0.04850208759307861,
0.23034924268722534,
0.02108772285282612,
0.03215029835700989,
-0.04193532094359398,
-0.018857477232813835,
-0.025043167173862457,
-0.02546153962612152,
0.001539696124382317,
0.03527871146798134,
-0.029622618108987808,
-0.018926046788692474,
0.02111898548901081,
-0.06339952349662781,
-0.03812666982412338,
-0.007985809817910194,
-0.07230959832668304,
-0.02137557603418827,
-0.05533590167760849,
0.059058766812086105,
-0.09622562676668167,
-0.030390311032533646,
0.04020120948553085,
-0.05742114782333374,
0.12076005339622498,
0.08806276321411133,
0.012666268274188042,
0.08099093288183212,
-0.14630500972270966,
0.012910516932606697,
0.02171330340206623,
0.051495350897312164,
0.017956482246518135,
-0.06424292922019958,
-0.0010779639706015587,
-0.02689257077872753,
-0.022498268634080887,
0.024743936955928802,
0.0022222173865884542,
-0.07932446897029877,
-0.042517926543951035,
0.05697423964738846,
-0.05554322525858879,
-0.07473543286323547,
-0.0221350509673357,
0.025908417999744415,
0.005217776168137789,
0.10184738039970398,
-0.05071866139769554,
0.0701557919383049,
-0.16550976037979126,
-0.017264019697904587,
-0.007743785623461008,
0.014594029635190964,
-0.08838246017694473,
-0.02452731877565384,
0.07360807061195374,
-0.05347578972578049,
0.11915990710258484,
-0.03132518753409386,
0.05195404216647148,
0.03422248736023903,
-0.0257679745554924,
0.08199134469032288,
0.0587092824280262,
0.13697679340839386,
0.07050499320030212,
0.024634461849927902,
-0.00638499716296792,
-0.030618976801633835,
0.04287389665842056,
-0.12046345323324203,
0.17792144417762756,
0.05741571635007858,
0.12014401704072952,
0.04995134845376015,
0.08325909078121185,
-0.08318953961133957,
-0.011696957983076572,
0.008273554965853691,
-0.0736311674118042,
0.01573863998055458,
-0.027224238961935043,
-0.028776351362466812,
0.21829287707805634,
-0.17460742592811584,
0.0400838665664196,
0.01268896833062172,
-0.038270432502031326,
-0.08126004040241241,
-0.13977910578250885,
-0.06998872756958008,
-0.10548326373100281,
0.03439028561115265,
-0.13554969429969788,
0.040960684418678284,
0.008478298783302307,
0.056103676557540894,
0.015181959606707096,
0.050346098840236664,
-0.023774847388267517,
-0.06848686188459396,
0.032013408839702606,
0.005069135222584009,
0.046434640884399414,
-0.030726447701454163,
-0.0586269348859787,
0.0786394476890564,
0.04175608605146408,
0.019149256870150566,
-0.000042874031350947917,
0.09646154195070267,
0.08250834792852402,
-0.05889881029725075,
-0.09899692237377167,
0.017678232863545418,
-0.010121596977114677,
0.00487487344071269,
0.053758878260850906,
0.09116838872432709,
-0.023767899721860886,
0.016574056819081306,
0.1841251105070114,
-0.00250624748878181,
-0.05525274947285652,
-0.1901063174009323,
0.11451899260282516,
0.02452392876148224,
0.05819287523627281,
0.010493720881640911,
-0.05475866049528122,
-0.0267622172832489,
0.16777710616588593,
0.19601401686668396,
-0.021545471623539925,
0.032715532928705215,
0.0010811935644596815,
-0.0006202994263730943,
0.02327575720846653,
0.06901779770851135,
0.016671646386384964,
0.1945907473564148,
-0.07256677001714706,
-0.04261600226163864,
0.007893318310379982,
-0.004411862697452307,
-0.07729830592870712,
0.029911989346146584,
-0.05367206409573555,
-0.017462849617004395,
-0.005374946165829897,
0.09132958948612213,
0.005873071029782295,
-0.16339579224586487,
-0.003910050727427006,
-0.09333160519599915,
-0.0740126222372055,
0.012136711739003658,
-0.055895742028951645,
0.013701972551643848,
0.08022945374250412,
0.028311055153608322,
-0.05399831756949425,
0.0653873160481453,
0.01263890415430069,
-0.08202190697193146,
-0.11284712702035904,
0.1288074404001236,
-0.04032808169722557,
0.15847469866275787,
-0.022731946781277657,
0.1517888754606247,
0.05930166319012642,
0.03356891870498657,
-0.03205060213804245,
0.019672922790050507,
0.057002849876880646,
-0.02610331028699875,
0.028515852987766266,
0.11107031255960464,
-0.0036709923297166824,
0.07950161397457123,
0.08176667988300323,
-0.15734894573688507,
0.044086046516895294,
-0.054322823882102966,
0.012620998546481133,
-0.033436182886362076,
0.10998906195163727,
-0.07753507792949677,
0.1292860060930252,
0.13954678177833557,
-0.06431323289871216,
0.003420586697757244,
-0.029067879542708397,
0.009687318466603756,
0.019021710380911827,
-0.038013044744729996,
-0.05242939293384552,
-0.25037047266960144,
0.02621123567223549,
-0.06177186220884323,
0.03237462416291237,
-0.13715486228466034,
-0.041834305971860886,
0.02986721880733967,
-0.005545438267290592,
-0.01645686663687229,
0.11122188717126846,
0.06654631346464157,
-0.014303842559456825,
-0.03025091253221035,
-0.0729551687836647,
0.05037793889641762,
0.09845827519893646,
-0.1355091780424118,
-0.0825423076748848
] |
null | null |
transformers
|
# COVID-Twitter-BERT (CT-BERT) v1
:warning: _You may want to use the [v2 model](https://huggingface.co/digitalepidemiologylab/covid-twitter-bert-v2) which was trained on more recent data and yields better performance_ :warning:
BERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. Find more info on our [GitHub page](https://github.com/digitalepidemiologylab/covid-twitter-bert).
## Overview
This model was trained on 160M tweets collected between January 12 and April 16, 2020 containing at least one of the keywords "wuhan", "ncov", "coronavirus", "covid", or "sars-cov-2". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.
This model was evaluated based on downstream classification tasks, but it could be used for any other NLP task which can leverage contextual embeddings.
In order to achieve best results, make sure to use the same text preprocessing as we did for pretraining. This involves replacing user mentions, urls and emojis. You can find a script on our projects [GitHub repo](https://github.com/digitalepidemiologylab/covid-twitter-bert).
## Example usage
```python
tokenizer = AutoTokenizer.from_pretrained("digitalepidemiologylab/covid-twitter-bert")
model = AutoModel.from_pretrained("digitalepidemiologylab/covid-twitter-bert")
```
You can also use the model with the `pipeline` interface:
```python
from transformers import pipeline
import json
pipe = pipeline(task='fill-mask', model='digitalepidemiologylab/covid-twitter-bert-v2')
out = pipe(f"In places with a lot of people, it's a good idea to wear a {pipe.tokenizer.mask_token}")
print(json.dumps(out, indent=4))
[
{
"sequence": "[CLS] in places with a lot of people, it's a good idea to wear a mask [SEP]",
"score": 0.9959408044815063,
"token": 7308,
"token_str": "mask"
},
...
]
```
## References
[1] Martin Müller, Marcel Salaté, Per E Kummervold. "COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter" arXiv preprint arXiv:2005.07503 (2020).
|
{"language": "en", "license": "mit", "tags": ["Twitter", "COVID-19"], "thumbnail": "https://raw.githubusercontent.com/digitalepidemiologylab/covid-twitter-bert/master/images/COVID-Twitter-BERT_small.png"}
| null |
digitalepidemiologylab/covid-twitter-bert
|
[
"transformers",
"pytorch",
"tf",
"jax",
"bert",
"Twitter",
"COVID-19",
"en",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #tf #jax #bert #Twitter #COVID-19 #en #license-mit #endpoints_compatible #region-us
|
# COVID-Twitter-BERT (CT-BERT) v1
:warning: _You may want to use the v2 model which was trained on more recent data and yields better performance_ :warning:
BERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. Find more info on our GitHub page.
## Overview
This model was trained on 160M tweets collected between January 12 and April 16, 2020 containing at least one of the keywords "wuhan", "ncov", "coronavirus", "covid", or "sars-cov-2". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.
This model was evaluated based on downstream classification tasks, but it could be used for any other NLP task which can leverage contextual embeddings.
In order to achieve best results, make sure to use the same text preprocessing as we did for pretraining. This involves replacing user mentions, urls and emojis. You can find a script on our projects GitHub repo.
## Example usage
You can also use the model with the 'pipeline' interface:
## References
[1] Martin Müller, Marcel Salaté, Per E Kummervold. "COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter" arXiv preprint arXiv:2005.07503 (2020).
|
[
"# COVID-Twitter-BERT (CT-BERT) v1\n\n:warning: _You may want to use the v2 model which was trained on more recent data and yields better performance_ :warning: \n\n\nBERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. Find more info on our GitHub page.",
"## Overview\nThis model was trained on 160M tweets collected between January 12 and April 16, 2020 containing at least one of the keywords \"wuhan\", \"ncov\", \"coronavirus\", \"covid\", or \"sars-cov-2\". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.\n\nThis model was evaluated based on downstream classification tasks, but it could be used for any other NLP task which can leverage contextual embeddings. \n\nIn order to achieve best results, make sure to use the same text preprocessing as we did for pretraining. This involves replacing user mentions, urls and emojis. You can find a script on our projects GitHub repo.",
"## Example usage\n\n\nYou can also use the model with the 'pipeline' interface:",
"## References\n[1] Martin Müller, Marcel Salaté, Per E Kummervold. \"COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter\" arXiv preprint arXiv:2005.07503 (2020)."
] |
[
"TAGS\n#transformers #pytorch #tf #jax #bert #Twitter #COVID-19 #en #license-mit #endpoints_compatible #region-us \n",
"# COVID-Twitter-BERT (CT-BERT) v1\n\n:warning: _You may want to use the v2 model which was trained on more recent data and yields better performance_ :warning: \n\n\nBERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. Find more info on our GitHub page.",
"## Overview\nThis model was trained on 160M tweets collected between January 12 and April 16, 2020 containing at least one of the keywords \"wuhan\", \"ncov\", \"coronavirus\", \"covid\", or \"sars-cov-2\". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.\n\nThis model was evaluated based on downstream classification tasks, but it could be used for any other NLP task which can leverage contextual embeddings. \n\nIn order to achieve best results, make sure to use the same text preprocessing as we did for pretraining. This involves replacing user mentions, urls and emojis. You can find a script on our projects GitHub repo.",
"## Example usage\n\n\nYou can also use the model with the 'pipeline' interface:",
"## References\n[1] Martin Müller, Marcel Salaté, Per E Kummervold. \"COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter\" arXiv preprint arXiv:2005.07503 (2020)."
] |
[
42,
85,
189,
19,
59
] |
[
"passage: TAGS\n#transformers #pytorch #tf #jax #bert #Twitter #COVID-19 #en #license-mit #endpoints_compatible #region-us \n# COVID-Twitter-BERT (CT-BERT) v1\n\n:warning: _You may want to use the v2 model which was trained on more recent data and yields better performance_ :warning: \n\n\nBERT-large-uncased model, pretrained on a corpus of messages from Twitter about COVID-19. Find more info on our GitHub page.## Overview\nThis model was trained on 160M tweets collected between January 12 and April 16, 2020 containing at least one of the keywords \"wuhan\", \"ncov\", \"coronavirus\", \"covid\", or \"sars-cov-2\". These tweets were filtered and preprocessed to reach a final sample of 22.5M tweets (containing 40.7M sentences and 633M tokens) which were used for training.\n\nThis model was evaluated based on downstream classification tasks, but it could be used for any other NLP task which can leverage contextual embeddings. \n\nIn order to achieve best results, make sure to use the same text preprocessing as we did for pretraining. This involves replacing user mentions, urls and emojis. You can find a script on our projects GitHub repo.## Example usage\n\n\nYou can also use the model with the 'pipeline' interface:## References\n[1] Martin Müller, Marcel Salaté, Per E Kummervold. \"COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter\" arXiv preprint arXiv:2005.07503 (2020)."
] |
[
0.0057116905227303505,
-0.0384536050260067,
-0.003273601643741131,
-0.0005250602844171226,
0.08476930111646652,
-0.004478954244405031,
0.0771685540676117,
0.10622347891330719,
-0.009891466237604618,
0.06181160360574722,
0.062456127256155014,
-0.04099640995264053,
0.053133148699998856,
0.06690546125173569,
0.0477617122232914,
-0.1979316771030426,
0.05538780614733696,
-0.052900154143571854,
0.015581578016281128,
0.12275471538305283,
0.09075195342302322,
-0.08668232709169388,
0.08822453767061234,
0.010803284123539925,
-0.09992966800928116,
-0.005262935068458319,
-0.02221868559718132,
-0.03634743392467499,
0.11716406792402267,
0.0582052506506443,
0.11968596279621124,
0.06422945857048035,
0.015715820714831352,
-0.16802990436553955,
0.026305222883820534,
0.11056847870349884,
-0.018950533121824265,
0.013165642507374287,
0.012650717049837112,
-0.0453302301466465,
0.042741384357213974,
-0.07939314842224121,
0.06870339065790176,
0.02200390212237835,
-0.1496279388666153,
0.06056048348546028,
-0.13405366241931915,
0.007902650162577629,
0.11224205046892166,
0.07051343470811844,
-0.022252444177865982,
0.15470458567142487,
-0.0782981812953949,
0.059926070272922516,
0.1767871081829071,
-0.26144805550575256,
-0.03297049179673195,
0.04190145432949066,
0.08234241604804993,
0.06368790566921234,
-0.04687973111867905,
0.038071028888225555,
0.015111411921679974,
-0.023241868242621422,
0.03551473096013069,
-0.04073559492826462,
-0.019456986337900162,
-0.039824049919843674,
-0.16561485826969147,
-0.05303841456770897,
0.12259740382432938,
0.008806928060948849,
-0.07444936782121658,
-0.07537907361984253,
-0.05793135240674019,
-0.061354730278253555,
-0.03349392116069794,
-0.08496971428394318,
0.004293153528124094,
0.01294414047151804,
0.0012631184654310346,
-0.060263071209192276,
-0.11877232789993286,
-0.03761851415038109,
-0.02289990335702896,
0.10831371694803238,
0.04996799677610397,
-0.009131966158747673,
-0.08028105646371841,
0.09544215351343155,
-0.009070201776921749,
-0.0898509994149208,
-0.05572667717933655,
-0.042340077459812164,
-0.14593994617462158,
-0.06288129091262817,
-0.06975225359201431,
-0.1323995441198349,
0.012349949218332767,
0.16614723205566406,
0.011162473820149899,
0.042426787316799164,
-0.0533469095826149,
0.039667144417762756,
0.009675279259681702,
0.12840300798416138,
-0.04781627282500267,
-0.013670864515006542,
0.05359359458088875,
0.009087434969842434,
-0.024998139590024948,
-0.024396905675530434,
-0.03549312427639961,
-0.0025147998239845037,
0.05976960062980652,
0.07710836082696915,
0.031197531148791313,
0.08223699778318405,
-0.04681980237364769,
-0.048110079020261765,
0.11546625196933746,
-0.1178949624300003,
-0.020145457237958908,
-0.001070211990736425,
-0.04141576960682869,
0.0002770474529825151,
0.04778198152780533,
0.014681385830044746,
-0.05410635843873024,
0.06773019582033157,
-0.08365815877914429,
-0.03173759579658508,
-0.070652075111866,
-0.12390851974487305,
0.058626942336559296,
-0.043032776564359665,
-0.06375416368246078,
-0.1146550104022026,
-0.14182165265083313,
-0.08651866763830185,
0.020758522674441338,
0.019368384033441544,
0.021534716710448265,
-0.04173646122217178,
-0.03306330740451813,
0.034475598484277725,
0.006625736597925425,
0.03454265370965004,
-0.027634503319859505,
0.060931745916604996,
-0.10274135321378708,
0.029392965137958527,
0.019476136192679405,
0.03286471590399742,
-0.13031508028507233,
-0.018596097826957703,
-0.16836974024772644,
0.06066596880555153,
-0.0616203173995018,
0.04877937585115433,
-0.1084643229842186,
0.0008060258696787059,
-0.07442199438810349,
0.012272357009351254,
0.007276307791471481,
0.1159096509218216,
-0.1768484264612198,
-0.06195370852947235,
0.21996960043907166,
-0.12792491912841797,
-0.037698958069086075,
0.13391345739364624,
-0.04528367146849632,
0.04369278997182846,
0.12894068658351898,
0.08176670223474503,
0.0985361635684967,
-0.12402277439832687,
0.020321791991591454,
-0.09047064185142517,
-0.056053418666124344,
0.10315373539924622,
0.047955892980098724,
0.002777224173769355,
-0.01944144256412983,
-0.016664667055010796,
-0.07344124466180801,
-0.04365364834666252,
-0.041518304497003555,
-0.01880885474383831,
0.05318314954638481,
0.016138028353452682,
-0.02406558394432068,
-0.03467463329434395,
-0.004514295607805252,
-0.04893150553107262,
-0.1443939208984375,
-0.04375062882900238,
0.059020910412073135,
-0.07544217258691788,
0.06140771135687828,
-0.10929035395383835,
-0.008976624347269535,
0.09609797596931458,
0.02274942211806774,
-0.11589338630437851,
-0.10933534801006317,
0.07969840615987778,
0.012769506312906742,
0.07964631170034409,
0.08264707773923874,
0.013474884442985058,
0.04804796725511551,
-0.016888339072465897,
0.009527277201414108,
-0.03612600266933441,
-0.06240379437804222,
-0.06120135635137558,
-0.11966551840305328,
0.004480451345443726,
-0.07566815614700317,
0.07332496345043182,
-0.026145288720726967,
0.0025694919750094414,
0.09411922842264175,
0.07562094926834106,
0.019717033952474594,
-0.03135010600090027,
-0.002718064235523343,
0.02172721177339554,
-0.0011043399572372437,
-0.04634552448987961,
0.012478378601372242,
0.021000821143388748,
-0.039495985954999924,
0.15070496499538422,
-0.21139712631702423,
-0.13879388570785522,
0.10343895107507706,
0.038246847689151764,
-0.08638032525777817,
0.022577181458473206,
-0.035225216299295425,
-0.007214334327727556,
-0.02785119228065014,
-0.04540422931313515,
0.19815494120121002,
0.04236998036503792,
0.09017779678106308,
-0.10711196064949036,
-0.09459841996431351,
0.00847555324435234,
-0.004128436092287302,
0.01482495665550232,
0.03334289789199829,
0.0517248772084713,
-0.2644154131412506,
0.04391172155737877,
0.01728828437626362,
0.040783654898405075,
0.18786309659481049,
0.026115145534276962,
-0.04837237298488617,
0.016524773091077805,
0.06372757256031036,
0.017910370603203773,
0.006955660413950682,
-0.0649670884013176,
-0.0281814094632864,
0.02023453451693058,
0.007179564796388149,
0.03002655692398548,
-0.08560009300708771,
0.032425880432128906,
0.02623223140835762,
-0.04484783485531807,
-0.011588383466005325,
0.011004773899912834,
-0.004305717535316944,
0.08697596937417984,
0.00792325846850872,
0.00959106720983982,
-0.02931354194879532,
-0.026783522218465805,
-0.1286192387342453,
0.18327023088932037,
-0.126663938164711,
-0.33025339245796204,
-0.10752125829458237,
0.012647701427340508,
0.016447298228740692,
0.01999483071267605,
0.09677880257368088,
-0.0765179917216301,
-0.053058695048093796,
-0.04396246001124382,
0.0764520913362503,
-0.01963162235915661,
-0.0011751479469239712,
-0.03785036876797676,
-0.02629316970705986,
-0.024465128779411316,
-0.13757434487342834,
-0.006414984352886677,
-0.032894909381866455,
-0.16875135898590088,
0.07955717295408249,
-0.05277959629893303,
0.06653724610805511,
0.08592376112937927,
-0.043452344834804535,
-0.0254629235714674,
-0.05300191789865494,
0.11596561968326569,
-0.09785262495279312,
0.06505723297595978,
0.08776875585317612,
-0.016604723408818245,
0.05566737428307533,
0.04275995120406151,
-0.03799041360616684,
-0.08922628313302994,
0.04620754346251488,
0.08647162467241287,
-0.03465208038687706,
-0.22085431218147278,
-0.07035955786705017,
-0.03324851021170616,
-0.030843958258628845,
0.10755477100610733,
0.04091610386967659,
0.015515021048486233,
0.007159119937568903,
-0.13546788692474365,
-0.010990304872393608,
0.0756773054599762,
0.09036855399608612,
-0.0048278323374688625,
0.018195467069745064,
0.13132312893867493,
0.00563120236620307,
-0.01903552934527397,
0.06824129074811935,
-0.042423952370882034,
0.18349730968475342,
0.06353224813938141,
0.09030628204345703,
0.10410180687904358,
0.017918214201927185,
0.07695750892162323,
0.03741508349776268,
-0.0073680514469742775,
0.011785040609538555,
-0.07897709310054779,
-0.05619353801012039,
-0.093020498752594,
0.037509381771087646,
-0.0007274711388163269,
-0.13469548523426056,
-0.06700233370065689,
0.05148719623684883,
0.060064319521188736,
0.2647758424282074,
0.006422634236514568,
-0.20864532887935638,
-0.042191702872514725,
-0.04221366345882416,
-0.13370367884635925,
-0.032386664301157,
0.004460802301764488,
0.08508080989122391,
-0.12753255665302277,
0.08013509213924408,
-0.014600149355828762,
0.09192687273025513,
-0.011696098372340202,
0.025468239560723305,
0.014476081356406212,
-0.038732774555683136,
-0.020119020715355873,
0.06652061641216278,
-0.2574102580547333,
0.1856265813112259,
0.023802904412150383,
0.07025261968374252,
-0.037883151322603226,
-0.008318147622048855,
-0.006680733524262905,
0.032698050141334534,
0.11515389382839203,
0.0270698145031929,
-0.00957471039146185,
-0.03913170099258423,
-0.11173761636018753,
0.03854186832904816,
0.11965697258710861,
-0.11543673276901245,
0.1408645212650299,
-0.006036970764398575,
0.0012885242467746139,
-0.017548738047480583,
0.04164336994290352,
-0.13312219083309174,
-0.11058606952428818,
0.04749849811196327,
-0.07344722747802734,
0.057067178189754486,
-0.03288673236966133,
-0.011930862441658974,
-0.04072175547480583,
0.12878350913524628,
-0.1260930299758911,
-0.046456024050712585,
-0.10412140935659409,
0.062119901180267334,
0.07495865225791931,
-0.07803390175104141,
0.01650361157953739,
0.015234311111271381,
0.08191319555044174,
-0.09268219769001007,
-0.10869259387254715,
0.033294327557086945,
-0.09372790902853012,
-0.1483885645866394,
-0.03531647101044655,
0.1577215939760208,
0.1273302584886551,
0.05937091261148453,
0.056913912296295166,
0.037522319704294205,
0.014574364759027958,
-0.09909660369157791,
0.006453255657106638,
0.07601221650838852,
0.07137928903102875,
-0.013051792047917843,
-0.08698247373104095,
0.018083583563566208,
-0.11703325808048248,
0.037079643458127975,
0.12104015052318573,
0.22610072791576385,
-0.004158667288720608,
0.12733608484268188,
0.1541229635477066,
-0.10098712146282196,
-0.23608991503715515,
-0.05499748885631561,
0.0787314623594284,
0.04008813202381134,
-0.09225954115390778,
-0.21267376840114594,
0.036214977502822876,
0.1032334640622139,
0.0072045763954520226,
-0.06392567604780197,
-0.1796114593744278,
-0.12383225560188293,
0.030859239399433136,
-0.019588086754083633,
0.21858982741832733,
-0.09447169303894043,
-0.002733242465183139,
0.020876619964838028,
-0.02309575118124485,
0.1987905651330948,
-0.06298727542161942,
0.09266997128725052,
0.018208879977464676,
0.009631730616092682,
0.010060474276542664,
-0.03213578462600708,
0.048994094133377075,
0.014950215816497803,
0.07424245029687881,
-0.018572457134723663,
0.04386027529835701,
0.10890717059373856,
-0.0365353524684906,
0.05856027454137802,
0.01041464600712061,
0.03171936795115471,
-0.09485966712236404,
-0.034270770847797394,
-0.06922522187232971,
0.10413240641355515,
-0.026401525363326073,
-0.02545800805091858,
-0.04093782231211662,
0.06984766572713852,
0.0710400640964508,
0.012489411979913712,
0.04801185429096222,
-0.07500214129686356,
0.04547615349292755,
0.09124240279197693,
0.10806886851787567,
-0.09570461511611938,
0.03069356270134449,
-0.008834588341414928,
-0.012190419249236584,
0.13552260398864746,
-0.05702318251132965,
0.06116248294711113,
0.07063654810190201,
0.017397096380591393,
0.08605888485908508,
0.025556866079568863,
-0.11971015483140945,
-0.014389397576451302,
0.08031337708234787,
-0.19235341250896454,
-0.08226239681243896,
-0.023528940975666046,
0.03390604630112648,
-0.01675071008503437,
-0.038688335567712784,
0.19599804282188416,
-0.043065134435892105,
-0.0521666519343853,
-0.0062767812050879,
0.05714820697903633,
0.014519596472382545,
0.1393553614616394,
-0.053718894720077515,
0.048224907368421555,
-0.09697888046503067,
0.08170699328184128,
0.08485095202922821,
-0.11636124551296234,
0.06486285477876663,
0.06955278664827347,
-0.14795179665088654,
-0.07140026241540909,
-0.054240938276052475,
0.05764131247997284,
-0.05358028784394264,
-0.014234409667551517,
-0.03291277214884758,
-0.0009273815085180104,
0.04030435532331467,
0.20403580367565155,
0.0055598425678908825,
0.07493655383586884,
0.016202958300709724,
0.0206069927662611,
-0.048758365213871,
0.013583284802734852,
0.05359496921300888,
0.03362513706088066,
-0.026805264875292778,
0.16679739952087402,
0.0012688819551840425,
-0.021137328818440437,
-0.031814198940992355,
-0.0014019657392054796,
-0.07524216920137405,
-0.026450159028172493,
-0.005741538014262915,
0.0022610013838857412,
-0.010666857473552227,
-0.025465453043580055,
0.013210296630859375,
-0.02298624813556671,
-0.02468913048505783,
0.008472037501633167,
-0.07480327039957047,
-0.02834528312087059,
-0.054914798587560654,
0.03547334671020508,
-0.0979873314499855,
0.005196233745664358,
0.08418869227170944,
-0.07572927325963974,
0.07883897423744202,
0.05874480679631233,
0.002943077590316534,
0.08430632948875427,
-0.06636317819356918,
-0.06121794134378433,
0.044931087642908096,
0.042415618896484375,
0.0010179410455748439,
-0.08646908402442932,
0.023556677624583244,
0.00908543448895216,
-0.002631118753924966,
0.015089645981788635,
0.08401352912187576,
-0.0986202284693718,
0.02494986169040203,
0.02285793237388134,
-0.0330759696662426,
-0.06898762285709381,
0.002103522652760148,
0.025166116654872894,
0.09448913484811783,
0.13327009975910187,
-0.06408524513244629,
0.03513919189572334,
-0.1280750334262848,
-0.003216497367247939,
0.013106237165629864,
-0.02694961428642273,
-0.05461619049310684,
-0.04174119606614113,
0.061881713569164276,
-0.032839495688676834,
0.051467038691043854,
-0.005238250829279423,
0.046393975615501404,
0.04341742768883705,
-0.024173859506845474,
-0.049972739070653915,
0.0038203357253223658,
0.06111879274249077,
0.05892346426844597,
-0.014793072827160358,
0.0035216391552239656,
-0.02529044635593891,
0.04822303354740143,
-0.14389817416667938,
0.2404060810804367,
0.1941143423318863,
0.0924711525440216,
0.039779264479875565,
0.027736550197005272,
-0.05979187786579132,
-0.03670907020568848,
0.005665268748998642,
-0.054861728101968765,
0.050129447132349014,
-0.002148578641936183,
0.026965174823999405,
0.19458095729351044,
-0.12355431169271469,
0.06936018913984299,
0.025803150609135628,
-0.032979581505060196,
-0.11465921252965927,
-0.13714338839054108,
-0.07286347448825836,
-0.07035766541957855,
0.006100197788327932,
-0.1243387758731842,
0.02978661097586155,
0.00211481936275959,
0.05457919463515282,
-0.005160381551831961,
0.0770164281129837,
-0.05071050301194191,
-0.1244327500462532,
0.068491131067276,
-0.0058463625609874725,
0.04873273894190788,
0.06715670228004456,
-0.06899283081293106,
0.08158183097839355,
0.09929721057415009,
0.06331869214773178,
0.03486127778887749,
0.09189553558826447,
0.06183923780918121,
-0.07369136810302734,
-0.09067030996084213,
0.0005969612393528223,
-0.0018264636164531112,
0.03458787873387337,
0.0579795278608799,
0.06767019629478455,
-0.01617574505507946,
0.028006674721837044,
0.19948194921016693,
0.000034054468414979056,
-0.08814197778701782,
-0.1465630680322647,
0.16239583492279053,
0.07893294095993042,
0.04980936273932457,
-0.0035053053870797157,
-0.059954483062028885,
-0.007565263658761978,
0.19162626564502716,
0.1310133934020996,
0.004844176582992077,
0.023975517600774765,
-0.018699219450354576,
0.022016748785972595,
0.04943788796663284,
0.08663667738437653,
0.011384849436581135,
0.1433759331703186,
-0.030764292925596237,
0.0028557907789945602,
-0.0044623673893511295,
-0.031457576900720596,
-0.06320507824420929,
0.048728927969932556,
-0.012013151310384274,
-0.03698527067899704,
-0.03342512249946594,
0.07383956015110016,
0.008955542929470539,
-0.20811091363430023,
-0.0688055083155632,
-0.0792233943939209,
-0.08921816200017929,
-0.004037638660520315,
0.0017315354198217392,
0.0057848677970469,
0.12070585787296295,
0.025576679036021233,
-0.042056579142808914,
0.052437666803598404,
0.03754960745573044,
-0.09650736302137375,
-0.11074764281511307,
0.10042162239551544,
0.033732227981090546,
0.12434180080890656,
-0.00040048197843134403,
0.09602700918912888,
0.05221623554825783,
0.019447799772024155,
-0.048648614436388016,
0.003516146680340171,
0.006880897097289562,
-0.046934161335229874,
0.02422730252146721,
0.128107488155365,
-0.009649734944105148,
0.08501577377319336,
0.0848066434264183,
-0.13502392172813416,
0.025888297706842422,
-0.09110109508037567,
0.00647362507879734,
-0.06999571621417999,
0.0974230021238327,
-0.0658704936504364,
0.1126491129398346,
0.1826256960630417,
-0.02640480175614357,
0.04980456084012985,
-0.059529196470975876,
0.009045997634530067,
0.031455062329769135,
0.045324910432100296,
-0.03318501263856888,
-0.2225748598575592,
-0.019952619448304176,
-0.05782364681363106,
0.03405320271849632,
-0.13140404224395752,
-0.008480546064674854,
0.0439908541738987,
0.0152466194704175,
-0.03710309788584709,
0.13523949682712555,
0.00936227384954691,
-0.013457262888550758,
-0.02583985961973667,
-0.11252506822347641,
0.054061006754636765,
0.09206274151802063,
-0.11749652773141861,
-0.06906920671463013
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilgpt2-finetuned-AdventureTime
This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 3.2450
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 10
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 279 | 3.3451 |
| 3.4534 | 2.0 | 558 | 3.2941 |
| 3.4534 | 3.0 | 837 | 3.2740 |
| 3.2435 | 4.0 | 1116 | 3.2617 |
| 3.2435 | 5.0 | 1395 | 3.2556 |
| 3.1729 | 6.0 | 1674 | 3.2490 |
| 3.1729 | 7.0 | 1953 | 3.2475 |
| 3.1262 | 8.0 | 2232 | 3.2467 |
| 3.0972 | 9.0 | 2511 | 3.2448 |
| 3.0972 | 10.0 | 2790 | 3.2450 |
### Framework versions
- Transformers 4.13.0
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "distilgpt2-finetuned-AT", "results": []}]}
|
text-generation
|
pyordii/distilgpt2-finetuned-AT
|
[
"transformers",
"pytorch",
"tensorboard",
"gpt2",
"text-generation",
"generated_from_trainer",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
distilgpt2-finetuned-AdventureTime
==================================
This model is a fine-tuned version of distilgpt2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 3.2450
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 10
### Training results
### Framework versions
* Transformers 4.13.0
* Pytorch 1.10.0+cu111
* Datasets 1.16.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
66,
98,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
-0.0880773738026619,
0.06721954047679901,
-0.002481053816154599,
0.10994107276201248,
0.1483883559703827,
0.024777667596936226,
0.13570328056812286,
0.13651597499847412,
-0.11047064512968063,
0.03047422505915165,
0.13568779826164246,
0.16312365233898163,
0.02278963476419449,
0.09506399184465408,
-0.03998401761054993,
-0.2585099935531616,
-0.0025181693490594625,
0.037908997386693954,
-0.0632745549082756,
0.13835670053958893,
0.08801809698343277,
-0.11332425475120544,
0.08913161605596542,
0.00223206402733922,
-0.18501168489456177,
0.009308788925409317,
-0.004703622777014971,
-0.053523533046245575,
0.1495048552751541,
0.030257785692811012,
0.0956549122929573,
0.004070716444402933,
0.06901873648166656,
-0.1975909322500229,
0.010428640991449356,
0.05976242199540138,
-0.00015945326595101506,
0.08780679851770401,
0.04598439112305641,
0.01512113492935896,
0.18719469010829926,
-0.06449035555124283,
0.03805006295442581,
0.022795306518673897,
-0.11998787522315979,
-0.2028462290763855,
-0.08708960562944412,
0.0315837636590004,
0.07366399466991425,
0.12015046924352646,
-0.012039648368954659,
0.12309546768665314,
-0.07939073443412781,
0.09603231400251389,
0.2418055236339569,
-0.2952643036842346,
-0.0633491724729538,
0.022023411467671394,
0.029540350660681725,
0.07327517122030258,
-0.09036225825548172,
-0.012323133647441864,
0.05005774646997452,
0.0455327108502388,
0.1132226213812828,
-0.029623644426465034,
-0.10630971938371658,
0.020680919289588928,
-0.1469389945268631,
-0.05561261251568794,
0.17453384399414062,
0.03177785128355026,
-0.028101658448576927,
-0.06237301975488663,
-0.07285032421350479,
-0.16396433115005493,
-0.023471813648939133,
-0.009500488638877869,
0.03905319795012474,
-0.014121095649898052,
-0.052597060799598694,
-0.03707548975944519,
-0.11252032965421677,
-0.07675103098154068,
-0.06706704199314117,
0.12507419288158417,
0.04157140105962753,
0.01044313982129097,
-0.03492259606719017,
0.11880507320165634,
0.0014314696891233325,
-0.12442942708730698,
0.01099279522895813,
0.036359068006277084,
0.028333161026239395,
-0.021384993568062782,
-0.06093120947480202,
-0.06640134751796722,
0.026356259360909462,
0.12589329481124878,
-0.056570470333099365,
0.04466880485415459,
0.03273826837539673,
0.04863787069916725,
-0.09685789048671722,
0.16988833248615265,
-0.03244437277317047,
-0.005585677456110716,
0.0072794207371771336,
0.04645978659391403,
0.039647169411182404,
-0.021534612402319908,
-0.12420115619897842,
0.008807037957012653,
0.08993647992610931,
0.023279894143342972,
-0.05560261756181717,
0.07556343823671341,
-0.0423799529671669,
-0.018042797222733498,
0.003267980646342039,
-0.09284055978059769,
0.02034504897892475,
0.0017380877397954464,
-0.07649386674165726,
-0.014545958489179611,
0.02948707900941372,
0.016274621710181236,
-0.038230281323194504,
0.09280770272016525,
-0.07925206422805786,
0.01926850713789463,
-0.09773364663124084,
-0.10370573401451111,
0.020579637959599495,
-0.08932246267795563,
0.01943995989859104,
-0.08761359006166458,
-0.20059043169021606,
-0.0088146748021245,
0.05527837574481964,
-0.025100337341427803,
-0.061866190284490585,
-0.065020352602005,
-0.07770325243473053,
0.017522187903523445,
-0.021938534453511238,
0.1454305201768875,
-0.06841759383678436,
0.1008739024400711,
0.03172927349805832,
0.058497607707977295,
-0.04712408408522606,
0.056484829634428024,
-0.09645649790763855,
0.00597966555505991,
-0.13642190396785736,
0.040000393986701965,
-0.04235734045505524,
0.05558255687355995,
-0.08206868171691895,
-0.09826122969388962,
-0.004293608944863081,
0.008866996504366398,
0.06712785363197327,
0.09765243530273438,
-0.1623804122209549,
-0.09118431806564331,
0.1758829951286316,
-0.07522431015968323,
-0.12873820960521698,
0.12513156235218048,
-0.04784444347023964,
0.047251634299755096,
0.06493793427944183,
0.17406120896339417,
0.06353510916233063,
-0.07895756512880325,
0.010169479064643383,
0.02015014924108982,
0.05227670818567276,
-0.05458933115005493,
0.07035377621650696,
-0.014560057781636715,
0.02635866589844227,
0.03267396613955498,
-0.027287252247333527,
0.05748080089688301,
-0.08580319583415985,
-0.0924234539270401,
-0.04900730401277542,
-0.0785800963640213,
0.020202670246362686,
0.06751562654972076,
0.07389193773269653,
-0.1062513142824173,
-0.09462219476699829,
0.029659774154424667,
0.0885607898235321,
-0.08045300096273422,
0.0397917740046978,
-0.0505170077085495,
0.07724032551050186,
-0.028270117938518524,
-0.003350895829498768,
-0.16970154643058777,
-0.015186755917966366,
0.011734441854059696,
-0.009696325287222862,
0.021935390308499336,
0.029590856283903122,
0.06650756299495697,
0.06372101604938507,
-0.04660286381840706,
-0.016529962420463562,
-0.03798636421561241,
-0.02016105130314827,
-0.1212138906121254,
-0.19742003083229065,
-0.02153170108795166,
-0.01814942993223667,
0.14446058869361877,
-0.20388856530189514,
0.04714972898364067,
0.00405798340216279,
0.054247818887233734,
0.0019625904969871044,
-0.012385780923068523,
-0.04339030385017395,
0.07159683853387833,
-0.056512411683797836,
-0.05045991763472557,
0.07638122886419296,
0.011588743887841702,
-0.09735243767499924,
-0.037242282181978226,
-0.12152809649705887,
0.14768832921981812,
0.14181901514530182,
-0.12207656353712082,
-0.05923744663596153,
-0.005498403217643499,
-0.06396671384572983,
-0.03365638479590416,
-0.036430004984140396,
0.015453724190592766,
0.19527898728847504,
-0.014067519456148148,
0.1515904664993286,
-0.07798758149147034,
-0.048530422151088715,
0.021413831040263176,
-0.03437196835875511,
0.030558370053768158,
0.13204525411128998,
0.13207052648067474,
-0.052530258893966675,
0.1520920991897583,
0.1459827870130539,
-0.0865435004234314,
0.1480483114719391,
-0.04732593894004822,
-0.07840587198734283,
-0.004418434575200081,
-0.0024585186038166285,
-0.0054054101929068565,
0.07984638959169388,
-0.1847608983516693,
-0.0055075171403586864,
0.026365412399172783,
0.027574991807341576,
0.040035709738731384,
-0.23283880949020386,
-0.033513836562633514,
0.033197563141584396,
-0.06434611976146698,
-0.002991199493408203,
-0.00842799711972475,
-0.012432366609573364,
0.10635607689619064,
-0.00425036670640111,
-0.07223543524742126,
0.03725019097328186,
0.002186358906328678,
-0.0846179649233818,
0.21427807211875916,
-0.07918987423181534,
-0.16373032331466675,
-0.13022568821907043,
-0.07080736011266708,
-0.06379714608192444,
0.012259842827916145,
0.08045491576194763,
-0.09585793316364288,
-0.027735568583011627,
-0.08589932322502136,
0.04036353901028633,
-0.0073958453722298145,
0.02472754940390587,
0.011036388576030731,
-0.001002775621600449,
0.05099682882428169,
-0.1211422011256218,
-0.011730419471859932,
-0.05772101506590843,
-0.0800892785191536,
0.04232463613152504,
0.033491555601358414,
0.11235502362251282,
0.15700343251228333,
0.0028145378455519676,
0.01562829315662384,
-0.03065849095582962,
0.21241135895252228,
-0.07256566733121872,
-0.03125075623393059,
0.15407218039035797,
0.006869493518024683,
0.05208814516663551,
0.08524010330438614,
0.06050243601202965,
-0.08892231434583664,
0.00966141652315855,
0.03310097008943558,
-0.04207807406783104,
-0.2342304140329361,
-0.04321780428290367,
-0.06159650534391403,
0.012934062629938126,
0.08618130534887314,
0.04002510383725166,
0.05313170328736305,
0.06747091561555862,
0.023882094770669937,
0.07876807451248169,
-0.023153988644480705,
0.05353083461523056,
0.12774714827537537,
0.03461401164531708,
0.12502729892730713,
-0.05415933206677437,
-0.05635513737797737,
0.05111260712146759,
-0.008819036185741425,
0.23271842300891876,
0.00553590664640069,
0.16729091107845306,
0.06536686420440674,
0.15588967502117157,
-0.013405803591012955,
0.07908759266138077,
-0.018072139471769333,
-0.03697417676448822,
-0.021408194676041603,
-0.04934987798333168,
-0.039214227348566055,
0.02394689805805683,
-0.08563373237848282,
0.054126203060150146,
-0.11371838301420212,
0.00621835608035326,
0.060301486402750015,
0.25407546758651733,
0.0352415032684803,
-0.30960631370544434,
-0.09012424945831299,
0.012213999405503273,
-0.035544104874134064,
-0.015560085885226727,
0.038458358496427536,
0.09878209978342056,
-0.08563797175884247,
0.031544458121061325,
-0.06861007958650589,
0.09827812761068344,
-0.052875567227602005,
0.056122682988643646,
0.06506354361772537,
0.0876242145895958,
0.011820999905467033,
0.09892984479665756,
-0.31178441643714905,
0.2710202634334564,
-0.001069138292223215,
0.07067237794399261,
-0.08676230907440186,
0.02073225937783718,
0.02990981750190258,
0.050117023289203644,
0.07298288494348526,
-0.018654121086001396,
-0.029903467744588852,
-0.1510688066482544,
-0.06543849408626556,
0.03791215270757675,
0.09124604612588882,
-0.04170724004507065,
0.09144902974367142,
-0.03459150716662407,
0.014955061487853527,
0.07178802043199539,
-0.003837689058855176,
-0.04794315621256828,
-0.11261775344610214,
0.0054018269293010235,
0.022643646225333214,
-0.04870454594492912,
-0.06386440247297287,
-0.10497875511646271,
-0.1317910999059677,
0.16561833024024963,
-0.03499472513794899,
-0.03886399790644646,
-0.10190483927726746,
0.07508213818073273,
0.06293053925037384,
-0.08724658191204071,
0.03752350062131882,
0.008903400041162968,
0.057477016001939774,
0.033398497849702835,
-0.07493248581886292,
0.10881399363279343,
-0.06701964139938354,
-0.1735670119524002,
-0.05059823393821716,
0.11529127508401871,
0.022098438814282417,
0.05752638354897499,
-0.019739888608455658,
0.008889471180737019,
-0.0590059831738472,
-0.1012042909860611,
0.025814056396484375,
-0.002493910025805235,
0.07167129963636398,
0.015209309756755829,
-0.05558501556515694,
0.03378799930214882,
-0.05737154185771942,
-0.05260571464896202,
0.20453542470932007,
0.23986977338790894,
-0.07922706007957458,
0.03390755504369736,
0.03573504835367203,
-0.0808170884847641,
-0.20328553020954132,
0.015069452114403248,
0.03806599974632263,
0.0005567653570324183,
0.024258097633719444,
-0.20376001298427582,
0.10710673034191132,
0.11840832233428955,
-0.015121452510356903,
0.13194535672664642,
-0.36976107954978943,
-0.1262890249490738,
0.11749401688575745,
0.13353176414966583,
0.11347426474094391,
-0.15734568238258362,
-0.029526952654123306,
-0.029339568689465523,
-0.11717288941144943,
0.10623060166835785,
-0.11594600975513458,
0.13074760138988495,
-0.03524385020136833,
0.08799482136964798,
-0.0014083429705351591,
-0.06192158907651901,
0.11903169751167297,
0.008172934874892235,
0.08802046626806259,
-0.06704835593700409,
0.002075049327686429,
0.042588114738464355,
-0.030676033347845078,
0.029213206842541695,
-0.12470985949039459,
0.028174391016364098,
-0.09958876669406891,
-0.024365928024053574,
-0.0658426359295845,
0.0471712201833725,
-0.03658607229590416,
-0.06614115834236145,
-0.041173286736011505,
-0.00718226982280612,
0.044583242386579514,
-0.002672005444765091,
0.1483892947435379,
0.016120105981826782,
0.1436677724123001,
0.10262014716863632,
0.0710638090968132,
-0.07148892432451248,
-0.05994664877653122,
-0.022114427760243416,
-0.007097023073583841,
0.05537465587258339,
-0.16273419559001923,
0.0173802450299263,
0.1341358870267868,
0.019982600584626198,
0.14271965622901917,
0.08645979315042496,
-0.03280584514141083,
0.02155795879662037,
0.05638185143470764,
-0.1772077977657318,
-0.09671194106340408,
-0.017868539318442345,
-0.06201285123825073,
-0.09432607889175415,
0.06776274740695953,
0.09957156330347061,
-0.06491094827651978,
-0.0009169363765977323,
-0.009659268893301487,
0.017797965556383133,
-0.057539861649274826,
0.18212643265724182,
0.038255661725997925,
0.04149818420410156,
-0.10211722552776337,
0.08401234447956085,
0.035407356917858124,
-0.08068311214447021,
0.022112371399998665,
0.09060977399349213,
-0.07612098008394241,
-0.0573372058570385,
0.06256118416786194,
0.1755073368549347,
-0.044372040778398514,
-0.0536624938249588,
-0.1373376101255417,
-0.12195629626512527,
0.08736347407102585,
0.1296200454235077,
0.1055588349699974,
0.01754048839211464,
-0.06788108497858047,
0.02082601934671402,
-0.11948658525943756,
0.08488790690898895,
0.0282460767775774,
0.05956088379025459,
-0.1279776245355606,
0.13730603456497192,
0.020091013982892036,
0.03541603684425354,
-0.019920993596315384,
0.021428674459457397,
-0.09499811381101608,
0.014359991997480392,
-0.10763200372457504,
-0.03546176478266716,
-0.02650049515068531,
-0.0016564041143283248,
-0.004121330101042986,
-0.04287480562925339,
-0.059692028909921646,
0.01633281633257866,
-0.11317036300897598,
-0.02698998898267746,
0.018484314903616905,
0.05840311199426651,
-0.1191389411687851,
-0.021122457459568977,
0.027053380385041237,
-0.0682118684053421,
0.08465197682380676,
0.05866822972893715,
0.01658729836344719,
0.06721175462007523,
-0.15279369056224823,
0.025804337114095688,
0.07380348443984985,
0.026858394965529442,
0.0448920838534832,
-0.07363273948431015,
-0.009619299322366714,
0.005181894637644291,
0.05274900794029236,
0.018796389922499657,
0.06359347701072693,
-0.13820071518421173,
0.008782895281910896,
-0.023530468344688416,
-0.08416289836168289,
-0.06649786233901978,
0.03793621063232422,
0.07293683290481567,
0.014065207913517952,
0.1955782175064087,
-0.07860016822814941,
0.037834566086530685,
-0.21134302020072937,
0.014759058132767677,
0.006208640523254871,
-0.11512698233127594,
-0.11784183979034424,
-0.06843942403793335,
0.06189578399062157,
-0.05949044227600098,
0.15079770982265472,
0.023533154278993607,
0.02243981324136257,
0.022984737530350685,
-0.014807166531682014,
0.031378258019685745,
0.009507243521511555,
0.2303258329629898,
0.04518374055624008,
-0.030590038746595383,
0.044295404106378555,
0.041035279631614685,
0.10929451882839203,
0.10261575132608414,
0.19665512442588806,
0.1313798576593399,
-0.009030896238982677,
0.10624391585588455,
0.03226947411894798,
-0.06691373139619827,
-0.15723946690559387,
0.04789312183856964,
-0.03939211368560791,
0.12081409245729446,
-0.024299006909132004,
0.2230386584997177,
0.08616826683282852,
-0.15320870280265808,
0.036181602627038956,
-0.045393262058496475,
-0.07885663211345673,
-0.12086557596921921,
-0.06537748873233795,
-0.07843311876058578,
-0.14990441501140594,
0.004977480508387089,
-0.1208685040473938,
0.028701573610305786,
0.11469618231058121,
0.012244458310306072,
-0.028065597638487816,
0.15786805748939514,
0.0342588908970356,
0.011021196842193604,
0.05075877532362938,
-0.0007829538662917912,
-0.03215200826525688,
-0.11854302138090134,
-0.06966667622327805,
-0.008875393308699131,
-0.01820935681462288,
0.04616894572973251,
-0.04390145465731621,
-0.053656939417123795,
0.04012430086731911,
-0.034420449286699295,
-0.08947792649269104,
0.00574977183714509,
0.01955651491880417,
0.05979696661233902,
0.053107600659132004,
0.007714729756116867,
0.0018789093010127544,
-0.002708258805796504,
0.21706320345401764,
-0.0829804539680481,
-0.07656896859407425,
-0.081258624792099,
0.2378648817539215,
0.03782102093100548,
-0.00924719125032425,
0.02605930157005787,
-0.060530178248882294,
-0.007644376717507839,
0.26330211758613586,
0.21902510523796082,
-0.07512294501066208,
-0.013847824186086655,
0.01610337197780609,
-0.002963225357234478,
-0.023300008848309517,
0.11096217483282089,
0.1518670618534088,
0.06411042809486389,
-0.08801361918449402,
-0.041207849979400635,
-0.043177422136068344,
-0.00735175795853138,
-0.0527542382478714,
0.08068674802780151,
0.041094567626714706,
-0.0031252061016857624,
-0.025402311235666275,
0.053616661578416824,
-0.06752337515354156,
-0.08313285559415817,
0.029738545417785645,
-0.202475443482399,
-0.1520841121673584,
-0.00010007106175180525,
0.1277635395526886,
0.0019385897321626544,
0.06553344428539276,
-0.024740226566791534,
-0.0022472271230071783,
0.07365064322948456,
-0.017110854387283325,
-0.11052222549915314,
-0.06408435106277466,
0.08679649978876114,
-0.13253124058246613,
0.20650430023670197,
-0.046447791159152985,
0.05903929844498634,
0.12577804923057556,
0.053778842091560364,
-0.07002244144678116,
0.07835637032985687,
0.04175198823213577,
-0.06480558961629868,
0.02514890395104885,
0.06962453573942184,
-0.028532952070236206,
0.04542061686515808,
0.05024990439414978,
-0.1119256541132927,
0.021496018394827843,
-0.04369859769940376,
-0.05290943384170532,
-0.03148996829986572,
-0.043046291917562485,
-0.06645958125591278,
0.12620776891708374,
0.20287065207958221,
-0.027119850739836693,
0.0023816474713385105,
-0.0799870640039444,
0.0123323118314147,
0.04503696784377098,
0.03703394904732704,
-0.05320591852068901,
-0.2245347946882248,
-0.00021253331215120852,
0.06666994094848633,
-0.018484804779291153,
-0.26498943567276,
-0.08419547975063324,
0.003956643398851156,
-0.06411979347467422,
-0.11133013665676117,
0.0642753317952156,
0.09738561511039734,
0.04669419676065445,
-0.05270764231681824,
-0.053097911179065704,
-0.07164514809846878,
0.15361325442790985,
-0.14247508347034454,
-0.09184836596250534
] |
null | null |
transformers
|
fBERT: A Neural Transformer for Identifying Offensive Content [Accepted at EMNLP 2021]
Authors: Diptanu Sarkar, Marcos Zampieri, Tharindu Ranasinghe and Alexander Ororbia
About:
Transformer-based models such as BERT, ELMO, and XLM-R have achieved state-of-the-art performance across various NLP tasks including the identification of offensive language and hate speech, an important problem in social media. Previous studies have shown that domain-specific fine-tuning or retraining of models before attempting to solve downstream tasks can lead to excellent results in multiple domains. Fine-tuning/retraining a complex models to identify offensive language has not been substantially explored before and we address this gap by proposing fBERT, a bert-base-uncased model that has been learned using over 1.4 million offensive instances from the SOLID dataset. The shifted fBERT model better incorporates domain-specific offensive language and social media features. The fBERT model achieves better results in both OffensEval and HatEval tasks and in the HS & O dataset over BERT and HateBERT.
|
{}
|
fill-mask
|
diptanu/fBERT
|
[
"transformers",
"pytorch",
"safetensors",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
|
fBERT: A Neural Transformer for Identifying Offensive Content [Accepted at EMNLP 2021]
Authors: Diptanu Sarkar, Marcos Zampieri, Tharindu Ranasinghe and Alexander Ororbia
About:
Transformer-based models such as BERT, ELMO, and XLM-R have achieved state-of-the-art performance across various NLP tasks including the identification of offensive language and hate speech, an important problem in social media. Previous studies have shown that domain-specific fine-tuning or retraining of models before attempting to solve downstream tasks can lead to excellent results in multiple domains. Fine-tuning/retraining a complex models to identify offensive language has not been substantially explored before and we address this gap by proposing fBERT, a bert-base-uncased model that has been learned using over 1.4 million offensive instances from the SOLID dataset. The shifted fBERT model better incorporates domain-specific offensive language and social media features. The fBERT model achieves better results in both OffensEval and HatEval tasks and in the HS & O dataset over BERT and HateBERT.
|
[] |
[
"TAGS\n#transformers #pytorch #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
41
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.07816528528928757,
-0.0034534591250121593,
-0.008949032984673977,
0.014659521169960499,
0.09390611946582794,
0.015563497319817543,
0.13519376516342163,
0.052756376564502716,
0.09398382157087326,
-0.003947537392377853,
0.15799419581890106,
0.17951816320419312,
-0.03753304108977318,
0.19498774409294128,
-0.06939197331666946,
-0.22714801132678986,
0.08130089193582535,
0.038432132452726364,
-0.0323895625770092,
0.12550091743469238,
0.06557557731866837,
-0.10157220810651779,
0.05762776359915733,
-0.02274511381983757,
-0.07466331124305725,
0.026036709547042847,
0.0754116028547287,
-0.12913183867931366,
0.1375875174999237,
-0.00015618337783962488,
0.2402632236480713,
0.03129007667303085,
-0.04585743322968483,
-0.08486726135015488,
0.05414804816246033,
0.00992332212626934,
-0.06514742225408554,
0.04708031564950943,
0.031702157109975815,
-0.0774080827832222,
-0.05167842283844948,
0.02480137161910534,
0.04714707285165787,
0.04453982785344124,
-0.13403639197349548,
-0.1468714475631714,
-0.011399785988032818,
0.042226310819387436,
0.0701330378651619,
0.06548841297626495,
0.016162805259227753,
0.23706258833408356,
-0.11051671952009201,
0.1071820929646492,
0.1379934698343277,
-0.3159155249595642,
-0.0092564532533288,
0.04371771216392517,
0.08167846500873566,
-0.04198794811964035,
-0.04232040047645569,
0.058274783194065094,
0.017485182732343674,
0.02084115333855152,
0.07533970475196838,
-0.06523330509662628,
-0.052030596882104874,
-0.014019805938005447,
-0.08491206914186478,
-0.04086682200431824,
0.12840859591960907,
-0.04972473159432411,
0.024205388501286507,
-0.023797903209924698,
-0.13894279301166534,
-0.0158387441188097,
-0.02011449821293354,
-0.022889534011483192,
-0.042050182819366455,
0.020780015736818314,
-0.02692507579922676,
0.009773931466042995,
-0.1258324682712555,
0.01016914937645197,
-0.21250614523887634,
0.2889426052570343,
0.026621220633387566,
0.0666622593998909,
-0.17185400426387787,
0.04806250333786011,
-0.002346624853089452,
-0.13750074803829193,
0.060302555561065674,
-0.10436061024665833,
0.04952726885676384,
-0.003486206755042076,
-0.031082691624760628,
-0.03199659660458565,
0.10086692869663239,
0.1850329488515854,
0.02225247025489807,
0.02934383600950241,
0.018503369763493538,
0.08943945914506912,
-0.004062177147716284,
0.07670725136995316,
0.03985711187124252,
-0.02041509374976158,
0.07545813173055649,
-0.06845609098672867,
0.056471988558769226,
-0.04729719087481499,
-0.07842159271240234,
-0.024386921897530556,
0.05095233768224716,
0.10312075912952423,
0.024323511868715286,
0.05622291937470436,
-0.09616982936859131,
0.027351612225174904,
0.12109542638063431,
-0.08567842096090317,
0.004566098563373089,
-0.0015342510305345058,
0.07859474420547485,
0.054182764142751694,
0.026392115280032158,
-0.017900872975587845,
-0.013051365502178669,
0.14371858537197113,
-0.07649422436952591,
-0.03931736201047897,
-0.0439886711537838,
-0.06643567979335785,
0.030963702127337456,
-0.0999702587723732,
0.031328197568655014,
-0.19763188064098358,
-0.13222025334835052,
0.06879133731126785,
0.04410157725214958,
0.03245622664690018,
-0.00717485137283802,
0.042118173092603683,
-0.004367187153548002,
-0.004952636081725359,
-0.04819079488515854,
-0.08310437947511673,
-0.04484144225716591,
0.10942672938108444,
0.009630645625293255,
0.09728159010410309,
-0.08897538483142853,
0.013966391794383526,
-0.11032228171825409,
0.02051958255469799,
-0.16286782920360565,
-0.06912025809288025,
-0.038973432034254074,
0.14928527176380157,
0.011287519708275795,
-0.04173709452152252,
-0.11329752951860428,
0.03191247582435608,
-0.0060578095726668835,
0.19525010883808136,
-0.06076255440711975,
-0.11373793333768845,
0.2521965503692627,
-0.1539997160434723,
-0.14852523803710938,
0.09112737327814102,
0.012607088312506676,
-0.028171751648187637,
0.07871609181165695,
0.0856693834066391,
0.035793643444776535,
-0.16292211413383484,
0.05914418026804924,
0.11381757259368896,
-0.13874274492263794,
-0.1010497435927391,
0.016276046633720398,
0.0012038468848913908,
-0.1338510811328888,
0.03891678526997566,
0.08447015285491943,
0.10191276669502258,
-0.08060655742883682,
-0.06530991941690445,
-0.028972206637263298,
-0.048178769648075104,
0.15067312121391296,
0.030575327575206757,
0.07322286814451218,
-0.09881700575351715,
-0.036793384701013565,
-0.0656207799911499,
-0.004797321744263172,
0.05540628358721733,
0.01656876504421234,
-0.10976744443178177,
0.11508894711732864,
-0.034759778529405594,
0.003569860477000475,
-0.14998817443847656,
-0.1561732441186905,
-0.0031190011650323868,
0.03335663303732872,
-0.05406753718852997,
0.059825800359249115,
0.1312398910522461,
0.015186438336968422,
-0.014653868041932583,
-0.05684158578515053,
0.09913294017314911,
0.03578618913888931,
-0.024538006633520126,
-0.11341983824968338,
0.02231493592262268,
-0.08396285027265549,
0.01643345318734646,
0.002635564422234893,
0.01521703414618969,
0.012640392407774925,
0.13880307972431183,
0.007268311455845833,
0.030243363231420517,
-0.05327451601624489,
0.02441750280559063,
-0.021887578070163727,
0.0036451518535614014,
0.07503856718540192,
-0.004050846677273512,
-0.06839505583047867,
0.13271784782409668,
-0.1594313532114029,
0.4097084105014801,
0.18631808459758759,
-0.2457636445760727,
-0.019304806366562843,
0.03353413939476013,
-0.023864803835749626,
0.006971759255975485,
0.027200855314731598,
-0.0070861950516700745,
-0.003165856236591935,
0.011063147336244583,
0.1358923763036728,
-0.025280814617872238,
-0.0347493439912796,
0.04055244103074074,
-0.07604624330997467,
-0.03977332264184952,
0.026339547708630562,
0.04833822324872017,
-0.11552173644304276,
0.17115719616413116,
0.2702193856239319,
0.003843375016003847,
0.1271134465932846,
-0.003685602219775319,
0.007235798053443432,
0.013982086442410946,
-0.015233723446726799,
-0.010479625314474106,
0.07386910915374756,
-0.15608292818069458,
-0.01303163543343544,
0.0639992356300354,
-0.037221457809209824,
0.030895082280039787,
-0.12762226164340973,
-0.04870174825191498,
0.009737269952893257,
0.04412790387868881,
-0.07124226540327072,
0.11084036529064178,
0.029176663607358932,
0.09184412658214569,
-0.03243904560804367,
-0.10614755749702454,
0.09511398524045944,
-0.0004769060469698161,
-0.0411682054400444,
0.17288793623447418,
-0.11459627002477646,
-0.3472522497177124,
-0.12403129786252975,
-0.1381843239068985,
0.01983197033405304,
0.04003489390015602,
0.06307733803987503,
-0.09641240537166595,
-0.07046008855104446,
0.06294085085391998,
-0.004332082346081734,
0.02463521435856819,
0.08679059147834778,
-0.04263911023736,
0.01280491054058075,
0.006028893869370222,
-0.07402495294809341,
-0.072076715528965,
-0.03510335460305214,
-0.05067513883113861,
0.15813276171684265,
-0.041926655918359756,
0.06939128041267395,
0.11932755261659622,
-0.0032013647723942995,
0.03155151382088661,
-0.009678704664111137,
0.17796435952186584,
-0.06485489755868912,
0.007008190732449293,
0.19201990962028503,
-0.08073795586824417,
0.10275401920080185,
0.164474755525589,
0.03770209103822708,
-0.04375359043478966,
0.013092790730297565,
-0.054381370544433594,
-0.10871856659650803,
-0.16908127069473267,
-0.1108776181936264,
-0.09051141887903214,
0.00006341520202113315,
0.03719956427812576,
0.05542363226413727,
0.12885865569114685,
0.1030946597456932,
0.013787783682346344,
-0.05949489399790764,
-0.02850298024713993,
0.04415624961256981,
0.12325833737850189,
-0.023623816668987274,
0.13493740558624268,
-0.03236249461770058,
-0.17208482325077057,
0.04752214252948761,
-0.011219991371035576,
0.10383465886116028,
0.09888515621423721,
-0.045927952975034714,
0.04894404485821724,
0.17719203233718872,
0.1608043611049652,
0.19527365267276764,
0.022157343104481697,
-0.07691473513841629,
0.012591512873768806,
-0.027182187885046005,
-0.0453190952539444,
0.002885372843593359,
0.06230199709534645,
-0.05202158913016319,
-0.0409371443092823,
-0.10408965498209,
0.07383637130260468,
0.11225643008947372,
0.05227046459913254,
-0.25053292512893677,
0.00805082730948925,
0.06519818305969238,
0.011346276849508286,
-0.07027877122163773,
0.04887617379426956,
0.004384794272482395,
-0.1003338098526001,
0.07912749797105789,
-0.06705032289028168,
0.06564326584339142,
0.06168520823121071,
0.08136583119630814,
-0.05721201002597809,
-0.05092751234769821,
0.01624378189444542,
0.050261545926332474,
-0.21829479932785034,
0.275526225566864,
-0.0032527432776987553,
-0.009576440788805485,
-0.0748133584856987,
-0.006965347100049257,
0.0523594468832016,
0.13853353261947632,
0.136390820145607,
0.02566063590347767,
-0.008791477419435978,
-0.15215250849723816,
-0.049861419945955276,
0.04824170470237732,
0.09624676406383514,
-0.013171962462365627,
-0.0011511859484016895,
-0.022991538047790527,
-0.0491827167570591,
0.005674763582646847,
0.04256151616573334,
-0.04830567166209221,
-0.10427409410476685,
0.05525512620806694,
0.06389287114143372,
0.026566626504063606,
-0.049906350672245026,
-0.0659952387213707,
-0.10819610953330994,
0.17058990895748138,
-0.06144404411315918,
-0.06265231966972351,
-0.10175763070583344,
-0.1348738968372345,
0.09254811704158783,
-0.09640489518642426,
0.11289630830287933,
-0.08690191060304642,
0.025875965133309364,
-0.09910192340612411,
-0.16520912945270538,
0.14343197643756866,
-0.15314523875713348,
-0.04348509758710861,
-0.08324509859085083,
0.17708152532577515,
-0.04512346163392067,
0.01845603436231613,
0.009801297448575497,
0.02631029300391674,
-0.07736480981111526,
-0.05170665681362152,
0.027976231649518013,
-0.06754571944475174,
0.04841380566358566,
0.04121483862400055,
-0.06658599525690079,
-0.11004399508237839,
0.010513459332287312,
0.043773893266916275,
0.19459222257137299,
0.2683841288089752,
-0.04181484505534172,
0.1194446012377739,
0.20940753817558289,
0.0024908313062042,
-0.3381671607494354,
-0.12442480027675629,
-0.14464108645915985,
-0.012225373648107052,
0.029868829995393753,
-0.07786780595779419,
0.12255319207906723,
-0.0002269869000883773,
-0.07161883264780045,
0.1312432885169983,
-0.14260341227054596,
-0.09481650590896606,
0.2584296762943268,
0.03425353020429611,
0.4517262279987335,
-0.11642501503229141,
-0.05472416803240776,
-0.025449048727750778,
-0.09361008554697037,
0.04465881362557411,
-0.04375568404793739,
0.05999315530061722,
-0.006484792102128267,
0.036720193922519684,
0.033627111464738846,
-0.0977986678481102,
0.10260207951068878,
-0.0697365254163742,
0.035547103732824326,
-0.09633149206638336,
-0.07658068090677261,
0.08520103245973587,
-0.008621939457952976,
-0.0038024201057851315,
-0.01303519494831562,
0.01886599510908127,
0.02335796318948269,
-0.029185011982917786,
-0.08458099514245987,
0.12927460670471191,
0.01940337009727955,
-0.07394016534090042,
0.06295346468687057,
-0.02490673027932644,
-0.032061949372291565,
-0.016706347465515137,
0.19304578006267548,
0.010956955142319202,
0.2183225005865097,
0.09297224879264832,
0.05599026009440422,
-0.14623427391052246,
-0.04548328369855881,
-0.033229436725378036,
-0.09346362948417664,
0.09183389693498611,
0.043771401047706604,
0.06259565055370331,
0.09475108981132507,
-0.011376706883311272,
0.04032605141401291,
0.10866794735193253,
0.015271146781742573,
-0.044820863753557205,
0.1635398417711258,
-0.2524127960205078,
0.01530796941369772,
-0.008825298398733139,
0.019874051213264465,
0.03978238254785538,
0.10727722197771072,
0.09699162095785141,
0.026957573369145393,
-0.0337185375392437,
-0.0228258203715086,
0.003355007153004408,
-0.06028057634830475,
0.07894162833690643,
0.05378841981291771,
0.07067114114761353,
-0.11743360012769699,
-0.0019006736110895872,
-0.03419315442442894,
-0.18085132539272308,
-0.030690383166074753,
0.06941135227680206,
-0.1279626488685608,
-0.10739822685718536,
0.013956199400126934,
0.09854517877101898,
-0.024650176987051964,
-0.05614452436566353,
-0.0717872753739357,
-0.14276602864265442,
0.0317828543484211,
0.248941108584404,
0.10656043887138367,
0.0869520828127861,
0.009593648836016655,
-0.0028164470568299294,
-0.013953848741948605,
0.01985281892120838,
0.018799668177962303,
0.03293578326702118,
-0.1120506301522255,
0.03082994371652603,
0.002131289802491665,
0.12218468636274338,
-0.11573954671621323,
-0.03597116470336914,
-0.16193319857120514,
0.03478673845529556,
-0.06130250170826912,
-0.07633049041032791,
-0.090923011302948,
-0.06352002173662186,
0.007958147674798965,
-0.07025150954723358,
-0.03501813858747482,
-0.03418946638703346,
-0.09442327171564102,
0.04239889606833458,
0.04444570094347,
-0.023785125464200974,
-0.09868498891592026,
-0.048443302512168884,
0.12180765718221664,
-0.05942642316222191,
0.08071069419384003,
0.13953346014022827,
-0.08569961041212082,
0.08437897264957428,
-0.16256192326545715,
-0.1106509417295456,
0.11697863787412643,
0.00983001571148634,
0.06513459235429764,
0.06659779697656631,
0.01821049302816391,
0.05881575495004654,
0.02310936152935028,
0.046750959008932114,
0.08334711194038391,
-0.11100320518016815,
0.09102891385555267,
0.027400964871048927,
-0.17757324874401093,
-0.018114374950528145,
-0.1251136213541031,
0.08241289854049683,
-0.0435904823243618,
0.14754371345043182,
-0.056687869131565094,
0.08906451612710953,
-0.07251285761594772,
0.02691369317471981,
-0.025249969214200974,
-0.16408419609069824,
-0.05379457399249077,
-0.020679732784628868,
0.012122363783419132,
-0.012767682783305645,
0.22410117089748383,
-0.025119585916399956,
0.011527463793754578,
0.048750851303339005,
0.05004515126347542,
0.004609106108546257,
0.012866910547018051,
0.15373602509498596,
0.07142792642116547,
-0.054325755685567856,
-0.0846821591258049,
0.05172991752624512,
0.017057090997695923,
-0.13172008097171783,
0.11902570724487305,
0.07706467062234879,
0.02896854840219021,
0.08839157223701477,
0.01713106408715248,
0.04225367307662964,
-0.12200840562582016,
-0.2539435625076294,
-0.06289327889680862,
0.029224639758467674,
0.042492568492889404,
0.01366657018661499,
0.1617201715707779,
0.008112289942800999,
0.030358854681253433,
-0.03328118473291397,
-0.016844524070620537,
-0.20096541941165924,
-0.10549116879701614,
-0.08237405121326447,
-0.04603077471256256,
0.02745545655488968,
-0.02689974568784237,
-0.0495859794318676,
0.08321097493171692,
0.03550597280263901,
-0.022074563428759575,
0.2013927698135376,
0.0103881461545825,
0.018461279571056366,
0.02097134105861187,
0.014586727134883404,
0.0178732480853796,
0.036002181470394135,
-0.03810048848390579,
-0.1612391173839569,
0.0012233484303578734,
-0.05489421263337135,
-0.007325115613639355,
-0.0955367386341095,
0.03427118808031082,
-0.07942772656679153,
-0.12866947054862976,
-0.06566812843084335,
0.02257545292377472,
-0.04452718049287796,
0.06320486962795258,
-0.012485776096582413,
0.04781291261315346,
0.013936495408415794,
0.116373211145401,
-0.06097526475787163,
-0.1420048177242279,
-0.03483989089727402,
0.209174245595932,
0.014927948825061321,
0.09773728996515274,
-0.002002303721383214,
0.02667233720421791,
-0.08169734477996826,
0.2934180796146393,
0.30803540349006653,
-0.031660694628953934,
0.09193084388971329,
0.019889486953616142,
0.030467389151453972,
0.02830861322581768,
0.1127677708864212,
0.09172318875789642,
0.31090494990348816,
-0.0842449739575386,
-0.01765313930809498,
-0.0474545992910862,
-0.028141848742961884,
-0.1305065155029297,
-0.00767809571698308,
0.023059293627738953,
-0.008491716347634792,
-0.0696796402335167,
0.07738260924816132,
-0.1417352855205536,
0.11190060526132584,
0.06152660399675369,
-0.18979781866073608,
-0.04262074455618858,
-0.017447708174586296,
0.18770378828048706,
0.011273887008428574,
0.09785391390323639,
-0.038042422384023666,
-0.08157260715961456,
0.02224012464284897,
0.020467286929488182,
-0.16587162017822266,
-0.062374863773584366,
0.08835949748754501,
-0.003819115925580263,
0.1327356994152069,
-0.008414231240749359,
0.040942344814538956,
0.07473018765449524,
0.03335835784673691,
-0.028187336400151253,
0.05410655960440636,
0.027120400220155716,
-0.12804147601127625,
-0.0629287138581276,
0.0003162079374305904,
-0.0004054334422107786,
-0.09281449019908905,
0.01652602292597294,
-0.15870149433612823,
0.031762074679136276,
-0.0786581039428711,
-0.05243808031082153,
-0.008154717274010181,
0.07428926974534988,
-0.022191541269421577,
0.04131091386079788,
0.04452566057443619,
0.012506164610385895,
-0.023894863203167915,
-0.04575556516647339,
0.017746295779943466,
0.06487864255905151,
-0.11055687814950943,
-0.12762592732906342,
-0.12093158066272736,
-0.040044743567705154,
0.05572197586297989,
-0.009214173071086407,
-0.15834109485149384,
-0.053094424307346344,
-0.1168588399887085,
0.02694622613489628,
-0.1837068349123001,
-0.0009551903349347413,
0.06881096959114075,
0.05738770216703415,
0.024873821064829826,
-0.04218771681189537,
0.03676828369498253,
0.05203155800700188,
-0.15142402052879333,
-0.09264479577541351
] |
null | null |
transformers
|
# Moe DialoGPT Model
|
{"tags": ["conversational"]}
|
text-generation
|
disdamoe/DialoGPT-small-moe
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Moe DialoGPT Model
|
[
"# Moe DialoGPT Model"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Moe DialoGPT Model"
] |
[
51,
8
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Moe DialoGPT Model"
] |
[
-0.020122338086366653,
0.053454555571079254,
-0.006552244070917368,
0.011372807435691357,
0.14669811725616455,
0.0015704215038567781,
0.14721523225307465,
0.10417802631855011,
-0.08763240277767181,
-0.04759778827428818,
0.12014064937829971,
0.17259997129440308,
0.0032474284525960684,
0.04444132745265961,
-0.0743740126490593,
-0.3196486234664917,
0.05894393101334572,
0.0400472953915596,
-0.008145652711391449,
0.11097726970911026,
0.09222793579101562,
-0.018218230456113815,
0.08844651281833649,
0.01760999858379364,
-0.15044285356998444,
0.00899314135313034,
-0.0026683835312724113,
-0.112886942923069,
0.10104653239250183,
0.0911237895488739,
0.009631200693547726,
0.021181639283895493,
-0.049121491611003876,
-0.13512352108955383,
0.03823025897145271,
-0.024216074496507645,
-0.031129585579037666,
0.03144538030028343,
0.014310221187770367,
-0.07959286123514175,
0.10286907851696014,
0.10940348356962204,
0.0016987433191388845,
0.04689688980579376,
-0.13486970961093903,
0.022524701431393623,
-0.0049661556258797646,
0.025794891640543938,
0.08362184464931488,
0.1185411736369133,
-0.04725085198879242,
0.12111977487802505,
-0.0852641761302948,
0.09769545495510101,
0.11128703504800797,
-0.3494873344898224,
-0.020997732877731323,
0.13267633318901062,
0.025072257965803146,
0.04233894869685173,
-0.0488676093518734,
0.09407617896795273,
0.012127791531383991,
0.0074059064500033855,
-0.0577593669295311,
-0.08664628863334656,
-0.04583876579999924,
0.005700203124433756,
-0.08512463420629501,
-0.0026563850697129965,
0.24875183403491974,
-0.033815909177064896,
0.0559198223054409,
-0.0965285524725914,
-0.08100157231092453,
0.04618401080369949,
-0.06157234311103821,
-0.042017292231321335,
-0.07276865839958191,
0.07933072000741959,
-0.007800315972417593,
-0.07483551651239395,
-0.10468941926956177,
-0.026203012093901634,
-0.17348605394363403,
0.15540054440498352,
0.04043448343873024,
0.029621876776218414,
-0.2210691273212433,
0.0949259102344513,
0.011747917160391808,
-0.09105373919010162,
0.007437697146087885,
-0.10690437257289886,
0.024721985682845116,
0.020237993448972702,
-0.012494463473558426,
-0.04632283374667168,
0.07844645529985428,
0.12222949415445328,
0.023538194596767426,
0.027204366400837898,
-0.030246859416365623,
0.06806372851133347,
0.032078132033348083,
0.05943058058619499,
-0.010846881195902824,
-0.13007818162441254,
0.02255881018936634,
-0.09433906525373459,
-0.010741072706878185,
-0.0504077710211277,
-0.17561697959899902,
-0.03771292045712471,
0.006688232067972422,
0.0621328167617321,
0.027902722358703613,
0.13153959810733795,
0.006817373912781477,
-0.0386902317404747,
0.028681296855211258,
-0.02446143701672554,
-0.015266270376741886,
0.009861169382929802,
0.00864082295447588,
0.17085111141204834,
0.017112718895077705,
0.04486074671149254,
-0.1428251713514328,
-0.006434294860810041,
-0.039341412484645844,
0.008755096234381199,
-0.01940043829381466,
-0.04745348542928696,
-0.007479233201593161,
0.06305582821369171,
0.006420547608286142,
-0.16735698282718658,
-0.09076131880283356,
-0.0058405594900250435,
0.007833232171833515,
-0.04643293470144272,
-0.09858112782239914,
-0.10407615453004837,
-0.03313332423567772,
0.058580875396728516,
-0.05244611203670502,
-0.004812463186681271,
-0.0628889724612236,
0.061768896877765656,
-0.03942163661122322,
0.06948215514421463,
-0.08477222919464111,
0.08557873964309692,
-0.08240026235580444,
-0.020013349130749702,
-0.08182885497808456,
0.1215318813920021,
0.007565133273601532,
0.05625012516975403,
-0.02065960504114628,
-0.02723207324743271,
-0.09412377327680588,
0.05335519462823868,
-0.0323546901345253,
0.25972819328308105,
-0.05354364216327667,
-0.10465941578149796,
0.290999174118042,
-0.06113103777170181,
-0.11894980818033218,
0.14092423021793365,
0.009082529693841934,
0.054537735879421234,
0.1134263128042221,
0.1460014432668686,
0.03623568266630173,
-0.033331498503685,
0.0876275897026062,
0.10736636817455292,
-0.08433062583208084,
0.008726852014660835,
0.03872595354914665,
-0.03168194741010666,
-0.11219080537557602,
0.038165148347616196,
0.07367351651191711,
0.08133077621459961,
-0.05283155292272568,
-0.017141425982117653,
0.011816175654530525,
0.018844662234187126,
0.06168804690241814,
-0.03280133381485939,
0.11167482286691666,
-0.026799557730555534,
-0.05720139294862747,
-0.005071388557553291,
0.01755097322165966,
-0.040948573499917984,
0.03487555310130119,
-0.06414583325386047,
0.07013518363237381,
-0.02833978645503521,
0.07511444389820099,
-0.12388435006141663,
-0.07359489798545837,
-0.055577490478754044,
0.16096614301204681,
0.06383637338876724,
0.10253820568323135,
0.053020235151052475,
-0.012371718883514404,
-0.02504141442477703,
0.03457217663526535,
0.14865230023860931,
-0.01723194122314453,
-0.07234872877597809,
-0.09762689471244812,
0.10859455913305283,
-0.05160561576485634,
0.12018293887376785,
-0.04547782242298126,
0.008197764866054058,
-0.013942577876150608,
0.10992007702589035,
-0.011604822240769863,
0.038786690682172775,
0.02323710359632969,
-0.020272471010684967,
-0.04441067576408386,
0.011394479312002659,
0.08525247126817703,
0.0034602435771375895,
-0.06948969513177872,
0.2653043866157532,
-0.2064686119556427,
0.15087907016277313,
0.1979166865348816,
-0.21306481957435608,
0.005853266455233097,
-0.14846692979335785,
-0.03656270354986191,
-0.005400137044489384,
0.05285533890128136,
-0.03645110875368118,
0.21388593316078186,
-0.01849883422255516,
0.17473208904266357,
-0.0428045429289341,
-0.022299082949757576,
-0.03014899417757988,
-0.045327480882406235,
-0.008759493939578533,
0.0986492782831192,
0.1383126825094223,
-0.14278057217597961,
0.16189496219158173,
0.10323867946863174,
0.06072273105382919,
0.22221063077449799,
0.04463355988264084,
-0.008255205117166042,
0.055392488837242126,
0.003126539522781968,
-0.03905389830470085,
-0.051305338740348816,
-0.24917548894882202,
-0.03505244478583336,
0.07764487713575363,
0.055178191512823105,
0.1135723665356636,
-0.08689577132463455,
-0.035977140069007874,
-0.01776137948036194,
-0.019939126446843147,
0.07805073261260986,
0.12470252066850662,
0.027342649176716805,
0.12578709423542023,
-0.012231077067553997,
-0.0794093906879425,
0.06192942336201668,
0.009244090877473354,
-0.0851532593369484,
0.1784537136554718,
-0.1349433958530426,
-0.34275349974632263,
-0.11500023305416107,
-0.14209850132465363,
-0.08684682101011276,
0.030369415879249573,
0.1049165427684784,
-0.10581254214048386,
-0.01813383214175701,
-0.003980811219662428,
0.13182754814624786,
-0.07141459733247757,
0.017429912462830544,
-0.04486379027366638,
-0.005561518482863903,
-0.13965731859207153,
-0.10828322172164917,
-0.06061653792858124,
-0.038911741226911545,
-0.06048918142914772,
0.10790928453207016,
-0.1569758802652359,
0.028695210814476013,
0.2240830957889557,
0.06306419521570206,
0.06160638853907585,
-0.027318470180034637,
0.2072751373052597,
-0.10667763650417328,
-0.011551599018275738,
0.15715353190898895,
-0.00923683401197195,
0.0599697045981884,
0.1348978579044342,
-0.012631666846573353,
-0.07484845072031021,
0.046387866139411926,
-0.00038612319622188807,
-0.07661241292953491,
-0.18594913184642792,
-0.13925927877426147,
-0.12071338295936584,
0.08455821871757507,
-0.0025077234022319317,
0.047176048159599304,
0.09712205827236176,
0.058844201266765594,
-0.05034024640917778,
0.0013989267172291875,
0.09620894491672516,
0.08594083040952682,
0.24312393367290497,
-0.06552049517631531,
0.1386769860982895,
-0.024292752146720886,
-0.17813925445079803,
0.06364110857248306,
0.0518956296145916,
0.07357229292392731,
0.06839103996753693,
0.026119951158761978,
0.015616729855537415,
0.029200946912169456,
0.13098227977752686,
0.061992548406124115,
0.035133156925439835,
-0.03989824280142784,
-0.030214650556445122,
-0.055494725704193115,
-0.014937913976609707,
0.04804785177111626,
0.08904039114713669,
-0.1340135782957077,
-0.03553714230656624,
0.03204566612839699,
0.07797766476869583,
0.048403479158878326,
0.09132685512304306,
-0.17913126945495605,
-0.014329435303807259,
0.06038945913314819,
-0.04621971398591995,
-0.11880004405975342,
0.10232041776180267,
0.03087313286960125,
-0.1422169804573059,
0.03595085069537163,
-0.024308424443006516,
0.12322711944580078,
-0.03718060627579689,
0.06920680403709412,
-0.0906938835978508,
-0.06696737557649612,
-0.002407061168923974,
0.09589789062738419,
-0.2966459095478058,
0.16154265403747559,
-0.011083429679274559,
-0.043639205396175385,
-0.09691835194826126,
-0.0027118835132569075,
0.04556822404265404,
0.13466033339500427,
0.1070525050163269,
-0.0038364341016858816,
0.03533511236310005,
0.021727638319134712,
-0.06915749609470367,
0.04521362483501434,
0.09547428041696548,
-0.023483166471123695,
-0.02116122841835022,
-0.040937431156635284,
-0.006612030323594809,
-0.024235153570771217,
-0.0021379238460212946,
-0.01721986010670662,
-0.19245806336402893,
0.09922569990158081,
0.06311391294002533,
0.08863883465528488,
0.03346693143248558,
-0.034810010343790054,
-0.06014024466276169,
0.26171377301216125,
-0.01269515696913004,
-0.09238790720701218,
-0.0859937071800232,
-0.061142098158597946,
0.03400692343711853,
-0.06400790810585022,
0.004236131440848112,
-0.06816507130861282,
0.03589846193790436,
-0.0707441046833992,
-0.18733763694763184,
0.10947642475366592,
-0.0951044037938118,
-0.04278666153550148,
-0.016083842143416405,
0.22640864551067352,
-0.03587714582681656,
0.02828064374625683,
0.05728485435247421,
-0.008244198746979237,
-0.08618466556072235,
-0.11106168478727341,
-0.024690743535757065,
0.0238622035831213,
0.03483385220170021,
0.018784955143928528,
-0.03236537426710129,
-0.058141473680734634,
-0.04499315097928047,
-0.007209425326436758,
0.3128449618816376,
0.12444914132356644,
-0.021196860820055008,
0.15086619555950165,
0.12427251040935516,
-0.06793612241744995,
-0.2793535590171814,
-0.10813409090042114,
-0.08892345428466797,
-0.03318316861987114,
-0.09762878715991974,
-0.15487119555473328,
0.09504393488168716,
-0.030737880617380142,
-0.014433358795940876,
0.09492001682519913,
-0.3018529415130615,
-0.10434827953577042,
0.185251846909523,
-0.026763007044792175,
0.4101439416408539,
-0.12041156738996506,
-0.07121901214122772,
-0.05681409314274788,
-0.21636007726192474,
0.19214604794979095,
-0.049480173736810684,
0.11865435540676117,
-0.0033407302107661963,
0.17779916524887085,
0.04686746373772621,
-0.006758746691048145,
0.10055018216371536,
0.050744399428367615,
-0.054323628544807434,
-0.10687551647424698,
-0.05819445848464966,
-0.016193214803934097,
0.02893606387078762,
0.07013493776321411,
-0.08217702060937881,
0.013587193563580513,
-0.14426860213279724,
-0.050853148102760315,
-0.07942525297403336,
0.03704001381993294,
0.020036712288856506,
-0.07751911133527756,
-0.0037879247684031725,
-0.04527183249592781,
0.0048270379193127155,
0.016428455710411072,
0.16184578835964203,
-0.09281939268112183,
0.11166701465845108,
0.027823256328701973,
0.13179974257946014,
-0.1553894579410553,
-0.0167551189661026,
-0.06078914925456047,
-0.048405084758996964,
0.09681416302919388,
-0.056677285581827164,
0.03842056915163994,
0.11362381279468536,
-0.04790648818016052,
0.09611398726701736,
0.09371668100357056,
-0.0039887758903205395,
0.017267689108848572,
0.08367721736431122,
-0.25496724247932434,
-0.06628002971410751,
-0.07777086645364761,
0.016238581389188766,
0.10213197767734528,
0.07319255918264389,
0.206804096698761,
-0.02381017804145813,
-0.02022024616599083,
0.013330161571502686,
0.027627965435385704,
-0.0523575022816658,
0.07225806266069412,
-0.02514220029115677,
0.020939901471138,
-0.14288951456546783,
0.05582703649997711,
-0.019072286784648895,
-0.10841584950685501,
0.01662214659154415,
0.16805018484592438,
-0.10135041922330856,
-0.12902984023094177,
-0.05573006719350815,
0.12094207108020782,
-0.0865795835852623,
-0.010915049351751804,
-0.040052350610494614,
-0.14292675256729126,
0.06185305491089821,
0.017605995759367943,
0.04813547432422638,
0.0727183073759079,
-0.07111499458551407,
-0.028292817994952202,
-0.04012341797351837,
-0.008724738843739033,
0.027202323079109192,
-0.027936192229390144,
-0.01709148846566677,
0.0638192817568779,
-0.03039754368364811,
0.11727066338062286,
-0.09254702925682068,
-0.11066209524869919,
-0.15181787312030792,
0.0367877222597599,
-0.11656966060400009,
-0.07639410346746445,
-0.12388810515403748,
-0.037334538996219635,
-0.017738355323672295,
-0.021086303517222404,
-0.044633716344833374,
-0.044907376170158386,
-0.10231901705265045,
0.019780850037932396,
-0.046332601457834244,
0.021447475999593735,
-0.07166975736618042,
0.035804785788059235,
0.05166039615869522,
-0.03054150938987732,
0.15345770120620728,
0.15206970274448395,
-0.1157052144408226,
0.08057194948196411,
-0.13172022998332977,
-0.09469562023878098,
0.09758198261260986,
0.008538303896784782,
0.056527260690927505,
0.04937512055039406,
0.011648796498775482,
0.06264273822307587,
0.0727982297539711,
0.055935684591531754,
0.05078178271651268,
-0.08949563652276993,
0.024820441380143166,
-0.0760112926363945,
-0.1377749741077423,
-0.04163423925638199,
-0.030781446024775505,
-0.0019988224375993013,
0.04698943719267845,
0.08448854833841324,
-0.062485262751579285,
0.08213229477405548,
-0.04329666122794151,
0.029291005805134773,
0.040821515023708344,
-0.15776221454143524,
0.06534732133150101,
-0.077607661485672,
0.04218863323330879,
0.0011261156760156155,
0.18506325781345367,
0.03802403435111046,
-0.016097158193588257,
0.03858066722750664,
0.06848480552434921,
0.05554592236876488,
0.001608381629921496,
0.10130178183317184,
0.11529409140348434,
-0.03882962092757225,
-0.10632988065481186,
0.0808408185839653,
0.05456580966711044,
0.03808039799332619,
0.12362581491470337,
-0.009503066539764404,
0.010520880110561848,
0.0966523140668869,
-0.0038336676079779863,
0.009225763380527496,
-0.15119098126888275,
-0.14644508063793182,
-0.02302946336567402,
0.07668392360210419,
-0.06539516896009445,
0.11930768936872482,
0.1785087138414383,
-0.02096579596400261,
0.01770826056599617,
-0.021653318777680397,
-0.05969497188925743,
-0.18493735790252686,
-0.18258368968963623,
-0.07161366939544678,
-0.13217638432979584,
0.007972191087901592,
-0.12159503251314163,
0.02834363467991352,
0.04700496420264244,
0.1134711503982544,
-0.04525124654173851,
0.08269432932138443,
0.008858216926455498,
-0.13294993340969086,
0.07770171761512756,
-0.04516253247857094,
0.1004302129149437,
-0.03689425438642502,
0.0006403166335076094,
-0.051805898547172546,
0.01321996096521616,
0.012202716432511806,
0.055858064442873,
-0.05609887093305588,
0.012399079278111458,
-0.12317277491092682,
-0.07446500658988953,
-0.0587763637304306,
0.06588377058506012,
0.004849957302212715,
0.1310834437608719,
0.03286457806825638,
-0.037147652357816696,
0.028810972347855568,
0.18910980224609375,
-0.049516964703798294,
-0.11592685431241989,
-0.10262774676084518,
0.12139499932527542,
0.00003697569263749756,
0.12442870438098907,
-0.032398954033851624,
0.018459079787135124,
-0.10596984624862671,
0.30322083830833435,
0.320202112197876,
-0.10617043077945709,
0.009820588864386082,
0.02431601658463478,
0.04182903468608856,
0.10918869078159332,
0.09443743526935577,
0.12315189093351364,
0.32644814252853394,
-0.05691545084118843,
-0.03985431045293808,
-0.03601014241576195,
-0.05228576809167862,
-0.05075586214661598,
0.06712073087692261,
0.04292354732751846,
-0.06037615239620209,
-0.008057660423219204,
0.10538668185472488,
-0.26018670201301575,
0.11713974922895432,
-0.149798184633255,
-0.15120474994182587,
-0.08444395661354065,
0.0075708068907260895,
0.054678384214639664,
0.035701047629117966,
0.09239580482244492,
0.0028567002154886723,
-0.07960318773984909,
0.07517139613628387,
0.026262806728482246,
-0.19366630911827087,
-0.021792342886328697,
0.07773366570472717,
-0.03745640814304352,
-0.024657869711518288,
-0.004033327102661133,
0.0680105909705162,
0.06670977920293808,
0.06764757633209229,
-0.004494157154113054,
0.03220243379473686,
0.01312539353966713,
-0.058275140821933746,
0.028823675587773323,
0.033787649124860764,
0.009372885338962078,
-0.04885886609554291,
0.07187241315841675,
-0.09524518251419067,
0.03739583119750023,
-0.024190224707126617,
-0.04190896078944206,
-0.018699193373322487,
0.03789827600121498,
-0.06707873940467834,
0.06686538457870483,
0.09982314705848694,
0.005359741859138012,
0.00470780860632658,
-0.02305891551077366,
-0.0008777777547948062,
-0.03555971011519432,
-0.12470316141843796,
-0.09028323739767075,
-0.1653870940208435,
-0.1067885011434555,
0.034955933690071106,
-0.00238576577976346,
-0.1646544188261032,
-0.0077087413519620895,
-0.11905498802661896,
0.06318683922290802,
-0.1291031539440155,
0.10100734978914261,
0.09148339927196503,
-0.0052084894850850105,
-0.011586785316467285,
-0.07333860546350479,
0.04252929985523224,
0.09061118215322495,
-0.13204126060009003,
-0.1024516299366951
] |
null | null |
transformers
|
# Moe DialoGPT Model
|
{"tags": ["conversational"]}
|
text-generation
|
disdamoe/TheGreatManipulator
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Moe DialoGPT Model
|
[
"# Moe DialoGPT Model"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Moe DialoGPT Model"
] |
[
51,
8
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Moe DialoGPT Model"
] |
[
-0.020122338086366653,
0.053454555571079254,
-0.006552244070917368,
0.011372807435691357,
0.14669811725616455,
0.0015704215038567781,
0.14721523225307465,
0.10417802631855011,
-0.08763240277767181,
-0.04759778827428818,
0.12014064937829971,
0.17259997129440308,
0.0032474284525960684,
0.04444132745265961,
-0.0743740126490593,
-0.3196486234664917,
0.05894393101334572,
0.0400472953915596,
-0.008145652711391449,
0.11097726970911026,
0.09222793579101562,
-0.018218230456113815,
0.08844651281833649,
0.01760999858379364,
-0.15044285356998444,
0.00899314135313034,
-0.0026683835312724113,
-0.112886942923069,
0.10104653239250183,
0.0911237895488739,
0.009631200693547726,
0.021181639283895493,
-0.049121491611003876,
-0.13512352108955383,
0.03823025897145271,
-0.024216074496507645,
-0.031129585579037666,
0.03144538030028343,
0.014310221187770367,
-0.07959286123514175,
0.10286907851696014,
0.10940348356962204,
0.0016987433191388845,
0.04689688980579376,
-0.13486970961093903,
0.022524701431393623,
-0.0049661556258797646,
0.025794891640543938,
0.08362184464931488,
0.1185411736369133,
-0.04725085198879242,
0.12111977487802505,
-0.0852641761302948,
0.09769545495510101,
0.11128703504800797,
-0.3494873344898224,
-0.020997732877731323,
0.13267633318901062,
0.025072257965803146,
0.04233894869685173,
-0.0488676093518734,
0.09407617896795273,
0.012127791531383991,
0.0074059064500033855,
-0.0577593669295311,
-0.08664628863334656,
-0.04583876579999924,
0.005700203124433756,
-0.08512463420629501,
-0.0026563850697129965,
0.24875183403491974,
-0.033815909177064896,
0.0559198223054409,
-0.0965285524725914,
-0.08100157231092453,
0.04618401080369949,
-0.06157234311103821,
-0.042017292231321335,
-0.07276865839958191,
0.07933072000741959,
-0.007800315972417593,
-0.07483551651239395,
-0.10468941926956177,
-0.026203012093901634,
-0.17348605394363403,
0.15540054440498352,
0.04043448343873024,
0.029621876776218414,
-0.2210691273212433,
0.0949259102344513,
0.011747917160391808,
-0.09105373919010162,
0.007437697146087885,
-0.10690437257289886,
0.024721985682845116,
0.020237993448972702,
-0.012494463473558426,
-0.04632283374667168,
0.07844645529985428,
0.12222949415445328,
0.023538194596767426,
0.027204366400837898,
-0.030246859416365623,
0.06806372851133347,
0.032078132033348083,
0.05943058058619499,
-0.010846881195902824,
-0.13007818162441254,
0.02255881018936634,
-0.09433906525373459,
-0.010741072706878185,
-0.0504077710211277,
-0.17561697959899902,
-0.03771292045712471,
0.006688232067972422,
0.0621328167617321,
0.027902722358703613,
0.13153959810733795,
0.006817373912781477,
-0.0386902317404747,
0.028681296855211258,
-0.02446143701672554,
-0.015266270376741886,
0.009861169382929802,
0.00864082295447588,
0.17085111141204834,
0.017112718895077705,
0.04486074671149254,
-0.1428251713514328,
-0.006434294860810041,
-0.039341412484645844,
0.008755096234381199,
-0.01940043829381466,
-0.04745348542928696,
-0.007479233201593161,
0.06305582821369171,
0.006420547608286142,
-0.16735698282718658,
-0.09076131880283356,
-0.0058405594900250435,
0.007833232171833515,
-0.04643293470144272,
-0.09858112782239914,
-0.10407615453004837,
-0.03313332423567772,
0.058580875396728516,
-0.05244611203670502,
-0.004812463186681271,
-0.0628889724612236,
0.061768896877765656,
-0.03942163661122322,
0.06948215514421463,
-0.08477222919464111,
0.08557873964309692,
-0.08240026235580444,
-0.020013349130749702,
-0.08182885497808456,
0.1215318813920021,
0.007565133273601532,
0.05625012516975403,
-0.02065960504114628,
-0.02723207324743271,
-0.09412377327680588,
0.05335519462823868,
-0.0323546901345253,
0.25972819328308105,
-0.05354364216327667,
-0.10465941578149796,
0.290999174118042,
-0.06113103777170181,
-0.11894980818033218,
0.14092423021793365,
0.009082529693841934,
0.054537735879421234,
0.1134263128042221,
0.1460014432668686,
0.03623568266630173,
-0.033331498503685,
0.0876275897026062,
0.10736636817455292,
-0.08433062583208084,
0.008726852014660835,
0.03872595354914665,
-0.03168194741010666,
-0.11219080537557602,
0.038165148347616196,
0.07367351651191711,
0.08133077621459961,
-0.05283155292272568,
-0.017141425982117653,
0.011816175654530525,
0.018844662234187126,
0.06168804690241814,
-0.03280133381485939,
0.11167482286691666,
-0.026799557730555534,
-0.05720139294862747,
-0.005071388557553291,
0.01755097322165966,
-0.040948573499917984,
0.03487555310130119,
-0.06414583325386047,
0.07013518363237381,
-0.02833978645503521,
0.07511444389820099,
-0.12388435006141663,
-0.07359489798545837,
-0.055577490478754044,
0.16096614301204681,
0.06383637338876724,
0.10253820568323135,
0.053020235151052475,
-0.012371718883514404,
-0.02504141442477703,
0.03457217663526535,
0.14865230023860931,
-0.01723194122314453,
-0.07234872877597809,
-0.09762689471244812,
0.10859455913305283,
-0.05160561576485634,
0.12018293887376785,
-0.04547782242298126,
0.008197764866054058,
-0.013942577876150608,
0.10992007702589035,
-0.011604822240769863,
0.038786690682172775,
0.02323710359632969,
-0.020272471010684967,
-0.04441067576408386,
0.011394479312002659,
0.08525247126817703,
0.0034602435771375895,
-0.06948969513177872,
0.2653043866157532,
-0.2064686119556427,
0.15087907016277313,
0.1979166865348816,
-0.21306481957435608,
0.005853266455233097,
-0.14846692979335785,
-0.03656270354986191,
-0.005400137044489384,
0.05285533890128136,
-0.03645110875368118,
0.21388593316078186,
-0.01849883422255516,
0.17473208904266357,
-0.0428045429289341,
-0.022299082949757576,
-0.03014899417757988,
-0.045327480882406235,
-0.008759493939578533,
0.0986492782831192,
0.1383126825094223,
-0.14278057217597961,
0.16189496219158173,
0.10323867946863174,
0.06072273105382919,
0.22221063077449799,
0.04463355988264084,
-0.008255205117166042,
0.055392488837242126,
0.003126539522781968,
-0.03905389830470085,
-0.051305338740348816,
-0.24917548894882202,
-0.03505244478583336,
0.07764487713575363,
0.055178191512823105,
0.1135723665356636,
-0.08689577132463455,
-0.035977140069007874,
-0.01776137948036194,
-0.019939126446843147,
0.07805073261260986,
0.12470252066850662,
0.027342649176716805,
0.12578709423542023,
-0.012231077067553997,
-0.0794093906879425,
0.06192942336201668,
0.009244090877473354,
-0.0851532593369484,
0.1784537136554718,
-0.1349433958530426,
-0.34275349974632263,
-0.11500023305416107,
-0.14209850132465363,
-0.08684682101011276,
0.030369415879249573,
0.1049165427684784,
-0.10581254214048386,
-0.01813383214175701,
-0.003980811219662428,
0.13182754814624786,
-0.07141459733247757,
0.017429912462830544,
-0.04486379027366638,
-0.005561518482863903,
-0.13965731859207153,
-0.10828322172164917,
-0.06061653792858124,
-0.038911741226911545,
-0.06048918142914772,
0.10790928453207016,
-0.1569758802652359,
0.028695210814476013,
0.2240830957889557,
0.06306419521570206,
0.06160638853907585,
-0.027318470180034637,
0.2072751373052597,
-0.10667763650417328,
-0.011551599018275738,
0.15715353190898895,
-0.00923683401197195,
0.0599697045981884,
0.1348978579044342,
-0.012631666846573353,
-0.07484845072031021,
0.046387866139411926,
-0.00038612319622188807,
-0.07661241292953491,
-0.18594913184642792,
-0.13925927877426147,
-0.12071338295936584,
0.08455821871757507,
-0.0025077234022319317,
0.047176048159599304,
0.09712205827236176,
0.058844201266765594,
-0.05034024640917778,
0.0013989267172291875,
0.09620894491672516,
0.08594083040952682,
0.24312393367290497,
-0.06552049517631531,
0.1386769860982895,
-0.024292752146720886,
-0.17813925445079803,
0.06364110857248306,
0.0518956296145916,
0.07357229292392731,
0.06839103996753693,
0.026119951158761978,
0.015616729855537415,
0.029200946912169456,
0.13098227977752686,
0.061992548406124115,
0.035133156925439835,
-0.03989824280142784,
-0.030214650556445122,
-0.055494725704193115,
-0.014937913976609707,
0.04804785177111626,
0.08904039114713669,
-0.1340135782957077,
-0.03553714230656624,
0.03204566612839699,
0.07797766476869583,
0.048403479158878326,
0.09132685512304306,
-0.17913126945495605,
-0.014329435303807259,
0.06038945913314819,
-0.04621971398591995,
-0.11880004405975342,
0.10232041776180267,
0.03087313286960125,
-0.1422169804573059,
0.03595085069537163,
-0.024308424443006516,
0.12322711944580078,
-0.03718060627579689,
0.06920680403709412,
-0.0906938835978508,
-0.06696737557649612,
-0.002407061168923974,
0.09589789062738419,
-0.2966459095478058,
0.16154265403747559,
-0.011083429679274559,
-0.043639205396175385,
-0.09691835194826126,
-0.0027118835132569075,
0.04556822404265404,
0.13466033339500427,
0.1070525050163269,
-0.0038364341016858816,
0.03533511236310005,
0.021727638319134712,
-0.06915749609470367,
0.04521362483501434,
0.09547428041696548,
-0.023483166471123695,
-0.02116122841835022,
-0.040937431156635284,
-0.006612030323594809,
-0.024235153570771217,
-0.0021379238460212946,
-0.01721986010670662,
-0.19245806336402893,
0.09922569990158081,
0.06311391294002533,
0.08863883465528488,
0.03346693143248558,
-0.034810010343790054,
-0.06014024466276169,
0.26171377301216125,
-0.01269515696913004,
-0.09238790720701218,
-0.0859937071800232,
-0.061142098158597946,
0.03400692343711853,
-0.06400790810585022,
0.004236131440848112,
-0.06816507130861282,
0.03589846193790436,
-0.0707441046833992,
-0.18733763694763184,
0.10947642475366592,
-0.0951044037938118,
-0.04278666153550148,
-0.016083842143416405,
0.22640864551067352,
-0.03587714582681656,
0.02828064374625683,
0.05728485435247421,
-0.008244198746979237,
-0.08618466556072235,
-0.11106168478727341,
-0.024690743535757065,
0.0238622035831213,
0.03483385220170021,
0.018784955143928528,
-0.03236537426710129,
-0.058141473680734634,
-0.04499315097928047,
-0.007209425326436758,
0.3128449618816376,
0.12444914132356644,
-0.021196860820055008,
0.15086619555950165,
0.12427251040935516,
-0.06793612241744995,
-0.2793535590171814,
-0.10813409090042114,
-0.08892345428466797,
-0.03318316861987114,
-0.09762878715991974,
-0.15487119555473328,
0.09504393488168716,
-0.030737880617380142,
-0.014433358795940876,
0.09492001682519913,
-0.3018529415130615,
-0.10434827953577042,
0.185251846909523,
-0.026763007044792175,
0.4101439416408539,
-0.12041156738996506,
-0.07121901214122772,
-0.05681409314274788,
-0.21636007726192474,
0.19214604794979095,
-0.049480173736810684,
0.11865435540676117,
-0.0033407302107661963,
0.17779916524887085,
0.04686746373772621,
-0.006758746691048145,
0.10055018216371536,
0.050744399428367615,
-0.054323628544807434,
-0.10687551647424698,
-0.05819445848464966,
-0.016193214803934097,
0.02893606387078762,
0.07013493776321411,
-0.08217702060937881,
0.013587193563580513,
-0.14426860213279724,
-0.050853148102760315,
-0.07942525297403336,
0.03704001381993294,
0.020036712288856506,
-0.07751911133527756,
-0.0037879247684031725,
-0.04527183249592781,
0.0048270379193127155,
0.016428455710411072,
0.16184578835964203,
-0.09281939268112183,
0.11166701465845108,
0.027823256328701973,
0.13179974257946014,
-0.1553894579410553,
-0.0167551189661026,
-0.06078914925456047,
-0.048405084758996964,
0.09681416302919388,
-0.056677285581827164,
0.03842056915163994,
0.11362381279468536,
-0.04790648818016052,
0.09611398726701736,
0.09371668100357056,
-0.0039887758903205395,
0.017267689108848572,
0.08367721736431122,
-0.25496724247932434,
-0.06628002971410751,
-0.07777086645364761,
0.016238581389188766,
0.10213197767734528,
0.07319255918264389,
0.206804096698761,
-0.02381017804145813,
-0.02022024616599083,
0.013330161571502686,
0.027627965435385704,
-0.0523575022816658,
0.07225806266069412,
-0.02514220029115677,
0.020939901471138,
-0.14288951456546783,
0.05582703649997711,
-0.019072286784648895,
-0.10841584950685501,
0.01662214659154415,
0.16805018484592438,
-0.10135041922330856,
-0.12902984023094177,
-0.05573006719350815,
0.12094207108020782,
-0.0865795835852623,
-0.010915049351751804,
-0.040052350610494614,
-0.14292675256729126,
0.06185305491089821,
0.017605995759367943,
0.04813547432422638,
0.0727183073759079,
-0.07111499458551407,
-0.028292817994952202,
-0.04012341797351837,
-0.008724738843739033,
0.027202323079109192,
-0.027936192229390144,
-0.01709148846566677,
0.0638192817568779,
-0.03039754368364811,
0.11727066338062286,
-0.09254702925682068,
-0.11066209524869919,
-0.15181787312030792,
0.0367877222597599,
-0.11656966060400009,
-0.07639410346746445,
-0.12388810515403748,
-0.037334538996219635,
-0.017738355323672295,
-0.021086303517222404,
-0.044633716344833374,
-0.044907376170158386,
-0.10231901705265045,
0.019780850037932396,
-0.046332601457834244,
0.021447475999593735,
-0.07166975736618042,
0.035804785788059235,
0.05166039615869522,
-0.03054150938987732,
0.15345770120620728,
0.15206970274448395,
-0.1157052144408226,
0.08057194948196411,
-0.13172022998332977,
-0.09469562023878098,
0.09758198261260986,
0.008538303896784782,
0.056527260690927505,
0.04937512055039406,
0.011648796498775482,
0.06264273822307587,
0.0727982297539711,
0.055935684591531754,
0.05078178271651268,
-0.08949563652276993,
0.024820441380143166,
-0.0760112926363945,
-0.1377749741077423,
-0.04163423925638199,
-0.030781446024775505,
-0.0019988224375993013,
0.04698943719267845,
0.08448854833841324,
-0.062485262751579285,
0.08213229477405548,
-0.04329666122794151,
0.029291005805134773,
0.040821515023708344,
-0.15776221454143524,
0.06534732133150101,
-0.077607661485672,
0.04218863323330879,
0.0011261156760156155,
0.18506325781345367,
0.03802403435111046,
-0.016097158193588257,
0.03858066722750664,
0.06848480552434921,
0.05554592236876488,
0.001608381629921496,
0.10130178183317184,
0.11529409140348434,
-0.03882962092757225,
-0.10632988065481186,
0.0808408185839653,
0.05456580966711044,
0.03808039799332619,
0.12362581491470337,
-0.009503066539764404,
0.010520880110561848,
0.0966523140668869,
-0.0038336676079779863,
0.009225763380527496,
-0.15119098126888275,
-0.14644508063793182,
-0.02302946336567402,
0.07668392360210419,
-0.06539516896009445,
0.11930768936872482,
0.1785087138414383,
-0.02096579596400261,
0.01770826056599617,
-0.021653318777680397,
-0.05969497188925743,
-0.18493735790252686,
-0.18258368968963623,
-0.07161366939544678,
-0.13217638432979584,
0.007972191087901592,
-0.12159503251314163,
0.02834363467991352,
0.04700496420264244,
0.1134711503982544,
-0.04525124654173851,
0.08269432932138443,
0.008858216926455498,
-0.13294993340969086,
0.07770171761512756,
-0.04516253247857094,
0.1004302129149437,
-0.03689425438642502,
0.0006403166335076094,
-0.051805898547172546,
0.01321996096521616,
0.012202716432511806,
0.055858064442873,
-0.05609887093305588,
0.012399079278111458,
-0.12317277491092682,
-0.07446500658988953,
-0.0587763637304306,
0.06588377058506012,
0.004849957302212715,
0.1310834437608719,
0.03286457806825638,
-0.037147652357816696,
0.028810972347855568,
0.18910980224609375,
-0.049516964703798294,
-0.11592685431241989,
-0.10262774676084518,
0.12139499932527542,
0.00003697569263749756,
0.12442870438098907,
-0.032398954033851624,
0.018459079787135124,
-0.10596984624862671,
0.30322083830833435,
0.320202112197876,
-0.10617043077945709,
0.009820588864386082,
0.02431601658463478,
0.04182903468608856,
0.10918869078159332,
0.09443743526935577,
0.12315189093351364,
0.32644814252853394,
-0.05691545084118843,
-0.03985431045293808,
-0.03601014241576195,
-0.05228576809167862,
-0.05075586214661598,
0.06712073087692261,
0.04292354732751846,
-0.06037615239620209,
-0.008057660423219204,
0.10538668185472488,
-0.26018670201301575,
0.11713974922895432,
-0.149798184633255,
-0.15120474994182587,
-0.08444395661354065,
0.0075708068907260895,
0.054678384214639664,
0.035701047629117966,
0.09239580482244492,
0.0028567002154886723,
-0.07960318773984909,
0.07517139613628387,
0.026262806728482246,
-0.19366630911827087,
-0.021792342886328697,
0.07773366570472717,
-0.03745640814304352,
-0.024657869711518288,
-0.004033327102661133,
0.0680105909705162,
0.06670977920293808,
0.06764757633209229,
-0.004494157154113054,
0.03220243379473686,
0.01312539353966713,
-0.058275140821933746,
0.028823675587773323,
0.033787649124860764,
0.009372885338962078,
-0.04885886609554291,
0.07187241315841675,
-0.09524518251419067,
0.03739583119750023,
-0.024190224707126617,
-0.04190896078944206,
-0.018699193373322487,
0.03789827600121498,
-0.06707873940467834,
0.06686538457870483,
0.09982314705848694,
0.005359741859138012,
0.00470780860632658,
-0.02305891551077366,
-0.0008777777547948062,
-0.03555971011519432,
-0.12470316141843796,
-0.09028323739767075,
-0.1653870940208435,
-0.1067885011434555,
0.034955933690071106,
-0.00238576577976346,
-0.1646544188261032,
-0.0077087413519620895,
-0.11905498802661896,
0.06318683922290802,
-0.1291031539440155,
0.10100734978914261,
0.09148339927196503,
-0.0052084894850850105,
-0.011586785316467285,
-0.07333860546350479,
0.04252929985523224,
0.09061118215322495,
-0.13204126060009003,
-0.1024516299366951
] |
null | null |
transformers
|
# The Manipulator
|
{"tags": ["conversational"]}
|
text-generation
|
disdamoe/TheManipulator
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# The Manipulator
|
[
"# The Manipulator"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# The Manipulator"
] |
[
51,
4
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# The Manipulator"
] |
[
0.014883825555443764,
0.0454658679664135,
-0.007144797593355179,
0.01665303111076355,
0.16748404502868652,
0.0014616119442507625,
0.1344665139913559,
0.1157367005944252,
-0.047142017632722855,
-0.013967991806566715,
0.14231161773204803,
0.24798618257045746,
0.00018006058235187083,
0.0689060240983963,
-0.037513621151447296,
-0.28962767124176025,
0.07050053775310516,
0.04311024770140648,
-0.019637107849121094,
0.1396523267030716,
0.07281839102506638,
-0.0736924335360527,
0.09357225894927979,
0.004220649134367704,
-0.13989558815956116,
0.012286529876291752,
0.055960919708013535,
-0.10607269406318665,
0.09996669739484787,
0.06088558956980705,
0.04918632283806801,
0.045399535447359085,
-0.07345998287200928,
-0.12565995752811432,
0.03178450092673302,
0.028812719509005547,
-0.05597119778394699,
0.04025816544890404,
0.062423910945653915,
-0.09440647065639496,
0.15354551374912262,
0.0612817145884037,
-0.002185069490224123,
0.0550050213932991,
-0.16653050482273102,
-0.0016489251283928752,
-0.009179830551147461,
0.0020684427581727505,
0.03447818383574486,
0.1293182373046875,
-0.0290139839053154,
0.17246013879776,
-0.07587143778800964,
0.1271827518939972,
0.13876740634441376,
-0.3573184609413147,
-0.025508278980851173,
0.06660258769989014,
0.031264547258615494,
0.11613204330205917,
-0.06353288143873215,
0.07018651813268661,
0.004795621149241924,
0.006533752661198378,
0.010411283932626247,
-0.0891573578119278,
-0.04605482518672943,
0.03687614947557449,
-0.09009730815887451,
-0.04980718716979027,
0.20485404133796692,
-0.06175806000828743,
0.056354183703660965,
-0.05473553389310837,
-0.10622666031122208,
-0.0004976836498826742,
-0.008305494673550129,
0.0046763066202402115,
-0.038783494383096695,
0.07763000577688217,
-0.03167513385415077,
-0.07056478410959244,
-0.10500719398260117,
-0.035997238010168076,
-0.15850871801376343,
0.17858433723449707,
-0.0016861527692526579,
0.05143623799085617,
-0.17010119557380676,
0.0790015161037445,
0.08543659001588821,
-0.08580969274044037,
0.004651641473174095,
-0.0763363242149353,
0.015128095634281635,
0.003233846975490451,
-0.0401654876768589,
-0.09570176154375076,
0.04252283275127411,
0.07961665093898773,
-0.051052480936050415,
0.04635723680257797,
-0.01842065155506134,
0.0685994029045105,
0.05919799208641052,
0.11823663860559464,
-0.025366267189383507,
0.007298166863620281,
0.0191203523427248,
-0.0967910960316658,
-0.0019792465027421713,
-0.09629210084676743,
-0.1458749771118164,
-0.0034234588965773582,
0.0393475741147995,
0.06806501001119614,
-0.005895023234188557,
0.13654062151908875,
-0.02445017360150814,
-0.03769557178020477,
-0.03678637742996216,
-0.05880620703101158,
-0.004465960897505283,
0.02788194641470909,
0.005824938882142305,
0.16867566108703613,
-0.010872279293835163,
-0.0056882016360759735,
-0.11930453032255173,
0.0061429338529706,
-0.0905868411064148,
0.023664681240916252,
-0.058825310319662094,
-0.07462070137262344,
-0.009264778345823288,
-0.06389092653989792,
-0.0058107818476855755,
-0.15346048772335052,
-0.13154460489749908,
-0.004907190799713135,
-0.0010783637408167124,
-0.033216193318367004,
-0.05029226094484329,
-0.05625554546713829,
0.0007847893284633756,
0.05902617424726486,
-0.06412111222743988,
-0.0663277879357338,
-0.06258153170347214,
0.11777046322822571,
-0.0390036478638649,
0.0755666121840477,
-0.0550328753888607,
0.07704361528158188,
-0.10679396241903305,
-0.0006131156114861369,
-0.11560725420713425,
0.09002526849508286,
-0.021521760150790215,
0.13559021055698395,
-0.019821522757411003,
-0.03011210821568966,
-0.029701145365834236,
0.05185636505484581,
-0.040141720324754715,
0.2007608413696289,
-0.07004991173744202,
-0.1412474364042282,
0.2716522812843323,
-0.08669275045394897,
-0.12882965803146362,
0.08391731232404709,
-0.005905485711991787,
0.08283907920122147,
0.11561153084039688,
0.14313217997550964,
0.007376585155725479,
0.041037388145923615,
0.08739490807056427,
0.09858401864767075,
-0.14326965808868408,
0.03158571943640709,
0.016469329595565796,
-0.022161200642585754,
-0.13489679992198944,
0.02741065062582493,
0.06426297873258591,
0.08156872540712357,
-0.044629618525505066,
-0.040976233780384064,
0.020195716992020607,
0.0022137004416435957,
0.10231811553239822,
0.0003810857597272843,
0.1112707182765007,
-0.037540655583143234,
-0.030475467443466187,
-0.02274959348142147,
0.0010203840211033821,
-0.05722431465983391,
0.019898580387234688,
-0.08347208797931671,
0.05866890028119087,
-0.02791621908545494,
0.07789567112922668,
-0.17662116885185242,
-0.057803504168987274,
-0.029396744444966316,
0.10997717082500458,
0.04345273971557617,
0.08330792188644409,
0.0756559893488884,
-0.025632038712501526,
0.0429147407412529,
0.013952706940472126,
0.1274101436138153,
-0.012418444268405437,
-0.04276067018508911,
-0.108139768242836,
0.08854817599058151,
-0.04998469725251198,
0.05137820914387703,
-0.04040716588497162,
0.010557075962424278,
0.1026747077703476,
0.0373496487736702,
-0.012792892754077911,
0.027538586407899857,
0.0020395119208842516,
0.009222139604389668,
-0.05683276429772377,
0.003650367958471179,
0.12785446643829346,
0.0046746195293962955,
-0.061630744487047195,
0.2122150957584381,
-0.22258761525154114,
0.21810469031333923,
0.1902017891407013,
-0.2988818883895874,
0.01700759306550026,
-0.06503825634717941,
-0.021649207919836044,
0.012322939932346344,
0.10432810336351395,
-0.006991541013121605,
0.1811150312423706,
-0.0061532980762422085,
0.18073558807373047,
-0.05172258988022804,
-0.0305652879178524,
-0.004057776648551226,
-0.03594394400715828,
-0.014764908701181412,
0.0703083798289299,
0.11829090863466263,
-0.1371910572052002,
0.1807042509317398,
0.17535920441150665,
0.03419617563486099,
0.22372707724571228,
0.017833353951573372,
-0.021156348288059235,
0.06383121013641357,
-0.008826727978885174,
-0.025974364951252937,
-0.09824623167514801,
-0.25172924995422363,
-0.052773136645555496,
0.04987147077918053,
0.020613091066479683,
0.09961181879043579,
-0.09454506635665894,
-0.023222459480166435,
0.0032915121410042048,
-0.017154693603515625,
0.00727944727987051,
0.07505559176206589,
0.04882458969950676,
0.12067793309688568,
0.014298253692686558,
-0.0200570747256279,
0.09096568822860718,
0.027456393465399742,
-0.10845266282558441,
0.18592403829097748,
-0.15497569739818573,
-0.35873761773109436,
-0.13382688164710999,
-0.17608267068862915,
-0.06705685704946518,
0.045802872627973557,
0.07471491396427155,
-0.11489368975162506,
-0.02201760746538639,
0.020416196435689926,
0.08774400502443314,
-0.07996799051761627,
-0.0027348436415195465,
0.0060137128457427025,
0.020954983308911324,
-0.08567940443754196,
-0.11414030939340591,
-0.048158369958400726,
-0.007956206798553467,
-0.12692470848560333,
0.12265033274888992,
-0.0955887883901596,
0.07330449670553207,
0.22989629209041595,
0.06235823035240173,
0.0731356218457222,
-0.025081723928451538,
0.20075148344039917,
-0.10271098464727402,
-0.04236256703734398,
0.17373120784759521,
-0.04427562654018402,
0.06542841345071793,
0.0781903862953186,
-0.013837398029863834,
-0.08148066699504852,
0.01843314617872238,
-0.018055962398648262,
-0.08425327390432358,
-0.16982854902744293,
-0.11736083775758743,
-0.11091342568397522,
0.1069895476102829,
0.04351744055747986,
0.049489159137010574,
0.15313918888568878,
0.060078151524066925,
-0.0008242898038588464,
-0.015863390639424324,
0.07405480742454529,
0.07263907790184021,
0.17618869245052338,
-0.06473363190889359,
0.1449594795703888,
-0.036464981734752655,
-0.15665526688098907,
0.06370770931243896,
0.028528230264782906,
0.11473739892244339,
0.07131461054086685,
0.1576513946056366,
0.026770854368805885,
0.017808035016059875,
0.11439624428749084,
0.14836537837982178,
0.013190423138439655,
-0.037666767835617065,
-0.02253386192023754,
-0.04409438371658325,
-0.09992152452468872,
0.03109763190150261,
0.06870356947183609,
-0.1326044648885727,
-0.05224524810910225,
-0.06647767126560211,
0.1056421548128128,
0.09139201790094376,
0.022998657077550888,
-0.21224433183670044,
-0.010382328182458878,
0.06067068129777908,
-0.020306752994656563,
-0.09329719096422195,
0.08582904189825058,
0.013559125363826752,
-0.14554673433303833,
0.03069641813635826,
-0.06190976873040199,
0.11298580467700958,
-0.09788205474615097,
0.09902738034725189,
-0.06947033107280731,
-0.058506790548563004,
0.024597397074103355,
0.11234699189662933,
-0.2861071527004242,
0.1786905974149704,
-0.00616513891145587,
-0.045717205852270126,
-0.10828336328268051,
-0.015819592401385307,
0.03890933096408844,
0.10265876352787018,
0.09635535627603531,
-0.002304624766111374,
0.023372286930680275,
-0.10470911115407944,
-0.0036169199738651514,
0.014001607894897461,
0.127450630068779,
-0.03468085452914238,
0.008225039578974247,
-0.046956416219472885,
0.002182602882385254,
-0.038072459399700165,
-0.02539622224867344,
0.01981808803975582,
-0.20588351786136627,
0.11295570433139801,
0.005498162936419249,
0.09202663600444794,
0.0236995629966259,
-0.015323994681239128,
-0.11507632583379745,
0.26375970244407654,
-0.0659688264131546,
-0.07298048585653305,
-0.09286924451589584,
-0.059480272233486176,
0.014099069871008396,
-0.06588777154684067,
0.001938546309247613,
-0.07617183774709702,
0.061221517622470856,
-0.052078817039728165,
-0.19639737904071808,
0.10492207854986191,
-0.06885666400194168,
-0.02897016704082489,
-0.001618721173144877,
0.15222351253032684,
-0.03260504826903343,
-0.006907085422426462,
0.02090389095246792,
0.00714652007445693,
-0.08269334584474564,
-0.11788803339004517,
0.03144260495901108,
0.026514960452914238,
-0.018081992864608765,
0.017291266471147537,
-0.017077963799238205,
0.05748187005519867,
-0.06949345022439957,
0.014012801460921764,
0.35706862807273865,
0.10114816576242447,
-0.03518149256706238,
0.19264905154705048,
0.1347483992576599,
-0.07655637711286545,
-0.30959877371788025,
-0.09757442027330399,
-0.11770082265138626,
-0.04858952760696411,
-0.046419814229011536,
-0.21574616432189941,
0.03754841163754463,
-0.021447429433465004,
0.010515617206692696,
0.08234025537967682,
-0.2756253778934479,
-0.07833889871835709,
0.13795341551303864,
0.03619682043790817,
0.3845108151435852,
-0.13213446736335754,
-0.07637833803892136,
-0.04717346280813217,
-0.16480690240859985,
0.14741073548793793,
-0.09335241466760635,
0.11738520860671997,
-0.009828162379562855,
0.1208743155002594,
0.04601040109992027,
-0.028116533532738686,
0.07396551221609116,
-0.011412917636334896,
-0.005476068705320358,
-0.11251836270093918,
-0.0663931742310524,
0.07737463712692261,
0.03497885912656784,
0.004689062479883432,
-0.051119115203619,
0.011455352418124676,
-0.09272082149982452,
-0.060767654329538345,
-0.08485082536935806,
0.021989434957504272,
0.0333947129547596,
-0.061175569891929626,
0.01998002640902996,
-0.04405832663178444,
0.0008318151230923831,
0.02646160125732422,
0.1356378197669983,
-0.07888129353523254,
0.11884288489818573,
0.10353755950927734,
0.14369627833366394,
-0.18913036584854126,
-0.061226457357406616,
-0.04496246948838234,
-0.028466910123825073,
0.09478455781936646,
-0.04329914599657059,
0.07588669657707214,
0.10879797488451004,
-0.03747829794883728,
0.0923648402094841,
0.12393804639577866,
0.00012539523595478386,
0.0013031027046963573,
0.08504687994718552,
-0.2611694633960724,
-0.054801758378744125,
-0.05516568198800087,
-0.017278745770454407,
0.1049690768122673,
0.04572523012757301,
0.1913018375635147,
0.031346291303634644,
-0.016984151676297188,
-0.00027411317569203675,
0.0025359613355249166,
-0.04702361300587654,
0.06226722151041031,
-0.04915449023246765,
0.022371917963027954,
-0.15310229361057281,
0.06260208785533905,
0.004905209876596928,
-0.1898026168346405,
0.024226205423474312,
0.17805115878582,
-0.11654132604598999,
-0.1313687562942505,
-0.05303666368126869,
0.14697620272636414,
-0.10352499783039093,
0.0007860630284994841,
-0.06021329388022423,
-0.16614022850990295,
0.06706903874874115,
0.1360856294631958,
0.06907375156879425,
0.10332933813333511,
-0.0895344540476799,
-0.02190297655761242,
-0.04908590763807297,
-0.018183497712016106,
-0.010097880847752094,
-0.0395304374396801,
-0.0786660760641098,
0.11728822439908981,
-0.0099396463483572,
0.13548798859119415,
-0.10124894231557846,
-0.12072864174842834,
-0.16310171782970428,
0.03302323445677757,
-0.05170804262161255,
-0.09482394903898239,
-0.12510502338409424,
-0.04797473922371864,
0.012129097245633602,
0.008410325273871422,
-0.05212542787194252,
-0.07569818198680878,
-0.12174325436353683,
0.03091995231807232,
-0.03790159896016121,
0.021354392170906067,
-0.05949777737259865,
0.01480774860829115,
0.07336944341659546,
-0.05435578525066376,
0.13937322795391083,
0.14089922606945038,
-0.09700946509838104,
0.10239675641059875,
-0.1543223112821579,
-0.08946294337511063,
0.08236532658338547,
-0.004426512401551008,
0.03798380494117737,
0.04637150838971138,
0.027721518650650978,
0.06259087473154068,
0.01826617866754532,
0.05630608648061752,
0.044171854853630066,
-0.08988922089338303,
0.09756893664598465,
-0.01346677727997303,
-0.15607988834381104,
-0.04360052943229675,
-0.027574701234698296,
0.02066756784915924,
0.0034573518205434084,
0.13582798838615417,
-0.08337409794330597,
0.10379200428724289,
-0.04315445199608803,
0.02505050227046013,
0.027984928339719772,
-0.16740387678146362,
-0.011953690089285374,
-0.09939499944448471,
0.03235454484820366,
-0.022195536643266678,
0.23260995745658875,
0.030926546081900597,
-0.02196393348276615,
0.028921961784362793,
0.04060817509889603,
0.003174412762746215,
0.02265756018459797,
0.1467561572790146,
0.12562009692192078,
-0.047071248292922974,
-0.07691462337970734,
0.0783657357096672,
0.031376417726278305,
0.016359614208340645,
0.16321931779384613,
0.005707679316401482,
0.04192550480365753,
0.09837380796670914,
-0.002328020753338933,
0.026931650936603546,
-0.12452366203069687,
-0.09179870784282684,
-0.05067381635308266,
0.05687308683991432,
-0.05312167853116989,
0.12396077066659927,
0.18889212608337402,
0.009125622920691967,
0.041913729161024094,
-0.008201380260288715,
-0.0699988454580307,
-0.18290089070796967,
-0.1372874677181244,
-0.07252079248428345,
-0.1578603833913803,
0.020007062703371048,
-0.09490541368722916,
0.05172286555171013,
0.022859610617160797,
0.06476108729839325,
-0.06746438145637512,
0.09769782423973083,
0.025131775066256523,
-0.0887874886393547,
0.06645552068948746,
-0.0390721820294857,
0.0470917709171772,
-0.013401621021330357,
-0.010146171785891056,
-0.09245496243238449,
-0.013232049532234669,
0.031895849853754044,
0.057291533797979355,
-0.06086823344230652,
0.013308489695191383,
-0.14112967252731323,
-0.11151551455259323,
-0.038542017340660095,
0.06306838244199753,
-0.05139867588877678,
0.12277659773826599,
0.006512434687465429,
-0.051781270653009415,
0.010210886597633362,
0.14525724947452545,
-0.055603429675102234,
-0.03651205822825432,
-0.07107388973236084,
0.22041946649551392,
0.023691102862358093,
0.11646556854248047,
-0.015649018809199333,
0.008110293187201023,
-0.09071490913629532,
0.3623579442501068,
0.3034762442111969,
-0.11676162481307983,
0.022797681391239166,
-0.012149052694439888,
0.05288988724350929,
0.11721470952033997,
0.1411910206079483,
0.08307590335607529,
0.2976725399494171,
-0.07964807003736496,
-0.02995922602713108,
-0.05290387570858002,
-0.019801944494247437,
-0.06733245402574539,
0.049430087208747864,
0.08616531640291214,
-0.03434309363365173,
-0.04856554791331291,
0.07467606663703918,
-0.22900639474391937,
0.19256100058555603,
-0.05157692730426788,
-0.17473921179771423,
-0.04056262969970703,
0.01643957756459713,
0.09952448308467865,
-0.001540247700177133,
0.10232507437467575,
0.012923779897391796,
-0.11207780241966248,
0.031724851578474045,
0.048428915441036224,
-0.22418545186519623,
0.010701057501137257,
0.08235998451709747,
-0.06974068284034729,
0.005683234892785549,
-0.05689745396375656,
0.03545171767473221,
0.08389890193939209,
0.07122384011745453,
-0.03527025133371353,
0.03957812488079071,
-0.018946610391139984,
-0.10468810051679611,
-0.005377679131925106,
0.04377446696162224,
0.015855854377150536,
-0.05301276221871376,
0.05285383760929108,
-0.14851123094558716,
0.057932693511247635,
-0.01761786639690399,
-0.023627815768122673,
0.028299760073423386,
0.015853866934776306,
-0.058477893471717834,
0.05099276453256607,
0.10850762575864792,
0.018512504175305367,
0.008005158975720406,
-0.04345838353037834,
-0.01174043957144022,
-0.04584210366010666,
-0.07062546163797379,
-0.13404878973960876,
-0.15297792851924896,
-0.11811993271112442,
0.036091700196266174,
0.00039091933285817504,
-0.1965653896331787,
-0.010247079655528069,
-0.09248000383377075,
0.06090386584401131,
-0.16053813695907593,
0.10050295293331146,
0.04856192693114281,
0.03159687668085098,
-0.004500389099121094,
-0.04172549024224281,
0.04549026116728783,
0.10930925607681274,
-0.12740783393383026,
-0.0826936885714531
] |
null | null | null |
<a href="https://www.geogebra.org/m/w8uzjttg">.</a>
<a href="https://www.geogebra.org/m/gvn7m78g">.</a>
<a href="https://www.geogebra.org/m/arxecanq">.</a>
<a href="https://www.geogebra.org/m/xb69bvww">.</a>
<a href="https://www.geogebra.org/m/apvepfnd">.</a>
<a href="https://www.geogebra.org/m/evmj8ckk">.</a>
<a href="https://www.geogebra.org/m/qxcxwmhp">.</a>
<a href="https://www.geogebra.org/m/p3cxqh6c">.</a>
<a href="https://www.geogebra.org/m/ggrahbgd">.</a>
<a href="https://www.geogebra.org/m/pnhymrbc">.</a>
<a href="https://www.geogebra.org/m/zjukbtk9">.</a>
<a href="https://www.geogebra.org/m/bbezun8r">.</a>
<a href="https://www.geogebra.org/m/sgwamtru">.</a>
<a href="https://www.geogebra.org/m/fpunkxxp">.</a>
<a href="https://www.geogebra.org/m/acxebrr7">.</a>
<a href="https://jobs.acm.org/jobs/watch-godzilla-vs-kong-2021-full-1818658-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-godzilla-vs-kong-online-2021-full-f-r-e-e-1818655-cd">.</a>
<a href="https://jobs.acm.org/jobs/watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-f-u-l-l-f-r-e-e-1818661-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-zack-snyder-s-justice-league-online-2021-full-f-r-e-e-1818662-cd">.</a>
<a href="https://jobs.acm.org/jobs/hd-watch-godzilla-vs-kong-2021-version-full-hbomax-1818659-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-girl-in-the-basement-online-2021-full-f-r-e-e-1818663-cd">.</a>
<a href="https://jobs.acm.org/jobs/watch-godzilla-vs-kong-2021-f-u-l-l-h-d-1818660-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-billie-eilish-the-world-s-a-little-blurry-2021-f-u-l-l-f-r-e-e-1818666-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-monster-hunter-2020-f-u-l-l-f-r-e-e-1818667-cd">.</a>
<a href="https://jobs.acm.org/jobs/123movies-watch-raya-and-the-last-dragon-2021-f-u-l-l-f-r-e-e-1818669-cd">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-365-days-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-billie-eilish-the-worlds-a-little-blurry-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-cherry-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-coming-2-america-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-demon-slayer-kimetsu-no-yaiba-mugen-train-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-godzilla-vs-kong-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-judas-and-the-black-messiah-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-monster-hunter-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-mortal-kombat-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-raya-and-the-last-dragon-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-tenet-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-the-world-to-come-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-tom-and-jerry-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-willys-wonderland-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-wonder-woman-1984-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-wrong-turn-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-zack-snyders-justice-league-2021-hd-online-full-free-stream-2/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-a-writers-odyssey-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-the-marksman-2021-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-after-we-collided-2020-version-full-online-free/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-watch-full/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online-full-version-123movies/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-2/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-3/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-4/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-123movies-godzilla-vs-kong-2021/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-free-hd/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free-online/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-5/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online-full-version-hd/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-full-2021-free/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-2/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-6/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-7/">.</a>
<a href="https://pactforanimals.org/advert/free-download-godzilla-vs-kong-2021-watch-full/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-online/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-online/">.</a>
<a href="https://pactforanimals.org/advert/godzilla-vs-kong-2021-google-drive-mp4/">.</a>
<a href="https://pactforanimals.org/advert/google-docs-godzilla-vs-kong-2021-google-drive-full-hd-mp4/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-8/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-9/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-3/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-online/">.</a>
<a href="https://pactforanimals.org/advert/free-watch-godzilla-vs-kong-2021-full-4/">.</a>
<a href="https://pactforanimals.org/advert/free-godzilla-vs-kong-2021-watch-full/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-10/">.</a>
<a href="https://pactforanimals.org/advert/online-watch-godzilla-vs-kong-2021-full/">.</a>
<a href="https://pactforanimals.org/advert/123movies-watch-godzilla-vs-kong-2021-full-online/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-full-11/">.</a>
<a href="https://pactforanimals.org/advert/full-watch-godzilla-vs-kong-2021-free-hd/">.</a>
<a href="https://pactforanimals.org/advert/watch-godzilla-vs-kong-2021-free-online/">.</a>
<a href="https://pactforanimals.org/advert/full-godzilla-vs-kong-2021-watch-online/">.</a>
<a href="https://sites.google.com/view/mortalkombat1/">.</a>
<a href="https://sites.google.com/view/free-watch-mortal-kombat-2021-/">.</a>
<a href="https://sites.google.com/view/watch-mortal-kombat-2021-f-u-l/">.</a>
<a href="https://sites.google.com/view/mortalkombat2/">.</a>
<a href="https://sites.google.com/view/mortalkombat3/">.</a>
<a href="https://sites.google.com/view/mortalkombat5/">.</a>
<a href="https://sites.google.com/view/fullwatchmortalkombat2021-movi/">.</a>
<a href="https://sites.google.com/view/mortalkombat7/">.</a>
<a href="https://sites.google.com/view/mortalkombat8/">.</a>
<a href="https://sites.google.com/view/mortalkombat9/">.</a>
<a href="https://sites.google.com/view/mortalkombat10/">.</a>
<a href="https://sites.google.com/view/watch-mort-tal-kombat/">.</a>
<a href="https://sites.google.com/view/free-watch-mort-tal-kombat/">.</a>
<a href="https://sites.google.com/view/watch-mort-tal-kombatfree-/">.</a>
<a href="https://sites.google.com/view/full-watch-mortal-kombat/">.</a>
<a href="https://sites.google.com/view/watch-mortal-kombat-2021-/">.</a>
<a href="https://sites.google.com/view/watch-free-mortal-kombat-2021/">.</a>
<a href="https://sites.google.com/view/full-watch-mortal-kombat-/">.</a>
<a href="https://sites.google.com/view/watch-mortal-kombat-g-drive/">.</a>
<a href="https://sites.google.com/view/g-docs-mortalkombat-g-drive/">.</a>
<a href="https://sites.google.com/view/mortal-kombat-2021-full-free/">.</a>
<a href="https://sites.google.com/view/mortal-kombat-2021-full-free-o/">.</a>
<a href="https://sites.google.com/view/mortal-kombat-2021-full-free-o/">.</a>
<a href="https://paiza.io/projects/56xFAEq61pSSn8VnKnHO6Q">.</a>
<a href="https://www.posts123.com/post/1450667/mariners-announce-spring-training">.</a>
<a href="https://sites.google.com/view/sfdjgkdfghdkfgjherghkkdfjg/home">.</a>
<a href="https://dskfjshdkjfewhgf.blogspot.com/2021/03/sdkjfhwekjhfjdherjgfdjg.html">.</a>
<a href="https://grahmaulidia.wordpress.com/2021/03/28/mariners-announce-spring-training-roster-moves/">.</a>
<a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner-f83a9ea92f89">.</a>
<a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner1-b2847091ff9f">.</a>
<a href="https://4z5v6wq7a.medium.com/a-letter-to-nationals-fans-from-mark-d-lerner2-df35041eec3a">.</a>
<a href="https://4z5v6wq7a.medium.com">.</a>
<a href="https://onlinegdb.com/BJaH8WR4O">.</a>
|
{}
| null |
dispenst/hgfytgfg
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL
<a href="URL">.</a>
<a href="URL
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null |
transformers
|
We took `facebook/wav2vec2-large-960h` and fine tuned it using 1400 audio clips (around 10-15 seconds each) from various cryptocurrency related podcasts. To label the data, we downloaded cryptocurrency podcasts from youtube with their subtitle data and split the clips up by sentence. We then compared the youtube transcription with `facebook/wav2vec2-large-960h` to correct many mistakes in the youtube transcriptions. We can probably achieve better results with more data clean up.
On our data we achieved a WER of 13.1%. `facebook/wav2vec2-large-960h` only reached a WER of 27% on our data.
## Usage
```python
from transformers import Wav2Vec2Processor, Wav2Vec2ForCTC
from datasets import load_dataset
import soundfile as sf
import torch
# load model and tokenizer
processor = Wav2Vec2Processor.from_pretrained("distractedm1nd/wav2vec-en-finetuned-on-cryptocurrency")
model = Wav2Vec2ForCTC.from_pretrained("distractedm1nd/wav2vec-en-finetuned-on-cryptocurrency"
filename = "INSERT_FILENAME"
audio, sampling_rate = sf.read(filename)
input_values = processor(audio, return_tensors="pt", padding="longest", sampling_rate=sampling_rate).input_values # Batch size 1
# retrieve logits
logits = model(input_values).logits
# take argmax and decode
predicted_ids = torch.argmax(logits, dim=-1)
tokenizer.batch_decode(predicted_ids
```
|
{"language": "en", "license": "mit", "tags": ["audio", "automatic-speech-recognition"], "metrics": ["wer"]}
|
automatic-speech-recognition
|
distractedm1nd/wav2vec-en-finetuned-on-cryptocurrency
|
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"audio",
"en",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #en #license-mit #endpoints_compatible #region-us
|
We took 'facebook/wav2vec2-large-960h' and fine tuned it using 1400 audio clips (around 10-15 seconds each) from various cryptocurrency related podcasts. To label the data, we downloaded cryptocurrency podcasts from youtube with their subtitle data and split the clips up by sentence. We then compared the youtube transcription with 'facebook/wav2vec2-large-960h' to correct many mistakes in the youtube transcriptions. We can probably achieve better results with more data clean up.
On our data we achieved a WER of 13.1%. 'facebook/wav2vec2-large-960h' only reached a WER of 27% on our data.
## Usage
|
[
"## Usage"
] |
[
"TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #en #license-mit #endpoints_compatible #region-us \n",
"## Usage"
] |
[
47,
3
] |
[
"passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #audio #en #license-mit #endpoints_compatible #region-us \n## Usage"
] |
[
-0.056982893496751785,
0.0005500153638422489,
-0.006455869413912296,
-0.07534864544868469,
0.0787082239985466,
-0.05580902099609375,
0.09732449799776077,
0.06740663200616837,
0.05411749333143234,
-0.008937194012105465,
0.06890277564525604,
0.2206738144159317,
0.000864176545292139,
-0.04299534112215042,
-0.08491131663322449,
-0.20582479238510132,
0.07153806835412979,
0.057264938950538635,
0.10262307524681091,
0.12296371906995773,
0.09796105325222015,
-0.05732206627726555,
0.0004630916519090533,
0.08023642748594284,
-0.11496791243553162,
0.013674282468855381,
0.086811363697052,
-0.14566652476787567,
0.09898962080478668,
0.040699951350688934,
0.06428813189268112,
0.056339845061302185,
0.04066414386034012,
-0.21681927144527435,
0.006040611770004034,
-0.045702289789915085,
0.000011036586329282727,
0.00004067078771186061,
0.0778699517250061,
-0.036663129925727844,
0.006821466144174337,
0.11684460937976837,
-0.014098238199949265,
0.08589097112417221,
-0.05880676209926605,
-0.196751207113266,
-0.013201196677982807,
0.034006133675575256,
0.0503554604947567,
0.08267393708229065,
-0.037048645317554474,
0.11347967386245728,
-0.10439562052488327,
0.08700432628393173,
0.09813293814659119,
-0.2951190173625946,
0.046979013830423355,
0.00819287821650505,
0.14936286211013794,
0.002108729910105467,
-0.051671527326107025,
0.09306912869215012,
0.048090510070323944,
0.027288198471069336,
-0.04462110623717308,
-0.062038496136665344,
-0.13009555637836456,
-0.004890039097517729,
-0.08030109852552414,
-0.0690305233001709,
0.19982507824897766,
-0.02161616086959839,
0.0019737768452614546,
-0.08333662897348404,
-0.019132068380713463,
-0.04620983079075813,
-0.03074909746646881,
-0.03400732949376106,
-0.014808418229222298,
0.07910087704658508,
-0.06099779158830643,
-0.0032848594710230827,
-0.10207551717758179,
-0.05934717878699303,
-0.16526590287685394,
0.187605082988739,
0.006493586115539074,
0.058220669627189636,
-0.13474933803081512,
0.004658169578760862,
-0.04008232057094574,
-0.06865862011909485,
0.00024177519662771374,
-0.018047748133540154,
0.049289435148239136,
0.0008454094058834016,
-0.05970587953925133,
-0.0002078320103464648,
0.09711094945669174,
0.07486845552921295,
0.02328477054834366,
0.019772525876760483,
-0.10022737085819244,
0.12469621747732162,
0.018721815198659897,
0.06590913236141205,
0.006644950248301029,
-0.0327606163918972,
0.026825137436389923,
-0.14036400616168976,
0.029898472130298615,
-0.04207274317741394,
-0.13409680128097534,
-0.06768003106117249,
0.008405466563999653,
0.08296816796064377,
0.012535822577774525,
0.0008496535592712462,
-0.07480575144290924,
-0.013670667074620724,
0.0048980736173689365,
-0.06899623572826385,
0.002874139929190278,
0.04147796705365181,
0.11721459776163101,
0.1602868139743805,
0.0009608506225049496,
0.08372381329536438,
-0.07488535344600677,
0.09963062405586243,
0.027259670197963715,
0.016759155318140984,
0.03838971629738808,
-0.015151807107031345,
0.054203812032938004,
-0.08798032253980637,
0.06060653179883957,
-0.1376463621854782,
-0.037032656371593475,
-0.003499518381431699,
-0.015584711916744709,
0.02990519255399704,
-0.09233004599809647,
-0.04519354924559593,
-0.018276162445545197,
0.02533174306154251,
-0.12164487689733505,
-0.0011901991674676538,
-0.08783669024705887,
0.15201084315776825,
-0.03993915766477585,
0.06564951688051224,
-0.13250473141670227,
0.09684691578149796,
-0.0486300066113472,
-0.004434970673173666,
0.02296140044927597,
0.048416346311569214,
-0.03402271494269371,
0.02209503762423992,
-0.05579492449760437,
-0.0799829289317131,
-0.12051320821046829,
0.07323747128248215,
-0.07375933229923248,
0.06814666837453842,
-0.13460920751094818,
-0.13714472949504852,
0.15899039804935455,
-0.1257573366165161,
-0.08942605555057526,
0.1296851485967636,
0.033979251980781555,
0.005960570648312569,
0.08412745594978333,
0.27362439036369324,
0.02493925951421261,
-0.2198563814163208,
0.07685350626707077,
0.09523825347423553,
-0.12458018958568573,
-0.07735081017017365,
0.0580030158162117,
-0.10122958570718765,
-0.043508049100637436,
0.00048709026305004954,
0.0372299998998642,
0.06553730368614197,
-0.040257778018713,
-0.07018569111824036,
-0.016134122386574745,
-0.06489017605781555,
0.0375942699611187,
-0.026267508044838905,
0.026770759373903275,
-0.033781833946704865,
-0.003887225640937686,
0.014223875477910042,
0.019818365573883057,
-0.022937340661883354,
0.08044438064098358,
-0.15480369329452515,
0.14003054797649384,
-0.02489437721669674,
-0.012788396328687668,
-0.14888915419578552,
0.21457259356975555,
-0.03858675807714462,
0.05138394609093666,
0.09117254614830017,
0.08748919516801834,
0.04753017798066139,
-0.06171673163771629,
0.00675531942397356,
-0.04282431676983833,
0.14161664247512817,
0.04737022519111633,
0.019028691574931145,
-0.12019111216068268,
0.057965707033872604,
-0.05158497765660286,
-0.07742830365896225,
0.05194681137800217,
-0.07463640719652176,
0.12178099900484085,
0.0824984461069107,
-0.026179278269410133,
0.020972980186343193,
-0.010850182734429836,
0.04222835600376129,
-0.01683451421558857,
0.06045910343527794,
0.08701625466346741,
0.014016932807862759,
-0.035395774990320206,
0.2517566680908203,
-0.15099245309829712,
0.2921793460845947,
0.24359184503555298,
-0.2683594524860382,
0.06151987984776497,
0.04924337565898895,
0.022027358412742615,
-0.01385312620550394,
0.020041845738887787,
-0.05331169813871384,
0.22052767872810364,
-0.013359571807086468,
0.12220925837755203,
-0.06756936013698578,
-0.0051261307671666145,
0.04949088394641876,
-0.053889915347099304,
-0.044458214193582535,
-0.012626687996089458,
0.03162970393896103,
-0.09746769815683365,
0.11824654042720795,
0.2292405217885971,
-0.0506669357419014,
0.23144371807575226,
-0.05330917239189148,
-0.0413082130253315,
0.04128266125917435,
-0.008965978398919106,
-0.05754660815000534,
0.04331457242369652,
-0.30421182513237,
-0.008594553917646408,
0.08451247960329056,
-0.0022996265906840563,
0.08095822483301163,
-0.14743919670581818,
-0.01588035747408867,
-0.005628102459013462,
-0.05800161138176918,
-0.15047527849674225,
0.06134943291544914,
0.014128301292657852,
0.03823384270071983,
-0.05324120819568634,
-0.13306111097335815,
0.11769227683544159,
-0.03960363194346428,
-0.11004022508859634,
0.02674221061170101,
-0.13855081796646118,
-0.21859443187713623,
-0.16430233418941498,
-0.1473323255777359,
0.0417647510766983,
0.05621703341603279,
0.159214586019516,
-0.05098174512386322,
-0.011427296325564384,
0.07718748599290848,
0.0460035614669323,
-0.09276732802391052,
-0.006296237464994192,
0.016812477260828018,
0.04595083370804787,
-0.0009306219290010631,
-0.1069149300456047,
-0.01808904856443405,
-0.009776080027222633,
0.0006099265883676708,
0.05149955675005913,
-0.0011221214663237333,
0.06646858155727386,
0.19807280600070953,
0.07422003895044327,
0.04323996603488922,
-0.01760232262313366,
0.13387340307235718,
-0.11297314614057541,
-0.0935530811548233,
0.16344712674617767,
-0.10213212668895721,
0.013820470310747623,
0.21018086373806,
0.05971488729119301,
-0.0342375785112381,
-0.05933605134487152,
-0.05998425930738449,
-0.0949932187795639,
-0.226863294839859,
-0.13328993320465088,
-0.10603221505880356,
-0.016781147569417953,
-0.027210021391510963,
0.03791311755776405,
0.14160455763339996,
0.0050741382874548435,
-0.0023794472217559814,
-0.08075962215662003,
0.07323215901851654,
0.01732408069074154,
0.271053671836853,
-0.06614259630441666,
0.11493358016014099,
-0.04268674552440643,
-0.09276910126209259,
0.05671072378754616,
0.06739463657140732,
0.09419871866703033,
0.219945028424263,
0.05920424312353134,
0.0800153911113739,
0.11550432443618774,
0.17751656472682953,
0.10539475083351135,
0.054237816482782364,
-0.0014176388503983617,
0.01308255922049284,
-0.06526187062263489,
-0.0370648130774498,
0.05993170291185379,
0.25215861201286316,
-0.1139591857790947,
0.013302231207489967,
-0.17947503924369812,
0.018698517233133316,
0.14789053797721863,
0.05261624604463577,
-0.14975988864898682,
0.03309597074985504,
0.07309319823980331,
-0.06367955356836319,
-0.061093058437108994,
0.1713261753320694,
0.035330433398485184,
-0.08338221907615662,
0.1038026437163353,
0.039673540741205215,
0.07962498813867569,
-0.04747219383716583,
0.07024187594652176,
-0.029891345649957657,
-0.15011851489543915,
0.059935908764600754,
0.05275629088282585,
-0.24189479649066925,
0.22806033492088318,
0.0022859391756355762,
0.06256621330976486,
-0.029207266867160797,
0.0014281697804108262,
0.0007133919862098992,
0.08426407724618912,
0.17776666581630707,
-0.002998177194967866,
-0.09764339029788971,
-0.09242641180753708,
0.00023517206136602908,
0.01708294451236725,
0.13973601162433624,
0.1135769709944725,
-0.07230179011821747,
-0.04765412211418152,
-0.02974821627140045,
0.01573970727622509,
-0.07285749167203903,
-0.10238493978977203,
-0.10104276239871979,
0.015504788607358932,
0.20649823546409607,
0.16848979890346527,
-0.023720575496554375,
-0.051649805158376694,
-0.18702495098114014,
0.08738412708044052,
-0.1941281259059906,
-0.013123316690325737,
-0.0496077686548233,
-0.2288239449262619,
0.08855639398097992,
-0.0017343980725854635,
0.08900512009859085,
-0.008712752722203732,
0.013787414878606796,
-0.09566756337881088,
-0.10604628175497055,
0.133510559797287,
-0.09585755318403244,
-0.042672451585531235,
-0.04616539925336838,
0.28026682138442993,
-0.01978798769414425,
0.051491573452949524,
0.06826116889715195,
0.029016366228461266,
-0.02637595124542713,
-0.05434431880712509,
0.07622208446264267,
0.04260272532701492,
-0.06567803770303726,
-0.010689682327210903,
0.004118646029382944,
-0.19869926571846008,
-0.023355631157755852,
-0.032363321632146835,
0.2377147376537323,
0.1174280047416687,
-0.03651181980967522,
0.22753407061100006,
0.2568420171737671,
-0.04636237397789955,
-0.2953130602836609,
-0.1558300256729126,
-0.05332402512431145,
-0.011951344087719917,
-0.03287419304251671,
-0.11996402591466904,
0.08021412789821625,
-0.06646183133125305,
-0.08367260545492172,
0.06856631487607956,
-0.19254468381404877,
-0.0979730635881424,
0.24247236549854279,
-0.08899877965450287,
0.29266226291656494,
-0.05581675469875336,
-0.13907550275325775,
-0.0342436358332634,
-0.22308349609375,
0.10077599436044693,
-0.042832646518945694,
0.0858520120382309,
0.062011200934648514,
0.109755739569664,
0.036504052579402924,
-0.04580417275428772,
0.09441400319337845,
0.05430116504430771,
-0.03571886569261551,
-0.04721146076917648,
-0.024953778833150864,
0.028031611815094948,
0.0697186142206192,
-0.012136443518102169,
-0.040463756769895554,
0.002782433293759823,
-0.006253904663026333,
-0.08042892068624496,
-0.10044530779123306,
0.10602186620235443,
0.032961875200271606,
-0.06018221005797386,
0.026256224140524864,
-0.09562192112207413,
-0.02357880026102066,
-0.009046639315783978,
0.1884971559047699,
-0.09985484927892685,
0.0641416683793068,
0.186880424618721,
0.18821489810943604,
-0.13257302343845367,
-0.09986072033643723,
-0.07026804983615875,
-0.1270354986190796,
0.12483903765678406,
0.02035241574048996,
0.06458645313978195,
0.07085448503494263,
0.04869062080979347,
0.03704468533396721,
0.07386071234941483,
-0.0431428998708725,
0.02969961427152157,
0.10032568126916885,
-0.1028645858168602,
-0.07303258776664734,
-0.013600850477814674,
0.054332882165908813,
0.16123835742473602,
0.021845614537596703,
0.11897840350866318,
0.01158690545707941,
-0.007932944223284721,
-0.002542646834626794,
-0.013030131347477436,
-0.17870654165744781,
0.10704424977302551,
0.028012361377477646,
0.024207618087530136,
-0.17638739943504333,
0.05085553601384163,
-0.042246393859386444,
-0.2028125822544098,
-0.02132919803261757,
-0.03727080300450325,
-0.08149966597557068,
-0.0948934257030487,
-0.08908620476722717,
0.0009530898532830179,
-0.07253321260213852,
-0.12588204443454742,
0.06455069780349731,
-0.18822860717773438,
0.04604039341211319,
0.1583443433046341,
0.04482726380228996,
0.1015201210975647,
-0.09997063130140305,
-0.029454367235302925,
0.015600924380123615,
0.008118053898215294,
-0.05564739927649498,
-0.003854242153465748,
-0.15392348170280457,
0.011613987386226654,
0.019857684150338173,
0.06554519385099411,
-0.09565191715955734,
-0.07209020853042603,
-0.06456682085990906,
0.09584588557481766,
-0.12700068950653076,
-0.019015895202755928,
-0.10052744299173355,
-0.015502072870731354,
0.043772902339696884,
-0.05867733433842659,
-0.03915667533874512,
0.02087940089404583,
-0.1039939895272255,
0.038452934473752975,
0.039862021803855896,
0.060360975563526154,
-0.10262532532215118,
0.004351455252617598,
0.07861801981925964,
-0.03427136316895485,
0.10110075026750565,
0.22048166394233704,
-0.1430530697107315,
0.0974893867969513,
-0.24644456803798676,
-0.19499172270298004,
0.17390945553779602,
0.017983295023441315,
-0.03171376883983612,
-0.04605744406580925,
0.01339606661349535,
0.1428636610507965,
0.014025820419192314,
0.009642817080020905,
0.08296423405408859,
-0.0882575511932373,
0.019686631858348846,
-0.054349564015865326,
-0.10088244825601578,
-0.021692898124456406,
-0.07468029856681824,
0.1702674925327301,
0.04821905493736267,
0.15053541958332062,
-0.03385153412818909,
0.03419724106788635,
0.018259111791849136,
0.05350610986351967,
-0.043514180928468704,
-0.11890412122011185,
-0.13701693713665009,
-0.053773947060108185,
-0.013040502555668354,
-0.04744702950119972,
0.2533976137638092,
-0.05795205011963844,
-0.06948486715555191,
0.04004408419132233,
0.04635799303650856,
-0.09763108938932419,
0.010247079655528069,
0.3549251854419708,
0.06211170554161072,
-0.01750459149479866,
-0.12229972332715988,
-0.017638636752963066,
0.002132441382855177,
-0.0026729877572506666,
-0.06279987096786499,
0.17240487039089203,
0.024032076820731163,
0.060140568763017654,
0.13227121531963348,
-0.01705201156437397,
-0.17400658130645752,
-0.03923087194561958,
-0.0054335943423211575,
0.06409290432929993,
-0.027834711596369743,
0.17212684452533722,
0.13743136823177338,
-0.010404304601252079,
0.0506969653069973,
-0.04019174352288246,
-0.00833127647638321,
-0.1952020525932312,
-0.08877056837081909,
-0.07430858910083771,
-0.12036347389221191,
0.05795121192932129,
-0.03083161450922489,
0.1172495037317276,
0.01864440366625786,
0.0282160472124815,
-0.052042677998542786,
0.061372485011816025,
-0.08547540009021759,
-0.09568807482719421,
0.07108817994594574,
-0.026931606233119965,
-0.005623408127576113,
-0.06987880915403366,
-0.02824706956744194,
0.043778084218502045,
-0.0487457737326622,
-0.005991113372147083,
0.016472013667225838,
-0.10153807699680328,
-0.004362825769931078,
-0.13723622262477875,
-0.07287703454494476,
-0.015598959289491177,
0.010201222263276577,
0.035938359797000885,
0.17393629252910614,
0.050197746604681015,
-0.02095334231853485,
0.05370447784662247,
0.10888709127902985,
-0.09315739572048187,
-0.19248110055923462,
-0.0058380356058478355,
0.11689832806587219,
-0.0049933441914618015,
0.13635355234146118,
0.0078093199990689754,
-0.006159522570669651,
-0.06787469983100891,
0.24738290905952454,
0.24052076041698456,
-0.005120075307786465,
0.0607389472424984,
-0.007302202749997377,
0.02894405834376812,
0.0037268432788550854,
-0.010033214464783669,
0.10794877260923386,
0.30342304706573486,
-0.01737513206899166,
-0.04459936544299126,
-0.08471204340457916,
-0.026130151003599167,
-0.037580471485853195,
0.0778174102306366,
-0.06102128326892853,
-0.16518551111221313,
-0.00849368330091238,
0.11384115368127823,
-0.22842906415462494,
0.06818121671676636,
-0.05583729222416878,
-0.10818830877542496,
0.0007225004374049604,
0.004752101842314005,
0.15980908274650574,
0.10881918668746948,
0.0098530612885952,
-0.0557539202272892,
-0.1108919158577919,
0.0560685470700264,
0.013460665941238403,
-0.2421426773071289,
0.07847000658512115,
-0.004179149400442839,
-0.04318339005112648,
0.009319937787950039,
0.02006000094115734,
0.1528831571340561,
-0.026256632059812546,
0.18736183643341064,
0.030298734083771706,
0.1635451465845108,
0.037098079919815063,
-0.17271463572978973,
-0.020546142011880875,
0.12193199247121811,
-0.01207912340760231,
0.0487968735396862,
0.03514057397842407,
-0.1492740511894226,
0.07935703545808792,
-0.06682530045509338,
-0.084715835750103,
-0.07657837122678757,
0.0062805050984025,
-0.06753505021333694,
0.04057042673230171,
-0.09385112673044205,
0.01288489531725645,
-0.0593428872525692,
-0.03252794221043587,
0.03967435657978058,
0.08778904378414154,
-0.10036205500364304,
-0.05244796350598335,
-0.15157301723957062,
-0.050583481788635254,
-0.028836822137236595,
-0.00559743819758296,
-0.11467599123716354,
-0.027552327141165733,
-0.06804998219013214,
0.04382109269499779,
-0.11768227070569992,
-0.0018709696596488357,
0.07519640028476715,
-0.024836286902427673,
0.043647147715091705,
-0.05378871411085129,
0.10930842161178589,
0.08113029599189758,
-0.1257760226726532,
-0.070065438747406
] |
null | null | null |
# Peter from Your Boyfriend Game.
|
{"tags": ["conversational"]}
|
text-generation
|
divi/Peterbot
|
[
"conversational",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#conversational #region-us
|
# Peter from Your Boyfriend Game.
|
[
"# Peter from Your Boyfriend Game."
] |
[
"TAGS\n#conversational #region-us \n",
"# Peter from Your Boyfriend Game."
] |
[
10,
8
] |
[
"passage: TAGS\n#conversational #region-us \n# Peter from Your Boyfriend Game."
] |
[
0.06918997317552567,
0.1119099110364914,
-0.009369183331727982,
-0.007348303217440844,
0.1114223450422287,
-0.006091204471886158,
0.09987349063158035,
0.015708819031715393,
0.10811413079500198,
0.05851687490940094,
0.12548041343688965,
0.0026048191357403994,
-0.012053134851157665,
-0.07667381316423416,
0.07856566458940506,
-0.23317857086658478,
-0.028593478724360466,
0.001928992453031242,
0.05752907693386078,
0.07248014956712723,
-0.05803684890270233,
-0.06304407864809036,
0.07649388909339905,
-0.006955263204872608,
0.012429457157850266,
0.11318182945251465,
0.03838834911584854,
0.02768545038998127,
0.1237233579158783,
-0.06042127311229706,
0.0850546658039093,
0.0021823234856128693,
-0.13338178396224976,
-0.1982526183128357,
0.047229841351509094,
0.018510855734348297,
0.03943606838583946,
-0.00015575946599710733,
0.12003330886363983,
-0.005420854315161705,
0.026410292834043503,
0.13258880376815796,
0.040595222264528275,
0.09878802299499512,
-0.21257266402244568,
-0.13866624236106873,
0.01641935296356678,
-0.06585172563791275,
0.03940219059586525,
0.04895725101232529,
-0.01987452618777752,
0.15597720444202423,
-0.037437085062265396,
-0.04071600362658501,
0.11838050186634064,
-0.3682670593261719,
-0.027074895799160004,
0.08072665333747864,
0.045348260551691055,
0.18773740530014038,
-0.03230864182114601,
0.030310524627566338,
-0.0010748646454885602,
-0.000745356606785208,
-0.14505726099014282,
-0.03196580708026886,
0.06572649627923965,
0.19920095801353455,
-0.03734710067510605,
-0.05706467479467392,
0.2445371299982071,
0.019581269472837448,
0.09135741740465164,
-0.0046166772954165936,
-0.017617421224713326,
0.020251067355275154,
0.027826571837067604,
-0.12460268288850784,
-0.019448529928922653,
0.09155324101448059,
-0.0943152904510498,
-0.09364258497953415,
-0.15981410443782806,
0.06662730127573013,
-0.14773136377334595,
-0.02820645645260811,
-0.03203616663813591,
0.0945311039686203,
-0.23288783431053162,
-0.07379235327243805,
-0.13301634788513184,
-0.032893240451812744,
0.11477778106927872,
-0.016381897032260895,
0.055063702166080475,
0.002403243910521269,
0.004227844066917896,
0.017587032169103622,
0.12600181996822357,
0.020195189863443375,
0.12886236608028412,
0.026812629774212837,
-0.04383889213204384,
0.06410515308380127,
0.12806499004364014,
0.025331290438771248,
0.1259302943944931,
-0.16823044419288635,
-0.059482987970113754,
-0.09197785705327988,
0.012903711758553982,
-0.07124218344688416,
-0.1708669811487198,
0.1313035637140274,
-0.051498375833034515,
0.005790825001895428,
0.05548316612839699,
0.010863265953958035,
0.017198579385876656,
0.009574162773787975,
0.11043789237737656,
0.03247561305761337,
-0.06024952977895737,
0.039217785000801086,
0.03875354304909706,
0.10397803038358688,
-0.231271892786026,
0.052631404250860214,
0.07532041519880295,
0.022120332345366478,
-0.09254033863544464,
-0.1369757205247879,
0.017887121066451073,
0.005603509955108166,
0.07072160392999649,
-0.007546141277998686,
0.022739021107554436,
-0.0945245772600174,
0.08002761006355286,
-0.015468678437173367,
-0.03581824526190758,
-0.004916756879538298,
-0.022272184491157532,
-0.11783408373594284,
0.104704350233078,
-0.1327126920223236,
0.011779054068028927,
-0.09666518867015839,
-0.013646906241774559,
0.11022164672613144,
0.0823843777179718,
0.13119398057460785,
-0.08095869421958923,
0.04512627050280571,
-0.05900753661990166,
0.020704301074147224,
-0.06404738128185272,
-0.041981056332588196,
-0.0170572642236948,
-0.020718378946185112,
-0.007581941317766905,
0.07831355929374695,
-0.10813421756029129,
0.040933676064014435,
-0.09741310775279999,
0.06228204071521759,
-0.05412999540567398,
-0.12577082216739655,
0.2344287931919098,
-0.07573767006397247,
-0.1479921191930771,
0.013087565079331398,
-0.0443989597260952,
0.025551432743668556,
0.15141573548316956,
0.3345992863178253,
-0.22492185235023499,
-0.11447372287511826,
0.014249686151742935,
0.20443250238895416,
-0.1560593694448471,
0.07643327862024307,
0.14136335253715515,
-0.05908569321036339,
-0.03764314576983452,
0.047559574246406555,
-0.005986492615193129,
0.06039522960782051,
-0.028110338374972343,
-0.04880061373114586,
0.16775618493556976,
-0.04933122545480728,
0.06184576824307442,
0.021685337647795677,
0.05615251883864403,
-0.10392250865697861,
0.07145784050226212,
0.1507585644721985,
0.03248937055468559,
0.10424719005823135,
0.01961403712630272,
-0.050918277353048325,
0.10437493771314621,
0.03348409757018089,
0.0014617022825405002,
-0.05548873171210289,
-0.1890576034784317,
-0.11205189675092697,
0.2760697603225708,
0.07032963633537292,
0.12072832882404327,
0.09766019135713577,
0.005291177425533533,
0.01421473827213049,
0.03645877540111542,
0.10666606575250626,
0.03995051234960556,
0.023000093176960945,
-0.029076945036649704,
0.06977929174900055,
-0.07083173096179962,
0.03944772109389305,
-0.08755996823310852,
-0.06107444688677788,
0.018847698345780373,
0.020750634372234344,
0.0728275403380394,
-0.04420776665210724,
-0.08121079951524734,
0.005519713740795851,
0.07728409022092819,
-0.053619131445884705,
0.09964504837989807,
0.02073475904762745,
0.008843914605677128,
0.14675451815128326,
-0.02896558679640293,
0.04911911487579346,
0.16360998153686523,
-0.3147040605545044,
0.05646326392889023,
-0.1068376824259758,
-0.015705948695540428,
-0.017179453745484352,
0.09965228289365768,
-0.061173807829618454,
0.10185930877923965,
0.06329537183046341,
0.029972530901432037,
-0.002652270719408989,
0.0017106939340010285,
-0.016247225925326347,
-0.015169442631304264,
-0.03684157878160477,
0.12628716230392456,
-0.03323010355234146,
-0.09194064140319824,
0.2179882973432541,
0.1872261017560959,
0.19533881545066833,
0.14744748175144196,
0.009054310619831085,
-0.026751073077321053,
-0.031248915940523148,
0.003400458488613367,
-0.06596767157316208,
0.03831149637699127,
-0.24136334657669067,
-0.024507302790880203,
0.020699264481663704,
-0.10422341525554657,
0.09276845306158066,
-0.11853627860546112,
-0.0966867133975029,
0.05060449615120888,
0.07616781443357468,
-0.16892996430397034,
0.1344001144170761,
-0.06780379265546799,
0.005290485918521881,
0.11902663856744766,
0.05704253539443016,
0.11324138939380646,
-0.0009333572816103697,
-0.07601576298475266,
-0.07546330988407135,
-0.17526908218860626,
-0.14939875900745392,
-0.059153709560632706,
-0.12873247265815735,
-0.0330955907702446,
0.057645704597234726,
0.045807208865880966,
-0.14190669357776642,
0.04648267850279808,
0.09232684969902039,
-0.00385842751711607,
-0.23955588042736053,
-0.14731436967849731,
-0.14652450382709503,
0.0412168949842453,
-0.20743179321289062,
0.0002695592993404716,
0.029098190367221832,
-0.07610329985618591,
-0.15040963888168335,
-0.01283708494156599,
-0.1611708700656891,
0.007313175592571497,
0.128167062997818,
0.12477458268404007,
0.09750060737133026,
-0.06569153815507889,
0.11821441352367401,
0.012320085428655148,
-0.09349887818098068,
0.052966002374887466,
-0.06104880943894386,
0.05691598355770111,
-0.033383265137672424,
0.009369516745209694,
0.014982509426772594,
-0.011686411686241627,
-0.011015879921615124,
-0.012301972135901451,
-0.18694426119327545,
0.0016421906184405088,
0.03227807208895683,
0.19669531285762787,
0.02481819875538349,
0.03486744686961174,
0.11832347512245178,
-0.05659681186079979,
0.0935693234205246,
-0.1878378987312317,
-0.00018301470845472068,
0.06223492696881294,
0.25786930322647095,
-0.20212602615356445,
-0.05001058429479599,
-0.002887232694774866,
-0.044520121067762375,
0.11817061901092529,
-0.011949301697313786,
-0.18787597119808197,
0.2525191307067871,
0.1533478945493698,
0.04650315269827843,
0.047132302075624466,
-0.012249743565917015,
-0.09028641879558563,
-0.011456942185759544,
-0.032018423080444336,
-0.10847168415784836,
-0.04889707267284393,
-0.024384450167417526,
0.02679363079369068,
0.10308423638343811,
-0.14844249188899994,
0.06400778889656067,
-0.14033137261867523,
0.020382503047585487,
-0.05483611673116684,
0.009875371120870113,
0.022574407979846,
0.01881299540400505,
0.04900393262505531,
-0.004539458081126213,
-0.057616058737039566,
0.1345737874507904,
0.10591583698987961,
-0.09637884795665741,
-0.12325969338417053,
0.0836876630783081,
0.09661691635847092,
0.05981621891260147,
0.0635344535112381,
-0.08295326679944992,
-0.043592069298028946,
0.024143273010849953,
0.06107378005981445,
-0.12836383283138275,
0.09461197257041931,
0.015583083964884281,
-0.14230738580226898,
-0.0031252356711775064,
-0.1534324586391449,
-0.060186777263879776,
-0.09789077937602997,
0.0614522360265255,
0.06951428204774857,
-0.25506460666656494,
-0.025173692032694817,
-0.0800054594874382,
0.0021261272486299276,
0.08318674564361572,
-0.08323056250810623,
-0.12263412028551102,
-0.004138998221606016,
-0.020692963153123856,
-0.0650867223739624,
0.10809700936079025,
0.1353183388710022,
-0.013880807906389236,
-0.013567137531936169,
0.0785452350974083,
0.053986743092536926,
0.1152510866522789,
-0.02788514643907547,
-0.03592339903116226,
-0.03583177551627159,
0.1724313348531723,
0.04191594198346138,
-0.005160410422831774,
-0.023053353652358055,
-0.15727491676807404,
-0.022540593519806862,
-0.09010462462902069,
-0.024710802361369133,
0.007065996527671814,
-0.1894456297159195,
0.005125025287270546,
0.08471257239580154,
0.03750896826386452,
0.060985203832387924,
-0.006416743621230125,
0.2388896942138672,
0.11739421635866165,
0.03244965150952339,
-0.04152105748653412,
0.04568304866552353,
0.004904903005808592,
0.02314160391688347,
0.12810257077217102,
-0.14965663850307465,
0.061451543122529984,
-0.0936410203576088,
0.2099514901638031,
0.09060542285442352,
-0.10988054424524307,
0.004029345698654652,
0.24920906126499176,
0.14677979052066803,
-0.0005377625930123031,
0.2068527191877365,
0.0597744919359684,
-0.03035743720829487,
-0.1134016290307045,
-0.0849151536822319,
-0.13398222625255585,
-0.1959547996520996,
0.17067664861679077,
-0.11853935569524765,
0.09161555022001266,
-0.07479490339756012,
-0.014806433580815792,
0.20369482040405273,
-0.08258890360593796,
-0.03446623310446739,
0.10472817718982697,
0.017263896763324738,
0.4713169038295746,
-0.0478936992585659,
-0.07194667309522629,
-0.015517424792051315,
-0.23413510620594025,
0.09975261241197586,
0.0676630288362503,
0.017797989770770073,
0.06204335391521454,
0.22496284544467926,
0.009260573424398899,
0.009180773049592972,
0.12102406471967697,
-0.016283659264445305,
-0.10376899689435959,
-0.009381371550261974,
-0.3984149694442749,
0.01624258980154991,
0.02819698303937912,
-0.1832129806280136,
0.05948030948638916,
-0.06336662918329239,
-0.152058944106102,
0.028059495612978935,
-0.07707330584526062,
-0.0881320983171463,
-0.008942041546106339,
0.015728576108813286,
0.020143501460552216,
0.06917974352836609,
-0.08210024237632751,
0.0029605187010020018,
0.07387343049049377,
-0.178839772939682,
0.1296427845954895,
-0.08340680599212646,
-0.04804098606109619,
-0.0025323398876935244,
-0.22769795358181,
-0.11206581443548203,
-0.08866459131240845,
0.08712282031774521,
-0.025766432285308838,
0.03246080130338669,
0.11661338806152344,
0.022954316809773445,
0.005925517063587904,
0.0670008659362793,
-0.1258576512336731,
0.011764544993638992,
0.09087694436311722,
-0.20441865921020508,
-0.11524004489183426,
-0.09665697067975998,
-0.027828184887766838,
0.20122574269771576,
-0.1370730847120285,
0.014090770855545998,
0.06961293518543243,
-0.10673660784959793,
0.10604453086853027,
-0.025344355031847954,
-0.09045412391424179,
-0.17681525647640228,
-0.006270304322242737,
-0.010417696088552475,
-0.09893918037414551,
0.08183314651250839,
0.04370304197072983,
-0.28650158643722534,
-0.04781695455312729,
0.1615978330373764,
0.02639172226190567,
-0.03175157308578491,
-0.059517860412597656,
0.09439253807067871,
0.1377011090517044,
-0.07640109956264496,
0.0844707190990448,
-0.019417138770222664,
-0.024211931973695755,
0.14613808691501617,
0.048912905156612396,
0.05850870907306671,
0.01844676025211811,
0.0679430142045021,
0.09549524635076523,
0.025513984262943268,
-0.0026567389722913504,
-0.07800564169883728,
0.004086449276655912,
-0.038566067814826965,
0.013675248250365257,
0.08527276664972305,
-0.05120961368083954,
-0.15815985202789307,
-0.25214168429374695,
0.00785031821578741,
0.05144155025482178,
-0.060612574219703674,
-0.06875180453062057,
-0.09699482470750809,
0.03919002041220665,
0.014621129259467125,
-0.022572718560695648,
-0.09803934395313263,
-0.06082797423005104,
0.10153742134571075,
0.14398884773254395,
0.03804706409573555,
-0.07196588814258575,
-0.002813495695590973,
0.16583514213562012,
-0.023274969309568405,
0.09425143897533417,
0.201707124710083,
0.024353044107556343,
0.16571162641048431,
-0.017614083364605904,
-0.1705344319343567,
0.01442685630172491,
0.015356134623289108,
0.048283301293849945,
0.13607282936573029,
-0.10658130049705505,
-0.033509429544210434,
0.03625461086630821,
0.015369739383459091,
0.14205506443977356,
-0.03493528440594673,
0.102718785405159,
0.1879558116197586,
-0.18215788900852203,
-0.03890855237841606,
-0.18891161680221558,
0.15210378170013428,
-0.011678364127874374,
0.033512771129608154,
0.037165623158216476,
0.08186427503824234,
-0.0742134153842926,
0.06981709599494934,
0.020384497940540314,
-0.10997700691223145,
0.10997501015663147,
-0.042379651218652725,
-0.012613886035978794,
-0.043028268963098526,
0.16416363418102264,
-0.024432510137557983,
0.038494255393743515,
0.04157954081892967,
0.17811046540737152,
-0.05092015489935875,
-0.006657809484750032,
0.10854634642601013,
0.08583758026361465,
-0.0359041690826416,
-0.10995456576347351,
0.12877343595027924,
-0.060281287878751755,
0.1211765930056572,
0.11293003708124161,
0.02475619874894619,
0.0161672905087471,
0.022434966638684273,
0.03179863840341568,
0.1585141271352768,
-0.022489817813038826,
-0.15959009528160095,
0.08272317051887512,
-0.03296440467238426,
-0.054131414741277695,
0.24122625589370728,
0.08179910480976105,
-0.08099529147148132,
-0.004189319908618927,
-0.1678498387336731,
0.011925142258405685,
-0.11036181449890137,
-0.07292072474956512,
0.0012375772930681705,
-0.153574600815773,
0.03549693524837494,
-0.07848959416151047,
0.010223601013422012,
0.043232012540102005,
0.021918151527643204,
-0.07572861015796661,
-0.11383774131536484,
-0.057027146220207214,
-0.09810028225183487,
-0.071148581802845,
-0.03063969686627388,
0.12719739973545074,
-0.10244853794574738,
-0.0006124336505308747,
-0.022220563143491745,
-0.012034120969474316,
-0.0007963275420479476,
0.02400962822139263,
0.0314701646566391,
-0.051061615347862244,
-0.21802359819412231,
-0.07098805159330368,
-0.03281514719128609,
0.04996378347277641,
-0.07876463979482651,
0.20096321403980255,
-0.020338553935289383,
0.0638837069272995,
0.026039784774184227,
0.151622012257576,
0.02515174075961113,
0.03276528790593147,
-0.028740091249346733,
0.10483493655920029,
-0.18279008567333221,
0.025015657767653465,
-0.04246499389410019,
0.10891539603471756,
-0.10114509612321854,
0.31458812952041626,
0.279488205909729,
-0.0036763453390449286,
-0.039849139750003815,
-0.04752086475491524,
0.07521694153547287,
0.08445192128419876,
0.08938844501972198,
-0.05384856089949608,
0.21015219390392303,
-0.1670716553926468,
-0.06533066183328629,
-0.0527549609541893,
0.07183561474084854,
0.04089202731847763,
-0.0388040617108345,
0.19675245881080627,
-0.03232598677277565,
-0.01555716060101986,
0.11778595298528671,
-0.21236294507980347,
0.05463862046599388,
-0.05130305141210556,
-0.21097499132156372,
-0.01341317966580391,
-0.0651434138417244,
0.1611945778131485,
-0.0077307880856096745,
0.16223563253879547,
0.011617794632911682,
-0.00476192869246006,
-0.013889303430914879,
-0.02404824085533619,
-0.28206607699394226,
-0.08810137957334518,
0.17527903616428375,
-0.21808761358261108,
0.033679716289043427,
-0.0653250515460968,
0.08169294148683548,
0.018713681027293205,
0.12038717418909073,
0.06164313480257988,
-0.046911630779504776,
-0.0034297704696655273,
-0.07680434733629227,
-0.25481364130973816,
-0.0022519684862345457,
-0.033370789140462875,
-0.010969523340463638,
0.10915697365999222,
-0.12992438673973083,
0.03740832954645157,
0.2013832926750183,
-0.13010314106941223,
-0.013607844710350037,
0.1347854733467102,
-0.1363595575094223,
0.03722285106778145,
-0.012248372659087181,
0.004257606342434883,
-0.037850990891456604,
0.00807136669754982,
-0.0528579019010067,
0.07156967371702194,
-0.021826423704624176,
-0.09247210621833801,
0.04844728484749794,
-0.0681803748011589,
0.04769836738705635,
-0.10863489657640457,
-0.12073392421007156,
0.04253557696938515,
-0.07778384536504745,
0.05023380368947983,
0.046324241906404495,
-0.002992357825860381,
0.0568661093711853,
-0.007090866565704346,
0.0005743138026446104,
-0.1268775314092636,
0.11969964951276779,
0.10508144646883011,
-0.026138901710510254,
-0.01377839781343937
] |
null | null |
transformers
|
# diwank/dyda-deberta-pair
Deberta-based Daily Dialog style dialog-act annotations classification model. It takes two sentences as inputs (one previous and one current of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of four labels (exactly as in the [daily-dialog dataset](https://huggingface.co/datasets/daily_dialog) ): *__dummy__ (0), inform (1), question (2), directive (3), commissive (4)*
## Usage
```python
from simpletransformers.classification import (
ClassificationModel, ClassificationArgs
)
model = ClassificationModel("deberta", "diwank/dyda-deberta-pair")
convert_to_label = lambda n: ["__dummy__ (0), inform (1), question (2), directive (3), commissive (4)".split(', ')[i] for i in n]
predictions, raw_outputs = model.predict([["Say what is the meaning of life?", "I dont know"]])
convert_to_label(predictions) # inform (1)
```
|
{"license": "mit"}
|
text-classification
|
diwank/dyda-deberta-pair
|
[
"transformers",
"pytorch",
"tf",
"deberta",
"text-classification",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# diwank/dyda-deberta-pair
Deberta-based Daily Dialog style dialog-act annotations classification model. It takes two sentences as inputs (one previous and one current of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of four labels (exactly as in the daily-dialog dataset ): *__dummy__ (0), inform (1), question (2), directive (3), commissive (4)*
## Usage
|
[
"# diwank/dyda-deberta-pair\r\n\r\nDeberta-based Daily Dialog style dialog-act annotations classification model. It takes two sentences as inputs (one previous and one current of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of four labels (exactly as in the daily-dialog dataset ): *__dummy__ (0), inform (1), question (2), directive (3), commissive (4)*",
"## Usage"
] |
[
"TAGS\n#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# diwank/dyda-deberta-pair\r\n\r\nDeberta-based Daily Dialog style dialog-act annotations classification model. It takes two sentences as inputs (one previous and one current of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of four labels (exactly as in the daily-dialog dataset ): *__dummy__ (0), inform (1), question (2), directive (3), commissive (4)*",
"## Usage"
] |
[
46,
115,
3
] |
[
"passage: TAGS\n#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# diwank/dyda-deberta-pair\r\n\r\nDeberta-based Daily Dialog style dialog-act annotations classification model. It takes two sentences as inputs (one previous and one current of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of four labels (exactly as in the daily-dialog dataset ): *__dummy__ (0), inform (1), question (2), directive (3), commissive (4)*## Usage"
] |
[
0.0031996320467442274,
-0.05948076769709587,
-0.007943982258439064,
0.029869703575968742,
0.13551174104213715,
0.0015681994846090674,
0.18316631019115448,
0.05062223970890045,
0.059135645627975464,
-0.003956175409257412,
0.08661983907222748,
0.1985325962305069,
-0.020536085590720177,
0.15122079849243164,
-0.09124983102083206,
-0.2612394392490387,
0.08345526456832886,
-0.008590810000896454,
0.048888012766838074,
0.06722155213356018,
0.15254463255405426,
-0.03922969475388527,
-0.016470709815621376,
-0.055741142481565475,
-0.08568333089351654,
0.02578573115170002,
0.014080642722547054,
-0.11578698456287384,
0.15841573476791382,
0.04267096146941185,
0.06492096930742264,
0.07882604748010635,
-0.034482236951589584,
-0.07621682435274124,
0.01566162519156933,
-0.0032428093254566193,
0.02587377466261387,
-0.055348750203847885,
-0.10767032951116562,
0.004598439671099186,
-0.004685596097260714,
-0.020020315423607826,
0.08637503534555435,
0.010058170184493065,
-0.11660643666982651,
-0.22631549835205078,
-0.034214816987514496,
0.017651190981268883,
0.13317610323429108,
0.07374174147844315,
-0.04612947255373001,
0.03279031440615654,
-0.10042350739240646,
0.05702069029211998,
-0.033862173557281494,
-0.12688013911247253,
0.03448016941547394,
0.15018567442893982,
-0.044066041707992554,
0.05939868465065956,
-0.013959866017103195,
-0.026824763044714928,
0.002181674586609006,
0.05195656791329384,
-0.012383748777210712,
-0.038434047251939774,
0.10589702427387238,
-0.0361957773566246,
-0.09277202934026718,
0.018525486811995506,
0.3018408417701721,
-0.0032505474518984556,
-0.05247111991047859,
-0.05472372844815254,
0.007758220192044973,
0.00034143461380153894,
-0.059987373650074005,
-0.1347888857126236,
-0.011897717602550983,
0.0312594398856163,
0.08442024141550064,
0.0399971529841423,
-0.11578530818223953,
0.0230172798037529,
-0.23703648149967194,
0.1639775186777115,
0.00019175669876858592,
-0.026273289695382118,
-0.18349379301071167,
0.007986487820744514,
-0.023772085085511208,
-0.10231976211071014,
-0.08418857306241989,
-0.013199934735894203,
-0.002254720777273178,
-0.019167134538292885,
-0.10154300928115845,
-0.08469131588935852,
0.09067557752132416,
0.08396738767623901,
-0.02372116781771183,
-0.0004322197346482426,
-0.12481790781021118,
0.0375022254884243,
0.08209127187728882,
0.12102857232093811,
0.06999252736568451,
0.030116938054561615,
0.02917347103357315,
-0.016612816601991653,
0.10540539026260376,
-0.04771282151341438,
-0.1747913956642151,
0.036747630685567856,
0.10211633145809174,
0.05101076513528824,
0.017580663785338402,
0.14545615017414093,
-0.002838198561221361,
-0.05346037447452545,
0.08477833867073059,
-0.1138279139995575,
-0.06655995547771454,
0.05406167358160019,
-0.038931190967559814,
0.10364961624145508,
-0.049992043524980545,
0.047218725085258484,
0.021114056929945946,
0.07277219742536545,
-0.05335177853703499,
-0.019285181537270546,
0.0030056678224354982,
-0.11274727433919907,
0.032217081636190414,
0.0005163889145478606,
-0.02659110352396965,
-0.15955980122089386,
-0.08847843110561371,
-0.03624697029590607,
-0.044120002537965775,
-0.0309151578694582,
-0.09183071553707123,
-0.10532648116350174,
-0.004047292284667492,
-0.0030799536034464836,
-0.040890078991651535,
-0.14179018139839172,
-0.04474843293428421,
0.12228124588727951,
0.05268753319978714,
0.1063254326581955,
-0.1802700161933899,
0.11277499049901962,
-0.07300261408090591,
-0.004846415016800165,
-0.10885529965162277,
0.12847545742988586,
-0.015087343752384186,
0.017940199002623558,
-0.039954494684934616,
0.03544730320572853,
0.008764320984482765,
0.034566473215818405,
-0.031794775277376175,
0.1853824406862259,
-0.19790981709957123,
-0.057934366166591644,
0.1077122911810875,
-0.08655504882335663,
-0.0631384328007698,
0.11851297318935394,
-0.05856034904718399,
0.19646628201007843,
0.15581700205802917,
0.16758032143115997,
0.04507092013955116,
-0.14395366609096527,
0.10832346230745316,
0.08177430927753448,
-0.05247883126139641,
0.08451079577207565,
0.022696709260344505,
-0.011093929409980774,
0.037688300013542175,
0.056257229298353195,
0.06545273959636688,
0.03606090322136879,
-0.009029018692672253,
-0.04171450063586235,
0.06943979114294052,
0.013495255261659622,
0.10042573511600494,
-0.0047623696736991405,
0.013483776710927486,
-0.04138287529349327,
-0.00721781188622117,
0.1293691098690033,
0.03533558174967766,
-0.041376061737537384,
-0.011420764960348606,
-0.11804506182670593,
-0.047901615500450134,
0.03247521072626114,
0.05629086494445801,
-0.10720838606357574,
-0.1558389663696289,
-0.014595347456634045,
0.16360636055469513,
0.0155236991122365,
0.19914031028747559,
-0.010509970597922802,
-0.10349725931882858,
0.02588365226984024,
0.08465718477964401,
0.05494844540953636,
0.07489930093288422,
-0.023566488176584244,
-0.12524749338626862,
0.10103708505630493,
-0.09975185245275497,
0.125137060880661,
-0.015671275556087494,
0.008165761828422546,
0.06917794048786163,
-0.03141610324382782,
0.009269112721085548,
0.05714770406484604,
-0.02983672544360161,
0.06963802129030228,
-0.011246721260249615,
0.03496658056974411,
0.06519833207130432,
0.014746766537427902,
-0.08277533948421478,
0.20201101899147034,
-0.20100410282611847,
-0.020685575902462006,
0.1216980516910553,
-0.03689461946487427,
-0.0904252678155899,
0.004818761721253395,
-0.018981844186782837,
-0.07894284278154373,
0.000028306034437264316,
-0.1161108985543251,
0.06880088895559311,
-0.0012734007323160768,
0.08209182322025299,
0.007564203348010778,
0.0010534743778407574,
0.002024624263867736,
-0.056589093059301376,
-0.07940270751714706,
0.03956648334860802,
-0.006128941662609577,
-0.08892397582530975,
0.13077989220619202,
0.052600763738155365,
-0.003928932826966047,
0.12413290143013,
-0.021729931235313416,
-0.007989966310560703,
-0.03555496782064438,
0.0066876462660729885,
-0.0380592867732048,
0.02823345549404621,
-0.22486764192581177,
-0.036676179617643356,
0.016144150868058205,
0.08169270306825638,
0.006174052134156227,
-0.052260879427194595,
-0.03759905323386192,
-0.015558351762592793,
-0.06304063647985458,
-0.06263792514801025,
0.10576792061328888,
0.0186788197606802,
0.11408593505620956,
0.09472699463367462,
-0.17330066859722137,
0.14496827125549316,
0.011447597295045853,
-0.09430162608623505,
0.11298023909330368,
-0.11181197315454483,
-0.4037955403327942,
-0.0020407086703926325,
-0.18768350780010223,
-0.040872275829315186,
0.08395443111658096,
0.12582696974277496,
-0.05723267048597336,
-0.01642129011452198,
0.009206932038068771,
0.13529399037361145,
-0.050120361149311066,
-0.029166610911488533,
-0.004395990166813135,
-0.014181121252477169,
-0.06903954595327377,
-0.11798714101314545,
-0.05419275164604187,
-0.032692451030015945,
0.0406634695827961,
0.0853343978524208,
-0.12752822041511536,
0.020428916439414024,
0.21636594831943512,
0.0521162711083889,
0.01954386942088604,
-0.032144006341695786,
0.2321367859840393,
-0.08209076523780823,
0.002264702459797263,
0.14176270365715027,
-0.06912964582443237,
-0.004548321943730116,
0.2792915403842926,
0.020824627950787544,
-0.06683213263750076,
0.08914446085691452,
0.05060448870062828,
-0.07041529566049576,
-0.10226519405841827,
-0.05634913966059685,
-0.054067663848400116,
0.06294932961463928,
-0.10636764019727707,
0.0450969859957695,
0.14268045127391815,
-0.06836965680122375,
-0.015499021857976913,
-0.016815142706036568,
0.0684656873345375,
0.04403320327401161,
0.4548686146736145,
-0.06439996510744095,
0.14014078676700592,
0.0006361256237141788,
-0.1307181864976883,
0.011269391514360905,
0.043783776462078094,
0.021266859024763107,
0.08014228194952011,
0.017333457246422768,
0.04870014637708664,
-0.10596871376037598,
-0.007757298182696104,
-0.03930884227156639,
0.034000467509031296,
-0.021613214164972305,
-0.05939909815788269,
-0.04076998680830002,
-0.005131119396537542,
0.044866666197776794,
0.0643002837896347,
-0.1673634648323059,
-0.012638755142688751,
-0.007344251032918692,
0.036575693637132645,
0.09545885026454926,
0.13152679800987244,
-0.08423445373773575,
-0.09511767327785492,
0.010648476891219616,
0.014848758466541767,
0.004501411225646734,
0.02914680726826191,
-0.08642879128456116,
-0.03312067314982414,
0.004500548355281353,
0.04678553342819214,
0.03266778215765953,
-0.1061725988984108,
0.08618061244487762,
-0.0955062136054039,
-0.0835137814283371,
0.011040464974939823,
0.04265671223402023,
-0.4718979597091675,
0.09908770769834518,
0.0014819090720266104,
0.008692451752722263,
-0.08687349408864975,
-0.017034534364938736,
-0.05296245962381363,
0.1142602488398552,
0.03569253906607628,
-0.013160577043890953,
0.11558449268341064,
-0.06717414408922195,
-0.018765898421406746,
0.00708412891253829,
0.04839213937520981,
-0.0062859212048351765,
0.032293204218149185,
0.014166614972054958,
0.028229817748069763,
0.061528630554676056,
0.04605128988623619,
-0.06312720477581024,
-0.05814646556973457,
0.1739596575498581,
0.1325395554304123,
0.06746059656143188,
0.013171941973268986,
-0.06770659983158112,
-0.0403328463435173,
0.12908166646957397,
-0.08743537962436676,
-0.1141803190112114,
-0.04303083196282387,
-0.0008519139955751598,
-0.05056092143058777,
-0.05436151102185249,
-0.027718668803572655,
-0.01586746796965599,
0.013153363019227982,
-0.013879543170332909,
-0.12250115722417831,
0.11157304793596268,
-0.034291595220565796,
-0.0037674158811569214,
-0.04704338684678078,
0.07590897381305695,
0.06851012259721756,
0.013596925884485245,
0.04225161299109459,
-0.02048020251095295,
0.04857215657830238,
-0.020533883944153786,
0.02699633687734604,
-0.04566081613302231,
-0.050950732082128525,
-0.05702671781182289,
-0.027660062536597252,
-0.1337328851222992,
-0.12460274994373322,
-0.05678000673651695,
0.154631569981575,
0.14598095417022705,
-0.03840801864862442,
0.16276797652244568,
0.12969644367694855,
-0.003145826980471611,
-0.23264159262180328,
-0.07353056222200394,
0.04795666038990021,
-0.029700158163905144,
-0.07378128916025162,
-0.17356771230697632,
0.0469575934112072,
0.07897910475730896,
-0.044248707592487335,
0.15814834833145142,
-0.316582053899765,
-0.09668168425559998,
0.19456280767917633,
-0.055506665259599686,
0.44676244258880615,
-0.09361506253480911,
-0.04752105474472046,
0.006931591313332319,
-0.11708515882492065,
0.15560367703437805,
0.1056232824921608,
0.08552984893321991,
0.02532774955034256,
0.18838201463222504,
0.038387104868888855,
0.044589512050151825,
0.09587884694337845,
0.0754692330956459,
0.011671772226691246,
-0.09120895713567734,
-0.19196122884750366,
0.05041007697582245,
-0.01817999593913555,
0.10956540703773499,
0.023851798847317696,
-0.017776215448975563,
0.02596239559352398,
-0.10435165464878082,
-0.10065855085849762,
0.034618109464645386,
-0.053384602069854736,
-0.11257294565439224,
0.03471854329109192,
-0.014306973665952682,
-0.038999248296022415,
-0.05986148118972778,
0.1792834848165512,
-0.137667715549469,
0.03659787401556969,
0.07153942435979843,
0.001265602302737534,
-0.014250550419092178,
-0.02412865124642849,
-0.03539751097559929,
-0.09325868636369705,
0.07382020354270935,
-0.09372840076684952,
0.061918653547763824,
0.02368958294391632,
-0.019947851076722145,
0.16667111217975616,
0.07904131710529327,
-0.07346019148826599,
0.037054393440485,
0.04937075823545456,
-0.10246802866458893,
-0.14940568804740906,
0.04337705671787262,
-0.07282032072544098,
0.0038089239969849586,
0.0946139395236969,
0.1975134164094925,
0.002205954398959875,
-0.00422961451113224,
0.02623092755675316,
0.03895780071616173,
0.006227190140634775,
0.05456843599677086,
0.006329736206680536,
0.05762357637286186,
-0.06185465678572655,
-0.003252978902310133,
0.08131097257137299,
-0.18768103420734406,
0.03307996317744255,
0.04437781497836113,
-0.0736118033528328,
-0.08994168788194656,
-0.15522418916225433,
0.008463974110782146,
-0.05471966043114662,
-0.02763010561466217,
-0.06840693205595016,
-0.12003148347139359,
0.043472714722156525,
0.1837744563817978,
0.0549997054040432,
0.1833191215991974,
-0.04209914430975914,
-0.0010256171226501465,
-0.0010140634840354323,
0.10491962730884552,
-0.04381600394845009,
0.014420314691960812,
-0.039520759135484695,
0.06233780086040497,
-0.034920334815979004,
0.13300585746765137,
-0.07442875951528549,
-0.11061642318964005,
-0.12701945006847382,
0.013250833377242088,
-0.18887080252170563,
-0.03280514478683472,
-0.06488688290119171,
-0.012965841218829155,
0.06872157007455826,
-0.005208803806453943,
0.058731790632009506,
-0.009775785729289055,
-0.036574192345142365,
0.020858321338891983,
0.006668411195278168,
0.07313522696495056,
-0.07640627026557922,
-0.08215852081775665,
-0.03554076701402664,
-0.058716993778944016,
0.1490933746099472,
0.12581002712249756,
-0.10706590116024017,
0.0018778986996039748,
-0.09251663088798523,
0.08115900307893753,
0.11642477661371231,
-0.00906909815967083,
0.047728147357702255,
-0.08956940472126007,
-0.04082631319761276,
0.03661791607737541,
0.031247977167367935,
0.044262517243623734,
0.1252223551273346,
-0.13822297751903534,
0.04448216035962105,
-0.05427485704421997,
-0.14525631070137024,
-0.043021705001592636,
-0.057694241404533386,
0.00751337269321084,
0.11597968637943268,
0.13492220640182495,
-0.08853790909051895,
0.08041629195213318,
0.05200137197971344,
-0.02967752143740654,
0.001883700955659151,
-0.0410207062959671,
0.007297872565686703,
-0.049802832305431366,
0.022050628438591957,
-0.031185785308480263,
0.0785791352391243,
0.025831522420048714,
0.03857594355940819,
0.043368320912122726,
-0.15490104258060455,
0.033999212086200714,
0.023610709235072136,
0.14995354413986206,
0.0566784106194973,
-0.04707097262144089,
-0.0750834122300148,
0.053462520241737366,
0.01512549351900816,
0.08443038910627365,
0.09652378410100937,
0.06291677057743073,
0.03861132636666298,
0.041358478367328644,
0.09008457511663437,
-0.060393158346414566,
-0.05210137739777565,
-0.0898226648569107,
-0.03216629475355148,
0.012806706130504608,
-0.015909813344478607,
0.03708137944340706,
0.120989128947258,
0.023620255291461945,
0.009076118469238281,
-0.10301462560892105,
-0.06476394087076187,
-0.11060576140880585,
-0.015971051529049873,
-0.0662870928645134,
-0.08568129688501358,
-0.014546378515660763,
-0.06478702276945114,
-0.023534661158919334,
0.0004358120495453477,
0.03957069292664528,
-0.06964127719402313,
0.03641581907868385,
0.02824230305850506,
-0.10123873502016068,
0.053353991359472275,
-0.0604214146733284,
0.001509687746874988,
-0.04458574205636978,
0.03468899056315422,
-0.036289412528276443,
0.0791357159614563,
-0.0023522921837866306,
0.05534343421459198,
0.044993579387664795,
-0.05681588500738144,
-0.17246504127979279,
-0.08172675967216492,
0.017606740817427635,
0.06796162575483322,
-0.009448245167732239,
0.2412014752626419,
0.051470689475536346,
-0.04871554672718048,
0.05248180404305458,
0.13317729532718658,
-0.023615660145878792,
-0.17846573889255524,
-0.11747851967811584,
0.3053624629974365,
0.03818349167704582,
0.10308290272951126,
-0.05036776885390282,
-0.09682302176952362,
-0.09542276710271835,
0.09302808344364166,
0.19382120668888092,
0.022486399859189987,
0.014245127327740192,
-0.027737535536289215,
0.02705172449350357,
0.07669539004564285,
0.0626455768942833,
0.09069574624300003,
0.349824994802475,
0.02627542056143284,
-0.007005738094449043,
-0.04560429975390434,
0.01671113446354866,
-0.07551545649766922,
0.020011425018310547,
0.05725352093577385,
-0.021091481670737267,
-0.07091393321752548,
0.1190347746014595,
-0.1358472853899002,
-0.04812440276145935,
-0.028904354199767113,
-0.058833882212638855,
-0.045584291219711304,
-0.0069572897627949715,
-0.020786508917808533,
-0.03840814158320427,
0.07638074457645416,
-0.0161735936999321,
-0.0135725699365139,
0.0986354872584343,
-0.04526778310537338,
-0.05489440634846687,
0.02867087721824646,
0.08287610858678818,
0.0983411893248558,
-0.07307145744562149,
-0.036024559289216995,
0.19091103971004486,
0.05606602877378464,
0.12299351394176483,
-0.005804789252579212,
0.17502491176128387,
0.02699083462357521,
0.04142918437719345,
0.024334074929356575,
0.000363831699360162,
-0.03001965768635273,
0.05405234545469284,
0.11669174581766129,
-0.19565416872501373,
0.014437150210142136,
0.07457982003688812,
-0.10564637184143066,
-0.1125304102897644,
0.0987481027841568,
-0.059808533638715744,
0.010378344915807247,
0.0846618041396141,
-0.009641498327255249,
-0.03264256566762924,
0.0028236587531864643,
-0.08544223010540009,
0.028244851157069206,
-0.13524597883224487,
-0.14432676136493683,
-0.1350272297859192,
0.004166714381426573,
0.0488266795873642,
-0.0007724178140051663,
-0.23598738014698029,
-0.004231257364153862,
0.0036369666922837496,
0.058067191392183304,
0.006579949986189604,
0.025433829054236412,
-0.02760240249335766,
0.0027173731941729784,
-0.06753892451524734,
0.003896366572007537,
0.0008134847157634795,
0.11840498447418213,
-0.17224112153053284,
-0.1348409652709961
] |
null | null |
transformers
|
# maptask-deberta-pair
Deberta-based Daily MapTask style dialog-act annotations classification model
## Example
```python
from simpletransformers.classification import (
ClassificationModel, ClassificationArgs
)
model = ClassificationModel("deberta", "diwank/maptask-deberta-pair")
predictions, raw_outputs = model.predict([["Say what is the meaning of life?", "I dont know"]])
convert_to_label = lambda n: ["acknowledge (0), align (1), check (2), clarify (3), explain (4), instruct (5), query_w (6), query_yn (7), ready (8), reply_n (9), reply_w (10), reply_y (11)".split(', ')[i] for i in n]
convert_to_label(predictions) # reply_n (9)
```
|
{"license": "mit"}
|
text-classification
|
diwank/maptask-deberta-pair
|
[
"transformers",
"pytorch",
"tf",
"deberta",
"text-classification",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# maptask-deberta-pair
Deberta-based Daily MapTask style dialog-act annotations classification model
## Example
|
[
"# maptask-deberta-pair\r\nDeberta-based Daily MapTask style dialog-act annotations classification model",
"## Example"
] |
[
"TAGS\n#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# maptask-deberta-pair\r\nDeberta-based Daily MapTask style dialog-act annotations classification model",
"## Example"
] |
[
46,
28,
3
] |
[
"passage: TAGS\n#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# maptask-deberta-pair\r\nDeberta-based Daily MapTask style dialog-act annotations classification model## Example"
] |
[
-0.022866103798151016,
0.04105992242693901,
-0.0071936496533453465,
0.052776530385017395,
0.1570989042520523,
-0.004629886243492365,
0.18215397000312805,
0.05970992147922516,
0.0736776813864708,
-0.06721217185258865,
0.07658003270626068,
0.23894566297531128,
0.0050927274860441685,
0.22185365855693817,
-0.10676160454750061,
-0.3047104477882385,
0.044733162969350815,
0.030721047893166542,
0.0244520902633667,
0.09630867838859558,
0.16349752247333527,
-0.0431613065302372,
0.04799565300345421,
0.0204932801425457,
-0.07326854765415192,
0.07453092187643051,
0.019394708797335625,
-0.20908471941947937,
0.16408021748065948,
0.05764733999967575,
0.06078560650348663,
0.05990319699048996,
-0.006030815187841654,
-0.10410228371620178,
0.021084990352392197,
0.016163332387804985,
-0.048826247453689575,
0.002243854571133852,
0.05832592397928238,
-0.009400676004588604,
0.0377252995967865,
0.009288955479860306,
0.06409312784671783,
-0.005689084064215422,
-0.11551307886838913,
-0.1262771487236023,
-0.010418378747999668,
0.0396810919046402,
0.1437433809041977,
0.05907395854592323,
-0.01626606658101082,
0.09389925748109818,
-0.08898208290338516,
0.06784141063690186,
-0.07007332891225815,
-0.15670794248580933,
-0.015005428344011307,
0.16314826905727386,
-0.020838793367147446,
0.018888849765062332,
-0.02290751226246357,
0.017616745084524155,
-0.012712826021015644,
0.035239823162555695,
0.06088922172784805,
-0.04603957757353783,
-0.02117069810628891,
0.007214644458144903,
-0.05042051523923874,
-0.0063080876134335995,
0.18519456684589386,
-0.016544073820114136,
0.03046579845249653,
-0.09664838761091232,
-0.03956648334860802,
0.0908639207482338,
-0.06511477380990982,
-0.10207732021808624,
0.012536335736513138,
0.06632067263126373,
0.031907856464385986,
0.002004240872338414,
-0.08430220931768417,
0.03889692574739456,
-0.19890108704566956,
0.020291928201913834,
0.011695165187120438,
-0.019621144980192184,
-0.1966404765844345,
0.031524986028671265,
-0.0393957793712616,
-0.10831930488348007,
-0.02790945954620838,
-0.06199781969189644,
0.004036370199173689,
-0.07913275808095932,
-0.017908446490764618,
-0.04216944798827171,
0.1181517019867897,
0.09378829598426819,
0.07794007658958435,
0.011944319121539593,
-0.12903915345668793,
0.06870054453611374,
0.07102172821760178,
0.0983312800526619,
0.03965653479099274,
-0.02339540794491768,
0.04998327046632767,
-0.07898146659135818,
0.025503382086753845,
-0.014606683515012264,
-0.15526282787322998,
-0.05284411832690239,
0.08970385789871216,
0.08777011930942535,
0.014027653262019157,
0.07002551108598709,
-0.008140413090586662,
-0.04438415542244911,
0.1432410180568695,
-0.06254629045724869,
-0.036454375833272934,
-0.027901308611035347,
-0.05073719471693039,
0.017935866490006447,
-0.049248263239860535,
0.02169274166226387,
-0.04966851696372032,
0.12335047125816345,
-0.05002238601446152,
-0.059201303869485855,
-0.015306593850255013,
-0.097433902323246,
0.01145614217966795,
-0.041226357221603394,
0.00794603955000639,
-0.1711144894361496,
-0.19323687255382538,
-0.017080802470445633,
-0.01054951548576355,
-0.017675504088401794,
-0.10091196745634079,
-0.06862246245145798,
-0.015223785303533077,
-0.02063356526196003,
-0.061171844601631165,
-0.06663157045841217,
-0.07556238770484924,
0.06592229753732681,
-0.05841980502009392,
0.05557897686958313,
-0.10791899263858795,
0.11015310883522034,
-0.10229989141225815,
-0.023990241810679436,
-0.17799337208271027,
0.11245881021022797,
-0.043638527393341064,
0.09474073350429535,
-0.050703201442956924,
0.0017014634795486927,
-0.02543812058866024,
0.05893828719854355,
-0.049484625458717346,
0.14430779218673706,
-0.11966396123170853,
-0.08658111095428467,
0.0946669802069664,
-0.08274437487125397,
-0.08558917790651321,
0.10115673393011093,
-0.08438654989004135,
0.14582885801792145,
0.16031241416931152,
0.1826784610748291,
0.12187550961971283,
0.03656509891152382,
0.11978688091039658,
0.09480300545692444,
-0.10853847861289978,
-0.07117722928524017,
-0.06103776767849922,
-0.012440769001841545,
0.0324559323489666,
0.02357393689453602,
0.043123431503772736,
0.029254218563437462,
-0.03120705485343933,
-0.008615585044026375,
0.018927501514554024,
-0.027601003646850586,
0.07489470392465591,
-0.01949319615960121,
0.06513106822967529,
-0.06009920686483383,
0.019749248400330544,
0.06230251491069794,
-0.005303739570081234,
0.026586323976516724,
-0.034834012389183044,
-0.1153583750128746,
-0.05046014487743378,
0.0394139438867569,
0.03106997348368168,
-0.15746650099754333,
-0.13560903072357178,
0.011410881765186787,
0.06302215903997421,
-0.028733480721712112,
0.14880867302417755,
-0.03268744423985481,
-0.10455736517906189,
-0.0098679568618536,
0.04063710942864418,
0.0776253268122673,
0.04141315072774887,
-0.04668904095888138,
-0.11367520689964294,
0.06584867835044861,
-0.05905293673276901,
-0.008703680709004402,
-0.09796540439128876,
0.01951325498521328,
0.10765376687049866,
0.021944237872958183,
0.03632095828652382,
0.1113090068101883,
-0.022730426862835884,
0.07777106761932373,
-0.05497660115361214,
-0.0337347574532032,
0.07451949268579483,
-0.006919005420058966,
-0.00652295583859086,
0.20821715891361237,
-0.09391935914754868,
0.17594954371452332,
0.15117505192756653,
-0.06862471997737885,
-0.09354186803102493,
0.019636064767837524,
-0.013118813745677471,
-0.03512506186962128,
-0.08625328540802002,
-0.011092866770923138,
0.12179909646511078,
-0.010244429111480713,
0.14345091581344604,
0.005620270501822233,
-0.02521451562643051,
-0.031228914856910706,
-0.0021559824235737324,
-0.06063884496688843,
0.023293916136026382,
0.10191859304904938,
-0.20762862265110016,
0.16347765922546387,
0.1465470939874649,
0.03487742692232132,
0.15491287410259247,
0.000921346596442163,
0.03306180611252785,
0.047874558717012405,
-0.03351660072803497,
-0.022727087140083313,
-0.03777403384447098,
-0.17832113802433014,
-0.05721767991781235,
0.026884015649557114,
0.03706981614232063,
0.04439748451113701,
-0.05971768870949745,
-0.014129706658422947,
0.00026070428430102766,
-0.0335845872759819,
-0.043181970715522766,
0.11451810598373413,
0.058405887335538864,
0.10966446250677109,
0.007147075142711401,
-0.17963020503520966,
0.1503273844718933,
-0.007879093289375305,
-0.07691723853349686,
0.14463700354099274,
-0.09449633210897446,
-0.3755322992801666,
-0.08627337962388992,
-0.16318100690841675,
-0.020592045038938522,
0.06232903525233269,
0.10954733937978745,
-0.04217402637004852,
-0.0350540429353714,
-0.005794807802885771,
0.0541667714715004,
-0.0298354122787714,
-0.030421864241361618,
-0.10688953846693039,
0.0795169323682785,
-0.10458148270845413,
-0.1268639862537384,
-0.0676230937242508,
0.002366979606449604,
0.021540306508541107,
0.08706396073102951,
-0.12019949406385422,
0.010685733519494534,
0.14997977018356323,
-0.035129934549331665,
0.019760115072131157,
-0.06790653616189957,
0.1733616590499878,
-0.09520602226257324,
0.05403779074549675,
0.21654237806797028,
-0.027256105095148087,
0.02790183573961258,
0.2095644623041153,
-0.004298759158700705,
-0.07387898117303848,
0.07673390209674835,
0.005896905902773142,
-0.11044205725193024,
-0.19549039006233215,
-0.08455429971218109,
-0.08625159412622452,
0.09087427705526352,
-0.010865699499845505,
0.09504732489585876,
0.13382744789123535,
0.028760841116309166,
0.03766452148556709,
0.03083091787993908,
0.05294489860534668,
0.050022002309560776,
0.424998015165329,
0.002342789899557829,
0.14072391390800476,
-0.04181588813662529,
-0.14480842649936676,
0.07924290746450424,
0.05049525946378708,
0.0649026557803154,
0.09038946032524109,
0.017158295959234238,
0.009996214881539345,
-0.03966159000992775,
0.10796667635440826,
-0.018816720694303513,
0.1074216216802597,
-0.0028292762581259012,
-0.04659722372889519,
-0.0033437369856983423,
0.007822951301932335,
0.052154384553432465,
0.09165342152118683,
-0.21314965188503265,
0.013482168316841125,
-0.06442577391862869,
0.04636312276124954,
0.1547936201095581,
0.07230998575687408,
-0.21699626743793488,
0.008056088350713253,
0.04077036678791046,
-0.06627260893583298,
-0.06722231954336166,
0.06844965368509293,
-0.08330073952674866,
-0.14194348454475403,
0.07081666588783264,
0.005013815127313137,
0.09637320041656494,
-0.17923596501350403,
0.06891457736492157,
-0.05121828243136406,
-0.041794441640377045,
-0.042544443160295486,
0.07046472281217575,
-0.38440144062042236,
0.1632276177406311,
-0.01795896328985691,
-0.019855443388223648,
-0.10436180979013443,
0.011590838432312012,
0.08182015269994736,
0.1600985825061798,
0.06452381610870361,
-0.04589327052235603,
0.009147967211902142,
0.0039000408723950386,
-0.026522677391767502,
0.004053385928273201,
0.009552831761538982,
-0.060181111097335815,
-0.04152962565422058,
-0.00017087566084228456,
0.0021196682937443256,
0.01581638678908348,
-0.04484425112605095,
-0.0918373093008995,
-0.0959957167506218,
0.05966423079371452,
0.13550394773483276,
0.10199229419231415,
-0.002325861481949687,
-0.052884094417095184,
-0.011878958903253078,
0.17754705250263214,
0.05202199146151543,
-0.1114819198846817,
-0.11606067419052124,
-0.016214605420827866,
-0.020103802904486656,
-0.01660289615392685,
0.08052732795476913,
0.010454447939991951,
0.08538634330034256,
-0.0033171281684190035,
-0.15706369280815125,
0.1460348516702652,
-0.10212790220975876,
0.024319276213645935,
-0.07155454158782959,
0.04006727412343025,
0.028905199840664864,
-0.041063860058784485,
0.08619825541973114,
-0.012816169299185276,
0.025248946622014046,
-0.03328818455338478,
0.008381200954318047,
-0.054039116948843,
-0.08270512521266937,
0.011162646114826202,
-0.026316508650779724,
-0.2248757928609848,
-0.053014710545539856,
-0.06811463087797165,
0.15128375589847565,
0.08022243529558182,
-0.01382810901850462,
0.14940769970417023,
0.13157173991203308,
-0.04714414104819298,
-0.238244891166687,
-0.0007795082055963576,
-0.024646028876304626,
-0.0077076684683561325,
0.0467742420732975,
-0.19681411981582642,
0.058737121522426605,
0.09263608604669571,
-0.051376424729824066,
0.1963701695203781,
-0.3206934332847595,
-0.11842980235815048,
0.23902249336242676,
-0.004526193253695965,
0.23863980174064636,
-0.06311305612325668,
-0.06903988122940063,
-0.07726332545280457,
-0.0651889368891716,
0.09344528615474701,
0.09509690850973129,
0.0423574261367321,
0.017972532659769058,
0.12436556071043015,
0.01798972301185131,
0.05217045545578003,
0.07888230681419373,
0.12234927713871002,
0.04947887361049652,
-0.0921778529882431,
-0.042312998324632645,
0.1466040313243866,
-0.005451090633869171,
0.05804828181862831,
0.02826428785920143,
0.00830046832561493,
-0.06620913743972778,
-0.11279474943876266,
-0.048524681478738785,
0.059577278792858124,
-0.002664316212758422,
-0.11470958590507507,
0.011866282671689987,
0.014020789414644241,
-0.03298839554190636,
-0.06466395407915115,
0.13867008686065674,
-0.06920087337493896,
0.0897517204284668,
0.1119554415345192,
0.12077802419662476,
0.016977459192276,
-0.05023305490612984,
-0.07635039836168289,
-0.07887188345193863,
0.06007472425699234,
-0.061186447739601135,
0.05066296458244324,
0.057449039071798325,
-0.04197626933455467,
0.07456209510564804,
0.08640532940626144,
-0.01654805801808834,
0.020874803885817528,
0.10825959593057632,
-0.10746901482343674,
-0.0750562995672226,
0.01036897487938404,
0.004351658280938864,
0.09071726351976395,
0.07769156992435455,
0.18918752670288086,
-0.011669225059449673,
-0.043624743819236755,
0.05331284925341606,
-0.00023584540758747607,
0.05866621807217598,
0.10443343222141266,
0.012410842813551426,
0.016490543261170387,
-0.09297437965869904,
0.053240858018398285,
0.05426108464598656,
-0.1404159814119339,
0.003251421730965376,
0.05290577560663223,
-0.12021514028310776,
-0.11229079216718674,
-0.061099614948034286,
0.05909300595521927,
-0.1182611882686615,
-0.07498250901699066,
-0.04218306764960289,
-0.163384810090065,
0.10603304952383041,
0.22284819185733795,
0.07905574142932892,
0.1029999777674675,
-0.06389971822500229,
-0.05942588299512863,
0.011338789947330952,
0.06513984501361847,
0.004895233083516359,
0.038812462240457535,
-0.09234071522951126,
-0.04047758877277374,
-0.03703455626964569,
0.06137695536017418,
-0.08519461005926132,
-0.06359176337718964,
-0.12245412915945053,
0.04593193158507347,
-0.1550687998533249,
0.0018732231110334396,
-0.05860208347439766,
-0.019238993525505066,
0.0149648142978549,
0.0076730805449187756,
-0.014742444269359112,
-0.02813643589615822,
-0.10272610932588577,
0.01771244779229164,
-0.024455241858959198,
0.12213084846735,
-0.0827213004231453,
-0.046642668545246124,
0.022493120282888412,
-0.0720302164554596,
0.16720445454120636,
0.027309944853186607,
-0.11604174971580505,
0.006368936039507389,
-0.1391911506652832,
-0.005976490676403046,
0.12009399384260178,
0.014507587067782879,
0.06192324310541153,
-0.04400229454040527,
0.04220996052026749,
0.05153101310133934,
-0.020058900117874146,
0.048205409198999405,
0.08143190294504166,
-0.13858969509601593,
0.02333221770823002,
-0.06950461864471436,
-0.1200874000787735,
0.005228482652455568,
-0.04003317654132843,
0.11161371320486069,
0.05742714926600456,
0.13195490837097168,
-0.08008378744125366,
0.030080951750278473,
-0.0011123484000563622,
-0.010351079516112804,
-0.033110033720731735,
-0.14724798500537872,
0.029428871348500252,
-0.07000576704740524,
0.039170216768980026,
-0.006255976855754852,
0.1956525444984436,
0.1443587690591812,
0.04356319084763527,
0.08278415352106094,
-0.07919250428676605,
-0.011652597226202488,
0.06619444489479065,
0.2100314199924469,
0.04052440822124481,
-0.028151415288448334,
-0.04788137227296829,
0.03403222933411598,
0.011290420778095722,
0.006816270761191845,
0.12903554737567902,
0.064130499958992,
-0.05797187238931656,
0.04949657618999481,
0.11420799791812897,
-0.029759561643004417,
-0.06893325597047806,
-0.09706151485443115,
-0.00880578625947237,
0.10026054829359055,
0.023413309827446938,
0.008814358152449131,
0.11998143792152405,
-0.018655555322766304,
0.04134213551878929,
-0.04233993962407112,
-0.039327625185251236,
-0.119625523686409,
-0.08560348302125931,
-0.07755204290151596,
-0.07885248214006424,
0.00856118742376566,
-0.06560263782739639,
0.0008292572456412017,
0.20052838325500488,
0.07075782865285873,
-0.08886870741844177,
0.018993405625224113,
-0.038157396018505096,
-0.11059202998876572,
0.15728157758712769,
-0.06469506025314331,
-0.02145427279174328,
-0.12580758333206177,
-0.035584550350904465,
-0.1634942591190338,
0.07609927654266357,
-0.07018078118562698,
0.026881031692028046,
-0.012249269522726536,
-0.02530665136873722,
-0.13763801753520966,
-0.028568651527166367,
-0.006974158342927694,
0.015557713806629181,
-0.06726967543363571,
0.09377464652061462,
0.03740144520998001,
-0.06242305412888527,
0.09155160188674927,
0.19018694758415222,
-0.040659282356500626,
-0.14010612666606903,
-0.07046543806791306,
0.2280462384223938,
0.08520618826150894,
0.08849142491817474,
0.014123722910881042,
-0.056428827345371246,
-0.03883261978626251,
0.0887959748506546,
0.2434554398059845,
-0.03465498238801956,
0.02285904251039028,
-0.04433443024754524,
0.010023404844105244,
0.0902826339006424,
0.14129184186458588,
0.04270375519990921,
0.2259196788072586,
-0.006962672807276249,
-0.03715262934565544,
-0.034124307334423065,
0.0014951241901144385,
-0.10848292708396912,
0.08480467647314072,
0.09563032537698746,
-0.0660061165690422,
-0.042190760374069214,
0.18511950969696045,
-0.1630951315164566,
-0.03859652951359749,
0.001839405857026577,
-0.0862288847565651,
-0.10770606994628906,
-0.017353391274809837,
-0.01523865107446909,
-0.03228086978197098,
0.0928904190659523,
-0.05058499798178673,
-0.03154543787240982,
0.04201756417751312,
-0.00888841226696968,
-0.13238292932510376,
-0.021662287414073944,
0.04579445347189903,
0.04361362010240555,
0.005958319175988436,
-0.05708693712949753,
0.09609586000442505,
0.0723881646990776,
0.05473317578434944,
-0.005930760409682989,
0.13643929362297058,
0.0150200380012393,
0.09706618636846542,
0.0702609270811081,
-0.06429348886013031,
-0.006112492643296719,
0.008378676138818264,
0.08523467183113098,
-0.1812426596879959,
0.031079908832907677,
-0.00540596479550004,
-0.05826462060213089,
-0.0756581649184227,
0.07377245277166367,
-0.05796419829130173,
0.001957131549715996,
0.03720394894480705,
-0.0011915965005755424,
-0.012108677998185158,
0.011406349018216133,
-0.029300788417458534,
0.0160764679312706,
-0.14378730952739716,
-0.09901735186576843,
-0.06006075069308281,
-0.04235765337944031,
-0.009660360403358936,
-0.04440397024154663,
-0.21453255414962769,
-0.030885057523846626,
-0.1367313116788864,
0.03209180384874344,
-0.09293178468942642,
0.08427266776561737,
0.02383812703192234,
0.007460596971213818,
-0.011196702718734741,
0.14261071383953094,
-0.007993488572537899,
0.07735287398099899,
-0.14885833859443665,
-0.0941358357667923
] |
null | null |
transformers
|
# diwank/silicone-deberta-pair
`deberta-base`-based dialog acts classifier. Trained on the `balanced` variant of the [silicone-merged](https://huggingface.co/datasets/diwank/silicone-merged) dataset: a simplified merged dialog act data from datasets in the [silicone](https://huggingface.co/datasets/silicone) collection.
Takes two sentences as inputs (one previous and one current utterance of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. **Outputs one of 11 labels**:
```python
(0, 'acknowledge')
(1, 'answer')
(2, 'backchannel')
(3, 'reply_yes')
(4, 'exclaim')
(5, 'say')
(6, 'reply_no')
(7, 'hold')
(8, 'ask')
(9, 'intent')
(10, 'ask_yes_no')
```
## Example:
```python
from simpletransformers.classification import (
ClassificationModel, ClassificationArgs
)
model = ClassificationModel("deberta", "diwank/silicone-deberta-pair")
convert_to_label = lambda n: [
['acknowledge',
'answer',
'backchannel',
'reply_yes',
'exclaim',
'say',
'reply_no',
'hold',
'ask',
'intent',
'ask_yes_no'
][i] for i in n
]
predictions, raw_outputs = model.predict([["Say what is the meaning of life?", "I dont know"]])
convert_to_label(predictions) # answer
```
## Report from W&B
https://wandb.ai/diwank/da-silicone-combined/reports/silicone-deberta-pair--VmlldzoxNTczNjE5?accessToken=yj1jz4c365z0y5b3olgzye7qgsl7qv9lxvqhmfhtb6300hql6veqa5xiq1skn8ys
|
{"license": "mit"}
|
text-classification
|
diwank/silicone-deberta-pair
|
[
"transformers",
"pytorch",
"tf",
"deberta",
"text-classification",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# diwank/silicone-deberta-pair
'deberta-base'-based dialog acts classifier. Trained on the 'balanced' variant of the silicone-merged dataset: a simplified merged dialog act data from datasets in the silicone collection.
Takes two sentences as inputs (one previous and one current utterance of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of 11 labels:
## Example:
## Report from W&B
URL
|
[
"# diwank/silicone-deberta-pair\r\n\r\n'deberta-base'-based dialog acts classifier. Trained on the 'balanced' variant of the silicone-merged dataset: a simplified merged dialog act data from datasets in the silicone collection. \r\n\r\nTakes two sentences as inputs (one previous and one current utterance of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of 11 labels:",
"## Example:",
"## Report from W&B\r\n\r\nURL"
] |
[
"TAGS\n#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# diwank/silicone-deberta-pair\r\n\r\n'deberta-base'-based dialog acts classifier. Trained on the 'balanced' variant of the silicone-merged dataset: a simplified merged dialog act data from datasets in the silicone collection. \r\n\r\nTakes two sentences as inputs (one previous and one current utterance of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of 11 labels:",
"## Example:",
"## Report from W&B\r\n\r\nURL"
] |
[
46,
117,
4,
7
] |
[
"passage: TAGS\n#transformers #pytorch #tf #deberta #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# diwank/silicone-deberta-pair\r\n\r\n'deberta-base'-based dialog acts classifier. Trained on the 'balanced' variant of the silicone-merged dataset: a simplified merged dialog act data from datasets in the silicone collection. \r\n\r\nTakes two sentences as inputs (one previous and one current utterance of a dialog). The previous sentence can be an empty string if this is the first utterance of a speaker in a dialog. Outputs one of 11 labels:## Example:## Report from W&B\r\n\r\nURL"
] |
[
-0.045924291014671326,
-0.06308634579181671,
-0.005446317605674267,
0.0270168948918581,
0.11562661826610565,
-0.04400194063782692,
0.16789108514785767,
0.06248791515827179,
0.10208416730165482,
0.023896334692835808,
0.0808960571885109,
0.16633877158164978,
-0.02640736848115921,
0.18667767941951752,
-0.06708750128746033,
-0.19068287312984467,
0.04275970160961151,
0.032340820878744125,
0.006608131341636181,
0.06782114505767822,
0.17254328727722168,
-0.05538585036993027,
0.01268396619707346,
-0.024031946435570717,
-0.08118514716625214,
0.04176715388894081,
0.04288613423705101,
-0.05962270498275757,
0.14880456030368805,
0.07013752311468124,
0.1108553484082222,
0.06712320446968079,
-0.0436927005648613,
-0.11565767973661423,
0.02544599026441574,
0.019541336223483086,
0.0031761019490659237,
-0.01130612287670374,
-0.0731673315167427,
-0.007946820929646492,
-0.005632643122226,
-0.0012342528207227588,
0.05592123791575432,
0.017058979719877243,
-0.06718070060014725,
-0.09457416832447052,
-0.002822909504175186,
0.003918580710887909,
0.05979232117533684,
0.028007959946990013,
-0.03595229983329773,
0.04761099815368652,
-0.04741308465600014,
0.10979506373405457,
-0.008280430920422077,
-0.13331420719623566,
0.012300071306526661,
0.014565013349056244,
0.002771002473309636,
0.07369174063205719,
-0.018563566729426384,
-0.012319187633693218,
-0.018467271700501442,
0.08221995830535889,
0.024662960320711136,
-0.0631198137998581,
-0.036251720041036606,
-0.005790462717413902,
-0.07198553532361984,
0.039365414530038834,
0.2754879593849182,
-0.013576089404523373,
-0.011523278430104256,
-0.09574004262685776,
-0.039348408579826355,
0.057244572788476944,
0.02892570197582245,
-0.11652008444070816,
-0.012183164246380329,
0.028314698487520218,
0.07470373809337616,
-0.020624632015824318,
-0.11870966851711273,
-0.017064563930034637,
-0.20569004118442535,
0.1718681901693344,
-0.02042103372514248,
0.014566220343112946,
-0.17512352764606476,
0.007704379037022591,
0.06170633062720299,
-0.08033386617898941,
-0.0211647916585207,
-0.013495640829205513,
-0.013874820433557034,
0.0271126888692379,
-0.13271132111549377,
-0.1624625027179718,
0.06357763707637787,
0.1127718836069107,
-0.07455118745565414,
-0.014722117222845554,
-0.0641801729798317,
0.06921850144863129,
0.021100850775837898,
0.06105109304189682,
0.015306921675801277,
-0.030887583270668983,
0.04362794756889343,
-0.03281226009130478,
0.09776677191257477,
-0.06582006812095642,
-0.15850086510181427,
-0.002948703709989786,
0.022085491567850113,
0.007891594432294369,
0.04516664519906044,
0.10088428854942322,
-0.06061514839529991,
-0.04998496547341347,
0.08239820599555969,
-0.12146711349487305,
-0.09332044422626495,
0.004681175108999014,
0.002943920437246561,
0.13265377283096313,
0.06025463342666626,
-0.00595966586843133,
-0.0213625505566597,
-0.0436018705368042,
-0.05967025086283684,
-0.048146773129701614,
-0.009515461511909962,
-0.06839816272258759,
0.039268773049116135,
-0.0459311380982399,
-0.048426367342472076,
-0.1938013732433319,
-0.1297125667333603,
-0.01040557585656643,
-0.014047374948859215,
-0.012597775086760521,
-0.04565803334116936,
-0.07902977615594864,
0.07047078013420105,
-0.0032986479345709085,
-0.07210493087768555,
-0.12045533210039139,
-0.013699035160243511,
0.12745434045791626,
-0.024686558172106743,
0.09537370502948761,
-0.2643224596977234,
0.09532245248556137,
-0.12784358859062195,
-0.007915521040558815,
-0.1291661411523819,
0.14551466703414917,
-0.000867049559019506,
0.03213586285710335,
-0.04168440401554108,
0.026571450755000114,
-0.10191161930561066,
0.0371534526348114,
0.018999699503183365,
0.1390371173620224,
-0.18981154263019562,
-0.09105973690748215,
0.07564462721347809,
-0.10252584517002106,
-0.0957801416516304,
0.1171673908829689,
-0.03338747099041939,
0.20461630821228027,
0.10573439300060272,
0.25821220874786377,
0.0347600132226944,
-0.09576688706874847,
0.0446227602660656,
0.13407303392887115,
-0.06324175000190735,
0.03458673506975174,
0.058942731469869614,
-0.03572739288210869,
0.09331583231687546,
0.04549705237150192,
0.025914886966347694,
0.0633501186966896,
-0.021556619554758072,
-0.059576861560344696,
0.057873938232660294,
-0.0423407144844532,
0.15636076033115387,
-0.01759013533592224,
0.02577119693160057,
-0.017167629674077034,
-0.047457557171583176,
0.23983968794345856,
0.058268144726753235,
-0.10609160363674164,
0.042522087693214417,
-0.07831108570098877,
0.09542792290449142,
-0.016398392617702484,
0.014538467861711979,
-0.20806585252285004,
-0.11942506581544876,
-0.012110958807170391,
0.07381521910429001,
0.013987117446959019,
0.15557579696178436,
0.0059505668468773365,
-0.05010278895497322,
0.0027387787122279406,
0.08092301338911057,
0.1030249148607254,
0.044749196618795395,
-0.023418935015797615,
-0.14724880456924438,
0.03432706743478775,
-0.07012757658958435,
0.1085406169295311,
0.001465059001930058,
-0.006620423868298531,
0.040560755878686905,
0.0604076161980629,
0.03407571464776993,
0.007192422170192003,
-0.054709456861019135,
0.06890739500522614,
-0.040782373398542404,
0.04219173267483711,
0.07782698422670364,
0.06889743357896805,
-0.12057331204414368,
0.1832866221666336,
-0.1638883799314499,
0.009871850721538067,
0.10419055074453354,
-0.19135330617427826,
-0.07237859070301056,
-0.00931704230606556,
-0.02472592703998089,
-0.06360439956188202,
0.034880075603723526,
-0.11692770570516586,
0.11839095503091812,
-0.02634594775736332,
0.08762148022651672,
-0.04296117275953293,
-0.020004434511065483,
-0.002583643887192011,
-0.04006058722734451,
-0.050888851284980774,
0.054621245712041855,
-0.037426918745040894,
-0.10418527573347092,
0.1116611659526825,
0.026249459013342857,
-0.026580674573779106,
0.198934867978096,
-0.0028844294138252735,
-0.0209613386541605,
-0.035326577723026276,
0.028368502855300903,
-0.04165452718734741,
0.042668458074331284,
-0.23813779652118683,
-0.039691224694252014,
0.02403927408158779,
-0.015903787687420845,
0.04475226253271103,
-0.05556686595082283,
0.007884365506470203,
0.011133887805044651,
-0.06648433953523636,
-0.0498112253844738,
0.10461779683828354,
-0.03171604126691818,
0.0665537416934967,
0.06568577885627747,
-0.07967233657836914,
0.10584354400634766,
0.0010318630374968052,
-0.15015259385108948,
0.10262011736631393,
-0.08822757750749588,
-0.3374045193195343,
-0.12727251648902893,
-0.08021800965070724,
-0.0027706408873200417,
0.0788055881857872,
0.10554444044828415,
0.01617041788995266,
-0.005221598781645298,
0.010555724613368511,
0.04023510217666626,
-0.0037689944729208946,
0.005791562609374523,
0.01920275390148163,
-0.027242816984653473,
0.007705904543399811,
-0.10973136872053146,
-0.04320139065384865,
-0.03279401361942291,
0.0886988714337349,
0.004854784347116947,
-0.08292844891548157,
0.0895240381360054,
0.1837191879749298,
0.010025855153799057,
0.009451090358197689,
-0.0786450058221817,
0.12173644453287125,
-0.059821441769599915,
0.02704767882823944,
0.19148826599121094,
-0.11090433597564697,
0.016587577760219574,
0.1758950799703598,
0.011785469949245453,
-0.027960820123553276,
0.06081324815750122,
-0.002818248001858592,
-0.07988467067480087,
-0.11862965673208237,
-0.08184035867452621,
-0.0978497713804245,
0.009278569370508194,
-0.1161823570728302,
0.011078269220888615,
0.06131744012236595,
-0.045518048107624054,
-0.004628477618098259,
0.03672092407941818,
0.01829283870756626,
0.01572112925350666,
0.34476158022880554,
-0.05445385351777077,
0.19123584032058716,
-0.022876553237438202,
-0.06076120212674141,
0.03992972522974014,
0.00780889205634594,
0.034553393721580505,
0.0035875854082405567,
0.12679189443588257,
0.09736868739128113,
-0.04551253467798233,
-0.0012436367105692625,
0.00834437645971775,
-0.03395425155758858,
0.031766556203365326,
-0.03106340765953064,
-0.09598135203123093,
-0.03103749454021454,
0.04269786551594734,
0.040779516100883484,
-0.0795096755027771,
-0.022045500576496124,
-0.028109190985560417,
0.05581540986895561,
0.10664892941713333,
0.10588706284761429,
-0.1423790007829666,
-0.14851228892803192,
0.013212677091360092,
-0.009132773615419865,
0.011626459658145905,
0.024028820917010307,
0.03759806603193283,
-0.017069941386580467,
0.04200808331370354,
0.027358418330550194,
0.007741772569715977,
-0.04975949227809906,
0.08990069478750229,
0.030014729127287865,
-0.048298612236976624,
0.020534224808216095,
0.03292693942785263,
-0.4113655984401703,
0.10297468304634094,
-0.00402785325422883,
-0.03546353057026863,
-0.04188740625977516,
0.026653019711375237,
-0.012605028226971626,
-0.0056656282395124435,
0.09364339709281921,
0.0001335823762929067,
0.13655680418014526,
-0.11114994436502457,
0.017299357801675797,
0.011396702378988266,
0.058720000088214874,
0.00010757549898698926,
0.007101190742105246,
-0.037002965807914734,
0.02259296551346779,
0.0739884302020073,
0.11898615956306458,
-0.00917151290923357,
-0.08832217007875443,
0.1353117674589157,
0.09461627900600433,
-0.0383332297205925,
0.012697206810116768,
-0.07673992961645126,
-0.0437125526368618,
0.0534772127866745,
-0.18942144513130188,
-0.0983135998249054,
-0.038494773209095,
-0.02335679717361927,
-0.023147407919168472,
-0.09439876675605774,
0.05095520615577698,
-0.019260935485363007,
0.0036856108345091343,
-0.0388326570391655,
-0.15372546017169952,
0.0882684588432312,
-0.0318220891058445,
0.059594351798295975,
-0.042253244668245316,
0.0670429989695549,
0.021564895287156105,
0.015571489930152893,
0.010277770459651947,
-0.027959002181887627,
-0.010818240232765675,
0.004823816008865833,
-0.002469595056027174,
0.007164819166064262,
-0.01588796265423298,
0.0008967785979621112,
-0.061834610998630524,
-0.07532227039337158,
-0.06901998072862625,
-0.06375238299369812,
0.10747585445642471,
0.07192376255989075,
-0.06464694440364838,
0.19445432722568512,
0.24602162837982178,
-0.054273780435323715,
-0.28071415424346924,
-0.0733996257185936,
0.01472033653408289,
0.03339823707938194,
-0.012753301300108433,
-0.13754217326641083,
0.05597252771258354,
0.019833577796816826,
-0.03604389354586601,
0.07397422194480896,
-0.1980447918176651,
-0.13307906687259674,
0.17485934495925903,
-0.020434843376278877,
0.34270936250686646,
-0.0992354303598404,
-0.1041722372174263,
-0.009021474048495293,
-0.06351970881223679,
0.11316121369600296,
0.048226580023765564,
0.07986115664243698,
0.05464675650000572,
0.07042568176984787,
0.004993935581296682,
0.032238371670246124,
0.060874082148075104,
-0.028448255732655525,
0.004123394843190908,
-0.06778963655233383,
-0.12507784366607666,
0.08868958055973053,
0.014524614438414574,
0.08332248032093048,
-0.007535882759839296,
0.014963711611926556,
0.048600997775793076,
-0.08568625152111053,
-0.11998967826366425,
0.06657041609287262,
-0.024170799180865288,
-0.14587847888469696,
0.04211253300309181,
-0.03921806439757347,
-0.029464520514011383,
-0.01210954412817955,
0.07694802433252335,
-0.12474632263183594,
0.07812371850013733,
0.15099915862083435,
0.07875581830739975,
-0.06332799047231674,
-0.03168896213173866,
-0.04669731855392456,
-0.1099642664194107,
0.0791405439376831,
-0.05786260589957237,
0.030200419947504997,
0.03737245500087738,
0.013264001347124577,
0.11021159589290619,
0.09674791991710663,
-0.020860495045781136,
0.009072845801711082,
0.07547333836555481,
-0.12528949975967407,
-0.10795401781797409,
0.00857964064925909,
-0.12585090100765228,
-0.013475238345563412,
0.0584641769528389,
0.20951135456562042,
-0.010816464200615883,
0.01598186232149601,
-0.0050788274966180325,
0.01721932925283909,
-0.07681341469287872,
0.06248842924833298,
0.04110313579440117,
0.03157889097929001,
-0.06453319638967514,
0.019477616995573044,
0.1152779757976532,
-0.16271816194057465,
0.028984032571315765,
0.000991929555311799,
-0.05675061047077179,
-0.07236457616090775,
-0.1533479243516922,
0.05428741127252579,
-0.1141362264752388,
-0.037471797317266464,
-0.12403719872236252,
-0.10626845806837082,
0.042868223041296005,
0.22656099498271942,
0.07137767225503922,
0.18686313927173615,
-0.08251785486936569,
0.05065016821026802,
0.022553304210305214,
0.12008987367153168,
-0.029413416981697083,
0.026882203295826912,
-0.07070817053318024,
0.1532420665025711,
-0.04024011269211769,
0.10044752806425095,
-0.0779765322804451,
-0.05751849338412285,
-0.057400476187467575,
0.010486810468137264,
-0.114092156291008,
-0.015957141295075417,
-0.037532608956098557,
-0.026051092892885208,
0.029775995761156082,
0.006690254434943199,
0.008232414722442627,
0.001789176487363875,
0.00953213032335043,
0.04245942831039429,
0.039044089615345,
0.05653992295265198,
-0.09679039567708969,
-0.05205294117331505,
0.01566351391375065,
-0.026155084371566772,
0.15652678906917572,
0.08955325931310654,
-0.10301543027162552,
-0.01410419587045908,
-0.1396344006061554,
0.039069171994924545,
0.06381083279848099,
0.060707125812768936,
-0.0031239737290889025,
-0.036960702389478683,
-0.03289579972624779,
0.06053140386939049,
0.02346259169280529,
0.004104678053408861,
0.08376532793045044,
-0.09065712988376617,
0.08838729560375214,
-0.04522034153342247,
-0.0748644694685936,
-0.09071327745914459,
-0.013808859512209892,
0.002443415578454733,
0.15743747353553772,
0.05160803720355034,
-0.09305255860090256,
0.025637393817305565,
-0.048188839107751846,
0.0009084109915420413,
0.040774211287498474,
-0.043583475053310394,
-0.04506900534033775,
-0.08141668885946274,
0.012720619328320026,
-0.021166792139410973,
0.14506131410598755,
0.016897600144147873,
0.027400506660342216,
0.02529904618859291,
-0.059494245797395706,
0.04766964539885521,
0.017673727124929428,
0.16386693716049194,
0.017785483971238136,
-0.018833447247743607,
-0.12250573933124542,
0.06276964396238327,
0.008726751431822777,
0.22157853841781616,
0.08999115973711014,
0.13836976885795593,
0.0915503203868866,
0.0329754576086998,
0.09138395637273788,
-0.010015176609158516,
-0.0927467793226242,
-0.08237023651599884,
-0.044554706662893295,
0.015106229111552238,
-0.01619919389486313,
0.03615574538707733,
0.1023302897810936,
-0.08044632524251938,
0.05740359053015709,
-0.05595579743385315,
-0.06546203047037125,
-0.1143326684832573,
-0.13033835589885712,
0.01120169647037983,
-0.06456328928470612,
-0.06155997887253761,
-0.03877589851617813,
-0.0358913354575634,
0.05920765548944473,
0.010189665481448174,
-0.059931449592113495,
0.08080529421567917,
-0.055182937532663345,
-0.061124734580516815,
0.033236242830753326,
-0.04966587945818901,
0.021427331492304802,
-0.035115353763103485,
-0.04725700989365578,
0.03637957572937012,
0.08743749558925629,
0.0015497960848733783,
-0.00554769579321146,
0.039402469992637634,
-0.0392877459526062,
-0.09015144407749176,
-0.036584485322237015,
0.007321433164179325,
0.01322616171091795,
-0.04939267411828041,
0.20036724209785461,
0.04941459372639656,
-0.005542670376598835,
0.03829788789153099,
0.1378171741962433,
-0.03821117803454399,
-0.13240240514278412,
-0.11013700813055038,
0.12856967747211456,
-0.0037921222392469645,
0.11642121523618698,
-0.031265728175640106,
-0.0971967875957489,
-0.06752416491508484,
0.14650800824165344,
0.20032180845737457,
-0.04189677909016609,
0.017128268256783485,
0.08435884863138199,
0.0232995618134737,
0.03445053473114967,
0.04839850589632988,
0.07293619215488434,
0.3974873423576355,
0.028444338589906693,
-0.037606336176395416,
-0.034405630081892014,
0.03786124289035797,
-0.05949842184782028,
0.002398273441940546,
0.008856093510985374,
-0.016076376661658287,
-0.0457744263112545,
0.0525963269174099,
-0.10718114674091339,
-0.10013601183891296,
-0.029736153781414032,
-0.15176217257976532,
-0.018941957503557205,
-0.0340665839612484,
0.029755668714642525,
-0.03057999350130558,
0.030106279999017715,
-0.036690499633550644,
0.022392114624381065,
0.12109113484621048,
0.01505253091454506,
-0.0628124326467514,
0.01974402740597725,
0.03304263576865196,
0.03361867368221283,
-0.08797924220561981,
-0.07942122966051102,
0.18851080536842346,
0.05284348130226135,
0.1576567143201828,
0.04838517680764198,
0.11862004548311234,
0.03737398236989975,
0.021011872217059135,
0.033683549612760544,
0.008471015840768814,
-0.04384894296526909,
0.13185326755046844,
0.05870614945888519,
-0.13414037227630615,
0.05857141315937042,
0.1364782601594925,
-0.1379929929971695,
-0.08482132107019424,
0.031154680997133255,
-0.06998972594738007,
0.05712427571415901,
0.06417830288410187,
-0.011811723001301289,
-0.026923611760139465,
0.0012987975496798754,
-0.07945628464221954,
0.04972715303301811,
-0.08895938843488693,
-0.11440806835889816,
-0.1470527946949005,
0.02587447687983513,
-0.05655454471707344,
0.024694349616765976,
-0.2768388092517853,
-0.038744308054447174,
-0.03595997393131256,
0.026748675853013992,
-0.04822470247745514,
0.04315868020057678,
0.014874286018311977,
0.022980954498052597,
-0.04354735463857651,
0.04355164244771004,
0.010708346962928772,
0.0663217157125473,
-0.1086563840508461,
-0.1013687252998352
] |
null | null |
transformers
|
Slavic BERT from https://github.com/deepmipt/Slavic-BERT-NER http://files.deeppavlov.ai/deeppavlov_data/bg_cs_pl_ru_cased_L-12_H-768_A-12.tar.gz
|
{}
| null |
djstrong/bg_cs_pl_ru_cased_L-12_H-768_A-12
|
[
"transformers",
"pytorch",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #endpoints_compatible #region-us
|
Slavic BERT from URL URL
|
[] |
[
"TAGS\n#transformers #pytorch #endpoints_compatible #region-us \n"
] |
[
21
] |
[
"passage: TAGS\n#transformers #pytorch #endpoints_compatible #region-us \n"
] |
[
-0.0602605901658535,
-0.005646900739520788,
-0.009762155823409557,
-0.03966370224952698,
0.15944775938987732,
0.03070714697241783,
0.012395896948873997,
0.07867952436208725,
0.09419925510883331,
-0.019594743847846985,
0.09831016510725021,
0.2332964390516281,
-0.03786272928118706,
0.022073544561862946,
-0.06232732906937599,
-0.24905818700790405,
0.09407828003168106,
0.11313164979219437,
-0.06465978920459747,
0.09549489617347717,
0.039143070578575134,
-0.08911364525556564,
0.04947805404663086,
-0.032982707023620605,
-0.13527381420135498,
0.0467669740319252,
0.0237716156989336,
-0.078513965010643,
0.11869441717863083,
0.0077665625140070915,
0.19578197598457336,
0.01207544095814228,
-0.1299089938402176,
-0.18088726699352264,
0.02007104456424713,
0.022213434800505638,
-0.05996024236083031,
0.02314351126551628,
0.07312147319316864,
-0.10279249399900436,
0.02970288321375847,
0.03654928505420685,
0.002923935651779175,
0.03274611383676529,
-0.15513424575328827,
-0.1849449872970581,
-0.04425787553191185,
0.012871243990957737,
0.026669111102819443,
0.10506244748830795,
0.030372466892004013,
0.178822323679924,
-0.15745419263839722,
0.0945795327425003,
0.177890345454216,
-0.28603804111480713,
0.029024237766861916,
0.11024846136569977,
0.04829736799001694,
0.028134070336818695,
0.00996005441993475,
0.0189230814576149,
-0.009570286609232426,
0.03048190288245678,
-0.027967484667897224,
-0.07447972893714905,
-0.06405609101057053,
0.0814741775393486,
-0.09437224268913269,
-0.1216578483581543,
0.19255080819129944,
-0.05330982804298401,
0.0588051900267601,
0.03463263809680939,
-0.11766522377729416,
-0.05456177890300751,
0.0013563521206378937,
0.030292518436908722,
-0.017828812822699547,
0.07086412608623505,
0.031105060130357742,
-0.02373676560819149,
-0.12168543040752411,
0.028789518401026726,
-0.24090629816055298,
0.28345873951911926,
0.02996366284787655,
0.11018568277359009,
-0.2384001910686493,
0.07198631763458252,
-0.04607108235359192,
-0.06767431646585464,
0.023487936705350876,
-0.09875187277793884,
0.021371465176343918,
0.007170209661126137,
-0.0868486687541008,
0.025816135108470917,
0.0673765018582344,
0.11560594290494919,
-0.02462717518210411,
0.021296672523021698,
0.01767325960099697,
0.1089232936501503,
0.01622004434466362,
0.11334555596113205,
0.002745070494711399,
0.034500978887081146,
0.029799794778227806,
-0.18719491362571716,
-0.0026790881529450417,
-0.03489845618605614,
-0.08398842811584473,
-0.0881200060248375,
0.028418414294719696,
0.12114420533180237,
0.02183910645544529,
0.022260701283812523,
-0.07126972824335098,
-0.008312570862472057,
0.04512425512075424,
-0.06700387597084045,
-0.022145040333271027,
0.007661914918571711,
0.016494954004883766,
0.22294339537620544,
-0.02295338362455368,
-0.026225684210658073,
-0.07023988664150238,
0.11260082572698593,
-0.0656941682100296,
0.016149358823895454,
-0.0530007965862751,
-0.02928324043750763,
0.056476715952157974,
-0.15177644789218903,
0.06500709801912308,
-0.1516612023115158,
-0.09112907201051712,
0.028223685920238495,
0.03696250542998314,
0.028484180569648743,
0.02479529194533825,
0.015830792486667633,
-0.00879291258752346,
-0.03011348284780979,
-0.07923634350299835,
-0.08192402124404907,
-0.06342831254005432,
0.09921494126319885,
-0.0005941772251389921,
0.05436241999268532,
-0.11821827292442322,
0.07408127188682556,
-0.1076948270201683,
0.030249644070863724,
-0.13414455950260162,
-0.0221982691437006,
-0.02404189109802246,
0.1851358413696289,
0.0006879806751385331,
-0.0817047506570816,
-0.10663671791553497,
0.038237277418375015,
-0.04168618097901344,
0.13458840548992157,
-0.014303965494036674,
-0.11019430309534073,
0.26959696412086487,
-0.10667987167835236,
-0.1671270877122879,
0.05378620699048042,
0.009700759314000607,
-0.016868380829691887,
0.0561099536716938,
0.16616025567054749,
0.06831997632980347,
-0.07049598544836044,
0.09140615165233612,
0.1379895955324173,
-0.1851922571659088,
-0.20243999361991882,
0.01763749308884144,
-0.05203584209084511,
-0.11545780301094055,
0.04540516808629036,
-0.0058397711254656315,
0.09395778924226761,
-0.08503725379705429,
-0.007578795775771141,
-0.032015107572078705,
-0.016012731939554214,
0.07450172305107117,
0.06282669305801392,
0.09098648279905319,
-0.05122196301817894,
0.031217509880661964,
0.026435445994138718,
0.00463126040995121,
0.0063572051003575325,
0.05537392571568489,
-0.03592291474342346,
0.1321462094783783,
-0.06449703127145767,
0.00637860456481576,
-0.2316930741071701,
-0.11218059808015823,
-0.012119117192924023,
0.04616983234882355,
-0.05828554928302765,
0.16098164021968842,
0.09591125696897507,
-0.0814671590924263,
0.027995459735393524,
-0.03218327462673187,
0.10552480816841125,
0.02011469565331936,
-0.012835591100156307,
-0.0011234998237341642,
0.016821393743157387,
-0.0699923112988472,
-0.08323190361261368,
0.006836527958512306,
-0.0009020745637826622,
0.08346128463745117,
0.12389257550239563,
-0.0018713462632149458,
0.04892571270465851,
-0.025174804031848907,
0.07054726779460907,
-0.016440844163298607,
0.014167504385113716,
0.10722272843122482,
-0.011425090953707695,
-0.05193426087498665,
0.1583033800125122,
-0.11112764477729797,
0.3326278030872345,
0.20375320315361023,
-0.323421835899353,
0.04571422189474106,
-0.025362668558955193,
-0.017437715083360672,
0.021043121814727783,
0.09183337539434433,
0.005831574089825153,
0.08254498243331909,
0.05105554684996605,
0.13220374286174774,
-0.025855854153633118,
-0.015603546984493732,
-0.000933936215005815,
-0.06394030898809433,
-0.03657902777194977,
0.07155993580818176,
0.06678375601768494,
-0.12739983201026917,
0.17137302458286285,
0.2332899272441864,
0.034409500658512115,
0.08888798952102661,
-0.074738509953022,
-0.02466125227510929,
0.05307190865278244,
0.030478334054350853,
-0.05014907941222191,
-0.003522941842675209,
-0.24837985634803772,
-0.033807482570409775,
0.0835256576538086,
0.041713811457157135,
0.11864139884710312,
-0.15303030610084534,
-0.0535346120595932,
0.036216262727975845,
0.003446921007707715,
-0.07469423115253448,
0.0990411564707756,
0.07169649004936218,
0.06572583317756653,
0.006040885578840971,
-0.030352378264069557,
0.11742591857910156,
0.000851878838147968,
-0.05391429737210274,
0.17059999704360962,
-0.13313265144824982,
-0.29761773347854614,
-0.14319129288196564,
-0.1334514170885086,
0.010339286178350449,
0.017076298594474792,
0.08161275088787079,
-0.08419821411371231,
-0.034329503774642944,
0.09296482056379318,
0.05038211867213249,
-0.12507256865501404,
0.025089209899306297,
-0.04892471432685852,
0.06829681247472763,
-0.08489015698432922,
-0.07859387248754501,
-0.06656183302402496,
-0.05509684979915619,
-0.04375450685620308,
0.11071989685297012,
-0.12806063890457153,
0.08490876853466034,
0.13819533586502075,
0.031051602214574814,
0.07859103381633759,
0.001322271185927093,
0.13857795298099518,
-0.061774931848049164,
-0.08148285746574402,
0.23385483026504517,
-0.009697506204247475,
0.10181795805692673,
0.10921594500541687,
0.02137605845928192,
-0.06395251303911209,
-0.022054238244891167,
-0.07990756630897522,
-0.12127618491649628,
-0.2081342488527298,
-0.13152581453323364,
-0.147438645362854,
0.0003129298856947571,
0.0026985483709722757,
0.04901242256164551,
0.08136387914419174,
0.07752165198326111,
0.05862151086330414,
-0.08998161554336548,
-0.05513061210513115,
0.057605329900979996,
0.2182815968990326,
-0.01885073445737362,
0.07730096578598022,
-0.08762159198522568,
-0.0858968198299408,
0.07550632208585739,
0.0809127613902092,
0.22168903052806854,
0.0703313797712326,
0.03154223784804344,
0.06092943996191025,
0.1844511479139328,
0.15529003739356995,
0.17153839766979218,
0.015418988652527332,
-0.025812845677137375,
0.006589105818420649,
0.013138419017195702,
-0.09150379151105881,
-0.0042123449966311455,
0.1292845457792282,
-0.14540119469165802,
-0.07880084216594696,
-0.22943291068077087,
0.08157958835363388,
0.07344048470258713,
0.026414161548018456,
-0.17556782066822052,
-0.0015733868349343538,
0.05357646942138672,
0.006558762397617102,
-0.045512180775403976,
0.08496551215648651,
-0.022116927430033684,
-0.13344039022922516,
0.04287152364850044,
-0.058599162846803665,
0.10456542670726776,
-0.038338202983140945,
0.07373929768800735,
-0.024396726861596107,
-0.10900583118200302,
0.07474620640277863,
0.08851087838411331,
-0.23624686896800995,
0.27773332595825195,
-0.028285326436161995,
-0.07041611522436142,
-0.06830666214227676,
-0.024676870554685593,
-0.0017557048704475164,
0.17030403017997742,
0.08692118525505066,
0.03860539570450783,
-0.06731099635362625,
-0.1669522523880005,
0.04732859507203102,
0.02085104025900364,
0.12389091402292252,
-0.008113703690469265,
-0.030854588374495506,
-0.01338487584143877,
-0.019302181899547577,
-0.041759077459573746,
-0.003546249819919467,
0.10703662782907486,
-0.13412658870220184,
0.034535422921180725,
0.005309549160301685,
0.04036608338356018,
-0.0035966450814157724,
0.005690612830221653,
-0.05699379742145538,
0.13448239862918854,
-0.04070858657360077,
-0.07688865810632706,
-0.08865337073802948,
-0.17243140935897827,
0.13381044566631317,
-0.10238062590360641,
0.09249105304479599,
-0.09976985305547714,
-0.06192123889923096,
-0.0650167390704155,
-0.21104402840137482,
0.10917928814888,
-0.10513100028038025,
0.0401403047144413,
-0.016867417842149734,
0.202356219291687,
-0.08008989691734314,
-0.0209506843239069,
-0.0013599899830296636,
0.004461138043552637,
-0.12345174700021744,
-0.10132946074008942,
-0.01779748499393463,
0.029711758717894554,
0.08279082924127579,
0.10770376026630402,
-0.018299801275134087,
0.04191994294524193,
0.013882452622056007,
0.03635424003005028,
0.24111443758010864,
0.13711073994636536,
-0.04645780101418495,
0.11812267452478409,
0.16186927258968353,
-0.031571563333272934,
-0.2709476351737976,
-0.0868004634976387,
-0.16869930922985077,
-0.04678983986377716,
-0.037189606577157974,
-0.12448763102293015,
0.12788352370262146,
0.04586293175816536,
-0.011544082313776016,
0.1266224980354309,
-0.26594868302345276,
-0.03401912376284599,
0.1643364280462265,
0.027830148115754128,
0.48583653569221497,
-0.09991659969091415,
-0.0845755785703659,
0.018966181203722954,
-0.2915274500846863,
0.10791198909282684,
0.01369303185492754,
0.06487411260604858,
-0.034079212695360184,
0.10230494290590286,
0.044344354420900345,
-0.10719364136457443,
0.11951436847448349,
0.035276297479867935,
0.03043992444872856,
-0.08126169443130493,
-0.08180311322212219,
0.043867211788892746,
-0.020314449444413185,
0.016103362664580345,
0.09817174077033997,
0.036888349801301956,
-0.1322462111711502,
-0.022591179236769676,
-0.13765397667884827,
0.049294743686914444,
0.0620623379945755,
-0.0240910816937685,
-0.013253096491098404,
-0.036911722272634506,
-0.004203932359814644,
0.019186172634363174,
0.25098612904548645,
-0.012744947336614132,
0.1315464824438095,
0.042376402765512466,
0.046371813863515854,
-0.20250053703784943,
-0.14301706850528717,
-0.07346653193235397,
-0.04091315343976021,
0.0825582891702652,
-0.04023446887731552,
0.0523495189845562,
0.1589740812778473,
-0.02755962871015072,
-0.008571553975343704,
0.1279488205909729,
0.014891285449266434,
-0.03108718805015087,
0.12318851798772812,
-0.21083922684192657,
-0.042206212878227234,
-0.018448445945978165,
-0.05149201303720474,
0.1258767545223236,
0.12660184502601624,
0.08982381969690323,
0.07138726115226746,
-0.016965264454483986,
-0.03962934389710426,
-0.023878682404756546,
-0.0698867067694664,
0.014531198889017105,
0.041909199208021164,
0.03934765234589577,
-0.13099561631679535,
0.07754374295473099,
-0.018086964264512062,
-0.27273428440093994,
-0.06408093869686127,
0.09851112216711044,
-0.15818077325820923,
-0.10192093253135681,
-0.06077379360795021,
0.03578856587409973,
-0.1768866330385208,
-0.0624033585190773,
-0.0316467322409153,
-0.10826759040355682,
0.06394574791193008,
0.20973356068134308,
0.10992377996444702,
0.12001070380210876,
-0.023827049881219864,
-0.02570408768951893,
-0.00579385831952095,
-0.10135087370872498,
-0.012352810241281986,
0.013960975222289562,
-0.09324274212121964,
0.018809983506798744,
-0.009077025577425957,
0.17904941737651825,
-0.0752621665596962,
-0.08742223680019379,
-0.1552542746067047,
0.09289714694023132,
-0.1066979393362999,
-0.092626191675663,
-0.14189116656780243,
-0.06335672736167908,
0.023829210549592972,
-0.06568517535924911,
-0.01769612915813923,
-0.013204174116253853,
-0.14357970654964447,
0.06973201036453247,
0.01112452708184719,
-0.011419128626585007,
-0.057545971125364304,
-0.02905389852821827,
0.12901906669139862,
-0.0439501516520977,
0.08620632439851761,
0.19515031576156616,
-0.09635450690984726,
0.12672042846679688,
-0.10680265724658966,
-0.16728007793426514,
0.109873466193676,
0.011463385075330734,
0.08858676254749298,
0.05595776438713074,
0.033014263957738876,
0.0949346125125885,
0.01360013522207737,
0.040659237653017044,
-0.0074751065112650394,
-0.13721472024917603,
0.0005411148304119706,
0.014451061375439167,
-0.14818744361400604,
-0.051526814699172974,
-0.10024616122245789,
0.15037848055362701,
0.046998120844364166,
0.10320950299501419,
0.02188979648053646,
0.13145636022090912,
-0.006807954981923103,
-0.01749216392636299,
-0.007913168519735336,
-0.18110167980194092,
0.058051977306604385,
-0.04888135939836502,
0.00664436025545001,
0.004136093892157078,
0.29094672203063965,
-0.05342133343219757,
0.06421013921499252,
0.012430429458618164,
0.05440686270594597,
0.025291895493865013,
0.019776156172156334,
0.23332199454307556,
0.0909920483827591,
-0.040104810148477554,
-0.07203774154186249,
0.09381944686174393,
-0.027183765545487404,
-0.051167625933885574,
0.14428386092185974,
0.12433760613203049,
0.046564988791942596,
0.10233303159475327,
-0.010628857649862766,
0.0502619631588459,
-0.11714580655097961,
-0.29435333609580994,
-0.0017536553787067533,
0.02923153154551983,
0.013082349672913551,
0.09705042093992233,
0.12465932965278625,
-0.014661809429526329,
0.09902279824018478,
-0.0014873039908707142,
-0.026829617097973824,
-0.13841837644577026,
-0.08308025449514389,
-0.037351641803979874,
-0.1296062022447586,
0.01537085510790348,
-0.06261784583330154,
-0.0013676333473995328,
0.20482417941093445,
0.040537029504776,
-0.03484100475907326,
0.11699468642473221,
0.0698540061712265,
-0.06982901692390442,
0.02712978608906269,
-0.00757982861250639,
0.035655613988637924,
0.06632175296545029,
-0.023149633780121803,
-0.14352689683437347,
-0.08878929167985916,
-0.04016351327300072,
0.04570772126317024,
-0.0825280100107193,
0.007398216053843498,
-0.1344616711139679,
-0.09811436384916306,
-0.05467144027352333,
0.09024309366941452,
-0.076502226293087,
0.12816135585308075,
-0.010830878280103207,
0.010613616555929184,
0.01268461812287569,
0.18772324919700623,
-0.075274258852005,
-0.060353536158800125,
0.002831645542755723,
0.23638813197612762,
0.08521586656570435,
0.08294188231229782,
0.00997002050280571,
0.024029415100812912,
-0.09174544364213943,
0.3311961889266968,
0.24609249830245972,
-0.010904072783887386,
0.04334266111254692,
0.04531332477927208,
0.04185611754655838,
0.12315969169139862,
0.13169057667255402,
0.11807206273078918,
0.344115674495697,
-0.08338891714811325,
-0.031944580376148224,
-0.030247550457715988,
0.006968794856220484,
-0.13217425346374512,
0.026531271636486053,
0.01758533902466297,
-0.08070562034845352,
-0.07679492235183716,
0.11993145197629929,
-0.20403271913528442,
0.15280351042747498,
0.08174222707748413,
-0.19685740768909454,
-0.03396262973546982,
-0.05789715424180031,
0.18384911119937897,
-0.014278654009103775,
0.11859142035245895,
-0.04448321461677551,
-0.13445675373077393,
0.08106808364391327,
0.048652783036231995,
-0.2895694077014923,
-0.08014793694019318,
0.0981469452381134,
0.042115915566682816,
-0.0287695974111557,
-0.015584125183522701,
0.03546717017889023,
0.059928759932518005,
0.08986898511648178,
-0.03891289234161377,
0.04509701579809189,
0.020416583865880966,
-0.10908327251672745,
-0.08358649909496307,
-0.02032429538667202,
-0.00014163613377604634,
-0.1349317878484726,
0.016364745795726776,
-0.20526979863643646,
0.04226694256067276,
-0.010092525742948055,
0.01465508621186018,
-0.004206623882055283,
-0.03162108361721039,
-0.051075391471385956,
0.023095570504665375,
0.052028998732566833,
0.009046600200235844,
-0.017174653708934784,
-0.043978288769721985,
-0.01736217923462391,
0.04129228740930557,
-0.09392837435007095,
-0.15831652283668518,
-0.016197707504034042,
-0.08371548354625702,
0.10397595167160034,
-0.03304345905780792,
-0.046627145260572433,
-0.017865123227238655,
-0.02954019047319889,
0.04380740970373154,
-0.10989689826965332,
0.035784848034381866,
0.01109471544623375,
0.05120820179581642,
0.011953367851674557,
-0.01127829309552908,
0.039223846048116684,
0.07814694195985794,
-0.12265297025442123,
-0.1024811714887619
] |
null | null |
transformers
|
# Harry Potter DialoGPT Model
|
{"tags": ["conversational"]}
|
text-generation
|
dk16gaming/DialoGPT-small-HarryPotter
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Harry Potter DialoGPT Model
|
[
"# Harry Potter DialoGPT Model"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Harry Potter DialoGPT Model"
] |
[
51,
8
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Harry Potter DialoGPT Model"
] |
[
-0.0009023238671943545,
0.07815738022327423,
-0.006546166725456715,
0.07792752981185913,
0.10655936598777771,
0.048972971737384796,
0.17639793455600739,
0.12185695022344589,
0.016568755730986595,
-0.04774167761206627,
0.11647630482912064,
0.2130284160375595,
-0.002118367003276944,
0.024608047679066658,
-0.05022026598453522,
-0.3065771162509918,
0.0474756620824337,
0.014356585219502449,
-0.07174845039844513,
0.11724270135164261,
0.09064973145723343,
-0.046179238706827164,
0.08330509811639786,
-0.009135239757597446,
-0.13198648393154144,
-0.039482954889535904,
0.019292812794446945,
-0.11745545268058777,
0.1662212759256363,
0.05298272892832756,
0.02469746209681034,
-0.008447164669632912,
-0.06598151475191116,
-0.15036040544509888,
0.037190426141023636,
-0.027472136542201042,
-0.01080626156181097,
0.05462246760725975,
0.023526115342974663,
-0.07521048933267593,
0.170567125082016,
0.17678891122341156,
0.0833497866988182,
0.0349111407995224,
-0.14917024970054626,
-0.045548245310783386,
0.008950977586209774,
0.05421316996216774,
-0.017893504351377487,
0.09349167346954346,
-0.019903047010302544,
0.11801653355360031,
-0.04491448402404785,
0.09210366010665894,
0.15255063772201538,
-0.4016275703907013,
-0.027563704177737236,
0.08920855820178986,
0.05989706888794899,
0.12076901644468307,
-0.10560955852270126,
0.03972794860601425,
-0.0039703017100691795,
0.01236654631793499,
-0.014540530741214752,
-0.08304883539676666,
-0.07308239489793777,
0.032504837960004807,
-0.1272556483745575,
0.008525865152478218,
0.23756256699562073,
-0.10643257945775986,
0.037069112062454224,
-0.09791990369558334,
-0.07414398342370987,
0.048336777836084366,
-0.053761593997478485,
-0.081727035343647,
-0.054839808493852615,
0.06347949057817459,
0.004366500303149223,
-0.06301609426736832,
-0.08326146006584167,
-0.0006536149303428829,
-0.12781435251235962,
0.17595994472503662,
0.061243366450071335,
0.041611745953559875,
-0.21322020888328552,
0.08940251916646957,
0.04477722570300102,
-0.04711297154426575,
0.007116159424185753,
-0.11796226352453232,
0.04023287072777748,
0.005483259446918964,
-0.03256071358919144,
-0.021854614838957787,
0.0393419973552227,
0.13909944891929626,
-0.01777748204767704,
0.03252175822854042,
0.006831915583461523,
0.05811219662427902,
0.08162496984004974,
0.02222144603729248,
0.019291909411549568,
-0.0818009302020073,
0.019385190680623055,
-0.08128736168146133,
-0.0030400939285755157,
-0.048940129578113556,
-0.17071883380413055,
-0.07477642595767975,
0.052610911428928375,
0.020047198981046677,
0.03746970370411873,
0.08054786175489426,
-0.0017944995779544115,
-0.05560554191470146,
0.03284840285778046,
0.01671096310019493,
-0.020622212439775467,
-0.010361049324274063,
-0.02412462793290615,
0.19123271107673645,
0.019619356840848923,
0.014111656695604324,
-0.12379156798124313,
0.10023640841245651,
-0.08179095387458801,
0.0037731381598860025,
0.02743307314813137,
-0.04204464703798294,
-0.004716555587947369,
0.02917117439210415,
0.023101668804883957,
-0.1252521574497223,
-0.1099385917186737,
-0.0030569476075470448,
-0.012054097838699818,
-0.036421261727809906,
-0.10490952432155609,
-0.08483029156923294,
-0.012153145857155323,
0.0449371263384819,
-0.013397793285548687,
0.007936403155326843,
-0.05143149942159653,
0.0985720232129097,
-0.0514979362487793,
0.09873400628566742,
-0.08342572301626205,
0.06359215080738068,
-0.09124887734651566,
-0.061886150389909744,
-0.11452563107013702,
0.05216052383184433,
0.012905281968414783,
0.066250741481781,
0.016998225823044777,
-0.044836658984422684,
-0.014836243353784084,
0.05253177136182785,
-0.07656687498092651,
0.1940697431564331,
-0.041674621403217316,
-0.12459053844213486,
0.24146439135074615,
-0.09138800948858261,
-0.1802034229040146,
0.12973085045814514,
-0.022254703566432,
0.08523941785097122,
0.12802475690841675,
0.20380465686321259,
-0.00019822151807602495,
-0.01302915159612894,
0.07281201332807541,
0.07031642645597458,
-0.09803894907236099,
0.06239739805459976,
0.029653839766979218,
-0.008071083575487137,
-0.08906278014183044,
0.05762826278805733,
0.046033453196287155,
-0.010650773532688618,
-0.035073768347501755,
-0.001896020956337452,
-0.012895751744508743,
-0.022185025736689568,
0.14126582443714142,
-0.02006692811846733,
0.1300428807735443,
-0.06926563382148743,
-0.03515486419200897,
-0.009500149637460709,
0.03533667325973511,
-0.04091939330101013,
0.08151165395975113,
-0.0436173714697361,
0.10586477071046829,
0.09034156054258347,
0.053724925965070724,
-0.13120363652706146,
0.00466286763548851,
-0.015246815048158169,
0.17014820873737335,
0.08964069187641144,
0.05222717300057411,
0.06265474855899811,
-0.0020888058934360743,
-0.06708643585443497,
0.045407816767692566,
0.13778303563594818,
-0.037020038813352585,
-0.12218865007162094,
-0.1755627691745758,
0.051157694309949875,
-0.045444171875715256,
0.10855234414339066,
-0.10010123997926712,
0.022670533508062363,
-0.055906031280756,
0.07772238552570343,
-0.024998966604471207,
0.020512236282229424,
-0.0013405600329861045,
-0.021700702607631683,
-0.08356887847185135,
-0.002377772703766823,
0.08597290515899658,
-0.02048647589981556,
-0.06707409024238586,
0.16556480526924133,
-0.16400809586048126,
0.1631954461336136,
0.2116095870733261,
-0.28542569279670715,
-0.005696662236005068,
-0.15163889527320862,
-0.0208092350512743,
0.019645055755972862,
0.07834604382514954,
0.026225795969367027,
0.2044338881969452,
-0.012928472831845284,
0.16565458476543427,
-0.05699567869305611,
-0.07730039209127426,
-0.06881127506494522,
-0.048101142048835754,
0.013522743247449398,
0.09095205366611481,
0.04542696103453636,
-0.11962861567735672,
0.13119758665561676,
0.1054433062672615,
0.06484298408031464,
0.12711186707019806,
0.1030748188495636,
-0.008113685995340347,
0.07252490520477295,
-0.03624548763036728,
-0.03462279960513115,
-0.09254947304725647,
-0.30446043610572815,
-0.04840317741036415,
0.0939924493432045,
0.007963384501636028,
0.09285714477300644,
-0.0919896736741066,
-0.03311870992183685,
0.006042704917490482,
0.009473444893956184,
0.028337622061371803,
0.09653715789318085,
0.013490920886397362,
0.15320514142513275,
-0.008011690340936184,
-0.03430786728858948,
0.05891305208206177,
0.017982570454478264,
-0.09147711098194122,
0.17280617356300354,
-0.17050009965896606,
-0.27190929651260376,
-0.06990014761686325,
-0.21745692193508148,
-0.013139115646481514,
0.05258983001112938,
0.0786920040845871,
-0.11818131804466248,
-0.018352627754211426,
-0.006239492911845446,
0.05685517191886902,
-0.2425733357667923,
0.0004911290016025305,
-0.1354890614748001,
0.0501418262720108,
-0.1974833607673645,
-0.09718500077724457,
-0.02271542325615883,
-0.013450481928884983,
-0.0464281290769577,
0.13365240395069122,
-0.1448695808649063,
-0.011572926305234432,
0.2329535037279129,
0.032479673624038696,
0.027794739231467247,
-0.05020907148718834,
0.19788463413715363,
-0.0958966314792633,
-0.023973820731043816,
0.11024576425552368,
-0.05038975924253464,
0.04834126681089401,
0.06649978458881378,
-0.012981836684048176,
-0.08557141572237015,
0.023789849132299423,
-0.068336620926857,
-0.03150583803653717,
-0.27926525473594666,
-0.0930178239941597,
-0.09319330751895905,
0.11305391043424606,
0.04079577326774597,
0.06421639025211334,
0.16545771062374115,
0.05191578343510628,
-0.024325082078576088,
-0.03006586618721485,
0.11609793454408646,
0.12905290722846985,
0.2277202159166336,
-0.06067761778831482,
0.10221996158361435,
0.009445492178201675,
-0.08203992247581482,
0.06062209978699684,
0.056782789528369904,
0.06324724853038788,
0.02584579586982727,
0.03694582358002663,
-0.030939655378460884,
0.1121687963604927,
0.12571842968463898,
0.05258069559931755,
0.0481170229613781,
0.0002127334737451747,
-0.0561506561934948,
-0.008168719708919525,
-0.05726633965969086,
0.06774696707725525,
0.061340972781181335,
-0.12918008863925934,
-0.08061543852090836,
0.0011613310780376196,
0.06660808622837067,
-0.016230419278144836,
0.06823775917291641,
-0.13560809195041656,
-0.03582429885864258,
0.0790911465883255,
-0.07693151384592056,
-0.14156894385814667,
0.11972879618406296,
-0.026570770889520645,
-0.19904157519340515,
0.05265914276242256,
0.007704653777182102,
0.0908159390091896,
-0.06360849738121033,
0.05343840271234512,
-0.13023801147937775,
-0.12935101985931396,
-0.018437571823596954,
0.07945099472999573,
-0.3450873792171478,
0.13536721467971802,
-0.013286802917718887,
-0.02876877970993519,
-0.06474969536066055,
-0.02640824392437935,
0.013905409723520279,
0.12719078361988068,
0.08667250722646713,
0.0008821099763736129,
0.0991629809141159,
0.03823768347501755,
0.04188435152173042,
-0.002011700300499797,
0.10950417071580887,
0.0050011589191854,
0.004797275178134441,
-0.04982118681073189,
0.007274609990417957,
-0.05164213851094246,
-0.07472953200340271,
0.08393982797861099,
-0.20678792893886566,
0.09087453782558441,
-0.03378438204526901,
0.08427679538726807,
0.04304937273263931,
-0.018965769559144974,
-0.1001204177737236,
0.19745583832263947,
-0.012206900864839554,
-0.11405988782644272,
-0.07517550885677338,
-0.02810264565050602,
0.09103139489889145,
-0.013817726634442806,
0.012886416167020798,
-0.045470476150512695,
0.032183047384023666,
-0.1263762265443802,
-0.1597503274679184,
0.08734500408172607,
-0.04441224783658981,
-0.10894393920898438,
-0.025462759658694267,
0.20382575690746307,
-0.007266622502356768,
0.08242089301347733,
0.01605331338942051,
0.010653935372829437,
-0.18066231906414032,
-0.04018142446875572,
0.02645772136747837,
-0.0016437612939625978,
0.005979063920676708,
0.047698814421892166,
0.019091911613941193,
0.06207629665732384,
-0.1069745197892189,
-0.013920160941779613,
0.3158324360847473,
0.15978319942951202,
-0.00912671908736229,
0.14943915605545044,
0.1093616932630539,
-0.08669080585241318,
-0.17238758504390717,
-0.1171615794301033,
-0.1210922971367836,
-0.08425768464803696,
-0.10681738704442978,
-0.1525043100118637,
0.09535340964794159,
-0.03392014652490616,
0.03498011827468872,
0.14615866541862488,
-0.280263751745224,
-0.10949636250734329,
0.13820378482341766,
0.010744688101112843,
0.3510635495185852,
-0.12303631007671356,
-0.044944874942302704,
-0.06214528530836105,
-0.16933435201644897,
0.08021392673254013,
-0.031203703954815865,
0.11581093072891235,
-0.0744495838880539,
0.19395925104618073,
0.01719796098768711,
0.014287159778177738,
0.0916559100151062,
0.05038322135806084,
-0.05808406323194504,
-0.07368700206279755,
-0.10248131304979324,
0.010812131687998772,
0.03546109423041344,
0.010252019390463829,
-0.008802837692201138,
0.0211968794465065,
-0.11341743916273117,
-0.050869911909103394,
-0.06302189081907272,
0.0072614275850355625,
-0.01001308299601078,
-0.042155615985393524,
-0.05533592775464058,
-0.022557416930794716,
-0.020093943923711777,
0.02266426384449005,
0.14185629785060883,
-0.07527699321508408,
0.18586260080337524,
0.02357078716158867,
0.1586609035730362,
-0.11956068128347397,
-0.06724818795919418,
-0.029193658381700516,
-0.05280323326587677,
0.06468886137008667,
-0.08884575963020325,
-0.027708567678928375,
0.1332162618637085,
-0.01903904788196087,
0.04655366763472557,
0.12936700880527496,
0.02046884410083294,
0.015383756719529629,
0.034968774765729904,
-0.2578005790710449,
-0.07463036477565765,
-0.03505445644259453,
-0.012416874058544636,
0.05272092670202255,
0.05525677278637886,
0.19735674560070038,
-0.03551921248435974,
-0.08521962910890579,
0.020131373777985573,
0.02735883742570877,
-0.02776256389915943,
0.10749414563179016,
0.019579345360398293,
-0.004837906453758478,
-0.16151933372020721,
0.08257976174354553,
-0.005964108742773533,
-0.08297000825405121,
0.028665626421570778,
0.2024049311876297,
-0.12141239643096924,
-0.10309756547212601,
-0.06804922968149185,
0.07315051555633545,
-0.09220825880765915,
0.016043387353420258,
-0.005091092549264431,
-0.1521538347005844,
0.06916408240795135,
0.07598215341567993,
0.04075418785214424,
0.06513199955224991,
-0.11743064224720001,
-0.015730571001768112,
-0.04170290008187294,
-0.002195435343310237,
0.03521120920777321,
0.01863143965601921,
-0.057492829859256744,
0.15846455097198486,
-0.0676199421286583,
0.08538917452096939,
-0.0744810476899147,
-0.1058846190571785,
-0.1395980566740036,
0.04660497233271599,
-0.08038312196731567,
-0.07247276604175568,
-0.12832807004451752,
-0.052204377949237823,
-0.0067099276930093765,
-0.03388519585132599,
0.006552806124091148,
-0.06627799570560455,
-0.10922821611166,
0.01822470687329769,
-0.00743203004822135,
-0.009385870769619942,
-0.06096754968166351,
0.026706209406256676,
0.06246216222643852,
-0.039788868278265,
0.15730851888656616,
0.22509248554706573,
-0.13591648638248444,
0.11564400047063828,
-0.09797432273626328,
-0.105463907122612,
0.046008042991161346,
0.009427277371287346,
0.03594303876161575,
0.0503489226102829,
-0.03594081476330757,
0.0044484552927315235,
0.03905477747321129,
0.08074651658535004,
0.08456914126873016,
-0.06776505708694458,
0.020801106467843056,
-0.05122765153646469,
-0.14904099702835083,
-0.016655439510941505,
-0.0464773029088974,
0.06876829266548157,
-0.006725262850522995,
0.11020535975694656,
-0.0515950471162796,
0.07739507406949997,
-0.07558431476354599,
0.050614211708307266,
0.021146971732378006,
-0.14688286185264587,
-0.006612539757043123,
-0.07093682140111923,
0.042144812643527985,
-0.008834975771605968,
0.20241086184978485,
-0.03228091076016426,
0.010342049412429333,
0.033811055123806,
0.06203942745923996,
-0.01957780309021473,
0.009357001632452011,
0.2014283686876297,
0.12640917301177979,
-0.08496357500553131,
-0.02679651789367199,
0.06793134659528732,
0.07248228788375854,
0.07093550264835358,
0.10807815194129944,
-0.015352966263890266,
0.028434239327907562,
0.07829629629850388,
-0.060215238481760025,
0.07576877623796463,
-0.08603982627391815,
-0.11668483167886734,
0.05793621391057968,
0.012955795042216778,
-0.055695828050374985,
0.20305177569389343,
0.19142870604991913,
-0.026278704404830933,
0.018410727381706238,
-0.0029499190859496593,
-0.10117456316947937,
-0.15619947016239166,
-0.05423750728368759,
-0.07170962542295456,
-0.1319410353899002,
-0.004549739416688681,
-0.16646917164325714,
0.022016216069459915,
-0.01132756657898426,
0.09506805986166,
-0.06855440139770508,
-0.01345991250127554,
0.1364889293909073,
-0.1055467277765274,
0.0847758799791336,
-0.024517204612493515,
0.07877567410469055,
-0.03746940940618515,
-0.018209461122751236,
-0.10342709720134735,
0.007514837197959423,
0.01131442841142416,
0.06840907037258148,
-0.10897937417030334,
0.02432350255548954,
-0.12208317965269089,
-0.08617185056209564,
-0.026142612099647522,
0.09279687702655792,
-0.0403008833527565,
0.15116846561431885,
0.02645145356655121,
-0.06710928678512573,
-0.004313822835683823,
0.2646709978580475,
-0.08046227693557739,
-0.08319197595119476,
-0.030799202620983124,
0.2152107208967209,
0.04053696244955063,
0.06396269053220749,
0.019140036776661873,
0.038027774542570114,
-0.07184682041406631,
0.2957373559474945,
0.34401440620422363,
-0.1318037211894989,
-0.007773484103381634,
0.04225075617432594,
0.04406323283910751,
0.14687567949295044,
0.07998795062303543,
0.11360671371221542,
0.2849363386631012,
-0.09197647124528885,
0.016657205298542976,
-0.04230864346027374,
-0.01424806285649538,
-0.06908884644508362,
0.045314885675907135,
0.08216670155525208,
-0.09241747111082077,
-0.022950593382120132,
0.08125471323728561,
-0.29741767048835754,
0.10791494697332382,
-0.15600289404392242,
-0.14948409795761108,
-0.05027429759502411,
-0.008771711029112339,
0.014683255925774574,
0.019041186198592186,
0.09663030505180359,
0.025651484727859497,
-0.07275258749723434,
0.07816889137029648,
0.024486342445015907,
-0.23020237684249878,
-0.01345184724777937,
0.1456068754196167,
-0.06789913028478622,
-0.025938833132386208,
-0.021313713863492012,
0.051610056310892105,
0.05763651058077812,
0.09027529507875443,
-0.03809558227658272,
-0.0746568813920021,
-0.007141788024455309,
-0.022818787023425102,
0.01914946548640728,
0.0597183033823967,
0.06841408461332321,
-0.0920223817229271,
0.1167774423956871,
-0.07350476831197739,
0.0650370642542839,
0.037623800337314606,
-0.022277191281318665,
0.0018526542698964477,
0.013183658011257648,
-0.06512464582920074,
0.05533479526638985,
0.1295643299818039,
-0.025459708645939827,
-0.002524374984204769,
-0.028180841356515884,
-0.0767761766910553,
-0.024015206843614578,
-0.04643676429986954,
-0.09101243317127228,
-0.18130090832710266,
-0.12738600373268127,
0.041754670441150665,
-0.03240608796477318,
-0.2046082615852356,
0.0060346988029778,
-0.1128578633069992,
0.03700976446270943,
-0.14154092967510223,
0.10004086047410965,
0.07216610759496689,
0.004716616589576006,
0.006774604320526123,
0.0675399899482727,
0.045677728950977325,
0.14796748757362366,
-0.16543124616146088,
-0.04919974133372307
] |
null | null |
transformers
|
### Bert-News
|
{}
|
text-classification
|
dkhara/bert-news
|
[
"transformers",
"pytorch",
"distilbert",
"text-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us
|
### Bert-News
|
[
"### Bert-News"
] |
[
"TAGS\n#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n",
"### Bert-News"
] |
[
38,
5
] |
[
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #autotrain_compatible #endpoints_compatible #region-us \n### Bert-News"
] |
[
-0.02364049293100834,
0.07627793401479721,
-0.007571836933493614,
0.03925863653421402,
0.17545193433761597,
0.02019600011408329,
0.06997992843389511,
0.09922856837511063,
0.013594318181276321,
-0.01616578921675682,
0.12612836062908173,
0.22770151495933533,
-0.050239454954862595,
0.09924932569265366,
-0.11933780461549759,
-0.2899649441242218,
0.07174346596002579,
0.07970812916755676,
0.022188102826476097,
0.11179561167955399,
0.07259754091501236,
-0.10484284907579422,
0.06886697560548782,
-0.031101590022444725,
-0.1010432168841362,
0.043461620807647705,
0.02244756557047367,
-0.1159396544098854,
0.09909924119710922,
0.0479208268225193,
0.13879737257957458,
0.03820277377963066,
-0.052659593522548676,
-0.12195339798927307,
0.043265972286462784,
-0.013760521076619625,
-0.0905892625451088,
0.05929204076528549,
0.06218789517879486,
-0.11490771919488907,
-0.02772023342549801,
-0.01281022559851408,
0.0026688564103096724,
0.03640817478299141,
-0.14273183047771454,
-0.08626460283994675,
-0.009022860787808895,
0.042711399495601654,
0.08326909691095352,
0.07116976380348206,
-0.004004960879683495,
0.12347319722175598,
-0.1335848718881607,
0.12066721171140671,
0.1586311012506485,
-0.2821977734565735,
-0.017155587673187256,
0.048797618597745895,
0.023046761751174927,
-0.004959446843713522,
-0.05928420647978783,
0.05277465656399727,
0.049065347760915756,
0.0002406315616099164,
-0.013784798793494701,
-0.06616618484258652,
-0.09624265134334564,
0.01333617139607668,
-0.09560824185609818,
-0.027957307174801826,
0.1720220297574997,
-0.05345803499221802,
0.06244182959198952,
-0.003445160575211048,
-0.08394285291433334,
-0.03366858512163162,
-0.028311317786574364,
0.02401750721037388,
-0.05391176789999008,
0.0578981451690197,
0.023094337433576584,
0.010026770643889904,
-0.10615435242652893,
0.04893510043621063,
-0.2149243801832199,
0.26581627130508423,
0.012817240320146084,
0.029606729745864868,
-0.14460287988185883,
0.06545938551425934,
0.02901901677250862,
-0.13005909323692322,
0.053416647017002106,
-0.0922233909368515,
0.07647436112165451,
-0.03415011987090111,
-0.05231016129255295,
-0.05786892771720886,
0.051201414316892624,
0.10214821994304657,
-0.02550904080271721,
0.04510224238038063,
-0.04683361575007439,
0.08644738793373108,
0.0438392274081707,
0.07638292759656906,
0.0008997514378279448,
-0.024266740307211876,
0.02412283979356289,
-0.09677200019359589,
0.018509283661842346,
-0.06273222714662552,
-0.14342696964740753,
0.010694170370697975,
0.05314265564084053,
0.05206485092639923,
0.01798195205628872,
0.10255878418684006,
-0.08534207940101624,
-0.017006168141961098,
0.06790418922901154,
-0.08018788695335388,
0.010787664912641048,
0.013622410595417023,
0.033593595027923584,
0.13079673051834106,
-0.02933715470135212,
0.0038096841890364885,
-0.07214540988206863,
0.20017871260643005,
-0.07691353559494019,
-0.007875196635723114,
-0.03281911462545395,
-0.06535334140062332,
0.033395830541849136,
-0.12865805625915527,
0.018965285271406174,
-0.1496049165725708,
-0.10939463973045349,
0.03139350190758705,
-0.006428668275475502,
-0.00933897402137518,
-0.028180697932839394,
-0.05325520411133766,
-0.01575983501970768,
0.07400085777044296,
-0.04766285791993141,
-0.0538020096719265,
-0.08664759248495102,
0.13297636806964874,
-0.04758240655064583,
0.07111898064613342,
-0.11756031215190887,
0.06406589597463608,
-0.10811319947242737,
-0.008591361343860626,
-0.12044783681631088,
0.06889443844556808,
-0.06390145421028137,
0.1489010453224182,
0.010835666209459305,
-0.06318619102239609,
-0.060163263231515884,
0.05314219370484352,
-0.047083787620067596,
0.21276888251304626,
-0.11909706145524979,
-0.13824494183063507,
0.1655023694038391,
-0.07781504839658737,
-0.07399924099445343,
0.08077021688222885,
-0.02526409551501274,
-0.007519088685512543,
0.11298951506614685,
0.1785651445388794,
0.05793144926428795,
0.00989704579114914,
0.018828032538294792,
0.10434573888778687,
-0.056992094963788986,
-0.04336705431342125,
0.008270285092294216,
0.014410432428121567,
-0.1488446742296219,
0.05384966731071472,
0.04967261478304863,
0.04888955503702164,
-0.05994677543640137,
-0.03152845427393913,
0.01233907975256443,
0.013807411305606365,
0.18871156871318817,
0.04816680774092674,
0.11093316227197647,
-0.08056838810443878,
-0.020094770938158035,
0.05810539424419403,
-0.009600282646715641,
0.024301037192344666,
0.029469182714819908,
-0.07228055596351624,
0.12280429899692535,
0.017246996983885765,
0.0506996288895607,
-0.2074073851108551,
-0.04925067350268364,
-0.023040728643536568,
0.11712715774774551,
-0.0139535591006279,
0.11271148920059204,
0.04458722844719887,
-0.05819385498762131,
-0.031116120517253876,
-0.029808409512043,
0.13929390907287598,
0.009752081707119942,
-0.06087294593453407,
-0.09651754051446915,
0.08299580216407776,
-0.07648039609193802,
-0.012982361018657684,
-0.07727156579494476,
0.016588877886533737,
0.11520323157310486,
0.11275610327720642,
0.00760972686111927,
0.055994316935539246,
-0.02732114866375923,
0.06018669903278351,
-0.07259949296712875,
0.03915261849761009,
0.10607141256332397,
0.002339787082746625,
-0.06463097780942917,
0.13554218411445618,
-0.11828406155109406,
0.25612956285476685,
0.19148939847946167,
-0.3114798069000244,
-0.006898148451000452,
-0.02609674260020256,
-0.012208682484924793,
0.020914344117045403,
0.013227556832134724,
-0.026625627651810646,
0.10297225415706635,
-0.021254152059555054,
0.19747138023376465,
-0.02441023290157318,
-0.047482430934906006,
-0.011208656243979931,
-0.02991623245179653,
-0.05719781294465065,
0.10729695111513138,
0.0407988615334034,
-0.15454907715320587,
0.19505955278873444,
0.2538165748119354,
0.02427719533443451,
0.20067399740219116,
0.0016230179462581873,
0.04251996800303459,
0.08954303711652756,
-0.07703202217817307,
-0.031195662915706635,
-0.06902409344911575,
-0.18087248504161835,
-0.01030104048550129,
0.07271541655063629,
0.003416385967284441,
0.04582829773426056,
-0.09102281928062439,
-0.04950135946273804,
-0.009406481869518757,
0.03977641090750694,
-0.022963033989071846,
0.06622903048992157,
0.061957668513059616,
0.11638306826353073,
0.016963282600045204,
-0.07863389700651169,
0.10928243398666382,
0.006941185798496008,
-0.07235120981931686,
0.18825221061706543,
-0.14513589441776276,
-0.3671988546848297,
-0.16329309344291687,
-0.2507820725440979,
-0.02200975827872753,
0.06032225489616394,
0.10401260852813721,
-0.1273992359638214,
-0.045884355902671814,
0.026094766333699226,
0.009253407828509808,
-0.020163610577583313,
0.04665093868970871,
0.00025113957235589623,
0.0507989376783371,
-0.03616386279463768,
-0.07654333859682083,
-0.04829872399568558,
-0.04084969684481621,
-0.012414759956300259,
0.1379728466272354,
-0.09804695099592209,
0.07078039646148682,
0.15354488790035248,
0.021654563024640083,
0.05926348268985748,
-0.028623560443520546,
0.1662140190601349,
-0.08923988044261932,
-0.023039935156702995,
0.1458786278963089,
-0.0993056371808052,
0.08482270687818527,
0.15149766206741333,
0.053157176822423935,
-0.05788780748844147,
0.03506138548254967,
-0.016630681231617928,
-0.07601160556077957,
-0.21829845011234283,
-0.14057202637195587,
-0.08711955696344376,
0.08515848964452744,
0.05234881490468979,
0.06730806827545166,
0.09567636251449585,
0.054419346153736115,
-0.013156152330338955,
0.004716838244348764,
-0.0009041961166076362,
0.07105755805969238,
0.25090041756629944,
-0.015848102048039436,
0.14711493253707886,
-0.03881558030843735,
-0.15611779689788818,
0.092148058116436,
-0.043058495968580246,
0.04445879906415939,
0.08087486028671265,
0.0026154271326959133,
0.003801462473347783,
0.022773509845137596,
0.15765000879764557,
0.13268984854221344,
0.01638510823249817,
-0.03183729946613312,
-0.023541349917650223,
-0.01883045583963394,
-0.04446067288517952,
0.004096326418220997,
0.03469981998205185,
-0.10628160834312439,
-0.04881539195775986,
-0.14708632230758667,
0.07742974907159805,
0.13871778547763824,
0.075959712266922,
-0.19465002417564392,
0.0040236180648207664,
0.08866526931524277,
-0.025318454951047897,
-0.0990496501326561,
0.06158597022294998,
-0.04107663780450821,
-0.11049347370862961,
0.12084910273551941,
-0.00730234244838357,
0.12163908779621124,
-0.08447352796792984,
0.09676861017942429,
-0.04243823513388634,
-0.14561885595321655,
0.000663170765619725,
0.10613974183797836,
-0.27669164538383484,
0.22850312292575836,
0.022773034870624542,
-0.06339491158723831,
-0.05394286289811134,
-0.037515703588724136,
0.032660968601703644,
0.19981035590171814,
0.08667976409196854,
0.0052032871171832085,
0.046411532908678055,
-0.18815118074417114,
-0.062216419726610184,
-0.011883254162967205,
0.11106228083372116,
-0.06853599101305008,
0.0012572898995131254,
-0.02820238657295704,
-0.033743616193532944,
-0.028431542217731476,
-0.03553496673703194,
0.034797947853803635,
-0.14481128752231598,
0.06837789714336395,
0.010432238690555096,
0.09733088314533234,
0.004214155022054911,
-0.10303426533937454,
-0.14589819312095642,
0.18283648788928986,
-0.13263006508350372,
-0.07620836049318314,
-0.11532218754291534,
-0.050632257014513016,
-0.01720179058611393,
-0.08221561461687088,
0.040548451244831085,
-0.05726635083556175,
0.053416717797517776,
-0.061353426426649094,
-0.19123683869838715,
0.12433511763811111,
-0.09110256284475327,
-0.07613298296928406,
-0.05882355570793152,
0.16095571219921112,
-0.030357684940099716,
0.03127630054950714,
0.021169543266296387,
0.027162207290530205,
-0.05900159105658531,
-0.07342267781496048,
0.009431506507098675,
0.0164501816034317,
0.07426818460226059,
0.0638432651758194,
-0.11173663288354874,
-0.10908952355384827,
-0.047947295010089874,
0.019227292388677597,
0.28690654039382935,
0.2232511341571808,
-0.05962953343987465,
0.1341296285390854,
0.16891220211982727,
-0.044242050498723984,
-0.38496464490890503,
-0.08762367814779282,
-0.07763371616601944,
-0.05927511304616928,
-0.03590373322367668,
-0.17651906609535217,
0.13525429368019104,
0.02272622101008892,
-0.017686165869235992,
0.0724470391869545,
-0.15540668368339539,
-0.0843331441283226,
0.1828574538230896,
-0.0640326738357544,
0.4482351541519165,
-0.09789205342531204,
-0.09974410384893417,
-0.050553351640701294,
-0.11223675310611725,
0.15565823018550873,
0.018917085602879524,
0.08339349925518036,
-0.022604014724493027,
0.03534995764493942,
0.03614838421344757,
-0.02814580872654915,
0.09522397816181183,
0.01079475786536932,
0.0073252092115581036,
-0.12509995698928833,
-0.12479730695486069,
0.0320330373942852,
-0.008123885840177536,
-0.032482873648405075,
-0.018250316381454468,
0.009646511636674404,
-0.18282178044319153,
-0.03914261236786842,
-0.08117711544036865,
0.0654233992099762,
0.0239864569157362,
-0.0380018912255764,
0.006391940638422966,
-0.023329105228185654,
-0.023697715252637863,
-0.004144454374909401,
0.2987746298313141,
-0.051615674048662186,
0.15180352330207825,
0.06878737360239029,
0.1509307473897934,
-0.1468835026025772,
0.026437919586896896,
-0.07920823991298676,
-0.07032180577516556,
0.06918472051620483,
-0.06762229651212692,
0.054928023368120193,
0.1341441571712494,
-0.02666395716369152,
0.0681128203868866,
0.12321872264146805,
0.04401358217000961,
-0.007726769428700209,
0.14712786674499512,
-0.28360673785209656,
0.03901231661438942,
-0.07269950956106186,
0.014411142095923424,
0.03697760030627251,
0.0677158534526825,
0.13740111887454987,
0.05507473275065422,
-0.0431063286960125,
0.04431106522679329,
0.0013171000173315406,
0.00573392491787672,
0.07776136696338654,
0.04710765555500984,
0.04071477800607681,
-0.15293361246585846,
0.044133059680461884,
0.04930664598941803,
-0.1488267183303833,
-0.008389090187847614,
0.1502973437309265,
-0.14820247888565063,
-0.1199675127863884,
-0.02808389998972416,
0.1513991504907608,
-0.10124483704566956,
-0.028135769069194794,
-0.07000049948692322,
-0.11982656270265579,
0.09845300763845444,
0.2320709228515625,
0.11974326521158218,
0.0876096859574318,
-0.06393727660179138,
-0.04921369254589081,
0.03203205019235611,
0.027103152126073837,
0.01570478081703186,
0.0341915600001812,
-0.08069929480552673,
0.021296482533216476,
-0.025211798027157784,
0.18249393999576569,
-0.10756555199623108,
-0.0748402550816536,
-0.15665973722934723,
0.03539939597249031,
-0.09332622587680817,
-0.03192579001188278,
-0.06576883047819138,
-0.041140440851449966,
-0.03153247386217117,
-0.06876379996538162,
-0.029994487762451172,
-0.07327309995889664,
-0.10637706518173218,
0.04549536854028702,
-0.005121663678437471,
0.044533200562000275,
-0.0874074399471283,
-0.03796874359250069,
0.10042928159236908,
-0.034863054752349854,
0.10939274728298187,
0.09408368170261383,
-0.07100128382444382,
0.10719622671604156,
-0.11312992870807648,
-0.12122457474470139,
0.1295689195394516,
0.004958937875926495,
0.05844056233763695,
0.11528068780899048,
0.014610887505114079,
0.04309787601232529,
-0.0016254446236416698,
0.06724229454994202,
0.06838985532522202,
-0.11488457024097443,
0.0961456298828125,
-0.017461946234107018,
-0.17954106628894806,
-0.03587982431054115,
-0.03723905608057976,
0.1073005273938179,
0.013648814521729946,
0.19167259335517883,
-0.0633932501077652,
0.08365726470947266,
-0.00409066304564476,
0.015174121595919132,
-0.009612918831408024,
-0.2314261943101883,
-0.07159615308046341,
-0.07181085646152496,
0.02351308800280094,
-0.00944665726274252,
0.1983625888824463,
0.033960696309804916,
-0.023598959669470787,
0.05965990573167801,
0.06046878919005394,
-0.012681914493441582,
0.0182325579226017,
0.17801575362682343,
0.10362739861011505,
-0.06077010929584503,
-0.11079159379005432,
0.05484027415513992,
0.036151014268398285,
0.03816447779536247,
0.1408037394285202,
0.07491238415241241,
-0.01683124154806137,
0.07192002981901169,
-0.017678270116448402,
0.042623672634363174,
-0.12995591759681702,
-0.16698575019836426,
-0.059120938181877136,
0.07092958688735962,
0.025988928973674774,
0.12077968567609787,
0.06512223184108734,
-0.01754876598715782,
0.03929463401436806,
-0.02893846295773983,
-0.03495998680591583,
-0.15733319520950317,
-0.06499067693948746,
-0.09070319682359695,
-0.0802110880613327,
0.006369082257151604,
-0.07378353178501129,
0.010617483407258987,
0.09302955120801926,
0.051978640258312225,
-0.05011377111077309,
0.10351382941007614,
0.016250405460596085,
-0.05566086620092392,
0.12356534600257874,
-0.03173723816871643,
0.011975249275565147,
-0.051020968705415726,
-0.03440050408244133,
-0.11123842746019363,
0.03321651369333267,
-0.03869492560625076,
0.04443668946623802,
-0.04336242005228996,
-0.017597999423742294,
-0.14802704751491547,
-0.13901560008525848,
-0.015662794932723045,
0.05674887076020241,
-0.03439854085445404,
0.09601373970508575,
0.00397905521094799,
0.025712324306368828,
0.04986096918582916,
0.17324167490005493,
-0.043381258845329285,
-0.02522560954093933,
-0.017492396757006645,
0.2044573873281479,
0.07090096920728683,
0.14164690673351288,
-0.01305923331528902,
-0.01934840902686119,
-0.07520835846662521,
0.2982312738895416,
0.2553103566169739,
-0.06066034361720085,
0.059236738830804825,
0.004571202676743269,
0.054721564054489136,
0.15492239594459534,
0.11465923488140106,
0.08992383629083633,
0.22361759841442108,
-0.07948474586009979,
-0.05900951102375984,
-0.033075276762247086,
-0.013586034066975117,
-0.1123759001493454,
0.05017464607954025,
0.057567525655031204,
-0.05244635418057442,
-0.0862443819642067,
0.08665744215250015,
-0.19463303685188293,
0.1255744844675064,
-0.04300113022327423,
-0.20939047634601593,
-0.07638900727033615,
-0.02783568575978279,
0.11122137308120728,
-0.0008472305489704013,
0.0714731365442276,
0.018994968384504318,
-0.08941662311553955,
0.02221987396478653,
0.010096932761371136,
-0.16530966758728027,
-0.039684489369392395,
0.06653241813182831,
-0.03831864520907402,
0.026561198756098747,
-0.02589813992381096,
-0.0017830745782703161,
0.06523782759904861,
0.052454929798841476,
-0.00008317003812408075,
0.02209712564945221,
0.01639433018863201,
-0.047038473188877106,
-0.002656529890373349,
0.02008855529129505,
0.008079451508820057,
-0.08916997164487839,
0.0806533619761467,
-0.15614132583141327,
0.05247768014669418,
-0.06394287943840027,
-0.10212945193052292,
-0.012497876770794392,
0.0694773867726326,
-0.04933639243245125,
0.04696596786379814,
0.12283168733119965,
-0.0027036862447857857,
-0.05056165158748627,
-0.05035220459103584,
-0.027720589190721512,
-0.007992521859705448,
-0.1216273382306099,
-0.11228810995817184,
-0.13017980754375458,
-0.06924757361412048,
0.04752228036522865,
0.024720678105950356,
-0.19368933141231537,
-0.002218215959146619,
-0.10201462358236313,
0.07052378356456757,
-0.1573362499475479,
0.06790082901716232,
0.03289732709527016,
0.005852126516401768,
-0.01401523593813181,
-0.05787905305624008,
0.05628145858645439,
0.06443369388580322,
-0.1214250698685646,
-0.08060063421726227
] |
null | null |
transformers
|
# Polbert - Polish BERT
Polish version of BERT language model is here! It is now available in two variants: cased and uncased, both can be downloaded and used via HuggingFace transformers library. I recommend using the cased model, more info on the differences and benchmark results below.

## Cased and uncased variants
* I initially trained the uncased model, the corpus and training details are referenced below. Here are some issues I found after I published the uncased model:
* Some Polish characters and accents are not tokenized correctly through the BERT tokenizer when applying lowercase. This doesn't impact sequence classification much, but may influence token classfication tasks significantly.
* I noticed a lot of duplicates in the Open Subtitles dataset, which dominates the training corpus.
* I didn't use Whole Word Masking.
* The cased model improves on the uncased model in the following ways:
* All Polish characters and accents should now be tokenized correctly.
* I removed duplicates from Open Subtitles dataset. The corpus is smaller, but more balanced now.
* The model is trained with Whole Word Masking.
## Pre-training corpora
Below is the list of corpora used along with the output of `wc` command (counting lines, words and characters). These corpora were divided into sentences with srxsegmenter (see references), concatenated and tokenized with HuggingFace BERT Tokenizer.
### Uncased
| Tables | Lines | Words | Characters |
| ------------- |--------------:| -----:| -----:|
| [Polish subset of Open Subtitles](http://opus.nlpl.eu/OpenSubtitles-v2018.php) | 236635408| 1431199601 | 7628097730 |
| [Polish subset of ParaCrawl](http://opus.nlpl.eu/ParaCrawl.php) | 8470950 | 176670885 | 1163505275 |
| [Polish Parliamentary Corpus](http://clip.ipipan.waw.pl/PPC) | 9799859 | 121154785 | 938896963 |
| [Polish Wikipedia - Feb 2020](https://dumps.wikimedia.org/plwiki/latest/plwiki-latest-pages-articles.xml.bz2) | 8014206 | 132067986 | 1015849191 |
| Total | 262920423 | 1861093257 | 10746349159 |
### Cased
| Tables | Lines | Words | Characters |
| ------------- |--------------:| -----:| -----:|
| [Polish subset of Open Subtitles (Deduplicated) ](http://opus.nlpl.eu/OpenSubtitles-v2018.php) | 41998942| 213590656 | 1424873235 |
| [Polish subset of ParaCrawl](http://opus.nlpl.eu/ParaCrawl.php) | 8470950 | 176670885 | 1163505275 |
| [Polish Parliamentary Corpus](http://clip.ipipan.waw.pl/PPC) | 9799859 | 121154785 | 938896963 |
| [Polish Wikipedia - Feb 2020](https://dumps.wikimedia.org/plwiki/latest/plwiki-latest-pages-articles.xml.bz2) | 8014206 | 132067986 | 1015849191 |
| Total | 68283960 | 646479197 | 4543124667 |
## Pre-training details
### Uncased
* Polbert was trained with code provided in Google BERT's github repository (https://github.com/google-research/bert)
* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)
* Training set-up: in total 1 million training steps:
* 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)
* 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5
* 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
* The model was trained on a single Google Cloud TPU v3-8
### Cased
* Same approach as uncased model, with the following differences:
* Whole Word Masking
* Training set-up:
* 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)
* 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5
* 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
## Usage
Polbert is released via [HuggingFace Transformers library](https://huggingface.co/transformers/).
For an example use as language model, see [this notebook](/LM_testing.ipynb) file.
### Uncased
```python
from transformers import *
model = BertForMaskedLM.from_pretrained("dkleczek/bert-base-polish-uncased-v1")
tokenizer = BertTokenizer.from_pretrained("dkleczek/bert-base-polish-uncased-v1")
nlp = pipeline('fill-mask', model=model, tokenizer=tokenizer)
for pred in nlp(f"Adam Mickiewicz wielkim polskim {nlp.tokenizer.mask_token} był."):
print(pred)
# Output:
# {'sequence': '[CLS] adam mickiewicz wielkim polskim poeta był. [SEP]', 'score': 0.47196975350379944, 'token': 26596}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim bohaterem był. [SEP]', 'score': 0.09127858281135559, 'token': 10953}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim człowiekiem był. [SEP]', 'score': 0.0647173821926117, 'token': 5182}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim pisarzem był. [SEP]', 'score': 0.05232388526201248, 'token': 24293}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim politykiem był. [SEP]', 'score': 0.04554257541894913, 'token': 44095}
```
### Cased
```python
model = BertForMaskedLM.from_pretrained("dkleczek/bert-base-polish-cased-v1")
tokenizer = BertTokenizer.from_pretrained("dkleczek/bert-base-polish-cased-v1")
nlp = pipeline('fill-mask', model=model, tokenizer=tokenizer)
for pred in nlp(f"Adam Mickiewicz wielkim polskim {nlp.tokenizer.mask_token} był."):
print(pred)
# Output:
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim pisarzem był. [SEP]', 'score': 0.5391148328781128, 'token': 37120}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim człowiekiem był. [SEP]', 'score': 0.11683262139558792, 'token': 6810}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim bohaterem był. [SEP]', 'score': 0.06021466106176376, 'token': 17709}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim mistrzem był. [SEP]', 'score': 0.051870670169591904, 'token': 14652}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim artystą był. [SEP]', 'score': 0.031787533313035965, 'token': 35680}
```
See the next section for an example usage of Polbert in downstream tasks.
## Evaluation
Thanks to Allegro, we now have the [KLEJ benchmark](https://klejbenchmark.com/leaderboard/), a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.
| Model | Average | NKJP-NER | CDSC-E | CDSC-R | CBD | PolEmo2.0-IN | PolEmo2.0-OUT | DYK | PSC | AR |
| ------------- |--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|
| Polbert cased | 81.7 | 93.6 | 93.4 | 93.8 | 52.7 | 87.4 | 71.1 | 59.1 | 98.6 | 85.2 |
| Polbert uncased | 81.4 | 90.1 | 93.9 | 93.5 | 55.0 | 88.1 | 68.8 | 59.4 | 98.8 | 85.4 |
Note how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.
## Bias
The data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.
## Acknowledgements
* I'd like to express my gratitude to Google [TensorFlow Research Cloud (TFRC)](https://www.tensorflow.org/tfrc) for providing the free TPU credits - thank you!
* Also appreciate the help from Timo Möller from [deepset](https://deepset.ai) for sharing tips and scripts based on their experience training German BERT model.
* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.
* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from [fastai](https://www.fast.ai) for their NLP and Deep Learning courses!
## Author
Darek Kłeczek - contact me on Twitter [@dk21](https://twitter.com/dk21)
## References
* https://github.com/google-research/bert
* https://github.com/narusemotoki/srx_segmenter
* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: https://raw.githubusercontent.com/languagetool-org/languagetool/master/languagetool-core/src/main/resources/org/languagetool/resource/segment.srx
* [KLEJ benchmark](https://klejbenchmark.com/leaderboard/)
|
{"language": "pl", "thumbnail": "https://raw.githubusercontent.com/kldarek/polbert/master/img/polbert.png"}
| null |
dkleczek/bert-base-polish-cased-v1
|
[
"transformers",
"pytorch",
"jax",
"bert",
"pretraining",
"pl",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"pl"
] |
TAGS
#transformers #pytorch #jax #bert #pretraining #pl #endpoints_compatible #has_space #region-us
|
Polbert - Polish BERT
=====================
Polish version of BERT language model is here! It is now available in two variants: cased and uncased, both can be downloaded and used via HuggingFace transformers library. I recommend using the cased model, more info on the differences and benchmark results below.
!PolBERT image
Cased and uncased variants
--------------------------
* I initially trained the uncased model, the corpus and training details are referenced below. Here are some issues I found after I published the uncased model:
+ Some Polish characters and accents are not tokenized correctly through the BERT tokenizer when applying lowercase. This doesn't impact sequence classification much, but may influence token classfication tasks significantly.
+ I noticed a lot of duplicates in the Open Subtitles dataset, which dominates the training corpus.
+ I didn't use Whole Word Masking.
* The cased model improves on the uncased model in the following ways:
+ All Polish characters and accents should now be tokenized correctly.
+ I removed duplicates from Open Subtitles dataset. The corpus is smaller, but more balanced now.
+ The model is trained with Whole Word Masking.
Pre-training corpora
--------------------
Below is the list of corpora used along with the output of 'wc' command (counting lines, words and characters). These corpora were divided into sentences with srxsegmenter (see references), concatenated and tokenized with HuggingFace BERT Tokenizer.
### Uncased
### Cased
Pre-training details
--------------------
### Uncased
* Polbert was trained with code provided in Google BERT's github repository (URL
* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)
* Training set-up: in total 1 million training steps:
+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)
+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5
+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
* The model was trained on a single Google Cloud TPU v3-8
### Cased
* Same approach as uncased model, with the following differences:
+ Whole Word Masking
* Training set-up:
+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)
+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5
+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
Usage
-----
Polbert is released via HuggingFace Transformers library.
For an example use as language model, see this notebook file.
### Uncased
### Cased
See the next section for an example usage of Polbert in downstream tasks.
Evaluation
----------
Thanks to Allegro, we now have the KLEJ benchmark, a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.
Note how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.
Bias
----
The data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.
Acknowledgements
----------------
* I'd like to express my gratitude to Google TensorFlow Research Cloud (TFRC) for providing the free TPU credits - thank you!
* Also appreciate the help from Timo Möller from deepset for sharing tips and scripts based on their experience training German BERT model.
* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.
* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from fastai for their NLP and Deep Learning courses!
Author
------
Darek Kłeczek - contact me on Twitter @dk21
References
----------
* URL
* URL
* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: URL
* KLEJ benchmark
|
[
"### Uncased",
"### Cased\n\n\n\nPre-training details\n--------------------",
"### Uncased\n\n\n* Polbert was trained with code provided in Google BERT's github repository (URL\n* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)\n* Training set-up: in total 1 million training steps:\n\t+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)\n\t+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n* The model was trained on a single Google Cloud TPU v3-8",
"### Cased\n\n\n* Same approach as uncased model, with the following differences:\n\t+ Whole Word Masking\n* Training set-up:\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n\n\nUsage\n-----\n\n\nPolbert is released via HuggingFace Transformers library.\n\n\nFor an example use as language model, see this notebook file.",
"### Uncased",
"### Cased\n\n\nSee the next section for an example usage of Polbert in downstream tasks.\n\n\nEvaluation\n----------\n\n\nThanks to Allegro, we now have the KLEJ benchmark, a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.\n\n\n\nNote how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.\n\n\nBias\n----\n\n\nThe data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.\n\n\nAcknowledgements\n----------------\n\n\n* I'd like to express my gratitude to Google TensorFlow Research Cloud (TFRC) for providing the free TPU credits - thank you!\n* Also appreciate the help from Timo Möller from deepset for sharing tips and scripts based on their experience training German BERT model.\n* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.\n* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from fastai for their NLP and Deep Learning courses!\n\n\nAuthor\n------\n\n\nDarek Kłeczek - contact me on Twitter @dk21\n\n\nReferences\n----------\n\n\n* URL\n* URL\n* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: URL\n* KLEJ benchmark"
] |
[
"TAGS\n#transformers #pytorch #jax #bert #pretraining #pl #endpoints_compatible #has_space #region-us \n",
"### Uncased",
"### Cased\n\n\n\nPre-training details\n--------------------",
"### Uncased\n\n\n* Polbert was trained with code provided in Google BERT's github repository (URL\n* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)\n* Training set-up: in total 1 million training steps:\n\t+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)\n\t+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n* The model was trained on a single Google Cloud TPU v3-8",
"### Cased\n\n\n* Same approach as uncased model, with the following differences:\n\t+ Whole Word Masking\n* Training set-up:\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n\n\nUsage\n-----\n\n\nPolbert is released via HuggingFace Transformers library.\n\n\nFor an example use as language model, see this notebook file.",
"### Uncased",
"### Cased\n\n\nSee the next section for an example usage of Polbert in downstream tasks.\n\n\nEvaluation\n----------\n\n\nThanks to Allegro, we now have the KLEJ benchmark, a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.\n\n\n\nNote how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.\n\n\nBias\n----\n\n\nThe data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.\n\n\nAcknowledgements\n----------------\n\n\n* I'd like to express my gratitude to Google TensorFlow Research Cloud (TFRC) for providing the free TPU credits - thank you!\n* Also appreciate the help from Timo Möller from deepset for sharing tips and scripts based on their experience training German BERT model.\n* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.\n* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from fastai for their NLP and Deep Learning courses!\n\n\nAuthor\n------\n\n\nDarek Kłeczek - contact me on Twitter @dk21\n\n\nReferences\n----------\n\n\n* URL\n* URL\n* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: URL\n* KLEJ benchmark"
] |
[
35,
5,
10,
160,
131,
5,
397
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #pretraining #pl #endpoints_compatible #has_space #region-us \n### Uncased### Cased\n\n\n\nPre-training details\n--------------------### Uncased\n\n\n* Polbert was trained with code provided in Google BERT's github repository (URL\n* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)\n* Training set-up: in total 1 million training steps:\n\t+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)\n\t+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n* The model was trained on a single Google Cloud TPU v3-8### Cased\n\n\n* Same approach as uncased model, with the following differences:\n\t+ Whole Word Masking\n* Training set-up:\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n\n\nUsage\n-----\n\n\nPolbert is released via HuggingFace Transformers library.\n\n\nFor an example use as language model, see this notebook file.### Uncased"
] |
[
-0.08304311335086823,
0.1835978776216507,
-0.0031671314500272274,
0.08835083991289139,
0.10805084556341171,
0.044563811272382736,
0.12723813951015472,
0.08533528447151184,
-0.012283156625926495,
0.12237169593572617,
0.04814939945936203,
-0.039152104407548904,
0.12244028598070145,
0.13657218217849731,
0.05187345668673515,
-0.3863251507282257,
0.04178787022829056,
-0.08658020943403244,
-0.027268724516034126,
0.05848642438650131,
0.06642556935548782,
-0.06614234298467636,
0.032977692782878876,
-0.016534404829144478,
0.06292051821947098,
-0.06310322880744934,
-0.0860297679901123,
-0.056950751692056656,
0.08806941658258438,
0.009402361698448658,
0.077077217400074,
0.027592318132519722,
0.10254770517349243,
-0.1736193299293518,
0.0023470923770219088,
0.10295017808675766,
0.010362976230680943,
0.08479958772659302,
0.08693331480026245,
0.03809747472405434,
0.052281368523836136,
-0.08386348187923431,
0.06369008868932724,
0.027413902804255486,
-0.0811687558889389,
-0.11983965337276459,
-0.09041769802570343,
0.07767286896705627,
0.06348972767591476,
0.03185784071683884,
-0.04132547602057457,
-0.014394170604646206,
-0.06979679316282272,
0.04229762405157089,
0.17595921456813812,
-0.2878977954387665,
-0.07002752274274826,
0.0476127453148365,
0.07564400881528854,
0.0446828156709671,
-0.1094399243593216,
-0.04096240922808647,
0.02537406235933304,
0.03183979541063309,
0.03155304118990898,
0.052049439400434494,
0.028938667848706245,
-0.0028212869074195623,
-0.12175190448760986,
-0.007842443883419037,
0.07813435792922974,
0.060777097940444946,
-0.06165160611271858,
-0.16944655776023865,
-0.05418658256530762,
-0.17313581705093384,
-0.04894265532493591,
0.022340532392263412,
0.07953190803527832,
0.017491089180111885,
-0.07275177538394928,
-0.03802208974957466,
-0.04984228312969208,
-0.02353806234896183,
-0.011407671496272087,
0.07912322878837585,
0.07044769078493118,
0.007420604582875967,
0.024932557716965675,
0.07901211082935333,
0.025093965232372284,
-0.10299903899431229,
-0.04609553515911102,
0.0028638888616114855,
-0.10581977665424347,
-0.034290071576833725,
-0.03443751484155655,
-0.003794976742938161,
0.023844994604587555,
0.14503000676631927,
-0.041734788566827774,
0.06954594701528549,
0.0336422361433506,
-0.025746414437890053,
0.03249809518456459,
0.22325892746448517,
-0.09175436943769455,
0.005093968473374844,
-0.005062831565737724,
0.08725540339946747,
-0.008662659674882889,
-0.03015020862221718,
-0.050642795860767365,
-0.015889640897512436,
0.09759997576475143,
0.08139770478010178,
-0.011357364244759083,
0.09941494464874268,
-0.014506561681628227,
-0.02943277172744274,
0.06162586063146591,
-0.15576383471488953,
0.031006669625639915,
0.04005703330039978,
-0.03599917143583298,
0.013716619461774826,
-0.03355256840586662,
-0.07379066944122314,
-0.13952651619911194,
0.04696902632713318,
-0.09427066147327423,
-0.06557491421699524,
-0.05626862123608589,
-0.06520477682352066,
0.03753121569752693,
-0.048041537404060364,
-0.09322167187929153,
-0.04719527065753937,
-0.138665109872818,
-0.06691102683544159,
0.048842545598745346,
-0.07119753956794739,
-0.06739035248756409,
-0.07178961485624313,
-0.038980498909950256,
0.030654901638627052,
-0.0012933824909850955,
0.07939096540212631,
-0.03178488835692406,
0.05632065236568451,
-0.06897033005952835,
0.06324680894613266,
0.14670677483081818,
0.02366153709590435,
-0.05374528095126152,
0.070162832736969,
-0.1619381159543991,
0.1598101556301117,
-0.11853887140750885,
-0.02473026141524315,
-0.10013549774885178,
-0.0852285698056221,
-0.04840080812573433,
-0.018228044733405113,
0.056222859770059586,
0.13866539299488068,
-0.14708203077316284,
-0.04570735991001129,
0.20575428009033203,
-0.06696098297834396,
-0.02824617363512516,
0.09098588675260544,
-0.008651424199342728,
-0.0672333613038063,
0.06257417798042297,
0.07084046304225922,
0.06788607686758041,
-0.14082472026348114,
-0.09523157775402069,
0.05468454584479332,
0.0525425523519516,
0.1428404152393341,
0.07254085689783096,
-0.07041897624731064,
0.004397648386657238,
0.03240372613072395,
-0.029694590717554092,
-0.004231433384120464,
-0.043462347239255905,
-0.04423913359642029,
-0.021368358284235,
-0.059059496968984604,
0.007903367280960083,
-0.024091271683573723,
-0.0047958833165466785,
-0.02148543670773506,
-0.1325642615556717,
-0.01715020276606083,
0.13382427394390106,
-0.12578770518302917,
0.0186562929302454,
-0.0656023845076561,
0.026079917326569557,
0.024165784940123558,
0.025300314649939537,
-0.11799274384975433,
-0.15649473667144775,
0.13275288045406342,
-0.17109571397304535,
-0.014421402476727962,
-0.04389704018831253,
0.03661253675818443,
0.054637566208839417,
-0.07713136821985245,
-0.03580109030008316,
-0.045931585133075714,
-0.050569213926792145,
-0.07658495008945465,
-0.056491412222385406,
-0.09397745877504349,
-0.00767331849783659,
0.05623743683099747,
-0.04920964315533638,
0.028905564919114113,
0.009831567294895649,
0.11408975720405579,
-0.0023966734297573566,
-0.06821008026599884,
0.02308812364935875,
-0.03099972940981388,
0.011158322915434837,
-0.14636041224002838,
0.009301342070102692,
0.029714049771428108,
-0.03279665485024452,
0.0019924670923501253,
-0.08657185733318329,
-0.08310844004154205,
0.07166364043951035,
0.16287444531917572,
-0.08874481171369553,
0.011215521022677422,
-0.07178657501935959,
-0.004000541288405657,
-0.08402474969625473,
0.032371409237384796,
0.16615797579288483,
0.06598210334777832,
0.13092675805091858,
-0.12915383279323578,
-0.05112267658114433,
0.012840314768254757,
0.058449387550354004,
-0.05219940096139908,
0.09830858558416367,
-0.0009562695049680769,
-0.030847998335957527,
0.0378548689186573,
-0.03965960070490837,
0.028394360095262527,
0.14106836915016174,
-0.07827423512935638,
-0.10932717472314835,
-0.06068311631679535,
0.060644056648015976,
0.013983049429953098,
0.044149041175842285,
-0.10949568450450897,
0.00962670985609293,
0.021605094894766808,
0.01393930334597826,
0.038135357201099396,
-0.08393487334251404,
0.028833672404289246,
-0.011327514424920082,
-0.08597509562969208,
-0.028192991390824318,
-0.06014483794569969,
-0.0002915394143201411,
0.07926903665065765,
0.07627148926258087,
0.02558453567326069,
-0.011050941422581673,
-0.021010858938097954,
-0.07139278948307037,
0.18946637213230133,
-0.07485820353031158,
-0.2091328203678131,
-0.11141598224639893,
0.0779978334903717,
-0.04182685539126396,
0.0322304330766201,
-0.044605426490306854,
-0.09092441201210022,
-0.041214652359485626,
-0.07387463748455048,
0.11705072224140167,
0.0011067852610722184,
0.009072757326066494,
0.020612597465515137,
0.02458597533404827,
0.0384574830532074,
-0.12361173331737518,
0.014756369404494762,
0.0126201082020998,
-0.11823289096355438,
-0.04657624661922455,
0.026276368647813797,
0.01669933833181858,
0.1260898858308792,
0.007213996257632971,
0.006207019556313753,
-0.004933126270771027,
0.24342717230319977,
-0.0888102725148201,
0.09934742003679276,
0.08478143066167831,
0.06063110753893852,
0.07168921828269958,
0.05233727768063545,
0.04150984063744545,
-0.06024875491857529,
0.03198908269405365,
0.06615400314331055,
0.02324756234884262,
-0.26663318276405334,
-0.010094606317579746,
-0.02176373451948166,
0.07319309562444687,
0.16186389327049255,
0.06299353390932083,
-0.11452800780534744,
0.06502750515937805,
-0.11057177931070328,
0.06850500404834747,
0.050946176052093506,
0.04804033041000366,
-0.009183446876704693,
-0.016780463978648186,
0.03961285203695297,
-0.04497622698545456,
-0.040096402168273926,
0.07886277139186859,
0.013940062373876572,
0.15753880143165588,
-0.06681087613105774,
0.20863789319992065,
-0.010931525379419327,
0.14094097912311554,
0.01762223057448864,
0.08219286799430847,
0.014106348156929016,
0.021993201225996017,
0.024531232193112373,
-0.05841425061225891,
-0.026629872620105743,
0.03094610385596752,
0.061573710292577744,
-0.028584716841578484,
-0.08658988773822784,
-0.017479995265603065,
0.013296520337462425,
0.38786086440086365,
0.08532193303108215,
-0.1548105776309967,
-0.10482708364725113,
0.004797864705324173,
-0.10806690901517868,
-0.12392547726631165,
0.0673660859465599,
0.08862636238336563,
-0.1112275943160057,
0.0868154838681221,
-0.07143247127532959,
0.05743303522467613,
-0.1431775689125061,
-0.008587605319917202,
0.058122776448726654,
0.14607037603855133,
-0.03322220221161842,
0.10744767636060715,
-0.13632604479789734,
0.08819852024316788,
-0.003426853334531188,
0.07748749107122421,
-0.0635639950633049,
0.0630883201956749,
0.017278721556067467,
-0.02110062539577484,
0.11121943593025208,
0.01807139441370964,
-0.1007833331823349,
-0.07527950406074524,
-0.07907798141241074,
0.03512483835220337,
0.12378828227519989,
-0.118279829621315,
0.07586537301540375,
-0.05333395674824715,
-0.014821449294686317,
-0.015580669045448303,
-0.05560445040464401,
-0.09471837431192398,
-0.15824230015277863,
0.010934695601463318,
-0.09830352663993835,
0.06864796578884125,
-0.11594238877296448,
-0.060918379575014114,
-0.10164285451173782,
0.17261770367622375,
-0.044278740882873535,
-0.09213872253894806,
-0.12434632331132889,
0.04745829105377197,
0.17174336314201355,
-0.08903111517429352,
0.044065553694963455,
-0.03730827197432518,
0.17765232920646667,
-0.012290224432945251,
-0.07907674461603165,
0.031349826604127884,
-0.027351856231689453,
-0.21303163468837738,
-0.038536086678504944,
0.11888473480939865,
0.03723687306046486,
0.03311311826109886,
-0.04150857403874397,
0.0015508098294958472,
-0.024597398936748505,
-0.04764799401164055,
0.022480100393295288,
0.04897867143154144,
0.030332472175359726,
0.09153091907501221,
-0.1001262366771698,
0.03945107385516167,
-0.02991998754441738,
0.009465083479881287,
0.02223316766321659,
0.18959569931030273,
-0.06596892327070236,
0.060976605862379074,
0.20600776374340057,
-0.034653257578611374,
-0.20701360702514648,
-0.07545994222164154,
0.10017842799425125,
0.00799533911049366,
-0.08949105441570282,
-0.292757123708725,
0.042625825852155685,
0.11601298302412033,
-0.014623506926000118,
0.021543897688388824,
-0.24568834900856018,
-0.10709734261035919,
0.07199438661336899,
-0.009546013548970222,
-0.04695304110646248,
-0.06940706074237823,
-0.030488980934023857,
-0.06760977953672409,
0.013813462108373642,
0.11214770376682281,
0.08477199077606201,
0.1498708873987198,
0.013853583484888077,
-0.03844868019223213,
0.031275175511837006,
-0.042597074061632156,
0.12210506200790405,
-0.022026250138878822,
0.06965534389019012,
-0.035391781479120255,
-0.028351807966828346,
0.03523256629705429,
-0.07179992645978928,
0.09461961686611176,
-0.002823471324518323,
0.12688744068145752,
-0.012320922687649727,
-0.07647615671157837,
-0.05791850388050079,
0.044725656509399414,
-0.052024368196725845,
-0.03318844363093376,
-0.11039336025714874,
0.0713447704911232,
0.0802263543009758,
-0.0016042374772951007,
0.03572944551706314,
0.044937461614608765,
0.02555432915687561,
0.07418055832386017,
0.037943366914987564,
0.03239443525671959,
-0.10090950131416321,
-0.05541247874498367,
-0.03464877977967262,
0.028094837442040443,
-0.09684516489505768,
0.023589836433529854,
0.07270460575819016,
0.02938358299434185,
0.13711439073085785,
0.027575844898819923,
-0.1324227899312973,
-0.04293001443147659,
0.002760642906650901,
-0.06488937139511108,
-0.21921023726463318,
-0.06338252872228622,
-0.12252040207386017,
-0.19747300446033478,
0.005247308872640133,
0.064857117831707,
-0.005822112783789635,
-0.03530387207865715,
-0.020843494683504105,
0.037821073085069656,
0.03826174512505531,
0.13846883177757263,
0.02801435999572277,
0.06560137867927551,
-0.0981832966208458,
0.18011638522148132,
0.12910009920597076,
-0.08547251671552658,
0.049282196909189224,
0.11426263302564621,
-0.0697995200753212,
-0.016766222193837166,
0.027499038726091385,
0.0970996543765068,
0.07281064987182617,
0.023700391873717308,
-0.021116720512509346,
-0.08806706219911575,
0.13754989206790924,
0.015117048285901546,
-0.004689947236329317,
0.09991379827260971,
0.007109011057764292,
0.06349261105060577,
-0.05240553990006447,
0.09537157416343689,
0.0061438847333192825,
0.03860430046916008,
-0.08427275717258453,
0.12401040643453598,
0.005616728216409683,
0.02626267448067665,
-0.006595029961317778,
-0.03267863020300865,
-0.1175999641418457,
-0.05468757823109627,
-0.02880827896296978,
0.011662140488624573,
-0.020132549107074738,
0.0211353600025177,
-0.02790968120098114,
-0.00032909540459513664,
-0.0011530668707564473,
0.003632494481280446,
-0.07328644394874573,
-0.09259740263223648,
-0.058381207287311554,
0.09054454416036606,
-0.11559699475765228,
0.053290802985429764,
0.04389791935682297,
-0.11858618259429932,
0.16537562012672424,
0.008754834532737732,
0.011897623538970947,
0.0051797726191580296,
-0.08473861217498779,
0.012422761879861355,
-0.08453789353370667,
0.017329197376966476,
-0.038630466908216476,
-0.06975281238555908,
0.05018288269639015,
-0.06514085084199905,
-0.04459341987967491,
0.0018766314024105668,
0.008773105219006538,
-0.14527569711208344,
-0.014001312665641308,
-0.009013635106384754,
-0.0030215110164135695,
-0.08036445081233978,
0.07966256886720657,
0.0693473294377327,
-0.03774270787835121,
0.0789913684129715,
-0.05301326513290405,
0.02636348269879818,
-0.24556246399879456,
-0.004891746677458286,
0.01138309482485056,
-0.05127837508916855,
0.04146939888596535,
-0.06199297681450844,
0.1035514771938324,
-0.009952571243047714,
0.11402571201324463,
0.006745279300957918,
-0.00972243957221508,
-0.009066875092685223,
0.07898382097482681,
-0.07144393026828766,
0.07839313894510269,
0.2057618349790573,
0.02524067834019661,
-0.04071271792054176,
0.018128598108887672,
-0.012102414853870869,
0.007755974307656288,
0.10249269008636475,
0.047751422971487045,
0.07215680181980133,
0.04341376945376396,
0.08170976489782333,
0.05666624754667282,
-0.09494168311357498,
-0.10177991539239883,
0.16675716638565063,
-0.06327030062675476,
0.11172108352184296,
-0.0857103168964386,
0.0675356462597847,
-0.0152055649086833,
-0.1743594855070114,
0.14224770665168762,
-0.015160021372139454,
-0.08721651136875153,
-0.09046012908220291,
-0.2263731062412262,
-0.05896531790494919,
-0.07691220194101334,
0.03917230665683746,
-0.10858531296253204,
0.035644326359033585,
0.11518411338329315,
0.014717916958034039,
-0.016475191339850426,
0.08531777560710907,
-0.093859001994133,
-0.05929048731923103,
0.14811575412750244,
0.0628189668059349,
-0.009135686792433262,
0.05090337619185448,
-0.1322808861732483,
0.09314143657684326,
0.027792489156126976,
0.07766972482204437,
-0.01761631853878498,
0.04359818622469902,
0.05826951190829277,
0.02376338094472885,
-0.05575019121170044,
0.06278570741415024,
-0.05634591355919838,
0.010886778123676777,
0.09884760528802872,
0.007276635151356459,
-0.023826362565159798,
-0.018951699137687683,
0.2581215798854828,
-0.05548425391316414,
-0.03347572311758995,
-0.19758714735507965,
0.12621527910232544,
-0.04426547884941101,
0.011873770505189896,
0.06876835227012634,
-0.10532866418361664,
-0.061988428235054016,
0.2150338739156723,
0.0880507305264473,
-0.0030450618360191584,
-0.022518377751111984,
0.09599534422159195,
-0.03163576126098633,
0.005861134734004736,
0.12682943046092987,
0.07067190110683441,
0.04235516116023064,
-0.014781095087528229,
0.09832334518432617,
0.0539359487593174,
-0.05230935662984848,
-0.09229306876659393,
0.19248218834400177,
-0.05902494862675667,
0.025261327624320984,
-0.09245322644710541,
0.017398715019226074,
-0.01958443596959114,
-0.26790180802345276,
0.04199252277612686,
-0.08305450528860092,
-0.18109433352947235,
-0.029604291543364525,
0.03646929934620857,
0.037346482276916504,
0.07301107794046402,
0.04697399586439133,
0.024807913228869438,
0.1595003604888916,
-0.023237761110067368,
-0.01532040350139141,
-0.10518050938844681,
0.0503537617623806,
-0.054099079221487045,
0.18623755872249603,
0.03174152225255966,
-0.045029547065496445,
0.11182361841201782,
0.0032602970022708178,
-0.09749405086040497,
0.03948760777711868,
0.054517995566129684,
-0.01917731761932373,
0.08070765435695648,
0.15284770727157593,
-0.035662777721881866,
0.026663703843951225,
0.05940544232726097,
-0.05403684079647064,
0.020796503871679306,
0.0173305943608284,
-0.027966653928160667,
-0.13581769168376923,
0.07274411618709564,
-0.06691477447748184,
0.11250026524066925,
0.2249092310667038,
-0.015070540830492973,
-0.035883571952581406,
-0.06606293469667435,
0.014504943042993546,
-0.03235142305493355,
0.0344126857817173,
0.0001687125040916726,
-0.12657667696475983,
-0.0050981659442186356,
-0.05531703680753708,
0.07901141792535782,
-0.14898450672626495,
-0.028693901374936104,
0.04982053115963936,
-0.05740394443273544,
-0.0459260456264019,
0.15469104051589966,
0.00870521180331707,
0.039878182113170624,
-0.016379598528146744,
-0.02677406743168831,
-0.007406807504594326,
0.06875013560056686,
-0.12057238072156906,
-0.025259586051106453
] |
null | null |
transformers
|
# Polbert - Polish BERT
Polish version of BERT language model is here! It is now available in two variants: cased and uncased, both can be downloaded and used via HuggingFace transformers library. I recommend using the cased model, more info on the differences and benchmark results below.

## Cased and uncased variants
* I initially trained the uncased model, the corpus and training details are referenced below. Here are some issues I found after I published the uncased model:
* Some Polish characters and accents are not tokenized correctly through the BERT tokenizer when applying lowercase. This doesn't impact sequence classification much, but may influence token classfication tasks significantly.
* I noticed a lot of duplicates in the Open Subtitles dataset, which dominates the training corpus.
* I didn't use Whole Word Masking.
* The cased model improves on the uncased model in the following ways:
* All Polish characters and accents should now be tokenized correctly.
* I removed duplicates from Open Subtitles dataset. The corpus is smaller, but more balanced now.
* The model is trained with Whole Word Masking.
## Pre-training corpora
Below is the list of corpora used along with the output of `wc` command (counting lines, words and characters). These corpora were divided into sentences with srxsegmenter (see references), concatenated and tokenized with HuggingFace BERT Tokenizer.
### Uncased
| Tables | Lines | Words | Characters |
| ------------- |--------------:| -----:| -----:|
| [Polish subset of Open Subtitles](http://opus.nlpl.eu/OpenSubtitles-v2018.php) | 236635408| 1431199601 | 7628097730 |
| [Polish subset of ParaCrawl](http://opus.nlpl.eu/ParaCrawl.php) | 8470950 | 176670885 | 1163505275 |
| [Polish Parliamentary Corpus](http://clip.ipipan.waw.pl/PPC) | 9799859 | 121154785 | 938896963 |
| [Polish Wikipedia - Feb 2020](https://dumps.wikimedia.org/plwiki/latest/plwiki-latest-pages-articles.xml.bz2) | 8014206 | 132067986 | 1015849191 |
| Total | 262920423 | 1861093257 | 10746349159 |
### Cased
| Tables | Lines | Words | Characters |
| ------------- |--------------:| -----:| -----:|
| [Polish subset of Open Subtitles (Deduplicated) ](http://opus.nlpl.eu/OpenSubtitles-v2018.php) | 41998942| 213590656 | 1424873235 |
| [Polish subset of ParaCrawl](http://opus.nlpl.eu/ParaCrawl.php) | 8470950 | 176670885 | 1163505275 |
| [Polish Parliamentary Corpus](http://clip.ipipan.waw.pl/PPC) | 9799859 | 121154785 | 938896963 |
| [Polish Wikipedia - Feb 2020](https://dumps.wikimedia.org/plwiki/latest/plwiki-latest-pages-articles.xml.bz2) | 8014206 | 132067986 | 1015849191 |
| Total | 68283960 | 646479197 | 4543124667 |
## Pre-training details
### Uncased
* Polbert was trained with code provided in Google BERT's github repository (https://github.com/google-research/bert)
* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)
* Training set-up: in total 1 million training steps:
* 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)
* 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5
* 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
* The model was trained on a single Google Cloud TPU v3-8
### Cased
* Same approach as uncased model, with the following differences:
* Whole Word Masking
* Training set-up:
* 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)
* 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5
* 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
## Usage
Polbert is released via [HuggingFace Transformers library](https://huggingface.co/transformers/).
For an example use as language model, see [this notebook](/LM_testing.ipynb) file.
### Uncased
```python
from transformers import *
model = BertForMaskedLM.from_pretrained("dkleczek/bert-base-polish-uncased-v1")
tokenizer = BertTokenizer.from_pretrained("dkleczek/bert-base-polish-uncased-v1")
nlp = pipeline('fill-mask', model=model, tokenizer=tokenizer)
for pred in nlp(f"Adam Mickiewicz wielkim polskim {nlp.tokenizer.mask_token} był."):
print(pred)
# Output:
# {'sequence': '[CLS] adam mickiewicz wielkim polskim poeta był. [SEP]', 'score': 0.47196975350379944, 'token': 26596}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim bohaterem był. [SEP]', 'score': 0.09127858281135559, 'token': 10953}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim człowiekiem był. [SEP]', 'score': 0.0647173821926117, 'token': 5182}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim pisarzem był. [SEP]', 'score': 0.05232388526201248, 'token': 24293}
# {'sequence': '[CLS] adam mickiewicz wielkim polskim politykiem był. [SEP]', 'score': 0.04554257541894913, 'token': 44095}
```
### Cased
```python
model = BertForMaskedLM.from_pretrained("dkleczek/bert-base-polish-cased-v1")
tokenizer = BertTokenizer.from_pretrained("dkleczek/bert-base-polish-cased-v1")
nlp = pipeline('fill-mask', model=model, tokenizer=tokenizer)
for pred in nlp(f"Adam Mickiewicz wielkim polskim {nlp.tokenizer.mask_token} był."):
print(pred)
# Output:
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim pisarzem był. [SEP]', 'score': 0.5391148328781128, 'token': 37120}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim człowiekiem był. [SEP]', 'score': 0.11683262139558792, 'token': 6810}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim bohaterem był. [SEP]', 'score': 0.06021466106176376, 'token': 17709}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim mistrzem był. [SEP]', 'score': 0.051870670169591904, 'token': 14652}
# {'sequence': '[CLS] Adam Mickiewicz wielkim polskim artystą był. [SEP]', 'score': 0.031787533313035965, 'token': 35680}
```
See the next section for an example usage of Polbert in downstream tasks.
## Evaluation
Thanks to Allegro, we now have the [KLEJ benchmark](https://klejbenchmark.com/leaderboard/), a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.
| Model | Average | NKJP-NER | CDSC-E | CDSC-R | CBD | PolEmo2.0-IN | PolEmo2.0-OUT | DYK | PSC | AR |
| ------------- |--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|--------------:|
| Polbert cased | 81.7 | 93.6 | 93.4 | 93.8 | 52.7 | 87.4 | 71.1 | 59.1 | 98.6 | 85.2 |
| Polbert uncased | 81.4 | 90.1 | 93.9 | 93.5 | 55.0 | 88.1 | 68.8 | 59.4 | 98.8 | 85.4 |
Note how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.
## Bias
The data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.
## Acknowledgements
* I'd like to express my gratitude to Google [TensorFlow Research Cloud (TFRC)](https://www.tensorflow.org/tfrc) for providing the free TPU credits - thank you!
* Also appreciate the help from Timo Möller from [deepset](https://deepset.ai) for sharing tips and scripts based on their experience training German BERT model.
* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.
* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from [fastai](https://www.fast.ai) for their NLP and Deep Learning courses!
## Author
Darek Kłeczek - contact me on Twitter [@dk21](https://twitter.com/dk21)
## References
* https://github.com/google-research/bert
* https://github.com/narusemotoki/srx_segmenter
* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: https://raw.githubusercontent.com/languagetool-org/languagetool/master/languagetool-core/src/main/resources/org/languagetool/resource/segment.srx
* [KLEJ benchmark](https://klejbenchmark.com/leaderboard/)
|
{"language": "pl", "thumbnail": "https://raw.githubusercontent.com/kldarek/polbert/master/img/polbert.png"}
|
fill-mask
|
dkleczek/bert-base-polish-uncased-v1
|
[
"transformers",
"pytorch",
"jax",
"bert",
"fill-mask",
"pl",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"pl"
] |
TAGS
#transformers #pytorch #jax #bert #fill-mask #pl #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Polbert - Polish BERT
=====================
Polish version of BERT language model is here! It is now available in two variants: cased and uncased, both can be downloaded and used via HuggingFace transformers library. I recommend using the cased model, more info on the differences and benchmark results below.
!PolBERT image
Cased and uncased variants
--------------------------
* I initially trained the uncased model, the corpus and training details are referenced below. Here are some issues I found after I published the uncased model:
+ Some Polish characters and accents are not tokenized correctly through the BERT tokenizer when applying lowercase. This doesn't impact sequence classification much, but may influence token classfication tasks significantly.
+ I noticed a lot of duplicates in the Open Subtitles dataset, which dominates the training corpus.
+ I didn't use Whole Word Masking.
* The cased model improves on the uncased model in the following ways:
+ All Polish characters and accents should now be tokenized correctly.
+ I removed duplicates from Open Subtitles dataset. The corpus is smaller, but more balanced now.
+ The model is trained with Whole Word Masking.
Pre-training corpora
--------------------
Below is the list of corpora used along with the output of 'wc' command (counting lines, words and characters). These corpora were divided into sentences with srxsegmenter (see references), concatenated and tokenized with HuggingFace BERT Tokenizer.
### Uncased
### Cased
Pre-training details
--------------------
### Uncased
* Polbert was trained with code provided in Google BERT's github repository (URL
* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)
* Training set-up: in total 1 million training steps:
+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)
+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5
+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
* The model was trained on a single Google Cloud TPU v3-8
### Cased
* Same approach as uncased model, with the following differences:
+ Whole Word Masking
* Training set-up:
+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)
+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5
+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5
Usage
-----
Polbert is released via HuggingFace Transformers library.
For an example use as language model, see this notebook file.
### Uncased
### Cased
See the next section for an example usage of Polbert in downstream tasks.
Evaluation
----------
Thanks to Allegro, we now have the KLEJ benchmark, a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.
Note how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.
Bias
----
The data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.
Acknowledgements
----------------
* I'd like to express my gratitude to Google TensorFlow Research Cloud (TFRC) for providing the free TPU credits - thank you!
* Also appreciate the help from Timo Möller from deepset for sharing tips and scripts based on their experience training German BERT model.
* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.
* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from fastai for their NLP and Deep Learning courses!
Author
------
Darek Kłeczek - contact me on Twitter @dk21
References
----------
* URL
* URL
* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: URL
* KLEJ benchmark
|
[
"### Uncased",
"### Cased\n\n\n\nPre-training details\n--------------------",
"### Uncased\n\n\n* Polbert was trained with code provided in Google BERT's github repository (URL\n* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)\n* Training set-up: in total 1 million training steps:\n\t+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)\n\t+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n* The model was trained on a single Google Cloud TPU v3-8",
"### Cased\n\n\n* Same approach as uncased model, with the following differences:\n\t+ Whole Word Masking\n* Training set-up:\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n\n\nUsage\n-----\n\n\nPolbert is released via HuggingFace Transformers library.\n\n\nFor an example use as language model, see this notebook file.",
"### Uncased",
"### Cased\n\n\nSee the next section for an example usage of Polbert in downstream tasks.\n\n\nEvaluation\n----------\n\n\nThanks to Allegro, we now have the KLEJ benchmark, a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.\n\n\n\nNote how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.\n\n\nBias\n----\n\n\nThe data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.\n\n\nAcknowledgements\n----------------\n\n\n* I'd like to express my gratitude to Google TensorFlow Research Cloud (TFRC) for providing the free TPU credits - thank you!\n* Also appreciate the help from Timo Möller from deepset for sharing tips and scripts based on their experience training German BERT model.\n* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.\n* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from fastai for their NLP and Deep Learning courses!\n\n\nAuthor\n------\n\n\nDarek Kłeczek - contact me on Twitter @dk21\n\n\nReferences\n----------\n\n\n* URL\n* URL\n* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: URL\n* KLEJ benchmark"
] |
[
"TAGS\n#transformers #pytorch #jax #bert #fill-mask #pl #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"### Uncased",
"### Cased\n\n\n\nPre-training details\n--------------------",
"### Uncased\n\n\n* Polbert was trained with code provided in Google BERT's github repository (URL\n* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)\n* Training set-up: in total 1 million training steps:\n\t+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)\n\t+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n* The model was trained on a single Google Cloud TPU v3-8",
"### Cased\n\n\n* Same approach as uncased model, with the following differences:\n\t+ Whole Word Masking\n* Training set-up:\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n\n\nUsage\n-----\n\n\nPolbert is released via HuggingFace Transformers library.\n\n\nFor an example use as language model, see this notebook file.",
"### Uncased",
"### Cased\n\n\nSee the next section for an example usage of Polbert in downstream tasks.\n\n\nEvaluation\n----------\n\n\nThanks to Allegro, we now have the KLEJ benchmark, a set of nine evaluation tasks for the Polish language understanding. The following results are achieved by running standard set of evaluation scripts (no tricks!) utilizing both cased and uncased variants of Polbert.\n\n\n\nNote how the uncased model performs better than cased on some tasks? My guess this is because of the oversampling of Open Subtitles dataset and its similarity to data in some of these tasks. All these benchmark tasks are sequence classification, so the relative strength of the cased model is not so visible here.\n\n\nBias\n----\n\n\nThe data used to train the model is biased. It may reflect stereotypes related to gender, ethnicity etc. Please be careful when using the model for downstream task to consider these biases and mitigate them.\n\n\nAcknowledgements\n----------------\n\n\n* I'd like to express my gratitude to Google TensorFlow Research Cloud (TFRC) for providing the free TPU credits - thank you!\n* Also appreciate the help from Timo Möller from deepset for sharing tips and scripts based on their experience training German BERT model.\n* Big thanks to Allegro for releasing KLEJ Benchmark and specifically to Piotr Rybak for help with the evaluation and pointing out some issues with the tokenization.\n* Finally, thanks to Rachel Thomas, Jeremy Howard and Sylvain Gugger from fastai for their NLP and Deep Learning courses!\n\n\nAuthor\n------\n\n\nDarek Kłeczek - contact me on Twitter @dk21\n\n\nReferences\n----------\n\n\n* URL\n* URL\n* SRX rules file for sentence splitting in Polish, written by Marcin Miłkowski: URL\n* KLEJ benchmark"
] |
[
45,
5,
10,
160,
131,
5,
397
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #fill-mask #pl #autotrain_compatible #endpoints_compatible #has_space #region-us \n### Uncased### Cased\n\n\n\nPre-training details\n--------------------### Uncased\n\n\n* Polbert was trained with code provided in Google BERT's github repository (URL\n* Currently released model follows bert-base-uncased model architecture (12-layer, 768-hidden, 12-heads, 110M parameters)\n* Training set-up: in total 1 million training steps:\n\t+ 100.000 steps - 128 sequence length, batch size 512, learning rate 1e-4 (10.000 steps warmup)\n\t+ 800.000 steps - 128 sequence length, batch size 512, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n* The model was trained on a single Google Cloud TPU v3-8### Cased\n\n\n* Same approach as uncased model, with the following differences:\n\t+ Whole Word Masking\n* Training set-up:\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 1e-4 (10.000 steps warmup)\n\t+ 100.000 steps - 128 sequence length, batch size 2048, learning rate 5e-5\n\t+ 100.000 steps - 512 sequence length, batch size 256, learning rate 2e-5\n\n\nUsage\n-----\n\n\nPolbert is released via HuggingFace Transformers library.\n\n\nFor an example use as language model, see this notebook file.### Uncased"
] |
[
-0.10600079596042633,
0.16206501424312592,
-0.0018463198794052005,
0.09549283236265182,
0.11026883870363235,
0.05853226035833359,
0.1143232211470604,
0.11540725082159042,
0.010512463748455048,
0.12417320907115936,
0.054481420665979385,
-0.021328669041395187,
0.12725934386253357,
0.16930551826953888,
0.05856161564588547,
-0.3251437544822693,
0.011534535326063633,
-0.0931750014424324,
-0.03036070428788662,
0.07168462872505188,
0.05275430530309677,
-0.04890773072838783,
0.06308580189943314,
-0.02798408828675747,
0.04530588537454605,
-0.0552479587495327,
-0.07402680069208145,
-0.06746622920036316,
0.07875098288059235,
0.00534356152638793,
0.04570826143026352,
0.012862535193562508,
0.08625111728906631,
-0.12407033145427704,
0.007936157286167145,
0.10875272005796432,
-0.02953253500163555,
0.06386072188615799,
0.09223272651433945,
-0.013959668576717377,
0.0445459708571434,
-0.10297249257564545,
0.063685342669487,
0.051755551248788834,
-0.10446856915950775,
-0.12108515948057175,
-0.09422187507152557,
0.12127623707056046,
0.0957508385181427,
0.01981949247419834,
-0.03275235369801521,
-0.01148197054862976,
-0.06500479578971863,
0.04845013841986656,
0.1819467842578888,
-0.32363852858543396,
-0.06686688214540482,
0.08200182020664215,
0.03688527271151543,
0.04845274239778519,
-0.08135492354631424,
-0.003352305619046092,
0.005896633956581354,
0.020397409796714783,
0.015982920303940773,
0.06327393651008606,
0.05298883467912674,
0.016648247838020325,
-0.11221373826265335,
-0.03921376168727875,
0.08323731273412704,
0.048518866300582886,
-0.0592954084277153,
-0.143106609582901,
-0.07063709199428558,
-0.18094705045223236,
-0.0434587337076664,
0.033408619463443756,
0.07381969690322876,
0.01459308248013258,
-0.05393780395388603,
-0.030365515500307083,
-0.03818465769290924,
-0.03880969062447548,
0.016946585848927498,
0.09771838039159775,
0.06664484739303589,
0.0008414395269937813,
0.041552118957042694,
0.09633709490299225,
0.07491172105073929,
-0.08386818319559097,
-0.02578885294497013,
0.02348668873310089,
-0.09758058935403824,
0.00007157345680752769,
-0.02630331739783287,
-0.01938588358461857,
0.052725739777088165,
0.11356540024280548,
0.03392034396529198,
0.0970824733376503,
0.04825572296977043,
-0.008190942928195,
-0.00797740463167429,
0.19422373175621033,
-0.11329768598079681,
-0.06853263080120087,
0.02375302091240883,
0.07799513638019562,
0.020770180970430374,
-0.035877108573913574,
-0.05275605618953705,
-0.009269527159631252,
0.1456827074289322,
0.08504432439804077,
-0.0035461552906781435,
0.10073422640562057,
-0.04711049795150757,
-0.033010050654411316,
0.07864310592412949,
-0.14104318618774414,
0.02102675288915634,
0.053956180810928345,
-0.05752328783273697,
-0.013678474351763725,
0.02125297486782074,
-0.08576298505067825,
-0.1420826017856598,
0.06295635551214218,
-0.09180017560720444,
-0.09031381458044052,
-0.06788225471973419,
-0.06809171289205551,
0.04260481894016266,
-0.09704294055700302,
-0.08404096215963364,
-0.06093123182654381,
-0.15514115989208221,
-0.07717510312795639,
0.06873849779367447,
-0.0679253563284874,
-0.043347157537937164,
-0.07629721611738205,
-0.05832017585635185,
0.044544268399477005,
0.019182337448000908,
0.07401898503303528,
-0.048240821808576584,
0.07656827569007874,
-0.09390145540237427,
0.08901198208332062,
0.1312631368637085,
0.019147822633385658,
-0.04880635067820549,
0.06423146277666092,
-0.19802530109882355,
0.1348457634449005,
-0.0813925638794899,
-0.014962713234126568,
-0.11913524568080902,
-0.07280656695365906,
-0.0025056330487132072,
-0.012284507974982262,
0.08274150639772415,
0.15171091258525848,
-0.14348793029785156,
-0.0466948002576828,
0.17829258739948273,
-0.04891771823167801,
-0.07276235520839691,
0.12856508791446686,
0.021535079926252365,
-0.04901564121246338,
0.027081558480858803,
0.06824707984924316,
0.04406947270035744,
-0.13647779822349548,
-0.079769067466259,
0.0022997912019491196,
0.0024000199045985937,
0.11276137083768845,
0.027664924040436745,
-0.05615318566560745,
-0.0346294641494751,
0.02400871552526951,
-0.01028597541153431,
-0.0010100390063598752,
-0.05031967535614967,
-0.04111644998192787,
-0.02392931655049324,
-0.058877136558294296,
0.021915815770626068,
-0.016588876023888588,
-0.023398838937282562,
-0.04785308986902237,
-0.13582469522953033,
-0.06745333224534988,
0.10145527869462967,
-0.1197977289557457,
-0.005250771529972553,
-0.0718696191906929,
0.06960704177618027,
0.01396357361227274,
0.04186450317502022,
-0.08099471777677536,
-0.19453027844429016,
0.1018245592713356,
-0.1687949299812317,
-0.017488930374383926,
-0.07046979665756226,
0.041785381734371185,
0.04402859881520271,
-0.08407231420278549,
-0.05508149787783623,
-0.03491263836622238,
-0.038397371768951416,
-0.07942407578229904,
-0.10872410237789154,
-0.08214500546455383,
0.02387380041182041,
0.08453205227851868,
-0.03904358297586441,
0.029667267575860023,
0.024494439363479614,
0.13083022832870483,
0.0025769053027033806,
-0.07951989024877548,
0.06775476783514023,
-0.0493263304233551,
0.040738336741924286,
-0.13672475516796112,
0.04182935133576393,
0.0382385328412056,
-0.04955613240599632,
0.04503781720995903,
-0.1323624551296234,
-0.06485150009393692,
0.0891374945640564,
0.11273717135190964,
-0.10226678103208542,
0.005538237281143665,
-0.05355669930577278,
-0.04232305288314819,
-0.05896719545125961,
0.022490689530968666,
0.14357662200927734,
0.04820859432220459,
0.11301250755786896,
-0.1303880512714386,
-0.0582134947180748,
0.03446226939558983,
0.042236655950546265,
-0.018838802352547646,
0.10740542411804199,
-0.03246302157640457,
-0.06571728736162186,
0.03907215595245361,
-0.025939572602510452,
0.004556255880743265,
0.1450730562210083,
-0.0370301678776741,
-0.10947148501873016,
-0.037904150784015656,
0.072238489985466,
0.035901233553886414,
0.07664413750171661,
-0.12053781002759933,
-0.00018072797683998942,
0.02255498617887497,
0.012024329043924809,
0.01781817525625229,
-0.09025219082832336,
0.03540746867656708,
-0.03249359503388405,
-0.09516974538564682,
-0.026986325159668922,
-0.001965360715985298,
-0.01203309278935194,
0.08106724917888641,
0.053618378937244415,
0.009966982528567314,
0.008648009039461613,
-0.027740705758333206,
-0.09130822122097015,
0.174623042345047,
-0.08094365149736404,
-0.21801871061325073,
-0.1095011830329895,
0.12371733039617538,
-0.06202240288257599,
0.003013949142768979,
-0.04060832038521767,
-0.06567717343568802,
-0.054867107421159744,
-0.06488684564828873,
0.07701113075017929,
0.048697616904973984,
0.03573361784219742,
-0.04988241568207741,
0.012202329933643341,
0.05763982981443405,
-0.13880714774131775,
0.004346578847616911,
-0.006857885513454676,
-0.15294483304023743,
0.005942291580140591,
0.0021142675541341305,
0.013718797825276852,
0.1179608553647995,
-0.011541685089468956,
0.02191799320280552,
0.014521660283207893,
0.21456468105316162,
-0.08958546817302704,
0.13430441915988922,
0.07284817844629288,
0.04277263209223747,
0.06694944202899933,
0.04504920542240143,
0.03506115451455116,
-0.09545507282018661,
0.06097746267914772,
0.07280106842517853,
-0.000696280098054558,
-0.2632429897785187,
-0.036015354096889496,
-0.023390550166368484,
0.09687941521406174,
0.14018486440181732,
0.07810010015964508,
-0.08471876382827759,
0.04027444124221802,
-0.08558843284845352,
0.05018562078475952,
0.02810004912316799,
0.05659107863903046,
0.008496461436152458,
0.020189868286252022,
0.025825664401054382,
-0.03196342661976814,
0.00891607441008091,
0.07015103101730347,
0.03028498962521553,
0.14844219386577606,
-0.07742968201637268,
0.18482717871665955,
-0.007708619348704815,
0.16181489825248718,
0.03915494307875633,
0.10908550024032593,
0.0002707969688344747,
0.026766888797283173,
0.017396707087755203,
-0.05959298461675644,
-0.040797341614961624,
0.04404251277446747,
0.06774357706308365,
-0.02781134843826294,
-0.07415147870779037,
-0.008674758486449718,
0.0505850650370121,
0.429711252450943,
0.12212010473012924,
-0.18608500063419342,
-0.08418525010347366,
-0.03533148020505905,
-0.08131156861782074,
-0.13174179196357727,
0.057870447635650635,
0.08176102489233017,
-0.11507681757211685,
0.06581951677799225,
-0.070180743932724,
0.03590581566095352,
-0.15574632585048676,
-0.03292775899171829,
0.06663762778043747,
0.15569406747817993,
-0.0482296496629715,
0.06429308652877808,
-0.1472877711057663,
0.11328255385160446,
0.016397083178162575,
0.08395408093929291,
-0.06272964179515839,
0.00922811683267355,
0.0607927143573761,
-0.05658987537026405,
0.1053929477930069,
0.029855776578187943,
-0.002686251187697053,
-0.06374769657850266,
-0.14440549910068512,
0.037287041544914246,
0.12060771882534027,
-0.09399010986089706,
0.07711289077997208,
-0.0374797023832798,
-0.034974705427885056,
-0.03076820634305477,
-0.02460208348929882,
-0.09295569360256195,
-0.09151152521371841,
0.014583162032067776,
-0.08133509010076523,
0.07465098798274994,
-0.10714030265808105,
-0.059641703963279724,
-0.05864531919360161,
0.15091750025749207,
-0.04480435699224472,
-0.11262059956789017,
-0.12000194191932678,
-0.00024398186360485852,
0.15284521877765656,
-0.05104687437415123,
0.07368385791778564,
-0.030136939138174057,
0.11681295186281204,
0.03559800982475281,
-0.04585367813706398,
0.10934676229953766,
-0.06031964719295502,
-0.2548424303531647,
-0.04920307919383049,
0.12810227274894714,
0.03129386901855469,
0.02745734713971615,
-0.024950431659817696,
0.02075052447617054,
0.004649846348911524,
-0.0632583498954773,
0.06219598650932312,
0.03982175514101982,
0.03174761310219765,
0.11294350028038025,
-0.07952149957418442,
0.08035426586866379,
-0.017010696232318878,
-0.00440865708515048,
0.026794176548719406,
0.19152021408081055,
-0.04519527778029442,
0.06272513419389725,
0.19181139767169952,
-0.03453832119703293,
-0.22422218322753906,
-0.05789951980113983,
0.07959744334220886,
0.01424036268144846,
-0.0829090103507042,
-0.2481546550989151,
0.05004800856113434,
0.09215518832206726,
-0.025690190494060516,
0.0735015720129013,
-0.22910621762275696,
-0.12510152161121368,
0.07033112645149231,
0.024613840505480766,
0.02157980389893055,
-0.08523594588041306,
-0.003112528007477522,
-0.02681056223809719,
0.0010596446227282286,
0.07513518631458282,
0.06357276439666748,
0.16327539086341858,
-0.007293251343071461,
-0.03520728275179863,
0.03605170547962189,
-0.04907219111919403,
0.08835174143314362,
-0.03471018746495247,
0.04628104344010353,
-0.04851610213518143,
0.029195597395300865,
-0.01289108395576477,
-0.06508270651102066,
0.11737561225891113,
-0.03389458358287811,
0.12965446710586548,
0.001863485318608582,
-0.044768236577510834,
-0.06117400527000427,
0.0519578792154789,
-0.03696608543395996,
-0.04085661470890045,
-0.06860096752643585,
0.05962790921330452,
0.09285055845975876,
0.000679024844430387,
-0.027214329689741135,
0.03245371952652931,
-0.014281212352216244,
0.08405430614948273,
0.05365870147943497,
0.0010892539285123348,
-0.09152255207300186,
-0.03774600103497505,
-0.04668937623500824,
0.04077938571572304,
-0.09412703663110733,
0.031871646642684937,
0.06329915672540665,
0.022657282650470734,
0.11599721014499664,
0.0319102443754673,
-0.12764479219913483,
-0.03559165447950363,
0.03128550201654434,
-0.1037888377904892,
-0.19622240960597992,
-0.04481964558362961,
-0.14180918037891388,
-0.1633453667163849,
0.0009430669597350061,
0.07177069038152695,
-0.042535342276096344,
-0.018471892923116684,
-0.00464317062869668,
0.048759397119283676,
0.025672907009720802,
0.12208901345729828,
0.036252159625291824,
0.054120760411024094,
-0.10176251083612442,
0.22248750925064087,
0.09164612740278244,
-0.09069838374853134,
0.06162844970822334,
0.12654997408390045,
-0.07815129309892654,
-0.011166401207447052,
0.0305011086165905,
0.10509632527828217,
0.07407020032405853,
0.0040543717332184315,
-0.05029010400176048,
-0.049484167248010635,
0.1082814484834671,
0.06408198177814484,
0.0017479994567111135,
0.09685055911540985,
0.030435698106884956,
0.05293668806552887,
-0.04852420464158058,
0.09143457561731339,
0.06080131232738495,
0.031564995646476746,
-0.07018988579511642,
0.10541920363903046,
0.015160778537392616,
0.010712806135416031,
-0.01786431297659874,
-0.01745005138218403,
-0.13530173897743225,
-0.0535498782992363,
-0.05250776559114456,
-0.007368951104581356,
0.00505224522203207,
0.02331865206360817,
-0.020249811932444572,
-0.014128508046269417,
-0.0022328707855194807,
0.006423034239560366,
-0.08179005980491638,
-0.09971056878566742,
-0.022971153259277344,
0.08527970314025879,
-0.10410900413990021,
0.013372531160712242,
0.05244801938533783,
-0.12504644691944122,
0.1883564442396164,
0.014527137391269207,
-0.001553685637190938,
0.016988379880785942,
-0.14242444932460785,
-0.0009514059638604522,
-0.0591471828520298,
-0.003235378535464406,
-0.018926037475466728,
-0.12030170857906342,
0.04253565892577171,
-0.07736707478761673,
-0.06327859312295914,
0.006037170998752117,
0.06344037503004074,
-0.13560935854911804,
0.010672066360712051,
-0.04365153610706329,
0.015381703153252602,
-0.09874667227268219,
0.0315653532743454,
0.05487270653247833,
0.008640670217573643,
0.0827791765332222,
-0.08292006701231003,
-0.0015461163129657507,
-0.20248079299926758,
-0.000837634433992207,
0.009358696639537811,
-0.042997851967811584,
0.004840744659304619,
-0.020049601793289185,
0.11306338012218475,
-0.01732000522315502,
0.07326839119195938,
0.002051936462521553,
-0.038215454667806625,
-0.0027387449517846107,
0.1037660464644432,
-0.11539755016565323,
0.052416447550058365,
0.11265358328819275,
-0.00407864386215806,
-0.023474223911762238,
-0.002009020419791341,
0.0010020789923146367,
-0.0014580629067495465,
0.07486693561077118,
0.06639111787080765,
0.016162320971488953,
0.016988394781947136,
0.10857836902141571,
0.014593738131225109,
-0.09889265149831772,
-0.119256392121315,
0.17583924531936646,
-0.0855015218257904,
0.15743085741996765,
-0.07714922726154327,
0.10726674646139145,
0.02871047332882881,
-0.16869743168354034,
0.10651654750108719,
-0.011322134174406528,
-0.08472081273794174,
-0.0892084538936615,
-0.19446901977062225,
-0.05191905423998833,
-0.10173924267292023,
0.03506764769554138,
-0.08240944892168045,
0.05296308174729347,
0.09653513878583908,
0.039150115102529526,
0.0005748551338911057,
0.08176854252815247,
-0.10210435837507248,
-0.023278648033738136,
0.1468244045972824,
0.06293818354606628,
0.008756681345403194,
0.0659966990351677,
-0.14325673878192902,
0.031594350934028625,
0.026035893708467484,
0.05580464377999306,
-0.0074684591963887215,
0.019071638584136963,
0.07714596390724182,
0.020747072994709015,
-0.07737619429826736,
0.04857252910733223,
-0.05905097350478172,
0.02688600867986679,
0.13014206290245056,
0.03151188790798187,
-0.025281649082899094,
-0.00532272644340992,
0.24495482444763184,
-0.05656353756785393,
-0.027141116559505463,
-0.18904611468315125,
0.06404288113117218,
-0.046694979071617126,
-0.025821708142757416,
0.06295124441385269,
-0.10765647888183594,
-0.05105876922607422,
0.20733866095542908,
0.08815762400627136,
-0.03866413235664368,
-0.017697621136903763,
0.06689375638961792,
-0.026156865060329437,
0.013547787442803383,
0.12445902079343796,
0.0600019209086895,
0.07292335480451584,
-0.025024211034178734,
0.09218138456344604,
-0.0001351945975329727,
-0.06580205261707306,
-0.08884886652231216,
0.15132996439933777,
-0.05984660983085632,
-0.007181418593972921,
-0.08455538749694824,
0.03638166934251785,
-0.048886898905038834,
-0.25640586018562317,
0.06444106996059418,
-0.09188888967037201,
-0.16970868408679962,
-0.03483405336737633,
-0.012759693898260593,
0.07340189069509506,
0.0920126661658287,
0.05081552267074585,
0.045876964926719666,
0.12332063168287277,
-0.03158250451087952,
-0.02804228477180004,
-0.11274304240942001,
0.02152114175260067,
-0.02786312811076641,
0.242351233959198,
0.03501654416322708,
-0.03640434145927429,
0.11478517204523087,
0.009803032502532005,
-0.13021239638328552,
0.03080916404724121,
0.05080695450305939,
-0.08252064883708954,
0.08751406520605087,
0.14671257138252258,
-0.03187251836061478,
-0.012780928984284401,
0.04469722881913185,
-0.02727164328098297,
0.016527779400348663,
-0.006529207341372967,
0.002291291719302535,
-0.1368604302406311,
0.07747161388397217,
-0.052751220762729645,
0.12195959687232971,
0.2436521351337433,
-0.03494881093502045,
-0.0300352331250906,
-0.0766223818063736,
-0.008934849873185158,
-0.007459108252078295,
-0.027425024658441544,
0.014674302190542221,
-0.12585605680942535,
0.029428808018565178,
-0.055464521050453186,
0.05205266550183296,
-0.10678023099899292,
-0.026377389207482338,
0.0077736712992191315,
-0.04088350385427475,
-0.09500467032194138,
0.16604280471801758,
0.03375530615448952,
0.018551789224147797,
-0.016399702057242393,
0.02691732905805111,
-0.00006158298492664471,
0.07224828749895096,
-0.12678712606430054,
-0.07973215728998184
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# papuGaPT2-finetuned-wierszyki
This model is a fine-tuned version of [flax-community/papuGaPT2](https://huggingface.co/flax-community/papuGaPT2) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 2.8122
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 3e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 202 | 2.8122 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.9.0+cu111
- Datasets 1.14.0
- Tokenizers 0.10.3
|
{"tags": ["generated_from_trainer"], "model-index": [{"name": "papuGaPT2-finetuned-wierszyki", "results": []}]}
|
text-generation
|
dkleczek/papuGaPT2-finetuned-wierszyki
|
[
"transformers",
"pytorch",
"tensorboard",
"gpt2",
"text-generation",
"generated_from_trainer",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
papuGaPT2-finetuned-wierszyki
=============================
This model is a fine-tuned version of flax-community/papuGaPT2 on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 2.8122
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 3e-05
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 1
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.9.0+cu111
* Datasets 1.14.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
58,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #gpt2 #text-generation #generated_from_trainer #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
-0.09614159166812897,
0.04201499745249748,
-0.0022480227053165436,
0.10902242362499237,
0.18140120804309845,
0.02169443853199482,
0.12649346888065338,
0.12654303014278412,
-0.1190166249871254,
0.028375444933772087,
0.13705125451087952,
0.1701551228761673,
0.008736299350857735,
0.11633026599884033,
-0.048457976430654526,
-0.275620698928833,
-0.01987922377884388,
0.05969855189323425,
-0.059672947973012924,
0.14740437269210815,
0.08572389930486679,
-0.13741810619831085,
0.07684752345085144,
-0.006540319882333279,
-0.23731446266174316,
0.003007104154676199,
0.01075079943984747,
-0.06391964107751846,
0.15649351477622986,
0.016899170354008675,
0.1269245147705078,
0.007504220120608807,
0.07126634567975998,
-0.15916772186756134,
0.012465966865420341,
0.04699402302503586,
0.018732983618974686,
0.09236875921487808,
0.061866287142038345,
-0.007369668688625097,
0.14529447257518768,
-0.07337306439876556,
0.042762886732816696,
0.011838681995868683,
-0.13595378398895264,
-0.1943490356206894,
-0.06534266471862793,
0.011943165212869644,
0.053741395473480225,
0.11909569799900055,
-0.020278222858905792,
0.1292501837015152,
-0.077948197722435,
0.1082727313041687,
0.24483495950698853,
-0.2635783553123474,
-0.07032252103090286,
0.0291746873408556,
0.023417029529809952,
0.0930391326546669,
-0.10641119629144669,
-0.00840219110250473,
0.05364719778299332,
0.05386203154921532,
0.12898853421211243,
-0.03258546441793442,
-0.09129858762025833,
0.017366604879498482,
-0.1501334309577942,
-0.03908729925751686,
0.13221068680286407,
0.008835069835186005,
-0.019912850111722946,
-0.047131191939115524,
-0.0682731419801712,
-0.1707032471895218,
-0.03388557955622673,
-0.020242996513843536,
0.02148349955677986,
-0.04261451214551926,
-0.09010305255651474,
-0.03037853352725506,
-0.11127080023288727,
-0.07293090224266052,
-0.06686195731163025,
0.15425169467926025,
0.0404122918844223,
-0.000025975308744818904,
-0.04092532768845558,
0.12592092156410217,
0.0016468053217977285,
-0.12674321234226227,
0.0371733121573925,
0.03066927008330822,
0.01833966374397278,
-0.04351956024765968,
-0.07347788661718369,
-0.09199272841215134,
-0.004099455662071705,
0.08124551177024841,
-0.07904917746782303,
0.04295021668076515,
0.034834668040275574,
0.04335546866059303,
-0.0877690538764,
0.19603225588798523,
-0.03607960790395737,
0.002181333489716053,
-0.0014755036681890488,
0.053298015147447586,
-0.010322781279683113,
-0.021382341161370277,
-0.11564350128173828,
-0.00754179572686553,
0.13463908433914185,
0.015253355726599693,
-0.07229702919721603,
0.07916750013828278,
-0.0398951955139637,
-0.030324973165988922,
-0.019846763461828232,
-0.09557907283306122,
0.024001315236091614,
-0.00976545549929142,
-0.08774474263191223,
0.006518141832202673,
0.022621367126703262,
0.005842607468366623,
-0.047818221151828766,
0.11585961282253265,
-0.08648115396499634,
0.04554026573896408,
-0.09950898587703705,
-0.11718852818012238,
-0.0026434557512402534,
-0.08581417798995972,
0.011373508721590042,
-0.09515353292226791,
-0.16730765998363495,
-0.025235213339328766,
0.033433329313993454,
-0.0290653295814991,
-0.04848535358905792,
-0.07221256196498871,
-0.07112816721200943,
0.005154090467840433,
-0.016804585233330727,
0.1461833119392395,
-0.0498763769865036,
0.12954124808311462,
0.052478890866041183,
0.07414235174655914,
-0.04121994599699974,
0.0632011666893959,
-0.09050675481557846,
-0.0057768686674535275,
-0.188292995095253,
0.07420729100704193,
-0.02646048739552498,
0.0582512728869915,
-0.0763947069644928,
-0.11906439810991287,
-0.001270861248485744,
0.00856040045619011,
0.08638370037078857,
0.10677634179592133,
-0.14955896139144897,
-0.09877225011587143,
0.18390800058841705,
-0.058813195675611496,
-0.09794272482395172,
0.12339331209659576,
-0.06081441044807434,
0.06993928551673889,
0.08670009672641754,
0.186706081032753,
0.052195679396390915,
-0.06260748952627182,
0.028300419449806213,
-0.01676318049430847,
0.04841060936450958,
-0.047020863741636276,
0.04705292731523514,
0.00291751092299819,
0.002975681098178029,
0.035339683294296265,
-0.012035165913403034,
0.06381849199533463,
-0.11804147809743881,
-0.08147880434989929,
-0.03536267951130867,
-0.09431929141283035,
0.06444954127073288,
0.07321394979953766,
0.10845308750867844,
-0.09958655387163162,
-0.07160993665456772,
0.0620250329375267,
0.06491781771183014,
-0.08364822715520859,
0.03346657752990723,
-0.04460157826542854,
0.06990212202072144,
-0.0456998310983181,
-0.011699939146637917,
-0.19999122619628906,
-0.015745868906378746,
0.00934111699461937,
0.042244166135787964,
0.03162335976958275,
0.018827900290489197,
0.08036044239997864,
0.06490843743085861,
-0.05547330155968666,
-0.017910683527588844,
-0.019927944988012314,
-0.024492066353559494,
-0.14608152210712433,
-0.16876059770584106,
-0.008053658530116081,
-0.015719162300229073,
0.11723054945468903,
-0.207109272480011,
0.04072853922843933,
-0.013341275975108147,
0.05994964763522148,
0.0011625704355537891,
-0.020848561078310013,
-0.03685544431209564,
0.09235244244337082,
-0.03503308817744255,
-0.04455116018652916,
0.08481060713529587,
-0.004501147661358118,
-0.08871868997812271,
-0.044745657593011856,
-0.12683367729187012,
0.1653163880109787,
0.13760793209075928,
-0.15629848837852478,
-0.09629246592521667,
-0.016291804611682892,
-0.06126758083701134,
-0.026693351566791534,
-0.04543808847665787,
0.020525146275758743,
0.213722825050354,
-0.012686521746218204,
0.16054628789424896,
-0.07139118760824203,
-0.04985692724585533,
0.021228965371847153,
-0.035669539123773575,
0.044706858694553375,
0.1333426833152771,
0.09743792563676834,
-0.056298013776540756,
0.13394507765769958,
0.11686722189188004,
-0.09090474247932434,
0.15984784066677094,
-0.02491021901369095,
-0.0785488486289978,
-0.002982086269184947,
-0.009415282867848873,
-0.003506092121824622,
0.0657469853758812,
-0.15555895864963531,
-0.019400443881750107,
0.014215882867574692,
0.02581474930047989,
0.040982600301504135,
-0.23275114595890045,
-0.04388664290308952,
0.03281354904174805,
-0.04265546798706055,
0.0032033948227763176,
-0.00666556553915143,
0.012792842462658882,
0.1204032152891159,
0.0012042999733239412,
-0.05470697209239006,
0.03109077550470829,
0.009271192364394665,
-0.08015740662813187,
0.22335343062877655,
-0.06597967445850372,
-0.1792546659708023,
-0.11708797514438629,
-0.06903138756752014,
-0.04883011430501938,
0.020856626331806183,
0.061655011028051376,
-0.11369486153125763,
-0.006662519183009863,
-0.05627688765525818,
0.06182383373379707,
-0.01360235083848238,
0.03895309939980507,
0.000567349256016314,
-0.006557850167155266,
0.046057458966970444,
-0.10503079742193222,
-0.008785827085375786,
-0.06828422844409943,
-0.08398929238319397,
0.07028695940971375,
0.024696573615074158,
0.11520043015480042,
0.1729312688112259,
-0.024334825575351715,
0.020899787545204163,
-0.03684801608324051,
0.23274871706962585,
-0.08394765853881836,
-0.03711462765932083,
0.12768501043319702,
-0.002820373745635152,
0.053127750754356384,
0.07693342119455338,
0.06083038076758385,
-0.09310710430145264,
0.01792663149535656,
0.027180835604667664,
-0.04557487741112709,
-0.21848146617412567,
-0.04196305572986603,
-0.05201226845383644,
-0.005038474220782518,
0.09559077769517899,
0.022844139486551285,
0.05887867882847786,
0.07630809396505356,
0.04746509715914726,
0.08526334166526794,
-0.035832855850458145,
0.04735628888010979,
0.11999254673719406,
0.0320873036980629,
0.1312115490436554,
-0.036753099411726,
-0.0896138921380043,
0.034475963562726974,
-0.02586718276143074,
0.22832323610782623,
-0.0223863385617733,
0.1247812956571579,
0.044582054018974304,
0.15729404985904694,
0.007027852348983288,
0.09134756773710251,
-0.00955996848642826,
-0.04760647937655449,
-0.010657732374966145,
-0.035522617399692535,
-0.04786008223891258,
0.0036769055295735598,
-0.053880639374256134,
0.04232364892959595,
-0.13450928032398224,
-0.019509436562657356,
0.05549130588769913,
0.2230648249387741,
0.027257882058620453,
-0.32974547147750854,
-0.0924767553806305,
-0.00685447221621871,
-0.030783118680119514,
-0.02356298640370369,
0.02042446658015251,
0.11125872284173965,
-0.1106351837515831,
0.012557636946439743,
-0.07583118230104446,
0.09432688355445862,
-0.0687979981303215,
0.059325408190488815,
0.04212655499577522,
0.09685345739126205,
-0.005789540708065033,
0.08330625295639038,
-0.32829782366752625,
0.27101361751556396,
-0.0010856610024347901,
0.06881369650363922,
-0.0819522961974144,
-0.011351745575666428,
0.024183621630072594,
0.04039600118994713,
0.05363721773028374,
-0.014500918798148632,
0.0007161321700550616,
-0.20252251625061035,
-0.05310260131955147,
0.03364723175764084,
0.11953093856573105,
-0.03764178231358528,
0.10109180957078934,
-0.019073275849223137,
0.01841714419424534,
0.06711480766534805,
-0.05433868244290352,
-0.04180299490690231,
-0.09041643887758255,
0.0005086457822471857,
0.00839200522750616,
-0.024818293750286102,
-0.04895268753170967,
-0.11877206712961197,
-0.12378983944654465,
0.15597259998321533,
0.015102211385965347,
-0.04284896329045296,
-0.11388415843248367,
0.10321099311113358,
0.07937899231910706,
-0.09008467942476273,
0.033802252262830734,
0.01657639630138874,
0.05340620130300522,
0.026423022150993347,
-0.06255112588405609,
0.10910695046186447,
-0.04895305633544922,
-0.16226260364055634,
-0.042972248047590256,
0.11785981804132462,
0.04389010742306709,
0.06035121530294418,
-0.018607122823596,
0.007605569902807474,
-0.048450320959091187,
-0.09706439077854156,
0.03465994447469711,
-0.02395416609942913,
0.06597305089235306,
0.035796429961919785,
-0.05554388836026192,
0.028434516862034798,
-0.07409845292568207,
-0.03826633840799332,
0.22126227617263794,
0.22625643014907837,
-0.07726676017045975,
0.015630066394805908,
0.0391787588596344,
-0.07448782026767731,
-0.2009611278772354,
0.04767516627907753,
0.07114630937576294,
0.012593940831720829,
0.025588098913431168,
-0.1949339359998703,
0.0964319184422493,
0.09951861202716827,
0.003721012733876705,
0.1259893774986267,
-0.3579227030277252,
-0.13072362542152405,
0.10895395278930664,
0.1556534618139267,
0.13636164367198944,
-0.1516668200492859,
-0.021786890923976898,
-0.013869802467525005,
-0.10279805213212967,
0.09972582012414932,
-0.07016841322183609,
0.14200840890407562,
-0.03182799369096756,
0.11835620552301407,
0.01566917449235916,
-0.06264495104551315,
0.10259044170379639,
0.016754992306232452,
0.0810830369591713,
-0.0691566988825798,
-0.03272610902786255,
0.036749809980392456,
-0.02753841131925583,
0.00963466614484787,
-0.056368667632341385,
0.024198921397328377,
-0.09628704935312271,
-0.03227876499295235,
-0.09326349198818207,
0.03631990775465965,
-0.026759909465909004,
-0.06683670729398727,
-0.029637739062309265,
0.005259402096271515,
0.03371401131153107,
-0.008436189964413643,
0.09819532185792923,
-0.00553361140191555,
0.16655616462230682,
0.09634551405906677,
0.07441635429859161,
-0.06011874973773956,
-0.048628319054841995,
-0.008710944093763828,
-0.005205092951655388,
0.049955904483795166,
-0.13382312655448914,
0.01785426400601864,
0.15495941042900085,
0.02155223675072193,
0.13910695910453796,
0.09405044466257095,
-0.022557666525244713,
0.02706213854253292,
0.05801645666360855,
-0.18664634227752686,
-0.07835401594638824,
-0.028466248884797096,
-0.09558767080307007,
-0.08849458396434784,
0.06292933225631714,
0.09979983419179916,
-0.06415245682001114,
-0.012962049804627895,
-0.019740043208003044,
-0.006627712398767471,
-0.05614665895700455,
0.21671395003795624,
0.05237327143549919,
0.04719313234090805,
-0.10145954042673111,
0.057343970984220505,
0.04355869069695473,
-0.07908100634813309,
0.017134197056293488,
0.11174745857715607,
-0.08111158758401871,
-0.04192052409052849,
0.0829811841249466,
0.18660344183444977,
-0.07446902990341187,
-0.01892934739589691,
-0.14836329221725464,
-0.11721950769424438,
0.0813269093632698,
0.15817861258983612,
0.10649773478507996,
0.012662569992244244,
-0.07212553173303604,
0.023333333432674408,
-0.150096595287323,
0.07351748645305634,
0.059041090309619904,
0.061878178268671036,
-0.11648673564195633,
0.20154309272766113,
0.007877027615904808,
0.05242728069424629,
-0.03361939266324043,
0.01018581073731184,
-0.1095910295844078,
0.027671154588460922,
-0.10369370132684708,
-0.049022216349840164,
-0.011153383180499077,
-0.011344040744006634,
-0.016820166260004044,
-0.051439233124256134,
-0.04036013409495354,
0.0017442642711102962,
-0.12312141060829163,
-0.018531642854213715,
0.02234707400202751,
0.031234167516231537,
-0.10819482058286667,
-0.02980238012969494,
0.018362974748015404,
-0.05456185340881348,
0.08365610986948013,
0.06481931358575821,
0.003381226444616914,
0.06570839136838913,
-0.14774887263774872,
0.007735785562545061,
0.07149913907051086,
-0.0036824943963438272,
0.054376281797885895,
-0.05427635461091995,
0.00015581227489747107,
-0.002660493366420269,
0.0887991189956665,
0.03405100479722023,
0.06713394820690155,
-0.13523425161838531,
0.019152352586388588,
-0.031213855370879173,
-0.07915440201759338,
-0.07852187007665634,
0.04527357593178749,
0.051753558218479156,
0.02544320374727249,
0.16361363232135773,
-0.08743884414434433,
0.058196183294057846,
-0.21129953861236572,
-0.00010934633610304445,
-0.0040875752456486225,
-0.12665489315986633,
-0.09266360849142075,
-0.0743592381477356,
0.08020300418138504,
-0.048090629279613495,
0.13096480071544647,
0.030766446143388748,
0.045663926750421524,
0.01678226701915264,
-0.022346053272485733,
0.011304822750389576,
0.018130434677004814,
0.21487317979335785,
0.042818449437618256,
-0.04911953955888748,
0.057515960186719894,
0.07763725519180298,
0.10717456042766571,
0.13358579576015472,
0.20776551961898804,
0.11977819353342056,
-0.02090967260301113,
0.08719053119421005,
0.014525890350341797,
-0.03653685748577118,
-0.15769600868225098,
0.029278365895152092,
-0.050307173281908035,
0.09102386981248856,
-0.029601961374282837,
0.20985136926174164,
0.06475508213043213,
-0.14861363172531128,
0.0457957424223423,
-0.04301391914486885,
-0.10353527218103409,
-0.10043830424547195,
-0.03640874847769737,
-0.07819696515798569,
-0.14328382909297943,
0.003700092202052474,
-0.12281376123428345,
0.02616656757891178,
0.10712455213069916,
0.021406006067991257,
-0.03436770290136337,
0.1888507753610611,
0.05286560207605362,
0.005114249885082245,
0.07879464328289032,
-0.0038691398221999407,
-0.009742936119437218,
-0.10882671177387238,
-0.06297849863767624,
-0.021827444434165955,
0.005160908214747906,
0.04896215349435806,
-0.048518210649490356,
-0.08337627351284027,
0.026081306859850883,
-0.03909663110971451,
-0.10220447182655334,
0.005902046337723732,
0.034687936305999756,
0.07156409323215485,
0.04333166033029556,
-0.001027342164888978,
-0.011525237932801247,
-0.024379530921578407,
0.23378123342990875,
-0.07445773482322693,
-0.08377130329608917,
-0.0816580206155777,
0.2721860706806183,
0.041219934821128845,
-0.005152586381882429,
0.019220838323235512,
-0.05799342319369316,
0.0036831984762102365,
0.27988868951797485,
0.2125694453716278,
-0.10016979277133942,
-0.009520173072814941,
0.010027931071817875,
-0.0031181492377072573,
0.004242825787514448,
0.12360801547765732,
0.13543415069580078,
0.07037138193845749,
-0.10718037188053131,
-0.032002780586481094,
-0.05087347328662872,
-0.009832074865698814,
-0.039275381714105606,
0.06516605615615845,
0.062194712460041046,
0.016189811751246452,
-0.051229942589998245,
0.06365080177783966,
-0.09938856214284897,
-0.08075062185525894,
0.01938428357243538,
-0.21821418404579163,
-0.1554875373840332,
-0.005112168844789267,
0.0996389091014862,
-0.005710593890398741,
0.07495123893022537,
-0.02667590230703354,
0.002516604494303465,
0.05160210654139519,
-0.019843513146042824,
-0.08460623770952225,
-0.056447453796863556,
0.08591074496507645,
-0.12586408853530884,
0.15994970500469208,
-0.049919493496418,
0.06286948919296265,
0.11928514391183853,
0.05176283419132233,
-0.05430569499731064,
0.06883741915225983,
0.02686990611255169,
-0.06406867504119873,
0.03090379759669304,
0.11983471363782883,
-0.024801693856716156,
0.021293139085173607,
0.04098745808005333,
-0.13693661987781525,
0.027778826653957367,
-0.07703226059675217,
-0.03497577831149101,
-0.029713600873947144,
-0.06208154186606407,
-0.054803818464279175,
0.12292929738759995,
0.2282559871673584,
-0.018451105803251266,
0.02471134625375271,
-0.0803251564502716,
0.004039809573441744,
0.04267621412873268,
0.04678300395607948,
-0.0781247690320015,
-0.25423663854599,
-0.01581280492246151,
0.08777163922786713,
-0.03649524226784706,
-0.2699001729488373,
-0.08304070681333542,
-0.0024273861199617386,
-0.06822919845581055,
-0.1016962006688118,
0.08843974769115448,
0.07106528431177139,
0.05470959469676018,
-0.045371852815151215,
-0.06084480881690979,
-0.07764372229576111,
0.17095789313316345,
-0.15647734701633453,
-0.08875411003828049
] |
null | null |
transformers
|
# papuGaPT2 - Polish GPT2 language model
[GPT2](https://d4mucfpksywv.cloudfront.net/better-language-models/language_models_are_unsupervised_multitask_learners.pdf) was released in 2019 and surprised many with its text generation capability. However, up until very recently, we have not had a strong text generation model in Polish language, which limited the research opportunities for Polish NLP practitioners. With the release of this model, we hope to enable such research.
Our model follows the standard GPT2 architecture and training approach. We are using a causal language modeling (CLM) objective, which means that the model is trained to predict the next word (token) in a sequence of words (tokens).
## Datasets
We used the Polish subset of the [multilingual Oscar corpus](https://www.aclweb.org/anthology/2020.acl-main.156) to train the model in a self-supervised fashion.
```
from datasets import load_dataset
dataset = load_dataset('oscar', 'unshuffled_deduplicated_pl')
```
## Intended uses & limitations
The raw model can be used for text generation or fine-tuned for a downstream task. The model has been trained on data scraped from the web, and can generate text containing intense violence, sexual situations, coarse language and drug use. It also reflects the biases from the dataset (see below for more details). These limitations are likely to transfer to the fine-tuned models as well. At this stage, we do not recommend using the model beyond research.
## Bias Analysis
There are many sources of bias embedded in the model and we caution to be mindful of this while exploring the capabilities of this model. We have started a very basic analysis of bias that you can see in [this notebook](https://huggingface.co/flax-community/papuGaPT2/blob/main/papuGaPT2_bias_analysis.ipynb).
### Gender Bias
As an example, we generated 50 texts starting with prompts "She/He works as". The image below presents the resulting word clouds of female/male professions. The most salient terms for male professions are: teacher, sales representative, programmer. The most salient terms for female professions are: model, caregiver, receptionist, waitress.

### Ethnicity/Nationality/Gender Bias
We generated 1000 texts to assess bias across ethnicity, nationality and gender vectors. We created prompts with the following scheme:
* Person - in Polish this is a single word that differentiates both nationality/ethnicity and gender. We assessed the following 5 nationalities/ethnicities: German, Romani, Jewish, Ukrainian, Neutral. The neutral group used generic pronounts ("He/She").
* Topic - we used 5 different topics:
* random act: *entered home*
* said: *said*
* works as: *works as*
* intent: Polish *niech* which combined with *he* would roughly translate to *let him ...*
* define: *is*
Each combination of 5 nationalities x 2 genders x 5 topics had 20 generated texts.
We used a model trained on [Polish Hate Speech corpus](https://huggingface.co/datasets/hate_speech_pl) to obtain the probability that each generated text contains hate speech. To avoid leakage, we removed the first word identifying the nationality/ethnicity and gender from the generated text before running the hate speech detector.
The following tables and charts demonstrate the intensity of hate speech associated with the generated texts. There is a very clear effect where each of the ethnicities/nationalities score higher than the neutral baseline.

Looking at the gender dimension we see higher hate score associated with males vs. females.

We don't recommend using the GPT2 model beyond research unless a clear mitigation for the biases is provided.
## Training procedure
### Training scripts
We used the [causal language modeling script for Flax](https://github.com/huggingface/transformers/blob/master/examples/flax/language-modeling/run_clm_flax.py). We would like to thank the authors of that script as it allowed us to complete this training in a very short time!
### Preprocessing and Training Details
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 512 consecutive tokens.
We have trained the model on a single TPUv3 VM, and due to unforeseen events the training run was split in 3 parts, each time resetting from the final checkpoint with a new optimizer state:
1. LR 1e-3, bs 64, linear schedule with warmup for 1000 steps, 10 epochs, stopped after 70,000 steps at eval loss 3.206 and perplexity 24.68
2. LR 3e-4, bs 64, linear schedule with warmup for 5000 steps, 7 epochs, stopped after 77,000 steps at eval loss 3.116 and perplexity 22.55
3. LR 2e-4, bs 64, linear schedule with warmup for 5000 steps, 3 epochs, stopped after 91,000 steps at eval loss 3.082 and perplexity 21.79
## Evaluation results
We trained the model on 95% of the dataset and evaluated both loss and perplexity on 5% of the dataset. The final checkpoint evaluation resulted in:
* Evaluation loss: 3.082
* Perplexity: 21.79
## How to use
You can use the model either directly for text generation (see example below), by extracting features, or for further fine-tuning. We have prepared a notebook with text generation examples [here](https://huggingface.co/flax-community/papuGaPT2/blob/main/papuGaPT2_text_generation.ipynb) including different decoding methods, bad words suppression, few- and zero-shot learning demonstrations.
### Text generation
Let's first start with the text-generation pipeline. When prompting for the best Polish poet, it comes up with a pretty reasonable text, highlighting one of the most famous Polish poets, Adam Mickiewicz.
```python
from transformers import pipeline, set_seed
generator = pipeline('text-generation', model='flax-community/papuGaPT2')
set_seed(42)
generator('Największym polskim poetą był')
>>> [{'generated_text': 'Największym polskim poetą był Adam Mickiewicz - uważany za jednego z dwóch geniuszów języka polskiego. "Pan Tadeusz" był jednym z najpopularniejszych dzieł w historii Polski. W 1801 został wystawiony publicznie w Teatrze Wilama Horzycy. Pod jego'}]
```
The pipeline uses `model.generate()` method in the background. In [our notebook](https://huggingface.co/flax-community/papuGaPT2/blob/main/papuGaPT2_text_generation.ipynb) we demonstrate different decoding methods we can use with this method, including greedy search, beam search, sampling, temperature scaling, top-k and top-p sampling. As an example, the below snippet uses sampling among the 50 most probable tokens at each stage (top-k) and among the tokens that jointly represent 95% of the probability distribution (top-p). It also returns 3 output sequences.
```python
from transformers import AutoTokenizer, AutoModelWithLMHead
model = AutoModelWithLMHead.from_pretrained('flax-community/papuGaPT2')
tokenizer = AutoTokenizer.from_pretrained('flax-community/papuGaPT2')
set_seed(42) # reproducibility
input_ids = tokenizer.encode('Największym polskim poetą był', return_tensors='pt')
sample_outputs = model.generate(
input_ids,
do_sample=True,
max_length=50,
top_k=50,
top_p=0.95,
num_return_sequences=3
)
print("Output:\
" + 100 * '-')
for i, sample_output in enumerate(sample_outputs):
print("{}: {}".format(i, tokenizer.decode(sample_output, skip_special_tokens=True)))
>>> Output:
>>> ----------------------------------------------------------------------------------------------------
>>> 0: Największym polskim poetą był Roman Ingarden. Na jego wiersze i piosenki oddziaływały jego zamiłowanie do przyrody i przyrody. Dlatego też jako poeta w czasie pracy nad utworami i wierszami z tych wierszy, a następnie z poezji własnej - pisał
>>> 1: Największym polskim poetą był Julian Przyboś, którego poematem „Wierszyki dla dzieci”.
>>> W okresie międzywojennym, pod hasłem „Papież i nie tylko” Polska, jak większość krajów europejskich, była państwem faszystowskim.
>>> Prócz
>>> 2: Największym polskim poetą był Bolesław Leśmian, który był jego tłumaczem, a jego poezja tłumaczyła na kilkanaście języków.
>>> W 1895 roku nakładem krakowskiego wydania "Scientio" ukazała się w języku polskim powieść W krainie kangurów
```
### Avoiding Bad Words
You may want to prevent certain words from occurring in the generated text. To avoid displaying really bad words in the notebook, let's pretend that we don't like certain types of music to be advertised by our model. The prompt says: *my favorite type of music is*.
```python
input_ids = tokenizer.encode('Mój ulubiony gatunek muzyki to', return_tensors='pt')
bad_words = [' disco', ' rock', ' pop', ' soul', ' reggae', ' hip-hop']
bad_word_ids = []
for bad_word in bad_words:
ids = tokenizer(bad_word).input_ids
bad_word_ids.append(ids)
sample_outputs = model.generate(
input_ids,
do_sample=True,
max_length=20,
top_k=50,
top_p=0.95,
num_return_sequences=5,
bad_words_ids=bad_word_ids
)
print("Output:\
" + 100 * '-')
for i, sample_output in enumerate(sample_outputs):
print("{}: {}".format(i, tokenizer.decode(sample_output, skip_special_tokens=True)))
>>> Output:
>>> ----------------------------------------------------------------------------------------------------
>>> 0: Mój ulubiony gatunek muzyki to muzyka klasyczna. Nie wiem, czy to kwestia sposobu, w jaki gramy,
>>> 1: Mój ulubiony gatunek muzyki to reggea. Zachwycają mnie piosenki i piosenki muzyczne o ducho
>>> 2: Mój ulubiony gatunek muzyki to rockabilly, ale nie lubię też punka. Moim ulubionym gatunkiem
>>> 3: Mój ulubiony gatunek muzyki to rap, ale to raczej się nie zdarza w miejscach, gdzie nie chodzi
>>> 4: Mój ulubiony gatunek muzyki to metal aranżeje nie mam pojęcia co mam robić. Co roku,
```
Ok, it seems this worked: we can see *classical music, rap, metal* among the outputs. Interestingly, *reggae* found a way through via a misspelling *reggea*. Take it as a caution to be careful with curating your bad word lists!
### Few Shot Learning
Let's see now if our model is able to pick up training signal directly from a prompt, without any finetuning. This approach was made really popular with GPT3, and while our model is definitely less powerful, maybe it can still show some skills! If you'd like to explore this topic in more depth, check out [the following article](https://huggingface.co/blog/few-shot-learning-gpt-neo-and-inference-api) which we used as reference.
```python
prompt = """Tekst: "Nienawidzę smerfów!"
Sentyment: Negatywny
###
Tekst: "Jaki piękny dzień 👍"
Sentyment: Pozytywny
###
Tekst: "Jutro idę do kina"
Sentyment: Neutralny
###
Tekst: "Ten przepis jest świetny!"
Sentyment:"""
res = generator(prompt, max_length=85, temperature=0.5, end_sequence='###', return_full_text=False, num_return_sequences=5,)
for x in res:
print(res[i]['generated_text'].split(' ')[1])
>>> Pozytywny
>>> Pozytywny
>>> Pozytywny
>>> Pozytywny
>>> Pozytywny
```
It looks like our model is able to pick up some signal from the prompt. Be careful though, this capability is definitely not mature and may result in spurious or biased responses.
### Zero-Shot Inference
Large language models are known to store a lot of knowledge in its parameters. In the example below, we can see that our model has learned the date of an important event in Polish history, the battle of Grunwald.
```python
prompt = "Bitwa pod Grunwaldem miała miejsce w roku"
input_ids = tokenizer.encode(prompt, return_tensors='pt')
# activate beam search and early_stopping
beam_outputs = model.generate(
input_ids,
max_length=20,
num_beams=5,
early_stopping=True,
num_return_sequences=3
)
print("Output:\
" + 100 * '-')
for i, sample_output in enumerate(beam_outputs):
print("{}: {}".format(i, tokenizer.decode(sample_output, skip_special_tokens=True)))
>>> Output:
>>> ----------------------------------------------------------------------------------------------------
>>> 0: Bitwa pod Grunwaldem miała miejsce w roku 1410, kiedy to wojska polsko-litewskie pod
>>> 1: Bitwa pod Grunwaldem miała miejsce w roku 1410, kiedy to wojska polsko-litewskie pokona
>>> 2: Bitwa pod Grunwaldem miała miejsce w roku 1410, kiedy to wojska polsko-litewskie,
```
## BibTeX entry and citation info
```bibtex
@misc{papuGaPT2,
title={papuGaPT2 - Polish GPT2 language model},
url={https://huggingface.co/flax-community/papuGaPT2},
author={Wojczulis, Michał and Kłeczek, Dariusz},
year={2021}
}
```
|
{"language": "pl", "tags": ["text-generation"], "widget": [{"text": "Najsmaczniejszy polski owoc to"}]}
|
text-generation
|
dkleczek/papuGaPT2
|
[
"transformers",
"pytorch",
"jax",
"tensorboard",
"gpt2",
"text-generation",
"pl",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"pl"
] |
TAGS
#transformers #pytorch #jax #tensorboard #gpt2 #text-generation #pl #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# papuGaPT2 - Polish GPT2 language model
GPT2 was released in 2019 and surprised many with its text generation capability. However, up until very recently, we have not had a strong text generation model in Polish language, which limited the research opportunities for Polish NLP practitioners. With the release of this model, we hope to enable such research.
Our model follows the standard GPT2 architecture and training approach. We are using a causal language modeling (CLM) objective, which means that the model is trained to predict the next word (token) in a sequence of words (tokens).
## Datasets
We used the Polish subset of the multilingual Oscar corpus to train the model in a self-supervised fashion.
## Intended uses & limitations
The raw model can be used for text generation or fine-tuned for a downstream task. The model has been trained on data scraped from the web, and can generate text containing intense violence, sexual situations, coarse language and drug use. It also reflects the biases from the dataset (see below for more details). These limitations are likely to transfer to the fine-tuned models as well. At this stage, we do not recommend using the model beyond research.
## Bias Analysis
There are many sources of bias embedded in the model and we caution to be mindful of this while exploring the capabilities of this model. We have started a very basic analysis of bias that you can see in this notebook.
### Gender Bias
As an example, we generated 50 texts starting with prompts "She/He works as". The image below presents the resulting word clouds of female/male professions. The most salient terms for male professions are: teacher, sales representative, programmer. The most salient terms for female professions are: model, caregiver, receptionist, waitress.
!gender bias
### Ethnicity/Nationality/Gender Bias
We generated 1000 texts to assess bias across ethnicity, nationality and gender vectors. We created prompts with the following scheme:
* Person - in Polish this is a single word that differentiates both nationality/ethnicity and gender. We assessed the following 5 nationalities/ethnicities: German, Romani, Jewish, Ukrainian, Neutral. The neutral group used generic pronounts ("He/She").
* Topic - we used 5 different topics:
* random act: *entered home*
* said: *said*
* works as: *works as*
* intent: Polish *niech* which combined with *he* would roughly translate to *let him ...*
* define: *is*
Each combination of 5 nationalities x 2 genders x 5 topics had 20 generated texts.
We used a model trained on Polish Hate Speech corpus to obtain the probability that each generated text contains hate speech. To avoid leakage, we removed the first word identifying the nationality/ethnicity and gender from the generated text before running the hate speech detector.
The following tables and charts demonstrate the intensity of hate speech associated with the generated texts. There is a very clear effect where each of the ethnicities/nationalities score higher than the neutral baseline.
!hate score by ethnicity
Looking at the gender dimension we see higher hate score associated with males vs. females.
!hate score by gender
We don't recommend using the GPT2 model beyond research unless a clear mitigation for the biases is provided.
## Training procedure
### Training scripts
We used the causal language modeling script for Flax. We would like to thank the authors of that script as it allowed us to complete this training in a very short time!
### Preprocessing and Training Details
The texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 512 consecutive tokens.
We have trained the model on a single TPUv3 VM, and due to unforeseen events the training run was split in 3 parts, each time resetting from the final checkpoint with a new optimizer state:
1. LR 1e-3, bs 64, linear schedule with warmup for 1000 steps, 10 epochs, stopped after 70,000 steps at eval loss 3.206 and perplexity 24.68
2. LR 3e-4, bs 64, linear schedule with warmup for 5000 steps, 7 epochs, stopped after 77,000 steps at eval loss 3.116 and perplexity 22.55
3. LR 2e-4, bs 64, linear schedule with warmup for 5000 steps, 3 epochs, stopped after 91,000 steps at eval loss 3.082 and perplexity 21.79
## Evaluation results
We trained the model on 95% of the dataset and evaluated both loss and perplexity on 5% of the dataset. The final checkpoint evaluation resulted in:
* Evaluation loss: 3.082
* Perplexity: 21.79
## How to use
You can use the model either directly for text generation (see example below), by extracting features, or for further fine-tuning. We have prepared a notebook with text generation examples here including different decoding methods, bad words suppression, few- and zero-shot learning demonstrations.
### Text generation
Let's first start with the text-generation pipeline. When prompting for the best Polish poet, it comes up with a pretty reasonable text, highlighting one of the most famous Polish poets, Adam Mickiewicz.
The pipeline uses 'model.generate()' method in the background. In our notebook we demonstrate different decoding methods we can use with this method, including greedy search, beam search, sampling, temperature scaling, top-k and top-p sampling. As an example, the below snippet uses sampling among the 50 most probable tokens at each stage (top-k) and among the tokens that jointly represent 95% of the probability distribution (top-p). It also returns 3 output sequences.
### Avoiding Bad Words
You may want to prevent certain words from occurring in the generated text. To avoid displaying really bad words in the notebook, let's pretend that we don't like certain types of music to be advertised by our model. The prompt says: *my favorite type of music is*.
Ok, it seems this worked: we can see *classical music, rap, metal* among the outputs. Interestingly, *reggae* found a way through via a misspelling *reggea*. Take it as a caution to be careful with curating your bad word lists!
### Few Shot Learning
Let's see now if our model is able to pick up training signal directly from a prompt, without any finetuning. This approach was made really popular with GPT3, and while our model is definitely less powerful, maybe it can still show some skills! If you'd like to explore this topic in more depth, check out the following article which we used as reference.
It looks like our model is able to pick up some signal from the prompt. Be careful though, this capability is definitely not mature and may result in spurious or biased responses.
### Zero-Shot Inference
Large language models are known to store a lot of knowledge in its parameters. In the example below, we can see that our model has learned the date of an important event in Polish history, the battle of Grunwald.
## BibTeX entry and citation info
|
[
"# papuGaPT2 - Polish GPT2 language model\nGPT2 was released in 2019 and surprised many with its text generation capability. However, up until very recently, we have not had a strong text generation model in Polish language, which limited the research opportunities for Polish NLP practitioners. With the release of this model, we hope to enable such research. \n\nOur model follows the standard GPT2 architecture and training approach. We are using a causal language modeling (CLM) objective, which means that the model is trained to predict the next word (token) in a sequence of words (tokens).",
"## Datasets\nWe used the Polish subset of the multilingual Oscar corpus to train the model in a self-supervised fashion.",
"## Intended uses & limitations\nThe raw model can be used for text generation or fine-tuned for a downstream task. The model has been trained on data scraped from the web, and can generate text containing intense violence, sexual situations, coarse language and drug use. It also reflects the biases from the dataset (see below for more details). These limitations are likely to transfer to the fine-tuned models as well. At this stage, we do not recommend using the model beyond research.",
"## Bias Analysis\nThere are many sources of bias embedded in the model and we caution to be mindful of this while exploring the capabilities of this model. We have started a very basic analysis of bias that you can see in this notebook.",
"### Gender Bias\nAs an example, we generated 50 texts starting with prompts \"She/He works as\". The image below presents the resulting word clouds of female/male professions. The most salient terms for male professions are: teacher, sales representative, programmer. The most salient terms for female professions are: model, caregiver, receptionist, waitress.\n\n!gender bias",
"### Ethnicity/Nationality/Gender Bias\nWe generated 1000 texts to assess bias across ethnicity, nationality and gender vectors. We created prompts with the following scheme: \n\n* Person - in Polish this is a single word that differentiates both nationality/ethnicity and gender. We assessed the following 5 nationalities/ethnicities: German, Romani, Jewish, Ukrainian, Neutral. The neutral group used generic pronounts (\"He/She\"). \n* Topic - we used 5 different topics: \n * random act: *entered home*\n * said: *said*\n * works as: *works as*\n * intent: Polish *niech* which combined with *he* would roughly translate to *let him ...*\n * define: *is*\n\nEach combination of 5 nationalities x 2 genders x 5 topics had 20 generated texts. \n\nWe used a model trained on Polish Hate Speech corpus to obtain the probability that each generated text contains hate speech. To avoid leakage, we removed the first word identifying the nationality/ethnicity and gender from the generated text before running the hate speech detector.\n \nThe following tables and charts demonstrate the intensity of hate speech associated with the generated texts. There is a very clear effect where each of the ethnicities/nationalities score higher than the neutral baseline. \n\n!hate score by ethnicity\n\nLooking at the gender dimension we see higher hate score associated with males vs. females. \n\n!hate score by gender\n\nWe don't recommend using the GPT2 model beyond research unless a clear mitigation for the biases is provided.",
"## Training procedure",
"### Training scripts\nWe used the causal language modeling script for Flax. We would like to thank the authors of that script as it allowed us to complete this training in a very short time!",
"### Preprocessing and Training Details\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 512 consecutive tokens.\n\nWe have trained the model on a single TPUv3 VM, and due to unforeseen events the training run was split in 3 parts, each time resetting from the final checkpoint with a new optimizer state: \n1. LR 1e-3, bs 64, linear schedule with warmup for 1000 steps, 10 epochs, stopped after 70,000 steps at eval loss 3.206 and perplexity 24.68\n2. LR 3e-4, bs 64, linear schedule with warmup for 5000 steps, 7 epochs, stopped after 77,000 steps at eval loss 3.116 and perplexity 22.55\n3. LR 2e-4, bs 64, linear schedule with warmup for 5000 steps, 3 epochs, stopped after 91,000 steps at eval loss 3.082 and perplexity 21.79",
"## Evaluation results\nWe trained the model on 95% of the dataset and evaluated both loss and perplexity on 5% of the dataset. The final checkpoint evaluation resulted in: \n* Evaluation loss: 3.082\n* Perplexity: 21.79",
"## How to use\nYou can use the model either directly for text generation (see example below), by extracting features, or for further fine-tuning. We have prepared a notebook with text generation examples here including different decoding methods, bad words suppression, few- and zero-shot learning demonstrations.",
"### Text generation\nLet's first start with the text-generation pipeline. When prompting for the best Polish poet, it comes up with a pretty reasonable text, highlighting one of the most famous Polish poets, Adam Mickiewicz.\n \n\n\nThe pipeline uses 'model.generate()' method in the background. In our notebook we demonstrate different decoding methods we can use with this method, including greedy search, beam search, sampling, temperature scaling, top-k and top-p sampling. As an example, the below snippet uses sampling among the 50 most probable tokens at each stage (top-k) and among the tokens that jointly represent 95% of the probability distribution (top-p). It also returns 3 output sequences.",
"### Avoiding Bad Words\nYou may want to prevent certain words from occurring in the generated text. To avoid displaying really bad words in the notebook, let's pretend that we don't like certain types of music to be advertised by our model. The prompt says: *my favorite type of music is*. \n\n\nOk, it seems this worked: we can see *classical music, rap, metal* among the outputs. Interestingly, *reggae* found a way through via a misspelling *reggea*. Take it as a caution to be careful with curating your bad word lists!",
"### Few Shot Learning\n\nLet's see now if our model is able to pick up training signal directly from a prompt, without any finetuning. This approach was made really popular with GPT3, and while our model is definitely less powerful, maybe it can still show some skills! If you'd like to explore this topic in more depth, check out the following article which we used as reference.\n\n\nIt looks like our model is able to pick up some signal from the prompt. Be careful though, this capability is definitely not mature and may result in spurious or biased responses.",
"### Zero-Shot Inference\n\nLarge language models are known to store a lot of knowledge in its parameters. In the example below, we can see that our model has learned the date of an important event in Polish history, the battle of Grunwald.",
"## BibTeX entry and citation info"
] |
[
"TAGS\n#transformers #pytorch #jax #tensorboard #gpt2 #text-generation #pl #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# papuGaPT2 - Polish GPT2 language model\nGPT2 was released in 2019 and surprised many with its text generation capability. However, up until very recently, we have not had a strong text generation model in Polish language, which limited the research opportunities for Polish NLP practitioners. With the release of this model, we hope to enable such research. \n\nOur model follows the standard GPT2 architecture and training approach. We are using a causal language modeling (CLM) objective, which means that the model is trained to predict the next word (token) in a sequence of words (tokens).",
"## Datasets\nWe used the Polish subset of the multilingual Oscar corpus to train the model in a self-supervised fashion.",
"## Intended uses & limitations\nThe raw model can be used for text generation or fine-tuned for a downstream task. The model has been trained on data scraped from the web, and can generate text containing intense violence, sexual situations, coarse language and drug use. It also reflects the biases from the dataset (see below for more details). These limitations are likely to transfer to the fine-tuned models as well. At this stage, we do not recommend using the model beyond research.",
"## Bias Analysis\nThere are many sources of bias embedded in the model and we caution to be mindful of this while exploring the capabilities of this model. We have started a very basic analysis of bias that you can see in this notebook.",
"### Gender Bias\nAs an example, we generated 50 texts starting with prompts \"She/He works as\". The image below presents the resulting word clouds of female/male professions. The most salient terms for male professions are: teacher, sales representative, programmer. The most salient terms for female professions are: model, caregiver, receptionist, waitress.\n\n!gender bias",
"### Ethnicity/Nationality/Gender Bias\nWe generated 1000 texts to assess bias across ethnicity, nationality and gender vectors. We created prompts with the following scheme: \n\n* Person - in Polish this is a single word that differentiates both nationality/ethnicity and gender. We assessed the following 5 nationalities/ethnicities: German, Romani, Jewish, Ukrainian, Neutral. The neutral group used generic pronounts (\"He/She\"). \n* Topic - we used 5 different topics: \n * random act: *entered home*\n * said: *said*\n * works as: *works as*\n * intent: Polish *niech* which combined with *he* would roughly translate to *let him ...*\n * define: *is*\n\nEach combination of 5 nationalities x 2 genders x 5 topics had 20 generated texts. \n\nWe used a model trained on Polish Hate Speech corpus to obtain the probability that each generated text contains hate speech. To avoid leakage, we removed the first word identifying the nationality/ethnicity and gender from the generated text before running the hate speech detector.\n \nThe following tables and charts demonstrate the intensity of hate speech associated with the generated texts. There is a very clear effect where each of the ethnicities/nationalities score higher than the neutral baseline. \n\n!hate score by ethnicity\n\nLooking at the gender dimension we see higher hate score associated with males vs. females. \n\n!hate score by gender\n\nWe don't recommend using the GPT2 model beyond research unless a clear mitigation for the biases is provided.",
"## Training procedure",
"### Training scripts\nWe used the causal language modeling script for Flax. We would like to thank the authors of that script as it allowed us to complete this training in a very short time!",
"### Preprocessing and Training Details\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 512 consecutive tokens.\n\nWe have trained the model on a single TPUv3 VM, and due to unforeseen events the training run was split in 3 parts, each time resetting from the final checkpoint with a new optimizer state: \n1. LR 1e-3, bs 64, linear schedule with warmup for 1000 steps, 10 epochs, stopped after 70,000 steps at eval loss 3.206 and perplexity 24.68\n2. LR 3e-4, bs 64, linear schedule with warmup for 5000 steps, 7 epochs, stopped after 77,000 steps at eval loss 3.116 and perplexity 22.55\n3. LR 2e-4, bs 64, linear schedule with warmup for 5000 steps, 3 epochs, stopped after 91,000 steps at eval loss 3.082 and perplexity 21.79",
"## Evaluation results\nWe trained the model on 95% of the dataset and evaluated both loss and perplexity on 5% of the dataset. The final checkpoint evaluation resulted in: \n* Evaluation loss: 3.082\n* Perplexity: 21.79",
"## How to use\nYou can use the model either directly for text generation (see example below), by extracting features, or for further fine-tuning. We have prepared a notebook with text generation examples here including different decoding methods, bad words suppression, few- and zero-shot learning demonstrations.",
"### Text generation\nLet's first start with the text-generation pipeline. When prompting for the best Polish poet, it comes up with a pretty reasonable text, highlighting one of the most famous Polish poets, Adam Mickiewicz.\n \n\n\nThe pipeline uses 'model.generate()' method in the background. In our notebook we demonstrate different decoding methods we can use with this method, including greedy search, beam search, sampling, temperature scaling, top-k and top-p sampling. As an example, the below snippet uses sampling among the 50 most probable tokens at each stage (top-k) and among the tokens that jointly represent 95% of the probability distribution (top-p). It also returns 3 output sequences.",
"### Avoiding Bad Words\nYou may want to prevent certain words from occurring in the generated text. To avoid displaying really bad words in the notebook, let's pretend that we don't like certain types of music to be advertised by our model. The prompt says: *my favorite type of music is*. \n\n\nOk, it seems this worked: we can see *classical music, rap, metal* among the outputs. Interestingly, *reggae* found a way through via a misspelling *reggea*. Take it as a caution to be careful with curating your bad word lists!",
"### Few Shot Learning\n\nLet's see now if our model is able to pick up training signal directly from a prompt, without any finetuning. This approach was made really popular with GPT3, and while our model is definitely less powerful, maybe it can still show some skills! If you'd like to explore this topic in more depth, check out the following article which we used as reference.\n\n\nIt looks like our model is able to pick up some signal from the prompt. Be careful though, this capability is definitely not mature and may result in spurious or biased responses.",
"### Zero-Shot Inference\n\nLarge language models are known to store a lot of knowledge in its parameters. In the example below, we can see that our model has learned the date of an important event in Polish history, the battle of Grunwald.",
"## BibTeX entry and citation info"
] |
[
56,
139,
31,
112,
56,
92,
367,
3,
43,
238,
54,
65,
177,
136,
126,
56,
10
] |
[
"passage: TAGS\n#transformers #pytorch #jax #tensorboard #gpt2 #text-generation #pl #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# papuGaPT2 - Polish GPT2 language model\nGPT2 was released in 2019 and surprised many with its text generation capability. However, up until very recently, we have not had a strong text generation model in Polish language, which limited the research opportunities for Polish NLP practitioners. With the release of this model, we hope to enable such research. \n\nOur model follows the standard GPT2 architecture and training approach. We are using a causal language modeling (CLM) objective, which means that the model is trained to predict the next word (token) in a sequence of words (tokens).## Datasets\nWe used the Polish subset of the multilingual Oscar corpus to train the model in a self-supervised fashion.## Intended uses & limitations\nThe raw model can be used for text generation or fine-tuned for a downstream task. The model has been trained on data scraped from the web, and can generate text containing intense violence, sexual situations, coarse language and drug use. It also reflects the biases from the dataset (see below for more details). These limitations are likely to transfer to the fine-tuned models as well. At this stage, we do not recommend using the model beyond research.## Bias Analysis\nThere are many sources of bias embedded in the model and we caution to be mindful of this while exploring the capabilities of this model. We have started a very basic analysis of bias that you can see in this notebook.### Gender Bias\nAs an example, we generated 50 texts starting with prompts \"She/He works as\". The image below presents the resulting word clouds of female/male professions. The most salient terms for male professions are: teacher, sales representative, programmer. The most salient terms for female professions are: model, caregiver, receptionist, waitress.\n\n!gender bias",
"passage: ### Ethnicity/Nationality/Gender Bias\nWe generated 1000 texts to assess bias across ethnicity, nationality and gender vectors. We created prompts with the following scheme: \n\n* Person - in Polish this is a single word that differentiates both nationality/ethnicity and gender. We assessed the following 5 nationalities/ethnicities: German, Romani, Jewish, Ukrainian, Neutral. The neutral group used generic pronounts (\"He/She\"). \n* Topic - we used 5 different topics: \n * random act: *entered home*\n * said: *said*\n * works as: *works as*\n * intent: Polish *niech* which combined with *he* would roughly translate to *let him ...*\n * define: *is*\n\nEach combination of 5 nationalities x 2 genders x 5 topics had 20 generated texts. \n\nWe used a model trained on Polish Hate Speech corpus to obtain the probability that each generated text contains hate speech. To avoid leakage, we removed the first word identifying the nationality/ethnicity and gender from the generated text before running the hate speech detector.\n \nThe following tables and charts demonstrate the intensity of hate speech associated with the generated texts. There is a very clear effect where each of the ethnicities/nationalities score higher than the neutral baseline. \n\n!hate score by ethnicity\n\nLooking at the gender dimension we see higher hate score associated with males vs. females. \n\n!hate score by gender\n\nWe don't recommend using the GPT2 model beyond research unless a clear mitigation for the biases is provided.## Training procedure### Training scripts\nWe used the causal language modeling script for Flax. We would like to thank the authors of that script as it allowed us to complete this training in a very short time!### Preprocessing and Training Details\nThe texts are tokenized using a byte-level version of Byte Pair Encoding (BPE) (for unicode characters) and a vocabulary size of 50,257. The inputs are sequences of 512 consecutive tokens.\n\nWe have trained the model on a single TPUv3 VM, and due to unforeseen events the training run was split in 3 parts, each time resetting from the final checkpoint with a new optimizer state: \n1. LR 1e-3, bs 64, linear schedule with warmup for 1000 steps, 10 epochs, stopped after 70,000 steps at eval loss 3.206 and perplexity 24.68\n2. LR 3e-4, bs 64, linear schedule with warmup for 5000 steps, 7 epochs, stopped after 77,000 steps at eval loss 3.116 and perplexity 22.55\n3. LR 2e-4, bs 64, linear schedule with warmup for 5000 steps, 3 epochs, stopped after 91,000 steps at eval loss 3.082 and perplexity 21.79## Evaluation results\nWe trained the model on 95% of the dataset and evaluated both loss and perplexity on 5% of the dataset. The final checkpoint evaluation resulted in: \n* Evaluation loss: 3.082\n* Perplexity: 21.79## How to use\nYou can use the model either directly for text generation (see example below), by extracting features, or for further fine-tuning. We have prepared a notebook with text generation examples here including different decoding methods, bad words suppression, few- and zero-shot learning demonstrations."
] |
[
-0.043703846633434296,
0.08110886812210083,
-0.004320123698562384,
0.03048545867204666,
0.09494422376155853,
-0.011698807589709759,
0.12131600081920624,
0.04922318458557129,
0.05196457356214523,
0.056183479726314545,
0.012787885963916779,
-0.027289969846606255,
0.06870420277118683,
0.08315939456224442,
0.1024385392665863,
-0.21541213989257812,
0.07318520545959473,
-0.08248644322156906,
0.01500428095459938,
0.07379712164402008,
0.1162431538105011,
-0.05786561220884323,
0.0627020001411438,
-0.0015398990362882614,
-0.04712842032313347,
0.007614552974700928,
-0.009362827986478806,
0.009059433825314045,
0.04619063809514046,
0.11611360311508179,
0.09751713275909424,
-0.009121641516685486,
-0.005347641184926033,
-0.15864121913909912,
0.0253908708691597,
0.08115408569574356,
-0.01075183879584074,
0.013215835206210613,
0.1065215989947319,
-0.05993329733610153,
0.19032686948776245,
-0.12295451760292053,
0.03544000908732414,
0.049765072762966156,
-0.1092802956700325,
-0.08012530952692032,
-0.08771008253097534,
0.10149869322776794,
0.09330558776855469,
0.03347732871770859,
-0.06546872109174728,
0.124644935131073,
0.0029432512819767,
0.0633905827999115,
0.07424991577863693,
-0.05731838196516037,
-0.016764571890234947,
-0.11735732853412628,
-0.027370238676667213,
0.11763311177492142,
-0.07781053334474564,
0.0043296730145812035,
0.000011604279279708862,
-0.01605280488729477,
0.0012920091394335032,
-0.01681063324213028,
0.08620146661996841,
-0.057046495378017426,
-0.14762341976165771,
-0.06902939081192017,
0.09492281079292297,
0.04584420472383499,
-0.09610029309988022,
-0.11007807403802872,
-0.003010291839018464,
-0.027215708047151566,
-0.002173994667828083,
-0.008611625991761684,
-0.04290200397372246,
-0.02160114049911499,
0.0946454405784607,
-0.050559863448143005,
-0.09001818299293518,
0.029870301485061646,
0.03385333716869354,
0.20346343517303467,
0.02211913652718067,
0.04238803684711456,
-0.011425759643316269,
0.013446779921650887,
0.012501154094934464,
-0.02955619804561138,
-0.017679112032055855,
-0.05432981997728348,
-0.1207793653011322,
-0.029043642804026604,
-0.02371813729405403,
-0.1022433489561081,
-0.030609209090471268,
0.018074888736009598,
-0.03376544266939163,
0.04407931864261627,
-0.0075912317261099815,
0.04747302085161209,
0.09959693998098373,
0.09685557335615158,
-0.0004753503599204123,
-0.05295278877019882,
-0.009197459556162357,
-0.05498560518026352,
0.030469978228211403,
-0.02031991071999073,
-0.038310594856739044,
-0.03039858490228653,
0.024336958304047585,
0.10124800354242325,
0.011493214406073093,
0.08133333921432495,
-0.0008586333133280277,
-0.04630789905786514,
0.07007494568824768,
-0.10129458457231522,
0.034297794103622437,
0.028114166110754013,
0.019425593316555023,
0.09246950596570969,
-0.004074095748364925,
-0.04175415635108948,
-0.11420268565416336,
0.014977717772126198,
-0.04868530109524727,
0.031499773263931274,
-0.11229818314313889,
-0.1344664990901947,
0.0362582765519619,
-0.029472187161445618,
-0.06596209853887558,
-0.12678559124469757,
-0.10589732974767685,
-0.08810566365718842,
-0.051780566573143005,
-0.02599972113966942,
0.021530229598283768,
-0.0702437236905098,
0.013754086568951607,
-0.014270886778831482,
0.034234799444675446,
-0.06590299308300018,
-0.011904716491699219,
0.024296334013342857,
-0.1599109023809433,
0.050509121268987656,
-0.015518310479819775,
0.005948659032583237,
-0.08755295723676682,
0.011345012113451958,
-0.173506498336792,
0.10682741552591324,
-0.04609207808971405,
0.04390367120504379,
-0.0694042444229126,
-0.046528562903404236,
-0.05627556890249252,
0.04912315681576729,
-0.027024611830711365,
0.16043582558631897,
-0.21724353730678558,
-0.04667232558131218,
0.13914203643798828,
-0.20845383405685425,
0.05059478431940079,
0.11960691213607788,
-0.016107093542814255,
0.08338604867458344,
0.12143844366073608,
0.16499082744121552,
0.05184773728251457,
0.028449775651097298,
-0.00810042954981327,
-0.05606996268033981,
-0.10631643235683441,
0.15918421745300293,
0.025299880653619766,
-0.09337307512760162,
0.005992702208459377,
-0.023891201242804527,
0.024556009098887444,
0.010864241048693657,
-0.009742263704538345,
-0.05885957553982735,
0.06004990637302399,
0.003364724572747946,
-0.004444616846740246,
-0.03412564843893051,
-0.06096576899290085,
-0.03867781162261963,
-0.08817215263843536,
0.008996503427624702,
0.11928094923496246,
-0.06871823221445084,
0.06038855016231537,
-0.10294251888990402,
0.02697494626045227,
0.03757036104798317,
-0.009886449202895164,
-0.1334751844406128,
-0.07319366931915283,
-0.0012776721268892288,
-0.1045655608177185,
0.08649803698062897,
0.020374106243252754,
0.03347852826118469,
0.05555134266614914,
0.005729401484131813,
0.031418222934007645,
-0.0006899368017911911,
0.009824088774621487,
-0.07327274978160858,
-0.16343677043914795,
-0.022657692432403564,
-0.02194800227880478,
0.12775135040283203,
-0.15424609184265137,
-0.010521641001105309,
0.08153949677944183,
0.04864736646413803,
0.041661061346530914,
-0.07525138556957245,
0.025812460109591484,
0.08629998564720154,
-0.006558036431670189,
-0.03947056084871292,
0.038689564913511276,
-0.017097463831305504,
-0.028194423764944077,
0.10944122821092606,
-0.19469845294952393,
-0.20960676670074463,
0.05577421188354492,
-0.040927693247795105,
-0.12154760956764221,
-0.021370528265833855,
-0.01698451302945614,
0.010027654469013214,
0.010522818192839622,
-0.07377464324235916,
0.05434497445821762,
0.008157686330378056,
0.01476906705647707,
-0.019674891605973244,
-0.013836227357387543,
-0.0004670051857829094,
-0.019649336114525795,
-0.05642356351017952,
0.06789328902959824,
0.05348926782608032,
-0.21976129710674286,
0.10273309051990509,
-0.006949612870812416,
0.02366930991411209,
0.2264425903558731,
0.013132026419043541,
-0.09893178939819336,
0.005068420432507992,
0.014844726771116257,
-0.0013055996969342232,
0.0696432963013649,
-0.10326121747493744,
-0.02231082320213318,
0.0050275735557079315,
-0.012353030033409595,
0.020238060504198074,
-0.04491607844829559,
0.023085568100214005,
0.02790287509560585,
-0.010944301262497902,
-0.08837972581386566,
-0.011779006570577621,
-0.009397917427122593,
0.09666126221418381,
-0.00045258132740855217,
0.04929511994123459,
-0.03779570013284683,
-0.05652415007352829,
-0.15576624870300293,
0.09531159698963165,
-0.08623082935810089,
-0.171195387840271,
-0.10734237730503082,
0.05014215409755707,
0.0024188850075006485,
0.021219627931714058,
0.013202077709138393,
-0.11445842683315277,
-0.0697578638792038,
-0.09388367831707001,
0.12307965755462646,
-0.008851217105984688,
-0.01534635666757822,
-0.09062948822975159,
0.03048034757375717,
-0.025953764095902443,
-0.07378841936588287,
0.0011863866820931435,
-0.011934173293411732,
-0.017524223774671555,
0.034136611968278885,
-0.02747921459376812,
0.07397828996181488,
0.10214035958051682,
0.019009413197636604,
-0.020587943494319916,
-0.04112707078456879,
0.20476223528385162,
-0.13651371002197266,
0.03752051293849945,
0.0251950453966856,
-0.13731369376182556,
0.03649622201919556,
0.07491970807313919,
-0.02306150645017624,
-0.024450013414025307,
0.013528016395866871,
0.060830358415842056,
-0.029268329963088036,
-0.13638608157634735,
-0.1210722103714943,
-0.04230330139398575,
-0.025979885831475258,
0.02466154843568802,
0.06058945879340172,
0.011688232421875,
0.016108741983771324,
-0.12012144923210144,
-0.04586353525519371,
0.04121929407119751,
0.0745311975479126,
0.010007747448980808,
0.008621415123343468,
0.027533767744898796,
-0.0954948365688324,
-0.02936876006424427,
0.07369054853916168,
-0.0864989161491394,
0.27218908071517944,
0.018539920449256897,
0.13199079036712646,
0.1021820455789566,
0.040830135345458984,
0.07666418701410294,
0.06822078675031662,
0.04611828550696373,
0.03166041523218155,
-0.032264839857816696,
-0.07872248440980911,
0.006074657663702965,
0.06550739705562592,
0.027216587215662003,
-0.08291491866111755,
-0.00312979007139802,
-0.0962001383304596,
0.11309510469436646,
0.1846887767314911,
-0.001285666599869728,
-0.10661232471466064,
-0.03728170692920685,
0.07039231061935425,
-0.07742323726415634,
-0.04135001078248024,
-0.06265676766633987,
0.0771736204624176,
-0.1448582261800766,
0.07488753646612167,
0.034256357699632645,
0.053604595363140106,
-0.1083282083272934,
0.00788462720811367,
-0.030613213777542114,
0.011337925679981709,
-0.05222685635089874,
0.09664299339056015,
-0.15805119276046753,
0.16646525263786316,
0.017202217131853104,
0.08311092853546143,
-0.08534634113311768,
-0.06650091707706451,
0.029568858444690704,
0.0290397759526968,
0.1314408779144287,
0.06827598810195923,
-0.11875079572200775,
-0.09263534843921661,
-0.027209119871258736,
0.023554416373372078,
0.09664035588502884,
-0.021974727511405945,
0.10021857917308807,
-0.010147842578589916,
0.013881865888834,
-0.04223651438951492,
-0.0005490314215421677,
-0.0998508632183075,
-0.1429334431886673,
0.024426590651273727,
-0.11089394986629486,
0.021481433883309364,
0.0008251641411334276,
-0.038266025483608246,
0.019600262865424156,
0.12454282492399216,
-0.15136681497097015,
-0.10702527314424515,
-0.09521076083183289,
-0.013979501090943813,
0.060265496373176575,
-0.08071576803922653,
0.003898800350725651,
0.07924391329288483,
0.1688145101070404,
-0.03333457559347153,
-0.05370891094207764,
0.01419980637729168,
-0.03544013202190399,
-0.11139164119958878,
-0.015397580340504646,
0.08387807756662369,
0.12979018688201904,
0.06867140531539917,
0.012341377325356007,
0.036548007279634476,
0.045052964240312576,
-0.11959174275398254,
0.00771815050393343,
0.14723017811775208,
-0.11586940288543701,
0.08244164288043976,
-0.03354376554489136,
-0.06939589232206345,
-0.16110149025917053,
-0.09971596300601959,
0.12482845038175583,
0.21726353466510773,
-0.047019582241773605,
0.1006845235824585,
0.09652765840291977,
-0.12088552117347717,
-0.22478249669075012,
0.01979278400540352,
0.024687252938747406,
0.004637602251023054,
0.061933696269989014,
-0.17755606770515442,
-0.07978878915309906,
0.03774534910917282,
0.04298004135489464,
0.024348057806491852,
-0.11117599159479141,
-0.10885591804981232,
0.10609830915927887,
0.05471402779221535,
0.0868791714310646,
-0.05674975365400314,
-0.018483523279428482,
-0.06457610428333282,
0.06692231446504593,
0.1649525910615921,
-0.07048321515321732,
0.016982119530439377,
0.04934948310256004,
0.050701040774583817,
0.05203808844089508,
0.008007253520190716,
0.1294821798801422,
0.03376635164022446,
0.06777805089950562,
-0.10442457348108292,
-0.003742864355444908,
-0.0127178356051445,
0.0064941104501485825,
0.06249253824353218,
0.036403901875019073,
-0.07955729961395264,
-0.05617548152804375,
-0.06752390414476395,
-0.08844846487045288,
0.028744785115122795,
-0.06132669001817703,
-0.014086197130382061,
-0.06463873386383057,
0.059135377407073975,
0.05189507454633713,
-0.022762693464756012,
-0.04025508090853691,
-0.09333085268735886,
-0.011118445545434952,
-0.0061739105731248856,
0.18532118201255798,
0.09995166212320328,
-0.07092778384685516,
0.03317587822675705,
-0.008631972596049309,
0.085755854845047,
-0.002254713326692581,
0.01912517100572586,
0.021563146263360977,
-0.02491813898086548,
0.1339760720729828,
-0.0020275190472602844,
-0.1659298837184906,
0.01317107118666172,
0.08271808177232742,
-0.09856647253036499,
-0.17374417185783386,
-0.01062911469489336,
-0.06677205860614777,
-0.045511212199926376,
-0.1116243302822113,
0.10154222697019577,
-0.05083278566598892,
-0.007191199343651533,
-0.04205708205699921,
0.0744641125202179,
-0.037083204835653305,
0.06367310136556625,
0.045873649418354034,
0.024888362735509872,
-0.0364479124546051,
0.07855764776468277,
0.048347875475883484,
-0.1041218712925911,
0.07171086221933365,
0.18939507007598877,
-0.09886154532432556,
-0.06621310114860535,
-0.06889142096042633,
0.10549198091030121,
-0.0857996940612793,
-0.08415402472019196,
0.030587786808609962,
-0.08212010562419891,
-0.033102378249168396,
0.1334528625011444,
-0.022792192175984383,
0.1049337387084961,
0.006901740562170744,
-0.008877073414623737,
-0.054171010851860046,
0.01854359544813633,
0.04192878305912018,
-0.04609876126050949,
0.01609196700155735,
0.19197584688663483,
0.024103038012981415,
0.008917379193007946,
-0.019534282386302948,
-0.05599864572286606,
-0.07528268545866013,
0.0004392620176076889,
-0.021715616807341576,
-0.04476979374885559,
-0.07286696135997772,
-0.03000224567949772,
0.017609553411602974,
0.00783462543040514,
0.012128408998250961,
0.0019147079437971115,
-0.06384631246328354,
0.0017141783609986305,
-0.02785537578165531,
0.03859655559062958,
-0.06838623434305191,
0.030277501791715622,
0.0754287987947464,
-0.046090275049209595,
0.07142061740159988,
0.04915186017751694,
-0.03307579830288887,
0.058339837938547134,
-0.09126468002796173,
0.061458054929971695,
-0.045159146189689636,
-0.020000282675027847,
-0.01871783658862114,
-0.09295807778835297,
-0.024467982351779938,
-0.020839616656303406,
0.01903102546930313,
0.01636916771531105,
0.06986551731824875,
-0.01932893693447113,
0.12680546939373016,
0.07413798570632935,
-0.06431984901428223,
-0.0650988519191742,
0.05679216980934143,
-0.008000800386071205,
0.010698433965444565,
0.11305718123912811,
-0.04201432690024376,
0.022821912541985512,
-0.08454671502113342,
0.022275907918810844,
0.027720503509044647,
0.055501848459243774,
-0.0027873495128005743,
-0.04351171851158142,
0.054218947887420654,
-0.00435543991625309,
0.11762000620365143,
0.04083869606256485,
-0.04421957582235336,
0.06473174691200256,
-0.008829392492771149,
-0.08901749551296234,
0.021350111812353134,
0.002293679863214493,
-0.022837508469820023,
-0.021717386320233345,
-0.08209840953350067,
-0.04735009744763374,
-0.06016729027032852,
0.02330237254500389,
0.13618695735931396,
0.08213087171316147,
0.15710145235061646,
0.052077069878578186,
-0.017200322821736336,
-0.03711063042283058,
-0.03280837461352348,
-0.03967444598674774,
0.04946359246969223,
-0.012139948084950447,
-0.09742511063814163,
0.09130887687206268,
0.17025184631347656,
-0.093373604118824,
0.10634453594684601,
-0.0029413094744086266,
-0.08558142930269241,
-0.09780357033014297,
-0.27275222539901733,
-0.005160459317266941,
-0.013254043646156788,
-0.004746858961880207,
-0.07088778913021088,
0.0643213614821434,
0.038774486631155014,
0.03338724002242088,
-0.047761015594005585,
0.036334067583084106,
-0.07806640863418579,
-0.1290837973356247,
0.04218696802854538,
0.0001436916645616293,
0.06918594241142273,
0.02955712005496025,
0.05900590121746063,
0.017434943467378616,
0.0006537716835737228,
0.05926328897476196,
0.09514893591403961,
0.02077058143913746,
0.05382133275270462,
-0.08765159547328949,
-0.08471409976482391,
0.010006936267018318,
-0.0011744345538318157,
0.06911572813987732,
0.24892430007457733,
0.05845668539404869,
-0.013359833508729935,
-0.012550881132483482,
0.1201838031411171,
0.046425819396972656,
-0.04961683228611946,
-0.09562806785106659,
0.14885477721691132,
0.017331603914499283,
0.015158158726990223,
-0.030715396627783775,
-0.08570965379476547,
0.06327028572559357,
0.14398524165153503,
0.14594846963882446,
0.03132219612598419,
0.04249895364046097,
-0.09202951192855835,
0.0241350457072258,
0.021799206733703613,
0.10426503419876099,
-0.04559400677680969,
0.20867963135242462,
-0.066401407122612,
0.16959646344184875,
-0.0438164658844471,
0.018056780099868774,
-0.02946927584707737,
0.10914464294910431,
-0.03558092936873436,
0.015917742624878883,
-0.0686364471912384,
0.11816470324993134,
-0.09212598204612732,
-0.19752231240272522,
-0.009340898133814335,
-0.0222167931497097,
-0.05786944180727005,
0.047097645699977875,
-0.05712805688381195,
0.07614298164844513,
0.09063155204057693,
0.007300886325538158,
-0.04331407696008682,
0.13652409613132477,
0.01700533553957939,
-0.052680231630802155,
-0.11790138483047485,
0.11038868129253387,
-0.026612823829054832,
0.13567204773426056,
0.02045121043920517,
0.08612316846847534,
0.049821630120277405,
0.018670540302991867,
-0.10405619442462921,
0.04413823038339615,
-0.0335775762796402,
0.025699453428387642,
0.013855572789907455,
0.18975171446800232,
-0.01924322545528412,
0.05406944453716278,
0.05867654085159302,
0.02289752848446369,
0.07765696197748184,
-0.04445230960845947,
-0.04763126000761986,
0.01878468506038189,
0.09777099639177322,
-0.12172301113605499,
0.12099117040634155,
0.14850082993507385,
-0.013466255739331245,
0.007816637866199017,
-0.06748060137033463,
-0.06388162076473236,
-0.01782246120274067,
0.03998098522424698,
-0.014348452910780907,
-0.1051555722951889,
-0.007127732038497925,
-0.006288532167673111,
0.06467360258102417,
-0.1387992799282074,
-0.00673518655821681,
-0.029554031789302826,
-0.015040837228298187,
0.005143836140632629,
0.08786292374134064,
0.012274719774723053,
-0.00363153126090765,
-0.03921676427125931,
-0.08806450664997101,
0.05988018214702606,
0.07870282977819443,
-0.03149237856268883,
-0.03898896276950836
] |
null | null |
transformers
|
# A certain person's AI
|
{"tags": ["conversational"]}
|
text-generation
|
dkminer81/Tromm
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# A certain person's AI
|
[
"# A certain person's AI"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# A certain person's AI"
] |
[
51,
7
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# A certain person's AI"
] |
[
-0.035198114812374115,
0.0469304658472538,
-0.00446815462782979,
0.007163135334849358,
0.17109315097332,
-0.004011659882962704,
0.18530695140361786,
0.12693984806537628,
0.1214396134018898,
0.007450040429830551,
0.13982674479484558,
0.16262395679950714,
0.036871179938316345,
0.07608704268932343,
-0.048611585050821304,
-0.2518514394760132,
0.10394511371850967,
0.058173663914203644,
0.11299961805343628,
0.1243879422545433,
0.0665886253118515,
-0.06705133616924286,
0.11630252003669739,
0.009779147803783417,
-0.15530124306678772,
-0.05608744919300079,
0.055698834359645844,
-0.11193270236253738,
0.1284058690071106,
0.08181609213352203,
0.028761418536305428,
0.0007674554944969714,
-0.0012652302393689752,
-0.15126775205135345,
0.014776096679270267,
-0.003506868612021208,
-0.016508186236023903,
0.04922477528452873,
0.03462466970086098,
-0.11058399081230164,
0.15620434284210205,
0.143255352973938,
-0.018826190382242203,
0.06765550374984741,
-0.1547175496816635,
0.03943731635808945,
0.04404149577021599,
0.01463172398507595,
0.039100732654333115,
0.1489238142967224,
-0.05056988075375557,
0.15835002064704895,
-0.057120680809020996,
0.10405954718589783,
0.12318249046802521,
-0.36430519819259644,
-0.005104995798319578,
0.040248043835163116,
0.044505078345537186,
0.0128471739590168,
-0.05906207114458084,
0.09489394724369049,
0.03861880674958229,
0.035065505653619766,
-0.007654786575585604,
-0.04465757682919502,
-0.0801616832613945,
0.02677430585026741,
-0.05203666165471077,
0.010147417895495892,
0.2663744390010834,
-0.0028728439938277006,
0.06282828748226166,
-0.07153231650590897,
-0.048933640122413635,
-0.05836174637079239,
0.002716550836339593,
-0.050865866243839264,
-0.10533720254898071,
0.04553958773612976,
0.04135221615433693,
-0.050225380808115005,
-0.12055492401123047,
-0.03739005699753761,
-0.23004436492919922,
0.21681633591651917,
0.0008893411722965539,
0.03486240282654762,
-0.1817416101694107,
0.054945558309555054,
0.09201400727033615,
-0.04277244210243225,
0.03500454127788544,
-0.07959670573472977,
-0.05371518060564995,
0.026231292635202408,
-0.09420671314001083,
-0.05119934305548668,
0.04365247115492821,
0.06117931380867958,
0.020936381071805954,
-0.0038051637820899487,
0.0015350664034485817,
0.07459457218647003,
0.09812962263822556,
0.11747066676616669,
-0.056703921407461166,
-0.00828268937766552,
-0.009511236101388931,
-0.10105898231267929,
-0.01120705809444189,
-0.09938152134418488,
-0.1698063164949417,
0.015638884156942368,
0.06430923938751221,
0.028943829238414764,
-0.0204752329736948,
0.16094551980495453,
-0.08216794580221176,
-0.04515322670340538,
-0.050318509340286255,
-0.018857957795262337,
-0.02065269835293293,
0.03155910223722458,
-0.057525720447301865,
0.08747141063213348,
-0.05182382091879845,
0.04976780340075493,
-0.12758222222328186,
-0.018895890563726425,
-0.05757734179496765,
0.00867556780576706,
-0.042316898703575134,
-0.036554981023073196,
0.007658772636204958,
-0.06091868504881859,
0.04097000136971474,
-0.19543084502220154,
-0.11969716101884842,
-0.009913885034620762,
-0.010011754930019379,
-0.07999030500650406,
-0.060222525149583817,
-0.09181389212608337,
0.019996875897049904,
0.014989025890827179,
-0.10039038211107254,
-0.10028917342424393,
-0.06118280068039894,
0.09285762906074524,
-0.02603219822049141,
0.11090680956840515,
-0.10578703135251999,
0.08391888439655304,
-0.09397562593221664,
-0.04542488604784012,
0.005973697640001774,
0.10109538584947586,
0.034269724041223526,
0.1128109022974968,
0.006389313377439976,
0.01827193610370159,
-0.066162109375,
0.07978504151105881,
-0.06307356059551239,
0.17402668297290802,
-0.08133305609226227,
-0.08200863748788834,
0.20719937980175018,
-0.02006005123257637,
-0.16623418033123016,
0.15689876675605774,
0.012580187991261482,
0.06864628940820694,
0.15483759343624115,
0.2551543414592743,
-0.11580507457256317,
0.04152026027441025,
0.022476185113191605,
0.06585537642240524,
-0.1283542662858963,
0.019892724230885506,
0.032489605247974396,
0.05231532081961632,
-0.0969710648059845,
0.02185097523033619,
0.10473942011594772,
0.06651466339826584,
-0.08566320687532425,
-0.05916200950741768,
0.05220145732164383,
-0.05299369618296623,
0.06648682802915573,
0.016429344192147255,
0.10566097497940063,
-0.028911136090755463,
-0.018433401361107826,
-0.13978224992752075,
0.0033492273651063442,
-0.05350077524781227,
0.010390050709247589,
-0.14685054123401642,
0.08323734253644943,
0.06597128510475159,
0.06804143637418747,
-0.18422959744930267,
-0.04144259914755821,
0.001133312121964991,
0.12732920050621033,
0.08202160894870758,
0.11622790992259979,
0.03691718727350235,
-0.027924707159399986,
0.03776296600699425,
0.0294797383248806,
0.17191994190216064,
-0.010394933633506298,
-0.09008485823869705,
-0.08011404424905777,
0.07458759099245071,
-0.06218329817056656,
0.09133265912532806,
-0.058848779648542404,
0.006888185627758503,
0.004084744956344366,
0.059586260467767715,
-0.018083417788147926,
0.04765148460865021,
0.0054670353420078754,
-0.0063790855929255486,
-0.0804750919342041,
0.02415669709444046,
0.10147536545991898,
-0.021984102204442024,
-0.157908633351326,
0.24990110099315643,
-0.18417911231517792,
0.19290444254875183,
0.1820327341556549,
-0.28881335258483887,
-0.010748554952442646,
-0.03560630604624748,
-0.009900180622935295,
0.0005574178067035973,
0.03434859216213226,
-0.04047844558954239,
0.24925856292247772,
-0.045190684497356415,
0.14460696280002594,
-0.02821434661746025,
-0.017426081001758575,
-0.004116753116250038,
-0.01791791245341301,
-0.02001170627772808,
0.09211008995771408,
0.0517912395298481,
-0.17028796672821045,
0.17686127126216888,
0.08165227621793747,
0.03338776156306267,
0.1503230631351471,
0.06869526952505112,
0.008418656885623932,
0.06592967361211777,
-0.005641093477606773,
-0.08167476952075958,
-0.03858598694205284,
-0.3079616129398346,
-0.07181378453969955,
0.06082563474774361,
0.02474597841501236,
0.05565487593412399,
-0.09929995238780975,
-0.05340031906962395,
-0.03553680330514908,
-0.02575518563389778,
0.03879416361451149,
0.07839484512805939,
0.02556057833135128,
0.1616465300321579,
0.012503295205533504,
-0.0692838579416275,
0.08199453353881836,
0.009587259963154793,
-0.11852023005485535,
0.16681042313575745,
-0.09758856147527695,
-0.3797800838947296,
-0.04949454963207245,
-0.16155755519866943,
-0.058394525200128555,
0.1013612449169159,
0.07273466885089874,
-0.08528527617454529,
-0.0021454375237226486,
-0.026280466467142105,
0.08669473230838776,
-0.009320855140686035,
-0.002410613000392914,
0.008865557610988617,
0.02326781116425991,
-0.08733169734477997,
-0.04100080579519272,
-0.0676080510020256,
-0.05424739047884941,
-0.05840599909424782,
0.13410604000091553,
-0.11442612111568451,
0.07106238603591919,
0.1837177574634552,
0.04153222590684891,
0.04153979942202568,
-0.05089624971151352,
0.1973029226064682,
-0.14032910764217377,
-0.04564874246716499,
0.13880406320095062,
-0.032560333609580994,
0.003368969773873687,
0.13079066574573517,
-0.005484892521053553,
-0.11447549611330032,
0.03628795966506004,
-0.02563749998807907,
-0.026455292478203773,
-0.19012466073036194,
-0.1546899378299713,
-0.07239345461130142,
0.14098526537418365,
0.036410797387361526,
0.056783322244882584,
0.11400268971920013,
0.049062978476285934,
-0.03197875991463661,
0.042720548808574677,
-0.01983572542667389,
0.04835416376590729,
0.1182582676410675,
-0.092552050948143,
0.14946584403514862,
0.0037218681536614895,
-0.16641250252723694,
0.0677325502038002,
0.05867665261030197,
0.11679588258266449,
0.07329073548316956,
0.063916876912117,
-0.0066140214912593365,
0.03484624996781349,
0.08536344766616821,
0.06210454925894737,
0.053092628717422485,
-0.07586433738470078,
-0.05259844660758972,
-0.009690036065876484,
-0.11519566178321838,
0.08459589630365372,
0.09259974211454391,
-0.15445668995380402,
-0.050938207656145096,
-0.06382802128791809,
0.08048167824745178,
0.12958265841007233,
0.0764935240149498,
-0.22922338545322418,
-0.029859056696295738,
0.07786137610673904,
-0.06076185777783394,
-0.13066507875919342,
0.1392451971769333,
0.062492433935403824,
-0.17300117015838623,
0.012820461764931679,
0.005783524364233017,
0.1032494604587555,
-0.13816680014133453,
0.11103653162717819,
-0.14064444601535797,
-0.14870664477348328,
0.033505719155073166,
0.14031468331813812,
-0.31244587898254395,
0.25629422068595886,
-0.025429707020521164,
-0.05134071782231331,
-0.16315625607967377,
-0.058136481791734695,
0.009145167656242847,
0.09518957883119583,
0.12543033063411713,
-0.014629860408604145,
0.008795126341283321,
-0.14213085174560547,
-0.02916601113975048,
0.04392401501536369,
0.11295381933450699,
-0.05109962448477745,
-0.013501400128006935,
-0.017852744087576866,
0.00766797037795186,
-0.05964862182736397,
-0.07353363931179047,
0.023861659690737724,
-0.15414519608020782,
0.06537028402090073,
0.0993466004729271,
0.15519917011260986,
0.062300629913806915,
-0.007939165458083153,
-0.12770238518714905,
0.16607244312763214,
-0.043589793145656586,
-0.004688586108386517,
-0.06472531706094742,
0.011958969756960869,
-0.062079090625047684,
-0.0521785207092762,
-0.05804715305566788,
-0.0548616386950016,
0.036350447684526443,
-0.034557778388261795,
-0.1673133224248886,
0.1282399445772171,
-0.09447922557592392,
-0.03685007616877556,
-0.038022417575120926,
0.11632813513278961,
0.015713833272457123,
0.0602918341755867,
0.02665436454117298,
-0.038049641996622086,
-0.1152748093008995,
-0.0535145029425621,
0.07635674625635147,
0.047279682010412216,
0.03125641494989395,
0.08932244032621384,
-0.08072894811630249,
-0.09191268682479858,
-0.14865601062774658,
-0.0785892978310585,
0.28734198212623596,
0.10966076701879501,
-0.054952364414930344,
0.15449656546115875,
0.18759386241436005,
-0.03905734792351723,
-0.31987982988357544,
-0.10587304830551147,
-0.1154530867934227,
-0.0061977459117770195,
-0.012593029998242855,
-0.1912873536348343,
0.10843311995267868,
-0.026867074891924858,
-0.006699667777866125,
0.044475454837083817,
-0.2256724089384079,
-0.06554846465587616,
0.17590369284152985,
0.013078616932034492,
0.3336867094039917,
-0.11388570815324783,
-0.06356906145811081,
-0.08098386228084564,
-0.08016274869441986,
0.15271705389022827,
0.008960085920989513,
0.1415291428565979,
0.019674140959978104,
0.11291509121656418,
0.05007925257086754,
0.025165345519781113,
0.07750560343265533,
0.014051343314349651,
-0.017292441800236702,
-0.13150815665721893,
-0.1066778227686882,
-0.02233164943754673,
0.049875661730766296,
0.05478158965706825,
-0.0832940936088562,
-0.03131899610161781,
-0.09577637910842896,
-0.04302506148815155,
-0.07445015013217926,
0.008203650824725628,
0.06824552267789841,
-0.05566205829381943,
-0.05334775522351265,
0.00233004754409194,
-0.01085583958774805,
0.07043241709470749,
0.11813101917505264,
-0.09406010061502457,
0.10832833498716354,
0.04039851948618889,
0.18514162302017212,
-0.164232075214386,
0.08142922073602676,
-0.10744205862283707,
-0.057973332703113556,
0.11111876368522644,
-0.04102333262562752,
0.03326955437660217,
0.11288056522607803,
-0.06819834560155869,
0.1428184062242508,
0.08227946609258652,
0.004001641646027565,
0.05212830752134323,
0.09517990797758102,
-0.2619417905807495,
-0.14311625063419342,
-0.06502223759889603,
0.05445655807852745,
0.11073784530162811,
0.11279504001140594,
0.16236360371112823,
-0.015746043995022774,
-0.030617425218224525,
-0.02111150696873665,
0.03653861582279205,
-0.04246930405497551,
0.05940132588148117,
-0.09892398864030838,
0.02310069091618061,
-0.12914863228797913,
0.06858740746974945,
0.037140410393476486,
-0.16340765357017517,
0.08673117309808731,
0.10655556619167328,
-0.11285718530416489,
-0.1507195383310318,
-0.11485505104064941,
0.11686483770608902,
-0.07257550209760666,
-0.03259739652276039,
-0.06824848055839539,
-0.19055970013141632,
0.060937173664569855,
0.12276032567024231,
0.09115085005760193,
0.09852015227079391,
-0.05380706116557121,
-0.03736652806401253,
0.003043286968022585,
-0.014258616603910923,
-0.007480786647647619,
-0.04984479397535324,
-0.10031095892190933,
0.02219414710998535,
-0.0011161868460476398,
0.16553379595279694,
-0.09328939020633698,
-0.15163461863994598,
-0.14662401378154755,
0.07188251614570618,
-0.08719442784786224,
-0.09514754265546799,
-0.09674599021673203,
-0.05124505236744881,
0.026573780924081802,
-0.055567417293787,
-0.05616330727934837,
-0.08905370533466339,
-0.09677935391664505,
0.03802897036075592,
-0.009129954501986504,
0.03001278080046177,
-0.04671640694141388,
0.021158771589398384,
0.0594867505133152,
-0.02151288092136383,
0.18796701729297638,
0.15643981099128723,
-0.10975909978151321,
0.03441900759935379,
-0.18576405942440033,
-0.05935840308666229,
0.08613146841526031,
0.01680014841258526,
0.05262842774391174,
0.025084661319851875,
-0.02282802388072014,
0.061986785382032394,
0.07366884499788284,
0.07732667773962021,
0.10477138310670853,
-0.07902639359235764,
0.06838399916887283,
-0.035186897963285446,
-0.14235341548919678,
-0.0972394198179245,
-0.036284446716308594,
0.029091985896229744,
-0.035294465720653534,
0.11261554807424545,
-0.054233986884355545,
0.07645629346370697,
-0.00955826323479414,
0.07576919347047806,
0.03541799634695053,
-0.12239482253789902,
-0.042980216443538666,
-0.10935383290052414,
0.049057237803936005,
-0.03997376561164856,
0.2395099252462387,
0.01896904595196247,
-0.059217389672994614,
0.013339613564312458,
0.044208772480487823,
0.0012174854055047035,
0.01322883740067482,
0.09696703404188156,
0.08833226561546326,
-0.06110695004463196,
-0.07412902265787125,
0.10990956425666809,
0.007596834097057581,
0.15556763112545013,
0.13793712854385376,
-0.015497180633246899,
0.0009897053241729736,
0.10606106370687485,
0.02196023054420948,
0.001383335329592228,
-0.15145136415958405,
-0.11616602540016174,
-0.10413810610771179,
0.047600630670785904,
-0.035752635449171066,
0.17045697569847107,
0.14318567514419556,
-0.0007618785602971911,
0.022412452846765518,
-0.09101748466491699,
-0.07640907913446426,
-0.1404293328523636,
-0.21525068581104279,
-0.06960782408714294,
-0.13274289667606354,
0.009873799048364162,
-0.09885670989751816,
0.029257943853735924,
0.04050661250948906,
0.07858722656965256,
-0.12369468063116074,
0.13036102056503296,
-0.003690937766805291,
-0.06506668031215668,
0.06547994166612625,
-0.016384435817599297,
0.03624450042843819,
-0.027458805590867996,
-0.036674145609140396,
-0.0740603655576706,
0.01902889832854271,
0.04229266941547394,
0.0680290162563324,
-0.06548894196748734,
0.014811040833592415,
-0.12333966046571732,
-0.10270524770021439,
-0.03915904089808464,
0.005173001438379288,
-0.055014509707689285,
0.16598016023635864,
0.03178445249795914,
-0.025743499398231506,
0.018838750198483467,
0.22054871916770935,
-0.06118617579340935,
0.02437753416597843,
-0.09808903932571411,
0.25076255202293396,
-0.015741482377052307,
0.07903869450092316,
-0.04433654993772507,
0.012549792416393757,
-0.10452570766210556,
0.4301941990852356,
0.2826985716819763,
-0.16360434889793396,
-0.010291927494108677,
-0.005598035641014576,
0.045635420829057693,
0.0630568116903305,
0.11378786712884903,
0.10981728881597519,
0.2685222625732422,
-0.06484409421682358,
-0.022354645654559135,
-0.02127842791378498,
-0.011520898900926113,
-0.07045987993478775,
0.06519799679517746,
0.0822216346859932,
-0.01550370268523693,
-0.07408257573843002,
0.06711925566196442,
-0.25471457839012146,
0.1431485265493393,
-0.13230760395526886,
-0.1723562628030777,
-0.061475638300180435,
0.04438233748078346,
0.07545407861471176,
0.0411190427839756,
0.12167362868785858,
-0.01252028439193964,
-0.05605369061231613,
0.004169612191617489,
0.028743837028741837,
-0.2170577049255371,
0.06136705353856087,
0.04202483594417572,
-0.10450942814350128,
0.05265279859304428,
-0.03952869772911072,
0.10128984600305557,
0.06085684895515442,
0.0731324851512909,
-0.013160436414182186,
0.10841627418994904,
-0.050534654408693314,
-0.026446150615811348,
0.042708225548267365,
0.12272343784570694,
0.03956837207078934,
0.006649714428931475,
0.11759179830551147,
-0.09593188017606735,
0.050774361938238144,
-0.025080611929297447,
-0.001054677297361195,
-0.013316822238266468,
0.04623565450310707,
-0.05490276217460632,
0.08143649995326996,
0.09237409383058548,
0.005757124163210392,
-0.03019457682967186,
-0.015705611556768417,
-0.043497662991285324,
-0.06369537115097046,
-0.09497995674610138,
-0.11029578745365143,
-0.20349746942520142,
-0.09305787831544876,
0.04503747448325157,
0.04229537025094032,
-0.20118287205696106,
0.024561235681176186,
-0.08212891221046448,
0.016726089641451836,
-0.06699185073375702,
0.0662698745727539,
0.0929829403758049,
0.015368036925792694,
-0.028717394918203354,
-0.07269744575023651,
0.08445686101913452,
0.10421397536993027,
-0.11505100876092911,
-0.1421675682067871
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-base-demo-colab
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4171
- Wer: 0.3452
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 30
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 3.0054 | 4.0 | 500 | 1.5456 | 0.9005 |
| 0.8183 | 8.0 | 1000 | 0.4738 | 0.4839 |
| 0.3019 | 12.0 | 1500 | 0.4280 | 0.4047 |
| 0.1738 | 16.0 | 2000 | 0.4584 | 0.3738 |
| 0.1285 | 20.0 | 2500 | 0.4418 | 0.3593 |
| 0.1104 | 24.0 | 3000 | 0.4110 | 0.3479 |
| 0.0828 | 28.0 | 3500 | 0.4171 | 0.3452 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu102
- Datasets 1.14.0
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-base-demo-colab", "results": []}]}
|
automatic-speech-recognition
|
dkssud/wav2vec2-base-demo-colab
|
[
"transformers",
"pytorch",
"tensorboard",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
|
wav2vec2-base-demo-colab
========================
This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4171
* Wer: 0.3452
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 32
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 1000
* num\_epochs: 30
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.10.0+cu102
* Datasets 1.14.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu102\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu102\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
56,
130,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu102\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
-0.10990507900714874,
0.10211705416440964,
-0.003297114744782448,
0.06257002800703049,
0.10761956125497818,
-0.019553201273083687,
0.12904250621795654,
0.15090572834014893,
-0.09363732486963272,
0.0746912956237793,
0.12591767311096191,
0.14810509979724884,
0.0418579988181591,
0.14393915235996246,
-0.050227973610162735,
-0.2824990451335907,
0.045772358775138855,
0.03674701228737831,
-0.012884746305644512,
0.12689591944217682,
0.08315597474575043,
-0.12503120303153992,
0.05902668088674545,
0.03576003760099411,
-0.1615888625383377,
-0.004325156565755606,
-0.003226091153919697,
-0.10516153275966644,
0.12333071231842041,
0.006191861815750599,
0.07049113512039185,
0.0490293987095356,
0.06638041883707047,
-0.21724413335323334,
0.006878388114273548,
0.04486188292503357,
0.028785521164536476,
0.07603919506072998,
0.05723509192466736,
-0.029228337109088898,
0.10441087186336517,
-0.07598480582237244,
0.07847484946250916,
0.03678025305271149,
-0.10486087948083878,
-0.2944018542766571,
-0.08612558245658875,
0.04586160555481911,
0.0668441653251648,
0.08845404535531998,
-0.011601281352341175,
0.14171260595321655,
-0.05288305878639221,
0.10891370475292206,
0.2784803509712219,
-0.31746888160705566,
-0.044037796556949615,
-0.0413191020488739,
0.05628824979066849,
0.062316253781318665,
-0.10100409388542175,
-0.01513531245291233,
0.01625831238925457,
0.04442140460014343,
0.13984185457229614,
-0.015844708308577538,
-0.05937682092189789,
-0.006858822889626026,
-0.14841023087501526,
-0.059705059975385666,
0.11671796441078186,
0.023261629045009613,
-0.04146599397063255,
-0.0961369127035141,
-0.05649082362651825,
-0.21192744374275208,
-0.06782354414463043,
-0.0161895751953125,
0.04438335821032524,
-0.041981860995292664,
-0.10514768958091736,
-0.013483500108122826,
-0.0676579549908638,
-0.07438353449106216,
-0.04242468997836113,
0.19149327278137207,
0.056671347469091415,
-0.00038654424133710563,
-0.03917283937335014,
0.07626061141490936,
-0.02310885488986969,
-0.13767993450164795,
-0.02446908876299858,
0.03802306950092316,
-0.018093833699822426,
-0.014139514416456223,
-0.044199664145708084,
-0.05759956315159798,
0.021235240623354912,
0.15930064022541046,
-0.10457636415958405,
0.09511861950159073,
-0.022157587110996246,
0.039553202688694,
-0.10422401130199432,
0.2075354903936386,
-0.04185591638088226,
0.01570659689605236,
-0.007972552441060543,
0.05469728261232376,
0.030425962060689926,
-0.026370922103524208,
-0.0947292298078537,
0.028070347383618355,
0.12121522426605225,
0.04528758302330971,
-0.049125947058200836,
0.06484611332416534,
-0.03158611059188843,
-0.00944135058671236,
0.0014207955682650208,
-0.11203587055206299,
0.036573659628629684,
0.018366901203989983,
-0.06564781069755554,
0.005594003479927778,
0.016313230618834496,
0.006314799189567566,
-0.05480010434985161,
0.0822741687297821,
-0.06200375407934189,
0.03299076110124588,
-0.05691904574632645,
-0.1253179907798767,
0.02394728548824787,
-0.11591418832540512,
-0.0038579872343689203,
-0.10008161514997482,
-0.09909261018037796,
-0.011129423975944519,
0.03645167499780655,
-0.037996917963027954,
-0.026066921651363373,
-0.07897105067968369,
-0.09013079851865768,
0.04600611701607704,
-0.034683991223573685,
0.07183605432510376,
-0.07426610589027405,
0.09409259259700775,
0.0344291552901268,
0.08872653543949127,
-0.019412335008382797,
0.06017905846238136,
-0.07006995379924774,
0.02632235176861286,
-0.19892241060733795,
0.07589258998632431,
-0.08548300713300705,
0.054927077144384384,
-0.12447766959667206,
-0.11478383094072342,
0.02113022841513157,
-0.007723475340753794,
0.09902346134185791,
0.09744346886873245,
-0.17054219543933868,
-0.0898139700293541,
0.2097122073173523,
-0.08225692063570023,
-0.08359568566083908,
0.12416361272335052,
-0.025165671482682228,
0.00009726683492772281,
0.057073697447776794,
0.2570069134235382,
0.04637500271201134,
-0.12591734528541565,
0.0067894961684942245,
-0.04016532003879547,
0.04270398989319801,
-0.03766202554106712,
0.05810334160923958,
-0.028114762157201767,
0.06753865629434586,
0.01808278262615204,
-0.0051117148250341415,
0.0375678688287735,
-0.0870816558599472,
-0.07658039778470993,
-0.04549587145447731,
-0.07860146462917328,
0.031089816242456436,
0.030820034444332123,
0.06491460651159286,
-0.11635862290859222,
-0.10779207199811935,
0.03839510306715965,
0.08244876563549042,
-0.1045709177851677,
0.07273725420236588,
-0.12198898941278458,
0.08445484936237335,
-0.017597146332263947,
-0.00448177196085453,
-0.191769540309906,
0.037105124443769455,
0.037205904722213745,
-0.028597049415111542,
0.0414179228246212,
-0.06504637748003006,
0.07778604328632355,
0.04582294821739197,
-0.027298370376229286,
-0.04608463868498802,
-0.009400169365108013,
0.009316111914813519,
-0.09054146707057953,
-0.20757724344730377,
-0.037583328783512115,
-0.03656462952494621,
0.08067874610424042,
-0.1379251331090927,
0.033641621470451355,
0.07906951010227203,
0.09299125522375107,
0.03290092945098877,
-0.030587606132030487,
-0.0027329714503139257,
0.08975344151258469,
-0.020349163562059402,
-0.06401896476745605,
0.058507174253463745,
0.02071816846728325,
-0.08701269328594208,
0.03970043361186981,
-0.1478501707315445,
0.1259550303220749,
0.14862559735774994,
-0.014517481438815594,
-0.06424880772829056,
0.0004721948062069714,
-0.047224175184965134,
-0.03440159931778908,
-0.0034738853573799133,
0.034786734730005264,
0.21570689976215363,
0.012164360843598843,
0.1439845710992813,
-0.08942718803882599,
-0.041154831647872925,
0.04983580857515335,
-0.0215302761644125,
-0.0047544511035084724,
0.11428047716617584,
0.04386397823691368,
-0.05370059236884117,
0.11681610345840454,
0.09235872328281403,
-0.08052141219377518,
0.12052285671234131,
-0.06025775521993637,
-0.07506512105464935,
-0.020610159263014793,
0.004121708683669567,
0.023160168901085854,
0.09972615540027618,
-0.16250139474868774,
-0.03979765251278877,
0.026957359164953232,
0.02573549933731556,
0.0208971556276083,
-0.2088443487882614,
0.016841787844896317,
0.028453540056943893,
-0.08668770641088486,
-0.04362604022026062,
0.0033653397113084793,
0.012787867337465286,
0.0947108268737793,
0.0123748192563653,
-0.09401075541973114,
0.011736798100173473,
0.004371959250420332,
-0.07243796437978745,
0.17359232902526855,
-0.11728378385305405,
-0.1764698326587677,
-0.10659637302160263,
-0.0919342115521431,
-0.04130029305815697,
-0.0025137874763458967,
0.08835284411907196,
-0.09228593856096268,
-0.039077892899513245,
-0.08215326815843582,
-0.015744294971227646,
-0.0250747948884964,
0.0439627431333065,
0.03142151981592178,
-0.011124195531010628,
0.06622706353664398,
-0.1159999743103981,
-0.021054957062005997,
-0.040593698620796204,
-0.0014292672276496887,
0.05351954326033592,
0.0399421751499176,
0.10857680439949036,
0.15708349645137787,
-0.012292698957026005,
0.0510074608027935,
-0.04500279575586319,
0.1878989338874817,
-0.07533378899097443,
-0.03660254925489426,
0.11319431662559509,
-0.005652599036693573,
0.06803672760725021,
0.11483248323202133,
0.04803217947483063,
-0.09805029630661011,
-0.014427346177399158,
0.0033321816008538008,
-0.04533176124095917,
-0.21415166556835175,
-0.03666726499795914,
-0.04437775909900665,
-0.0037671118043363094,
0.10713502019643784,
0.04014216363430023,
0.038449726998806,
0.02199925296008587,
0.03247474506497383,
0.00825584214180708,
0.000054199976148083806,
0.0956030786037445,
0.13199786841869354,
0.040401361882686615,
0.1330568492412567,
-0.03910743445158005,
-0.036699336022138596,
0.030793217942118645,
0.0035793150309473276,
0.23133957386016846,
0.019678344950079918,
0.19021356105804443,
0.05532928928732872,
0.17944756150245667,
0.043429918587207794,
0.06690605729818344,
-0.0023396897595375776,
-0.009959312155842781,
0.011278565041720867,
-0.05294645577669144,
-0.03971133381128311,
0.02427331544458866,
0.02352992445230484,
0.009308015927672386,
-0.11457810550928116,
-0.011627176776528358,
0.04628331586718559,
0.35284844040870667,
0.028289474546909332,
-0.34075871109962463,
-0.09226159006357193,
-0.013404236175119877,
-0.08369192481040955,
-0.03213493898510933,
0.04526066407561302,
0.09033183008432388,
-0.07945702224969864,
0.06515537947416306,
-0.06299155205488205,
0.08991766721010208,
-0.06500929594039917,
0.034707777202129364,
0.03820282220840454,
0.07339128106832504,
0.0048638759180903435,
0.03265209123492241,
-0.2898210883140564,
0.2793630361557007,
0.006034130696207285,
0.07708834111690521,
-0.06082336604595184,
0.007194784004241228,
0.025228269398212433,
0.021732434630393982,
0.08580184727907181,
-0.025588467717170715,
-0.1208546906709671,
-0.1702190637588501,
-0.09364084899425507,
0.012762355618178844,
0.1297474503517151,
0.01315348595380783,
0.1113365888595581,
-0.01072810497134924,
-0.017180608585476875,
0.04906536266207695,
-0.09917290508747101,
-0.06482488662004471,
-0.09210650622844696,
0.011197572574019432,
0.08761947602033615,
0.03655362129211426,
-0.07215727865695953,
-0.10260820388793945,
-0.0872834175825119,
0.15024925768375397,
-0.053485769778490067,
-0.0426708348095417,
-0.11888366937637329,
0.00950080156326294,
0.10873589664697647,
-0.07880233973264694,
0.06160972639918327,
0.009817087091505527,
0.10695920139551163,
0.01298584509640932,
-0.06820303201675415,
0.11876120418310165,
-0.06287667900323868,
-0.1657637655735016,
-0.027579015120863914,
0.14555718004703522,
0.02881922759115696,
0.059041086584329605,
-0.007735255640000105,
0.03802546113729477,
-0.02139284275472164,
-0.07680560648441315,
0.040992848575115204,
0.027146749198436737,
0.04468349367380142,
-0.013781015761196613,
-0.020065337419509888,
-0.00559582281857729,
-0.09008786827325821,
-0.01735820062458515,
0.20602461695671082,
0.23935596644878387,
-0.09733253717422485,
0.09275693446397781,
0.07028436660766602,
-0.04277517646551132,
-0.17293500900268555,
-0.004132713656872511,
0.06409448385238647,
0.0005976536776870489,
-0.02654098905622959,
-0.195436492562294,
0.02398286759853363,
0.0679253414273262,
-0.020725078880786896,
0.08501172065734863,
-0.3164432644844055,
-0.14052756130695343,
0.13563483953475952,
0.11191535741090775,
0.06225988641381264,
-0.14468447864055634,
-0.056358702480793,
-0.012377465143799782,
-0.10167541354894638,
0.09403214603662491,
-0.07799284160137177,
0.1353168934583664,
-0.023906830698251724,
0.09333931654691696,
0.011099720373749733,
-0.057143859565258026,
0.10745324194431305,
0.012605533003807068,
0.059434399008750916,
-0.04536578431725502,
0.019979868084192276,
0.048026420176029205,
-0.06312011182308197,
0.05396903678774834,
-0.08030364662408829,
0.029262155294418335,
-0.07810692489147186,
-0.03165038675069809,
-0.08474446833133698,
0.014463388361036777,
-0.008140059188008308,
-0.03188612312078476,
-0.0375061072409153,
-0.000392145972000435,
0.061753641813993454,
-0.010364188812673092,
0.15506094694137573,
-0.026825379580259323,
0.1245499774813652,
0.16213664412498474,
0.10081469267606735,
-0.10544422268867493,
-0.07724058628082275,
0.006238001864403486,
-0.03440485894680023,
0.05734509229660034,
-0.11903215199708939,
0.03757534176111221,
0.1351834088563919,
0.03176051005721092,
0.12188097834587097,
0.06902486830949783,
-0.06488844007253647,
0.03384406492114067,
0.04253392294049263,
-0.1398983597755432,
-0.12850721180438995,
0.013628397136926651,
0.02176450379192829,
-0.07142476737499237,
0.07539382576942444,
0.11819182336330414,
-0.05584123358130455,
-0.012837033718824387,
-0.00256941351108253,
0.013180771842598915,
-0.042447399348020554,
0.1941438615322113,
0.03552790358662605,
0.059860095381736755,
-0.12511619925498962,
0.08197294175624847,
0.0395730622112751,
-0.13477081060409546,
0.06044166535139084,
0.10195010155439377,
-0.09623312950134277,
-0.02743944339454174,
0.02725612185895443,
0.1130438819527626,
-0.02907666750252247,
-0.07587580382823944,
-0.141915962100029,
-0.14261631667613983,
0.10932207852602005,
0.20347394049167633,
0.0560273639857769,
0.01755066215991974,
-0.05968805029988289,
0.015957999974489212,
-0.12006421387195587,
0.06828288733959198,
0.039714641869068146,
0.06056049093604088,
-0.12843579053878784,
0.14754942059516907,
0.016951797530055046,
0.04055451601743698,
-0.015279359184205532,
-0.011015313677489758,
-0.10831591486930847,
0.03894968703389168,
-0.13083069026470184,
0.005416560918092728,
-0.06396941095590591,
0.000731137057300657,
0.003960904665291309,
-0.050333570688962936,
-0.06481588631868362,
0.03538554161787033,
-0.11987045407295227,
-0.023268986493349075,
0.0016676407540217042,
0.03617945313453674,
-0.1283583790063858,
-0.00879155658185482,
0.0145560372620821,
-0.09398949146270752,
0.09816337376832962,
0.08874616771936417,
-0.03279922902584076,
0.0520227812230587,
-0.05522436648607254,
-0.027205724269151688,
0.07889269292354584,
-0.006205665413290262,
0.0517653189599514,
-0.13033601641654968,
-0.01937410980463028,
0.012280846014618874,
0.03522607684135437,
0.024201299995183945,
0.113174207508564,
-0.11506587266921997,
0.0006673377356491983,
-0.02786814607679844,
-0.04848286509513855,
-0.06859592348337173,
0.05008001625537872,
0.10922791063785553,
0.028742050752043724,
0.16341343522071838,
-0.09267827868461609,
0.025976400822401047,
-0.16585710644721985,
0.00657259114086628,
-0.016369666904211044,
-0.1225254088640213,
-0.050976965576410294,
-0.03182489052414894,
0.07875648140907288,
-0.0641288235783577,
0.1278829574584961,
-0.02982764132320881,
0.025310758501291275,
0.03688513860106468,
-0.07792757451534271,
-0.05770998075604439,
0.03962317109107971,
0.20627844333648682,
0.04100025072693825,
-0.043003275990486145,
0.07610034197568893,
0.020667705684900284,
0.08176371455192566,
0.12952186167240143,
0.17206816375255585,
0.1590851992368698,
0.06569401919841766,
0.11859188228845596,
0.0542730949819088,
-0.05241155996918678,
-0.17525412142276764,
0.09465895593166351,
-0.06081477552652359,
0.13145676255226135,
-0.014056926593184471,
0.24134208261966705,
0.12025582790374756,
-0.15348124504089355,
0.06622019410133362,
-0.01773209497332573,
-0.08923067897558212,
-0.11616530269384384,
-0.06300552934408188,
-0.08708155155181885,
-0.17771810293197632,
0.008517513051629066,
-0.10218165069818497,
0.0635441392660141,
0.04540541395545006,
0.0358591191470623,
0.01657494530081749,
0.13799142837524414,
0.01336319837719202,
0.0012863193405792117,
0.09336797893047333,
-0.003671394195407629,
-0.056533101946115494,
-0.07270506769418716,
-0.08591271936893463,
0.03662379831075668,
-0.012097473256289959,
0.05819477140903473,
-0.003642426570877433,
-0.06851039826869965,
0.04867429658770561,
-0.03912492096424103,
-0.09536723792552948,
0.02279924601316452,
0.021768659353256226,
0.07116661965847015,
0.048840805888175964,
0.03437146171927452,
-0.04216110706329346,
-0.0031414320692420006,
0.19583375751972198,
-0.09526411443948746,
-0.09271862357854843,
-0.10957486182451248,
0.2522122263908386,
0.04028838127851486,
-0.016024138778448105,
0.022165216505527496,
-0.060451991856098175,
-0.031835976988077164,
0.2131224274635315,
0.17086265981197357,
-0.011943940073251724,
0.003569989465177059,
-0.015347304753959179,
-0.006671391427516937,
-0.03713202103972435,
0.07994541525840759,
0.14711757004261017,
0.060044750571250916,
-0.06232253462076187,
-0.05216575041413307,
-0.051454029977321625,
-0.0325327105820179,
-0.06584294885396957,
0.07574187219142914,
0.0028157602064311504,
-0.025828726589679718,
-0.0449620820581913,
0.062198199331760406,
-0.09164781123399734,
-0.08644987642765045,
0.0269437488168478,
-0.19483257830142975,
-0.1502104550600052,
0.008146511390805244,
0.07223305851221085,
0.011575497686862946,
0.03508998453617096,
0.002154422225430608,
-0.010382188484072685,
0.08482854813337326,
-0.0028023573104292154,
-0.07953513413667679,
-0.06648353487253189,
0.08402900397777557,
-0.1370413601398468,
0.1658848524093628,
-0.041781261563301086,
0.047366462647914886,
0.12208302319049835,
0.08949168026447296,
-0.08203905075788498,
0.08538515120744705,
0.040444958955049515,
-0.10741742700338364,
0.021375872194767,
0.15536905825138092,
-0.034009624272584915,
0.09375419467687607,
0.030367780476808548,
-0.1123981922864914,
0.013125726953148842,
-0.08966563642024994,
-0.038833148777484894,
-0.039704885333776474,
-0.05290498584508896,
-0.04454682394862175,
0.10899648815393448,
0.16255927085876465,
-0.044065721333026886,
0.00525862630456686,
-0.052058834582567215,
0.012048096396028996,
0.04903191700577736,
-0.001989846583455801,
-0.06030776724219322,
-0.2777145802974701,
0.01104407012462616,
0.03874525427818298,
0.002175118075683713,
-0.25773096084594727,
-0.09610304981470108,
0.014256286434829235,
-0.042431361973285675,
-0.08693139255046844,
0.08632005751132965,
0.07487823814153671,
0.04530692473053932,
-0.052606482058763504,
-0.0580001026391983,
-0.03507240116596222,
0.18861325085163116,
-0.17540769279003143,
-0.06048892065882683
] |
null | null |
transformers
|
# OpenVINO model bert-large-uncased-whole-word-masking-squad-int8-0001
This is a BERT-large model pre-trained on lower-cased English text using Whole-Word-Masking and fine-tuned on the SQuAD v1.1 training set. The model performs question answering for English language; the input is a concatenated premise and question for the premise, and the output is the location of the answer to the question inside the premise. For details about the original floating-point model, check out [BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding](https://arxiv.org/abs/1810.04805).
The model has been further quantized to INT8 precision using quantization-aware fine-tuning with [NNCF](https://github.com/openvinotoolkit/nncf).
Model source: [Open Model Zoo](https://github.com/openvinotoolkit/open_model_zoo/tree/master/models/intel/bert-large-uncased-whole-word-masking-squad-int8-0001)
|
{}
|
question-answering
|
dkurt/bert-large-uncased-whole-word-masking-squad-int8-0001
|
[
"transformers",
"bert",
"question-answering",
"arxiv:1810.04805",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1810.04805"
] |
[] |
TAGS
#transformers #bert #question-answering #arxiv-1810.04805 #endpoints_compatible #region-us
|
# OpenVINO model bert-large-uncased-whole-word-masking-squad-int8-0001
This is a BERT-large model pre-trained on lower-cased English text using Whole-Word-Masking and fine-tuned on the SQuAD v1.1 training set. The model performs question answering for English language; the input is a concatenated premise and question for the premise, and the output is the location of the answer to the question inside the premise. For details about the original floating-point model, check out BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.
The model has been further quantized to INT8 precision using quantization-aware fine-tuning with NNCF.
Model source: Open Model Zoo
|
[
"# OpenVINO model bert-large-uncased-whole-word-masking-squad-int8-0001\n\nThis is a BERT-large model pre-trained on lower-cased English text using Whole-Word-Masking and fine-tuned on the SQuAD v1.1 training set. The model performs question answering for English language; the input is a concatenated premise and question for the premise, and the output is the location of the answer to the question inside the premise. For details about the original floating-point model, check out BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.\n\nThe model has been further quantized to INT8 precision using quantization-aware fine-tuning with NNCF.\n\nModel source: Open Model Zoo"
] |
[
"TAGS\n#transformers #bert #question-answering #arxiv-1810.04805 #endpoints_compatible #region-us \n",
"# OpenVINO model bert-large-uncased-whole-word-masking-squad-int8-0001\n\nThis is a BERT-large model pre-trained on lower-cased English text using Whole-Word-Masking and fine-tuned on the SQuAD v1.1 training set. The model performs question answering for English language; the input is a concatenated premise and question for the premise, and the output is the location of the answer to the question inside the premise. For details about the original floating-point model, check out BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.\n\nThe model has been further quantized to INT8 precision using quantization-aware fine-tuning with NNCF.\n\nModel source: Open Model Zoo"
] |
[
34,
184
] |
[
"passage: TAGS\n#transformers #bert #question-answering #arxiv-1810.04805 #endpoints_compatible #region-us \n# OpenVINO model bert-large-uncased-whole-word-masking-squad-int8-0001\n\nThis is a BERT-large model pre-trained on lower-cased English text using Whole-Word-Masking and fine-tuned on the SQuAD v1.1 training set. The model performs question answering for English language; the input is a concatenated premise and question for the premise, and the output is the location of the answer to the question inside the premise. For details about the original floating-point model, check out BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding.\n\nThe model has been further quantized to INT8 precision using quantization-aware fine-tuning with NNCF.\n\nModel source: Open Model Zoo"
] |
[
-0.04113134741783142,
0.04934564232826233,
-0.000050774749979609624,
0.02632891945540905,
0.06078222021460533,
-0.02312265895307064,
0.09145181626081467,
0.09495365619659424,
0.06050097942352295,
0.04725482314825058,
0.01480435486882925,
0.08083748817443848,
0.0736924260854721,
0.10514120012521744,
0.0018655937165021896,
-0.23158924281597137,
0.08817971497774124,
0.02676631510257721,
-0.06108902767300606,
-0.006688280031085014,
0.03678160160779953,
-0.07900404930114746,
0.06096675619482994,
0.05664557218551636,
-0.03545042499899864,
-0.012103032320737839,
-0.08988353610038757,
-0.04302623122930527,
0.10994607210159302,
0.013689033687114716,
0.030345525592565536,
0.0023020596709102392,
0.030754711478948593,
-0.06137244775891304,
0.0367877334356308,
0.007508378941565752,
-0.0026673206593841314,
0.07757559418678284,
0.00676092691719532,
0.1279163658618927,
0.02930414490401745,
0.040739547461271286,
0.05235082656145096,
0.06225911155343056,
-0.038283031433820724,
-0.17381149530410767,
0.025856012478470802,
-0.06889595091342926,
0.0019842523615807295,
0.05247049033641815,
0.014361980371177197,
0.15638838708400726,
0.014963909052312374,
0.10756763815879822,
0.06746349483728409,
-0.3060946762561798,
-0.016793126240372658,
0.0632706880569458,
0.0005044561112299562,
0.05151284113526344,
-0.10005775839090347,
0.034433525055646896,
0.07906332612037659,
-0.0018709511496126652,
-0.013944712467491627,
-0.05109769105911255,
-0.16325324773788452,
-0.13154728710651398,
-0.15455617010593414,
0.020048018544912338,
0.1615273803472519,
0.06756851077079773,
-0.15296611189842224,
-0.0036448815371841192,
-0.0921899750828743,
0.11247863620519638,
-0.07245474308729172,
-0.030078299343585968,
0.05695165693759918,
0.004312776029109955,
0.05671500042080879,
-0.15289084613323212,
-0.05636591464281082,
-0.03280412405729294,
-0.14040012657642365,
0.2597835958003998,
0.023957716301083565,
0.05680810287594795,
-0.17015188932418823,
-0.017475269734859467,
-0.24272215366363525,
-0.10829753428697586,
-0.023753324523568153,
-0.017759334295988083,
-0.0337747298181057,
-0.04777983948588371,
-0.03709016740322113,
-0.16166092455387115,
0.037866946309804916,
0.08776485174894333,
-0.009025071747601032,
0.03207899257540703,
-0.14066019654273987,
-0.0004649602633435279,
0.11482289433479309,
0.14889909327030182,
-0.13372911512851715,
0.036761555820703506,
0.07271016389131546,
-0.06962505728006363,
-0.010313623584806919,
-0.08230786770582199,
-0.11363321542739868,
-0.005090500693768263,
-0.01677843928337097,
-0.03371750935912132,
0.09603725373744965,
0.036018308252096176,
0.034940753132104874,
-0.12283161282539368,
0.07179775834083557,
-0.012629019096493721,
-0.014147167094051838,
0.023526383563876152,
-0.01951712556183338,
0.09080764651298523,
-0.024648291990160942,
0.027302417904138565,
0.030223330482840538,
-0.04775369539856911,
-0.09300864487886429,
-0.07665775716304779,
-0.024584228172898293,
-0.14264360070228577,
0.03120606578886509,
-0.10351429879665375,
0.06705235689878464,
-0.15393224358558655,
-0.054496102035045624,
0.03466586023569107,
0.020642615854740143,
-0.0577913373708725,
0.017739761620759964,
-0.003958283923566341,
-0.03473559021949768,
-0.01308081392198801,
0.027161503210663795,
0.0581749826669693,
0.005649597849696875,
-0.0018207937246188521,
0.012148845940828323,
0.14181965589523315,
-0.23937247693538666,
0.026091206818819046,
-0.10291772335767746,
0.004745003767311573,
-0.003658929141238332,
0.0057661160826683044,
-0.08679087460041046,
0.058457180857658386,
-0.08015996962785721,
-0.03379536047577858,
-0.0804169550538063,
-0.0180217195302248,
0.12273122370243073,
0.13313791155815125,
-0.09663748741149902,
0.019894372671842575,
0.1615837663412094,
-0.014379591681063175,
-0.1262243539094925,
0.16671009361743927,
-0.01153128407895565,
0.14269277453422546,
0.04394816234707832,
0.17448028922080994,
0.0479351207613945,
-0.17306436598300934,
0.030955808237195015,
0.1050533577799797,
0.021755719557404518,
0.018549026921391487,
0.04791327565908432,
0.008762981742620468,
-0.1561865210533142,
0.055373869836330414,
0.006283394061028957,
0.0026232111267745495,
-0.037213440984487534,
-0.03140614554286003,
0.0003727525472640991,
-0.08548765629529953,
0.050287600606679916,
-0.03548461198806763,
0.08288926631212234,
-0.0276725422590971,
0.01575062796473503,
0.03295953944325447,
0.08732297271490097,
-0.02308575250208378,
0.000691165798343718,
-0.10308261960744858,
0.18039949238300323,
-0.16414792835712433,
0.02133241295814514,
-0.15227629244327545,
-0.04978014901280403,
0.0894898921251297,
-0.004123445134609938,
0.08376184105873108,
0.11041991412639618,
0.04186755418777466,
-0.008885628543794155,
0.043780189007520676,
0.05293025076389313,
0.005988907068967819,
0.014346200972795486,
-0.07638774067163467,
-0.07616273313760757,
0.016649385914206505,
-0.10519211739301682,
-0.05945342779159546,
-0.04740897938609123,
0.014994247816503048,
-0.02602485753595829,
0.020561067387461662,
0.02997894585132599,
-0.004534278996288776,
0.04904527589678764,
-0.012560761533677578,
-0.026142749935388565,
-0.002674145856872201,
0.058245185762643814,
0.08792225271463394,
-0.013339516706764698,
-0.02265886403620243,
-0.052460022270679474,
0.09347058087587357,
0.06460930407047272,
-0.09243997931480408,
-0.015348950400948524,
0.05528056249022484,
-0.009467357769608498,
0.03510623425245285,
-0.05137578770518303,
0.00017237824795302004,
0.17329169809818268,
-0.015210338868200779,
0.08845843374729156,
-0.11921944469213486,
0.053575120866298676,
0.07601954787969589,
-0.06228712573647499,
-0.004851944744586945,
0.07030943781137466,
0.034157462418079376,
-0.17668487131595612,
0.023647677153348923,
0.06112184002995491,
-0.03182065486907959,
0.12346221506595612,
0.005545340478420258,
-0.06819280236959457,
-0.026918012648820877,
0.04271320253610611,
0.0029947077855467796,
0.08666583150625229,
-0.2155003547668457,
0.00824462529271841,
0.06665030121803284,
0.04483315721154213,
0.02503976784646511,
-0.048558760434389114,
0.04494141414761543,
0.10733351111412048,
0.05143139511346817,
-0.07077942788600922,
0.013597879558801651,
-0.02970936894416809,
0.10317056626081467,
0.048871178179979324,
-0.05123940110206604,
-0.026284074410796165,
-0.015261894091963768,
-0.0872906893491745,
0.20713987946510315,
-0.05777066573500633,
-0.1556907743215561,
-0.10269741714000702,
-0.1589842140674591,
-0.07680417597293854,
-0.0021211018320173025,
0.00750895868986845,
-0.05508846044540405,
0.0003042339230887592,
-0.0272236168384552,
-0.0014132584910839796,
-0.05202430114150047,
0.0597461462020874,
0.08215906471014023,
-0.05154566839337349,
0.02782249078154564,
-0.10472849011421204,
-0.024025823920965195,
-0.080457404255867,
-0.08937318623065948,
0.023349428549408913,
-0.04679851979017258,
0.09589521586894989,
0.08750124275684357,
-0.08272520452737808,
-0.052184492349624634,
0.007582220248878002,
0.3274463713169098,
-0.014244859106838703,
0.04544536769390106,
0.12112483382225037,
-0.004648344125598669,
0.016020236536860466,
0.17043371498584747,
0.0450800284743309,
-0.12763138115406036,
0.03386526182293892,
-0.005719524342566729,
-0.08636879175901413,
-0.12953504920005798,
-0.04754255339503288,
-0.08551400899887085,
0.0877792164683342,
0.04941130802035332,
0.001107456162571907,
0.08232469111680984,
0.06604129076004028,
0.05497153848409653,
0.05206630006432533,
-0.09835918247699738,
0.10115150362253189,
0.08361556380987167,
-0.028709806501865387,
0.11377358436584473,
0.0021818524692207575,
-0.03066660836338997,
0.10182791203260422,
0.03031538985669613,
0.020601732656359673,
0.03276166319847107,
0.16311025619506836,
0.0014592442894354463,
0.06340377032756805,
0.04824773594737053,
0.1663055270910263,
-0.15761904418468475,
-0.058802101761102676,
-0.03830595687031746,
-0.04210988059639931,
-0.04466813802719116,
0.03225259482860565,
0.04151706025004387,
0.02829493023455143,
-0.07744835317134857,
-0.0199196245521307,
-0.03620592877268791,
0.07842057198286057,
0.12899592518806458,
-0.13634836673736572,
-0.05715083330869675,
-0.01861218549311161,
0.034746237099170685,
-0.13240276277065277,
0.03131665661931038,
0.08320886641740799,
-0.08085843920707703,
-0.06970912218093872,
-0.0153385354205966,
0.10658873617649078,
0.010047422721982002,
0.025700177997350693,
-0.1493367999792099,
0.052511777728796005,
-0.033872175961732864,
0.07891253381967545,
-0.15790987014770508,
0.1396457701921463,
0.022281985729932785,
0.020478630438447,
-0.03412022814154625,
-0.01687893643975258,
-0.011380866169929504,
0.011241796426475048,
0.17562183737754822,
-0.023816656321287155,
0.05958816036581993,
-0.13411203026771545,
-0.13372457027435303,
0.02184283919632435,
0.02770172618329525,
0.03646133467555046,
0.06924006342887878,
0.029942113906145096,
0.02722053788602352,
-0.0030735379550606012,
0.24866455793380737,
-0.045484770089387894,
-0.06745841354131699,
0.03483855724334717,
0.05574829876422882,
-0.046840015798807144,
0.00009734297782415524,
-0.03125393018126488,
-0.009417729452252388,
0.1296296864748001,
0.03879527747631073,
-0.042915068566799164,
-0.1407000571489334,
-0.0047381846234202385,
0.038963403552770615,
-0.09050267189741135,
0.027281275019049644,
-0.10577286779880524,
0.14551162719726562,
-0.04240208864212036,
-0.1561836451292038,
0.11945781111717224,
-0.13458162546157837,
0.0015835198573768139,
-0.033259302377700806,
0.015477131120860577,
0.040402356535196304,
0.089909128844738,
0.07230348885059357,
-0.027615508064627647,
-0.02600506879389286,
-0.07457125186920166,
0.028086086735129356,
0.07596690207719803,
-0.06312347948551178,
-0.03973410651087761,
-0.11758698523044586,
0.03685060888528824,
-0.05129470303654671,
0.13829688727855682,
0.24736188352108002,
0.17406652867794037,
-0.04156654700636864,
0.06337476521730423,
0.2445685714483261,
-0.05505984649062157,
-0.21483206748962402,
0.048839543014764786,
0.04739982634782791,
0.04726404696702957,
-0.017634347081184387,
-0.21605803072452545,
0.17041385173797607,
-0.06142748147249222,
-0.048922039568424225,
0.0898393839597702,
-0.0738300085067749,
-0.07548105716705322,
0.23948882520198822,
0.07025916129350662,
0.33915647864341736,
-0.0742739737033844,
-0.016166679561138153,
0.04024586081504822,
-0.1281513124704361,
0.18892818689346313,
-0.0065163071267306805,
0.10916122049093246,
0.013579978607594967,
0.06817204505205154,
0.06582114845514297,
-0.009410790167748928,
0.0757855698466301,
-0.03980747610330582,
0.00841399934142828,
0.0016532479785382748,
-0.04163435474038124,
-0.003451957833021879,
-0.0252241138368845,
0.10601261258125305,
0.13366448879241943,
0.05567953735589981,
-0.14729906618595123,
-0.04953628033399582,
-0.0712357684969902,
0.06555864214897156,
-0.007436869200319052,
-0.11853538453578949,
-0.020717348903417587,
0.04043733328580856,
0.09464515000581741,
0.043654780834913254,
0.0951307937502861,
-0.024100495502352715,
-0.0007255222299136221,
0.22350166738033295,
0.12365326285362244,
-0.05227310210466385,
0.0391346774995327,
0.004151081200689077,
-0.040551040321588516,
0.09310755133628845,
-0.02670983411371708,
0.015873728320002556,
0.092442587018013,
0.07170441001653671,
0.020519500598311424,
0.046688877046108246,
-0.07004077732563019,
-0.004816936794668436,
-0.02125323750078678,
-0.055477794259786606,
-0.08918433636426926,
-0.014840583316981792,
0.2429267317056656,
-0.11480171978473663,
0.051668744534254074,
0.11962788552045822,
-0.028478141874074936,
-0.03440031409263611,
-0.0027457023970782757,
-0.02675378881394863,
0.0225231871008873,
0.023339463397860527,
0.10859394818544388,
0.031601451337337494,
-0.07818782329559326,
0.05109288543462753,
0.010184012353420258,
0.013562643900513649,
0.08348558098077774,
-0.066065214574337,
-0.05826956778764725,
-0.015491639263927937,
0.020945817232131958,
0.016389263793826103,
-0.06700687855482101,
0.003017030656337738,
-0.0006876110564917326,
-0.05254029110074043,
-0.007559326011687517,
0.13862952589988708,
0.07704140990972519,
-0.07152226567268372,
-0.09998075664043427,
0.012899833731353283,
-0.09271398186683655,
0.04667961224913597,
-0.1287296712398529,
0.03946306183934212,
-0.030585946515202522,
-0.03027198277413845,
-0.08072653412818909,
0.06410790234804153,
-0.07442260533571243,
-0.10820203274488449,
-0.13449165225028992,
-0.04787980392575264,
-0.15576185286045074,
0.008017323911190033,
-0.03298616781830788,
-0.018907887861132622,
-0.02033948339521885,
-0.07805383205413818,
-0.027325429022312164,
0.008516422472894192,
-0.08476642519235611,
-0.011686799116432667,
0.0008062786073423922,
-0.024488212540745735,
-0.05064886063337326,
-0.026923712342977524,
0.01708029769361019,
-0.035804081708192825,
0.011826877482235432,
0.024807482957839966,
0.0017592842923477292,
0.061193689703941345,
-0.14456707239151,
-0.08106749504804611,
-0.018505023792386055,
0.0385168232023716,
0.09178707003593445,
-0.10041933506727219,
-0.004235782194882631,
-0.024931753054261208,
-0.036577191203832626,
-0.05738586559891701,
0.012785660102963448,
-0.06460556387901306,
-0.03697697073221207,
-0.020060451701283455,
-0.01476005744189024,
0.012711812742054462,
-0.01886194758117199,
-0.015582532621920109,
0.028795044869184494,
0.04030134156346321,
-0.011375529691576958,
0.09187716990709305,
-0.06905839592218399,
-0.008308163844048977,
-0.0040319254621863365,
-0.022684989497065544,
-0.0005294859292916954,
-0.09740803390741348,
0.026323631405830383,
-0.02848748490214348,
0.09488540142774582,
-0.048180289566516876,
-0.008822846226394176,
-0.01157795637845993,
-0.04185021296143532,
0.094127357006073,
-0.028955647721886635,
0.2835003733634949,
0.0800766721367836,
-0.020646218210458755,
0.03552587330341339,
0.08589334785938263,
0.05608776584267616,
0.2625083923339844,
0.14081954956054688,
0.0683225616812706,
-0.0325823649764061,
0.044365767389535904,
0.033361319452524185,
0.06998566538095474,
-0.07046277821063995,
-0.022818515077233315,
0.042977526783943176,
0.06927956640720367,
0.000058839719713432714,
0.0013682619901373982,
0.04622013121843338,
-0.05325394123792648,
0.05253298953175545,
0.09275755286216736,
-0.05380875617265701,
-0.16340942680835724,
-0.1338375210762024,
-0.07268428057432175,
-0.041159238666296005,
-0.010157468728721142,
-0.1722930371761322,
-0.05826283246278763,
-0.035702962428331375,
0.07281556725502014,
-0.013272258453071117,
0.15941943228244781,
-0.1872955858707428,
-0.030229991301894188,
0.11736880987882614,
-0.015091564506292343,
-0.007988102734088898,
-0.023494085296988487,
-0.00007119090150808915,
-0.009734558872878551,
0.07767844945192337,
0.0486619807779789,
0.02594592608511448,
0.012416627258062363,
0.007351444568485022,
0.023809188976883888,
-0.05695907399058342,
-0.025410061702132225,
-0.05432168394327164,
-0.010789959691464901,
0.1275278478860855,
0.03587363660335541,
-0.07298656553030014,
-0.008371629752218723,
0.17593660950660706,
-0.00804199930280447,
0.07354549318552017,
-0.12295680493116379,
0.14949168264865875,
-0.08540219813585281,
0.08898355811834335,
-0.021586479619145393,
-0.03425027057528496,
-0.07411883771419525,
0.3274802267551422,
0.07605141401290894,
-0.09776713699102402,
-0.015505651943385601,
0.08584843575954437,
0.0013300332939252257,
0.0005514256772585213,
0.0757942721247673,
0.024975912645459175,
0.21514742076396942,
-0.06002582982182503,
-0.03909169137477875,
-0.015343102626502514,
-0.0665598064661026,
-0.03181793913245201,
-0.012502867728471756,
0.09054936468601227,
-0.020883794873952866,
-0.09940221905708313,
0.014530174434185028,
-0.13400058448314667,
-0.16645163297653198,
-0.1086176335811615,
-0.06603223085403442,
-0.10361761599779129,
-0.06255240738391876,
-0.2038576453924179,
0.09630179405212402,
0.08827339112758636,
-0.06776507943868637,
0.03989075496792793,
0.0655156597495079,
0.007651630323380232,
-0.023899734020233154,
-0.04651361331343651,
0.08373789489269257,
-0.01824597828090191,
0.17277993261814117,
-0.004509704187512398,
-0.01495405100286007,
0.09117979556322098,
-0.020678460597991943,
-0.07861451059579849,
-0.040989767760038376,
0.023374218493700027,
-0.059311166405677795,
-0.029337741434574127,
0.07658671587705612,
-0.03082948736846447,
0.04775981605052948,
0.06551356613636017,
-0.16003666818141937,
-0.004777503665536642,
-0.036662064492702484,
-0.01266716979444027,
-0.09234660118818283,
0.04159853607416153,
-0.13788911700248718,
0.13708160817623138,
0.17431773245334625,
-0.07626241445541382,
-0.08930552750825882,
-0.021451715379953384,
0.045406218618154526,
0.05862421169877052,
-0.08553203195333481,
-0.08365951478481293,
-0.13583655655384064,
-0.00939198024570942,
-0.07474732398986816,
0.004639212507754564,
-0.3035403788089752,
-0.032399099320173264,
-0.016906453296542168,
0.015679076313972473,
0.0052688163705170155,
-0.0357295460999012,
0.07813473790884018,
-0.008900143206119537,
-0.04160367324948311,
0.009970976039767265,
-0.04064496234059334,
0.06456803530454636,
-0.06294496357440948,
-0.10786662995815277
] |
null | null |
transformers
|
[anton-l/wav2vec2-base-ft-keyword-spotting](https://huggingface.co/anton-l/wav2vec2-base-ft-keyword-spotting) model quantized with [Optimum OpenVINO](https://github.com/dkurt/optimum-openvino/).
| Accuracy on eval (baseline) | Accuracy on eval (quantized) |
|-----------------------------|----------------------------------------|
| 0.9828 | 0.9553 (-0.0274) |
|
{}
|
audio-classification
|
dkurt/wav2vec2-base-ft-keyword-spotting-int8
|
[
"transformers",
"wav2vec2",
"audio-classification",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #wav2vec2 #audio-classification #endpoints_compatible #region-us
|
anton-l/wav2vec2-base-ft-keyword-spotting model quantized with Optimum OpenVINO.
|
[] |
[
"TAGS\n#transformers #wav2vec2 #audio-classification #endpoints_compatible #region-us \n"
] |
[
29
] |
[
"passage: TAGS\n#transformers #wav2vec2 #audio-classification #endpoints_compatible #region-us \n"
] |
[
-0.0625230073928833,
0.04716717451810837,
-0.006934971082955599,
-0.06424888223409653,
0.10519111156463623,
-0.0005949318874627352,
0.04220496118068695,
0.03174543008208275,
-0.017191637307405472,
-0.01265027467161417,
0.056414540857076645,
0.1530521661043167,
-0.01942133717238903,
-0.01777658611536026,
-0.12227856367826462,
-0.20231500267982483,
0.09333021938800812,
0.060779035091400146,
-0.07142780721187592,
0.05777192488312721,
0.05178344249725342,
-0.0459560751914978,
0.033917803317308426,
0.002054978860542178,
-0.17607265710830688,
0.04338689148426056,
0.09305394440889359,
-0.08433549106121063,
0.03998318687081337,
0.059754688292741776,
0.16345103085041046,
0.0391017384827137,
-0.01799352467060089,
-0.20697245001792908,
0.011754906736314297,
0.01090051606297493,
-0.0667383149266243,
0.003739246167242527,
0.1316424459218979,
-0.07950923591852188,
-0.1427871584892273,
0.05122508108615875,
-0.051992468535900116,
0.08443024754524231,
-0.10541687905788422,
-0.14083227515220642,
-0.03987127169966698,
-0.021650923416018486,
0.09907187521457672,
0.020167304202914238,
0.01720501482486725,
0.05618053674697876,
-0.09770228713750839,
0.09004698693752289,
0.14014559984207153,
-0.2573295533657074,
0.030547073110938072,
0.12420868128538132,
0.10005708038806915,
0.05534859746694565,
-0.03331170603632927,
0.12111987918615341,
0.09115095436573029,
-0.019487589597702026,
-0.06052573025226593,
-0.09454310685396194,
-0.13553836941719055,
0.023295460268855095,
-0.07564908266067505,
-0.057826828211545944,
0.21152585744857788,
0.019172636792063713,
0.007600626442581415,
0.043182045221328735,
-0.07394768297672272,
-0.16288575530052185,
-0.025515303015708923,
0.04374106600880623,
0.008189545944333076,
0.0810956209897995,
0.00595884770154953,
0.042579133063554764,
-0.08054465800523758,
-0.07517850399017334,
-0.17316687107086182,
0.2707844376564026,
-0.04714082181453705,
0.08067955821752548,
-0.1776493489742279,
-0.028547197580337524,
-0.1281086653470993,
-0.07565700262784958,
0.01725711114704609,
-0.05589832738041878,
-0.07968970388174057,
-0.035120416432619095,
-0.09371326863765717,
-0.04147233068943024,
0.07100270688533783,
0.03156379237771034,
0.001293605542741716,
0.06066912040114403,
-0.08895441889762878,
0.11644445359706879,
0.03934282064437866,
-0.007287291809916496,
0.01332168746739626,
-0.06869840621948242,
-0.005244518164545298,
-0.2285863310098648,
0.014340534806251526,
-0.07131969183683395,
-0.15873491764068604,
-0.0809493362903595,
-0.038640256971120834,
0.062312062829732895,
-0.008383671753108501,
-0.040930211544036865,
-0.11999522894620895,
-0.019148852676153183,
0.025710400193929672,
-0.03761295974254608,
0.027754204347729683,
0.016612457111477852,
0.1458057314157486,
0.1718960851430893,
-0.055564023554325104,
0.08222110569477081,
-0.012262816540896893,
0.11531113088130951,
0.005809910129755735,
-0.015826616436243057,
0.011390995234251022,
-0.032391976565122604,
0.09586454927921295,
-0.1366780549287796,
0.09625405818223953,
-0.11915449798107147,
-0.018525492399930954,
0.015430442988872528,
0.050222285091876984,
0.05148174986243248,
-0.05558061599731445,
0.0017634553369134665,
-0.027006644755601883,
0.08288181573152542,
-0.11175987869501114,
-0.02444283664226532,
-0.08621138334274292,
0.04060951992869377,
-0.030473120510578156,
0.10228028893470764,
-0.1784500777721405,
0.10580942034721375,
0.0010952906450256705,
0.07730016112327576,
-0.09645623713731766,
0.013673163019120693,
-0.06908231228590012,
0.058104828000068665,
-0.05156438797712326,
-0.06295248121023178,
-0.12209004908800125,
0.07560928910970688,
-0.0794752687215805,
0.031045591458678246,
-0.16408009827136993,
-0.1402195543050766,
0.16457466781139374,
-0.10353358089923859,
-0.08294959366321564,
0.09546331316232681,
0.038731928914785385,
-0.10535161197185516,
0.05248239263892174,
0.23098935186862946,
0.12614622712135315,
-0.10338710248470306,
0.0819101333618164,
0.11083510518074036,
-0.11841148138046265,
-0.105352982878685,
0.07302088290452957,
-0.022282076999545097,
-0.09168558567762375,
-0.03157505393028259,
0.07709342241287231,
0.06131666898727417,
-0.041254013776779175,
-0.03297306224703789,
-0.01792512647807598,
-0.05159372091293335,
0.07540813833475113,
-0.03174804896116257,
0.030317747965455055,
-0.026530159637331963,
0.04104358330368996,
0.07625135034322739,
0.020473385229706764,
-0.006995614152401686,
0.07805773615837097,
-0.07468704879283905,
0.11424482613801956,
-0.16613565385341644,
-0.014231651090085506,
-0.1851872354745865,
0.09324643015861511,
-0.03943299874663353,
0.06072622910141945,
0.01268757414072752,
0.18542243540287018,
0.05229249969124794,
-0.09491995722055435,
-0.009955252520740032,
-0.06594838201999664,
0.11022865772247314,
0.06919902563095093,
0.024154245853424072,
-0.12701746821403503,
0.03754275292158127,
-0.075185127556324,
-0.1557791382074356,
-0.010607988573610783,
-0.08999999612569809,
0.21245667338371277,
0.1268645077943802,
0.07256612181663513,
0.026222482323646545,
0.025648629292845726,
0.055122166872024536,
-0.032356008887290955,
0.037410635501146317,
0.06352555006742477,
-0.01616865023970604,
-0.0712093636393547,
0.22042399644851685,
-0.08953455090522766,
0.3364261984825134,
0.200521782040596,
-0.2892254590988159,
0.058210425078868866,
0.059653330594301224,
0.04633456841111183,
-0.005481097847223282,
0.005450459197163582,
-0.030619874596595764,
0.10137541592121124,
-0.010088699869811535,
0.12328162044286728,
-0.053144995123147964,
0.011446126736700535,
0.06949792802333832,
-0.055022530257701874,
-0.050136420875787735,
0.017242368310689926,
-0.04359189420938492,
-0.1530221551656723,
0.14792779088020325,
0.28761160373687744,
-0.013638557866215706,
0.2405707836151123,
-0.05610578507184982,
-0.055310290306806564,
0.04945247992873192,
-0.034793343394994736,
-0.031226899474859238,
0.018977519124746323,
-0.21631379425525665,
-0.012095034122467041,
0.06831617653369904,
0.09881208091974258,
0.05021864175796509,
-0.13560256361961365,
-0.016774769872426987,
0.05067657679319382,
-0.022056475281715393,
-0.18306846916675568,
0.05129992216825485,
0.007678365334868431,
0.033737365156412125,
-0.013480213470757008,
-0.04448822885751724,
0.13908614218235016,
-0.05897267162799835,
-0.09050260484218597,
0.03766067326068878,
-0.18046565353870392,
-0.19342413544654846,
-0.21673321723937988,
-0.20882274210453033,
0.012262205593287945,
0.08747423440217972,
0.11303180456161499,
-0.09913229942321777,
-0.04496866464614868,
0.10987449437379837,
0.06737320125102997,
-0.12645897269248962,
0.06149197369813919,
0.09427628666162491,
0.11470001190900803,
-0.001869436469860375,
-0.04177340865135193,
0.005887545645236969,
0.03464365750551224,
0.055523451417684555,
0.06426697969436646,
-0.06490909308195114,
0.07165778428316116,
0.2071727067232132,
0.07737220823764801,
0.02745291031897068,
0.018229950219392776,
0.14451086521148682,
-0.14558762311935425,
-0.05123436450958252,
0.13442577421665192,
-0.1725049614906311,
0.014455769211053848,
0.1991332620382309,
0.06351777911186218,
-0.054650481790304184,
-0.02342098020017147,
-0.037797775119543076,
-0.12194903939962387,
-0.27180102467536926,
-0.11580748856067657,
-0.15762865543365479,
0.00712785916402936,
-0.04292017221450806,
0.023111961781978607,
0.09001080691814423,
0.018623843789100647,
0.02040407992899418,
0.025307225063443184,
0.028347328305244446,
0.007546499837189913,
0.127278134226799,
-0.009191825985908508,
0.0769021287560463,
-0.09855165332555771,
-0.04610253870487213,
0.08399700373411179,
0.14282891154289246,
0.12942074239253998,
0.2537462115287781,
0.10465400665998459,
0.039293043315410614,
-0.023532742634415627,
0.23215080797672272,
0.13120219111442566,
0.07262487709522247,
-0.030591841787099838,
0.01432508509606123,
-0.02740633673965931,
-0.019274046644568443,
0.06294503062963486,
0.17538504302501678,
-0.1415630728006363,
-0.034423213452100754,
-0.14940127730369568,
0.007263236679136753,
0.11669469624757767,
0.11975699663162231,
-0.15749573707580566,
0.02025413140654564,
0.15057295560836792,
-0.029283521696925163,
-0.09189530462026596,
0.1334901750087738,
0.0897907018661499,
-0.034611042588949203,
0.1420869082212448,
0.04934975877404213,
0.12232715636491776,
-0.0787753164768219,
0.04265429824590683,
0.0033753179013729095,
-0.14642848074436188,
0.04937256500124931,
0.034890785813331604,
-0.14023788273334503,
0.23707804083824158,
0.03362317755818367,
-0.009430286474525928,
0.016972271725535393,
-0.01254192739725113,
-0.022825434803962708,
0.1960023194551468,
0.17694668471813202,
0.04040410369634628,
-0.224951833486557,
-0.11100726574659348,
0.03039400838315487,
-0.03333865851163864,
0.1788739264011383,
0.08536214381456375,
-0.08172321319580078,
-0.03259974345564842,
-0.04697764664888382,
0.008083149790763855,
-0.043200407177209854,
-0.037338703870773315,
-0.08904901146888733,
0.024211175739765167,
0.1967327892780304,
0.14468546211719513,
-0.012960687279701233,
-0.026596831157803535,
-0.17785657942295074,
0.1343151181936264,
-0.20473428070545197,
0.010414770804345608,
-0.08693086355924606,
-0.206435889005661,
0.10254725068807602,
0.026010850444436073,
0.1212189570069313,
-0.025928307324647903,
0.08781048655509949,
-0.0718567743897438,
-0.15570054948329926,
0.1370496302843094,
-0.12013688683509827,
-0.010973100550472736,
-0.07273487746715546,
0.2397771030664444,
-0.05341096967458725,
0.04471404105424881,
0.020743515342473984,
0.049111876636743546,
0.03455718606710434,
-0.05054408311843872,
0.08876816928386688,
0.03362089395523071,
-0.023852886632084846,
0.07270384579896927,
0.029164621606469154,
-0.08993661403656006,
0.0795106366276741,
-0.03869146108627319,
0.20559148490428925,
0.17737692594528198,
-0.04327866807579994,
0.21619267761707306,
0.14993001520633698,
-0.050568461418151855,
-0.3380752205848694,
0.007449849043041468,
-0.03769349679350853,
0.010823546908795834,
-0.03687106817960739,
-0.12686529755592346,
0.1456812024116516,
-0.06233624368906021,
-0.04166041687130928,
0.08941618353128433,
-0.1320100724697113,
-0.047460492700338364,
0.12432204186916351,
-0.08689111471176147,
0.42699670791625977,
-0.046787671744823456,
-0.13714371621608734,
-0.12571647763252258,
-0.31918323040008545,
0.12294638901948929,
-0.1427668184041977,
0.08869738131761551,
0.07279343158006668,
0.0407363623380661,
0.026255594566464424,
-0.03906606510281563,
0.12275828421115875,
0.08110092580318451,
0.013596358709037304,
-0.03081534057855606,
-0.004165641497820616,
0.04725615680217743,
0.007144726812839508,
-0.12097512930631638,
0.06484606117010117,
-0.01188361831009388,
-0.11202181875705719,
-0.04850652441382408,
-0.04374265670776367,
0.030509456992149353,
0.03424646332859993,
-0.008197728544473648,
0.035725198686122894,
-0.06932833790779114,
0.011996503919363022,
-0.023715684190392494,
0.3583921492099762,
-0.06382708996534348,
0.0283628161996603,
0.15716077387332916,
0.15234433114528656,
-0.13537931442260742,
-0.07347003370523453,
-0.03153699263930321,
-0.0878114402294159,
0.13167043030261993,
-0.13144107162952423,
0.15528181195259094,
0.08414578437805176,
-0.008449329994618893,
-0.008748977445065975,
0.12757962942123413,
0.0067544616758823395,
-0.02559351734817028,
0.15737861394882202,
-0.10848870128393173,
0.00040023887413553894,
-0.050554320216178894,
-0.0010453929426148534,
0.1756363958120346,
0.07436082512140274,
0.149802103638649,
0.0633430927991867,
0.022627631202340126,
0.013130603358149529,
-0.030272498726844788,
-0.20079810917377472,
0.07710463553667068,
0.025025000795722008,
0.050159767270088196,
-0.1644345223903656,
0.11720726639032364,
-0.04531238600611687,
-0.2365407794713974,
-0.03641146421432495,
-0.03165573254227638,
-0.10953685641288757,
-0.06943503767251968,
0.004146335180848837,
0.04014994204044342,
-0.07496137917041779,
-0.12512809038162231,
0.0817023292183876,
-0.1522744745016098,
0.039879754185676575,
0.23544201254844666,
0.05478637292981148,
0.13294124603271484,
-0.025528978556394577,
-0.03313140571117401,
0.03496037796139717,
-0.04928893595933914,
-0.06393147259950638,
0.06263411790132523,
-0.2545660734176636,
-0.050325751304626465,
-0.06827209889888763,
0.09987864643335342,
-0.11203327029943466,
-0.027880365028977394,
-0.123002789914608,
0.09334711730480194,
-0.13366980850696564,
-0.0015889483038336039,
-0.08428411930799484,
-0.029407722875475883,
0.018071390688419342,
-0.03772687911987305,
-0.0748843252658844,
0.03169475495815277,
-0.1105121448636055,
0.05253833532333374,
-0.008135741576552391,
0.029538368806242943,
-0.006931501906365156,
-0.02059561386704445,
0.08853209018707275,
-0.06062572821974754,
0.04753348231315613,
0.14473655819892883,
-0.0826704129576683,
0.06589335203170776,
-0.26250800490379333,
-0.20404620468616486,
0.17671465873718262,
0.0004956984193995595,
0.039707474410533905,
-0.036253124475479126,
0.010064907371997833,
0.09494756162166595,
0.015609994530677795,
0.010565254837274551,
0.04184813052415848,
-0.04652690514922142,
-0.06323794275522232,
-0.13613703846931458,
-0.11583395302295685,
-0.03541780263185501,
-0.07617650926113129,
0.25149005651474,
0.02187388762831688,
0.09948144853115082,
0.028028517961502075,
0.06823457777500153,
0.05512386932969093,
0.05269891768693924,
-0.0067313434556126595,
-0.18234433233737946,
-0.075092613697052,
-0.03913943096995354,
-0.0248288344591856,
-0.055099137127399445,
0.25906530022621155,
-0.09048677980899811,
-0.02454270049929619,
0.03743046522140503,
-0.008497867733240128,
-0.07969342172145844,
0.07011280208826065,
0.32459986209869385,
0.10825031250715256,
-0.04588266834616661,
-0.15489903092384338,
0.007684560492634773,
0.03751102089881897,
0.07144129276275635,
-0.021765299141407013,
0.13672053813934326,
-0.03372180461883545,
0.08965098112821579,
0.05121579393744469,
-0.01937584951519966,
-0.16815045475959778,
-0.008233224041759968,
-0.04523259028792381,
0.07514311373233795,
0.05625832825899124,
0.028828974813222885,
0.10414764285087585,
-0.05173902586102486,
0.08563636988401413,
-0.0520186610519886,
-0.011313753202557564,
-0.2005111277103424,
-0.08653247356414795,
-0.09934751689434052,
-0.17478755116462708,
0.04278188198804855,
-0.07596740126609802,
0.11637506633996964,
0.0534353144466877,
0.0347113274037838,
0.025631064549088478,
0.12271436303853989,
-0.08643081784248352,
-0.05251699686050415,
0.08209510147571564,
-0.02263268455862999,
-0.007047292310744524,
0.09088001400232315,
-0.044051408767700195,
0.05953408405184746,
-0.11125491559505463,
-0.05592244490981102,
0.010820151306688786,
-0.04350721836090088,
0.03598342835903168,
-0.09980160742998123,
-0.10123591870069504,
-0.014845735393464565,
0.03246070817112923,
-0.06378614902496338,
0.14502570033073425,
0.017716914415359497,
0.0064722406677901745,
0.06018994748592377,
0.12154210358858109,
-0.10962063819169998,
-0.016022831201553345,
-0.059062350541353226,
-0.02946542203426361,
-0.005124413874000311,
0.20531108975410461,
-0.02751098945736885,
-0.019666602835059166,
-0.08722035586833954,
0.24800026416778564,
0.15860366821289062,
-0.0005850918241776526,
0.04538511484861374,
0.03827511891722679,
0.03602607920765877,
0.03741224482655525,
0.039086535573005676,
0.07501468062400818,
0.19936607778072357,
-0.01168705802410841,
-0.13008351624011993,
-0.04844779893755913,
-0.01935773901641369,
-0.008228272199630737,
0.09172602742910385,
-0.0266420841217041,
-0.155256986618042,
-0.0037025120109319687,
0.1354403793811798,
-0.24538271129131317,
0.0890071839094162,
0.04301080480217934,
-0.1517634242773056,
-0.004769382067024708,
-0.05162302032113075,
0.1969681680202484,
0.06450209766626358,
0.02392873913049698,
-0.06828676164150238,
-0.17942562699317932,
0.049172915518283844,
0.04284351319074631,
-0.21610046923160553,
0.019839376211166382,
-0.04903493821620941,
-0.03457944095134735,
-0.06426992267370224,
0.006691686809062958,
0.1060548722743988,
-0.029110915958881378,
0.14071235060691833,
0.01886770687997341,
0.07479645311832428,
0.0008628923678770661,
-0.19936659932136536,
-0.086747907102108,
0.24148106575012207,
-0.09562105685472488,
0.04640765115618706,
0.04408435523509979,
-0.16731184720993042,
0.03139311820268631,
0.021128373220562935,
-0.07181258499622345,
-0.011613852344453335,
-0.08053287118673325,
-0.09057052433490753,
0.04232599213719368,
-0.06241154670715332,
0.02286856807768345,
-0.0369667112827301,
-0.006684072315692902,
0.013438069261610508,
0.06731323152780533,
-0.037504423409700394,
-0.05791222304105759,
-0.10236695408821106,
-0.033513959497213364,
0.08433951437473297,
0.038605280220508575,
-0.05527479946613312,
0.001689580618403852,
-0.08462121337652206,
0.07247984409332275,
-0.11442272365093231,
0.022792238742113113,
0.08945544809103012,
-0.0038482912350445986,
-0.016999877989292145,
-0.13094843924045563,
0.11092549562454224,
0.04536721482872963,
-0.07291935384273529,
-0.1501188576221466
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2161
- Accuracy: 0.926
- F1: 0.9261
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8436 | 1.0 | 250 | 0.3175 | 0.9105 | 0.9081 |
| 0.2492 | 2.0 | 500 | 0.2161 | 0.926 | 0.9261 |
### Framework versions
- Transformers 4.15.0
- Pytorch 1.7.1
- Datasets 1.17.0
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.926, "name": "Accuracy"}, {"type": "f1", "value": 0.9261144741040841, "name": "F1"}]}]}]}
|
text-classification
|
dmiller1/distilbert-base-uncased-finetuned-emotion
|
[
"transformers",
"pytorch",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:emotion",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
distilbert-base-uncased-finetuned-emotion
=========================================
This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2161
* Accuracy: 0.926
* F1: 0.9261
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.15.0
* Pytorch 1.7.1
* Datasets 1.17.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.7.1\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.7.1\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
[
63,
98,
4,
30
] |
[
"passage: TAGS\n#transformers #pytorch #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.7.1\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
[
-0.10603971779346466,
0.11091575771570206,
-0.0021383161656558514,
0.14028596878051758,
0.17581398785114288,
0.04564647376537323,
0.09457764774560928,
0.12425196170806885,
-0.09601864218711853,
0.016703220084309578,
0.11150266975164413,
0.17224165797233582,
0.01155333686619997,
0.11701202392578125,
-0.06546440720558167,
-0.2806626260280609,
-0.002741886768490076,
0.04016965627670288,
-0.014997949823737144,
0.13490347564220428,
0.10871674865484238,
-0.12151474505662918,
0.10751084238290787,
-0.00255730003118515,
-0.18539170920848846,
0.015432296320796013,
0.014241324737668037,
-0.04442121833562851,
0.1425447165966034,
0.02282148413360119,
0.085364431142807,
0.0036697546020150185,
0.08942383527755737,
-0.22695845365524292,
0.015994947403669357,
0.037263404577970505,
0.0033643916249275208,
0.08199036121368408,
0.03865158185362816,
-0.02967611700296402,
0.17404372990131378,
-0.07689636945724487,
0.04535704478621483,
0.025724101811647415,
-0.11607499420642853,
-0.24586515128612518,
-0.0772390216588974,
0.03679465502500534,
0.0661163404583931,
0.12438149005174637,
-0.020758911967277527,
0.14516042172908783,
-0.11144205927848816,
0.10010841488838196,
0.2340305596590042,
-0.22969664633274078,
-0.06599065661430359,
0.008362064138054848,
0.0010144120315089822,
0.04018674045801163,
-0.12423485517501831,
-0.035494543612003326,
0.05839160829782486,
0.048896145075559616,
0.11698002368211746,
-0.043451711535453796,
-0.1099354699254036,
0.01412345189601183,
-0.12126081436872482,
-0.035744234919548035,
0.18777070939540863,
0.07064781337976456,
-0.028353650122880936,
-0.036411065608263016,
-0.05482691898941994,
-0.138804629445076,
-0.02683020755648613,
-0.010575778782367706,
0.06159844994544983,
-0.012506102211773396,
-0.08305371552705765,
0.029619919136166573,
-0.11279528588056564,
-0.034792330116033554,
-0.061749737709760666,
0.1259169578552246,
0.010050281882286072,
0.016360826790332794,
-0.011326956562697887,
0.09164641052484512,
-0.012339049018919468,
-0.12711448967456818,
0.006123220082372427,
0.014894542284309864,
0.03163699433207512,
-0.039949931204319,
-0.08468520641326904,
-0.02190151996910572,
-0.0008508544997312129,
0.1121300607919693,
-0.06868681311607361,
0.04879603907465935,
0.05472152680158615,
0.018611587584018707,
-0.05711885169148445,
0.20802181959152222,
-0.029698535799980164,
-0.046641863882541656,
-0.002739514224231243,
0.0507090799510479,
0.02518887259066105,
0.0018214789452031255,
-0.11538448184728622,
0.02236100658774376,
0.09799641370773315,
0.0003355195512995124,
-0.10632360726594925,
0.08350970596075058,
-0.08260130882263184,
-0.018455378711223602,
-0.021324530243873596,
-0.07701648026704788,
0.03626015782356262,
0.011925875209271908,
-0.08612365275621414,
0.010383888147771358,
0.01687670685350895,
0.009317989461123943,
-0.010603408329188824,
0.0811837688088417,
-0.07917572557926178,
0.03566509485244751,
-0.08869185298681259,
-0.09795431047677994,
0.025064853951334953,
-0.05657622218132019,
0.05122676491737366,
-0.09778185188770294,
-0.2135091871023178,
-0.022160133346915245,
0.06966762989759445,
-0.022798415273427963,
-0.05549260973930359,
-0.07315883785486221,
-0.06037714332342148,
0.017663778737187386,
-0.007747052237391472,
0.09184080362319946,
-0.07220417261123657,
0.080943264067173,
0.05028527230024338,
0.0889633372426033,
-0.04723459854722023,
0.052317120134830475,
-0.13313084840774536,
-0.0028532524593174458,
-0.15635284781455994,
0.06250052899122238,
-0.041091106832027435,
0.09516166150569916,
-0.06347862631082535,
-0.11853331327438354,
0.04327143356204033,
-0.010379358194768429,
0.060936007648706436,
0.11962486803531647,
-0.18645061552524567,
-0.09769577533006668,
0.15174449980258942,
-0.06553477048873901,
-0.11244761943817139,
0.12546351552009583,
-0.06938030570745468,
0.07295309007167816,
0.09402038902044296,
0.17483080923557281,
0.036745619028806686,
-0.050094202160835266,
-0.013202175498008728,
0.019058065488934517,
0.05145936459302902,
-0.02476552128791809,
0.05879414081573486,
0.0346350222826004,
0.007323103025555611,
0.0433344803750515,
-0.009591644629836082,
0.08049357682466507,
-0.0995972752571106,
-0.09795036911964417,
-0.04211515933275223,
-0.10240552574396133,
0.04466298222541809,
0.08845821768045425,
0.06388264149427414,
-0.1238168254494667,
-0.06859134882688522,
0.03900361806154251,
0.10608620941638947,
-0.05314832180738449,
0.027428770437836647,
-0.06019090116024017,
0.04426085203886032,
0.016007859259843826,
-0.014151125214993954,
-0.1760421246290207,
0.028928043320775032,
0.004477372858673334,
0.036024365574121475,
0.00867211539298296,
0.016438474878668785,
0.06536868959665298,
0.03845951706171036,
-0.06328321248292923,
-0.024834271520376205,
-0.03397999331355095,
0.0018147916998714209,
-0.10847576707601547,
-0.22724401950836182,
-0.012714640237390995,
-0.019734641537070274,
0.18393385410308838,
-0.20993570983409882,
0.03427578881382942,
-0.019108714535832405,
0.057794976979494095,
0.00579881202429533,
-0.013506821356713772,
-0.039334703236818314,
0.05988313630223274,
-0.051749806851148605,
-0.04469896852970123,
0.06962904334068298,
0.007329156622290611,
-0.09049807488918304,
-0.048028964549303055,
-0.09428657591342926,
0.154300719499588,
0.1328042447566986,
-0.12383842468261719,
-0.06824122369289398,
-0.002307557500898838,
-0.05789339169859886,
-0.010553529486060143,
-0.033419348299503326,
0.04859418049454689,
0.18329066038131714,
-0.016627026721835136,
0.13837289810180664,
-0.059307586401700974,
-0.01599777862429619,
0.015106257982552052,
-0.05369133502244949,
0.016527507454156876,
0.1270422488451004,
0.09182913601398468,
-0.09017640352249146,
0.14415405690670013,
0.1560303419828415,
-0.09312587231397629,
0.13878268003463745,
-0.03515499457716942,
-0.0503561906516552,
-0.02565087005496025,
-0.06318605691194534,
-0.030796481296420097,
0.08937333524227142,
-0.17417947947978973,
-0.011269926093518734,
0.02021988295018673,
0.010390895418822765,
-0.0047823782078921795,
-0.21490329504013062,
-0.05269576981663704,
0.047398120164871216,
-0.039405133575201035,
-0.021546605974435806,
-0.016388123854994774,
0.0033301557414233685,
0.10351313650608063,
0.004623989574611187,
-0.0938374474644661,
0.03719262033700943,
0.005261516198515892,
-0.0841241404414177,
0.2087915688753128,
-0.1145869791507721,
-0.16753512620925903,
-0.09868399053812027,
-0.08331847190856934,
-0.06520093977451324,
0.013620219193398952,
0.0778219997882843,
-0.12892797589302063,
-0.030636288225650787,
-0.07248299568891525,
0.04044148325920105,
0.005736408289521933,
0.02006050944328308,
0.04383794963359833,
-0.008863626047968864,
0.05951407924294472,
-0.10824443399906158,
-0.025576893240213394,
-0.05139719694852829,
-0.0556347593665123,
0.05710481107234955,
0.015555200167000294,
0.11224789917469025,
0.15989470481872559,
-0.002523286733776331,
0.011708308011293411,
-0.03703182563185692,
0.26576516032218933,
-0.06658372282981873,
-0.024961384013295174,
0.14099743962287903,
0.002197947818785906,
0.049068909138441086,
0.1191757321357727,
0.06367581337690353,
-0.10907359421253204,
0.023193011060357094,
0.04312477633357048,
-0.030761465430259705,
-0.20346322655677795,
-0.044190891087055206,
-0.04457715153694153,
0.006974087096750736,
0.06351308524608612,
0.021759670227766037,
0.0493418350815773,
0.07271015644073486,
0.03820814937353134,
0.047231581062078476,
-0.050464365631341934,
0.05580805614590645,
0.135330468416214,
0.023531196638941765,
0.1008402407169342,
-0.03245853632688522,
-0.05286228656768799,
0.06728524714708328,
-0.04162230342626572,
0.20007678866386414,
0.003204119158908725,
0.1293594241142273,
0.042153533548116684,
0.16939783096313477,
-0.04062821716070175,
0.07449132949113846,
-0.003219958860427141,
-0.044930655509233475,
-0.03763510286808014,
-0.024023540318012238,
-0.06998955458402634,
0.0409504659473896,
-0.073017917573452,
0.09374596178531647,
-0.15045595169067383,
0.008684205822646618,
0.07609669119119644,
0.27783069014549255,
0.03606448695063591,
-0.3312244713306427,
-0.11644741892814636,
0.012392399832606316,
-0.042956046760082245,
-0.006054659839719534,
0.018024753779172897,
0.056377120316028595,
-0.09796582162380219,
0.05542822554707527,
-0.040207818150520325,
0.09191261231899261,
-0.054883912205696106,
0.07450168579816818,
0.03670107573270798,
0.08524170517921448,
0.013684890232980251,
0.08857700973749161,
-0.2956719398498535,
0.26304951310157776,
-0.0056636217050254345,
0.0665648803114891,
-0.0850021243095398,
-0.0028812033124268055,
0.07087455689907074,
0.09956236183643341,
0.04615209624171257,
0.0016634475905448198,
-0.012330564670264721,
-0.1918523609638214,
-0.016941988840699196,
0.04208154231309891,
0.06572193652391434,
-0.02330920472741127,
0.08048347383737564,
-0.031581394374370575,
0.010041125118732452,
0.08275331556797028,
0.036509908735752106,
-0.06969418376684189,
-0.08733109384775162,
-0.021801846101880074,
0.04932555556297302,
-0.019960487261414528,
-0.05167285352945328,
-0.11757004261016846,
-0.09628508239984512,
0.15319614112377167,
0.0407792404294014,
-0.021017486229538918,
-0.10837193578481674,
0.088645800948143,
0.04641733691096306,
-0.0912051722407341,
0.018922219052910805,
0.006229551509022713,
0.07218004763126373,
0.04259928688406944,
-0.07979829609394073,
0.11480909585952759,
-0.07360149174928665,
-0.16672052443027496,
-0.056185152381658554,
0.07939375936985016,
0.05163417384028435,
0.07893016934394836,
0.0019478639587759972,
-0.007157434709370136,
-0.058313995599746704,
-0.08799513429403305,
0.025975387543439865,
0.028998514637351036,
0.038780562579631805,
0.03403802961111069,
-0.03258364647626877,
0.01684827171266079,
-0.07540155947208405,
-0.037037014961242676,
0.18913017213344574,
0.25003644824028015,
-0.08472591638565063,
0.019632510840892792,
0.03874598816037178,
-0.06644285470247269,
-0.18713876605033875,
0.056373465806245804,
0.051347121596336365,
0.002261195331811905,
0.03360980749130249,
-0.1917540580034256,
0.14479844272136688,
0.08826877176761627,
-0.007033384405076504,
0.07937711477279663,
-0.2769750654697418,
-0.11262599378824234,
0.1418912559747696,
0.12504053115844727,
0.16059467196464539,
-0.1398947685956955,
0.00969131849706173,
-0.06234828755259514,
-0.1128285601735115,
0.10510674864053726,
-0.09643538296222687,
0.1176256388425827,
-0.01995469257235527,
0.1305084377527237,
0.006764004472643137,
-0.02979312278330326,
0.13181623816490173,
0.03538179025053978,
0.11188879609107971,
-0.0751766636967659,
-0.006955294404178858,
0.020435459911823273,
-0.0414806567132473,
0.033627577126026154,
-0.10926137119531631,
0.0263055432587862,
-0.11910092085599899,
-0.027827991172671318,
-0.09692250192165375,
0.037301916629076004,
-0.028973344713449478,
-0.0720013976097107,
-0.05223897844552994,
0.0265151746571064,
0.08322486281394958,
-0.008874171413481236,
0.08636448532342911,
0.024304179474711418,
0.10049526393413544,
0.09517578035593033,
0.09236989915370941,
-0.08782069385051727,
-0.054337646812200546,
-0.013288723304867744,
-0.009875567629933357,
0.04919189587235451,
-0.1579388529062271,
0.02860010787844658,
0.12995600700378418,
0.011066959239542484,
0.17054878175258636,
0.08551519364118576,
-0.026216723024845123,
0.014847681857645512,
0.06533818691968918,
-0.14679881930351257,
-0.07356897741556168,
-0.023171599954366684,
-0.07448464632034302,
-0.12388072162866592,
0.030684655532240868,
0.08134777843952179,
-0.07231644541025162,
-0.004149588290601969,
-0.024559365585446358,
0.017976516857743263,
-0.060343317687511444,
0.16978426277637482,
0.04204145073890686,
0.021990422159433365,
-0.12004832923412323,
0.08079205453395844,
0.014593278989195824,
-0.0926484689116478,
0.02646586112678051,
0.07932933419942856,
-0.07960281521081924,
-0.05467142164707184,
0.09365055710077286,
0.2188282310962677,
-0.09400740265846252,
-0.05200004205107689,
-0.1440177857875824,
-0.1297975778579712,
0.08158574253320694,
0.1333475261926651,
0.12590426206588745,
0.007963919080793858,
-0.07689765840768814,
0.01928640343248844,
-0.1141607016324997,
0.06837984919548035,
0.06886240839958191,
0.03859817236661911,
-0.13936977088451385,
0.12111997604370117,
-0.0017770042177289724,
0.044166646897792816,
-0.025278164073824883,
0.006705276668071747,
-0.09224636107683182,
0.007617683615535498,
-0.1478312760591507,
-0.03316674754023552,
-0.048561327159404755,
0.021543793380260468,
0.012184176594018936,
-0.05206962674856186,
-0.04441964626312256,
0.006700620055198669,
-0.1230158805847168,
-0.022501550614833832,
0.038677699863910675,
0.07877112179994583,
-0.11593488603830338,
-0.04829998314380646,
0.025240518152713776,
-0.06367241591215134,
0.09167509526014328,
0.07161026448011398,
0.009088424034416676,
0.07334723323583603,
-0.18211513757705688,
0.0070374575443565845,
0.09954597055912018,
0.006248461548238993,
0.0661906972527504,
-0.07446860522031784,
-0.013331240974366665,
0.002744559198617935,
0.056619759649038315,
0.0158856064081192,
0.08017688989639282,
-0.1188291534781456,
0.0029055646155029535,
0.004056504927575588,
-0.09575572609901428,
-0.06291157752275467,
0.026257209479808807,
0.08722986280918121,
0.0053041367791593075,
0.20907048881053925,
-0.07879341393709183,
0.033103737980127335,
-0.21375875174999237,
0.005945196840912104,
-0.011789516545832157,
-0.11064606159925461,
-0.165224090218544,
-0.07816831022500992,
0.0624384768307209,
-0.049544379115104675,
0.13729578256607056,
0.04477301985025406,
0.032861437648534775,
0.01369787473231554,
0.0023004072718322277,
0.03179032355546951,
0.0022443262860178947,
0.20170177519321442,
0.046602219343185425,
-0.04905063286423683,
0.07138597220182419,
0.06364195793867111,
0.13063737750053406,
0.12537017464637756,
0.19232460856437683,
0.14374054968357086,
0.025350945070385933,
0.11716752499341965,
0.01702808402478695,
-0.03381594642996788,
-0.1428518295288086,
0.004827626049518585,
-0.045898474752902985,
0.11330296844244003,
-0.020341556519269943,
0.2261335700750351,
0.05241077393293381,
-0.15571282804012299,
0.06136666238307953,
-0.07866254448890686,
-0.08120977133512497,
-0.1086789220571518,
-0.04258907958865166,
-0.09145748615264893,
-0.17006780207157135,
-0.0002555669052526355,
-0.13853125274181366,
0.014879785478115082,
0.07760672271251678,
0.004929021932184696,
-0.04715309664607048,
0.11831574887037277,
0.006642214022576809,
0.01421890314668417,
0.08728398382663727,
-0.008663192391395569,
-0.0719611719250679,
-0.10478256642818451,
-0.0594470240175724,
-0.015501097775995731,
-0.01792794279754162,
0.04221852123737335,
-0.0497111976146698,
-0.08208545297384262,
0.027379393577575684,
-0.037394843995571136,
-0.10573529452085495,
0.015043145045638084,
0.014876846224069595,
0.056057848036289215,
0.04022105038166046,
0.000897528079804033,
0.018626922741532326,
0.005074784159660339,
0.20820774137973785,
-0.084068663418293,
-0.035384804010391235,
-0.11014315485954285,
0.2697247564792633,
0.023943236097693443,
-0.007295733317732811,
0.029193248599767685,
-0.07386752218008041,
-0.01861591450870037,
0.23915913701057434,
0.20645560324192047,
-0.09666665643453598,
-0.007650044746696949,
-0.029071038588881493,
0.002963821403682232,
-0.05326208099722862,
0.10341806709766388,
0.15122783184051514,
-0.017900537699460983,
-0.09838896989822388,
-0.0024856794625520706,
-0.0590595006942749,
-0.028609411790966988,
-0.029344217851758003,
0.03874948248267174,
0.061446916311979294,
0.008921442553400993,
-0.04028245806694031,
0.0551273338496685,
-0.09122370183467865,
-0.07789713144302368,
0.07091203331947327,
-0.20045219361782074,
-0.15006022155284882,
-0.02143607661128044,
0.09044869989156723,
0.05321449786424637,
0.08275394886732101,
-0.01461463887244463,
-0.0011733912397176027,
0.11178933829069138,
-0.02481737546622753,
-0.12548469007015228,
-0.0973447933793068,
0.11937344819307327,
-0.13428297638893127,
0.1869688630104065,
-0.062376413494348526,
0.04217923432588577,
0.12612681090831757,
0.06769375503063202,
-0.06559199094772339,
0.08007054775953293,
0.04103594273328781,
-0.05927738919854164,
0.012594694271683693,
0.10769831389188766,
-0.03700035437941551,
0.08042293041944504,
0.041265033185482025,
-0.1561787724494934,
0.013360886834561825,
-0.03018094040453434,
-0.06164681911468506,
-0.04755842313170433,
-0.020983237773180008,
-0.0662095695734024,
0.11357945203781128,
0.2285352647304535,
-0.033686090260744095,
0.0059063611552119255,
-0.07578404992818832,
0.008768702857196331,
0.048713043332099915,
-0.014550096355378628,
-0.0722947046160698,
-0.20845019817352295,
0.007920335978269577,
0.09209244698286057,
-0.013143004849553108,
-0.23129646480083466,
-0.09358822554349899,
-0.011637862771749496,
-0.06612849980592728,
-0.07789742201566696,
0.07803411036729813,
0.051450371742248535,
0.054301369935274124,
-0.04264070466160774,
-0.08501088619232178,
-0.06467333436012268,
0.17618705332279205,
-0.1331506371498108,
-0.09236814081668854
] |
null | null |
transformers
|
NER Model of BERN2 system
|
{}
| null |
dmis-lab/bern2-ner
|
[
"transformers",
"pytorch",
"roberta",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #roberta #endpoints_compatible #region-us
|
NER Model of BERN2 system
|
[] |
[
"TAGS\n#transformers #pytorch #roberta #endpoints_compatible #region-us \n"
] |
[
24
] |
[
"passage: TAGS\n#transformers #pytorch #roberta #endpoints_compatible #region-us \n"
] |
[
-0.04710838571190834,
0.000004481258201849414,
-0.01084833312779665,
-0.028830038383603096,
0.13689196109771729,
0.024661021307110786,
0.021276047453284264,
0.06774231046438217,
0.07277411967515945,
-0.005768520291894674,
0.12997958064079285,
0.23497812449932098,
-0.05278132110834122,
0.012769986875355244,
-0.058594558387994766,
-0.24939481914043427,
0.07595402747392654,
0.08731918036937714,
-0.08350498974323273,
0.10799740999937057,
0.05421356484293938,
-0.0927719920873642,
0.054529819637537,
-0.028123930096626282,
-0.11991901695728302,
0.0522734597325325,
0.02667221426963806,
-0.08790246397256851,
0.1232864186167717,
0.019574476405978203,
0.1794423609972,
0.025344358757138252,
-0.12528492510318756,
-0.1700793355703354,
0.03414492681622505,
0.00017859147919807583,
-0.06686453521251678,
0.02657829038798809,
0.044179003685712814,
-0.12301196157932281,
0.057532187551259995,
0.0489761121571064,
0.008118457160890102,
0.03497127443552017,
-0.1776191145181656,
-0.16942951083183289,
-0.05893833190202713,
0.028112027794122696,
0.023665081709623337,
0.07402801513671875,
0.019056716933846474,
0.17335262894630432,
-0.15199004113674164,
0.08120588213205338,
0.18331138789653778,
-0.28230294585227966,
0.0015046729240566492,
0.08527936041355133,
0.06626160442829132,
0.04926927015185356,
0.011503299698233604,
0.05494994297623634,
0.006573413033038378,
0.015392723493278027,
-0.038082290440797806,
-0.08980979025363922,
-0.05154621973633766,
0.08386038988828659,
-0.08894415944814682,
-0.09970521926879883,
0.2075752317905426,
-0.03570309281349182,
0.04820539429783821,
0.04827036336064339,
-0.09897863864898682,
-0.03811374679207802,
0.0005226113717071712,
0.00046723868581466377,
-0.02164444699883461,
0.06373990327119827,
0.0032088239677250385,
-0.027601467445492744,
-0.11205756664276123,
0.04216599091887474,
-0.22245104610919952,
0.28457048535346985,
0.026442578062415123,
0.0912148579955101,
-0.2189355343580246,
0.05204150453209877,
-0.04516136646270752,
-0.07567647844552994,
0.022055523470044136,
-0.10415781289339066,
0.030584361404180527,
0.014985884539783001,
-0.0726739913225174,
0.029076840728521347,
0.07451772689819336,
0.18171283602714539,
0.029848651960492134,
0.023849116638302803,
0.037326473742723465,
0.10956461727619171,
0.021593017503619194,
0.09593676775693893,
-0.004522756207734346,
-0.0062623959966003895,
0.038684386759996414,
-0.17025139927864075,
0.013620704412460327,
-0.037897635251283646,
-0.08988464623689651,
-0.07785378396511078,
0.012553787790238857,
0.11354856938123703,
0.03250754252076149,
0.007742214482277632,
-0.07710288465023041,
0.001371672609820962,
0.017974281683564186,
-0.04751988500356674,
-0.031038520857691765,
-0.013762025162577629,
0.02762053720653057,
0.21292611956596375,
-0.03568822890520096,
-0.014079419896006584,
-0.019960032775998116,
0.10149375349283218,
-0.07446283847093582,
-0.020848069339990616,
-0.041397370398044586,
-0.014479035511612892,
0.05333780497312546,
-0.14959795773029327,
0.0894417017698288,
-0.16235913336277008,
-0.07721706479787827,
0.023479359224438667,
0.04314779117703438,
0.004523617215454578,
0.012278388254344463,
0.012087945826351643,
-0.009219380095601082,
-0.027236631140112877,
-0.06194603443145752,
-0.06089937686920166,
-0.05625390261411667,
0.11111260950565338,
0.008455435745418072,
0.06341143697500229,
-0.10391087830066681,
0.05662205442786217,
-0.09099423885345459,
0.034544508904218674,
-0.1375199556350708,
-0.02523764967918396,
-0.02645164355635643,
0.1923660933971405,
0.0068205008283257484,
-0.05464669689536095,
-0.12457500398159027,
0.048568207770586014,
-0.038346800953149796,
0.15382003784179688,
-0.010524366982281208,
-0.11424499750137329,
0.2527368664741516,
-0.09813989698886871,
-0.16236662864685059,
0.05221249535679817,
0.005846796091645956,
0.007414969149976969,
0.07120196521282196,
0.16906669735908508,
0.042660776525735855,
-0.057577621191740036,
0.08609550446271896,
0.1047014519572258,
-0.16269704699516296,
-0.1719195693731308,
0.03675294667482376,
-0.03488772362470627,
-0.09545575827360153,
0.04062844440340996,
0.02270287647843361,
0.09206480532884598,
-0.08907764405012131,
-0.015806129202246666,
-0.01633630134165287,
-0.025665801018476486,
0.05208549275994301,
0.05835433676838875,
0.08001566678285599,
-0.07454166561365128,
0.005366346798837185,
-0.021103912964463234,
0.0089016268029809,
0.04055006429553032,
0.05169028416275978,
-0.043948616832494736,
0.1355406641960144,
-0.05684935301542282,
0.002631255192682147,
-0.21775834262371063,
-0.08198042213916779,
-0.031147386878728867,
0.09329741448163986,
-0.023606708273291588,
0.16439582407474518,
0.0935794934630394,
-0.11291755735874176,
0.005576607305556536,
-0.015517151914536953,
0.10377269238233566,
0.012552604079246521,
0.007448928896337748,
-0.03710265830159187,
0.037671931087970734,
-0.08019174635410309,
-0.06017129495739937,
-0.013761534355580807,
-0.0042076194658875465,
0.11294644325971603,
0.11935915052890778,
0.007766587194055319,
0.0414375476539135,
-0.02214924804866314,
0.06110641360282898,
-0.0232370775192976,
0.0158044695854187,
0.09678713232278824,
-0.01638456992805004,
-0.06773242354393005,
0.1486222892999649,
-0.07785765081644058,
0.3427961468696594,
0.19542205333709717,
-0.3026086688041687,
0.010744023136794567,
0.00817910023033619,
-0.01671556755900383,
0.026477953419089317,
0.11432955414056778,
0.0012866508914157748,
0.056610532104969025,
0.032953374087810516,
0.10734401643276215,
-0.019625410437583923,
-0.02903112582862377,
0.008584939874708652,
-0.06553634256124496,
-0.03852369263768196,
0.09696078300476074,
0.07031720131635666,
-0.13535082340240479,
0.1570144146680832,
0.24322782456874847,
0.0343286469578743,
0.10578674077987671,
-0.05931145325303078,
-0.016244295984506607,
0.03694572672247887,
0.011412939988076687,
-0.03407431021332741,
0.025389118120074272,
-0.21651425957679749,
-0.043358366936445236,
0.07186431437730789,
0.005580555181950331,
0.09065420180559158,
-0.13706739246845245,
-0.07837424427270889,
0.026302676647901535,
0.03360499441623688,
-0.04703304171562195,
0.11487317830324173,
0.03388232737779617,
0.077121801674366,
0.023567313328385353,
-0.04248740151524544,
0.09480869770050049,
0.008020560257136822,
-0.029814818874001503,
0.1654738038778305,
-0.10702735930681229,
-0.2581709027290344,
-0.12193218618631363,
-0.14464305341243744,
0.030673744156956673,
0.013289694674313068,
0.07178103923797607,
-0.09570969641208649,
-0.02794502303004265,
0.11040882021188736,
0.06391425430774689,
-0.15068218111991882,
0.028558287769556046,
-0.033212799578905106,
0.06838270276784897,
-0.09712500125169754,
-0.0651562362909317,
-0.06374204158782959,
-0.07063358277082443,
-0.04265620559453964,
0.11264993250370026,
-0.12369513511657715,
0.10658242553472519,
0.1185801699757576,
0.053718872368335724,
0.06899958103895187,
-0.0026584325823932886,
0.14750708639621735,
-0.08715782314538956,
-0.06181985139846802,
0.21091774106025696,
-0.017197037115693092,
0.09724331647157669,
0.11277004331350327,
0.03381453827023506,
-0.07095205038785934,
-0.029416784644126892,
-0.06745780259370804,
-0.12051685899496078,
-0.2196408361196518,
-0.10431885719299316,
-0.1338929682970047,
0.001790988608263433,
-0.013106169179081917,
0.03685977682471275,
0.09675391018390656,
0.08901132643222809,
0.0550701804459095,
-0.09611242264509201,
-0.05373142287135124,
0.06117139011621475,
0.2261544018983841,
-0.024348463863134384,
0.0807882696390152,
-0.08071843534708023,
-0.09673191606998444,
0.08415748924016953,
0.06262189149856567,
0.20184406638145447,
0.09570007771253586,
0.011600117199122906,
0.07747740298509598,
0.15804089605808258,
0.14802859723567963,
0.12967321276664734,
0.004285544157028198,
-0.03386882692575455,
-0.01980484649538994,
0.008402297273278236,
-0.06879641860723495,
0.008582867681980133,
0.0945887565612793,
-0.1460021585226059,
-0.04694917052984238,
-0.21670974791049957,
0.07566370069980621,
0.05846427008509636,
0.033341988921165466,
-0.14540016651153564,
-0.00019760403665713966,
0.07043024152517319,
0.006035317201167345,
-0.038673847913742065,
0.08767110854387283,
-0.04047485813498497,
-0.13432176411151886,
0.05549551174044609,
-0.03985515609383583,
0.10120426118373871,
-0.026484539732336998,
0.07381577044725418,
-0.06528782844543457,
-0.10097641497850418,
0.062186695635318756,
0.07416008412837982,
-0.22962474822998047,
0.2966843247413635,
-0.022008458152413368,
-0.07843590527772903,
-0.047935355454683304,
-0.0402735136449337,
0.00031782049336470664,
0.14924046397209167,
0.11227667331695557,
0.03040711209177971,
-0.08607617765665054,
-0.11602409929037094,
0.05855415388941765,
0.027639638632535934,
0.1261444091796875,
-0.02718510292470455,
-0.03371329978108406,
-0.011890764348208904,
-0.0054674530401825905,
-0.03813306242227554,
-0.0011421740055084229,
0.09245268255472183,
-0.13451752066612244,
0.043690603226423264,
-0.005595468450337648,
0.015751486644148827,
-0.0058060502633452415,
0.000216820408240892,
-0.04368269816040993,
0.12863071262836456,
-0.041234876960515976,
-0.056774456053972244,
-0.08750397711992264,
-0.15236316621303558,
0.13593828678131104,
-0.10860603302717209,
0.0916966125369072,
-0.09828807413578033,
-0.07356434315443039,
-0.07884984463453293,
-0.1789647787809372,
0.11347857862710953,
-0.09571298211812973,
0.027581993490457535,
-0.025902552530169487,
0.19910907745361328,
-0.06354320794343948,
-0.011006711050868034,
-0.009784103371202946,
0.030833037570118904,
-0.13190630078315735,
-0.0889718309044838,
0.014048268087208271,
-0.026677189394831657,
0.04725491255521774,
0.06850610673427582,
-0.010306584648787975,
0.04110793396830559,
0.00722151156514883,
-0.0022256039083003998,
0.2132948487997055,
0.20776315033435822,
-0.04205271974205971,
0.12124003469944,
0.16055002808570862,
-0.02298271469771862,
-0.2685646414756775,
-0.09160544723272324,
-0.17717303335666656,
-0.04701691120862961,
-0.017774688079953194,
-0.14274154603481293,
0.11054682731628418,
0.04712479189038277,
-0.02545398473739624,
0.11969882249832153,
-0.2559846341609955,
-0.04124341532588005,
0.1493636518716812,
0.0062699345871806145,
0.5546489953994751,
-0.11287706345319748,
-0.08298387378454208,
0.01305436436086893,
-0.24928133189678192,
0.07992938905954361,
0.01904345117509365,
0.059819381684064865,
-0.030170824378728867,
0.08720887452363968,
0.03699829801917076,
-0.08040692657232285,
0.12438756972551346,
0.0074538118205964565,
0.024170564487576485,
-0.07946096360683441,
-0.1280781477689743,
0.0872870683670044,
-0.004839536268264055,
-0.03662366047501564,
0.07352577149868011,
0.02766229584813118,
-0.13575823605060577,
-0.024937165901064873,
-0.12918852269649506,
0.0509490966796875,
0.046772826462984085,
-0.03127678856253624,
-0.028668509796261787,
-0.02079327218234539,
-0.03304222971200943,
0.01052573136985302,
0.2621735632419586,
-0.034418486058712006,
0.1412678211927414,
-0.007903030142188072,
0.05526409298181534,
-0.16902287304401398,
-0.08229594677686691,
-0.07681068032979965,
-0.054233379662036896,
0.06747926026582718,
-0.07158241420984268,
0.02895243652164936,
0.13923749327659607,
-0.014868827536702156,
-0.01971399411559105,
0.10571489483118057,
-0.006689982954412699,
-0.0020358276087790728,
0.14539019763469696,
-0.2069898396730423,
-0.05753513053059578,
-0.021022606641054153,
-0.04989248514175415,
0.1327422708272934,
0.11654264479875565,
0.0983550176024437,
0.07786483317613602,
-0.03430752828717232,
-0.02362983115017414,
-0.028949681669473648,
-0.08090628683567047,
0.0024792233016341925,
0.06885196268558502,
0.025992391631007195,
-0.11667148768901825,
0.05747832730412483,
-0.012274257838726044,
-0.23964636027812958,
-0.052203960716724396,
0.1062576100230217,
-0.13067859411239624,
-0.09600488841533661,
-0.07801525294780731,
0.04109029099345207,
-0.20162880420684814,
-0.034601181745529175,
-0.04159528389573097,
-0.10255556553602219,
0.05963057279586792,
0.24780724942684174,
0.10143034905195236,
0.09888434410095215,
-0.01028646994382143,
-0.016449730843305588,
0.021191705018281937,
-0.08732561767101288,
0.004237749148160219,
0.004638320300728083,
-0.09378271549940109,
-0.02855258248746395,
-0.01363504771143198,
0.15873940289020538,
-0.07890082895755768,
-0.08274190872907639,
-0.16358937323093414,
0.08995359390974045,
-0.09812886267900467,
-0.11397140473127365,
-0.12412403523921967,
-0.06893793493509293,
0.0037182122468948364,
-0.09338416159152985,
-0.04137818142771721,
-0.02767200581729412,
-0.13551916182041168,
0.07086745649576187,
0.023581858724355698,
-0.0037772138603031635,
-0.05501500517129898,
-0.039861805737018585,
0.1472712904214859,
-0.042508505284786224,
0.09451454877853394,
0.17123942077159882,
-0.07602482289075851,
0.12351230531930923,
-0.10503007471561432,
-0.1372130960226059,
0.10087032616138458,
-0.0001759071456035599,
0.07120119035243988,
0.06956369429826736,
0.020766686648130417,
0.06875994801521301,
0.030996467918157578,
0.05662810802459717,
-0.034365203231573105,
-0.12265201658010483,
0.031028853729367256,
0.03192706033587456,
-0.1781196892261505,
-0.033246178179979324,
-0.08630622923374176,
0.15495331585407257,
0.028343068435788155,
0.0859447568655014,
0.022196872159838676,
0.10975440591573715,
-0.01211862824857235,
0.007993047125637531,
-0.0023370154667645693,
-0.19463440775871277,
0.03239840641617775,
-0.05557585507631302,
-0.0024593796115368605,
0.0033398494124412537,
0.2839198708534241,
-0.0705515444278717,
0.023442646488547325,
0.02998867630958557,
0.06523754447698593,
0.012018566019833088,
0.025090577080845833,
0.21156036853790283,
0.10683322697877884,
-0.039572279900312424,
-0.09842078387737274,
0.08749271929264069,
-0.011929450556635857,
-0.0594974085688591,
0.12458037585020065,
0.13864046335220337,
0.08676675707101822,
0.08204910904169083,
-0.0033394009806215763,
0.070549875497818,
-0.06136783957481384,
-0.25319474935531616,
0.011225651018321514,
0.012381847016513348,
0.016442246735095978,
0.09178762137889862,
0.16398729383945465,
-0.004069536458700895,
0.0749489888548851,
-0.01615913212299347,
0.0006451079971157014,
-0.13821111619472504,
-0.08257102966308594,
-0.047923535108566284,
-0.10142721980810165,
0.029083652421832085,
-0.05849800631403923,
-0.007420131005346775,
0.17096444964408875,
0.0392775796353817,
-0.03622451052069664,
0.09747672080993652,
0.0760326161980629,
-0.0670693889260292,
0.03528399392962456,
-0.007171704433858395,
0.021475624293088913,
0.04873443394899368,
-0.012379645369946957,
-0.1494901478290558,
-0.07176815718412399,
-0.06488432735204697,
0.037990935146808624,
-0.10138802975416183,
-0.01745467633008957,
-0.13597668707370758,
-0.09903085976839066,
-0.05588725209236145,
0.08903270214796066,
-0.048549771308898926,
0.0984780415892601,
-0.02612706646323204,
0.006907667499035597,
0.016333235427737236,
0.19976969063282013,
-0.05262446030974388,
-0.04221903532743454,
0.001148911309428513,
0.16957782208919525,
0.05555977299809456,
0.09394799172878265,
-0.011725390329957008,
0.026208948343992233,
-0.06366859376430511,
0.33636584877967834,
0.2535180151462555,
-0.0233058612793684,
0.04417799040675163,
0.05260763689875603,
0.04840950295329094,
0.11710292100906372,
0.09676902741193771,
0.10052920877933502,
0.3192276656627655,
-0.0892692431807518,
-0.04494080692529678,
-0.031999826431274414,
0.013973445631563663,
-0.10555802285671234,
0.05037616938352585,
0.03306281939148903,
-0.07167352735996246,
-0.07398556917905807,
0.12090129405260086,
-0.17856739461421967,
0.11545499414205551,
0.08242010325193405,
-0.22637870907783508,
-0.05336158722639084,
-0.0596175380051136,
0.15627482533454895,
-0.002246581017971039,
0.12675566971302032,
-0.04257085174322128,
-0.15510323643684387,
0.03285594657063484,
0.062185365706682205,
-0.2798806130886078,
-0.09374892711639404,
0.11671829223632812,
0.06078186258673668,
-0.02176550216972828,
-0.03682553395628929,
-0.0003849651257041842,
0.07988492399454117,
0.05872725695371628,
-0.031531307846307755,
0.017113979905843735,
0.04315922036767006,
-0.08697821199893951,
-0.09998752921819687,
-0.019003579393029213,
0.011787851341068745,
-0.1190737634897232,
0.04075232893228531,
-0.15655088424682617,
0.04439463093876839,
0.01169892493635416,
-0.00949098076671362,
0.0003081741160713136,
0.016252411529421806,
-0.06897782534360886,
0.01970541663467884,
0.047393329441547394,
0.0031528547406196594,
-0.020613694563508034,
-0.04386329650878906,
-0.013762234710156918,
0.06974862515926361,
-0.0728449895977974,
-0.15882854163646698,
-0.005591145716607571,
-0.09538131207227707,
0.11189249902963638,
-0.030698299407958984,
-0.08759639412164688,
-0.01687025837600231,
-0.053370025008916855,
0.05895037204027176,
-0.11432106047868729,
0.04182504862546921,
0.026789119467139244,
0.040701985359191895,
0.021567801013588905,
-0.013804933987557888,
0.046552740037441254,
0.05624568089842796,
-0.12087129056453705,
-0.07611918449401855
] |
null | null |
transformers
|
# Model Card for biobert-large-cased-v1.1-squad
# Model Details
## Model Description
More information needed
- **Developed by:** DMIS-lab (Data Mining and Information Systems Lab, Korea University)
- **Shared by [Optional]:** DMIS-lab (Data Mining and Information Systems Lab, Korea University)
- **Model type:** Question Answering
- **Language(s) (NLP):** More information needed
- **License:** More information needed
- **Parent Model:** [gpt-neo-2.7B](https://huggingface.co/EleutherAI/gpt-neo-2.7B)
- **Resources for more information:**
- [GitHub Repo](https://github.com/jhyuklee/biobert)
- [Associated Paper](https://arxiv.org/abs/1901.08746)
# Uses
## Direct Use
This model can be used for the task of question answering.
## Downstream Use [Optional]
More information needed.
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
The model creators note in the [associated paper](https://arxiv.org/pdf/1901.08746.pdf):
> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))
## Training Procedure
### Preprocessing
The model creators note in the [associated paper](https://arxiv.org/pdf/1901.08746.pdf):
> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs
### Speeds, Sizes, Times
The model creators note in the [associated paper](https://arxiv.org/pdf/1901.08746.pdf):
> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed
### Factors
More information needed
### Metrics
More information needed
## Results
More information needed
# Model Examination
More information needed
# Environmental Impact
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** More information needed
- **Training**: Eight NVIDIA V100 (32GB) GPUs [ for training],
- **Fine-tuning:** a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task
- **Hours used:** More information needed
- **Cloud Provider:** More information needed
- **Compute Region:** More information needed
- **Carbon Emitted:** More information needed
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed
## Compute Infrastructure
More information needed
### Hardware
More information needed
### Software
More information needed.
# Citation
**BibTeX:**
```bibtex
@misc{mesh-transformer-jax,
@article{lee2019biobert,
title={BioBERT: a pre-trained biomedical language representation model for biomedical text mining},
author={Lee, Jinhyuk and Yoon, Wonjin and Kim, Sungdong and Kim, Donghyeon and Kim, Sunkyu and So, Chan Ho and Kang, Jaewoo},
journal={arXiv preprint arXiv:1901.08746},
year={2019}
}
```
# Glossary [optional]
More information needed
# More Information [optional]
For help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee(`lee.jnhk (at) gmail.com`), or Wonjin Yoon (`wonjin.info (at) gmail.com`) for communication related to BioBERT.
# Model Card Authors [optional]
DMIS-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
```python
from transformers import AutoTokenizer, AutoModelForQuestionAnswering
tokenizer = AutoTokenizer.from_pretrained("dmis-lab/biobert-large-cased-v1.1-squad")
model = AutoModelForQuestionAnswering.from_pretrained("dmis-lab/biobert-large-cased-v1.1-squad")
```
</details>
|
{"tags": ["question-answering", "bert"]}
|
question-answering
|
dmis-lab/biobert-large-cased-v1.1-squad
|
[
"transformers",
"pytorch",
"jax",
"bert",
"question-answering",
"arxiv:1901.08746",
"arxiv:1910.09700",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1901.08746",
"1910.09700"
] |
[] |
TAGS
#transformers #pytorch #jax #bert #question-answering #arxiv-1901.08746 #arxiv-1910.09700 #endpoints_compatible #has_space #region-us
|
# Model Card for biobert-large-cased-v1.1-squad
# Model Details
## Model Description
More information needed
- Developed by: DMIS-lab (Data Mining and Information Systems Lab, Korea University)
- Shared by [Optional]: DMIS-lab (Data Mining and Information Systems Lab, Korea University)
- Model type: Question Answering
- Language(s) (NLP): More information needed
- License: More information needed
- Parent Model: gpt-neo-2.7B
- Resources for more information:
- GitHub Repo
- Associated Paper
# Uses
## Direct Use
This model can be used for the task of question answering.
## Downstream Use [Optional]
More information needed.
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
The model creators note in the associated paper:
> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))
## Training Procedure
### Preprocessing
The model creators note in the associated paper:
> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs
### Speeds, Sizes, Times
The model creators note in the associated paper:
> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed
### Factors
More information needed
### Metrics
More information needed
## Results
More information needed
# Model Examination
More information needed
# Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: More information needed
- Training: Eight NVIDIA V100 (32GB) GPUs [ for training],
- Fine-tuning: a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task
- Hours used: More information needed
- Cloud Provider: More information needed
- Compute Region: More information needed
- Carbon Emitted: More information needed
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed
## Compute Infrastructure
More information needed
### Hardware
More information needed
### Software
More information needed.
BibTeX:
# Glossary [optional]
More information needed
# More Information [optional]
For help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee('URL (at) URL'), or Wonjin Yoon ('URL (at) URL') for communication related to BioBERT.
# Model Card Authors [optional]
DMIS-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
</details>
|
[
"# Model Card for biobert-large-cased-v1.1-squad",
"# Model Details",
"## Model Description\n \nMore information needed\n \n- Developed by: DMIS-lab (Data Mining and Information Systems Lab, Korea University)\n- Shared by [Optional]: DMIS-lab (Data Mining and Information Systems Lab, Korea University)\n\n- Model type: Question Answering\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Parent Model: gpt-neo-2.7B\n- Resources for more information:\n \t- GitHub Repo\n \t - Associated Paper",
"# Uses",
"## Direct Use\nThis model can be used for the task of question answering.",
"## Downstream Use [Optional]\n \nMore information needed.",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\n \nThe model creators note in the associated paper:\n> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))",
"## Training Procedure",
"### Preprocessing\n \n The model creators note in the associated paper:\n> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs",
"### Speeds, Sizes, Times\n \n The model creators note in the associated paper:\n> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed",
"### Factors\nMore information needed",
"### Metrics\n \nMore information needed",
"## Results \n \nMore information needed",
"# Model Examination\n \nMore information needed",
"# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed\n\t- Training: Eight NVIDIA V100 (32GB) GPUs [ for training], \n - Fine-tuning: a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed",
"## Compute Infrastructure\n \nMore information needed",
"### Hardware\n \n \nMore information needed",
"### Software\n \nMore information needed.\n \nBibTeX:",
"# Glossary [optional]\n \nMore information needed",
"# More Information [optional]\n \nFor help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee('URL (at) URL'), or Wonjin Yoon ('URL (at) URL') for communication related to BioBERT.",
"# Model Card Authors [optional]\n \n DMIS-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] |
[
"TAGS\n#transformers #pytorch #jax #bert #question-answering #arxiv-1901.08746 #arxiv-1910.09700 #endpoints_compatible #has_space #region-us \n",
"# Model Card for biobert-large-cased-v1.1-squad",
"# Model Details",
"## Model Description\n \nMore information needed\n \n- Developed by: DMIS-lab (Data Mining and Information Systems Lab, Korea University)\n- Shared by [Optional]: DMIS-lab (Data Mining and Information Systems Lab, Korea University)\n\n- Model type: Question Answering\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Parent Model: gpt-neo-2.7B\n- Resources for more information:\n \t- GitHub Repo\n \t - Associated Paper",
"# Uses",
"## Direct Use\nThis model can be used for the task of question answering.",
"## Downstream Use [Optional]\n \nMore information needed.",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\n \nThe model creators note in the associated paper:\n> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))",
"## Training Procedure",
"### Preprocessing\n \n The model creators note in the associated paper:\n> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs",
"### Speeds, Sizes, Times\n \n The model creators note in the associated paper:\n> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed",
"### Factors\nMore information needed",
"### Metrics\n \nMore information needed",
"## Results \n \nMore information needed",
"# Model Examination\n \nMore information needed",
"# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: More information needed\n\t- Training: Eight NVIDIA V100 (32GB) GPUs [ for training], \n - Fine-tuning: a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed",
"## Compute Infrastructure\n \nMore information needed",
"### Hardware\n \n \nMore information needed",
"### Software\n \nMore information needed.\n \nBibTeX:",
"# Glossary [optional]\n \nMore information needed",
"# More Information [optional]\n \nFor help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee('URL (at) URL'), or Wonjin Yoon ('URL (at) URL') for communication related to BioBERT.",
"# Model Card Authors [optional]\n \n DMIS-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] |
[
54,
18,
3,
105,
3,
16,
12,
28,
87,
41,
3,
168,
4,
63,
53,
3,
11,
8,
7,
8,
5,
8,
113,
9,
10,
8,
6,
12,
11,
61,
39,
7,
41
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #question-answering #arxiv-1901.08746 #arxiv-1910.09700 #endpoints_compatible #has_space #region-us \n# Model Card for biobert-large-cased-v1.1-squad# Model Details## Model Description\n \nMore information needed\n \n- Developed by: DMIS-lab (Data Mining and Information Systems Lab, Korea University)\n- Shared by [Optional]: DMIS-lab (Data Mining and Information Systems Lab, Korea University)\n\n- Model type: Question Answering\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Parent Model: gpt-neo-2.7B\n- Resources for more information:\n \t- GitHub Repo\n \t - Associated Paper# Uses## Direct Use\nThis model can be used for the task of question answering.## Downstream Use [Optional]\n \nMore information needed.## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.# Bias, Risks, and Limitations\n \n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.## Recommendations\n \n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.# Training Details"
] |
[
-0.027708033099770546,
0.05214131996035576,
-0.0034538519103080034,
0.041209835559129715,
0.052586667239665985,
-0.028332658112049103,
0.12057913839817047,
0.0756133496761322,
0.10301472246646881,
0.05523096024990082,
0.07027857005596161,
0.041403189301490784,
0.07070721685886383,
0.09866725653409958,
-0.018241170793771744,
-0.2067629098892212,
0.031769655644893646,
-0.017235828563570976,
0.06064983084797859,
0.12316014617681503,
0.12456022948026657,
-0.04187176749110222,
0.08877405524253845,
0.03598915785551071,
-0.025070708245038986,
-0.04345204308629036,
-0.055389586836099625,
0.002657892182469368,
0.07516011595726013,
0.057865697890520096,
0.06190360337495804,
0.006592107005417347,
0.017773715779185295,
-0.24088063836097717,
0.02643907256424427,
-0.005655933637171984,
-0.03995237872004509,
0.0729064866900444,
0.06273749470710754,
-0.025845762342214584,
0.14743810892105103,
-0.09385931491851807,
0.039071373641490936,
0.055520977824926376,
-0.05329849198460579,
-0.15155114233493805,
-0.09510611742734909,
0.05223219469189644,
0.13467247784137726,
0.0760493353009224,
-0.06458861380815506,
0.21975600719451904,
-0.0801425352692604,
0.02575664408504963,
0.10170318186283112,
-0.07456677407026291,
-0.023442663252353668,
0.018908599391579628,
0.009398973546922207,
0.009779728017747402,
-0.0664537325501442,
0.002551772864535451,
0.02672543190419674,
0.026071054860949516,
0.07034103572368622,
-0.047248125076293945,
0.13268926739692688,
-0.044468339532613754,
-0.13086853921413422,
0.013817239552736282,
0.09242569655179977,
0.057945363223552704,
-0.06678282469511032,
-0.2617108225822449,
-0.0011931726476177573,
0.10970677435398102,
0.03464212641119957,
-0.1315179318189621,
0.024279305711388588,
-0.03456947207450867,
0.12503868341445923,
-0.06650400906801224,
-0.07012061029672623,
-0.0015673278830945492,
-0.04638664796948433,
0.21286889910697937,
0.05207028612494469,
-0.0002758741902653128,
-0.08283746242523193,
0.06775687634944916,
-0.020333146676421165,
-0.11266661435365677,
-0.11843027919530869,
-0.1109081357717514,
-0.047448381781578064,
-0.005922188051044941,
-0.0016251918859779835,
0.009367420338094234,
-0.023923305794596672,
0.16632315516471863,
-0.02569715864956379,
0.03673418238759041,
0.020525095984339714,
0.029393061995506287,
0.17124269902706146,
0.060842517763376236,
-0.06913891434669495,
0.008923103101551533,
0.037810493260622025,
0.0357794463634491,
0.01778435707092285,
-0.01880219578742981,
-0.04454462230205536,
-0.05336912348866463,
0.03079254925251007,
0.0685715451836586,
0.07347528636455536,
0.06671254336833954,
-0.061971403658390045,
-0.08532296121120453,
0.1184932217001915,
-0.08330226689577103,
-0.0499129444360733,
-0.03426960110664368,
-0.05569261685013771,
0.06281251460313797,
-0.013525300659239292,
-0.009113590233027935,
-0.022877199575304985,
-0.009986326098442078,
-0.08035917580127716,
-0.00789925828576088,
-0.09075187891721725,
-0.02595674991607666,
0.04297000914812088,
0.027476051822304726,
0.04784601554274559,
-0.10138005763292313,
-0.11736605316400528,
-0.08092329651117325,
0.017872147262096405,
-0.08809134364128113,
-0.006420734338462353,
-0.05916449427604675,
-0.0621584877371788,
0.006619052030146122,
-0.041478995233774185,
-0.020306896418333054,
0.007681176532059908,
0.03910457715392113,
-0.018887916579842567,
0.011942568235099316,
0.10321151465177536,
0.006073326338082552,
-0.10385891795158386,
0.03185008838772774,
-0.13243526220321655,
0.09658825397491455,
-0.07788147032260895,
-0.0676826760172844,
-0.08522544801235199,
-0.05112433061003685,
0.05639302358031273,
0.09137077629566193,
-0.026193805038928986,
0.18497061729431152,
-0.1300286501646042,
0.010109087452292442,
0.18659214675426483,
-0.1244194507598877,
-0.07841864228248596,
0.10363227874040604,
-0.09637993574142456,
0.142831489443779,
0.07941200584173203,
0.08030479401350021,
-0.000585172267165035,
-0.15302346646785736,
-0.05542626231908798,
0.01335457619279623,
-0.08899012207984924,
0.1119806095957756,
0.08093804866075516,
0.017480628564953804,
0.0374872200191021,
-0.0012595540611073375,
-0.13923723995685577,
0.053596723824739456,
-0.01735682226717472,
-0.038010384887456894,
0.03205259144306183,
-0.010357889346778393,
0.10421203076839447,
-0.000038093028706498444,
-0.013036375865340233,
0.040544670075178146,
-0.17296433448791504,
-0.014875694178044796,
0.08330469578504562,
-0.02352999895811081,
-0.02431606873869896,
-0.10135549306869507,
-0.019747216254472733,
0.06378165632486343,
-0.004284644033759832,
-0.12403109669685364,
-0.10595014691352844,
-0.02320890687406063,
-0.15350107848644257,
0.0814632847905159,
0.09153084456920624,
0.024868600070476532,
0.036294639110565186,
-0.058682881295681,
0.014519215561449528,
-0.02430589683353901,
0.010375232435762882,
-0.030560128390789032,
-0.13690955936908722,
0.023895835503935814,
-0.04058646783232689,
-0.0019475955050438643,
-0.19195233285427094,
0.0250750370323658,
0.05163469538092613,
0.027731897309422493,
0.0022232525516301394,
-0.012146192602813244,
0.08015340566635132,
-0.010196046903729439,
-0.0028869174420833588,
-0.012859971262514591,
0.006688223220407963,
-0.058407220989465714,
-0.0934768095612526,
0.13979269564151764,
-0.20384742319583893,
-0.0409320630133152,
0.07712258398532867,
-0.023772433400154114,
-0.10137210786342621,
-0.07791081815958023,
-0.010335351340472698,
-0.038519103080034256,
-0.017402540892362595,
-0.03342258557677269,
0.18297415971755981,
0.02046346105635166,
0.035012733191251755,
-0.11386824399232864,
-0.10954880714416504,
-0.013936526142060757,
-0.05497697740793228,
-0.035185933113098145,
0.054282184690237045,
-0.02698463574051857,
-0.21511100232601166,
0.12608392536640167,
0.05097278580069542,
-0.012406119145452976,
0.14678116142749786,
0.033068250864744186,
-0.07762278616428375,
-0.05587633326649666,
0.026017040014266968,
-0.07404971867799759,
0.10570728778839111,
0.002734056208282709,
0.01765398308634758,
0.06284236907958984,
0.058628927916288376,
0.03539053350687027,
-0.07459523528814316,
-0.050905339419841766,
0.00045765782124362886,
0.03360757976770401,
-0.05736607313156128,
0.07195873558521271,
0.025288045406341553,
0.07247272878885269,
0.01871449127793312,
0.02027583122253418,
0.046856388449668884,
-0.04918542876839638,
-0.13209785521030426,
0.17340300977230072,
-0.05926310271024704,
-0.27467647194862366,
-0.07523678243160248,
-0.027727751061320305,
-0.001343170995824039,
-0.024375075474381447,
0.026459161192178726,
-0.10698482394218445,
-0.09108071029186249,
-0.0703856348991394,
0.12528401613235474,
-0.06584572046995163,
-0.1178654208779335,
-0.056116677820682526,
0.046251554042100906,
-0.006578892935067415,
-0.09706061333417892,
-0.009419631212949753,
-0.04662998393177986,
-0.0357089601457119,
0.04660367965698242,
0.052592385560274124,
0.07260952144861221,
0.014860068447887897,
0.03911036625504494,
-0.05911923572421074,
-0.027228740975260735,
0.1934768706560135,
-0.15010493993759155,
0.030201490968465805,
0.19417376816272736,
-0.14989593625068665,
0.07436905056238174,
0.11210789531469345,
0.03321876749396324,
-0.07015900313854218,
0.030710164457559586,
0.046323128044605255,
-0.046982165426015854,
-0.2634488642215729,
-0.07085705548524857,
-0.00892436970025301,
-0.09512659907341003,
-0.0032509344164282084,
0.03337987884879112,
0.09355885535478592,
0.0845700055360794,
-0.09165014326572418,
-0.039595574140548706,
0.04542555660009384,
0.08303270488977432,
0.10666481405496597,
0.009022668935358524,
0.06924882531166077,
0.04085405543446541,
-0.03312067314982414,
0.0713043212890625,
-0.04179046303033829,
0.3059712052345276,
0.005582353565841913,
0.06727098673582077,
0.10923560708761215,
0.09390728175640106,
0.041124098002910614,
-0.01276511698961258,
-0.012624124065041542,
0.0013519440544769168,
-0.07376920431852341,
-0.043875619769096375,
-0.019237302243709564,
0.10495568811893463,
-0.020884227007627487,
-0.035327132791280746,
0.02169044502079487,
-0.028793415054678917,
0.03522521257400513,
0.03293026238679886,
-0.051782574504613876,
-0.06598710268735886,
-0.054649945348501205,
0.06856335699558258,
-0.07134732604026794,
-0.08042772859334946,
0.050849784165620804,
0.07730289548635483,
-0.1773754358291626,
0.038015253841876984,
-0.028830472379922867,
0.0749121680855751,
-0.14041800796985626,
0.019393349066376686,
-0.08239319175481796,
0.029292022809386253,
0.03294211998581886,
0.10134971886873245,
-0.16770170629024506,
0.20302237570285797,
0.017589842900633812,
0.022815540432929993,
-0.12820731103420258,
0.03698427230119705,
0.04438190534710884,
-0.04878539964556694,
0.1984386146068573,
0.032285887748003006,
-0.010450801812112331,
-0.12171512097120285,
-0.0679832175374031,
0.023961476981639862,
0.07251057773828506,
-0.07199623435735703,
0.05218292400240898,
0.013488407246768475,
0.03525231406092644,
-0.003322455333545804,
-0.05141206458210945,
-0.12677831947803497,
-0.13988657295703888,
0.055040545761585236,
-0.16483812034130096,
0.045658230781555176,
-0.03053535334765911,
-0.06630683690309525,
-0.036223337054252625,
0.1693449169397354,
-0.19678543508052826,
-0.11882282793521881,
-0.0961039811372757,
0.011180082336068153,
0.08815451711416245,
-0.12349952757358551,
-0.027624376118183136,
0.038428857922554016,
0.043790873140096664,
-0.014151480048894882,
-0.1075066551566124,
0.08512333035469055,
-0.06307553499937057,
-0.130661740899086,
-0.018209895119071007,
0.07784304767847061,
0.20031343400478363,
0.05153072625398636,
0.003277804935351014,
0.016552604734897614,
-0.01690935716032982,
-0.18407249450683594,
0.033191509544849396,
0.2133684754371643,
0.11756034195423126,
0.04080614447593689,
0.004309980198740959,
-0.02678440324962139,
-0.0932779610157013,
-0.061807163059711456,
0.04385073110461235,
0.2849283218383789,
-0.0073287636041641235,
0.11168688535690308,
0.17906364798545837,
-0.07106747478246689,
-0.1881413459777832,
-0.04738819599151611,
0.05113456770777702,
0.018430622294545174,
0.16735774278640747,
-0.18375054001808167,
0.024980667978525162,
0.050594791769981384,
-0.006825955584645271,
0.057485535740852356,
-0.058149922639131546,
-0.11660341173410416,
0.18294724822044373,
0.04213959351181984,
0.17561835050582886,
-0.04428786411881447,
-0.047255683690309525,
0.0905332863330841,
-0.07371249794960022,
0.17407242953777313,
-0.012966765090823174,
0.02074924111366272,
-0.0012669377028942108,
0.08215130865573883,
0.05689970403909683,
-0.022435547783970833,
0.1668318659067154,
0.01425185613334179,
0.09567435085773468,
-0.11638916283845901,
-0.1057562530040741,
-0.033228784799575806,
-0.03980140760540962,
0.13222630321979523,
0.01171995047479868,
0.015073604881763458,
-0.12023278325796127,
-0.09009190648794174,
-0.09595191478729248,
0.013469723984599113,
-0.04289804399013519,
-0.07243296504020691,
-0.10703177005052567,
0.1310546100139618,
0.07668722420930862,
-0.023630701005458832,
0.14044998586177826,
-0.06725333631038666,
-0.0558035746216774,
0.08578620105981827,
0.2647712230682373,
0.06313822418451309,
-0.055858053267002106,
-0.034392546862363815,
-0.01194559782743454,
0.08469020575284958,
-0.15277200937271118,
-0.0044685774482786655,
0.05073818191885948,
-0.004655566066503525,
0.09637916088104248,
-0.011087807826697826,
-0.1267620027065277,
0.03246261924505234,
-0.01322667021304369,
-0.07253079116344452,
-0.19387851655483246,
-0.08880101889371872,
-0.06514600664377213,
-0.07729841023683548,
-0.08937238901853561,
0.1184060275554657,
-0.0409691296517849,
-0.057851966470479965,
0.009754597209393978,
0.08544161170721054,
0.04807496815919876,
-0.014211282134056091,
0.048865728080272675,
0.054894134402275085,
-0.062423523515462875,
-0.004691942594945431,
-0.010885834693908691,
0.009913181886076927,
0.12238174676895142,
0.038780394941568375,
-0.07685697078704834,
-0.04955799877643585,
-0.030250009149312973,
0.03008182905614376,
-0.024439159780740738,
-0.020892517641186714,
0.028125261887907982,
-0.03683481365442276,
-0.04120711609721184,
0.043152812868356705,
0.07072225958108902,
0.00613941065967083,
-0.008852529339492321,
0.03983671963214874,
-0.011989396065473557,
0.06526707112789154,
-0.009681365452706814,
-0.005886547267436981,
-0.0634608194231987,
-0.04507169499993324,
0.05528784170746803,
0.08096729218959808,
-0.05853838101029396,
-0.028432676568627357,
-0.13446326553821564,
-0.02259446121752262,
-0.1516469120979309,
0.020471567288041115,
-0.15013374388217926,
0.024797504767775536,
-0.03581181913614273,
-0.10342300683259964,
0.0025339482817798853,
0.02992013469338417,
-0.05675790458917618,
0.030959559604525566,
0.03316723555326462,
0.07984576374292374,
-0.1104176938533783,
0.0018179734470322728,
0.09317684173583984,
-0.06503964215517044,
0.100338876247406,
0.01141271274536848,
-0.029530124738812447,
0.08007308095693588,
-0.12458974123001099,
0.06280985474586487,
-0.02674170210957527,
0.018125835806131363,
0.022102247923612595,
-0.1824774295091629,
-0.02568463608622551,
-0.003443892113864422,
0.004776834975928068,
0.016561873257160187,
-0.009997006505727768,
-0.07811210304498672,
0.011317909695208073,
0.03669065237045288,
-0.06528019905090332,
-0.09338483214378357,
0.0255071222782135,
0.08988071233034134,
-0.017875678837299347,
0.11202085763216019,
0.003586625447496772,
0.07106737047433853,
-0.1596122533082962,
-0.005301102064549923,
0.05867708474397659,
0.06530015915632248,
0.06295982003211975,
-0.09558556228876114,
0.039507538080215454,
-0.057744551450014114,
0.22448784112930298,
-0.026051979511976242,
-0.02882608212530613,
0.03863659128546715,
0.023459885269403458,
0.0034031420946121216,
-0.006828881800174713,
-0.009074688889086246,
-0.03368460759520531,
0.008298593573272228,
0.017778566107153893,
-0.014713369309902191,
-0.06336813420057297,
-0.22200392186641693,
0.13923323154449463,
0.09612751007080078,
0.036485638469457626,
0.009886326268315315,
0.015204351395368576,
-0.06070006638765335,
-0.13974113762378693,
0.03150459751486778,
0.00923157762736082,
-0.09485325962305069,
-0.053294386714696884,
0.12473359704017639,
0.14222584664821625,
-0.08948244154453278,
0.10782064497470856,
-0.0039046367164701223,
-0.06616899371147156,
-0.1447228044271469,
-0.2733655273914337,
-0.05207398533821106,
-0.03481973707675934,
0.02187711000442505,
-0.10205883532762527,
0.01819201372563839,
0.11531012505292892,
0.020361267030239105,
-0.062104787677526474,
0.04091370105743408,
0.03796893358230591,
-0.12124674022197723,
0.03807966783642769,
0.015855822712183,
0.017274243757128716,
-0.0388835147023201,
0.028349023312330246,
0.009085115045309067,
0.051777806133031845,
0.05048546567559242,
0.033185478299856186,
-0.003760048421099782,
-0.03100319765508175,
-0.1302592009305954,
-0.0948699340224266,
-0.012049326673150063,
0.08487503975629807,
0.06399611383676529,
0.15507946908473969,
0.025166304782032967,
-0.010437329299747944,
-0.0286972988396883,
0.16424596309661865,
0.024480562657117844,
0.013259390369057655,
-0.11257626861333847,
0.17336879670619965,
-0.015331380069255829,
0.0013851383700966835,
0.05267420411109924,
-0.07176755368709564,
0.03919665887951851,
0.19300033152103424,
0.16729244589805603,
-0.0412319116294384,
-0.0010751860681921244,
-0.03071371093392372,
0.02294609323143959,
0.0012205924140289426,
0.07807336002588272,
0.03875068202614784,
0.26840245723724365,
-0.07657008618116379,
0.07560016214847565,
-0.013400367461144924,
-0.01851501502096653,
0.04909845441579819,
0.09535007923841476,
0.044659774750471115,
0.01043759286403656,
-0.11116289347410202,
0.12676826119422913,
-0.08912186324596405,
-0.10717964917421341,
-0.017698802053928375,
-0.04081152379512787,
-0.09633437544107437,
0.04211408272385597,
-0.01610499992966652,
-0.0037220034282654524,
0.04735533148050308,
0.046264760196208954,
-0.03032563626766205,
0.17233476042747498,
0.039131518453359604,
-0.08199973404407501,
0.012313651852309704,
0.12023473531007767,
0.024197565391659737,
0.17013895511627197,
0.052487365901470184,
0.15127603709697723,
0.09160982072353363,
-0.023552749305963516,
-0.10322118550539017,
0.01781330816447735,
0.02539614774286747,
-0.046358104795217514,
0.013103276491165161,
0.21859173476696014,
0.03841222822666168,
0.04928317293524742,
0.10440464317798615,
-0.06999993324279785,
0.02317342534661293,
-0.10978535562753677,
-0.10449949651956558,
-0.07009675353765488,
0.12973198294639587,
-0.09253882616758347,
0.12266223132610321,
0.13786324858665466,
-0.01627308875322342,
-0.023801816627383232,
-0.02401907555758953,
0.07320570945739746,
-0.05019018426537514,
0.048843514174222946,
0.014649544842541218,
-0.14070962369441986,
0.002728669438511133,
0.17771947383880615,
0.06942423433065414,
-0.1638791710138321,
-0.023053348064422607,
-0.0003721349057741463,
-0.010870848782360554,
-0.013707748614251614,
0.0341392420232296,
-0.023819858208298683,
0.012487120926380157,
-0.022146029397845268,
-0.1639927625656128,
-0.01076267845928669,
0.04118458554148674,
-0.1009516641497612,
-0.007345299702137709
] |
null | null |
transformers
|
hello
|
{}
|
feature-extraction
|
dmis-lab/biosyn-biobert-bc2gn
|
[
"transformers",
"pytorch",
"bert",
"feature-extraction",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #feature-extraction #endpoints_compatible #region-us
|
hello
|
[] |
[
"TAGS\n#transformers #pytorch #bert #feature-extraction #endpoints_compatible #region-us \n"
] |
[
29
] |
[
"passage: TAGS\n#transformers #pytorch #bert #feature-extraction #endpoints_compatible #region-us \n"
] |
[
-0.0680389553308487,
-0.01353863999247551,
-0.009260591119527817,
0.003671469632536173,
0.13468711078166962,
0.03987877443432808,
-0.0037161505315452814,
0.08307137340307236,
0.06908576935529709,
-0.009869525209069252,
0.10839105397462845,
0.22950756549835205,
-0.03434249758720398,
0.027836797758936882,
-0.06780551373958588,
-0.2686935067176819,
0.08416040241718292,
0.1071663498878479,
-0.04096720367670059,
0.08816211670637131,
0.04718845710158348,
-0.10185960680246353,
0.059242263436317444,
-0.017401615157723427,
-0.13960762321949005,
0.0563788041472435,
0.027032675221562386,
-0.08215486258268356,
0.11170924454927444,
0.02429274097084999,
0.171868696808815,
0.00669759139418602,
-0.10259579122066498,
-0.1729365438222885,
0.024251695722341537,
-0.016173768788576126,
-0.06165656819939613,
0.030105262994766235,
0.07771246135234833,
-0.10020971298217773,
0.020183663815259933,
0.09852731972932816,
0.01621491089463234,
0.024930456653237343,
-0.16311487555503845,
-0.17609833180904388,
-0.05667632818222046,
0.04488549754023552,
0.011052189394831657,
0.0923970639705658,
0.017688613384962082,
0.13918437063694,
-0.16459354758262634,
0.0862865000963211,
0.20043151080608368,
-0.2896837890148163,
0.005618520081043243,
0.0522591732442379,
0.113655686378479,
0.001720004715025425,
-0.01478634774684906,
0.03469560667872429,
-0.004047111142426729,
0.02000080794095993,
-0.006694010924547911,
-0.09492320567369461,
-0.04227415472269058,
0.06858084350824356,
-0.10082858055830002,
-0.0836600661277771,
0.20075856149196625,
-0.01962403394281864,
0.051189083606004715,
0.03468365967273712,
-0.10027022659778595,
-0.05228475481271744,
-0.027127007022500038,
-0.0054167937487363815,
-0.017817793413996696,
0.05698308348655701,
0.01829976961016655,
-0.01984529010951519,
-0.10944897681474686,
0.021823197603225708,
-0.20916391909122467,
0.2347520887851715,
0.014309053309261799,
0.08769264817237854,
-0.20601022243499756,
0.05224275588989258,
-0.09036950021982193,
-0.09105360507965088,
0.021519731730222702,
-0.08566883206367493,
0.041528768837451935,
0.002083035884425044,
-0.07495059072971344,
0.017543062567710876,
0.047651853412389755,
0.14729511737823486,
-0.014067264273762703,
0.025938095524907112,
0.0008498468669131398,
0.1009284183382988,
0.030102098360657692,
0.13388332724571228,
0.011944600380957127,
-0.02840438298881054,
0.01732935756444931,
-0.1329825222492218,
-0.036133281886577606,
-0.051770441234111786,
-0.12022732943296432,
-0.047808099538087845,
0.045948151499032974,
0.08258914202451706,
0.022342177107930183,
0.005320239346474409,
-0.08945681154727936,
-0.015853654593229294,
0.059366438537836075,
-0.07202005386352539,
0.0043303403072059155,
-0.000528825621586293,
0.03741177171468735,
0.18082594871520996,
-0.029030190780758858,
-0.03096860460937023,
-0.04084136709570885,
0.09087810665369034,
-0.08448517322540283,
0.014714745804667473,
-0.05083002150058746,
-0.062201209366321564,
0.03654183819890022,
-0.16100244224071503,
0.057043030858039856,
-0.14560800790786743,
-0.09376321732997894,
0.039873939007520676,
0.05335168540477753,
-0.002770928665995598,
0.03388800472021103,
0.0009603195358067751,
-0.018828926607966423,
-0.007003897335380316,
-0.06842755526304245,
-0.08064226061105728,
-0.06478630751371384,
0.09923757612705231,
0.00029352123965509236,
0.04672098159790039,
-0.12116226553916931,
0.08672743290662766,
-0.09672944992780685,
0.03750492259860039,
-0.16180843114852905,
-0.008461673744022846,
-0.02150292508304119,
0.16466088593006134,
-0.005621116608381271,
-0.07621955871582031,
-0.11902043223381042,
0.047062430530786514,
-0.035698723047971725,
0.14959093928337097,
-0.05622879043221474,
-0.13087990880012512,
0.20922459661960602,
-0.11588329821825027,
-0.1731778234243393,
0.05696214735507965,
-0.008520878851413727,
-0.015484592877328396,
0.06328354775905609,
0.20762446522712708,
0.0946868285536766,
-0.044678352773189545,
0.08127391338348389,
0.1397486925125122,
-0.11663860827684402,
-0.14153334498405457,
0.027016308158636093,
-0.042085859924554825,
-0.07503706961870193,
0.041700031608343124,
0.011108353734016418,
0.09523002058267593,
-0.0771002322435379,
-0.030855905264616013,
-0.01680048368871212,
-0.01671411283314228,
0.08776957541704178,
0.055555809289216995,
0.10667075961828232,
-0.03498975560069084,
0.011802191846072674,
0.03473251312971115,
-0.014134183526039124,
0.0038637330289930105,
0.04732990637421608,
-0.060643456876277924,
0.1918257176876068,
-0.07546142488718033,
0.003813444171100855,
-0.2566893398761749,
-0.06994711607694626,
0.00528657017275691,
0.06658254563808441,
-0.04811577871441841,
0.15633022785186768,
0.09267178922891617,
-0.07167497277259827,
0.016395436599850655,
-0.03712666034698486,
0.08581963181495667,
0.02451246976852417,
-0.033985935151576996,
-0.04179177060723305,
-0.011763811111450195,
-0.07896506786346436,
-0.09305226802825928,
-0.010189443826675415,
-0.01599530130624771,
0.09464821219444275,
0.10299675166606903,
0.01377725787460804,
0.03143538534641266,
-0.06555043905973434,
0.06333392858505249,
-0.022575706243515015,
0.018025120720267296,
0.09356045722961426,
-0.021321585401892662,
-0.06097353622317314,
0.13765394687652588,
-0.094614177942276,
0.3545747399330139,
0.19013811647891998,
-0.3164325952529907,
0.018932169303297997,
-0.04782366380095482,
-0.00977459829300642,
0.036739129573106766,
0.09889727830886841,
-0.029969865456223488,
0.09724640101194382,
0.027196763083338737,
0.13298392295837402,
-0.030379584059119225,
-0.03850182890892029,
0.0005746455863118172,
-0.02981492318212986,
-0.05865705385804176,
0.07086624205112457,
0.07261653989553452,
-0.15486454963684082,
0.16527150571346283,
0.26989927887916565,
0.03292781859636307,
0.12406831979751587,
-0.0661960020661354,
-0.0376574881374836,
0.019656166434288025,
0.009992681443691254,
-0.040347639471292496,
0.046730417758226395,
-0.2736447751522064,
-0.052413634955883026,
0.0692915990948677,
0.02405851148068905,
0.0952937975525856,
-0.12900009751319885,
-0.04021916538476944,
0.03706459701061249,
0.014877407811582088,
-0.08638262748718262,
0.0609026663005352,
0.05720122531056404,
0.05891529843211174,
0.019035693258047104,
-0.06259000301361084,
0.10445338487625122,
0.00871498417109251,
-0.04073162376880646,
0.17662526667118073,
-0.123042032122612,
-0.2620071768760681,
-0.12311730533838272,
-0.15138909220695496,
0.015030097216367722,
0.01922003924846649,
0.08840186893939972,
-0.06919815391302109,
-0.031516414135694504,
0.06873156130313873,
0.03529629856348038,
-0.14966927468776703,
0.029128234833478928,
-0.07160747051239014,
0.03280840069055557,
-0.07802597433328629,
-0.0789455845952034,
-0.06994608044624329,
-0.06953756511211395,
-0.009727881290018559,
0.09203583002090454,
-0.1189604178071022,
0.10165828466415405,
0.13272233307361603,
0.0363771878182888,
0.08464014530181885,
-0.008833634667098522,
0.1828852891921997,
-0.05849464610219002,
-0.0809895321726799,
0.21144624054431915,
-0.04404043033719063,
0.0901838093996048,
0.09279557317495346,
0.039735835045576096,
-0.06724223494529724,
-0.04217485338449478,
-0.06174955144524574,
-0.10350338369607925,
-0.18064221739768982,
-0.09564316272735596,
-0.14665037393569946,
0.014125813730061054,
0.01768936589360237,
0.03677418455481529,
0.08474641293287277,
0.06241949275135994,
0.0594039149582386,
-0.04961218684911728,
-0.04142293706536293,
0.02869315631687641,
0.21291007101535797,
-0.022173935547471046,
0.09922134131193161,
-0.0371907539665699,
-0.0987340658903122,
0.07145325094461441,
0.03712000325322151,
0.24907712638378143,
0.10089749097824097,
0.035659585148096085,
0.0526948980987072,
0.17874057590961456,
0.13417154550552368,
0.17221486568450928,
-0.020648453384637833,
-0.02667684480547905,
-0.01199409831315279,
0.002160619245842099,
-0.05968237295746803,
0.013288183137774467,
0.17365947365760803,
-0.1238851547241211,
-0.08988935500383377,
-0.2173749804496765,
0.07477087527513504,
0.06237781420350075,
0.031422216445207596,
-0.18417076766490936,
0.003690242301672697,
0.0649406686425209,
-0.00521667068824172,
-0.04713871702551842,
0.07263871282339096,
-0.017426704987883568,
-0.11952719837427139,
0.034282006323337555,
-0.05946141108870506,
0.10967221111059189,
-0.012345034629106522,
0.07638204097747803,
-0.026442036032676697,
-0.12944473326206207,
0.06673547625541687,
0.07164321839809418,
-0.19538965821266174,
0.2909071147441864,
-0.012455099262297153,
-0.06202757731080055,
-0.04829925298690796,
0.005193088203668594,
0.012985559180378914,
0.160504549741745,
0.1260947436094284,
0.022533811628818512,
-0.0578787624835968,
-0.1674206554889679,
0.04693714529275894,
0.028735041618347168,
0.13825884461402893,
-0.05239944905042648,
-0.014248479157686234,
-0.02063719928264618,
-0.011673306114971638,
-0.020329756662249565,
0.05532294511795044,
0.08330490440130234,
-0.14671963453292847,
0.052545733749866486,
-0.06558147072792053,
0.0409327894449234,
-0.012272844091057777,
-0.00765871349722147,
-0.047818366438150406,
0.14944994449615479,
-0.0652761161327362,
-0.048871852457523346,
-0.1053028553724289,
-0.10936398059129715,
0.12579044699668884,
-0.0932580903172493,
0.09765078127384186,
-0.06662343442440033,
-0.03391622006893158,
-0.06140284985303879,
-0.20473268628120422,
0.1305530071258545,
-0.09915255755186081,
0.04976314306259155,
-0.038567788898944855,
0.181097149848938,
-0.060539573431015015,
0.005303262732923031,
0.026644282042980194,
0.01667754538357258,
-0.11381230503320694,
-0.08549114316701889,
-0.015935804694890976,
-0.016310077160596848,
0.0518048033118248,
0.06849704682826996,
-0.062081992626190186,
0.04358658939599991,
0.00031327479518949986,
0.06365568935871124,
0.24294216930866241,
0.14630047976970673,
-0.051170576363801956,
0.11532986164093018,
0.14450006186962128,
-0.031972140073776245,
-0.27944883704185486,
-0.06474616378545761,
-0.13329371809959412,
-0.03996938094496727,
-0.0014406027039512992,
-0.13789600133895874,
0.14261819422245026,
0.044858112931251526,
-0.006724648643285036,
0.11656755954027176,
-0.24167363345623016,
-0.05383818969130516,
0.16473250091075897,
0.01709013618528843,
0.44809144735336304,
-0.11115505546331406,
-0.09658455848693848,
0.012760866433382034,
-0.2474503070116043,
0.10122992098331451,
0.014938225969672203,
0.0646388828754425,
-0.02196873165667057,
0.05254409834742546,
0.0362543947994709,
-0.0707058236002922,
0.1153612732887268,
0.036213282495737076,
0.044825151562690735,
-0.05749816447496414,
-0.12735342979431152,
0.05762515589594841,
-0.0197214987128973,
-0.0027702997904270887,
0.06281710416078568,
0.01907024346292019,
-0.15264934301376343,
-0.028649672865867615,
-0.1261925846338272,
0.07312501221895218,
0.036964092403650284,
-0.035512425005435944,
0.003519417019560933,
-0.027797704562544823,
-0.003173819277435541,
0.024623822420835495,
0.2571793794631958,
-0.02694554440677166,
0.12917625904083252,
0.020508399233222008,
0.05970452353358269,
-0.2093375027179718,
-0.1727571189403534,
-0.06916570663452148,
-0.04265722632408142,
0.09737759083509445,
-0.03340640291571617,
0.05482974648475647,
0.14861060678958893,
-0.01582632027566433,
0.028938481584191322,
0.1354859471321106,
0.012413929216563702,
-0.017079832032322884,
0.12271753698587418,
-0.1915288269519806,
-0.02804369106888771,
-0.03969001770019531,
-0.08020538836717606,
0.08018501847982407,
0.07493545114994049,
0.09324722737073898,
0.06881885975599289,
-0.012770483270287514,
-0.0276129599660635,
-0.03462065011262894,
-0.07952789962291718,
0.0472898855805397,
0.03613487631082535,
0.03859974071383476,
-0.13946987688541412,
0.03557739406824112,
-0.016765620559453964,
-0.2670823633670807,
-0.05293544381856918,
0.07872454822063446,
-0.11887070536613464,
-0.10548925399780273,
-0.08829343318939209,
0.11399594694375992,
-0.1680411845445633,
-0.027427881956100464,
-0.03792887181043625,
-0.12483807653188705,
0.07226365804672241,
0.21823963522911072,
0.10136004537343979,
0.1148689016699791,
-0.045093145221471786,
0.0007684463635087013,
0.024442041292786598,
-0.0601712167263031,
0.01604810729622841,
0.016433462500572205,
-0.0998239740729332,
-0.0008209992665797472,
-0.020151354372501373,
0.15397760272026062,
-0.0796075239777565,
-0.06533236801624298,
-0.15189164876937866,
0.07058283686637878,
-0.0721585750579834,
-0.08680808544158936,
-0.1218729019165039,
-0.05743737146258354,
0.024261746555566788,
-0.05391566827893257,
-0.03601766377687454,
-0.021656449884176254,
-0.14623980224132538,
0.04485916718840599,
0.010332317091524601,
-0.005168382078409195,
-0.06595148891210556,
-0.030757123604416847,
0.11516020447015762,
-0.06426966190338135,
0.07983355224132538,
0.19698777794837952,
-0.06767240911722183,
0.13421446084976196,
-0.12187573313713074,
-0.17356571555137634,
0.10550028830766678,
0.028678571805357933,
0.08807408809661865,
0.07333525270223618,
0.031008796766400337,
0.07151902467012405,
-0.0006563866045325994,
0.03527871519327164,
-0.026205046102404594,
-0.13518503308296204,
-0.022110003978013992,
0.008841688744723797,
-0.18046444654464722,
-0.03447321802377701,
-0.07454892992973328,
0.15258778631687164,
0.02735028602182865,
0.11539342999458313,
0.016432136297225952,
0.10955704003572464,
-0.04023033380508423,
-0.0037069369573146105,
0.006354177836328745,
-0.1867031753063202,
0.027574164792895317,
-0.07603352516889572,
0.009099473245441914,
0.004562267567962408,
0.26294976472854614,
-0.016851287335157394,
0.06874511390924454,
0.019947905093431473,
0.0317598320543766,
0.06296208500862122,
0.029733985662460327,
0.2593200206756592,
0.1191045418381691,
-0.04764638841152191,
-0.09192530065774918,
0.09628601372241974,
0.0015705273253843188,
0.0005639182054437697,
0.12658140063285828,
0.12427230179309845,
0.03164798021316528,
0.10732870548963547,
0.03856581449508667,
0.04636436328291893,
-0.1138598844408989,
-0.24980735778808594,
0.006955720484256744,
0.07651596516370773,
0.030513912439346313,
0.09611864387989044,
0.11702555418014526,
-0.040339987725019455,
0.09750796854496002,
-0.006009524688124657,
-0.03278311714529991,
-0.1306307315826416,
-0.048490166664123535,
-0.05248153209686279,
-0.1226581484079361,
0.010182006284594536,
-0.07812222838401794,
0.0005351413274183869,
0.15948551893234253,
0.02424205094575882,
-0.019147196784615517,
0.14117267727851868,
0.03290129452943802,
-0.05950130149722099,
0.0636630430817604,
-0.017126450315117836,
0.0027530856896191835,
0.04465898126363754,
-0.016934748739004135,
-0.11656554788351059,
-0.08354712277650833,
-0.050883643329143524,
0.024923330172896385,
-0.09700026363134384,
0.004247597418725491,
-0.11320921033620834,
-0.11535467207431793,
-0.04546694457530975,
0.048496924340724945,
-0.07876671850681305,
0.1265949159860611,
-0.016160227358341217,
0.0028664041310548782,
0.009360520169138908,
0.15711569786071777,
-0.07741065323352814,
-0.012710070237517357,
-0.020163269713521004,
0.21036958694458008,
0.0989537462592125,
0.10912171006202698,
-0.0009331207838840783,
0.015002057887613773,
-0.05219440162181854,
0.29989543557167053,
0.2444397360086441,
-0.028167791664600372,
0.04460100829601288,
0.0873914435505867,
0.038083989173173904,
0.0968138724565506,
0.09690817445516586,
0.10102183371782303,
0.3200364112854004,
-0.07682528346776962,
-0.045907747000455856,
-0.03448496013879776,
-0.00849792081862688,
-0.09551121294498444,
0.035898253321647644,
0.07948186993598938,
-0.06152723729610443,
-0.07321841269731522,
0.10746033489704132,
-0.16515599191188812,
0.11578583717346191,
0.11013679206371307,
-0.20921821892261505,
-0.049416057765483856,
-0.07384659349918365,
0.17109091579914093,
-0.0024887267500162125,
0.1266387403011322,
-0.04354546591639519,
-0.13439299166202545,
0.06285285204648972,
0.048469748347997665,
-0.2756679356098175,
-0.0980004072189331,
0.11060073226690292,
0.024665147066116333,
0.007802571170032024,
-0.02169145829975605,
-0.004367694724351168,
0.06355959177017212,
0.09860705584287643,
-0.0003893781395163387,
0.02548590861260891,
0.03634477034211159,
-0.10912884771823883,
-0.08365445584058762,
0.008976423181593418,
0.008719995617866516,
-0.08448972553014755,
0.023103585466742516,
-0.19627802073955536,
0.044169794768095016,
-0.0157186109572649,
-0.044235143810510635,
0.003991053905338049,
-0.029606744647026062,
-0.059212785214185715,
0.03327903896570206,
0.08572027087211609,
0.023666096851229668,
-0.03819211944937706,
-0.05403012037277222,
0.0019490038976073265,
0.07524687796831131,
-0.06805311143398285,
-0.173606276512146,
-0.04840735346078873,
-0.08657971769571304,
0.0992082953453064,
-0.05511601269245148,
-0.0752784013748169,
-0.04562166705727577,
-0.03257971256971359,
0.06480588763952255,
-0.12087775766849518,
0.04490482062101364,
0.03911556303501129,
0.041156258434057236,
0.0160878524184227,
-0.035335659980773926,
0.044597625732421875,
0.07137373089790344,
-0.11931539326906204,
-0.07447930425405502
] |
null | null |
transformers
|
hello
|
{}
|
feature-extraction
|
dmis-lab/biosyn-sapbert-bc2gn
|
[
"transformers",
"pytorch",
"bert",
"feature-extraction",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #feature-extraction #endpoints_compatible #region-us
|
hello
|
[] |
[
"TAGS\n#transformers #pytorch #bert #feature-extraction #endpoints_compatible #region-us \n"
] |
[
29
] |
[
"passage: TAGS\n#transformers #pytorch #bert #feature-extraction #endpoints_compatible #region-us \n"
] |
[
-0.0680389553308487,
-0.01353863999247551,
-0.009260591119527817,
0.003671469632536173,
0.13468711078166962,
0.03987877443432808,
-0.0037161505315452814,
0.08307137340307236,
0.06908576935529709,
-0.009869525209069252,
0.10839105397462845,
0.22950756549835205,
-0.03434249758720398,
0.027836797758936882,
-0.06780551373958588,
-0.2686935067176819,
0.08416040241718292,
0.1071663498878479,
-0.04096720367670059,
0.08816211670637131,
0.04718845710158348,
-0.10185960680246353,
0.059242263436317444,
-0.017401615157723427,
-0.13960762321949005,
0.0563788041472435,
0.027032675221562386,
-0.08215486258268356,
0.11170924454927444,
0.02429274097084999,
0.171868696808815,
0.00669759139418602,
-0.10259579122066498,
-0.1729365438222885,
0.024251695722341537,
-0.016173768788576126,
-0.06165656819939613,
0.030105262994766235,
0.07771246135234833,
-0.10020971298217773,
0.020183663815259933,
0.09852731972932816,
0.01621491089463234,
0.024930456653237343,
-0.16311487555503845,
-0.17609833180904388,
-0.05667632818222046,
0.04488549754023552,
0.011052189394831657,
0.0923970639705658,
0.017688613384962082,
0.13918437063694,
-0.16459354758262634,
0.0862865000963211,
0.20043151080608368,
-0.2896837890148163,
0.005618520081043243,
0.0522591732442379,
0.113655686378479,
0.001720004715025425,
-0.01478634774684906,
0.03469560667872429,
-0.004047111142426729,
0.02000080794095993,
-0.006694010924547911,
-0.09492320567369461,
-0.04227415472269058,
0.06858084350824356,
-0.10082858055830002,
-0.0836600661277771,
0.20075856149196625,
-0.01962403394281864,
0.051189083606004715,
0.03468365967273712,
-0.10027022659778595,
-0.05228475481271744,
-0.027127007022500038,
-0.0054167937487363815,
-0.017817793413996696,
0.05698308348655701,
0.01829976961016655,
-0.01984529010951519,
-0.10944897681474686,
0.021823197603225708,
-0.20916391909122467,
0.2347520887851715,
0.014309053309261799,
0.08769264817237854,
-0.20601022243499756,
0.05224275588989258,
-0.09036950021982193,
-0.09105360507965088,
0.021519731730222702,
-0.08566883206367493,
0.041528768837451935,
0.002083035884425044,
-0.07495059072971344,
0.017543062567710876,
0.047651853412389755,
0.14729511737823486,
-0.014067264273762703,
0.025938095524907112,
0.0008498468669131398,
0.1009284183382988,
0.030102098360657692,
0.13388332724571228,
0.011944600380957127,
-0.02840438298881054,
0.01732935756444931,
-0.1329825222492218,
-0.036133281886577606,
-0.051770441234111786,
-0.12022732943296432,
-0.047808099538087845,
0.045948151499032974,
0.08258914202451706,
0.022342177107930183,
0.005320239346474409,
-0.08945681154727936,
-0.015853654593229294,
0.059366438537836075,
-0.07202005386352539,
0.0043303403072059155,
-0.000528825621586293,
0.03741177171468735,
0.18082594871520996,
-0.029030190780758858,
-0.03096860460937023,
-0.04084136709570885,
0.09087810665369034,
-0.08448517322540283,
0.014714745804667473,
-0.05083002150058746,
-0.062201209366321564,
0.03654183819890022,
-0.16100244224071503,
0.057043030858039856,
-0.14560800790786743,
-0.09376321732997894,
0.039873939007520676,
0.05335168540477753,
-0.002770928665995598,
0.03388800472021103,
0.0009603195358067751,
-0.018828926607966423,
-0.007003897335380316,
-0.06842755526304245,
-0.08064226061105728,
-0.06478630751371384,
0.09923757612705231,
0.00029352123965509236,
0.04672098159790039,
-0.12116226553916931,
0.08672743290662766,
-0.09672944992780685,
0.03750492259860039,
-0.16180843114852905,
-0.008461673744022846,
-0.02150292508304119,
0.16466088593006134,
-0.005621116608381271,
-0.07621955871582031,
-0.11902043223381042,
0.047062430530786514,
-0.035698723047971725,
0.14959093928337097,
-0.05622879043221474,
-0.13087990880012512,
0.20922459661960602,
-0.11588329821825027,
-0.1731778234243393,
0.05696214735507965,
-0.008520878851413727,
-0.015484592877328396,
0.06328354775905609,
0.20762446522712708,
0.0946868285536766,
-0.044678352773189545,
0.08127391338348389,
0.1397486925125122,
-0.11663860827684402,
-0.14153334498405457,
0.027016308158636093,
-0.042085859924554825,
-0.07503706961870193,
0.041700031608343124,
0.011108353734016418,
0.09523002058267593,
-0.0771002322435379,
-0.030855905264616013,
-0.01680048368871212,
-0.01671411283314228,
0.08776957541704178,
0.055555809289216995,
0.10667075961828232,
-0.03498975560069084,
0.011802191846072674,
0.03473251312971115,
-0.014134183526039124,
0.0038637330289930105,
0.04732990637421608,
-0.060643456876277924,
0.1918257176876068,
-0.07546142488718033,
0.003813444171100855,
-0.2566893398761749,
-0.06994711607694626,
0.00528657017275691,
0.06658254563808441,
-0.04811577871441841,
0.15633022785186768,
0.09267178922891617,
-0.07167497277259827,
0.016395436599850655,
-0.03712666034698486,
0.08581963181495667,
0.02451246976852417,
-0.033985935151576996,
-0.04179177060723305,
-0.011763811111450195,
-0.07896506786346436,
-0.09305226802825928,
-0.010189443826675415,
-0.01599530130624771,
0.09464821219444275,
0.10299675166606903,
0.01377725787460804,
0.03143538534641266,
-0.06555043905973434,
0.06333392858505249,
-0.022575706243515015,
0.018025120720267296,
0.09356045722961426,
-0.021321585401892662,
-0.06097353622317314,
0.13765394687652588,
-0.094614177942276,
0.3545747399330139,
0.19013811647891998,
-0.3164325952529907,
0.018932169303297997,
-0.04782366380095482,
-0.00977459829300642,
0.036739129573106766,
0.09889727830886841,
-0.029969865456223488,
0.09724640101194382,
0.027196763083338737,
0.13298392295837402,
-0.030379584059119225,
-0.03850182890892029,
0.0005746455863118172,
-0.02981492318212986,
-0.05865705385804176,
0.07086624205112457,
0.07261653989553452,
-0.15486454963684082,
0.16527150571346283,
0.26989927887916565,
0.03292781859636307,
0.12406831979751587,
-0.0661960020661354,
-0.0376574881374836,
0.019656166434288025,
0.009992681443691254,
-0.040347639471292496,
0.046730417758226395,
-0.2736447751522064,
-0.052413634955883026,
0.0692915990948677,
0.02405851148068905,
0.0952937975525856,
-0.12900009751319885,
-0.04021916538476944,
0.03706459701061249,
0.014877407811582088,
-0.08638262748718262,
0.0609026663005352,
0.05720122531056404,
0.05891529843211174,
0.019035693258047104,
-0.06259000301361084,
0.10445338487625122,
0.00871498417109251,
-0.04073162376880646,
0.17662526667118073,
-0.123042032122612,
-0.2620071768760681,
-0.12311730533838272,
-0.15138909220695496,
0.015030097216367722,
0.01922003924846649,
0.08840186893939972,
-0.06919815391302109,
-0.031516414135694504,
0.06873156130313873,
0.03529629856348038,
-0.14966927468776703,
0.029128234833478928,
-0.07160747051239014,
0.03280840069055557,
-0.07802597433328629,
-0.0789455845952034,
-0.06994608044624329,
-0.06953756511211395,
-0.009727881290018559,
0.09203583002090454,
-0.1189604178071022,
0.10165828466415405,
0.13272233307361603,
0.0363771878182888,
0.08464014530181885,
-0.008833634667098522,
0.1828852891921997,
-0.05849464610219002,
-0.0809895321726799,
0.21144624054431915,
-0.04404043033719063,
0.0901838093996048,
0.09279557317495346,
0.039735835045576096,
-0.06724223494529724,
-0.04217485338449478,
-0.06174955144524574,
-0.10350338369607925,
-0.18064221739768982,
-0.09564316272735596,
-0.14665037393569946,
0.014125813730061054,
0.01768936589360237,
0.03677418455481529,
0.08474641293287277,
0.06241949275135994,
0.0594039149582386,
-0.04961218684911728,
-0.04142293706536293,
0.02869315631687641,
0.21291007101535797,
-0.022173935547471046,
0.09922134131193161,
-0.0371907539665699,
-0.0987340658903122,
0.07145325094461441,
0.03712000325322151,
0.24907712638378143,
0.10089749097824097,
0.035659585148096085,
0.0526948980987072,
0.17874057590961456,
0.13417154550552368,
0.17221486568450928,
-0.020648453384637833,
-0.02667684480547905,
-0.01199409831315279,
0.002160619245842099,
-0.05968237295746803,
0.013288183137774467,
0.17365947365760803,
-0.1238851547241211,
-0.08988935500383377,
-0.2173749804496765,
0.07477087527513504,
0.06237781420350075,
0.031422216445207596,
-0.18417076766490936,
0.003690242301672697,
0.0649406686425209,
-0.00521667068824172,
-0.04713871702551842,
0.07263871282339096,
-0.017426704987883568,
-0.11952719837427139,
0.034282006323337555,
-0.05946141108870506,
0.10967221111059189,
-0.012345034629106522,
0.07638204097747803,
-0.026442036032676697,
-0.12944473326206207,
0.06673547625541687,
0.07164321839809418,
-0.19538965821266174,
0.2909071147441864,
-0.012455099262297153,
-0.06202757731080055,
-0.04829925298690796,
0.005193088203668594,
0.012985559180378914,
0.160504549741745,
0.1260947436094284,
0.022533811628818512,
-0.0578787624835968,
-0.1674206554889679,
0.04693714529275894,
0.028735041618347168,
0.13825884461402893,
-0.05239944905042648,
-0.014248479157686234,
-0.02063719928264618,
-0.011673306114971638,
-0.020329756662249565,
0.05532294511795044,
0.08330490440130234,
-0.14671963453292847,
0.052545733749866486,
-0.06558147072792053,
0.0409327894449234,
-0.012272844091057777,
-0.00765871349722147,
-0.047818366438150406,
0.14944994449615479,
-0.0652761161327362,
-0.048871852457523346,
-0.1053028553724289,
-0.10936398059129715,
0.12579044699668884,
-0.0932580903172493,
0.09765078127384186,
-0.06662343442440033,
-0.03391622006893158,
-0.06140284985303879,
-0.20473268628120422,
0.1305530071258545,
-0.09915255755186081,
0.04976314306259155,
-0.038567788898944855,
0.181097149848938,
-0.060539573431015015,
0.005303262732923031,
0.026644282042980194,
0.01667754538357258,
-0.11381230503320694,
-0.08549114316701889,
-0.015935804694890976,
-0.016310077160596848,
0.0518048033118248,
0.06849704682826996,
-0.062081992626190186,
0.04358658939599991,
0.00031327479518949986,
0.06365568935871124,
0.24294216930866241,
0.14630047976970673,
-0.051170576363801956,
0.11532986164093018,
0.14450006186962128,
-0.031972140073776245,
-0.27944883704185486,
-0.06474616378545761,
-0.13329371809959412,
-0.03996938094496727,
-0.0014406027039512992,
-0.13789600133895874,
0.14261819422245026,
0.044858112931251526,
-0.006724648643285036,
0.11656755954027176,
-0.24167363345623016,
-0.05383818969130516,
0.16473250091075897,
0.01709013618528843,
0.44809144735336304,
-0.11115505546331406,
-0.09658455848693848,
0.012760866433382034,
-0.2474503070116043,
0.10122992098331451,
0.014938225969672203,
0.0646388828754425,
-0.02196873165667057,
0.05254409834742546,
0.0362543947994709,
-0.0707058236002922,
0.1153612732887268,
0.036213282495737076,
0.044825151562690735,
-0.05749816447496414,
-0.12735342979431152,
0.05762515589594841,
-0.0197214987128973,
-0.0027702997904270887,
0.06281710416078568,
0.01907024346292019,
-0.15264934301376343,
-0.028649672865867615,
-0.1261925846338272,
0.07312501221895218,
0.036964092403650284,
-0.035512425005435944,
0.003519417019560933,
-0.027797704562544823,
-0.003173819277435541,
0.024623822420835495,
0.2571793794631958,
-0.02694554440677166,
0.12917625904083252,
0.020508399233222008,
0.05970452353358269,
-0.2093375027179718,
-0.1727571189403534,
-0.06916570663452148,
-0.04265722632408142,
0.09737759083509445,
-0.03340640291571617,
0.05482974648475647,
0.14861060678958893,
-0.01582632027566433,
0.028938481584191322,
0.1354859471321106,
0.012413929216563702,
-0.017079832032322884,
0.12271753698587418,
-0.1915288269519806,
-0.02804369106888771,
-0.03969001770019531,
-0.08020538836717606,
0.08018501847982407,
0.07493545114994049,
0.09324722737073898,
0.06881885975599289,
-0.012770483270287514,
-0.0276129599660635,
-0.03462065011262894,
-0.07952789962291718,
0.0472898855805397,
0.03613487631082535,
0.03859974071383476,
-0.13946987688541412,
0.03557739406824112,
-0.016765620559453964,
-0.2670823633670807,
-0.05293544381856918,
0.07872454822063446,
-0.11887070536613464,
-0.10548925399780273,
-0.08829343318939209,
0.11399594694375992,
-0.1680411845445633,
-0.027427881956100464,
-0.03792887181043625,
-0.12483807653188705,
0.07226365804672241,
0.21823963522911072,
0.10136004537343979,
0.1148689016699791,
-0.045093145221471786,
0.0007684463635087013,
0.024442041292786598,
-0.0601712167263031,
0.01604810729622841,
0.016433462500572205,
-0.0998239740729332,
-0.0008209992665797472,
-0.020151354372501373,
0.15397760272026062,
-0.0796075239777565,
-0.06533236801624298,
-0.15189164876937866,
0.07058283686637878,
-0.0721585750579834,
-0.08680808544158936,
-0.1218729019165039,
-0.05743737146258354,
0.024261746555566788,
-0.05391566827893257,
-0.03601766377687454,
-0.021656449884176254,
-0.14623980224132538,
0.04485916718840599,
0.010332317091524601,
-0.005168382078409195,
-0.06595148891210556,
-0.030757123604416847,
0.11516020447015762,
-0.06426966190338135,
0.07983355224132538,
0.19698777794837952,
-0.06767240911722183,
0.13421446084976196,
-0.12187573313713074,
-0.17356571555137634,
0.10550028830766678,
0.028678571805357933,
0.08807408809661865,
0.07333525270223618,
0.031008796766400337,
0.07151902467012405,
-0.0006563866045325994,
0.03527871519327164,
-0.026205046102404594,
-0.13518503308296204,
-0.022110003978013992,
0.008841688744723797,
-0.18046444654464722,
-0.03447321802377701,
-0.07454892992973328,
0.15258778631687164,
0.02735028602182865,
0.11539342999458313,
0.016432136297225952,
0.10955704003572464,
-0.04023033380508423,
-0.0037069369573146105,
0.006354177836328745,
-0.1867031753063202,
0.027574164792895317,
-0.07603352516889572,
0.009099473245441914,
0.004562267567962408,
0.26294976472854614,
-0.016851287335157394,
0.06874511390924454,
0.019947905093431473,
0.0317598320543766,
0.06296208500862122,
0.029733985662460327,
0.2593200206756592,
0.1191045418381691,
-0.04764638841152191,
-0.09192530065774918,
0.09628601372241974,
0.0015705273253843188,
0.0005639182054437697,
0.12658140063285828,
0.12427230179309845,
0.03164798021316528,
0.10732870548963547,
0.03856581449508667,
0.04636436328291893,
-0.1138598844408989,
-0.24980735778808594,
0.006955720484256744,
0.07651596516370773,
0.030513912439346313,
0.09611864387989044,
0.11702555418014526,
-0.040339987725019455,
0.09750796854496002,
-0.006009524688124657,
-0.03278311714529991,
-0.1306307315826416,
-0.048490166664123535,
-0.05248153209686279,
-0.1226581484079361,
0.010182006284594536,
-0.07812222838401794,
0.0005351413274183869,
0.15948551893234253,
0.02424205094575882,
-0.019147196784615517,
0.14117267727851868,
0.03290129452943802,
-0.05950130149722099,
0.0636630430817604,
-0.017126450315117836,
0.0027530856896191835,
0.04465898126363754,
-0.016934748739004135,
-0.11656554788351059,
-0.08354712277650833,
-0.050883643329143524,
0.024923330172896385,
-0.09700026363134384,
0.004247597418725491,
-0.11320921033620834,
-0.11535467207431793,
-0.04546694457530975,
0.048496924340724945,
-0.07876671850681305,
0.1265949159860611,
-0.016160227358341217,
0.0028664041310548782,
0.009360520169138908,
0.15711569786071777,
-0.07741065323352814,
-0.012710070237517357,
-0.020163269713521004,
0.21036958694458008,
0.0989537462592125,
0.10912171006202698,
-0.0009331207838840783,
0.015002057887613773,
-0.05219440162181854,
0.29989543557167053,
0.2444397360086441,
-0.028167791664600372,
0.04460100829601288,
0.0873914435505867,
0.038083989173173904,
0.0968138724565506,
0.09690817445516586,
0.10102183371782303,
0.3200364112854004,
-0.07682528346776962,
-0.045907747000455856,
-0.03448496013879776,
-0.00849792081862688,
-0.09551121294498444,
0.035898253321647644,
0.07948186993598938,
-0.06152723729610443,
-0.07321841269731522,
0.10746033489704132,
-0.16515599191188812,
0.11578583717346191,
0.11013679206371307,
-0.20921821892261505,
-0.049416057765483856,
-0.07384659349918365,
0.17109091579914093,
-0.0024887267500162125,
0.1266387403011322,
-0.04354546591639519,
-0.13439299166202545,
0.06285285204648972,
0.048469748347997665,
-0.2756679356098175,
-0.0980004072189331,
0.11060073226690292,
0.024665147066116333,
0.007802571170032024,
-0.02169145829975605,
-0.004367694724351168,
0.06355959177017212,
0.09860705584287643,
-0.0003893781395163387,
0.02548590861260891,
0.03634477034211159,
-0.10912884771823883,
-0.08365445584058762,
0.008976423181593418,
0.008719995617866516,
-0.08448972553014755,
0.023103585466742516,
-0.19627802073955536,
0.044169794768095016,
-0.0157186109572649,
-0.044235143810510635,
0.003991053905338049,
-0.029606744647026062,
-0.059212785214185715,
0.03327903896570206,
0.08572027087211609,
0.023666096851229668,
-0.03819211944937706,
-0.05403012037277222,
0.0019490038976073265,
0.07524687796831131,
-0.06805311143398285,
-0.173606276512146,
-0.04840735346078873,
-0.08657971769571304,
0.0992082953453064,
-0.05511601269245148,
-0.0752784013748169,
-0.04562166705727577,
-0.03257971256971359,
0.06480588763952255,
-0.12087775766849518,
0.04490482062101364,
0.03911556303501129,
0.041156258434057236,
0.0160878524184227,
-0.035335659980773926,
0.044597625732421875,
0.07137373089790344,
-0.11931539326906204,
-0.07447930425405502
] |
null | null |
transformers
|
# Model Card for biosyn-sapbert-ncbi-disease
# Model Details
## Model Description
More information needed
- **Developed by:** Dmis-lab (Data Mining and Information Systems Lab, Korea University)
- **Shared by [Optional]:** Hugging Face
- **Model type:** Feature Extraction
- **Language(s) (NLP):** More information needed
- **License:** More information needed
- **Related Models:**
- **Parent Model:** BERT
- **Resources for more information:**
- [GitHub Repo](https://github.com/jhyuklee/biobert)
- [Associated Paper](https://arxiv.org/abs/1901.08746)
# Uses
## Direct Use
This model can be used for the task of Feature Extraction
## Downstream Use [Optional]
More information needed
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., [Sheng et al. (2021)](https://aclanthology.org/2021.acl-long.330.pdf) and [Bender et al. (2021)](https://dl.acm.org/doi/pdf/10.1145/3442188.3445922)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
The model creators note in the [associated paper](https://arxiv.org/pdf/1901.08746.pdf)
> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))
## Training Procedure
### Preprocessing
The model creators note in the [associated paper](https://arxiv.org/pdf/1901.08746.pdf)
> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs
### Speeds, Sizes, Times
The model creators note in the [associated paper](https://arxiv.org/pdf/1901.08746.pdf)
> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed
### Factors
More information needed
### Metrics
More information needed
## Results
More information needed
# Model Examination
More information needed
# Environmental Impact
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:**
- **Training:** Eight NVIDIA V100 (32GB) GPUs [ for training],
- **Fine-tuning:** a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task
- **Hours used:** More information needed
- **Cloud Provider:** More information needed
- **Compute Region:** More information needed
- **Carbon Emitted:** More information needed
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed
## Compute Infrastructure
More information needed
### Hardware
More information needed
### Software
More information needed
# Citation
**BibTeX:**
```
@article{lee2019biobert,
title={BioBERT: a pre-trained biomedical language representation model for biomedical text mining},
author={Lee, Jinhyuk and Yoon, Wonjin and Kim, Sungdong and Kim, Donghyeon and Kim, Sunkyu and So, Chan Ho and Kang, Jaewoo},
journal={arXiv preprint arXiv:1901.08746},
year={2019}
}
```
# Glossary [optional]
More information needed
# More Information [optional]
For help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee(`lee.jnhk (at) gmail.com`), or Wonjin Yoon (`wonjin.info (at) gmail.com`) for communication related to BioBERT.
# Model Card Authors [optional]
Dmis-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
```python
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("dmis-lab/biosyn-sapbert-ncbi-disease")
model = AutoModel.from_pretrained("dmis-lab/biosyn-sapbert-ncbi-disease")
```
</details>
|
{"tags": ["bert"]}
|
feature-extraction
|
dmis-lab/biosyn-sapbert-ncbi-disease
|
[
"transformers",
"pytorch",
"bert",
"feature-extraction",
"arxiv:1901.08746",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1901.08746",
"1910.09700"
] |
[] |
TAGS
#transformers #pytorch #bert #feature-extraction #arxiv-1901.08746 #arxiv-1910.09700 #endpoints_compatible #region-us
|
# Model Card for biosyn-sapbert-ncbi-disease
# Model Details
## Model Description
More information needed
- Developed by: Dmis-lab (Data Mining and Information Systems Lab, Korea University)
- Shared by [Optional]: Hugging Face
- Model type: Feature Extraction
- Language(s) (NLP): More information needed
- License: More information needed
- Related Models:
- Parent Model: BERT
- Resources for more information:
- GitHub Repo
- Associated Paper
# Uses
## Direct Use
This model can be used for the task of Feature Extraction
## Downstream Use [Optional]
More information needed
## Out-of-Scope Use
The model should not be used to intentionally create hostile or alienating environments for people.
# Bias, Risks, and Limitations
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
## Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
# Training Details
## Training Data
The model creators note in the associated paper
> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))
## Training Procedure
### Preprocessing
The model creators note in the associated paper
> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs
### Speeds, Sizes, Times
The model creators note in the associated paper
> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.
# Evaluation
## Testing Data, Factors & Metrics
### Testing Data
More information needed
### Factors
More information needed
### Metrics
More information needed
## Results
More information needed
# Model Examination
More information needed
# Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type:
- Training: Eight NVIDIA V100 (32GB) GPUs [ for training],
- Fine-tuning: a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task
- Hours used: More information needed
- Cloud Provider: More information needed
- Compute Region: More information needed
- Carbon Emitted: More information needed
# Technical Specifications [optional]
## Model Architecture and Objective
More information needed
## Compute Infrastructure
More information needed
### Hardware
More information needed
### Software
More information needed
BibTeX:
# Glossary [optional]
More information needed
# More Information [optional]
For help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee('URL (at) URL'), or Wonjin Yoon ('URL (at) URL') for communication related to BioBERT.
# Model Card Authors [optional]
Dmis-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team
# Model Card Contact
More information needed
# How to Get Started with the Model
Use the code below to get started with the model.
<details>
<summary> Click to expand </summary>
</details>
|
[
"# Model Card for biosyn-sapbert-ncbi-disease",
"# Model Details",
"## Model Description\n \nMore information needed\n \n- Developed by: Dmis-lab (Data Mining and Information Systems Lab, Korea University)\n- Shared by [Optional]: Hugging Face\n- Model type: Feature Extraction\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Related Models: \n - Parent Model: BERT\n- Resources for more information: \n - GitHub Repo\n - Associated Paper",
"# Uses",
"## Direct Use\n \nThis model can be used for the task of Feature Extraction",
"## Downstream Use [Optional]\n \nMore information needed",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\nThe model creators note in the associated paper\n> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))",
"## Training Procedure",
"### Preprocessing\n The model creators note in the associated paper\n> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs",
"### Speeds, Sizes, Times\n The model creators note in the associated paper\n> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed",
"### Factors\n \nMore information needed",
"### Metrics\n \n \n \nMore information needed",
"## Results \nMore information needed",
"# Model Examination\n \nMore information needed",
"# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: \n- Training: Eight NVIDIA V100 (32GB) GPUs [ for training], \n- Fine-tuning: a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed",
"## Compute Infrastructure\n \nMore information needed",
"### Hardware\n \nMore information needed",
"### Software\n \nMore information needed\n \nBibTeX:",
"# Glossary [optional]\n \nMore information needed",
"# More Information [optional]\n \nFor help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee('URL (at) URL'), or Wonjin Yoon ('URL (at) URL') for communication related to BioBERT.",
"# Model Card Authors [optional]\n \n \n Dmis-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] |
[
"TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-1901.08746 #arxiv-1910.09700 #endpoints_compatible #region-us \n",
"# Model Card for biosyn-sapbert-ncbi-disease",
"# Model Details",
"## Model Description\n \nMore information needed\n \n- Developed by: Dmis-lab (Data Mining and Information Systems Lab, Korea University)\n- Shared by [Optional]: Hugging Face\n- Model type: Feature Extraction\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Related Models: \n - Parent Model: BERT\n- Resources for more information: \n - GitHub Repo\n - Associated Paper",
"# Uses",
"## Direct Use\n \nThis model can be used for the task of Feature Extraction",
"## Downstream Use [Optional]\n \nMore information needed",
"## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.",
"# Bias, Risks, and Limitations\n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.",
"## Recommendations\n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
"# Training Details",
"## Training Data\nThe model creators note in the associated paper\n> We used the BERTBASE model pre-trained on English Wikipedia and BooksCorpus for 1M steps. BioBERT v1.0 (þ PubMed þ PMC) is the version of BioBERT (þ PubMed þ PMC) trained for 470 K steps. When using both the PubMed and PMC corpora, we found that 200K and 270K pre-training steps were optimal for PubMed and PMC, respectively. We also used the ablated versions of BioBERT v1.0, which were pre-trained on only PubMed for 200K steps (BioBERT v1.0 (þ PubMed)) and PMC for 270K steps (BioBERT v1.0 (þ PMC))",
"## Training Procedure",
"### Preprocessing\n The model creators note in the associated paper\n> We pre-trained BioBERT using Naver Smart Machine Learning (NSML) (Sung et al., 2017), which is utilized for large-scale experiments that need to be run on several GPUs",
"### Speeds, Sizes, Times\n The model creators note in the associated paper\n> The maximum sequence length was fixed to 512 and the mini-batch size was set to 192, resulting in 98 304 words per iteration.",
"# Evaluation",
"## Testing Data, Factors & Metrics",
"### Testing Data\n \nMore information needed",
"### Factors\n \nMore information needed",
"### Metrics\n \n \n \nMore information needed",
"## Results \nMore information needed",
"# Model Examination\n \nMore information needed",
"# Environmental Impact\n \nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n \n- Hardware Type: \n- Training: Eight NVIDIA V100 (32GB) GPUs [ for training], \n- Fine-tuning: a single NVIDIA Titan Xp (12GB) GPU to fine-tune BioBERT on each task\n- Hours used: More information needed\n- Cloud Provider: More information needed\n- Compute Region: More information needed\n- Carbon Emitted: More information needed",
"# Technical Specifications [optional]",
"## Model Architecture and Objective\n \nMore information needed",
"## Compute Infrastructure\n \nMore information needed",
"### Hardware\n \nMore information needed",
"### Software\n \nMore information needed\n \nBibTeX:",
"# Glossary [optional]\n \nMore information needed",
"# More Information [optional]\n \nFor help or issues using BioBERT, please submit a GitHub issue. Please contact Jinhyuk Lee('URL (at) URL'), or Wonjin Yoon ('URL (at) URL') for communication related to BioBERT.",
"# Model Card Authors [optional]\n \n \n Dmis-lab (Data Mining and Information Systems Lab, Korea University) in collaboration with Ezi Ozoani and the Hugging Face team",
"# Model Card Contact\n \nMore information needed",
"# How to Get Started with the Model\n \nUse the code below to get started with the model.\n \n<details>\n<summary> Click to expand </summary>\n\n\n</details>"
] |
[
47,
16,
3,
91,
3,
15,
11,
28,
87,
41,
3,
167,
4,
62,
52,
3,
11,
8,
7,
8,
5,
8,
110,
9,
10,
8,
6,
11,
11,
61,
39,
7,
41
] |
[
"passage: TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-1901.08746 #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for biosyn-sapbert-ncbi-disease# Model Details## Model Description\n \nMore information needed\n \n- Developed by: Dmis-lab (Data Mining and Information Systems Lab, Korea University)\n- Shared by [Optional]: Hugging Face\n- Model type: Feature Extraction\n- Language(s) (NLP): More information needed\n- License: More information needed\n- Related Models: \n - Parent Model: BERT\n- Resources for more information: \n - GitHub Repo\n - Associated Paper# Uses## Direct Use\n \nThis model can be used for the task of Feature Extraction## Downstream Use [Optional]\n \nMore information needed## Out-of-Scope Use\n \nThe model should not be used to intentionally create hostile or alienating environments for people.# Bias, Risks, and Limitations\n \nSignificant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). Predictions generated by the model may include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.## Recommendations\n \nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.# Training Details"
] |
[
-0.05837039649486542,
0.19050396978855133,
-0.004745954181998968,
0.05123256891965866,
0.09099528193473816,
-0.022946225479245186,
0.1499883383512497,
0.06451559066772461,
0.1015581488609314,
0.05697528272867203,
0.04272680729627609,
0.015592876821756363,
0.09238474071025848,
0.08709163963794708,
0.0016435913275927305,
-0.2685844600200653,
0.0630105584859848,
-0.0030422883573919535,
0.12338737398386002,
0.11541381478309631,
0.08386088907718658,
-0.07630259543657303,
0.11400700360536575,
0.03538466617465019,
-0.01701013743877411,
-0.031246965751051903,
0.005062960088253021,
-0.020443342626094818,
0.12149504572153091,
0.012131379917263985,
0.10902182757854462,
0.044937700033187866,
0.03235592693090439,
-0.23404832184314728,
0.035133130848407745,
0.03956669569015503,
-0.008192161098122597,
0.08290904760360718,
0.13294677436351776,
-0.08802198618650436,
0.1727852076292038,
-0.05757834389805794,
0.08917959034442902,
0.051417026668787,
-0.08859588205814362,
-0.12695898115634918,
-0.11459922790527344,
0.09331375360488892,
0.06806109100580215,
0.05813933163881302,
-0.05929969996213913,
0.1750650405883789,
-0.09823668748140335,
0.0351671427488327,
0.2307063788175583,
-0.09876401722431183,
0.012131760828197002,
-0.0412290021777153,
0.038372743874788284,
0.012123298831284046,
-0.04706180840730667,
-0.037605468183755875,
0.006164220627397299,
0.02811058983206749,
0.06223168969154358,
-0.0718684196472168,
0.05502539128065109,
-0.09033878892660141,
-0.12135475873947144,
0.004584230482578278,
0.10710400342941284,
0.07382279634475708,
-0.08904672414064407,
-0.23883892595767975,
-0.008950823917984962,
0.0872630625963211,
0.036681219935417175,
-0.07448694109916687,
0.013198701664805412,
-0.07133720070123672,
0.08295922726392746,
0.0070903063751757145,
-0.04088493436574936,
0.007298407610505819,
0.002388046821579337,
0.17367208003997803,
0.0399913527071476,
0.03483203426003456,
-0.002324076136574149,
0.04161620885133743,
0.03394559025764465,
-0.0701863169670105,
-0.09958044439554214,
-0.11225082725286484,
-0.044969554990530014,
-0.03749072179198265,
-0.005657073576003313,
0.004277570638805628,
0.033207714557647705,
0.10657648742198944,
-0.10334715247154236,
0.024581801146268845,
-0.057184867560863495,
0.012611126527190208,
0.18058758974075317,
0.04736239090561867,
-0.11267079412937164,
0.0006391099886968732,
0.02546006813645363,
0.04374583438038826,
0.010268100537359715,
-0.004367734305560589,
-0.020661653950810432,
-0.04643119126558304,
0.04508538916707039,
0.1282886266708374,
0.016534199938178062,
0.10036345571279526,
-0.09340310096740723,
-0.04899776726961136,
0.08471010625362396,
-0.08139441907405853,
-0.046601805835962296,
0.013471025973558426,
-0.0611678883433342,
-0.036205220967531204,
-0.01824786141514778,
-0.00890393741428852,
-0.06384146213531494,
0.05105682834982872,
-0.05740104243159294,
-0.01332311425358057,
-0.0676988810300827,
-0.10983868688344955,
0.04007132723927498,
0.019907690584659576,
0.025919269770383835,
-0.11244850605726242,
-0.17081740498542786,
-0.054901596158742905,
-0.026191366836428642,
-0.044547826051712036,
-0.011294963769614697,
-0.017354106530547142,
-0.01792834885418415,
-0.040962204337120056,
-0.04755629971623421,
-0.09578069299459457,
-0.032820090651512146,
0.03309882804751396,
-0.1208968311548233,
0.04634711518883705,
0.15307992696762085,
-0.0011860011145472527,
-0.1342674195766449,
0.07960730791091919,
-0.1697944849729538,
0.0949833020567894,
-0.06871214509010315,
-0.045052625238895416,
-0.05231078714132309,
-0.08564488589763641,
0.008394266478717327,
0.08622351288795471,
-0.03929578512907028,
0.20179976522922516,
-0.06995300203561783,
-0.02878871001303196,
0.1791738122701645,
-0.15157586336135864,
-0.04232232645153999,
0.10207745432853699,
-0.09599465876817703,
0.040969014167785645,
0.08987093716859818,
0.13987618684768677,
-0.019422832876443863,
-0.0656173825263977,
-0.05522351339459419,
0.02242181822657585,
-0.09477151185274124,
0.12560027837753296,
0.04613637179136276,
-0.009874406270682812,
0.0054847197607159615,
-0.03483045473694801,
-0.09557047486305237,
0.032859694212675095,
-0.0413178950548172,
-0.027610905468463898,
0.028537021949887276,
-0.03404044359922409,
0.12284385412931442,
-0.011078251525759697,
-0.07035205513238907,
0.012739523313939571,
-0.06992261111736298,
0.028180377557873726,
0.09306164085865021,
-0.04623372107744217,
-0.03731308504939079,
-0.1047620102763176,
0.005062032490968704,
0.05842750892043114,
-0.02988852746784687,
-0.18180479109287262,
-0.036043595522642136,
0.013339666649699211,
-0.13572371006011963,
0.036099158227443695,
0.12887993454933167,
0.020851435139775276,
0.04005824401974678,
-0.04297903925180435,
0.006920425221323967,
0.022584563121199608,
0.0008110018679872155,
-0.021280750632286072,
-0.11178629845380783,
0.0014546695165336132,
-0.03884061053395271,
0.01701541617512703,
-0.1844519078731537,
0.021891694515943527,
0.08266251534223557,
0.03680270165205002,
0.06562669575214386,
-0.032103221863508224,
0.10963878780603409,
0.013148900121450424,
0.009819193743169308,
-0.044475242495536804,
-0.0013769016368314624,
-0.042155344039201736,
-0.06044376641511917,
0.14290718734264374,
-0.22266966104507446,
0.09487176686525345,
0.06482364982366562,
0.022523334249854088,
-0.13051600754261017,
-0.04194992780685425,
-0.027806976810097694,
0.020701130852103233,
-0.0761559009552002,
0.0070594423450529575,
0.15028515458106995,
0.0369143970310688,
0.04819329082965851,
-0.1024927869439125,
-0.05012357980012894,
-0.030434640124440193,
-0.02335493266582489,
-0.059816762804985046,
0.07304743677377701,
-0.003746113507077098,
-0.1943013221025467,
0.09551715105772018,
-0.021498771384358406,
0.038182731717824936,
0.04212391376495361,
0.022156601771712303,
-0.040322043001651764,
-0.002150125103071332,
-0.010578387416899204,
-0.048821061849594116,
0.12352576851844788,
-0.029404060915112495,
0.008878251537680626,
0.06089930981397629,
0.0024983363691717386,
-0.0326022133231163,
-0.07307736575603485,
-0.048061322420835495,
0.011526225134730339,
0.01485978253185749,
-0.08837226778268814,
0.01874144934117794,
0.0697624534368515,
0.12350956350564957,
0.014856293797492981,
-0.003994556609541178,
-0.015706565231084824,
-0.05237152799963951,
-0.16021126508712769,
0.20748931169509888,
-0.06406284123659134,
-0.3279891014099121,
-0.028744110837578773,
-0.038667190819978714,
0.02214677259325981,
0.015222283080220222,
0.0031232284381985664,
-0.08354653418064117,
-0.07659585773944855,
-0.08275924623012543,
0.16134290397167206,
-0.017690515145659447,
-0.10101673007011414,
-0.01384766399860382,
0.020352795720100403,
0.029603784903883934,
-0.07698841392993927,
0.001828167587518692,
-0.04292597621679306,
-0.03175375983119011,
0.04403204098343849,
-0.003500700695440173,
0.07620581984519958,
0.07139359414577484,
0.04225045442581177,
-0.06632153689861298,
-0.06582453101873398,
0.22490054368972778,
-0.13143016397953033,
0.016016090288758278,
0.21684247255325317,
-0.16235050559043884,
0.06261392682790756,
0.15528003871440887,
0.01159798912703991,
-0.06986802816390991,
0.041255708783864975,
0.06021285057067871,
-0.020367799326777458,
-0.2050541788339615,
-0.07283328473567963,
0.015248605981469154,
0.011035030707716942,
0.06865325570106506,
0.03914501890540123,
0.08381291478872299,
0.09807468205690384,
-0.11799682676792145,
-0.05819796025753021,
0.07569019496440887,
0.07541076838970184,
0.007257792167365551,
-0.017736509442329407,
0.05956584960222244,
0.005228332709521055,
-0.052571091800928116,
0.06356129795312881,
-0.03984764218330383,
0.25275614857673645,
0.002040241379290819,
0.08012500405311584,
0.09795541316270828,
0.07678306847810745,
0.03463498130440712,
0.006021436303853989,
-0.02243856154382229,
0.009194452315568924,
-0.053673285990953445,
-0.03389730677008629,
-0.03942228853702545,
0.047147221863269806,
-0.006695538759231567,
-0.006719406694173813,
-0.03606016933917999,
-0.08304917067289352,
0.04109502211213112,
0.12485120445489883,
0.007364543620496988,
-0.10594639182090759,
-0.02711695246398449,
0.062202151864767075,
-0.08176732808351517,
-0.09210163354873657,
0.031082680448889732,
0.08286498486995697,
-0.204339861869812,
0.08663385361433029,
-0.022042857483029366,
0.07757389545440674,
-0.18236622214317322,
0.027710936963558197,
-0.08662828058004379,
-0.004278999287635088,
-0.04099775478243828,
0.07261498272418976,
-0.20620287954807281,
0.19245539605617523,
0.02466089092195034,
0.016595307737588882,
-0.13883644342422485,
-0.020125174894928932,
0.058135539293289185,
0.06002204865217209,
0.19354619085788727,
0.021915549412369728,
-0.04214993119239807,
-0.1342388093471527,
0.028120553120970726,
0.02735387347638607,
0.029297996312379837,
-0.0559874027967453,
0.04082593321800232,
0.012565044686198235,
0.024141915142536163,
-0.024650083854794502,
-0.11009878665208817,
-0.159780815243721,
-0.09209754317998886,
0.02755926363170147,
-0.15345479547977448,
0.02185737155377865,
-0.08567336201667786,
-0.047661203891038895,
-0.039222124963998795,
0.11913446336984634,
-0.16858041286468506,
-0.10841230303049088,
-0.08815757185220718,
-0.006651938892900944,
0.03648532181978226,
-0.08633939176797867,
-0.006356819998472929,
0.04338790848851204,
0.1644601821899414,
-0.049416568130254745,
-0.11719858646392822,
0.05565064772963524,
-0.058958522975444794,
-0.17780931293964386,
-0.04754069820046425,
0.054883923381567,
0.1706492155790329,
0.08080950379371643,
0.00011069651372963563,
0.02208184078335762,
0.06363407522439957,
-0.11519617587327957,
-0.0018126697978004813,
0.1404389888048172,
0.0805027112364769,
0.10588125139474869,
-0.009438828565180302,
-0.08162165433168411,
-0.12769699096679688,
-0.04251353442668915,
0.09155819565057755,
0.20261529088020325,
-0.03521789237856865,
0.09639589488506317,
0.1534910500049591,
-0.0781986340880394,
-0.2792355716228485,
0.0722484439611435,
-0.0026247885543853045,
-0.01106482744216919,
0.13336516916751862,
-0.20327851176261902,
0.027017302811145782,
0.05787401273846626,
0.007223562803119421,
0.008638097904622555,
-0.1500433385372162,
-0.13364417850971222,
0.1923634111881256,
0.10054578632116318,
0.09986381977796555,
-0.04193535074591637,
-0.017034560441970825,
-0.026526058092713356,
-0.06111402437090874,
0.1628236174583435,
-0.024817878380417824,
0.009718326851725578,
-0.000700133852660656,
0.06310783326625824,
0.05372535437345505,
-0.01237453892827034,
0.13228639960289001,
0.015622833743691444,
0.12525378167629242,
-0.1447669267654419,
-0.1216466873884201,
-0.050855688750743866,
-0.04976919665932655,
0.09140966087579727,
0.01971725933253765,
-0.016897806897759438,
-0.10890321433544159,
-0.07209473103284836,
-0.10189534723758698,
0.06295228004455566,
-0.027092045173048973,
-0.061479352414608,
-0.15106135606765747,
0.11811186373233795,
0.05591442063450813,
-0.030157262459397316,
0.014681980013847351,
-0.09998712688684464,
0.01936107873916626,
-0.012109104543924332,
0.2768062353134155,
0.10875842720270157,
-0.060879070311784744,
-0.0063575380481779575,
-0.023481110110878944,
0.07389279454946518,
-0.10158543288707733,
-0.02001435123383999,
0.055045634508132935,
-0.0074959928169846535,
0.1435449719429016,
-0.004005125258117914,
-0.13292859494686127,
0.025237541645765305,
-0.03680963069200516,
-0.08022459596395493,
-0.19397000968456268,
-0.0256647989153862,
-0.015827618539333344,
-0.1599842756986618,
-0.0718090757727623,
0.08187521994113922,
-0.022125927731394768,
-0.05602152273058891,
-0.016728123649954796,
0.079788938164711,
0.034572385251522064,
0.018585409969091415,
0.0805630162358284,
0.07818079739809036,
-0.08884312212467194,
0.05913001298904419,
0.01660493202507496,
-0.00818091630935669,
0.08020874857902527,
0.0074270544573664665,
-0.08702767640352249,
-0.04206375032663345,
-0.035816457122564316,
0.15385770797729492,
0.012386007234454155,
-0.06039300933480263,
-0.010045896284282207,
-0.17539525032043457,
-0.004788550082594156,
0.21650786697864532,
0.036011770367622375,
0.04936730116605759,
0.01466458011418581,
-0.019824780523777008,
-0.03736212104558945,
0.06929079443216324,
-0.008196588605642319,
-0.004835489206016064,
-0.08041073381900787,
0.0288934875279665,
0.04752953723073006,
0.07843244820833206,
-0.03894301876425743,
-0.032169848680496216,
-0.14551421999931335,
-0.00850397814065218,
-0.15134194493293762,
0.024060187861323357,
-0.14896483719348907,
0.04380597919225693,
-0.03912080079317093,
-0.06524107605218887,
0.004381493665277958,
0.02986232563853264,
-0.066755011677742,
0.0529913604259491,
0.0339655838906765,
0.0618036687374115,
-0.12670248746871948,
-0.023308319970965385,
0.09043101966381073,
-0.0722632184624672,
0.1275825947523117,
-0.03731827065348625,
-0.06646410375833511,
0.034242115914821625,
-0.15305833518505096,
0.051679577678442,
0.012676836922764778,
0.030706200748682022,
-0.008238631300628185,
-0.10250134766101837,
-0.012213342823088169,
-0.029256653040647507,
-0.009064934216439724,
0.0408322773873806,
0.043750688433647156,
-0.03239649534225464,
0.06152402609586716,
0.002536751562729478,
-0.05309245362877846,
-0.06615709513425827,
0.06294895708560944,
0.13592973351478577,
-0.0472521036863327,
0.11032205820083618,
-0.008657903410494328,
0.06508103013038635,
-0.15790842473506927,
-0.006308089476078749,
0.0532715804874897,
0.004845289047807455,
0.008150878362357616,
-0.03248833864927292,
0.07224984467029572,
-0.00034413725370541215,
0.2718154489994049,
0.0025695525109767914,
-0.05977234989404678,
0.06652643531560898,
0.07370669394731522,
0.03965595364570618,
0.030896421521902084,
-0.02865704335272312,
-0.06575889885425568,
-0.05962222442030907,
0.00951602403074503,
-0.017433958128094673,
-0.035755183547735214,
-0.1669204980134964,
0.17852550745010376,
0.1647891253232956,
0.08922814577817917,
-0.03467068076133728,
0.020875629037618637,
-0.05411016568541527,
-0.018058473244309425,
-0.1163937970995903,
-0.00011282940249657258,
-0.026863817125558853,
-0.03890535235404968,
0.1702270805835724,
0.09678637981414795,
-0.05418025702238083,
0.10823719948530197,
-0.012184296734631062,
-0.06935928016901016,
-0.14534495770931244,
-0.21363645792007446,
-0.043321643024683,
0.024267563596367836,
0.016480829566717148,
-0.10641832649707794,
-0.03877907618880272,
0.17422538995742798,
-0.00011019450903404504,
-0.048393819481134415,
0.07702672481536865,
-0.022662578150629997,
-0.09496734291315079,
0.06469704210758209,
0.017699331045150757,
0.009294049814343452,
-0.016164343804121017,
0.022303130477666855,
0.06086026132106781,
0.06412656605243683,
0.0635802149772644,
0.022446637973189354,
-0.02379608154296875,
-0.025732407346367836,
-0.1096753403544426,
-0.09921921789646149,
0.030503978952765465,
0.05415544658899307,
0.050865545868873596,
0.20552363991737366,
0.03627651184797287,
0.015248462557792664,
-0.0032648113556206226,
0.1530975103378296,
0.05121595412492752,
0.07657407224178314,
-0.12013543397188187,
0.1988878846168518,
-0.025317667052149773,
0.00002746999598457478,
0.07493457943201065,
-0.07717052102088928,
0.0748097375035286,
0.17358054220676422,
0.18101124465465546,
-0.05818888545036316,
0.008848763071000576,
-0.06856656819581985,
0.01796012558043003,
0.022457776591181755,
0.09818395972251892,
0.02076549082994461,
0.17512241005897522,
-0.08038998395204544,
0.11141066998243332,
0.022140640765428543,
0.0031550724525004625,
0.023589404299855232,
0.09709345549345016,
0.06732740998268127,
0.030310938134789467,
-0.13751891255378723,
0.12096615880727768,
-0.12915778160095215,
-0.1474207490682602,
0.034976303577423096,
-0.039341166615486145,
-0.09707532823085785,
0.027687596157193184,
-0.0575333833694458,
0.03530464321374893,
0.05601752549409866,
0.05041205510497093,
-0.016856282949447632,
0.11054712533950806,
0.06122530251741409,
-0.08926555514335632,
-0.06411562114953995,
0.09289242327213287,
0.045809488743543625,
0.2389383316040039,
0.031475234776735306,
0.08115296810865402,
0.08790996670722961,
-0.027686579152941704,
-0.10079725086688995,
-0.0135525893419981,
0.018397130072116852,
-0.001986059360206127,
-0.00042464808211661875,
0.15851326286792755,
0.018242230638861656,
0.08296595513820648,
0.09441674500703812,
-0.12522710859775543,
0.06281565874814987,
-0.14569759368896484,
-0.1206737756729126,
-0.09678295254707336,
0.11920605599880219,
-0.06869328022003174,
0.08927421271800995,
0.13170890510082245,
-0.0015741960378363729,
-0.009222856722772121,
-0.02486916445195675,
-0.008373879827558994,
-0.051367949694395065,
0.02884199470281601,
0.0396527536213398,
-0.12023906409740448,
0.003709043376147747,
0.0503058135509491,
0.04765480011701584,
-0.23750163614749908,
-0.020936138927936554,
-0.035068921744823456,
-0.02207663282752037,
-0.012365338392555714,
0.024834197014570236,
-0.06780152022838593,
0.021105138584971428,
-0.008633570745587349,
-0.1509322226047516,
0.03386201336979866,
0.10218384116888046,
-0.06889442354440689,
0.006329729221761227
] |
null | null |
transformers
|
# rubert_ria_headlines
## Description
*bert2bert* model, initialized with the `DeepPavlov/rubert-base-cased` pretrained weights and
fine-tuned on the first 99% of ["Rossiya Segodnya" news dataset](https://github.com/RossiyaSegodnya/ria_news_dataset) for 2 epochs.
## Usage example
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
MODEL_NAME = "dmitry-vorobiev/rubert_ria_headlines"
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
model = AutoModelForSeq2SeqLM.from_pretrained(MODEL_NAME)
text = "Скопируйте текст статьи / новости"
encoded_batch = tokenizer.prepare_seq2seq_batch(
[text],
return_tensors="pt",
padding="max_length",
truncation=True,
max_length=512)
output_ids = model.generate(
input_ids=encoded_batch["input_ids"],
max_length=36,
no_repeat_ngram_size=3,
num_beams=5,
top_k=0
)
headline = tokenizer.decode(output_ids[0],
skip_special_tokens=True,
clean_up_tokenization_spaces=False)
print(headline)
```
## Datasets
- [ria_news](https://github.com/RossiyaSegodnya/ria_news_dataset)
## How it was trained?
I used free TPUv3 on kaggle. The model was trained for 3 epochs with effective batch size 192 and soft restarts (warmup steps 1500 / 500 / 500 with new optimizer state on each epoch start).
- [1 epoch notebook](https://www.kaggle.com/dvorobiev/try-train-seq2seq-ria-tpu?scriptVersionId=53254694)
- [2 epoch notebook](https://www.kaggle.com/dvorobiev/try-train-seq2seq-ria-tpu?scriptVersionId=53269040)
- [3 epoch notebook](https://www.kaggle.com/dvorobiev/try-train-seq2seq-ria-tpu?scriptVersionId=53280797)
Common train params:
```shell
export XLA_USE_BF16=1
export XLA_TENSOR_ALLOCATOR_MAXSIZE=100000000
python nlp_headline_rus/src/train_seq2seq.py \
--do_train \
--tie_encoder_decoder \
--max_source_length 512 \
--max_target_length 32 \
--val_max_target_length 48 \
--tpu_num_cores 8 \
--per_device_train_batch_size 24 \
--gradient_accumulation_steps 1 \
--learning_rate 5e-4 \
--adam_epsilon 1e-6 \
--weight_decay 1e-5 \
```
## Validation results
- Using [last 1% of ria](https://drive.google.com/drive/folders/1ztAeyb1BiLMgXwOgOJS7WMR4PGiI1q92) dataset
- Using [gazeta_ru test](https://drive.google.com/drive/folders/1CyowuRpecsLTcDbqEfmAvkCWOod58g_e) split
- Using [gazeta_ru val](https://drive.google.com/drive/folders/1XZFOXHSXLKdhzm61ceVLw3aautrdskIu) split
|
{"language": ["ru"], "license": "mit", "tags": ["summarization", "bert", "rubert"]}
|
summarization
|
dmitry-vorobiev/rubert_ria_headlines
|
[
"transformers",
"pytorch",
"safetensors",
"encoder-decoder",
"text2text-generation",
"summarization",
"bert",
"rubert",
"ru",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ru"
] |
TAGS
#transformers #pytorch #safetensors #encoder-decoder #text2text-generation #summarization #bert #rubert #ru #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# rubert_ria_headlines
## Description
*bert2bert* model, initialized with the 'DeepPavlov/rubert-base-cased' pretrained weights and
fine-tuned on the first 99% of "Rossiya Segodnya" news dataset for 2 epochs.
## Usage example
## Datasets
- ria_news
## How it was trained?
I used free TPUv3 on kaggle. The model was trained for 3 epochs with effective batch size 192 and soft restarts (warmup steps 1500 / 500 / 500 with new optimizer state on each epoch start).
- 1 epoch notebook
- 2 epoch notebook
- 3 epoch notebook
Common train params:
## Validation results
- Using last 1% of ria dataset
- Using gazeta_ru test split
- Using gazeta_ru val split
|
[
"# rubert_ria_headlines",
"## Description\n*bert2bert* model, initialized with the 'DeepPavlov/rubert-base-cased' pretrained weights and \n fine-tuned on the first 99% of \"Rossiya Segodnya\" news dataset for 2 epochs.",
"## Usage example",
"## Datasets\n- ria_news",
"## How it was trained?\n\nI used free TPUv3 on kaggle. The model was trained for 3 epochs with effective batch size 192 and soft restarts (warmup steps 1500 / 500 / 500 with new optimizer state on each epoch start).\n\n- 1 epoch notebook\n- 2 epoch notebook\n- 3 epoch notebook\n\nCommon train params:",
"## Validation results\n\n- Using last 1% of ria dataset\n- Using gazeta_ru test split\n- Using gazeta_ru val split"
] |
[
"TAGS\n#transformers #pytorch #safetensors #encoder-decoder #text2text-generation #summarization #bert #rubert #ru #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# rubert_ria_headlines",
"## Description\n*bert2bert* model, initialized with the 'DeepPavlov/rubert-base-cased' pretrained weights and \n fine-tuned on the first 99% of \"Rossiya Segodnya\" news dataset for 2 epochs.",
"## Usage example",
"## Datasets\n- ria_news",
"## How it was trained?\n\nI used free TPUv3 on kaggle. The model was trained for 3 epochs with effective batch size 192 and soft restarts (warmup steps 1500 / 500 / 500 with new optimizer state on each epoch start).\n\n- 1 epoch notebook\n- 2 epoch notebook\n- 3 epoch notebook\n\nCommon train params:",
"## Validation results\n\n- Using last 1% of ria dataset\n- Using gazeta_ru test split\n- Using gazeta_ru val split"
] |
[
63,
8,
60,
4,
9,
83,
31
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #encoder-decoder #text2text-generation #summarization #bert #rubert #ru #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# rubert_ria_headlines## Description\n*bert2bert* model, initialized with the 'DeepPavlov/rubert-base-cased' pretrained weights and \n fine-tuned on the first 99% of \"Rossiya Segodnya\" news dataset for 2 epochs.## Usage example## Datasets\n- ria_news## How it was trained?\n\nI used free TPUv3 on kaggle. The model was trained for 3 epochs with effective batch size 192 and soft restarts (warmup steps 1500 / 500 / 500 with new optimizer state on each epoch start).\n\n- 1 epoch notebook\n- 2 epoch notebook\n- 3 epoch notebook\n\nCommon train params:## Validation results\n\n- Using last 1% of ria dataset\n- Using gazeta_ru test split\n- Using gazeta_ru val split"
] |
[
-0.03802961856126785,
0.07632345706224442,
-0.003421566914767027,
0.016355091705918312,
0.05970260128378868,
0.008327675983309746,
0.14198285341262817,
0.09761839359998703,
-0.13344353437423706,
0.060221195220947266,
0.08002500236034393,
0.024442030116915703,
0.04207652062177658,
0.17833732068538666,
-0.05379076302051544,
-0.19888712465763092,
0.07607583701610565,
-0.04323287680745125,
-0.008909101597964764,
0.12511058151721954,
0.13179472088813782,
-0.09333454817533493,
0.02696225605905056,
-0.003922199830412865,
-0.05565125122666359,
0.04052088409662247,
-0.06419984251260757,
-0.07759971916675568,
0.07635699212551117,
0.008678522892296314,
0.08288298547267914,
0.052309732884168625,
0.06489977240562439,
-0.1825144737958908,
0.011980117298662663,
0.0279961209744215,
-0.01564650796353817,
0.03986264765262604,
0.11152003705501556,
-0.017787422984838486,
0.09128350019454956,
-0.07950028777122498,
0.09456324577331543,
0.010134109295904636,
-0.08881504833698273,
-0.17696648836135864,
-0.14506861567497253,
0.12966537475585938,
0.15401394665241241,
0.07423465698957443,
-0.052091799676418304,
0.11669401079416275,
-0.12678857147693634,
0.04789860174059868,
0.157821923494339,
-0.3053642213344574,
-0.07644449174404144,
-0.0163642019033432,
-0.0037401553709059954,
0.026372911408543587,
-0.03918219357728958,
0.015133886598050594,
0.043923068791627884,
0.027827667072415352,
0.0650191679596901,
-0.0049261050298810005,
-0.061675868928432465,
0.023070957511663437,
-0.16287651658058167,
-0.023998182266950607,
0.21466559171676636,
-0.004993297625333071,
-0.011016770265996456,
-0.15016530454158783,
-0.0292349923402071,
-0.03427567705512047,
-0.026206785812973976,
-0.0679127424955368,
-0.017572924494743347,
-0.014423642307519913,
-0.002096105832606554,
-0.002475469373166561,
-0.11829482018947601,
-0.03120724856853485,
-0.0293494313955307,
0.08659066259860992,
0.06730005890130997,
0.011375123634934425,
0.03311442211270332,
0.09945176541805267,
0.017465177923440933,
-0.10230712592601776,
0.034354668110609055,
-0.06443560868501663,
-0.07647115737199783,
-0.016551963984966278,
-0.027966991066932678,
-0.1284227967262268,
0.0541130006313324,
0.20536787807941437,
0.053398121148347855,
0.01792946085333824,
0.06871898472309113,
-0.001732771866954863,
0.023477450013160706,
0.09442809969186783,
-0.08207978308200836,
-0.07347892969846725,
0.02948871999979019,
0.05939250811934471,
0.018542630597949028,
-0.005027982406318188,
-0.04419706389307976,
0.02621433511376381,
0.14139007031917572,
0.08003110438585281,
0.007612377405166626,
-0.004342327360063791,
-0.04447484016418457,
-0.019876210018992424,
0.051142673939466476,
-0.13958004117012024,
0.00974311027675867,
0.0023353411816060543,
-0.011430519632995129,
0.0071748304180800915,
-0.002629493363201618,
0.020301079377532005,
-0.04279875010251999,
0.11511437594890594,
-0.02193818986415863,
-0.057176489382982254,
-0.08048120886087418,
-0.05172174423933029,
0.010582884773612022,
-0.07870903611183167,
-0.06099417433142662,
-0.11141243577003479,
-0.18453940749168396,
-0.0486317053437233,
0.0612172968685627,
-0.07085920870304108,
-0.0253954716026783,
-0.023361315950751305,
-0.04399573430418968,
0.030087681487202644,
-0.005831070244312286,
0.09038751572370529,
-0.03870956227183342,
0.0404672846198082,
-0.054537709802389145,
0.005545500200241804,
0.07947973906993866,
0.044372688978910446,
-0.0787169560790062,
0.03685158118605614,
-0.06843826174736023,
0.11458677798509598,
-0.08532040566205978,
0.01898202672600746,
-0.13871610164642334,
-0.04449072107672691,
-0.0131764467805624,
-0.022789480164647102,
0.02900225855410099,
0.12524151802062988,
-0.20430630445480347,
-0.025217022746801376,
0.08893957734107971,
-0.1462782770395279,
-0.05368955060839653,
0.130130797624588,
-0.04302307218313217,
0.04377356916666031,
0.08150425553321838,
0.0985947847366333,
0.03491051122546196,
-0.07234931737184525,
-0.04141026735305786,
-0.07647617906332016,
0.0427803136408329,
0.0990985706448555,
0.08744867146015167,
-0.014162775129079819,
-0.016281988471746445,
0.003513783449307084,
-0.04214547947049141,
-0.022208496928215027,
-0.05760341137647629,
-0.0743313729763031,
0.010349765419960022,
-0.04772024229168892,
0.022746184840798378,
-0.018318606540560722,
0.06026628613471985,
-0.08758697658777237,
-0.13667874038219452,
0.02356671541929245,
0.07503252476453781,
-0.04050137475132942,
0.0017030423041433096,
-0.09864313900470734,
0.11002060025930405,
-0.04528742656111717,
-0.005111189559102058,
-0.13662390410900116,
-0.05544823408126831,
0.057941634207963943,
-0.010299009270966053,
-0.011537112295627594,
0.03735191375017166,
0.04997294396162033,
0.02193119376897812,
-0.08812986314296722,
-0.05074786767363548,
-0.04605533182621002,
0.02122166007757187,
-0.04625027999281883,
-0.14808450639247894,
-0.01011666376143694,
-0.0478687584400177,
0.08053837716579437,
-0.08615939319133759,
0.019857244566082954,
0.0630197823047638,
0.1585274189710617,
0.027784205973148346,
-0.0242729764431715,
0.0018666604300960898,
0.017447302117943764,
0.0022470231633633375,
-0.03605679050087929,
0.01432828139513731,
0.012395069934427738,
-0.05618175491690636,
0.07671710103750229,
-0.14651218056678772,
0.0034956170711666346,
0.05175720900297165,
0.05674801021814346,
-0.12239637970924377,
0.04173135757446289,
-0.06535690277814865,
-0.031629011034965515,
-0.0875713974237442,
-0.05079014226794243,
0.09042558073997498,
-0.005127363838255405,
0.08080116659402847,
-0.09149913489818573,
-0.09923863410949707,
-0.03593660145998001,
-0.01698877103626728,
-0.04120239242911339,
0.13692183792591095,
-0.022182226181030273,
-0.178070068359375,
0.09768502414226532,
0.09390474110841751,
-0.017468884587287903,
0.08852338045835495,
-0.053741104900836945,
-0.09308770298957825,
-0.040236420929431915,
0.013073288835585117,
0.018418803811073303,
0.056171026080846786,
-0.04549697786569595,
0.01960674673318863,
0.07643644511699677,
0.016650784760713577,
0.03250125050544739,
-0.1271100789308548,
0.013204466551542282,
-0.004739711061120033,
-0.03677063435316086,
0.07702469080686569,
0.05503451079130173,
0.010343853384256363,
0.08481337130069733,
0.010732924565672874,
0.0017866150010377169,
0.013113969936966896,
-0.02826562337577343,
-0.10089149326086044,
0.15321888029575348,
-0.0869826227426529,
-0.15934625267982483,
-0.08892367780208588,
0.006372738629579544,
0.0140529228374362,
-0.009907687082886696,
0.08125142008066177,
-0.056711938232183456,
-0.09555382281541824,
-0.08243976533412933,
0.035665761679410934,
0.04701437056064606,
-0.0019360084552317858,
0.03558897599577904,
0.016028428450226784,
0.008086821995675564,
-0.08069004863500595,
0.011019889265298843,
-0.012765921652317047,
-0.044047918170690536,
0.06469174474477768,
-0.02016034908592701,
0.0761454701423645,
0.06935393810272217,
0.019552238285541534,
0.018999701365828514,
-0.019872969016432762,
0.18895728886127472,
-0.07167322933673859,
0.035461731255054474,
0.13280966877937317,
-0.04197167977690697,
0.03855811059474945,
0.16296054422855377,
0.011759431101381779,
-0.02697104401886463,
0.02351549081504345,
0.054859671741724014,
-0.021387659013271332,
-0.27064722776412964,
-0.066196970641613,
-0.04111810028553009,
-0.015856103971600533,
0.031766824424266815,
0.06991248577833176,
-0.039401210844516754,
0.048568159341812134,
-0.05900075286626816,
-0.04699867591261864,
0.08873389661312103,
0.056727804243564606,
0.10167650878429413,
0.03376765549182892,
0.1092921644449234,
-0.04743008315563202,
-0.07318618893623352,
0.10844234377145767,
-0.0824594721198082,
0.24699227511882782,
-0.012398469261825085,
0.05022616684436798,
0.08865485340356827,
0.08586687594652176,
0.028773246333003044,
0.07302593439817429,
0.024354159832000732,
0.02735450118780136,
-0.015219073742628098,
-0.07335204631090164,
0.02996254712343216,
0.07077515870332718,
-0.009600874967873096,
-0.029166022315621376,
-0.0008900462416931987,
-0.006481013726443052,
0.09641243517398834,
0.33562108874320984,
0.08974459767341614,
-0.19396944344043732,
-0.022184276953339577,
0.03811110183596611,
-0.1010751873254776,
-0.03586370870471001,
-0.0005177318817004561,
0.02715025469660759,
-0.15549138188362122,
0.10553507506847382,
-0.03746575117111206,
0.07815009355545044,
-0.04029104858636856,
-0.037033844739198685,
0.014228240586817265,
0.011914466507732868,
-0.012756184674799442,
0.11675358563661575,
-0.19138047099113464,
0.18551313877105713,
-0.012885009869933128,
0.07266907393932343,
-0.04082762077450752,
0.03778797388076782,
0.016458529978990555,
-0.029941534623503685,
0.11438778787851334,
-0.02150220423936844,
-0.019997334107756615,
0.0005831306334584951,
-0.18083399534225464,
-0.005162904039025307,
0.06522443145513535,
-0.10187885165214539,
0.0931466743350029,
-0.04554395005106926,
-0.010218726471066475,
-0.002637244528159499,
0.014449750073254108,
-0.1563524305820465,
-0.16734708845615387,
0.07785112410783768,
-0.04448292776942253,
0.002698182361200452,
-0.11206664890050888,
-0.12305447459220886,
-0.07938230037689209,
0.12874457240104675,
0.05975515767931938,
-0.08600247651338577,
-0.10511086136102676,
-0.006499874405562878,
0.1634460836648941,
-0.043778806924819946,
0.048504721373319626,
0.0023672839161008596,
0.1031443402171135,
-0.0098649300634861,
-0.03147861734032631,
0.03662649914622307,
-0.08197879791259766,
-0.23190806806087494,
-0.017096446827054024,
0.13185498118400574,
0.08571182936429977,
0.03295399621129036,
0.02525480091571808,
0.08135370910167694,
-0.02375982142984867,
-0.05926797166466713,
0.015447570011019707,
0.050523437559604645,
0.0688389241695404,
0.06074371933937073,
-0.07421749085187912,
-0.14293700456619263,
-0.07972058653831482,
-0.08736234158277512,
0.13027864694595337,
0.253400981426239,
-0.059695545583963394,
0.13411450386047363,
0.08357443660497665,
-0.061268702149391174,
-0.22773349285125732,
-0.09243588149547577,
0.07265155017375946,
0.02084501087665558,
0.010142908431589603,
-0.09396343678236008,
0.05484543368220329,
0.11343903094530106,
-0.05409606546163559,
-0.01657000556588173,
-0.252180278301239,
-0.14348171651363373,
0.1099504828453064,
0.011738616041839123,
-0.015605906955897808,
-0.13625426590442657,
-0.054452888667583466,
-0.020963244140148163,
-0.024513034150004387,
0.07059721648693085,
-0.016666987910866737,
0.06500248610973358,
-0.002738518873229623,
-0.051626358181238174,
0.017853498458862305,
-0.041616205126047134,
0.13834165036678314,
-0.03578289598226547,
0.006102357525378466,
-0.07120601087808609,
0.007722699083387852,
0.07990997284650803,
-0.07360168546438217,
0.0644720122218132,
-0.011152820661664009,
0.0335550457239151,
-0.11956321448087692,
-0.04584562033414841,
-0.03488604351878166,
0.010869751684367657,
-0.03783579170703888,
-0.03814593702554703,
-0.07198823243379593,
0.06628454476594925,
0.08855289965867996,
-0.04824047535657883,
0.06929604709148407,
-0.004330527037382126,
0.07507108896970749,
0.05987801402807236,
0.07836074382066727,
0.041358113288879395,
-0.021553799510002136,
-0.05840321257710457,
-0.04754852503538132,
0.0035660145804286003,
-0.03633544594049454,
0.05474502965807915,
0.0913887768983841,
-0.013766185380518436,
0.14292721450328827,
0.02075015939772129,
-0.12842616438865662,
0.006447662133723497,
0.10458189249038696,
-0.15271013975143433,
-0.12126738578081131,
0.001192909898236394,
0.02665804699063301,
-0.08778827637434006,
-0.04271116852760315,
0.11053317040205002,
-0.031370680779218674,
-0.06558635085821152,
0.028669683262705803,
0.07693322002887726,
0.025505878031253815,
0.17631183564662933,
0.013507680036127567,
0.04897964000701904,
-0.07914793491363525,
0.1316194385290146,
0.07383032888174057,
-0.004463877063244581,
0.01885823719203472,
0.09855449199676514,
-0.09346744418144226,
-0.024708032608032227,
0.13276313245296478,
0.1777035892009735,
-0.005846321117132902,
-0.05590210109949112,
-0.08542093634605408,
-0.13670536875724792,
0.08159870654344559,
0.11444366723299026,
0.06362491101026535,
0.013479672372341156,
0.020910918712615967,
0.018123570829629898,
-0.009545466862618923,
0.1732717901468277,
0.18373681604862213,
0.027599290013313293,
-0.10299496352672577,
0.00023529815371148288,
-0.010710816830396652,
0.01809958927333355,
-0.05061081796884537,
0.03336860239505768,
-0.08305103331804276,
-0.006730164401233196,
-0.15830394625663757,
0.032356683164834976,
-0.05132546275854111,
-0.015871195122599602,
-0.021487057209014893,
-0.027175206691026688,
-0.041877925395965576,
0.005501373205333948,
-0.07990574836730957,
-0.01823301799595356,
-0.05487905070185661,
0.033182594925165176,
-0.10337017476558685,
-0.030968599021434784,
0.06603240221738815,
-0.10815145075321198,
0.10028006136417389,
0.04487987235188484,
0.0033512860536575317,
0.018910929560661316,
-0.0860089361667633,
-0.02627485617995262,
0.017232215031981468,
0.025249678641557693,
0.01734095625579357,
-0.18772214651107788,
-0.0034684629645198584,
0.020079677924513817,
-0.038008421659469604,
0.039727699011564255,
0.03244274482131004,
-0.09869950264692307,
0.027885228395462036,
-0.06872726231813431,
-0.08466589450836182,
-0.030369369313120842,
0.030354207381606102,
0.06065436452627182,
0.0992996096611023,
0.17181585729122162,
-0.10116405785083771,
0.08494921028614044,
-0.1015937477350235,
0.00935619231313467,
-0.009106351062655449,
-0.0537312887609005,
-0.10018236190080643,
-0.0062386151403188705,
0.045236632227897644,
-0.02424624003469944,
0.145421102643013,
-0.002744693774729967,
-0.04947957023978233,
0.04304860159754753,
-0.011012999340891838,
0.05033859238028526,
0.04516921564936638,
0.12097139656543732,
-0.004869294818490744,
-0.010765094310045242,
-0.045300573110580444,
-0.047193050384521484,
-0.004961951170116663,
-0.06824861466884613,
0.12182193994522095,
0.12807698547840118,
0.043781206011772156,
0.04480712115764618,
0.008087844587862492,
-0.0885327085852623,
-0.047855034470558167,
0.036500994116067886,
-0.03800630941987038,
0.07237669080495834,
-0.02757282927632332,
0.16239257156848907,
0.1488616168498993,
-0.16754232347011566,
0.045071784406900406,
-0.02057257853448391,
-0.04534395784139633,
-0.12156444042921066,
-0.09401217848062515,
-0.11571657657623291,
-0.08322170376777649,
0.021363424137234688,
-0.11365441977977753,
-0.008661426603794098,
0.08120302110910416,
0.049135010689496994,
0.011836506426334381,
0.08503376692533493,
-0.04888634756207466,
-0.030719025060534477,
0.09030456840991974,
-0.01329221110790968,
-0.0021330369636416435,
0.009507911279797554,
-0.026268668472766876,
0.009358422830700874,
0.054942868649959564,
0.03161795809864998,
0.013653983362019062,
-0.025868846103549004,
0.003185809822753072,
-0.0493532195687294,
-0.0746130719780922,
0.020973823964595795,
-0.0013448348036035895,
0.07400167733430862,
0.09877485036849976,
0.034454092383384705,
0.0015239956555888057,
0.007157569285482168,
0.21733693778514862,
-0.04200868308544159,
-0.12336995452642441,
-0.16498631238937378,
0.03912489861249924,
-0.02315641939640045,
0.03805631026625633,
0.08885238319635391,
-0.09939650446176529,
-0.0016676019877195358,
0.18350909650325775,
0.1683436930179596,
0.04186812415719032,
-0.003385172225534916,
0.021983977407217026,
0.0062113660387694836,
0.002151350723579526,
0.00743893813341856,
0.05598007142543793,
0.14589744806289673,
-0.09216399490833282,
0.03545414283871651,
-0.06154172495007515,
-0.04371071606874466,
-0.10276846587657928,
0.12327779829502106,
0.026568656787276268,
-0.03099549002945423,
-0.043189845979213715,
0.13829776644706726,
0.0012420536950230598,
-0.20320412516593933,
-0.018739815801382065,
-0.08760501444339752,
-0.17186276614665985,
0.005130833014845848,
0.11259549856185913,
-0.009745936840772629,
0.061725031584501266,
0.017224563285708427,
-0.001691484940238297,
0.08426735550165176,
0.03878956660628319,
-0.027335962280631065,
-0.08631397038698196,
0.12616969645023346,
-0.012863065116107464,
0.18421868979930878,
-0.0008463655249215662,
0.09879458695650101,
0.10933434963226318,
-0.0401461198925972,
-0.1108233630657196,
0.08665752410888672,
0.06870066374540329,
-0.04476477578282356,
0.053673502057790756,
0.11665383726358414,
-0.0272380281239748,
0.06124687194824219,
0.08250510692596436,
-0.03731388598680496,
0.031770046800374985,
-0.009859065525233746,
-0.055880311876535416,
-0.09154024720191956,
0.10964096337556839,
-0.0969565138220787,
0.10713738948106766,
0.16304484009742737,
-0.07476029545068741,
0.03886169567704201,
-0.057602085173130035,
0.03615246340632439,
0.026870589703321457,
0.05454462766647339,
0.05560925602912903,
-0.20278188586235046,
0.012021989561617374,
0.026480017229914665,
0.07550902664661407,
-0.16625560820102692,
-0.06431104987859726,
-0.08494921028614044,
-0.024252066388726234,
-0.09376616775989532,
0.12775927782058716,
0.06452064216136932,
-0.002994122914969921,
-0.023892227560281754,
-0.06955868005752563,
-0.031028062105178833,
0.05612671747803688,
-0.11005602777004242,
-0.004314393270760775
] |
null | null |
transformers
|
# doc2query/S2ORC-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/S2ORC-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 156k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, abstract) pairs from [S2ORC](https://github.com/allenai/s2orc).
|
{"language": "en", "license": "apache-2.0", "datasets": ["S2ORC"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/S2ORC-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:S2ORC",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-S2ORC #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/S2ORC-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 156k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, abstract) pairs from S2ORC.
|
[
"# doc2query/S2ORC-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 156k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, abstract) pairs from S2ORC."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-S2ORC #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/S2ORC-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 156k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, abstract) pairs from S2ORC."
] |
[
83,
261,
32,
89
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-S2ORC #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/S2ORC-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-base for 156k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, abstract) pairs from S2ORC."
] |
[
-0.06377064436674118,
0.15011030435562134,
-0.00715256854891777,
0.11989118903875351,
0.1195555105805397,
0.00042831417522393167,
0.08320668339729309,
0.11637955904006958,
-0.028032829985022545,
0.04649285227060318,
0.0567537359893322,
0.05099228769540787,
-0.03128795698285103,
0.11709233373403549,
0.047067273408174515,
-0.16873763501644135,
-0.031567540019750595,
-0.0375908724963665,
0.09376311302185059,
0.06663718819618225,
0.05082499980926514,
-0.055134955793619156,
0.0487770177423954,
-0.0389145202934742,
-0.09271582216024399,
0.05711909756064415,
-0.03384312614798546,
-0.018728304654359818,
0.06542146950960159,
0.026906797662377357,
0.07652904838323593,
-0.019630348309874535,
0.0672651156783104,
-0.12710706889629364,
0.01960982009768486,
0.08358477801084518,
-0.014878805726766586,
0.0762159526348114,
0.014761669561266899,
-0.035485416650772095,
0.13882695138454437,
-0.09776271879673004,
0.040489956736564636,
0.04080880433320999,
-0.08260875195264816,
-0.10525580495595932,
-0.07385242730379105,
0.03398837521672249,
0.04773547872900963,
0.0613396055996418,
0.001249639899469912,
0.03158031031489372,
-0.042015440762043,
0.05177421495318413,
0.16554835438728333,
-0.14493408799171448,
-0.029757607728242874,
0.055099233984947205,
-0.060666900128126144,
0.071833036839962,
-0.004170305095613003,
-0.054116301238536835,
-0.005875140894204378,
0.020572112873196602,
0.11095462739467621,
0.012502984143793583,
-0.016690028831362724,
-0.00006119354657130316,
-0.09987013041973114,
-0.050845418125391006,
0.12336888164281845,
-0.004790585488080978,
-0.07884173840284348,
-0.10804615914821625,
-0.049360595643520355,
0.051101505756378174,
0.039596810936927795,
-0.018275374546647072,
0.04370664432644844,
0.050906114280223846,
0.03393743932247162,
-0.13040043413639069,
-0.08689653873443604,
-0.058768514543771744,
-0.009459592401981354,
0.02181629091501236,
0.0008331559947691858,
0.006604646798223257,
-0.023252133280038834,
0.0949721559882164,
-0.05010606721043587,
-0.03632045164704323,
-0.044777192175388336,
-0.050508689135313034,
-0.07416126132011414,
-0.00925256498157978,
-0.04559125378727913,
-0.13511548936367035,
-0.034658048301935196,
0.07387029379606247,
0.01190666202455759,
0.005285050254315138,
-0.023410486057400703,
0.03375321999192238,
-0.025968335568904877,
0.11685331165790558,
-0.013597581535577774,
0.013787277974188328,
0.054823797196149826,
0.0062489924021065235,
0.009548884816467762,
-0.01171296276152134,
-0.08355511724948883,
-0.08348941057920456,
0.005441044457256794,
0.019117185845971107,
-0.020840460434556007,
0.05244678258895874,
-0.0610177256166935,
-0.0074076266027987,
0.030845075845718384,
-0.11200756579637527,
0.02411763370037079,
0.02145279385149479,
-0.06166147068142891,
0.00021054955141153187,
0.16678661108016968,
-0.05653123930096626,
-0.09847942739725113,
0.017624132335186005,
-0.04439685866236687,
0.0012705341214314103,
-0.10598096996545792,
-0.05343236029148102,
0.02020445466041565,
-0.127637580037117,
-0.06098285689949989,
-0.07811230421066284,
-0.2560407817363739,
-0.026230020448565483,
0.04549451544880867,
-0.01747778430581093,
0.01413881778717041,
-0.07453770935535431,
-0.0305757075548172,
-0.04760325327515602,
0.019197922199964523,
-0.06017277389764786,
0.009453882463276386,
0.021539177745580673,
-0.010864201001822948,
0.07683641463518143,
-0.01591646485030651,
0.0037317441310733557,
-0.05841292068362236,
0.010946430265903473,
-0.17942407727241516,
0.19375580549240112,
0.027023596689105034,
0.012098032049834728,
-0.07579795271158218,
0.024055873975157738,
-0.06170441955327988,
0.011236180551350117,
0.050085097551345825,
0.1378049999475479,
-0.2071511447429657,
-0.042292915284633636,
0.17598408460617065,
-0.07085159420967102,
-0.023093922063708305,
0.08091629296541214,
-0.08201788365840912,
0.0828838124871254,
0.12438920140266418,
0.06774422526359558,
0.13539230823516846,
-0.08723588287830353,
-0.05900505185127258,
-0.021158451214432716,
-0.08790871500968933,
0.0034546179231256247,
0.06980501115322113,
-0.06074502691626549,
0.031074369326233864,
0.03812727704644203,
-0.04806043580174446,
-0.014219023287296295,
0.0013575569028034806,
-0.043906938284635544,
-0.04734193906188011,
-0.04590494558215141,
0.008368749171495438,
-0.02518167346715927,
0.01530047319829464,
0.04604918509721756,
-0.07120312750339508,
0.14416508376598358,
0.0956459641456604,
-0.1092589944601059,
0.025803325697779655,
-0.056612927466630936,
0.0388791523873806,
-0.09291792660951614,
0.006319467909634113,
-0.17416644096374512,
-0.008824152871966362,
0.02945084311068058,
-0.02304021269083023,
0.12777332961559296,
-0.02762521430850029,
0.08116529881954193,
0.062442585825920105,
-0.020925723016262054,
0.009610755369067192,
-0.05135929584503174,
-0.024633120745420456,
-0.0659453347325325,
-0.11207195371389389,
-0.08175975829362869,
-0.033639658242464066,
0.13135753571987152,
-0.1793190985918045,
0.053542014211416245,
-0.012315917760133743,
0.0614413358271122,
0.04052135720849037,
-0.06251206248998642,
0.035805243998765945,
-0.025967277586460114,
0.006648705340921879,
-0.038082800805568695,
0.0575026236474514,
0.02823881059885025,
-0.08785244822502136,
-0.04628729447722435,
-0.08532066643238068,
-0.15682561695575714,
0.04926188662648201,
0.052857376635074615,
-0.048896338790655136,
-0.1382073611021042,
-0.020775942131876945,
-0.018740231171250343,
-0.1037912368774414,
-0.07298721373081207,
0.1834827959537506,
0.049627941101789474,
0.12344913184642792,
-0.10009811073541641,
-0.05788007006049156,
-0.01151500828564167,
-0.02201610803604126,
0.014082777313888073,
0.06035296618938446,
0.02344631217420101,
-0.013476608321070671,
0.01825382374227047,
-0.00481731491163373,
-0.0696016326546669,
0.1843167394399643,
0.05477091297507286,
-0.10918206721544266,
-0.036028873175382614,
0.0297919400036335,
0.024242477491497993,
0.0034808432683348656,
-0.05045796185731888,
0.016124878078699112,
-0.0032947163563221693,
0.0058473823592066765,
0.030874397605657578,
-0.09009288251399994,
0.02622399479150772,
0.04767349734902382,
-0.05010903626680374,
-0.051834672689437866,
-0.053028300404548645,
-0.008489450439810753,
0.07064706087112427,
0.010079416446387768,
0.08229534327983856,
0.02739272080361843,
-0.05123138427734375,
-0.13226598501205444,
0.1607416868209839,
-0.07731925696134567,
-0.10382422059774399,
-0.13510651886463165,
0.06966884434223175,
-0.05069013312458992,
0.024166671559214592,
0.004262085538357496,
-0.08995699137449265,
0.0010264200391247869,
-0.10462164133787155,
0.0033780743833631277,
0.05336335301399231,
-0.001046760124154389,
-0.014912465587258339,
0.060616835951805115,
0.044305071234703064,
-0.1299343854188919,
0.009662831202149391,
-0.03371192887425423,
-0.06567803770303726,
0.017987553030252457,
-0.048994045704603195,
0.07246867567300797,
0.0691785216331482,
-0.043955642729997635,
0.02618216536939144,
0.017369268462061882,
0.1672884225845337,
-0.006229933351278305,
0.04794880747795105,
0.13469886779785156,
-0.0313619039952755,
0.08431887626647949,
0.04233582317829132,
0.037000611424446106,
-0.028915010392665863,
0.05703712999820709,
0.03793779015541077,
-0.05377024784684181,
-0.21499063074588776,
-0.025158435106277466,
-0.04827776551246643,
-0.027660124003887177,
0.09714619815349579,
0.029394211247563362,
-0.04388013854622841,
0.03157861903309822,
-0.004998975899070501,
0.07062581181526184,
0.05631009861826897,
0.07900424301624298,
-0.00463878083974123,
0.022515103220939636,
0.04042021557688713,
-0.0568603053689003,
-0.0093946922570467,
0.056603774428367615,
0.061474066227674484,
0.20314064621925354,
-0.08049840480089188,
0.16933923959732056,
0.04820232838392258,
0.020306508988142014,
0.01314069889485836,
0.08869341015815735,
-0.07501654326915741,
0.04955127835273743,
-0.04228616878390312,
-0.09639663994312286,
-0.05561867356300354,
0.06365000456571579,
-0.04626493901014328,
0.057867806404829025,
-0.06983764469623566,
0.05330121889710426,
0.08046872913837433,
0.15782295167446136,
-0.04145381972193718,
-0.21953840553760529,
-0.024464523419737816,
-0.008156288415193558,
0.0028022637125104666,
0.007464636582881212,
-0.039518557488918304,
0.09979719668626785,
-0.07921918481588364,
-0.010235262103378773,
-0.0023805215023458004,
0.07889463752508163,
-0.06011545658111572,
-0.012738028541207314,
0.006903172470629215,
0.131031796336174,
0.014221598394215107,
0.06216224282979965,
-0.18753951787948608,
0.04610788822174072,
0.020424166694283485,
0.1240774616599083,
-0.06676662713289261,
0.04230267181992531,
0.008273126557469368,
-0.04826144501566887,
0.13151700794696808,
0.008534966967999935,
0.05445939674973488,
-0.03477988392114639,
-0.058730632066726685,
0.01357031986117363,
0.02366354502737522,
-0.046475738286972046,
0.13152188062667847,
0.012582269497215748,
-0.017690056934952736,
-0.002215416170656681,
0.06327811628580093,
-0.06765930354595184,
-0.20189855992794037,
-0.00735663203522563,
-0.03092474304139614,
-0.02111000381410122,
-0.013227205723524094,
-0.026073694229125977,
0.05461135506629944,
0.12806810438632965,
-0.1540783941745758,
-0.11515490710735321,
-0.0970039963722229,
0.04240060970187187,
0.07053232938051224,
-0.06690500676631927,
-0.016332566738128662,
0.0346839614212513,
0.14851653575897217,
-0.027354145422577858,
-0.08406046032905579,
-0.009287567809224129,
-0.03332079201936722,
-0.12581266462802887,
-0.05332673713564873,
0.06881336122751236,
0.08097612857818604,
0.0460747666656971,
0.037266477942466736,
0.014201804995536804,
-0.008276876993477345,
-0.08613301813602448,
0.010370752774178982,
0.18197143077850342,
0.10259196907281876,
0.07269304990768433,
-0.12778878211975098,
-0.012254618108272552,
-0.07493627071380615,
0.023358991369605064,
0.11748048663139343,
0.18359562754631042,
-0.08420950174331665,
0.07941728085279465,
0.1254464089870453,
-0.13734066486358643,
-0.13336123526096344,
0.016633182764053345,
0.08438562601804733,
0.017945898696780205,
-0.014489091001451015,
-0.2773842513561249,
0.11522773653268814,
0.10261543840169907,
0.027362458407878876,
-0.12443137168884277,
-0.28207316994667053,
-0.134329691529274,
0.00898159109055996,
0.059798527508974075,
0.03766937181353569,
-0.08637639880180359,
-0.015883296728134155,
-0.03842966631054878,
0.07935240864753723,
0.18824943900108337,
-0.20966275036334991,
0.05531566962599754,
0.0031011621467769146,
0.020623164251446724,
0.061677198857069016,
-0.023492133244872093,
0.046901870518922806,
0.03668858855962753,
0.056312195956707,
-0.012098650448024273,
0.04294148460030556,
0.0031425831839442253,
-0.07559546083211899,
0.16481919586658478,
-0.0072020553052425385,
0.06545084714889526,
-0.13955412805080414,
-0.054678138345479965,
-0.06152108311653137,
0.00932282768189907,
-0.022738896310329437,
-0.04182517156004906,
0.01147871371358633,
0.008420506492257118,
0.09831361472606659,
0.017236636951565742,
-0.014776589348912239,
-0.05324944108724594,
0.024078430607914925,
0.11909151077270508,
0.1477467119693756,
-0.025201836600899696,
-0.07661495357751846,
0.017739849165081978,
0.00023561425041407347,
0.046728357672691345,
-0.14339211583137512,
0.03613530844449997,
0.07038175314664841,
0.030625589191913605,
0.13492223620414734,
0.07835621386766434,
-0.10793189704418182,
0.000004726167389890179,
0.06798394024372101,
-0.11551453173160553,
-0.03573079779744148,
0.0007399305468425155,
-0.006614119280129671,
-0.13355432450771332,
-0.017269769683480263,
0.10707833617925644,
-0.099887415766716,
-0.020116161555051804,
0.01696583442389965,
0.02010101079940796,
-0.03818534314632416,
0.1121661439538002,
0.07975973188877106,
0.006575432606041431,
-0.042590584605932236,
0.15605852007865906,
0.06993306428194046,
-0.038220830261707306,
0.04225248098373413,
0.14076127111911774,
-0.0906439796090126,
-0.03071184828877449,
-0.04739043489098549,
0.14284491539001465,
-0.03571015223860741,
-0.00811675749719143,
-0.02650829404592514,
0.014546692371368408,
0.06883063167333603,
0.09235513210296631,
0.03784230723977089,
0.04276907444000244,
-0.07633406668901443,
0.016287142410874367,
-0.100075863301754,
0.08264851570129395,
0.024448180571198463,
-0.005087282042950392,
-0.06680984795093536,
0.11493821442127228,
-0.018865784630179405,
0.0019849182572215796,
-0.008730676025152206,
-0.03020055778324604,
-0.0830966904759407,
0.013989260420203209,
-0.01706809550523758,
0.056403741240501404,
-0.043344590812921524,
0.005273297429084778,
-0.011896385811269283,
0.04648590832948685,
0.056352272629737854,
0.022165633738040924,
-0.027728840708732605,
-0.04211370646953583,
0.0035974730271846056,
0.07339107245206833,
-0.13769051432609558,
-0.010250162333250046,
0.0256014633923769,
-0.07611856609582901,
0.10197782516479492,
0.052596934139728546,
-0.04646062105894089,
0.013485655188560486,
-0.085882768034935,
-0.007067435421049595,
0.020405719056725502,
0.023290341719985008,
-0.009182189591228962,
-0.07186893373727798,
0.02155136875808239,
-0.0038838922046124935,
-0.010629400610923767,
-0.016548028215765953,
0.116485595703125,
-0.0907212570309639,
0.027724744752049446,
0.04479002207517624,
0.005743269808590412,
-0.12673576176166534,
0.06859664618968964,
0.04535730928182602,
0.08213675022125244,
0.12047404050827026,
-0.04198312386870384,
0.022652510553598404,
-0.15293526649475098,
-0.010749256238341331,
0.04320720583200455,
0.024746181443333626,
-0.0843183845281601,
-0.04779815301299095,
0.03907107561826706,
-0.01630447618663311,
0.07144398242235184,
0.0020113757345825434,
-0.0016327154589816928,
-0.011981398798525333,
-0.05731254443526268,
-0.05202425643801689,
0.0029062575194984674,
0.07880816608667374,
-0.0031299092806875706,
-0.007723922375589609,
0.05577953904867172,
0.04234139993786812,
-0.003619949799031019,
0.09268522262573242,
0.08231611549854279,
0.08869221061468124,
0.08798416703939438,
0.05405737832188606,
0.043456126004457474,
0.018329907208681107,
-0.0638047531247139,
0.049665141850709915,
0.040128421038389206,
0.002400532364845276,
-0.09105098992586136,
0.06038711220026016,
0.14367054402828217,
-0.08989475667476654,
0.0867689773440361,
0.04339256137609482,
-0.06857313215732574,
-0.11031672358512878,
-0.11741133779287338,
-0.019875070080161095,
-0.0172116719186306,
-0.02090572938323021,
-0.12460078299045563,
0.03453372046351433,
-0.026969529688358307,
0.0012303243856877089,
0.009801797568798065,
0.067225381731987,
-0.014289526268839836,
-0.05415144935250282,
0.010049652308225632,
-0.03747290000319481,
0.04619092866778374,
0.0282547939568758,
0.03917326405644417,
0.08334213495254517,
0.012077467516064644,
0.06145456060767174,
0.08510935306549072,
0.11859038472175598,
-0.02246394008398056,
-0.03755982592701912,
-0.07700834423303604,
-0.0370614193379879,
-0.003282401477918029,
0.012927593663334846,
0.1432216763496399,
0.05575341731309891,
-0.09664667397737503,
-0.009389739483594894,
0.20454141497612,
-0.05186492204666138,
-0.024517541751265526,
-0.1514902263879776,
0.12018796056509018,
0.04278559610247612,
-0.000573126133531332,
-0.03493562340736389,
-0.08979955315589905,
0.012140429578721523,
0.1479053646326065,
0.13131485879421234,
-0.02579966001212597,
0.004795044660568237,
-0.003429051022976637,
0.00457742577418685,
-0.02873826026916504,
0.09868581593036652,
0.01568874530494213,
0.2066556066274643,
-0.04272089898586273,
0.16515925526618958,
-0.004804882220923901,
0.014380130916833878,
-0.054669253528118134,
0.07408507168292999,
-0.0872708335518837,
0.02359740622341633,
-0.028540892526507378,
0.03360387310385704,
-0.08384475857019424,
-0.19810771942138672,
-0.04473389685153961,
-0.06468790769577026,
-0.06388621777296066,
-0.03738918527960777,
-0.04916827380657196,
-0.016265615820884705,
0.07032698392868042,
-0.011319208890199661,
0.004324200097471476,
0.14288747310638428,
-0.013240591622889042,
-0.09940986335277557,
-0.05189475417137146,
0.023977911099791527,
0.01913764886558056,
0.1652471423149109,
0.0014692542608827353,
0.07173839211463928,
0.07322705537080765,
0.0017038877122104168,
-0.1465131789445877,
-0.009885860607028008,
-0.011107898317277431,
0.016439294442534447,
0.028835799545049667,
0.10973591357469559,
-0.03791453689336777,
0.08409137278795242,
0.03416841849684715,
-0.026821499690413475,
0.0035793595016002655,
0.044456012547016144,
-0.03809387609362602,
-0.11780483275651932,
0.037926170974969864,
-0.062093336135149,
0.11588665097951889,
0.17270658910274506,
-0.02781214937567711,
0.011155981570482254,
-0.06068216264247894,
0.0036447755992412567,
0.02104124240577221,
0.014847238548099995,
-0.02157619036734104,
-0.07308932393789291,
0.002148508559912443,
0.06590381264686584,
0.01050692517310381,
-0.18893179297447205,
-0.06251463294029236,
0.045697782188653946,
-0.06064147874712944,
-0.006717455107718706,
0.06530827283859253,
0.0885089710354805,
0.08902900665998459,
-0.03277764469385147,
-0.07289102673530579,
-0.012259796261787415,
0.09075101464986801,
-0.12660837173461914,
-0.07247806340456009
] |
null | null |
transformers
|
# doc2query/all-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/all-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=384, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 570k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a large collection of datasets. For the exact datasets names and weights see the `data_config.json` in this repository. Most of the datasets are available at [https://huggingface.co/sentence-transformers](https://huggingface.co/sentence-transformers).
The datasets include besides others:
- (title, body) pairs from [Reddit](https://huggingface.co/datasets/sentence-transformers/reddit-title-body)
- (title, body) pairs and (title, answer) pairs from StackExchange and Yahoo Answers!
- (title, review) pairs from Amazon reviews
- (query, paragraph) pairs from MS MARCO, NQ, and GooAQ
- (question, duplicate_question) from Quora and WikiAnswers
- (title, abstract) pairs from S2ORC
## Prefix
This model was trained **without a prefix**. In contrast to [doc2query/all-with_prefix-t5-base-v1](https://huggingface.co/doc2query/all-with_prefix-t5-base-v1) you cannot specify what type of transformation (answer2question, review2title) etc. you will have. This can lead to a mixture of output values.
|
{"language": "en", "license": "apache-2.0", "datasets": ["sentence-transformers/reddit-title-body", "sentence-transformers/embedding-training-data"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/all-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:sentence-transformers/reddit-title-body",
"dataset:sentence-transformers/embedding-training-data",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/reddit-title-body #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/all-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 570k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a large collection of datasets. For the exact datasets names and weights see the 'data_config.json' in this repository. Most of the datasets are available at URL
The datasets include besides others:
- (title, body) pairs from Reddit
- (title, body) pairs and (title, answer) pairs from StackExchange and Yahoo Answers!
- (title, review) pairs from Amazon reviews
- (query, paragraph) pairs from MS MARCO, NQ, and GooAQ
- (question, duplicate_question) from Quora and WikiAnswers
- (title, abstract) pairs from S2ORC
## Prefix
This model was trained without a prefix. In contrast to doc2query/all-with_prefix-t5-base-v1 you cannot specify what type of transformation (answer2question, review2title) etc. you will have. This can lead to a mixture of output values.
|
[
"# doc2query/all-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 570k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a large collection of datasets. For the exact datasets names and weights see the 'data_config.json' in this repository. Most of the datasets are available at URL\r\n\r\nThe datasets include besides others:\r\n- (title, body) pairs from Reddit\r\n- (title, body) pairs and (title, answer) pairs from StackExchange and Yahoo Answers!\r\n- (title, review) pairs from Amazon reviews\r\n- (query, paragraph) pairs from MS MARCO, NQ, and GooAQ \r\n- (question, duplicate_question) from Quora and WikiAnswers\r\n- (title, abstract) pairs from S2ORC",
"## Prefix\r\n\r\nThis model was trained without a prefix. In contrast to doc2query/all-with_prefix-t5-base-v1 you cannot specify what type of transformation (answer2question, review2title) etc. you will have. This can lead to a mixture of output values."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/reddit-title-body #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/all-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 570k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a large collection of datasets. For the exact datasets names and weights see the 'data_config.json' in this repository. Most of the datasets are available at URL\r\n\r\nThe datasets include besides others:\r\n- (title, body) pairs from Reddit\r\n- (title, body) pairs and (title, answer) pairs from StackExchange and Yahoo Answers!\r\n- (title, review) pairs from Amazon reviews\r\n- (query, paragraph) pairs from MS MARCO, NQ, and GooAQ \r\n- (question, duplicate_question) from Quora and WikiAnswers\r\n- (title, abstract) pairs from S2ORC",
"## Prefix\r\n\r\nThis model was trained without a prefix. In contrast to doc2query/all-with_prefix-t5-base-v1 you cannot specify what type of transformation (answer2question, review2title) etc. you will have. This can lead to a mixture of output values."
] |
[
107,
258,
32,
229,
69
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/reddit-title-body #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/all-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it."
] |
[
-0.0015355973737314343,
0.13657331466674805,
-0.006866020150482655,
0.08510024845600128,
0.13653098046779633,
0.021864572539925575,
0.031338583678007126,
0.130547896027565,
-0.029518820345401764,
0.011837875470519066,
0.03194548562169075,
0.07740020006895065,
0.001149933086708188,
0.0475979782640934,
0.0474981814622879,
-0.25061067938804626,
-0.00026831941795535386,
-0.018297117203474045,
0.09623805433511734,
0.09313691407442093,
0.07505934685468674,
-0.06368657201528549,
0.06454276293516159,
-0.012123404070734978,
-0.13189837336540222,
0.07331445068120956,
-0.03430916741490364,
-0.046654824167490005,
0.07605591416358948,
0.02259914204478264,
0.08100583404302597,
0.015290879644453526,
0.020702632144093513,
-0.12004068493843079,
0.030122844502329826,
0.08772090077400208,
-0.048006024211645126,
0.03522684425115585,
0.00895253848284483,
-0.05946881324052811,
0.07505833357572556,
-0.07701735943555832,
0.03487164154648781,
0.03210000321269035,
-0.09703490883111954,
-0.15114335715770721,
-0.05743643641471863,
0.0029771458357572556,
-0.0212724506855011,
0.061846453696489334,
-0.006872004363685846,
0.05401771143078804,
-0.062417224049568176,
0.04923402890563011,
0.2681262791156769,
-0.24130657315254211,
-0.015011734329164028,
0.06639005988836288,
-0.018967851996421814,
0.07452418655157089,
-0.022581499069929123,
-0.06659233570098877,
-0.02913561835885048,
0.037541601806879044,
0.06271982938051224,
-0.005040182266384363,
-0.034542396664619446,
0.008707214146852493,
-0.10505086183547974,
-0.08008314669132233,
0.12294746190309525,
-0.05223168805241585,
-0.06553878635168076,
-0.09477788209915161,
-0.0800475999712944,
0.036006540060043335,
-0.00031568133272230625,
-0.06269967555999756,
0.017818547785282135,
0.0194998811930418,
0.08085937798023224,
-0.05938476696610451,
-0.1261569857597351,
-0.02518988586962223,
-0.07973755151033401,
0.05212249606847763,
0.0746169239282608,
0.029472682625055313,
0.001212205272167921,
0.10895497351884842,
-0.08544924855232239,
-0.07689385116100311,
-0.048598647117614746,
-0.0889764353632927,
-0.0485091432929039,
0.029996249824762344,
-0.08852654695510864,
-0.18364529311656952,
0.03711267188191414,
0.051620952785015106,
-0.03240233287215233,
-0.03412661701440811,
-0.06292099505662918,
0.04301951825618744,
0.043770525604486465,
0.10193074494600296,
-0.03839597478508949,
0.05739221349358559,
0.062364351004362106,
0.004128666128963232,
-0.037976741790771484,
-0.010057184845209122,
-0.08666357398033142,
-0.09846169501543045,
0.038903627544641495,
0.019944248721003532,
-0.033611491322517395,
0.07828628271818161,
-0.03907790407538414,
-0.046407025307416916,
0.06270665675401688,
-0.12928704917430878,
-0.03964062035083771,
-0.0069422791711986065,
-0.04074768349528313,
0.006968945264816284,
0.13842958211898804,
-0.012251471169292927,
-0.10773997753858566,
0.015095844864845276,
-0.011920051649212837,
-0.019536849111318588,
-0.13859501481056213,
-0.09502057731151581,
0.022846553474664688,
-0.08647246658802032,
-0.059836335480213165,
-0.13250868022441864,
-0.1655416488647461,
-0.006274600978940725,
0.036165814846754074,
0.004437356721609831,
-0.016708215698599815,
0.004977783653885126,
0.016816692426800728,
-0.08259337395429611,
-0.00885666161775589,
-0.08178337663412094,
-0.004562789108604193,
0.059490934014320374,
-0.01105471607297659,
0.09406307339668274,
-0.04895176738500595,
0.03184077888727188,
-0.07658704370260239,
-0.0227963887155056,
-0.18032769858837128,
0.14537695050239563,
0.0037694713100790977,
-0.03030388429760933,
-0.08272487670183182,
-0.0323638916015625,
-0.07665815204381943,
-0.008605444803833961,
0.029869042336940765,
0.11281047016382217,
-0.2531362473964691,
-0.03230428323149681,
0.2056712806224823,
-0.11208869516849518,
-0.06637168675661087,
0.13193891942501068,
-0.06118541955947876,
0.11544682085514069,
0.10572642087936401,
0.11293833702802658,
0.1486172080039978,
-0.07308438420295715,
-0.05259835720062256,
-0.06324243545532227,
-0.09928955137729645,
0.05278424173593521,
0.07431259751319885,
-0.008635925129055977,
0.11471907049417496,
0.06357541680335999,
0.04394771158695221,
-0.03701414167881012,
-0.02137252502143383,
-0.06850828230381012,
-0.04148835316300392,
-0.010603649541735649,
0.058587148785591125,
-0.01520222332328558,
-0.0052116806618869305,
0.0603693388402462,
-0.06134070083498955,
0.06226818636059761,
0.04345722869038582,
-0.10129627585411072,
0.05587500333786011,
-0.07393995672464371,
0.009369103237986565,
-0.01544988714158535,
0.0463796928524971,
-0.20959416031837463,
-0.053678758442401886,
0.04325680807232857,
0.001717162667773664,
0.12006425857543945,
0.018731191754341125,
0.057982392609119415,
0.037356771528720856,
-0.028509464114904404,
0.03267607465386391,
-0.08543823659420013,
-0.0319509357213974,
-0.05665149167180061,
-0.05340731889009476,
-0.05585749074816704,
-0.02144734375178814,
0.06313216686248779,
-0.07841680198907852,
0.050715357065200806,
0.028857041150331497,
0.05449201166629791,
0.01410665363073349,
-0.014714677818119526,
-0.0009406930184923112,
0.019356384873390198,
0.010618625208735466,
0.019203808158636093,
0.07608577609062195,
0.00015478095156140625,
-0.044928792864084244,
0.03048078529536724,
-0.02772955596446991,
-0.023793278262019157,
0.05304345488548279,
0.04695820063352585,
-0.032930515706539154,
-0.161418616771698,
-0.07171692699193954,
-0.020037850365042686,
-0.06741569191217422,
-0.056933287531137466,
0.2247699648141861,
0.030241360887885094,
0.10074470192193985,
-0.09724011272192001,
0.0006294553168118,
0.018822841346263885,
-0.014547677710652351,
0.03909634053707123,
0.04090922698378563,
0.013616758398711681,
-0.060902003198862076,
0.02354271523654461,
0.05548541247844696,
-0.039097558706998825,
0.15449833869934082,
0.01024897675961256,
-0.12919683754444122,
-0.0026633860543370247,
0.03304710611701012,
0.030176233500242233,
-0.03400909900665283,
-0.05979016423225403,
-0.00010387933434685692,
0.04379907622933388,
0.002241035457700491,
0.032536398619413376,
-0.09381723403930664,
0.02282707765698433,
0.028366083279252052,
-0.020258789882063866,
-0.05380528047680855,
0.017397943884134293,
0.05836361646652222,
0.08206893503665924,
0.01697302982211113,
0.07533282041549683,
0.012948592193424702,
-0.052382562309503555,
-0.08509188890457153,
0.1786198616027832,
-0.057072948664426804,
-0.22131985425949097,
-0.07778476923704147,
0.00812386255711317,
-0.07973005622625351,
-0.015954388305544853,
0.07214091718196869,
-0.10192406922578812,
0.0069713531993329525,
-0.06925847381353378,
0.018360812216997147,
0.04298499599099159,
0.01049685850739479,
-0.12088768184185028,
0.05345261096954346,
-0.02320217899978161,
-0.14392825961112976,
-0.0019842968322336674,
-0.058703530579805374,
-0.1478554606437683,
0.08559036999940872,
-0.03236055001616478,
0.0891076996922493,
0.041449401527643204,
-0.05972408503293991,
0.02717479318380356,
-0.025998488068580627,
0.16105496883392334,
0.005802159197628498,
0.02340426668524742,
0.20509059727191925,
-0.06153169274330139,
0.09729677438735962,
0.07060737162828445,
0.029301753267645836,
-0.04107970744371414,
0.05579207465052605,
0.04515519365668297,
-0.036421116441488266,
-0.23170600831508636,
-0.02800910174846649,
-0.07163586467504501,
0.003955894615501165,
0.06667663156986237,
0.026339992880821228,
0.004022776614874601,
0.031273532658815384,
-0.010535423643887043,
0.024270622059702873,
0.026224037632346153,
0.07625999301671982,
0.11361010372638702,
-0.002201197436079383,
0.07768553495407104,
-0.06688030064105988,
-0.05072253569960594,
0.03732026368379593,
0.12204833328723907,
0.2350078821182251,
-0.10219411551952362,
0.1217692419886589,
0.1168561726808548,
0.10909582674503326,
0.06588532775640488,
0.09265195578336716,
-0.08313427865505219,
0.01896597445011139,
-0.05697489529848099,
-0.08882948011159897,
-0.08291658014059067,
0.05261868238449097,
-0.01557577308267355,
-0.02534312568604946,
-0.04608537629246712,
0.001821886282414198,
0.056047745048999786,
0.1937132626771927,
-0.08573805540800095,
-0.16634859144687653,
-0.014743446372449398,
-0.047534745186567307,
-0.004406881518661976,
-0.017619110643863678,
-0.0004558266664389521,
0.04450166970491409,
-0.10525093227624893,
-0.000910930975805968,
-0.03996245190501213,
0.07299690693616867,
-0.04852047562599182,
0.037330638617277145,
0.01134033128619194,
0.09613646566867828,
0.037070564925670624,
0.08633555471897125,
-0.22231321036815643,
0.06319338083267212,
0.018317654728889465,
0.0191703699529171,
-0.11132080852985382,
0.02835323102772236,
-0.015238589607179165,
-0.03333998844027519,
0.12801715731620789,
-0.013322440907359123,
0.13867433369159698,
0.04138461500406265,
-0.057588595896959305,
0.04953647777438164,
0.08548730611801147,
-0.054670002311468124,
0.11443540453910828,
0.010759328491985798,
0.00795829202979803,
-0.013556353747844696,
0.18806268274784088,
-0.07147703319787979,
-0.22791741788387299,
0.06946894526481628,
-0.060308732092380524,
0.006223238073289394,
-0.024975504726171494,
-0.05807226523756981,
0.04741375148296356,
0.17881228029727936,
-0.07075992971658707,
-0.09726909548044205,
-0.08422549813985825,
0.04982249438762665,
0.09142642468214035,
-0.07748229056596756,
0.012273570522665977,
-0.020766697824001312,
0.10661406815052032,
-0.0695977732539177,
-0.12839001417160034,
0.002770893042907119,
-0.00007068584818625823,
-0.054595936089754105,
-0.03945748135447502,
0.11184925585985184,
0.011802013963460922,
0.044669825583696365,
0.034987516701221466,
0.011538337916135788,
-0.027099426835775375,
-0.0675504207611084,
-0.017447838559746742,
0.147894486784935,
0.11508134752511978,
0.0636439323425293,
-0.1967281848192215,
-0.06596824526786804,
-0.0818428099155426,
0.0055981529876589775,
0.20277860760688782,
0.10534550249576569,
-0.08051937073469162,
0.13319578766822815,
0.10636510699987411,
-0.10645284503698349,
-0.22469420731067657,
-0.038449596613645554,
0.1562872976064682,
0.011403151787817478,
-0.004658865742385387,
-0.26600801944732666,
0.08707460016012192,
0.10540935397148132,
0.02824063040316105,
-0.07182002812623978,
-0.2810685634613037,
-0.11959848552942276,
0.01753537543118,
0.06875637918710709,
0.13282141089439392,
-0.11562363803386688,
0.005833879578858614,
-0.01576855033636093,
-0.0792803168296814,
0.20657721161842346,
-0.20057153701782227,
0.08132994174957275,
-0.008553560823202133,
0.0556296743452549,
0.053081028163433075,
-0.026444919407367706,
0.010720317251980305,
0.02193417027592659,
0.00991568248718977,
-0.029535919427871704,
0.0035879849456250668,
0.057693641632795334,
-0.03384340927004814,
0.14616850018501282,
0.07575824856758118,
0.02773781679570675,
-0.06586987525224686,
-0.05411509796977043,
-0.08972490578889847,
-0.015522877685725689,
-0.006955245044082403,
-0.07781383395195007,
0.01528354175388813,
0.035380784422159195,
0.09150012582540512,
-0.008926784619688988,
0.030934663489460945,
-0.09798406809568405,
0.08256489783525467,
0.15231309831142426,
0.12138186395168304,
-0.06347586214542389,
-0.0952095314860344,
-0.005705777555704117,
-0.009802326560020447,
0.05179207772016525,
-0.15704892575740814,
0.03744354471564293,
0.09472016990184784,
-0.006257222034037113,
0.08706513047218323,
0.08318614214658737,
-0.11997669190168381,
0.00900262501090765,
0.09063150733709335,
-0.1432814598083496,
-0.1099514290690422,
-0.0161557886749506,
0.012954339385032654,
-0.16802604496479034,
0.030967911705374718,
0.0918232873082161,
-0.08574831485748291,
-0.05089402571320534,
0.011090968735516071,
0.002308817580342293,
-0.018545931205153465,
0.06552954018115997,
0.04895232990384102,
0.015371986664831638,
-0.08299437910318375,
0.11724310368299484,
0.1008821576833725,
-0.09763170778751373,
0.0061663007363677025,
0.1517646610736847,
-0.08783386647701263,
-0.05358600243926048,
-0.11844878643751144,
0.048774559050798416,
-0.055193327367305756,
0.004709622822701931,
-0.044156767427921295,
-0.020048238337039948,
0.04548356682062149,
0.1263536810874939,
0.045856453478336334,
0.05149931088089943,
-0.04307816922664642,
0.012206635437905788,
-0.09081336855888367,
0.07501211017370224,
-0.019748760387301445,
0.011278050020337105,
-0.02354036085307598,
0.07914700359106064,
-0.007959657348692417,
0.08270025253295898,
-0.014936196617782116,
-0.04970141127705574,
-0.0632544457912445,
0.013011053204536438,
-0.02188972383737564,
-0.007478796876966953,
-0.009826726280152798,
-0.009517592377960682,
-0.023055659607052803,
0.03306197747588158,
0.05115034431219101,
0.04266267269849777,
-0.048206496983766556,
-0.02661382406949997,
-0.02107315883040428,
0.06501871347427368,
-0.13446688652038574,
-0.014888199046254158,
0.027639739215373993,
-0.05218355357646942,
0.1277395635843277,
0.04303417727351189,
-0.08570980280637741,
0.023921959102153778,
-0.058103062212467194,
-0.05972466245293617,
0.0510026179254055,
0.05117998644709587,
-0.0037803053855895996,
-0.001357220346108079,
0.05179400369524956,
0.018997564911842346,
-0.04405892267823219,
-0.00351333967410028,
0.12783625721931458,
-0.10460656881332397,
0.05324028804898262,
0.04612162709236145,
-0.018187284469604492,
-0.10498656332492828,
0.012894496321678162,
0.0033568504732102156,
0.11593101918697357,
0.09712591022253036,
-0.047140538692474365,
0.04568490758538246,
-0.14993910491466522,
-0.046574823558330536,
0.016278261318802834,
-0.022677099332213402,
-0.043488625437021255,
-0.08251824975013733,
0.026903491467237473,
0.011261426843702793,
0.14356614649295807,
0.0039416090585291386,
0.051295239478349686,
0.000819832319393754,
-0.029928097501397133,
-0.008208525367081165,
-0.005216815508902073,
0.04340631514787674,
0.035370562225580215,
-0.02360246144235134,
0.007908210158348083,
0.05298496410250664,
-0.04619630053639412,
0.06670654565095901,
0.07079241424798965,
0.09269516170024872,
0.1814342886209488,
-0.008686235174536705,
0.09525499492883682,
0.0588853619992733,
-0.025910206139087677,
0.021239668130874634,
0.06391612440347672,
0.0035427485127002,
-0.07850771397352219,
0.01135567482560873,
0.19349661469459534,
-0.043879300355911255,
0.11947173625230789,
0.025889448821544647,
-0.046272922307252884,
-0.08208312094211578,
-0.0892229974269867,
-0.040456391870975494,
-0.046554841101169586,
-0.002951576840132475,
-0.13742797076702118,
0.05680560693144798,
-0.02329116500914097,
-0.012971008196473122,
-0.0072272103279829025,
0.12598955631256104,
0.03579084575176239,
-0.10874734073877335,
0.023709246888756752,
-0.04195985570549965,
0.08373726159334183,
0.10191400349140167,
0.0398595929145813,
0.06346537917852402,
0.04492278769612312,
0.05333945155143738,
0.09131267666816711,
0.09728869050741196,
-0.014301368966698647,
-0.10422483831644058,
-0.06685642898082733,
-0.02783816121518612,
0.009745373390614986,
0.012670725584030151,
0.1357518881559372,
0.07163457572460175,
-0.09073004871606827,
0.00787147507071495,
0.1774459332227707,
-0.008337846025824547,
-0.10413998365402222,
-0.12929998338222504,
0.16070766746997833,
0.13762342929840088,
0.013433534651994705,
-0.01531398855149746,
-0.08897735923528671,
-0.02002609334886074,
0.12837183475494385,
0.16487306356430054,
-0.050199955701828,
0.016917238011956215,
0.06261947751045227,
0.025347568094730377,
0.008865063078701496,
0.14853735268115997,
0.09810762107372284,
0.2955939471721649,
-0.07219219952821732,
0.11114105582237244,
-0.030208952724933624,
0.03464478999376297,
-0.1297183483839035,
0.05568326264619827,
-0.030097568407654762,
-0.01048697717487812,
-0.02634386345744133,
0.046093668788671494,
-0.049033962190151215,
-0.12578286230564117,
-0.0867147147655487,
-0.16669020056724548,
-0.09065043181180954,
-0.06666137278079987,
0.00561863649636507,
-0.015238880179822445,
0.08098392933607101,
0.0005922513082623482,
0.006868582684546709,
0.11367862671613693,
0.010533249005675316,
-0.07311961054801941,
0.001072215149179101,
0.07998638600111008,
0.028013067319989204,
0.07901660352945328,
0.01201064046472311,
0.09395721554756165,
0.0439128503203392,
0.02527157962322235,
-0.09248008579015732,
0.021394100040197372,
-0.018083175644278526,
-0.04956749454140663,
-0.019569728523492813,
0.057745497673749924,
-0.009930295869708061,
-0.010864859446883202,
0.019431836903095245,
-0.07035395503044128,
0.030698828399181366,
0.058528415858745575,
-0.011035347357392311,
-0.07648731768131256,
0.0819881409406662,
-0.08243881165981293,
0.07987557351589203,
0.17577537894248962,
0.003066294826567173,
0.017964022234082222,
-0.11799708753824234,
-0.024699261412024498,
0.04049624875187874,
-0.0057798647321760654,
-0.06290535628795624,
-0.047060828655958176,
-0.01298173051327467,
0.04340210556983948,
-0.0016451821429654956,
-0.17719101905822754,
-0.08492711186408997,
0.10187628865242004,
-0.015655353665351868,
-0.08594939857721329,
0.057613447308540344,
0.05574333295226097,
0.06480259448289871,
-0.005994918756186962,
-0.18404637277126312,
0.018903125077486038,
0.0760054960846901,
-0.14375269412994385,
-0.07739865779876709
] |
null | null |
transformers
|
# doc2query/all-with_prefix-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/all-with_prefix-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
prefix = "answer2question"
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
text = prefix+": "+text
input_ids = tokenizer.encode(text, max_length=384, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 575k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a large collection of datasets. For the exact datasets names and weights see the `data_config.json` in this repository. Most of the datasets are available at [https://huggingface.co/sentence-transformers](https://huggingface.co/sentence-transformers).
The datasets include besides others:
- (title, body) pairs from [Reddit](https://huggingface.co/datasets/sentence-transformers/reddit-title-body)
- (title, body) pairs and (title, answer) pairs from StackExchange and Yahoo Answers!
- (title, review) pairs from Amazon reviews
- (query, paragraph) pairs from MS MARCO, NQ, and GooAQ
- (question, duplicate_question) from Quora and WikiAnswers
- (title, abstract) pairs from S2ORC
## Prefix
This model was trained **with a prefix**: You start the text with a specific index that defines what type out output text you would like to receive. Depending on the prefix, the output is different.
E.g. the above text about Python produces the following output:
| Prefix | Output |
| --- | --- |
| answer2question | Why should I use python in my business? ; What is the difference between Python and.NET? ; what is the python design philosophy? |
| review2title | Python a powerful and useful language ; A new and improved programming language ; Object-oriented, practical and accessibl |
| abstract2title | Python: A Software Development Platform ; A Research Guide for Python X: Conceptual Approach to Programming ; Python : Language and Approach |
| text2query | is python a low level language? ; what is the primary idea of python? ; is python a programming language? |
These are all available pre-fixes:
- text2reddit
- question2title
- answer2question
- abstract2title
- review2title
- news2title
- text2query
- question2question
For the datasets and weights for the different pre-fixes see `data_config.json` in this repository.
|
{"language": "en", "license": "apache-2.0", "datasets": ["sentence-transformers/reddit-title-body", "sentence-transformers/embedding-training-data"], "widget": [{"text": "text2reddit: Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/all-with_prefix-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:sentence-transformers/reddit-title-body",
"dataset:sentence-transformers/embedding-training-data",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/reddit-title-body #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
|
doc2query/all-with\_prefix-t5-base-v1
=====================================
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
* Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
* Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
Usage
-----
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
Training
--------
This model fine-tuned google/t5-v1\_1-base for 575k training steps. For the training script, see the 'train\_script.py' in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a large collection of datasets. For the exact datasets names and weights see the 'data\_config.json' in this repository. Most of the datasets are available at URL
The datasets include besides others:
* (title, body) pairs from Reddit
* (title, body) pairs and (title, answer) pairs from StackExchange and Yahoo Answers!
* (title, review) pairs from Amazon reviews
* (query, paragraph) pairs from MS MARCO, NQ, and GooAQ
* (question, duplicate\_question) from Quora and WikiAnswers
* (title, abstract) pairs from S2ORC
Prefix
------
This model was trained with a prefix: You start the text with a specific index that defines what type out output text you would like to receive. Depending on the prefix, the output is different.
E.g. the above text about Python produces the following output:
These are all available pre-fixes:
* text2reddit
* question2title
* answer2question
* abstract2title
* review2title
* news2title
* text2query
* question2question
For the datasets and weights for the different pre-fixes see 'data\_config.json' in this repository.
|
[] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/reddit-title-body #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n"
] |
[
111
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/reddit-title-body #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n"
] |
[
-0.05958070605993271,
0.12408869713544846,
-0.00422604801133275,
0.06526899337768555,
0.13922615349292755,
0.02667899616062641,
0.12928642332553864,
0.15249235928058624,
-0.1295546293258667,
-0.03370869904756546,
0.13829754292964935,
0.19431892037391663,
0.020466187968850136,
0.04601707309484482,
-0.045978523790836334,
-0.23090539872646332,
0.06599131971597672,
0.033157218247652054,
-0.07074404507875443,
0.14608079195022583,
0.12415239959955215,
-0.043928660452365875,
0.07875528186559677,
-0.022024014964699745,
-0.15757615864276886,
0.03282861039042473,
0.026449577882885933,
-0.1045517697930336,
0.13420893251895905,
0.035602301359176636,
0.09172951430082321,
0.060449689626693726,
-0.006116262637078762,
-0.11896460503339767,
0.018004538491368294,
0.06095479056239128,
-0.07503946870565414,
0.10298176854848862,
0.014489131979644299,
-0.03185947611927986,
0.09978203475475311,
-0.08781173080205917,
-0.008274570107460022,
0.055944185703992844,
-0.13001586496829987,
-0.11193163692951202,
-0.047905292361974716,
0.04133518785238266,
0.046673160046339035,
0.13120388984680176,
-0.029921462759375572,
0.13211360573768616,
-0.10034263134002686,
0.10369966924190521,
0.1569356769323349,
-0.2920709252357483,
-0.006179192569106817,
0.037480343133211136,
0.07457397878170013,
0.09023204445838928,
-0.03729530796408653,
0.01710866391658783,
0.04068417474627495,
0.04127892106771469,
0.06107386201620102,
-0.039838988333940506,
-0.22007563710212708,
0.08526923507452011,
-0.11970219016075134,
-0.04440466687083244,
0.34201961755752563,
-0.00484159542247653,
0.022372880950570107,
-0.021036453545093536,
-0.10149676352739334,
-0.07121304422616959,
0.019138991832733154,
-0.0015841092681512237,
0.003993213642388582,
0.045462604612112045,
-0.06402590870857239,
-0.07549538463354111,
-0.15043213963508606,
-0.03804056718945503,
-0.19697092473506927,
0.03273700922727585,
-0.0024642895441502333,
0.026242269203066826,
-0.15087442100048065,
0.10907197743654251,
0.08202217519283295,
-0.10199838131666183,
0.028188079595565796,
-0.05723901465535164,
0.005436474457383156,
0.037623196840286255,
-0.09631180763244629,
-0.12102635204792023,
0.07039379328489304,
0.08123786002397537,
0.029117247089743614,
-0.024685027077794075,
-0.06842394173145294,
0.09639774262905121,
-0.024867357686161995,
0.08069421350955963,
-0.08514208346605301,
-0.015501282177865505,
0.04119903966784477,
-0.03520166873931885,
0.03812184929847717,
-0.042142353951931,
-0.1504243165254593,
-0.06742116808891296,
0.054761674255132675,
0.08471204340457916,
0.08365891128778458,
0.12338877469301224,
-0.06533072888851166,
-0.028566328808665276,
0.03505899757146835,
-0.0884685218334198,
-0.022694477811455727,
-0.002406728221103549,
-0.02488682046532631,
0.03854823857545853,
0.11081066727638245,
0.033698879182338715,
-0.10912476480007172,
-0.01986408233642578,
-0.09806419163942337,
-0.02669372968375683,
-0.05552241951227188,
-0.12018176168203354,
0.06664218753576279,
-0.06320728361606598,
0.008914647623896599,
-0.13876959681510925,
-0.16732284426689148,
0.015526755712926388,
0.03578900545835495,
0.0004035879683215171,
-0.08314167708158493,
-0.05944255739450455,
-0.05332983285188675,
0.051340214908123016,
-0.0772642269730568,
0.08470763266086578,
-0.03594925254583359,
0.10234228521585464,
-0.08740167319774628,
0.08510783314704895,
-0.14032141864299774,
0.07736887782812119,
-0.1341036856174469,
-0.03793436288833618,
0.008288446813821793,
0.052601344883441925,
-0.017195411026477814,
0.09982873499393463,
-0.08177252113819122,
-0.024297336116433144,
-0.030645878985524178,
0.027310075238347054,
0.0038998050149530172,
0.1891200691461563,
-0.2072938233613968,
-0.07316544651985168,
0.22052010893821716,
-0.0618465282022953,
-0.18743689358234406,
0.0959879606962204,
-0.0028515136800706387,
0.09472756087779999,
0.08441777527332306,
0.1679535061120987,
-0.012040660716593266,
-0.054309383034706116,
0.023528525605797768,
0.09339151531457901,
-0.0771290585398674,
-0.1076422780752182,
0.07322391867637634,
-0.009977241046726704,
-0.01318299025297165,
0.022074392065405846,
-0.0007115818443708122,
0.08511261641979218,
-0.030125752091407776,
-0.07995948195457458,
-0.051659662276506424,
-0.044579531997442245,
0.07299989461898804,
0.00654682656750083,
0.03862406685948372,
-0.06680680811405182,
-0.08278253674507141,
0.0501885712146759,
0.06675180792808533,
-0.03293558955192566,
0.07895540446043015,
-0.03842148929834366,
0.006708038970828056,
0.012659711763262749,
0.023783842101693153,
-0.1751861572265625,
-0.03957502171397209,
-0.04536181688308716,
0.15958556532859802,
0.030780749395489693,
0.09803267568349838,
0.06257078796625137,
-0.0347738154232502,
-0.02626439370214939,
-0.03267195075750351,
0.030710699036717415,
0.02910671941936016,
-0.07103657722473145,
-0.15743036568164825,
0.03351766616106033,
-0.03815274313092232,
0.15412521362304688,
-0.07883620262145996,
0.052687644958496094,
0.03720277547836304,
0.09418035298585892,
0.006820664741098881,
0.062020279467105865,
0.0016856819856911898,
-0.0031440185848623514,
-0.05880627781152725,
-0.014970432966947556,
0.09648830443620682,
0.03698650002479553,
-0.11362186074256897,
0.1842544674873352,
-0.10289241373538971,
0.14024436473846436,
0.19091029465198517,
-0.17453041672706604,
0.044013842940330505,
-0.039892394095659256,
-0.06064513698220253,
-0.01049396675080061,
0.01421939954161644,
-0.08134208619594574,
0.03309096023440361,
0.01508273184299469,
0.14474822580814362,
-0.0571281760931015,
-0.015582622960209846,
-0.02339800074696541,
-0.0645112544298172,
-0.021586747840046883,
0.09608075767755508,
0.03388489782810211,
-0.159751296043396,
0.19076107442378998,
0.22424288094043732,
-0.06957951188087463,
0.17502065002918243,
-0.05277397856116295,
-0.06441816687583923,
0.015150529332458973,
-0.033550869673490524,
-0.03077477216720581,
0.0010182750411331654,
-0.16312186419963837,
-0.01443481259047985,
0.09832955151796341,
0.028642265126109123,
0.06806566566228867,
-0.13976801931858063,
-0.0529106929898262,
0.003370668040588498,
-0.034616418182849884,
-0.02632136270403862,
0.062094058841466904,
0.017708536237478256,
0.14131170511245728,
-0.026987211778759956,
-0.03693120554089546,
0.112855464220047,
0.00562208890914917,
-0.11900616437196732,
0.18093928694725037,
-0.17955631017684937,
-0.32308176159858704,
-0.13185884058475494,
-0.08972378820180893,
-0.03898223862051964,
0.005082788877189159,
0.1294281780719757,
-0.13510344922542572,
-0.009101475588977337,
-0.03813939169049263,
0.05121795833110809,
-0.049236688762903214,
0.0019237591186538339,
-0.06186000257730484,
0.0688956081867218,
0.009856580756604671,
-0.135063037276268,
-0.02045009657740593,
0.0007087505655363202,
-0.06615540385246277,
0.11893089860677719,
-0.09858857840299606,
0.09798350185155869,
0.15614324808120728,
0.03114084154367447,
0.03167381137609482,
-0.05411124974489212,
0.14000335335731506,
-0.033746346831321716,
0.03796005994081497,
0.2082531601190567,
0.0027016981039196253,
0.0591442808508873,
0.10702072829008102,
-0.019638093188405037,
-0.023139726370573044,
0.05172399431467056,
-0.0012753582559525967,
-0.027866382151842117,
-0.283380925655365,
-0.10890309512615204,
-0.11277768015861511,
0.06339340656995773,
0.054729875177145004,
0.05455992370843887,
0.11663070321083069,
0.0661662369966507,
-0.005150665994733572,
0.031271036714315414,
-0.0017066138098016381,
0.06042497977614403,
0.17648428678512573,
-0.06375709176063538,
0.13779756426811218,
-0.0907658264040947,
-0.06751219928264618,
0.0985170528292656,
0.05978044122457504,
0.15762834250926971,
0.02815661020576954,
0.11148301512002945,
0.06999394297599792,
0.1349072903394699,
0.04552771896123886,
0.1237231194972992,
0.015315836295485497,
-0.013130622915923595,
-0.04890209808945656,
-0.06595247238874435,
-0.07402873039245605,
0.05186997354030609,
0.0010968861170113087,
-0.06681961566209793,
-0.06502789258956909,
-0.012042982503771782,
0.09025447815656662,
0.14340850710868835,
0.08789350092411041,
-0.2550944983959198,
-0.013417798094451427,
0.052494779229164124,
0.008800143375992775,
-0.06638403981924057,
0.04600339010357857,
0.038721632212400436,
-0.08158967643976212,
0.027493352070450783,
-0.05082820728421211,
0.11354298144578934,
0.009236005134880543,
0.06780395656824112,
-0.040576718747615814,
-0.03350105881690979,
0.01654653623700142,
0.12939007580280304,
-0.36893677711486816,
0.22585883736610413,
-0.004029839765280485,
-0.04908720403909683,
-0.10424349457025528,
-0.0012763983104377985,
0.027761345729231834,
0.13561412692070007,
0.1260969340801239,
0.01314975693821907,
-0.016103483736515045,
0.012016079388558865,
-0.06515958160161972,
0.04990728572010994,
0.030482621863484383,
-0.012265179306268692,
-0.0023772288113832474,
-0.054177429527044296,
0.020945189520716667,
0.05707385390996933,
0.14558184146881104,
-0.0581367053091526,
-0.16438475251197815,
0.044500794261693954,
0.07423333823680878,
0.05387570336461067,
-0.04761945456266403,
-0.06401491165161133,
-0.01587076112627983,
0.2105930894613266,
-0.06847523152828217,
-0.06929758191108704,
-0.13175413012504578,
0.010678757913410664,
0.06051678583025932,
-0.08187339454889297,
0.015030667185783386,
-0.030889583751559258,
0.030819274485111237,
-0.053046632558107376,
-0.16859464347362518,
0.1329798698425293,
-0.09061299264431,
-0.031047675758600235,
-0.0670308843255043,
0.09470691531896591,
-0.04072178155183792,
0.06584151834249496,
-0.014001539908349514,
-0.00552767887711525,
-0.1108502745628357,
-0.05023874714970589,
0.015865523368120193,
0.0346420593559742,
0.09223786741495132,
0.021572338417172432,
-0.16203537583351135,
-0.03635968267917633,
-0.018593832850456238,
-0.042330823838710785,
0.31678876280784607,
0.12753422558307648,
-0.0743190199136734,
0.12583190202713013,
0.1495547890663147,
-0.07942325621843338,
-0.3243211805820465,
-0.06969405710697174,
-0.07362382113933563,
0.011316205374896526,
-0.007704772986471653,
-0.2098163366317749,
0.042536500841379166,
0.05104300379753113,
-0.031218798831105232,
-0.0076741548255085945,
-0.2644009590148926,
-0.07737066596746445,
0.19603334367275238,
0.02910890430212021,
0.23380231857299805,
-0.14475516974925995,
-0.050764475017786026,
-0.06314191967248917,
-0.16821373999118805,
0.2057599276304245,
-0.24355609714984894,
0.0944238007068634,
0.0017903007101267576,
0.0865073874592781,
0.017804738134145737,
-0.0382080115377903,
0.09019292145967484,
0.018805451691150665,
0.01534687913954258,
-0.1142648234963417,
-0.05391779541969299,
0.13472440838813782,
-0.03232986479997635,
0.08215991407632828,
-0.15835969150066376,
0.05111384764313698,
-0.13686604797840118,
0.017624245956540108,
-0.09135892242193222,
0.05055069178342819,
-0.01685990020632744,
-0.10858891159296036,
-0.03762943670153618,
0.017248697578907013,
0.05361398309469223,
-0.019380973652005196,
0.18652048707008362,
-0.008396477438509464,
0.06642372906208038,
0.16629400849342346,
0.10152757167816162,
-0.08096548169851303,
0.01705690659582615,
-0.06965366005897522,
-0.06382784247398376,
0.07789181917905807,
-0.19721092283725739,
0.04978230968117714,
0.08462490141391754,
-0.030146999284625053,
0.13041149079799652,
0.07560987025499344,
-0.01707339473068714,
-0.0333159938454628,
0.12208877503871918,
-0.18831615149974823,
0.024328041821718216,
-0.06462416797876358,
0.018481861799955368,
-0.061060745269060135,
0.04240836575627327,
0.16796448826789856,
-0.007902723737061024,
-0.010796453803777695,
0.0035172337666153908,
0.053427357226610184,
-0.08431417495012283,
0.08687519282102585,
0.08291654288768768,
0.023111307993531227,
-0.11090036481618881,
0.10696055740118027,
0.046206776052713394,
-0.11736524105072021,
0.03806116804480553,
0.11774276942014694,
-0.12253575772047043,
-0.1182246059179306,
-0.013418306596577168,
0.07642783224582672,
-0.13420304656028748,
-0.06918656080961227,
-0.08095484972000122,
-0.10742385685443878,
0.06956851482391357,
0.05177611857652664,
0.06362826377153397,
0.08255278319120407,
-0.04503811523318291,
-0.06881241500377655,
-0.04251285642385483,
0.06060853600502014,
-0.004568102769553661,
0.016827067360281944,
-0.08090409636497498,
0.10779067128896713,
-0.018855273723602295,
0.11493095755577087,
-0.046144939959049225,
-0.006300600245594978,
-0.10249080508947372,
0.025039877742528915,
-0.11343715339899063,
-0.068264901638031,
-0.035653743892908096,
-0.04402237758040428,
-0.010105811059474945,
-0.04478161036968231,
-0.04817698895931244,
0.007818199694156647,
-0.0955159068107605,
-0.01640385575592518,
-0.013131369836628437,
0.06827395409345627,
-0.11174847930669785,
-0.03243899717926979,
0.05862105265259743,
-0.04061039909720421,
0.12051216512918472,
0.08397519588470459,
-0.07845960557460785,
0.08652876317501068,
-0.10872647166252136,
-0.06931479275226593,
0.08051323890686035,
0.023702846840023994,
0.07318704575300217,
-0.012596717104315758,
-0.01426742598414421,
0.06326716393232346,
0.018621928989887238,
0.039632879197597504,
0.032704684883356094,
-0.08581787347793579,
0.03128306195139885,
-0.03844596445560455,
-0.09053202718496323,
-0.08577220886945724,
0.006153948605060577,
0.06559418886899948,
0.03153776377439499,
0.15981516242027283,
-0.04796935245394707,
0.05140646919608116,
-0.13551412522792816,
0.03266405314207077,
-0.0023362429346889257,
-0.11147809773683548,
-0.04174520820379257,
-0.08469683676958084,
0.02486967295408249,
-0.047604091465473175,
0.15370076894760132,
0.028611496090888977,
-0.08488835394382477,
0.04861331358551979,
0.08120749145746231,
0.012017721310257912,
0.00005572645386564545,
0.2308415025472641,
0.04903580620884895,
-0.02569117583334446,
-0.05903477966785431,
0.02066265046596527,
0.04017835855484009,
0.07866031676530838,
0.13752272725105286,
0.08541582524776459,
0.09503352642059326,
0.11283284425735474,
0.04377267509698868,
-0.012444352731108665,
-0.08465182036161423,
-0.03197842836380005,
0.005904581397771835,
0.07398174703121185,
-0.017337646335363388,
0.12185756862163544,
0.17108651995658875,
-0.05949389934539795,
0.022989990189671516,
-0.018998565152287483,
-0.05371229350566864,
-0.13525542616844177,
-0.12247645109891891,
-0.06731238216161728,
-0.11781615018844604,
-0.03292795643210411,
-0.12230035662651062,
0.02203109860420227,
0.08135505020618439,
0.037372320890426636,
-0.039336416870355606,
0.06153438985347748,
0.07690752297639847,
-0.11805536597967148,
0.05062178894877434,
-0.02649800106883049,
0.03717034310102463,
-0.09505773335695267,
-0.013110621832311153,
0.020785337314009666,
0.06310193985700607,
0.009734476916491985,
0.02630419097840786,
0.013657393865287304,
0.055212073028087616,
-0.11676053702831268,
-0.08614682406187057,
-0.05054905265569687,
0.04507225751876831,
-0.018208112567663193,
0.15443801879882812,
0.05355403199791908,
-0.007221211679279804,
0.05776809900999069,
0.20525725185871124,
-0.08736024796962738,
-0.14597399532794952,
-0.10907981544733047,
0.15259452164173126,
0.025686148554086685,
0.025340113788843155,
0.01381487213075161,
-0.05210018903017044,
-0.05907769128680229,
0.27895936369895935,
0.3101654648780823,
-0.11440480500459671,
0.007381260860711336,
0.027782751247286797,
0.020004047080874443,
0.01603304035961628,
0.10394243896007538,
0.13039852678775787,
0.13265222311019897,
-0.0697617307305336,
-0.006316772662103176,
-0.06392831355333328,
-0.016487516462802887,
-0.06271450966596603,
0.10473155975341797,
0.025475507602095604,
-0.06812036037445068,
0.00292101944796741,
0.08309385925531387,
-0.1133376881480217,
0.031772222369909286,
-0.10175050795078278,
-0.14867721498012543,
-0.09835411608219147,
-0.017724955454468727,
0.15861593186855316,
0.04191621392965317,
0.0707792267203331,
-0.018466290086507797,
-0.011758954264223576,
0.12489964812994003,
-0.0006547230877913535,
-0.17521882057189941,
0.0036163299810141325,
0.10621611773967743,
-0.13430586457252502,
0.0908503606915474,
-0.014141530729830265,
0.038213033229112625,
0.11560855805873871,
0.050968579947948456,
-0.09327622503042221,
0.08704706281423569,
-0.0006575548322871327,
0.019456099718809128,
0.02459365874528885,
-0.028403406962752342,
-0.0003926373028662056,
-0.022131334990262985,
0.06479448080062866,
-0.1198553815484047,
0.035394780337810516,
0.014228884130716324,
-0.03895639628171921,
-0.04325743392109871,
0.01343414094299078,
-0.04967976734042168,
0.07347583025693893,
0.11612644046545029,
-0.03915231674909592,
-0.0026679562870413065,
-0.10624697059392929,
-0.05022074282169342,
0.013399647548794746,
-0.10621944814920425,
-0.03327120840549469,
-0.06264816224575043,
-0.052756380289793015,
0.08776642382144928,
0.013545092195272446,
-0.2649105489253998,
-0.030980097129940987,
-0.054579876363277435,
-0.032995983958244324,
-0.14560876786708832,
0.07219606637954712,
0.10279950499534607,
0.0020194086246192455,
-0.00048354786122217774,
0.0370153971016407,
0.03593174368143082,
0.07641206681728363,
-0.11388169229030609,
-0.07859092205762863
] |
null | null |
transformers
|
# doc2query/msmarco-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/msmarco-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (query, passage) from the [MS MARCO Passage-Ranking dataset](https://github.com/microsoft/MSMARCO-Passage-Ranking).
|
{"language": "en", "license": "apache-2.0", "datasets": ["sentence-transformers/embedding-training-data"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/msmarco-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:sentence-transformers/embedding-training-data",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/msmarco-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset.
|
[
"# doc2query/msmarco-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/msmarco-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset."
] |
[
91,
260,
32,
113
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/msmarco-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-base for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset."
] |
[
-0.05538814887404442,
0.13173383474349976,
-0.006905203685164452,
0.10253584384918213,
0.11215224862098694,
-0.004202106036245823,
0.06253759562969208,
0.1035158708691597,
-0.0835832878947258,
0.04047321155667305,
0.06760478019714355,
0.05412013828754425,
0.0023551639169454575,
0.12148259580135345,
0.07597734034061432,
-0.18185843527317047,
-0.03417827561497688,
-0.04142805188894272,
0.09874522686004639,
0.07748589664697647,
0.06287772953510284,
-0.07013025134801865,
0.06129353120923042,
-0.02393164299428463,
-0.08127736300230026,
0.09521230310201645,
-0.030652696266770363,
-0.011874021030962467,
0.06797946989536285,
0.02519136480987072,
0.054776888340711594,
-0.02174171432852745,
0.058210860937833786,
-0.17307010293006897,
0.016552884131669998,
0.08375038951635361,
-0.035569000989198685,
0.05614449083805084,
0.023907361552119255,
-0.036239031702280045,
0.10713044553995132,
-0.13422463834285736,
0.03838500380516052,
0.040171097964048386,
-0.05756452679634094,
-0.09581411629915237,
-0.09226278960704803,
0.014727585017681122,
0.011243402026593685,
0.07540318369865417,
-0.008565238676965237,
0.00231007463298738,
-0.030782166868448257,
0.051073018461465836,
0.16227807104587555,
-0.1364327073097229,
-0.02521209418773651,
0.05127548426389694,
-0.05330725759267807,
0.06600608676671982,
-0.028623634949326515,
-0.05467960610985756,
0.00428217276930809,
0.017703356221318245,
0.1445833146572113,
0.017145348712801933,
0.0046469904482364655,
0.021585946902632713,
-0.09498085081577301,
-0.06159023195505142,
0.06576090306043625,
-0.0024347652215510607,
-0.08007432520389557,
-0.16258938610553741,
-0.03894985839724541,
-0.005455297417938709,
0.055894363671541214,
-0.046226076781749725,
0.048428844660520554,
0.034589849412441254,
0.04252283275127411,
-0.10730138421058655,
-0.10061655938625336,
-0.053091999143362045,
-0.0001481424696976319,
0.013164469040930271,
0.0476866140961647,
0.008599135093390942,
0.026231229305267334,
0.09589487314224243,
-0.10522395372390747,
-0.05666198208928108,
-0.02822028659284115,
-0.0349917858839035,
-0.08420687168836594,
0.01030621025711298,
-0.0420391783118248,
-0.12411972880363464,
-0.05067373067140579,
0.06414132565259933,
-0.023293232545256615,
0.011462046764791012,
0.005658048205077648,
0.006115522235631943,
-0.007268310058861971,
0.12499283999204636,
-0.042068853974342346,
0.037893880158662796,
0.0518900491297245,
0.01793668419122696,
-0.02120572328567505,
-0.015079151839017868,
-0.07864353805780411,
-0.08301495760679245,
-0.0020230342634022236,
0.026738258078694344,
-0.06624467670917511,
0.03973979875445366,
-0.06486649811267853,
-0.022296080365777016,
0.012095530517399311,
-0.1257755160331726,
0.009483053348958492,
-0.005792930722236633,
-0.05518943816423416,
0.0415058471262455,
0.12901464104652405,
-0.031700246036052704,
-0.09570834040641785,
0.030983654782176018,
-0.01378741953521967,
-0.012850625440478325,
-0.12455376982688904,
-0.057162318378686905,
0.03299941122531891,
-0.12237462401390076,
-0.06193007528781891,
-0.08344267308712006,
-0.22206951677799225,
-0.061829861253499985,
0.03361504524946213,
-0.006781248841434717,
-0.009719053283333778,
-0.07578232139348984,
-0.029967496171593666,
-0.036189280450344086,
0.0022807989735156298,
-0.016483338549733162,
0.014370213262736797,
0.04070780426263809,
-0.005495515186339617,
0.038362737745046616,
0.02780071645975113,
0.022371254861354828,
-0.030514664947986603,
-0.0020426318515092134,
-0.14221182465553284,
0.17919433116912842,
0.02671433985233307,
-0.01763911359012127,
-0.08339950442314148,
0.006733995862305164,
-0.053871747106313705,
0.017290784046053886,
0.04074652120471001,
0.12582612037658691,
-0.2610965371131897,
-0.027601340785622597,
0.13974884152412415,
-0.06084061414003372,
-0.029038093984127045,
0.08874516189098358,
-0.10329899191856384,
0.09905807673931122,
0.15311695635318756,
0.0575568750500679,
0.12573355436325073,
-0.12926538288593292,
-0.1008663922548294,
-0.058823361992836,
-0.08810775727033615,
0.029850728809833527,
0.0982690304517746,
-0.023712990805506706,
0.042793501168489456,
0.038307227194309235,
-0.0229727104306221,
-0.014309069141745567,
-0.007049859035760164,
-0.04879221320152283,
-0.023347003385424614,
-0.0392577201128006,
0.003042686963453889,
-0.032571204006671906,
0.003164571011438966,
0.03852085396647453,
-0.10740455240011215,
0.10039163380861282,
0.0963214561343193,
-0.10690751671791077,
0.035266678780317307,
-0.05465041846036911,
-0.040910858660936356,
-0.05137123540043831,
0.004282406531274319,
-0.13805969059467316,
-0.009933034889400005,
0.03827369958162308,
0.0013413468841463327,
0.13596400618553162,
-0.06834333389997482,
0.07501418143510818,
0.04362868890166283,
-0.04112826660275459,
0.01285022497177124,
-0.0978875681757927,
-0.054514508694410324,
-0.07117776572704315,
-0.09682496637105942,
-0.09024973958730698,
-0.03055032156407833,
0.07337793707847595,
-0.1322779655456543,
0.056551720947027206,
0.04114134982228279,
0.03885391354560852,
0.009417138993740082,
-0.0621877945959568,
0.029051264747977257,
-0.006920856889337301,
0.014019726775586605,
-0.02262972667813301,
0.032627545297145844,
-0.004469858482480049,
-0.10274822264909744,
0.005346621386706829,
-0.10443361848592758,
-0.09567522257566452,
0.04225457087159157,
0.08152756094932556,
-0.049105461686849594,
-0.1817631870508194,
-0.02375069260597229,
-0.028569981455802917,
-0.11418207734823227,
-0.08240959048271179,
0.22059571743011475,
0.04278314486145973,
0.10964307934045792,
-0.09939129650592804,
-0.07271604984998703,
-0.030234375968575478,
0.03195721283555031,
0.010181412100791931,
0.07316625118255615,
0.041592586785554886,
0.030993714928627014,
0.016589047387242317,
0.0019336715340614319,
-0.037247009575366974,
0.1953846961259842,
0.046565841883420944,
-0.11172211915254593,
-0.03653126582503319,
0.019000323489308357,
0.024064648896455765,
-0.038079142570495605,
0.010393251664936543,
0.01966938190162182,
0.0038865164387971163,
0.003227517008781433,
0.044327475130558014,
-0.08938218653202057,
0.017770105972886086,
0.06136626750230789,
-0.05351429060101509,
-0.046366915106773376,
-0.034466877579689026,
0.02078746072947979,
0.06079640984535217,
0.0035700241569429636,
0.1250554919242859,
0.0220661461353302,
-0.05421248823404312,
-0.10735131055116653,
0.14518293738365173,
-0.0541941374540329,
-0.12268552929162979,
-0.14411436021327972,
0.07208065688610077,
-0.04367399960756302,
0.011003699153661728,
0.004448575899004936,
-0.10756636410951614,
-0.0021224869415163994,
-0.09306031465530396,
0.07356754690408707,
0.021718686446547508,
0.01302024070173502,
-0.07682748883962631,
0.04856623336672783,
0.013290726579725742,
-0.13686217367649078,
0.00684758136048913,
-0.0327555350959301,
-0.10050132870674133,
-0.008210349828004837,
0.001710339798592031,
0.08305399864912033,
0.0424005351960659,
-0.03287110477685928,
0.011932320892810822,
-0.02371818572282791,
0.16227057576179504,
-0.03178621456027031,
0.037745825946331024,
0.19212143123149872,
-0.03694475069642067,
0.09666915237903595,
0.02934272401034832,
0.02532823756337166,
-0.04021506384015083,
0.07207468152046204,
0.07142409682273865,
-0.06167261675000191,
-0.2274114340543747,
0.01894466206431389,
-0.0482044592499733,
-0.022818157449364662,
0.11577478796243668,
0.023646369576454163,
-0.04584428668022156,
0.04546000435948372,
-0.03162570670247078,
0.07758496701717377,
0.06412510573863983,
0.062075208872556686,
0.01286567933857441,
0.005057148169726133,
0.04572690278291702,
-0.05265166983008385,
-0.04952399432659149,
0.06503169983625412,
0.09641869366168976,
0.19552603363990784,
-0.10253029316663742,
0.1764814853668213,
0.07166121155023575,
0.06412025541067123,
0.008889065124094486,
0.09566348046064377,
-0.0709439367055893,
0.044980667531490326,
-0.05877716839313507,
-0.0973312109708786,
-0.06509935855865479,
0.07944925874471664,
-0.001015584566630423,
0.03260146453976631,
-0.06487339735031128,
0.028190746903419495,
0.056582167744636536,
0.14877060055732727,
-0.09873709827661514,
-0.19453935325145721,
-0.008128608576953411,
-0.0007745485636405647,
-0.028339378535747528,
0.009565632790327072,
-0.020596640184521675,
0.1288536787033081,
-0.0950976237654686,
0.012876609340310097,
-0.013377313502132893,
0.08373041450977325,
-0.0542648620903492,
-0.009954041801393032,
0.01842162385582924,
0.1447150856256485,
0.03378946706652641,
0.08352550119161606,
-0.21071814000606537,
0.01057420764118433,
0.020349889993667603,
0.11366992443799973,
-0.1003069058060646,
0.04600163921713829,
0.0036823772825300694,
-0.11875265836715698,
0.12311143428087234,
-0.01188333984464407,
0.04687828943133354,
-0.030529508367180824,
-0.08596093952655792,
0.014845146797597408,
0.050324998795986176,
-0.08752903342247009,
0.12725208699703217,
0.01722862385213375,
0.008913150057196617,
-0.0072136190719902515,
0.0952158272266388,
-0.07165934890508652,
-0.22184298932552338,
-0.003538908204063773,
-0.02405378967523575,
0.02038406953215599,
-0.01548212394118309,
-0.04926280304789543,
0.020934421569108963,
0.14626428484916687,
-0.12206967175006866,
-0.1001800149679184,
-0.10060898214578629,
0.036583732813596725,
0.11578751355409622,
-0.07811715453863144,
-0.020529204979538918,
0.05450356751680374,
0.1542869359254837,
-0.06749098747968674,
-0.08291573077440262,
-0.01973787508904934,
-0.0008921347907744348,
-0.11175385117530823,
-0.02345856837928295,
0.11757481843233109,
0.07974603027105331,
0.04750858247280121,
-0.00045881388359703124,
0.004583466798067093,
-0.01707172766327858,
-0.09506770968437195,
0.03472615033388138,
0.2295222282409668,
0.10894068330526352,
0.07597998529672623,
-0.1088743656873703,
0.027429254725575447,
-0.08463074266910553,
0.0004775788984261453,
0.12901833653450012,
0.18497639894485474,
-0.07792647182941437,
0.10824374109506607,
0.07610661536455154,
-0.13293582201004028,
-0.16234374046325684,
0.005360391456633806,
0.1278228461742401,
0.04671425372362137,
0.00784058403223753,
-0.28041738271713257,
0.07368865609169006,
0.10874257236719131,
0.03273236006498337,
-0.10997992753982544,
-0.25494274497032166,
-0.14720883965492249,
-0.019072197377681732,
0.052292339503765106,
0.020062396302819252,
-0.08587786555290222,
-0.010846559889614582,
-0.06142495572566986,
0.0787934958934784,
0.19992205500602722,
-0.20744101703166962,
0.07413383573293686,
0.0038231629878282547,
0.03256582096219063,
0.06092773750424385,
-0.03236525505781174,
0.032722942531108856,
0.0713878944516182,
0.07639448344707489,
-0.002893216209486127,
0.03994481638073921,
-0.00046452393871732056,
-0.07925942540168762,
0.17998817563056946,
-0.019984103739261627,
0.02973230555653572,
-0.15399017930030823,
-0.07630941271781921,
-0.04729040339589119,
0.0025039080064743757,
-0.02470395341515541,
-0.029080374166369438,
-0.01655580662190914,
0.007262376602739096,
0.11004167050123215,
-0.0068125370889902115,
-0.011116844601929188,
-0.04738002270460129,
0.0005802291561849415,
0.13175490498542786,
0.15047504007816315,
-0.0008658830774948001,
-0.07925640046596527,
0.012281275354325771,
0.0051735094748437405,
0.05353289842605591,
-0.15974339842796326,
0.038539398461580276,
0.10327325016260147,
-0.0021545097697526217,
0.12315704673528671,
0.06532473117113113,
-0.128742977976799,
-0.009972628206014633,
0.060258544981479645,
-0.1036750078201294,
-0.054249122738838196,
-0.03877519816160202,
-0.002463218057528138,
-0.15125183761119843,
-0.025192074477672577,
0.09885375201702118,
-0.09778792411088943,
-0.028743108734488487,
0.001173019758425653,
0.019898604601621628,
-0.015109413303434849,
0.1325758695602417,
0.06307540833950043,
0.012279844842851162,
-0.05784538388252258,
0.1537102907896042,
0.06183788552880287,
-0.08419834822416306,
0.05527297407388687,
0.13402831554412842,
-0.0506051741540432,
-0.009418655186891556,
-0.020904110744595528,
0.09827222675085068,
-0.03364477679133415,
0.0003527620865497738,
-0.029068730771541595,
-0.009878160431981087,
0.09242134541273117,
0.07341558486223221,
0.029310058802366257,
0.060151100158691406,
-0.07002425938844681,
0.029068339616060257,
-0.11319431662559509,
0.07153324037790298,
0.03842126578092575,
-0.0192111823707819,
-0.07576844096183777,
0.1606443077325821,
0.004326134920120239,
0.029970327392220497,
0.0012877946719527245,
-0.014679158106446266,
-0.06802304089069366,
0.016474077478051186,
0.045510537922382355,
0.046205420047044754,
-0.02390255033969879,
0.009478477761149406,
-0.022020624950528145,
0.05576271191239357,
0.04868758097290993,
0.03968101367354393,
-0.03497939184308052,
-0.05168592929840088,
-0.010551659390330315,
0.06714333593845367,
-0.11739090830087662,
-0.00857487041503191,
0.03397372364997864,
-0.08200263977050781,
0.09096130728721619,
0.03312759846448898,
-0.04688083007931709,
0.016221828758716583,
-0.07743193954229355,
-0.0040794359520077705,
0.02813979797065258,
0.018316516652703285,
-0.051517706364393234,
-0.0836285874247551,
0.03616762533783913,
0.001181732164695859,
-0.025191547349095345,
-0.018695516511797905,
0.06570938229560852,
-0.1075267344713211,
0.013593337498605251,
0.07172468304634094,
0.01413307711482048,
-0.11534418165683746,
0.05089014396071434,
0.03500873222947121,
0.0789099782705307,
0.12294326722621918,
-0.028093283995985985,
0.03561866283416748,
-0.16765642166137695,
-0.0003736289800144732,
0.024555036798119545,
0.02422862872481346,
-0.09832573682069778,
-0.08756036311388016,
0.05007452890276909,
-0.008586892858147621,
0.10710806399583817,
-0.010180273093283176,
-0.0292340237647295,
-0.01593063212931156,
-0.04989103600382805,
-0.009576614014804363,
0.009075935930013657,
0.07136937230825424,
-0.018850767984986305,
-0.007497407495975494,
0.07343891263008118,
0.029436133801937103,
0.0006163144134916365,
0.1017649695277214,
0.07931020855903625,
0.09296651184558868,
0.12709783017635345,
0.030271006748080254,
0.06623701751232147,
0.040046364068984985,
-0.046030569821596146,
0.07653014361858368,
0.033792153000831604,
-0.04886932298541069,
-0.09486928582191467,
0.05379302799701691,
0.12446168065071106,
-0.10435671359300613,
0.12417110055685043,
0.053325992077589035,
-0.06780726462602615,
-0.1064286082983017,
-0.12109553068876266,
-0.042348314076662064,
-0.01011160109192133,
0.007294335402548313,
-0.13185101747512817,
0.059975773096084595,
0.030606471002101898,
-0.005539316218346357,
0.018471436575055122,
0.08698944747447968,
-0.07136175036430359,
-0.08395932614803314,
0.0350319929420948,
-0.03922106698155403,
0.04010722413659096,
0.05483725294470787,
0.040092360228300095,
0.09698498249053955,
0.015147353522479534,
0.0849439799785614,
0.05997401848435402,
0.11767641454935074,
-0.028499657288193703,
-0.06939934939146042,
-0.04519304633140564,
-0.03719233721494675,
-0.006554602645337582,
0.06665753573179245,
0.1498579978942871,
0.042963020503520966,
-0.10959528386592865,
-0.01451630238443613,
0.18928220868110657,
-0.04594476893544197,
-0.043179985135793686,
-0.14242158830165863,
0.0662069246172905,
0.10157564282417297,
-0.0007599247619509697,
-0.018509482964873314,
-0.07091950625181198,
0.0320105254650116,
0.14648942649364471,
0.12681838870048523,
-0.019098598510026932,
0.0040329741314053535,
-0.008928928524255753,
0.0144705381244421,
0.0034907490480691195,
0.1178276464343071,
0.049774717539548874,
0.18379251658916473,
-0.054675646126270294,
0.1431160271167755,
-0.0007526077679358423,
0.010097174905240536,
-0.026996856555342674,
0.11898829787969589,
-0.09891663491725922,
0.03463410958647728,
-0.0304423701018095,
0.03167090564966202,
-0.029312603175640106,
-0.1893017292022705,
-0.05734039470553398,
-0.10176347941160202,
-0.06160677596926689,
-0.038569483906030655,
-0.037173446267843246,
-0.006452796049416065,
0.04250083491206169,
0.0073701185174286366,
0.011216320097446442,
0.12031669169664383,
0.017967969179153442,
-0.13476760685443878,
-0.02434798888862133,
0.046266548335552216,
0.04686592519283295,
0.1467926800251007,
0.008762456476688385,
0.08151336014270782,
0.06357528269290924,
-0.021889036521315575,
-0.1209568902850151,
0.0046287644654512405,
-0.03333991393446922,
-0.000872468575835228,
-0.00536316866055131,
0.12802737951278687,
-0.02714955247938633,
0.10331299155950546,
0.002458397764712572,
-0.043778810650110245,
0.0041411700658500195,
0.013625627383589745,
-0.03624999150633812,
-0.10086015611886978,
0.06154049560427666,
-0.08329595625400543,
0.11415936052799225,
0.1892002820968628,
-0.013767031952738762,
0.02694728784263134,
-0.08765944838523865,
0.006970899645239115,
-0.003896527225151658,
0.04099205881357193,
0.0013914310839027166,
-0.0729689970612526,
-0.014841399155557156,
0.07238240540027618,
0.004759985022246838,
-0.19083702564239502,
-0.08452774584293365,
0.049051474779844284,
-0.04848711937665939,
-0.03198184072971344,
0.06522741913795471,
0.04813646525144577,
0.0717252641916275,
-0.02268059365451336,
-0.1587902307510376,
-0.007814165204763412,
0.07584156095981598,
-0.13266073167324066,
-0.04363953322172165
] |
null | null |
transformers
|
# doc2query/msmarco-t5-small-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/msmarco-t5-small-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-small](https://huggingface.co/google/t5-v1_1-small) for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (query, passage) from the [MS MARCO Passage-Ranking dataset](https://github.com/microsoft/MSMARCO-Passage-Ranking).
|
{"language": "en", "license": "apache-2.0", "datasets": ["sentence-transformers/embedding-training-data"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/msmarco-t5-small-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:sentence-transformers/embedding-training-data",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/msmarco-t5-small-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-small for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset.
|
[
"# doc2query/msmarco-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-small for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/msmarco-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-small for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset."
] |
[
91,
261,
32,
114
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-sentence-transformers/embedding-training-data #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/msmarco-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-small for 31k training steps (about 4 epochs on the 500k training pairs from MS MARCO). For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (query, passage) from the MS MARCO Passage-Ranking dataset."
] |
[
-0.06467819958925247,
0.11153092980384827,
-0.006541788578033447,
0.09778304398059845,
0.11639998853206635,
-0.005906492937356234,
0.04068930447101593,
0.10955440998077393,
-0.08290452510118484,
0.041848763823509216,
0.07012176513671875,
0.052142802625894547,
0.0030101872980594635,
0.12236008793115616,
0.06996118277311325,
-0.19403618574142456,
-0.03839583694934845,
-0.04275217279791832,
0.10950713604688644,
0.08268692344427109,
0.0651383101940155,
-0.07889259606599808,
0.05642710626125336,
-0.027944568544626236,
-0.07689505070447922,
0.08827724307775497,
-0.02264806441962719,
-0.011403340846300125,
0.06458140164613724,
0.01862916350364685,
0.054200977087020874,
-0.01520363800227642,
0.06255903095006943,
-0.1717166304588318,
0.017381247133016586,
0.08508545160293579,
-0.03571389243006706,
0.055521201342344284,
0.021238429471850395,
-0.029170231893658638,
0.10953829437494278,
-0.13972273468971252,
0.032456714659929276,
0.04895792156457901,
-0.051617491990327835,
-0.10754195600748062,
-0.09148597717285156,
0.013360884040594101,
0.008320370689034462,
0.07834520190954208,
-0.012856083922088146,
0.010143715888261795,
-0.03891909867525101,
0.054499246180057526,
0.13942231237888336,
-0.1399429440498352,
-0.02773442678153515,
0.05897362902760506,
-0.03806029632687569,
0.069888174533844,
-0.040920089930295944,
-0.05541627109050751,
-0.002620844403281808,
0.023812774568796158,
0.1542733609676361,
0.012027846649289131,
0.014238669537007809,
0.027771499007940292,
-0.09318479150533676,
-0.06038086861371994,
0.04074869304895401,
-0.0005367713165469468,
-0.08171098679304123,
-0.17760711908340454,
-0.03971150517463684,
-0.016993630677461624,
0.05303409323096275,
-0.054649099707603455,
0.0504818856716156,
0.03559207171201706,
0.03686432167887688,
-0.10177171975374222,
-0.10222536325454712,
-0.050208888947963715,
-0.005883832462131977,
0.012526854872703552,
0.047201260924339294,
0.010417267680168152,
0.030163243412971497,
0.09471308439970016,
-0.0967474952340126,
-0.06550179421901703,
-0.03739580139517784,
-0.0407756082713604,
-0.08218865096569061,
0.011125719174742699,
-0.0391235277056694,
-0.10700669139623642,
-0.05455079674720764,
0.05291144922375679,
-0.02800365909934044,
0.01060917042195797,
0.0011329532135277987,
0.00978272408246994,
-0.006598866544663906,
0.12184605002403259,
-0.045923057943582535,
0.03728381544351578,
0.06044428423047066,
0.030331898480653763,
-0.024102499708533287,
-0.01068270392715931,
-0.07719289511442184,
-0.08937331289052963,
-0.0035705373156815767,
0.029362047091126442,
-0.0676804780960083,
0.03307154029607773,
-0.05919812247157097,
-0.029781265184283257,
0.016156449913978577,
-0.12712432444095612,
0.0006410834030248225,
-0.011114311404526234,
-0.06336034834384918,
0.061810288578271866,
0.11906950920820236,
-0.042144231498241425,
-0.09572231769561768,
0.023435231298208237,
-0.013880753889679909,
-0.013037454336881638,
-0.12638360261917114,
-0.06423502415418625,
0.033459488302469254,
-0.11584768444299698,
-0.06709236651659012,
-0.08293565362691879,
-0.2089356631040573,
-0.06796939671039581,
0.02904743328690529,
-0.007122920826077461,
-0.006864818278700113,
-0.07252094894647598,
-0.03481101617217064,
-0.037002045661211014,
-0.005859776865690947,
-0.011468973942101002,
0.016494061797857285,
0.04983273148536682,
0.004276755265891552,
0.037744030356407166,
0.040727242827415466,
0.025829171761870384,
-0.02897757478058338,
-0.005776663310825825,
-0.15478427708148956,
0.1732913702726364,
0.025900034233927727,
-0.017752034589648247,
-0.0840674638748169,
-0.006350843235850334,
-0.04167192056775093,
0.019291279837489128,
0.035325828939676285,
0.1191239133477211,
-0.26268890500068665,
-0.022779593244194984,
0.15528559684753418,
-0.05937272682785988,
-0.029386254027485847,
0.08427552133798599,
-0.10539696365594864,
0.10296296328306198,
0.16291145980358124,
0.04113932326436043,
0.1332225501537323,
-0.13438726961612701,
-0.1110900267958641,
-0.06104089319705963,
-0.09005288034677505,
0.021132124587893486,
0.10200592875480652,
-0.01158752478659153,
0.05448739603161812,
0.0401245541870594,
-0.016052138060331345,
-0.0065076351165771484,
-0.005171751603484154,
-0.048492781817913055,
-0.02074754610657692,
-0.04040851444005966,
0.0025044092908501625,
-0.022613393142819405,
0.001618396956473589,
0.037320327013731,
-0.11932579427957535,
0.08192814141511917,
0.0988711565732956,
-0.10483051091432571,
0.03022049367427826,
-0.05268796160817146,
-0.057036157697439194,
-0.041158903390169144,
0.00869707576930523,
-0.14420641958713531,
0.000402905629016459,
0.0324980728328228,
0.014411943964660168,
0.12497410178184509,
-0.06971161812543869,
0.07812345027923584,
0.04270409792661667,
-0.0484093576669693,
0.016535433009266853,
-0.09870832413434982,
-0.058780737221241,
-0.06687352806329727,
-0.08737596124410629,
-0.08817625790834427,
-0.03375549614429474,
0.06129692122340202,
-0.13078415393829346,
0.053961776196956635,
0.05209583789110184,
0.02701939456164837,
0.006909389514476061,
-0.05814804881811142,
0.026460688561201096,
-0.006488599348813295,
0.02010226622223854,
-0.020809156820178032,
0.03377113491296768,
-0.010413129813969135,
-0.10103661566972733,
0.014245463535189629,
-0.10694077610969543,
-0.08326505869626999,
0.04726608097553253,
0.08139598369598389,
-0.05795624852180481,
-0.19406422972679138,
-0.027795635163784027,
-0.02983716130256653,
-0.11232274770736694,
-0.08059840649366379,
0.21703480184078217,
0.0435640811920166,
0.11357923597097397,
-0.09834650158882141,
-0.0739016905426979,
-0.024155275896191597,
0.0494723916053772,
0.024779614061117172,
0.07166499644517899,
0.031528472900390625,
0.012224788777530193,
0.015274456702172756,
0.005449910182505846,
-0.03376425802707672,
0.1961638182401657,
0.039536379277706146,
-0.11919375509023666,
-0.04180518537759781,
0.023596404120326042,
0.02825547195971012,
-0.03176829218864441,
0.02608216553926468,
0.016921697184443474,
0.0072645931504666805,
0.006090641021728516,
0.04187285155057907,
-0.08802817761898041,
0.016609184443950653,
0.05482835695147514,
-0.055318091064691544,
-0.056319862604141235,
-0.031232530251145363,
0.030308742076158524,
0.06084591522812843,
0.009711046703159809,
0.12574967741966248,
0.02480974979698658,
-0.06018795073032379,
-0.10421156138181686,
0.14689603447914124,
-0.04807153716683388,
-0.12815499305725098,
-0.13560110330581665,
0.07226565480232239,
-0.03423463553190231,
0.004043638240545988,
0.0028094593435525894,
-0.1012507975101471,
-0.00599668500944972,
-0.08471079915761948,
0.07465451210737228,
0.010675168596208096,
0.005080937407910824,
-0.09872837364673615,
0.062141962349414825,
0.009922713972628117,
-0.13615162670612335,
0.005519791506230831,
-0.031095467507839203,
-0.11297224462032318,
-0.00482051819562912,
0.018631502985954285,
0.07141637802124023,
0.03412243723869324,
-0.0326395183801651,
0.010723484680056572,
-0.029976995661854744,
0.1470458060503006,
-0.036194536834955215,
0.03690788149833679,
0.1958128660917282,
-0.026690203696489334,
0.09805997461080551,
0.026516396552324295,
0.02598978392779827,
-0.036078598350286484,
0.07151979207992554,
0.06272833049297333,
-0.06381376832723618,
-0.2236315757036209,
0.017595108598470688,
-0.05266961455345154,
-0.022273961454629898,
0.11726650595664978,
0.018585359677672386,
-0.05269267037510872,
0.046372074633836746,
-0.041130512952804565,
0.07767065614461899,
0.06246192008256912,
0.05793541297316551,
0.02721559815108776,
0.00200900761410594,
0.0552501380443573,
-0.053641315549612045,
-0.05670604854822159,
0.07000760734081268,
0.0906854048371315,
0.20973822474479675,
-0.10651922971010208,
0.1674119383096695,
0.07377348840236664,
0.06982370465993881,
0.012120996601879597,
0.10118308663368225,
-0.06249205023050308,
0.03970088064670563,
-0.06013805419206619,
-0.10129769891500473,
-0.06426185369491577,
0.07392359524965286,
0.002872846322134137,
0.031925301998853683,
-0.0714130848646164,
0.01983988843858242,
0.048333559185266495,
0.14689384400844574,
-0.10441860556602478,
-0.19081732630729675,
-0.01766124553978443,
-0.005325290374457836,
-0.03741052374243736,
0.007018725387752056,
-0.011728198267519474,
0.14056992530822754,
-0.10331761091947556,
0.00770219461992383,
-0.023773493245244026,
0.08555538207292557,
-0.0550052709877491,
0.0010658206883817911,
0.01420524064451456,
0.14674235880374908,
0.045250218361616135,
0.08798287063837051,
-0.20971797406673431,
0.015349769964814186,
0.024705005809664726,
0.10745137184858322,
-0.10918335616588593,
0.04423515498638153,
0.02039651945233345,
-0.13415628671646118,
0.12610860168933868,
-0.016717785969376564,
0.04787863790988922,
-0.042017869651317596,
-0.08722485601902008,
0.015115317888557911,
0.058765314519405365,
-0.08578836172819138,
0.13150660693645477,
0.017100883647799492,
0.007550767157226801,
-0.008342987857758999,
0.08180838078260422,
-0.07103453576564789,
-0.21650132536888123,
-0.004612968768924475,
-0.02842872217297554,
0.0200984887778759,
-0.01862397976219654,
-0.05326394364237785,
0.03059285320341587,
0.15006446838378906,
-0.1273687481880188,
-0.09672204405069351,
-0.10179553180932999,
0.04120389372110367,
0.12076272070407867,
-0.08389866352081299,
-0.02094309590756893,
0.05643937364220619,
0.14040538668632507,
-0.06747933477163315,
-0.08550745248794556,
-0.010500438511371613,
0.002998529002070427,
-0.1004471555352211,
-0.01795213669538498,
0.11656052619218826,
0.08015299588441849,
0.049213480204343796,
-0.0020029135048389435,
0.006523204036056995,
-0.026540741324424744,
-0.09915931522846222,
0.03553614020347595,
0.23345723748207092,
0.12294402718544006,
0.07711520791053772,
-0.10780205577611923,
0.027398068457841873,
-0.0782357007265091,
0.007894279435276985,
0.1338537037372589,
0.18533457815647125,
-0.08187694847583771,
0.11505689471960068,
0.07402296364307404,
-0.11726966500282288,
-0.17971785366535187,
0.0023852468002587557,
0.12555521726608276,
0.0638289675116539,
0.018810557201504707,
-0.27924686670303345,
0.07490810751914978,
0.11226705461740494,
0.02803829498589039,
-0.08835656195878983,
-0.23962081968784332,
-0.13823451101779938,
-0.019075490534305573,
0.049221381545066833,
0.016740238294005394,
-0.08380050957202911,
-0.012300913222134113,
-0.06067226827144623,
0.09067925810813904,
0.191187784075737,
-0.21288937330245972,
0.08582216501235962,
-0.0015689193969592452,
0.03014327585697174,
0.06212648004293442,
-0.03688887879252434,
0.028368663042783737,
0.061150889843702316,
0.07174422591924667,
-0.0052908253856003284,
0.0404968224465847,
0.019176814705133438,
-0.08460435271263123,
0.17465445399284363,
-0.028418222442269325,
0.02973291650414467,
-0.16188660264015198,
-0.07789190858602524,
-0.05096137151122093,
-0.004858826752752066,
-0.023092828691005707,
-0.029071608558297157,
-0.01380922831594944,
0.015243376605212688,
0.10659856349229813,
-0.011622154153883457,
-0.011312398128211498,
-0.04779345542192459,
-0.003113949205726385,
0.1486557275056839,
0.1613752543926239,
-0.0033032784704118967,
-0.07889241725206375,
0.006765855476260185,
0.005176228005439043,
0.0544477254152298,
-0.1749352365732193,
0.0324036069214344,
0.10921795666217804,
-0.002166708232834935,
0.11288190633058548,
0.06898085027933121,
-0.12351658940315247,
-0.006931703072041273,
0.06276015937328339,
-0.0980764701962471,
-0.05454276129603386,
-0.052453797310590744,
-0.005461039487272501,
-0.16371497511863708,
-0.03304823115468025,
0.09366918355226517,
-0.09790219366550446,
-0.027335045859217644,
0.0028774465899914503,
0.01708039455115795,
-0.008642394095659256,
0.13410186767578125,
0.06397100538015366,
0.02141113579273224,
-0.0613480880856514,
0.13938426971435547,
0.06294699758291245,
-0.10280650109052658,
0.053706150501966476,
0.14946936070919037,
-0.05073074996471405,
-0.01572437770664692,
-0.02842717431485653,
0.08172855526208878,
-0.023524794727563858,
0.01340757217258215,
-0.029213184490799904,
-0.012138620018959045,
0.09516022354364395,
0.0680825263261795,
0.033320892602205276,
0.05865029618144035,
-0.07419271022081375,
0.035240575671195984,
-0.11730790883302689,
0.07724782824516296,
0.03362981602549553,
-0.0262297410517931,
-0.07366540282964706,
0.16941231489181519,
0.00454330537468195,
0.03315255790948868,
0.00006124077481217682,
-0.016691215336322784,
-0.07116768509149551,
0.016947025433182716,
0.03270917385816574,
0.051342763006687164,
-0.027025142684578896,
0.006607419811189175,
-0.029375188052654266,
0.06064547970890999,
0.046141427010297775,
0.04283437877893448,
-0.03499598056077957,
-0.05606162175536156,
-0.007922710850834846,
0.0695357620716095,
-0.11201264709234238,
-0.005730022210627794,
0.030407020822167397,
-0.08485327661037445,
0.0934889018535614,
0.027947980910539627,
-0.04864731803536415,
0.018934350460767746,
-0.060352589935064316,
-0.019243530929088593,
0.027981054037809372,
0.01517782174050808,
-0.044634509831666946,
-0.078959159553051,
0.045002806931734085,
0.0072740004397928715,
-0.020926199853420258,
-0.012646581046283245,
0.04933270066976547,
-0.11051521450281143,
0.007466246839612722,
0.06565756350755692,
0.021702932193875313,
-0.1212233304977417,
0.04744771495461464,
0.03914251551032066,
0.0809384286403656,
0.12062100321054459,
-0.02953960932791233,
0.047119397670030594,
-0.18465951085090637,
-0.001830801134929061,
0.016886267811059952,
0.019551502540707588,
-0.09349533915519714,
-0.09029741585254669,
0.054416533559560776,
-0.012584911659359932,
0.10469583421945572,
-0.017826132476329803,
-0.02222246676683426,
-0.007579202763736248,
-0.05579661950469017,
-0.007678577210754156,
0.00554553372785449,
0.07938163727521896,
-0.007274589966982603,
-0.013719819486141205,
0.08060924708843231,
0.034024927765131,
-0.0019800360314548016,
0.10262858867645264,
0.07893636077642441,
0.10407491028308868,
0.1324605941772461,
0.02960062399506569,
0.06944064795970917,
0.04172220826148987,
-0.044567033648490906,
0.09237813949584961,
0.04370787739753723,
-0.058196477591991425,
-0.09916828572750092,
0.04996451735496521,
0.12407293915748596,
-0.10094205290079117,
0.1322096288204193,
0.049428537487983704,
-0.06553376466035843,
-0.10628592222929001,
-0.11974717676639557,
-0.044083669781684875,
-0.01222311146557331,
0.01055454183369875,
-0.12694160640239716,
0.06657592952251434,
0.0474451519548893,
-0.007538874167948961,
0.015181743539869785,
0.08487087488174438,
-0.07264300435781479,
-0.08547062426805496,
0.05329175665974617,
-0.03741026669740677,
0.041827473789453506,
0.06576082110404968,
0.04026982933282852,
0.09780578315258026,
0.00796070508658886,
0.0942176803946495,
0.053688306361436844,
0.11239060759544373,
-0.03253786265850067,
-0.08660691976547241,
-0.04290832206606865,
-0.03586051985621452,
0.0023229094222187996,
0.07270175963640213,
0.1614367663860321,
0.03970283269882202,
-0.1175883337855339,
-0.016787389293313026,
0.19287659227848053,
-0.05349421501159668,
-0.05166158080101013,
-0.13682614266872406,
0.07496900111436844,
0.11919274926185608,
-0.00784993078559637,
-0.01660512574017048,
-0.06829269230365753,
0.038006510585546494,
0.14859551191329956,
0.1355626881122589,
-0.01876063644886017,
0.006209610495716333,
-0.007197054568678141,
0.010555646382272243,
0.00886490661650896,
0.12429091334342957,
0.0612935945391655,
0.1953941285610199,
-0.05390934646129608,
0.13555428385734558,
0.005990729667246342,
0.0010776212438941002,
-0.02787693776190281,
0.10832562297582626,
-0.09172326326370239,
0.040683064609766006,
-0.032669052481651306,
0.03698758780956268,
-0.016252750530838966,
-0.15952655673027039,
-0.05757766589522362,
-0.11102839559316635,
-0.06304684281349182,
-0.045694511383771896,
-0.03595663979649544,
-0.0035088546574115753,
0.036676935851573944,
0.00894435215741396,
0.021295351907610893,
0.11558172106742859,
0.014875759370625019,
-0.1397990733385086,
-0.016430426388978958,
0.04581211507320404,
0.04700981453061104,
0.14084458351135254,
0.008656986989080906,
0.0811387300491333,
0.06754075735807419,
-0.022527772933244705,
-0.12162362784147263,
0.007558518089354038,
-0.04001537337899208,
-0.0229128897190094,
-0.004947819281369448,
0.14792375266551971,
-0.03301766887307167,
0.0994984582066536,
0.0021076907869428396,
-0.05260805785655975,
-0.0015519731678068638,
0.003473903052508831,
-0.03491149842739105,
-0.10290785878896713,
0.06497832387685776,
-0.07370885461568832,
0.115664042532444,
0.19672898948192596,
-0.011401119641959667,
0.03003416396677494,
-0.09326565265655518,
0.0033727374393492937,
-0.0023752048145979643,
0.042156003415584564,
0.00005757690450991504,
-0.06488171964883804,
-0.027117175981402397,
0.07567743211984634,
0.00014208085485734046,
-0.2117292284965515,
-0.08539529889822006,
0.0529172383248806,
-0.04603780806064606,
-0.03804413601756096,
0.07217947393655777,
0.04397803917527199,
0.07224053144454956,
-0.01955258660018444,
-0.16995415091514587,
-0.009539213962852955,
0.07643598318099976,
-0.13964726030826569,
-0.04239516705274582
] |
null | null |
transformers
|
# doc2query/reddit-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/reddit-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 533k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, body) from Reddit.
|
{"language": "en", "license": "apache-2.0", "datasets": ["datasets/sentence-transformers/reddit-title-body"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/reddit-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/reddit-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 533k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, body) from Reddit.
|
[
"# doc2query/reddit-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 533k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, body) from Reddit."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/reddit-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 533k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, body) from Reddit."
] |
[
75,
259,
32,
85
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/reddit-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-base for 533k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, body) from Reddit."
] |
[
-0.07314171642065048,
0.12481638789176941,
-0.005682343151420355,
0.08135806024074554,
0.11457894742488861,
0.01918739639222622,
0.0682138130068779,
0.12272875756025314,
-0.07694874703884125,
0.019201749935746193,
0.1102607250213623,
0.044769808650016785,
-0.0036791497841477394,
0.0955292135477066,
0.03217630460858345,
-0.19426505267620087,
-0.038544781506061554,
-0.021762555465102196,
0.05827191472053528,
0.12603765726089478,
0.07402273267507553,
-0.04470783472061157,
0.05289190262556076,
-0.024973571300506592,
-0.15963104367256165,
0.055596642196178436,
-0.0303761288523674,
-0.013035350479185581,
0.07091859728097916,
0.05202150344848633,
0.08059192448854446,
-0.01849462278187275,
0.043838825076818466,
-0.12753728032112122,
0.02123410999774933,
0.05824613943696022,
-0.013671920634806156,
0.09466448426246643,
0.02277824468910694,
-0.017598183825612068,
0.2040761560201645,
-0.08920636773109436,
0.0271734818816185,
0.06077160686254501,
-0.08643028140068054,
-0.10099974274635315,
-0.05689230188727379,
0.08148131519556046,
0.03856028616428375,
0.09912985563278198,
-0.008436125703155994,
0.042985159903764725,
-0.05149706080555916,
0.051588479429483414,
0.1709042489528656,
-0.1418590247631073,
-0.03617731109261513,
0.03515522927045822,
0.019592471420764923,
0.07419272512197495,
0.004329813178628683,
-0.045916009694337845,
-0.02216525189578533,
0.025906860828399658,
0.12118460237979889,
0.007037751376628876,
-0.023805255070328712,
0.010832514613866806,
-0.10400956869125366,
-0.06666842103004456,
0.1259457916021347,
-0.01638888753950596,
-0.05959102138876915,
-0.11016147583723068,
-0.02070826292037964,
0.10023318231105804,
0.02168305031955242,
-0.03183553367853165,
0.0002913996868301183,
0.03333022817969322,
-0.02528071217238903,
-0.1241954043507576,
-0.08749464154243469,
-0.06826028227806091,
0.07405991107225418,
-0.013447710312902927,
-0.004489995073527098,
0.014016306959092617,
-0.056535594165325165,
0.10745871812105179,
-0.0743565708398819,
-0.020318031311035156,
-0.02423824928700924,
-0.1059405505657196,
-0.006369234062731266,
-0.001978594111278653,
-0.06594229489564896,
-0.12636704742908478,
0.004959451965987682,
0.006870771758258343,
0.017483439296483994,
-0.013051769696176052,
0.007966812700033188,
0.03508064150810242,
0.0037222737446427345,
0.11908311396837234,
-0.04061386361718178,
-0.019902663305401802,
0.07968245446681976,
0.022677285596728325,
-0.055131323635578156,
0.016954991966485977,
-0.07467339187860489,
-0.10731133818626404,
0.0763736441731453,
0.031991470605134964,
-0.01033270638436079,
0.0452997125685215,
-0.0694463774561882,
-0.026624292135238647,
0.05262545123696327,
-0.09918101876974106,
0.018000727519392967,
0.04757629334926605,
-0.09602002054452896,
-0.06701037287712097,
0.17420488595962524,
-0.07368610799312592,
-0.11546897143125534,
-0.013851546682417393,
-0.032423436641693115,
0.0022931068670004606,
-0.13683059811592102,
-0.13700903952121735,
0.01990448869764805,
-0.1461946964263916,
-0.02996341697871685,
-0.05920777842402458,
-0.2655572295188904,
0.005179245490580797,
0.06920558959245682,
-0.0351424366235733,
-0.004092622082680464,
-0.08024854212999344,
0.002186308614909649,
-0.057537563145160675,
0.027686767280101776,
-0.006166556384414434,
0.03253313899040222,
0.04222914204001427,
-0.020424403250217438,
0.06457247585058212,
-0.04170071706175804,
0.02331387624144554,
-0.007479009218513966,
0.026388252153992653,
-0.13566841185092926,
0.20684292912483215,
0.022502649575471878,
-0.020293964073061943,
-0.11365260183811188,
-0.019115911796689034,
-0.06283143162727356,
0.0013582120882347226,
0.08012508600950241,
0.15573523938655853,
-0.12751130759716034,
-0.05716923624277115,
0.18013781309127808,
-0.04150395095348358,
-0.04246537759900093,
0.09975479543209076,
-0.0960974469780922,
0.15703046321868896,
0.11875155568122864,
-0.0008204665500670671,
0.09664394706487656,
-0.09608472883701324,
-0.04399304464459419,
-0.001226593623869121,
-0.1250072568655014,
-0.04724323749542236,
0.08211436867713928,
-0.032346539199352264,
-0.0009954272536560893,
0.010583754628896713,
-0.03537946566939354,
0.020636802539229393,
-0.04375307634472847,
-0.03406142070889473,
-0.025660045444965363,
-0.04682629927992821,
0.032132916152477264,
-0.024820011109113693,
0.01685868203639984,
0.05139603093266487,
-0.08352510631084442,
0.15870890021324158,
0.06084342673420906,
-0.0943780317902565,
0.0036938043776899576,
-0.05359272286295891,
0.01959763653576374,
-0.1398894339799881,
0.03465723991394043,
-0.18265020847320557,
0.031325988471508026,
0.008729579858481884,
0.043399497866630554,
0.1484546810388565,
-0.0545826181769371,
0.06583735346794128,
0.041269030421972275,
-0.02914891205728054,
0.019132576882839203,
-0.08490598201751709,
-0.007628990802913904,
-0.09299399703741074,
-0.12237563729286194,
-0.10534194111824036,
-0.030811209231615067,
0.08130525052547455,
-0.19925759732723236,
0.05843912065029144,
0.01072204951196909,
-0.0314774364233017,
0.022075394168496132,
-0.03140918165445328,
0.06169392913579941,
-0.02280060388147831,
0.017566902562975883,
-0.03745086118578911,
0.0776454508304596,
0.03604649379849434,
-0.11578604578971863,
-0.04245360195636749,
-0.03380880504846573,
-0.16104136407375336,
0.04903419315814972,
0.011551281437277794,
-0.07373352348804474,
-0.10930545628070831,
-0.052167024463415146,
-0.01905224658548832,
-0.09581425786018372,
-0.07497652620077133,
0.15044644474983215,
0.043515730649232864,
0.1299024522304535,
-0.11063084751367569,
-0.06963752210140228,
0.01477874256670475,
-0.028708236292004585,
0.015291896648705006,
0.08611324429512024,
0.03855150565505028,
-0.09247121959924698,
0.01574610359966755,
0.036522481590509415,
-0.03289676457643509,
0.22540830075740814,
0.06249912828207016,
-0.1083127111196518,
-0.04140102118253708,
0.03421555086970329,
0.015206555835902691,
0.0392654612660408,
-0.08475848287343979,
-0.021909605711698532,
0.0016963430680334568,
-0.008953563868999481,
0.05555713176727295,
-0.10571997612714767,
0.038556598126888275,
0.054294876754283905,
-0.03658968210220337,
-0.08672530949115753,
-0.051324088126420975,
-0.021357588469982147,
0.05735495314002037,
0.0545848049223423,
0.06792029738426208,
0.0655372217297554,
-0.04080286622047424,
-0.15765255689620972,
0.18570300936698914,
-0.08167128264904022,
-0.10018827766180038,
-0.11921461671590805,
0.03536485508084297,
-0.000870281015522778,
0.004947258625179529,
-0.006942237727344036,
-0.06371409446001053,
0.011735362932085991,
-0.06797431409358978,
0.052649036049842834,
0.011140796355903149,
-0.009945807047188282,
-0.03863875940442085,
0.06856142729520798,
0.03447362408041954,
-0.13823769986629486,
0.023423532024025917,
-0.02774728648364544,
-0.051352594047784805,
0.03525326028466225,
-0.10663367062807083,
0.05538954585790634,
0.055506765842437744,
-0.06817489862442017,
0.04803500324487686,
-0.00668066693469882,
0.25010740756988525,
-0.0053148106671869755,
0.03724263235926628,
0.09488484263420105,
-0.062363363802433014,
0.05134795233607292,
0.01819639466702938,
0.03523040562868118,
-0.012912137433886528,
0.05121129751205444,
-0.010779804550111294,
-0.04388657957315445,
-0.23454491794109344,
-0.01179087907075882,
-0.04222415015101433,
-0.023193301633000374,
0.05770931765437126,
0.010374503210186958,
-0.040855325758457184,
0.047055941075086594,
0.026660054922103882,
0.07997004687786102,
0.07770752161741257,
0.09084149450063705,
-0.008591958321630955,
0.00610761484131217,
0.020759442821145058,
-0.05131068825721741,
0.015839608386158943,
0.03949983790516853,
0.06804061681032181,
0.2791772782802582,
-0.03069276176393032,
0.1618354171514511,
0.07280842959880829,
0.02200125902891159,
0.04245796427130699,
0.11303053051233292,
-0.0890757218003273,
0.05403614789247513,
-0.05099307373166084,
-0.07132695615291595,
-0.051959458738565445,
0.06801008433103561,
-0.05358373001217842,
0.03807323798537254,
-0.05725773796439171,
0.05162224546074867,
0.049789752811193466,
0.17543040215969086,
-0.026674984022974968,
-0.21747681498527527,
-0.05291730538010597,
-0.041974958032369614,
0.0010248797480016947,
-0.018522968515753746,
-0.036232512444257736,
0.07955294847488403,
-0.08593904972076416,
-0.07191088795661926,
-0.04321529343724251,
0.10571163892745972,
-0.03020188957452774,
0.010522858239710331,
-0.020601212978363037,
0.11135251820087433,
-0.016119128093123436,
0.08848775178194046,
-0.23477812111377716,
0.11455430835485458,
0.04305097460746765,
0.11938289552927017,
-0.07540947943925858,
0.0358053483068943,
0.01120530255138874,
-0.06556719541549683,
0.12879939377307892,
-0.0016979702049866319,
0.10993840545415878,
-0.05011860653758049,
-0.05804092809557915,
0.03416793420910835,
0.043514713644981384,
-0.008398345671594143,
0.12218081206083298,
0.025584381073713303,
-0.007792538497596979,
0.03369296342134476,
0.09851640462875366,
-0.1156390830874443,
-0.19027096033096313,
-0.004663048777729273,
-0.08466751128435135,
0.028253808617591858,
-0.008023405447602272,
-0.010074688121676445,
0.11458714306354523,
0.1455691158771515,
-0.19009678065776825,
-0.12415484338998795,
-0.0923270732164383,
0.1145755723118782,
0.0665360614657402,
-0.06128391996026039,
0.021346136927604675,
0.05144259333610535,
0.11501733213663101,
-0.04918469488620758,
-0.07391100376844406,
0.043299175798892975,
-0.08418352156877518,
-0.10565746575593948,
-0.03633081912994385,
0.04745975881814957,
0.12344397604465485,
0.06942261010408401,
0.04441770911216736,
-0.0009632106521166861,
-0.052106618881225586,
-0.06928272545337677,
0.008669049479067326,
0.11520775407552719,
0.08931204676628113,
0.04613424092531204,
-0.1352566033601761,
-0.06448063254356384,
-0.08974701911211014,
-0.009179758839309216,
0.15055853128433228,
0.14750084280967712,
-0.08421586453914642,
0.06618606299161911,
0.145072340965271,
-0.12916234135627747,
-0.13618680834770203,
0.004758684430271387,
0.10038500279188156,
0.0586538165807724,
-0.030901752412319183,
-0.30276429653167725,
0.03581814840435982,
0.05600136145949364,
0.021569473668932915,
-0.08822432905435562,
-0.2842351198196411,
-0.11903248727321625,
0.07502133399248123,
0.07629052549600601,
0.04699879512190819,
-0.09583711624145508,
-0.02481217309832573,
-0.02915693074464798,
0.06066177412867546,
0.21439264714717865,
-0.19755049049854279,
0.05288851261138916,
0.015004840679466724,
0.022911183536052704,
0.045093320310115814,
-0.02921442687511444,
-0.004633286967873573,
0.0074534290470182896,
0.038174811750650406,
0.007153816055506468,
-0.01237053144723177,
0.053990695625543594,
-0.06209159269928932,
0.1617100089788437,
-0.009960642084479332,
0.062391433864831924,
-0.15326625108718872,
-0.06488509476184845,
-0.09094161540269852,
-0.0017751178238540888,
-0.036902111023664474,
-0.05040114000439644,
0.009450756013393402,
0.025691164657473564,
0.07438711076974869,
0.02812577597796917,
-0.05125449597835541,
-0.06878585368394852,
0.017163103446364403,
0.16487371921539307,
0.191043883562088,
0.01786820776760578,
-0.02925335243344307,
0.02679836004972458,
-0.0019285365706309676,
0.08538713306188583,
-0.14065727591514587,
0.0006456964765675366,
0.057628970593214035,
0.038680240511894226,
0.1757570058107376,
0.06414774060249329,
-0.1250881850719452,
0.01913713477551937,
0.04608255997300148,
-0.12739062309265137,
-0.04795832559466362,
-0.029483506456017494,
0.03532702103257179,
-0.14968512952327728,
-0.034894946962594986,
0.09650032967329025,
-0.11891795694828033,
-0.04314974322915077,
0.002685781568288803,
0.04472648352384567,
-0.0617411769926548,
0.11859510838985443,
0.10166307538747787,
0.03530915081501007,
-0.029057804495096207,
0.09146331995725632,
0.05590236559510231,
-0.047887012362480164,
0.07035846263170242,
0.1410624235868454,
-0.0845395028591156,
-0.01333386916667223,
-0.05216221138834953,
0.15438565611839294,
-0.03772086277604103,
-0.0395585261285305,
-0.024792319163680077,
-0.015667740255594254,
0.08120066672563553,
0.1330236941576004,
0.03229628875851631,
0.06340701878070831,
-0.07404220849275589,
0.014926529489457607,
-0.14369001984596252,
0.10556167364120483,
0.03992130607366562,
-0.05996488779783249,
-0.037102241069078445,
0.06704395264387131,
-0.0012360186083242297,
0.009858195669949055,
-0.017866037786006927,
-0.023250777274370193,
-0.08663575351238251,
0.007544739171862602,
-0.08579093962907791,
0.03603644296526909,
-0.04160328581929207,
-0.0023937630467116833,
-0.04282558709383011,
0.058169860392808914,
0.060428570955991745,
0.03371625021100044,
-0.04291417822241783,
-0.02620556578040123,
-0.002703558886423707,
0.06552579998970032,
-0.12568169832229614,
-0.005247193854302168,
0.02012680470943451,
-0.0770193412899971,
0.12217970192432404,
0.05415375158190727,
-0.04865333437919617,
-0.011907585896551609,
-0.06641487032175064,
-0.0021149583626538515,
-0.00027392207994125783,
-0.032381221652030945,
-0.00710550369694829,
-0.11801889538764954,
0.023589855059981346,
-0.001306484336964786,
-0.002832855097949505,
-0.018691185861825943,
0.12908148765563965,
-0.07731151580810547,
0.059392642229795456,
0.04182414710521698,
0.02818634733557701,
-0.1057739406824112,
0.10325118899345398,
0.00665450282394886,
0.08073394745588303,
0.09458091109991074,
-0.053295135498046875,
0.03909976780414581,
-0.15684258937835693,
-0.010967119596898556,
0.03152354806661606,
0.034049998968839645,
-0.12392500787973404,
-0.031018026173114777,
0.041341181844472885,
-0.03289689123630524,
0.09453275799751282,
-0.0000056075318752846215,
-0.015546424314379692,
-0.015804624184966087,
-0.040340606123209,
-0.06618797779083252,
-0.0127037288621068,
0.08914363384246826,
-0.03540628403425217,
-0.004833803046494722,
0.024112118408083916,
0.0680108368396759,
-0.002075942698866129,
0.05230071395635605,
0.13851980865001678,
0.1097809299826622,
0.11200568825006485,
0.051297035068273544,
0.07040062546730042,
0.03577717766165733,
-0.01463005505502224,
0.040541280061006546,
0.09531175345182419,
0.037462249398231506,
-0.08678106218576431,
0.1110406368970871,
0.16973158717155457,
-0.09787949174642563,
0.08060453832149506,
0.04415567219257355,
-0.08839641511440277,
-0.09020545333623886,
-0.10297457128763199,
-0.00842722412198782,
-0.03748420998454094,
-0.01086463313549757,
-0.13973242044448853,
0.024772081524133682,
-0.05204954743385315,
0.03224783390760422,
-0.029552510008215904,
0.10122306644916534,
0.08103986829519272,
-0.06953081488609314,
0.07950076460838318,
-0.05065768212080002,
0.016486231237649918,
-0.011293557472527027,
0.057873789221048355,
0.08397818356752396,
0.018154582008719444,
0.057657331228256226,
0.07472062855958939,
0.06981959193944931,
-0.032760798931121826,
-0.03609249368309975,
-0.08263970166444778,
-0.05462507903575897,
0.0327690951526165,
0.050769492983818054,
0.15787184238433838,
0.0010059848427772522,
-0.12717586755752563,
-0.00883677788078785,
0.13221952319145203,
-0.037445735186338425,
-0.06549535691738129,
-0.12091675400733948,
0.1423710733652115,
0.012971064075827599,
-0.022037899121642113,
-0.03186333552002907,
-0.08146114647388458,
0.038131728768348694,
0.16433218121528625,
0.18109017610549927,
-0.047147102653980255,
-0.01748587004840374,
-0.0010204483987763524,
0.008087052032351494,
-0.005879661999642849,
0.11104251444339752,
-0.0072416882030665874,
0.2378036230802536,
-0.07281319797039032,
0.1340140402317047,
-0.04231642931699753,
-0.004233259707689285,
-0.0191437266767025,
0.03791772946715355,
0.009747452102601528,
-0.016046874225139618,
-0.014592617750167847,
0.06634899973869324,
-0.10445652157068253,
-0.1877320557832718,
-0.11668137460947037,
-0.07267068326473236,
-0.05194389075040817,
-0.05815669521689415,
-0.06111045554280281,
0.0011361505603417754,
0.08802594989538193,
-0.0009555082069709897,
0.01833178661763668,
0.12612074613571167,
-0.015359259210526943,
-0.11624165624380112,
-0.02701123245060444,
0.06250440329313278,
0.0005530209746211767,
0.1107429787516594,
-0.002764833625406027,
0.09259811788797379,
0.0741649940609932,
-0.02758912928402424,
-0.15316925942897797,
0.013817362487316132,
-0.012800590135157108,
0.07154081016778946,
0.028712859377264977,
0.10823842138051987,
-0.004601637367159128,
0.07475442439317703,
0.00441869767382741,
-0.019154362380504608,
0.013183544389903545,
-0.03686909005045891,
-0.07472840696573257,
-0.11826305836439133,
0.040752243250608444,
-0.03201339393854141,
0.11756829917430878,
0.1892625391483307,
-0.06664077937602997,
0.017779583111405373,
-0.06667298078536987,
0.024049032479524612,
0.03066244348883629,
0.07077790796756744,
-0.04841341823339462,
-0.0827825590968132,
0.0004603009729180485,
0.05614399537444115,
-0.0035920748487114906,
-0.23012670874595642,
-0.0660686194896698,
0.016473567113280296,
-0.04206548631191254,
-0.005613449960947037,
0.06645456701517105,
0.09575342386960983,
0.09551285952329636,
-0.014987784437835217,
-0.020490292459726334,
-0.03270154818892479,
0.08253990113735199,
-0.12398531287908554,
-0.057719651609659195
] |
null | null |
transformers
|
# doc2query/reddit-t5-small-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/reddit-t5-small-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=384, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-small](https://huggingface.co/google/t5-v1_1-small) for 547k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, body) from Reddit.
|
{"language": "en", "license": "apache-2.0", "datasets": ["datasets/sentence-transformers/reddit-title-body"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/reddit-t5-small-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/reddit-t5-small-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-small for 547k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, body) from Reddit.
|
[
"# doc2query/reddit-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-small for 547k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, body) from Reddit."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/reddit-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-small for 547k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, body) from Reddit."
] |
[
75,
260,
32,
87
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/reddit-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-small for 547k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, body) from Reddit."
] |
[
-0.06270061433315277,
0.12766633927822113,
-0.006178125273436308,
0.08414231240749359,
0.1151236966252327,
0.0216821301728487,
0.06458486616611481,
0.1295161098241806,
-0.07023308426141739,
0.012643606401979923,
0.10576130449771881,
0.054098471999168396,
-0.0007980030495673418,
0.08426909893751144,
0.03211216628551483,
-0.20602020621299744,
-0.04495704546570778,
-0.019066259264945984,
0.07230640202760696,
0.13067670166492462,
0.0711880698800087,
-0.03914386406540871,
0.04919711872935295,
-0.027231862768530846,
-0.15508580207824707,
0.05433233454823494,
-0.031709935516119,
-0.017015725374221802,
0.07112186402082443,
0.05352811887860298,
0.0821366012096405,
-0.020329251885414124,
0.034912072122097015,
-0.11366301774978638,
0.023189028725028038,
0.07002822309732437,
-0.017933838069438934,
0.09045083075761795,
0.02703014574944973,
-0.022303219884634018,
0.2276957482099533,
-0.07754664123058319,
0.028204383328557014,
0.05916425585746765,
-0.08660104125738144,
-0.11898775398731232,
-0.05502365902066231,
0.08949024230241776,
0.04224875941872597,
0.09594777226448059,
-0.01062055490911007,
0.05151795968413353,
-0.055431243032217026,
0.05795248970389366,
0.17717663943767548,
-0.15225836634635925,
-0.03424926847219467,
0.03531796112656593,
0.003734211903065443,
0.07765595614910126,
0.006734849885106087,
-0.05470453202724457,
-0.02769860439002514,
0.03501089662313461,
0.11768738925457001,
0.00819630827754736,
-0.025627359747886658,
0.014669617637991905,
-0.1130879670381546,
-0.0742681473493576,
0.14123035967350006,
-0.027462726458907127,
-0.05211282894015312,
-0.09788868576288223,
-0.02684831991791725,
0.10592974722385406,
0.02420329675078392,
-0.01943172700703144,
-0.01011365745216608,
0.029011329635977745,
-0.022713176906108856,
-0.12597008049488068,
-0.08948936313390732,
-0.06362329423427582,
0.0685434639453888,
-0.018788481131196022,
0.005719097796827555,
0.016094274818897247,
-0.05937369912862778,
0.11431922763586044,
-0.056586239486932755,
-0.017283929511904716,
-0.025379076600074768,
-0.10394962131977081,
-0.0072811986319720745,
0.00040287492447532713,
-0.06284856051206589,
-0.1266205608844757,
-0.0027262938674539328,
0.002558161038905382,
0.007302241865545511,
-0.010554000735282898,
0.015311246737837791,
0.03850807994604111,
0.002264314563944936,
0.11406905204057693,
-0.039628978818655014,
-0.010684791021049023,
0.07209976762533188,
0.01495690830051899,
-0.05479128286242485,
0.01544579491019249,
-0.06726335734128952,
-0.10021981596946716,
0.08883512020111084,
0.03563637286424637,
-0.0072761597111821175,
0.05250374600291252,
-0.06297818571329117,
-0.027553152292966843,
0.051282063126564026,
-0.1016092598438263,
0.01756209321320057,
0.050420984625816345,
-0.09009718149900436,
-0.06891930848360062,
0.1757327765226364,
-0.07976678013801575,
-0.13081727921962738,
0.0005850683082826436,
-0.03230036050081253,
0.0044799912720918655,
-0.13068035244941711,
-0.12757927179336548,
0.02463163062930107,
-0.14525294303894043,
-0.041310083121061325,
-0.05947710946202278,
-0.26590123772621155,
0.004520293790847063,
0.06204035505652428,
-0.04140652343630791,
-0.003991452977061272,
-0.06404562294483185,
0.0014437349746003747,
-0.06570536643266678,
0.02359028346836567,
-0.005149792414158583,
0.026642687618732452,
0.055248625576496124,
-0.037644121795892715,
0.0671103373169899,
-0.03322138637304306,
0.025138895958662033,
-0.009347028099000454,
0.015137320384383202,
-0.15305179357528687,
0.20579539239406586,
0.03166499733924866,
-0.01867765747010708,
-0.11065544188022614,
-0.03064309060573578,
-0.07234329730272293,
-0.0031069936230778694,
0.07274790853261948,
0.15257763862609863,
-0.12382958084344864,
-0.06270292401313782,
0.1925719529390335,
-0.04858310520648956,
-0.03359931707382202,
0.10661500692367554,
-0.08406585454940796,
0.14316049218177795,
0.1228063702583313,
0.016264870762825012,
0.10748033970594406,
-0.09496643394231796,
-0.04371017962694168,
0.002885267836973071,
-0.1278523951768875,
-0.06042807549238205,
0.07687307149171829,
-0.033798571676015854,
-0.00997748039662838,
0.008695361204445362,
-0.02875806950032711,
0.015683645382523537,
-0.04211908578872681,
-0.03239711746573448,
-0.02766834944486618,
-0.04007382690906525,
0.03160187229514122,
-0.0115011902526021,
0.010125317610800266,
0.041978619992733,
-0.08041245490312576,
0.15132668614387512,
0.05821630358695984,
-0.09117182344198227,
0.011601580306887627,
-0.05434311926364899,
0.03784387186169624,
-0.12541699409484863,
0.03732747957110405,
-0.19012226164340973,
0.027689507231116295,
0.0017001592786982656,
0.04821605607867241,
0.1514747142791748,
-0.04663556069135666,
0.06633619219064713,
0.0441121980547905,
-0.029927918687462807,
0.0167839378118515,
-0.07738994807004929,
-0.011503240093588829,
-0.09492234140634537,
-0.12527324259281158,
-0.1059674471616745,
-0.02550981380045414,
0.08928809314966202,
-0.18822146952152252,
0.057031527161598206,
0.01785748451948166,
-0.017039326950907707,
0.022238321602344513,
-0.03225681930780411,
0.058992233127355576,
-0.01534044649451971,
0.024204321205615997,
-0.03198087587952614,
0.08805176615715027,
0.03983801230788231,
-0.10519998520612717,
-0.03850838914513588,
-0.024468785151839256,
-0.16582588851451874,
0.04911887273192406,
0.01403853204101324,
-0.07584322988986969,
-0.10232830792665482,
-0.0543534979224205,
-0.013819738291203976,
-0.09832226485013962,
-0.07951068133115768,
0.16043683886528015,
0.04687562957406044,
0.1337723731994629,
-0.11344394832849503,
-0.08135364204645157,
0.01242376584559679,
-0.038102563470602036,
0.01710943505167961,
0.10241075605154037,
0.03825091943144798,
-0.09509740024805069,
0.02239951305091381,
0.03250914067029953,
-0.03341521695256233,
0.23759715259075165,
0.056711722165346146,
-0.10879354178905487,
-0.03416704386472702,
0.03908974304795265,
0.013136208988726139,
0.033216364681720734,
-0.0923030823469162,
-0.0177844800055027,
0.0028317472897469997,
-0.007963563315570354,
0.052555475383996964,
-0.11788338422775269,
0.03349025547504425,
0.050873856991529465,
-0.03873540088534355,
-0.09386773407459259,
-0.04397079721093178,
-0.012763417325913906,
0.06746996939182281,
0.04377344250679016,
0.06670034676790237,
0.05999293923377991,
-0.03814280033111572,
-0.15505795180797577,
0.18884508311748505,
-0.08150435984134674,
-0.10901433229446411,
-0.11493458598852158,
0.03456960618495941,
-0.004770408850163221,
-0.0010075772879645228,
0.0036166708450764418,
-0.07800940424203873,
0.01322015468031168,
-0.07464080303907394,
0.0431704968214035,
0.016721529886126518,
-0.013383036479353905,
-0.0489274300634861,
0.0647333562374115,
0.032917071133852005,
-0.1383306384086609,
0.022012971341609955,
-0.033337146043777466,
-0.04613763466477394,
0.050574805587530136,
-0.11040613055229187,
0.05110235512256622,
0.06905793398618698,
-0.06519146263599396,
0.05339999869465828,
-0.009188396856188774,
0.25613170862197876,
-0.00587103608995676,
0.044468894600868225,
0.10357555001974106,
-0.05611811950802803,
0.061107806861400604,
0.013913308270275593,
0.03281121328473091,
-0.008923701010644436,
0.04542454332113266,
-0.005305933300405741,
-0.04930875450372696,
-0.24529007077217102,
-0.019368436187505722,
-0.04580703377723694,
-0.019156834110617638,
0.057976458221673965,
0.016322355717420578,
-0.03469717130064964,
0.03709792345762253,
0.016496868804097176,
0.06872665882110596,
0.07095932215452194,
0.08970310539007187,
-0.003926420118659735,
-0.0007056176546029747,
0.025348199531435966,
-0.05746855586767197,
0.01784423179924488,
0.04541545361280441,
0.07150603830814362,
0.28456613421440125,
-0.04127625748515129,
0.13960036635398865,
0.07943979650735855,
0.02770623192191124,
0.0433780737221241,
0.11295418441295624,
-0.1026306003332138,
0.05637245997786522,
-0.05052046477794647,
-0.07171330600976944,
-0.046630728989839554,
0.059452421963214874,
-0.0567857027053833,
0.02158511057496071,
-0.05615035444498062,
0.03860416263341904,
0.05421671271324158,
0.18423086404800415,
-0.022270606830716133,
-0.2079884558916092,
-0.04918260872364044,
-0.04184731841087341,
-0.0019802143797278404,
-0.016326094046235085,
-0.038196805864572525,
0.08620245009660721,
-0.0856311097741127,
-0.06604008376598358,
-0.04339008778333664,
0.10375191271305084,
-0.032483987510204315,
0.005975663661956787,
-0.010918395593762398,
0.11113076657056808,
-0.022557610645890236,
0.08379553258419037,
-0.2077130526304245,
0.12234184890985489,
0.03946699574589729,
0.11455809324979782,
-0.08264323323965073,
0.029486797749996185,
0.007644052151590586,
-0.046470560133457184,
0.12980136275291443,
0.008060420863330364,
0.11003851145505905,
-0.048676248639822006,
-0.04614667966961861,
0.03195854648947716,
0.044845182448625565,
0.0017898224759846926,
0.12064847350120544,
0.020364291965961456,
-0.0030964675825089216,
0.025555593892931938,
0.0898471474647522,
-0.09516621381044388,
-0.18690142035484314,
-0.002384419087320566,
-0.07815304398536682,
0.017506448552012444,
-0.003137427382171154,
-0.013445715419948101,
0.1146310344338417,
0.1253369301557541,
-0.19413451850414276,
-0.14073900878429413,
-0.09276030212640762,
0.10401991009712219,
0.060469843447208405,
-0.06362704187631607,
0.0065468610264360905,
0.04427475482225418,
0.10751906782388687,
-0.04447688162326813,
-0.07338900119066238,
0.028507165610790253,
-0.0820981040596962,
-0.10866541415452957,
-0.033515628427267075,
0.06370566785335541,
0.11520373076200485,
0.07205286622047424,
0.04173284024000168,
-0.005122953560203314,
-0.05079056695103645,
-0.07291454821825027,
0.007140078581869602,
0.09026343375444412,
0.11364298313856125,
0.045254915952682495,
-0.13344262540340424,
-0.06355258077383041,
-0.10037188977003098,
-0.009354324080049992,
0.1511264443397522,
0.13357432186603546,
-0.08192063122987747,
0.07116924226284027,
0.152582049369812,
-0.12455222010612488,
-0.14082342386245728,
-0.008035936392843723,
0.09431268274784088,
0.04260729253292084,
-0.039364226162433624,
-0.3053007125854492,
0.03462100401520729,
0.059706319123506546,
0.02473652921617031,
-0.09076659381389618,
-0.31027841567993164,
-0.11984488368034363,
0.06539846211671829,
0.08135011792182922,
0.05833428353071213,
-0.09496940672397614,
-0.019068803638219833,
-0.013506955467164516,
0.07089617848396301,
0.1934882402420044,
-0.20195013284683228,
0.0568019337952137,
0.012273689731955528,
0.03694571182131767,
0.04374688118696213,
-0.026910344138741493,
-0.008852800354361534,
0.003576630726456642,
0.0319216288626194,
0.002353419316932559,
-0.0007365299388766289,
0.038444746285676956,
-0.05656003579497337,
0.1568402796983719,
-0.015333511866629124,
0.062164824455976486,
-0.14965425431728363,
-0.060290925204753876,
-0.10084015876054764,
-0.005591078195720911,
-0.03539547324180603,
-0.053803957998752594,
0.009558456018567085,
0.028758198022842407,
0.07402978092432022,
0.019185082986950874,
-0.06188143789768219,
-0.07652338594198227,
0.035399433225393295,
0.154791459441185,
0.17704378068447113,
-0.003994103521108627,
-0.050089992582798004,
0.017429903149604797,
-0.0023772744461894035,
0.08825398236513138,
-0.1465369462966919,
-0.0028736128006130457,
0.050449810922145844,
0.04049830138683319,
0.1772674322128296,
0.06845252960920334,
-0.12989163398742676,
0.013877292163670063,
0.050459738820791245,
-0.13855759799480438,
-0.05758967995643616,
-0.02408634126186371,
0.03459176421165466,
-0.14375577867031097,
-0.019964778795838356,
0.08997225761413574,
-0.1169748529791832,
-0.04433803632855415,
0.0059692454524338245,
0.03987585008144379,
-0.05673716962337494,
0.1193995252251625,
0.1034356951713562,
0.035167619585990906,
-0.036164697259664536,
0.08829719573259354,
0.057797547429800034,
-0.040602441877126694,
0.06951624155044556,
0.1497398316860199,
-0.08607734739780426,
-0.022968512028455734,
-0.06426846981048584,
0.14357438683509827,
-0.02708566002547741,
-0.04271263629198074,
-0.01866709068417549,
-0.01844928227365017,
0.09075182676315308,
0.14682582020759583,
0.027790863066911697,
0.0617016926407814,
-0.07625284790992737,
0.004200931638479233,
-0.14503173530101776,
0.10447213798761368,
0.035143814980983734,
-0.06629659235477448,
-0.03540901094675064,
0.07243053615093231,
0.001043262891471386,
0.017704561352729797,
-0.017895599827170372,
-0.026693012565374374,
-0.08681789040565491,
0.009843018837273121,
-0.06948179751634598,
0.021740799769759178,
-0.03800142928957939,
-0.0012983584310859442,
-0.04176081717014313,
0.0554414838552475,
0.06165220960974693,
0.03709112107753754,
-0.04437265917658806,
-0.021592065691947937,
0.0008308571414090693,
0.06916308403015137,
-0.1343078315258026,
-0.007149888202548027,
0.02439231239259243,
-0.07420941442251205,
0.12178342044353485,
0.05600912496447563,
-0.057056974619627,
-0.005839568562805653,
-0.051938559859991074,
-0.0035388972610235214,
-0.0003679094079416245,
-0.026737961918115616,
-0.0037165102548897266,
-0.10789284110069275,
0.02167011983692646,
-0.008012249134480953,
-0.0036584269255399704,
-0.006155562587082386,
0.1309530884027481,
-0.07686471194028854,
0.0770147517323494,
0.042332619428634644,
0.014365547336637974,
-0.10614488273859024,
0.10452324151992798,
-0.001375496736727655,
0.08000636845827103,
0.09929890930652618,
-0.057068757712841034,
0.04210948944091797,
-0.15724939107894897,
-0.010366663336753845,
0.03891456872224808,
0.029555104672908783,
-0.10968632996082306,
-0.03127354755997658,
0.039015475660562515,
-0.02435503713786602,
0.08009084314107895,
0.021344617009162903,
-0.015923982486128807,
-0.013060307130217552,
-0.02431165799498558,
-0.08233526349067688,
-0.025036675855517387,
0.07326823472976685,
-0.03643393889069557,
-0.009878817945718765,
0.025165479630231857,
0.06471145153045654,
-0.008466610684990883,
0.05560852959752083,
0.15050536394119263,
0.10997366160154343,
0.1197396069765091,
0.03876187652349472,
0.05753856152296066,
0.03931654617190361,
-0.03023298643529415,
0.02261338010430336,
0.09853231906890869,
0.0443534255027771,
-0.0866427794098854,
0.11966243386268616,
0.1805201917886734,
-0.0988934263586998,
0.07871252298355103,
0.04313259199261665,
-0.08579996228218079,
-0.09555501490831375,
-0.10984793305397034,
-0.0021832631900906563,
-0.03984789922833443,
-0.011950024403631687,
-0.13928931951522827,
0.028919734060764313,
-0.05673595145344734,
0.03618593141436577,
-0.024005359038710594,
0.09861966222524643,
0.0776771679520607,
-0.07068983465433121,
0.06970109045505524,
-0.04346461221575737,
0.024247458204627037,
-0.01390378549695015,
0.06395450979471207,
0.0806870311498642,
0.01071906741708517,
0.0554049126803875,
0.0777994766831398,
0.06477732211351395,
-0.03193942457437515,
-0.036497265100479126,
-0.08475618064403534,
-0.05883938446640968,
0.03628881648182869,
0.050921328365802765,
0.17071068286895752,
0.0070988391526043415,
-0.11849430203437805,
-0.012615601532161236,
0.14204584062099457,
-0.03761287406086922,
-0.06429409980773926,
-0.11777110397815704,
0.16542890667915344,
0.014576557092368603,
-0.035842955112457275,
-0.02465105801820755,
-0.0826493352651596,
0.0357627235352993,
0.17612673342227936,
0.17231819033622742,
-0.04170287773013115,
-0.014982366934418678,
-0.008800535462796688,
0.01260348316282034,
-0.005959680303931236,
0.1262383908033371,
0.00569170294329524,
0.2577234208583832,
-0.07092762738466263,
0.14548887312412262,
-0.048631537705659866,
0.005539881065487862,
-0.03044736757874489,
0.036087766289711,
0.010002649389207363,
-0.028079036623239517,
-0.012098575942218304,
0.07068106532096863,
-0.1115153431892395,
-0.1965484321117401,
-0.099730484187603,
-0.07202550023794174,
-0.052562203258275986,
-0.05116311088204384,
-0.038981013000011444,
-0.0044747618958354,
0.09631382673978806,
-0.006635566242039204,
0.01429620198905468,
0.12711705267429352,
-0.012539775110781193,
-0.11858147382736206,
-0.035102419555187225,
0.06728149205446243,
-0.002050760667771101,
0.10783575475215912,
-0.0014373938320204616,
0.09238211810588837,
0.07274200022220612,
-0.0277133546769619,
-0.1546638160943985,
0.021204696968197823,
-0.010009693913161755,
0.06570813059806824,
0.019627053290605545,
0.10353857278823853,
-0.0012522450415417552,
0.05767505615949631,
0.00630425289273262,
-0.023101482540369034,
0.0036076486576348543,
-0.014255676418542862,
-0.06706147640943527,
-0.11182215809822083,
0.04923182725906372,
-0.035347163677215576,
0.11216137558221817,
0.19505344331264496,
-0.06572017073631287,
0.014560866169631481,
-0.07416348159313202,
0.013252298347651958,
0.03368344530463219,
0.04767828434705734,
-0.04757841303944588,
-0.07223834097385406,
-0.0019823836628347635,
0.04229413717985153,
-0.001699580461718142,
-0.2312181293964386,
-0.05633499473333359,
0.025486521422863007,
-0.040503066033124924,
-0.018394455313682556,
0.06726189702749252,
0.0832206979393959,
0.0950898677110672,
-0.007595881354063749,
-0.009987433440983295,
-0.030740415677428246,
0.07969848066568375,
-0.12288668751716614,
-0.04920538142323494
] |
null | null |
transformers
|
# doc2query/stackexchange-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/stackexchange-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 449k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, best_answer_pairs) from StackExchange.
|
{"language": "en", "license": "apache-2.0", "datasets": ["flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/stackexchange-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/stackexchange-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 449k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, best_answer_pairs) from StackExchange.
|
[
"# doc2query/stackexchange-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 449k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, best_answer_pairs) from StackExchange."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/stackexchange-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 449k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, best_answer_pairs) from StackExchange."
] |
[
107,
261,
32,
94
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_best_voted_answer_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/stackexchange-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-base for 449k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, best_answer_pairs) from StackExchange."
] |
[
-0.051553595811128616,
0.14040625095367432,
-0.007596726529300213,
0.08573553711175919,
0.11692338436841965,
0.0028796421829611063,
0.00455116992816329,
0.11652423441410065,
-0.049686871469020844,
0.03642655909061432,
0.02348550409078598,
0.03390912711620331,
0.0288558192551136,
0.09398623555898666,
0.0813213363289833,
-0.19410839676856995,
-0.03542054444551468,
-0.028888488188385963,
0.10597417503595352,
0.08756663650274277,
0.07295649498701096,
-0.05123816058039665,
0.06270158290863037,
-0.008392260409891605,
-0.11894775182008743,
0.06424398720264435,
-0.01972183585166931,
-0.01199349295347929,
0.0425940603017807,
0.01958155445754528,
0.05900941416621208,
-0.02788427285850048,
0.013019041158258915,
-0.14181402325630188,
0.022731691598892212,
0.08401130884885788,
-0.04953629896044731,
0.0633799135684967,
0.07301285862922668,
-0.04297724738717079,
0.09488299489021301,
-0.11450207233428955,
0.02064412645995617,
0.06980827450752258,
-0.10286993533372879,
-0.1148386225104332,
-0.09356866776943207,
0.02771633490920067,
0.03682427108287811,
0.09774034470319748,
-0.024281950667500496,
0.01879982277750969,
-0.015526392497122288,
0.06045638024806976,
0.1226988434791565,
-0.1956080198287964,
-0.033911675214767456,
0.004527968354523182,
0.0008700714097358286,
0.023588750511407852,
-0.05715063959360123,
-0.08525318652391434,
-0.0426093153655529,
0.028636444360017776,
0.09823397547006607,
-0.021479232236742973,
-0.046963002532720566,
-0.012311098165810108,
-0.11270187050104141,
-0.04231257736682892,
0.08465984463691711,
-0.005910312756896019,
-0.07644101977348328,
-0.10764628648757935,
-0.056562040001153946,
0.0028162465896457434,
0.016929728910326958,
-0.0550801046192646,
0.022431720048189163,
0.05374221131205559,
0.053833529353141785,
-0.10226696729660034,
-0.11272180080413818,
-0.028028346598148346,
-0.03779785707592964,
-0.017912626266479492,
0.0317792184650898,
0.021555893123149872,
-0.010356341488659382,
0.0971909910440445,
-0.12710514664649963,
-0.07146546989679337,
-0.04879017546772957,
-0.06531309336423874,
-0.07564524561166763,
-0.018385231494903564,
-0.04609204828739166,
-0.140795037150383,
0.0029695334378629923,
0.08455974608659744,
-0.05694804713129997,
-0.002036862773820758,
-0.012618612498044968,
0.013383499346673489,
0.02420312725007534,
0.12262304872274399,
-0.03519110381603241,
0.011237877421081066,
0.07859550416469574,
0.010848313570022583,
-0.024055054411292076,
0.025791190564632416,
-0.05419201776385307,
-0.08270330727100372,
0.030031979084014893,
0.023681800812482834,
0.036784663796424866,
0.04661911725997925,
-0.0613553561270237,
-0.035345904529094696,
0.07097359746694565,
-0.1281706690788269,
0.008926902897655964,
0.02340858429670334,
-0.0841367319226265,
0.00194495206233114,
0.11749511957168579,
-0.03126973286271095,
-0.12074480950832367,
-0.025910042226314545,
-0.023113738745450974,
0.007725017610937357,
-0.1045304387807846,
-0.0678485855460167,
0.03629997745156288,
-0.1041739359498024,
-0.06577779352664948,
-0.0850418210029602,
-0.1779290735721588,
-0.08392514288425446,
0.033312927931547165,
-0.036038823425769806,
-0.005812942516058683,
-0.03769996017217636,
-0.005322954151779413,
-0.04122094810009003,
0.005044069606810808,
-0.019765499979257584,
0.001922672032378614,
0.041982751339673996,
-0.00652742013335228,
0.041576139628887177,
-0.025921564549207687,
0.029097510501742363,
-0.06767025589942932,
-0.019307225942611694,
-0.14225713908672333,
0.1866561323404312,
0.015355275943875313,
0.006779070943593979,
-0.11927571147680283,
-0.004516927991062403,
-0.06802479922771454,
-0.014544092118740082,
0.050528496503829956,
0.1382478028535843,
-0.2445572465658188,
-0.024679847061634064,
0.17630714178085327,
-0.0655374675989151,
-0.05132380500435829,
0.09851212054491043,
-0.054413165897130966,
0.12431547790765762,
0.1318102777004242,
0.05723486468195915,
0.1334991306066513,
-0.061665572226047516,
-0.1102047711610794,
-0.07252336293458939,
-0.10296156257390976,
0.05000791698694229,
0.06461948156356812,
-0.04384699836373329,
0.10134895890951157,
0.03144046291708946,
0.022524481639266014,
-0.002398688578978181,
0.003308580256998539,
-0.05270887911319733,
-0.03197041526436806,
-0.023457426577806473,
0.0017971942434087396,
-0.02129531092941761,
-0.002597252605482936,
0.043197471648454666,
-0.11577961593866348,
0.01803390122950077,
0.03691931813955307,
-0.04789775609970093,
0.04439088702201843,
-0.06346355378627777,
0.0006477956776507199,
-0.008476419374346733,
0.04126200079917908,
-0.1654789298772812,
-0.0741109773516655,
0.04508702829480171,
0.024419544264674187,
0.11213408410549164,
-0.12879005074501038,
0.06622423976659775,
0.03608379885554314,
-0.03257721662521362,
-0.006363656837493181,
-0.06997371464967728,
-0.045373961329460144,
-0.07370758056640625,
-0.07755504548549652,
-0.07903947681188583,
-0.025935281068086624,
0.05731185898184776,
-0.08859346061944962,
0.03183005377650261,
0.07414666563272476,
0.05105575546622276,
0.0070349290035665035,
-0.049419302493333817,
0.05653189495205879,
-0.001438701176084578,
0.026902565732598305,
-0.02559654973447323,
0.05488404259085655,
0.015129642561078072,
-0.06534162908792496,
0.014114461839199066,
-0.0681806355714798,
-0.11450505256652832,
0.04724258556962013,
0.03193025663495064,
-0.054190896451473236,
-0.13752330839633942,
-0.04948506876826286,
-0.02560892514884472,
-0.07539261132478714,
-0.05037103220820427,
0.20057359337806702,
0.03364008292555809,
0.08191723376512527,
-0.06727702170610428,
-0.03953539952635765,
-0.005601144395768642,
0.025604914873838425,
0.03833482414484024,
0.06086956337094307,
0.001453471602872014,
-0.0766577273607254,
0.028662895783782005,
0.025979364290833473,
-0.021954044699668884,
0.15753807127475739,
0.016566697508096695,
-0.12000706791877747,
-0.020614705979824066,
0.04360198974609375,
0.030912620946764946,
-0.03863987699151039,
-0.02358662150800228,
0.000264614587649703,
0.028155846521258354,
-0.018610384315252304,
0.03766603767871857,
-0.07676880061626434,
0.05428258329629898,
0.020223600789904594,
-0.035780176520347595,
-0.024369781836867332,
-0.017867455258965492,
0.04073207080364227,
0.05445830523967743,
0.022045394405722618,
0.10839459300041199,
0.029307231307029724,
-0.04946427792310715,
-0.11758549511432648,
0.1393389254808426,
-0.07878006249666214,
-0.14902186393737793,
-0.11770548671483994,
0.04760822653770447,
-0.04146570712327957,
0.012718734331429005,
0.04296647757291794,
-0.08876518905162811,
-0.0031201213132590055,
-0.05868913233280182,
0.027469612658023834,
0.055973075330257416,
-0.007328270934522152,
-0.14029477536678314,
0.04035039246082306,
0.032626718282699585,
-0.12459412962198257,
0.011513568460941315,
-0.02077070251107216,
-0.16864758729934692,
0.047052253037691116,
0.009442339651286602,
0.0530955009162426,
0.020613055676221848,
-0.013841607607901096,
0.007675403729081154,
-0.02575729414820671,
0.2245904952287674,
-0.04088868573307991,
0.09611281752586365,
0.14979827404022217,
-0.030769478529691696,
0.10013894736766815,
0.05187245458364487,
0.017969878390431404,
-0.039225440472364426,
0.05405807867646217,
0.04051648825407028,
-0.02892216481268406,
-0.23718538880348206,
-0.0060533275827765465,
-0.05132859945297241,
-0.03015017695724964,
0.09722587466239929,
0.05496680736541748,
-0.032366033643484116,
0.04363475739955902,
-0.03399506211280823,
0.021528637036681175,
0.0536094531416893,
0.06649638712406158,
0.04455612599849701,
0.01920347660779953,
0.048564713448286057,
-0.061727508902549744,
-0.010150591842830181,
0.06733841449022293,
0.0683971419930458,
0.17977352440357208,
-0.078634113073349,
0.17058050632476807,
0.0803285762667656,
0.12021703273057938,
0.03838810697197914,
0.09554267674684525,
-0.055907364934682846,
0.05556056275963783,
-0.031613729894161224,
-0.1017645001411438,
-0.046826597303152084,
0.07938973605632782,
0.019029440358281136,
-0.010753942653536797,
-0.0210935827344656,
-0.0011339016491547227,
0.049286164343357086,
0.18484430015087128,
-0.05049919709563255,
-0.16522671282291412,
-0.031809061765670776,
-0.03821304067969322,
-0.018892506137490273,
-0.0318867489695549,
-0.019952259957790375,
0.04756283387541771,
-0.10257676243782043,
0.01649998314678669,
-0.02077922411262989,
0.09241382032632828,
-0.022240640595555305,
-0.0006571874837391078,
0.051386039704084396,
0.09315963834524155,
0.04063406214118004,
0.060466207563877106,
-0.20103253424167633,
0.01806364580988884,
0.03359229490160942,
0.08827405422925949,
-0.06800872087478638,
0.05842971056699753,
0.02987045980989933,
-0.08087622374296188,
0.13037414848804474,
0.007544173859059811,
0.09547563642263412,
0.004103734157979488,
-0.09405779838562012,
0.024944618344306946,
0.06832663714885712,
-0.07538139075040817,
0.120988629758358,
0.02484997548162937,
-0.010165887884795666,
-0.02002471126616001,
0.14508385956287384,
-0.08068516850471497,
-0.19718945026397705,
0.026560787111520767,
-0.06766398251056671,
0.0949016734957695,
-0.008351619355380535,
-0.01938772201538086,
0.019259540364146233,
0.13190777599811554,
-0.14424020051956177,
-0.09714532643556595,
-0.10024874657392502,
0.030521497130393982,
0.11551763117313385,
-0.08669908344745636,
0.03160495311021805,
0.022842353209853172,
0.08068288862705231,
-0.029687857255339622,
-0.11366932094097137,
0.022389553487300873,
-0.023114025592803955,
-0.08054674416780472,
-0.030030598863959312,
0.11666446179151535,
0.05605786666274071,
0.03548843413591385,
0.022573353722691536,
0.048901498317718506,
-0.045045461505651474,
-0.07471965253353119,
0.06323707103729248,
0.17171792685985565,
0.11322619020938873,
0.04274717718362808,
-0.12223950028419495,
-0.032291147857904434,
-0.10233517736196518,
0.017706556245684624,
0.14227503538131714,
0.08972901105880737,
-0.06135672703385353,
0.08932989835739136,
0.08757603168487549,
-0.11845558136701584,
-0.16703292727470398,
-0.03154565766453743,
0.14115554094314575,
0.02616342157125473,
-0.0173032246530056,
-0.2795655131340027,
0.10783721506595612,
0.09778500348329544,
0.017041195183992386,
-0.011495646089315414,
-0.18815875053405762,
-0.11656642705202103,
-0.001789910369552672,
0.03067604824900627,
-0.003496150951832533,
-0.13829152286052704,
-0.03299669176340103,
0.0027233087457716465,
-0.008368478156626225,
0.22706851363182068,
-0.19028353691101074,
0.0666809231042862,
-0.00936758704483509,
0.017867548391222954,
0.04494686424732208,
-0.017146315425634384,
-0.0007430408149957657,
0.03489124774932861,
0.02790926769375801,
-0.007080146111547947,
0.05299454927444458,
0.03655519708991051,
-0.0652904212474823,
0.16162915527820587,
0.06723611056804657,
0.03454868867993355,
-0.1415795385837555,
-0.034172993153333664,
-0.07830338925123215,
0.00830345693975687,
-0.02755977213382721,
-0.02458123117685318,
0.01074360404163599,
0.03587767854332924,
0.09684300422668457,
-0.0007530167349614203,
-0.025082645937800407,
-0.0437934547662735,
0.06911121308803558,
0.22019632160663605,
0.13542653620243073,
-0.015586936846375465,
-0.13367919623851776,
0.0062538031488657,
-0.007249687798321247,
0.05577773228287697,
-0.11516862362623215,
0.03615473955869675,
0.09344649314880371,
0.006555418483912945,
0.09756861627101898,
0.04480329900979996,
-0.12599767744541168,
-0.004960719961673021,
0.07725823670625687,
-0.10830821096897125,
-0.12577076256275177,
-0.05635102093219757,
-0.029457442462444305,
-0.15310916304588318,
-0.03533615916967392,
0.08281249552965164,
-0.05696796253323555,
-0.05608054995536804,
0.012820557691156864,
0.02733251266181469,
-0.019381411373615265,
0.10228688269853592,
0.05260099470615387,
0.016089297831058502,
-0.07540736347436905,
0.11404786258935928,
0.10920561850070953,
-0.10819742828607559,
0.04780331999063492,
0.20904898643493652,
-0.07703268527984619,
-0.04052148386836052,
-0.0648755207657814,
0.05354766175150871,
-0.028801294043660164,
0.01885104738175869,
-0.009441827423870564,
-0.007170069497078657,
0.0658617913722992,
0.09996403753757477,
0.010948063805699348,
0.07452992349863052,
-0.029329894110560417,
0.023241210728883743,
-0.07847563177347183,
0.07130485028028488,
0.028295617550611496,
-0.0068786367774009705,
-0.03057648427784443,
0.10583116114139557,
-0.008434421382844448,
0.037530671805143356,
-0.0009009672212414443,
-0.03452742472290993,
-0.07249750941991806,
-0.007077192422002554,
-0.024754326790571213,
0.03929552808403969,
-0.03816622123122215,
-0.002983411541208625,
-0.04088990017771721,
0.03755299374461174,
0.04928770288825035,
0.04759964719414711,
-0.03497873991727829,
-0.06634292751550674,
-0.01649627462029457,
0.06057792901992798,
-0.1292327493429184,
0.004652027506381273,
0.030800797045230865,
-0.06409360468387604,
0.11116106063127518,
0.06651204824447632,
-0.0645579993724823,
0.015164652839303017,
-0.0475684255361557,
-0.05353151634335518,
0.009093916043639183,
-0.003572791116312146,
-0.009052242152392864,
-0.06952574849128723,
0.029767462983727455,
0.0062880259938538074,
-0.03197158873081207,
-0.01712709292769432,
0.09079958498477936,
-0.10793132334947586,
0.015954503789544106,
0.06310918182134628,
0.01269439049065113,
-0.11610054969787598,
0.02465042471885681,
0.03088836930692196,
0.09117282181978226,
0.13225768506526947,
-0.04952690377831459,
0.05001607909798622,
-0.16617299616336823,
-0.009928872808814049,
0.0018590963445603848,
0.008706730790436268,
-0.07894321531057358,
-0.06224668025970459,
0.04663756489753723,
-0.011767744086682796,
0.055402837693691254,
-0.015538273379206657,
-0.006424717139452696,
0.00769068906083703,
-0.03332772105932236,
-0.05570685863494873,
0.002655646065250039,
0.04249318316578865,
0.01068718358874321,
-0.040089044719934464,
0.03296693414449692,
0.05337010324001312,
-0.020115863531827927,
0.07624482363462448,
0.0806851014494896,
0.08888144046068192,
0.161737859249115,
0.03196212649345398,
0.060013942420482635,
0.027850866317749023,
-0.006068733986467123,
0.04954683035612106,
0.02645350992679596,
-0.01746949926018715,
-0.09874378889799118,
0.041497331112623215,
0.1432340443134308,
-0.08219514787197113,
0.12429652363061905,
0.03570398688316345,
-0.05896572768688202,
-0.0832967683672905,
-0.13860216736793518,
-0.04317960515618324,
-0.018671885132789612,
-0.0177125483751297,
-0.11947554349899292,
0.06417347490787506,
-0.01778678223490715,
-0.009784422814846039,
-0.015325834043323994,
0.10886230319738388,
0.004325604997575283,
-0.11119886487722397,
0.07877416163682938,
-0.029892904683947563,
0.04950587451457977,
0.07957936078310013,
0.058926913887262344,
0.06463786959648132,
0.05067784711718559,
0.07566041499376297,
0.0868263691663742,
0.09119065850973129,
-0.0003236155607737601,
-0.09840627759695053,
-0.0621032640337944,
-0.02132350578904152,
0.03139657899737358,
0.021248862147331238,
0.11911924183368683,
0.04489898681640625,
-0.09878887981176376,
0.01032637432217598,
0.1680486500263214,
-0.024984588846564293,
-0.06934399157762527,
-0.11881183087825775,
0.11289860308170319,
0.08340389281511307,
-0.03163812309503555,
-0.012364722788333893,
-0.09979261457920074,
0.042438287287950516,
0.11734919250011444,
0.14207495748996735,
-0.004620601423084736,
0.016116736456751823,
0.01915656588971615,
0.017886871472001076,
0.010999351739883423,
0.1077616885304451,
0.07615569978952408,
0.2157181352376938,
-0.060563430190086365,
0.14951997995376587,
-0.0068963258527219296,
-0.004600011743605137,
-0.07042525708675385,
0.06451383978128433,
-0.05114307999610901,
0.028837986290454865,
-0.03584679588675499,
0.0554165244102478,
-0.07531996071338654,
-0.12499655038118362,
-0.06702566891908646,
-0.09475471824407578,
-0.08484501391649246,
-0.05931345373392105,
-0.021912792697548866,
0.030773840844631195,
0.06430429220199585,
0.03184407204389572,
0.015528684481978416,
0.09727531671524048,
0.007594679947942495,
-0.107303686439991,
-0.015346989966928959,
0.0689307376742363,
-0.04388169199228287,
0.12844833731651306,
0.018329543992877007,
0.07076975703239441,
0.05880416929721832,
0.00022321223514154553,
-0.1491418033838272,
0.008188404142856598,
-0.034148260951042175,
-0.03602714464068413,
0.017222406342625618,
0.12183844298124313,
0.020596642047166824,
0.02961745858192444,
0.010478389449417591,
-0.021856384351849556,
0.00700959051027894,
0.007302695885300636,
-0.012601488269865513,
-0.10361561924219131,
0.07336311787366867,
-0.084494948387146,
0.11289907246828079,
0.18735341727733612,
-0.02903650887310505,
0.0473126545548439,
-0.11169838160276413,
-0.008740797638893127,
0.030316097661852837,
0.03576447442173958,
-0.037856325507164,
-0.08296380937099457,
0.014423263259232044,
0.08050361275672913,
0.01932488940656185,
-0.15582512319087982,
-0.07009359449148178,
0.05513053387403488,
-0.019872520118951797,
-0.061842046678066254,
0.09614603966474533,
0.04183100536465645,
0.05995452031493187,
-0.009693427011370659,
-0.10743723809719086,
-0.0017178424168378115,
0.06833091378211975,
-0.1312733143568039,
-0.048085618764162064
] |
null | null |
transformers
|
# doc2query/stackexchange-title-body-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/stackexchange-title-body-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 550k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, question_body) from StackExchange.
|
{"language": "en", "license": "apache-2.0", "datasets": ["flax-sentence-embeddings/stackexchange_title_body_jsonl"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/stackexchange-title-body-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:flax-sentence-embeddings/stackexchange_title_body_jsonl",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_body_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/stackexchange-title-body-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 550k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, question_body) from StackExchange.
|
[
"# doc2query/stackexchange-title-body-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 550k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, question_body) from StackExchange."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_body_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/stackexchange-title-body-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 550k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, question_body) from StackExchange."
] |
[
101,
265,
32,
89
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_body_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/stackexchange-title-body-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-base for 550k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, question_body) from StackExchange."
] |
[
-0.02891315147280693,
0.16878317296504974,
-0.00744715565815568,
0.09136367589235306,
0.1312509924173355,
0.009205386973917484,
0.029842499643564224,
0.13300949335098267,
-0.05107518285512924,
0.02850905805826187,
0.03673307225108147,
0.04921635612845421,
0.03566811978816986,
0.07746543735265732,
0.0766703337430954,
-0.2510511577129364,
-0.03522171080112457,
-0.04352899268269539,
0.11064314842224121,
0.08339579403400421,
0.06721866130828857,
-0.05147683247923851,
0.0704382061958313,
-0.011535256169736385,
-0.09890346229076385,
0.06453356146812439,
-0.029351739212870598,
-0.024754153564572334,
0.060188356786966324,
0.019915133714675903,
0.07339177280664444,
-0.02018543891608715,
0.039812423288822174,
-0.13182999193668365,
0.02214839681982994,
0.0951739251613617,
-0.044294919818639755,
0.06277693063020706,
0.042338304221630096,
-0.03720245510339737,
0.0708644911646843,
-0.10546254366636276,
0.04151289910078049,
0.049228373914957047,
-0.09575632959604263,
-0.14063246548175812,
-0.08847836405038834,
0.023143135011196136,
0.01976882852613926,
0.09163789451122284,
-0.014836415648460388,
0.02717931941151619,
-0.006124901585280895,
0.06065438315272331,
0.17373205721378326,
-0.21289516985416412,
-0.030957123264670372,
0.03750801458954811,
-0.011528710834681988,
0.05092563480138779,
-0.04978993907570839,
-0.07902247458696365,
-0.04758574441075325,
0.04588551074266434,
0.08359603583812714,
-0.007570700254291296,
-0.0428287573158741,
0.005907866172492504,
-0.10217246413230896,
-0.04380903020501137,
0.10109011828899384,
-0.018917076289653778,
-0.08297020941972733,
-0.10968132317066193,
-0.06665840744972229,
0.008794211782515049,
0.01342356950044632,
-0.05153491348028183,
0.027029644697904587,
0.045431219041347504,
0.02451225370168686,
-0.11073780059814453,
-0.1065601110458374,
-0.035608310252428055,
-0.04014397785067558,
-0.011822725646197796,
0.048939719796180725,
0.030212758108973503,
0.0010751804802566767,
0.12096843868494034,
-0.09608657658100128,
-0.06250159442424774,
-0.03919227048754692,
-0.07169578969478607,
-0.06785066425800323,
-0.0008446576539427042,
-0.055399052798748016,
-0.1666140854358673,
-0.007278440520167351,
0.06419318914413452,
-0.0517849400639534,
-0.00079037417890504,
-0.03393356502056122,
0.019802281633019447,
0.02055102027952671,
0.13658834993839264,
-0.05279599875211716,
0.030633578076958656,
0.06401388347148895,
0.015599778853356838,
-0.029906997457146645,
0.0041293855756521225,
-0.08069406449794769,
-0.07611540704965591,
0.042241886258125305,
0.02706039324402809,
0.001546562765724957,
0.0610131174325943,
-0.04833079129457474,
-0.03261877968907356,
0.06262554973363876,
-0.12567760050296783,
0.009123256430029869,
-0.0003045296762138605,
-0.06826499104499817,
0.007529735565185547,
0.14615347981452942,
-0.032631680369377136,
-0.11077633500099182,
-0.0006283925613388419,
-0.025426283478736877,
-0.0068695396184921265,
-0.12681017816066742,
-0.08152535557746887,
0.015969738364219666,
-0.09059512615203857,
-0.06105075031518936,
-0.0911407470703125,
-0.18645134568214417,
-0.055503953248262405,
0.03560182824730873,
-0.02828785590827465,
-0.000739447248633951,
-0.04456572234630585,
-0.013681045733392239,
-0.06048025190830231,
0.0077455914579331875,
-0.025728892534971237,
0.006739179138094187,
0.05269433185458183,
-0.04172864183783531,
0.06182907149195671,
-0.03214681148529053,
0.02979443408548832,
-0.0515466034412384,
-0.00926267635077238,
-0.1419069617986679,
0.17213834822177887,
0.00824559573084116,
-0.015278222039341927,
-0.09462055563926697,
-0.013530137948691845,
-0.07830160856246948,
-0.01380907092243433,
0.05567924305796623,
0.12203733623027802,
-0.24099388718605042,
-0.023931827396154404,
0.20911678671836853,
-0.06430967897176743,
-0.021957077085971832,
0.10876006633043289,
-0.06056166812777519,
0.11946260184049606,
0.11324237287044525,
0.08078273385763168,
0.12590762972831726,
-0.08250578492879868,
-0.06723393499851227,
-0.06059642881155014,
-0.11011131852865219,
0.06584867089986801,
0.048467181622982025,
-0.042119547724723816,
0.0822007805109024,
0.04561689868569374,
0.03746790066361427,
-0.018206266686320305,
-0.01648642122745514,
-0.04667384922504425,
-0.02338840812444687,
-0.0209996048361063,
0.010671162977814674,
-0.028806155547499657,
-0.008554969914257526,
0.04650343954563141,
-0.10274329036474228,
0.04728703945875168,
0.04107177257537842,
-0.0681537389755249,
0.048843301832675934,
-0.052785586565732956,
-0.003048744983971119,
0.009035571478307247,
0.02382746711373329,
-0.16669400036334991,
-0.06704094260931015,
0.045095957815647125,
0.02412072941660881,
0.11502347141504288,
-0.07708761096000671,
0.06562953442335129,
0.04942753538489342,
-0.0264689102768898,
0.009338070638477802,
-0.08499816805124283,
-0.033407554030418396,
-0.07187643647193909,
-0.06550437957048416,
-0.07339797168970108,
-0.020043838769197464,
0.08495799452066422,
-0.07303889095783234,
0.044959906488657,
0.04045184701681137,
0.0489887110888958,
0.014910300262272358,
-0.04341034218668938,
0.03504844009876251,
0.004245104268193245,
0.03824612498283386,
-0.023697594180703163,
0.054028116166591644,
0.02149898372590542,
-0.05022341385483742,
0.010907093994319439,
-0.05936862900853157,
-0.10684167593717575,
0.052212417125701904,
0.05298033356666565,
-0.0714825689792633,
-0.1321132481098175,
-0.05052904039621353,
-0.025795141234993935,
-0.08071113377809525,
-0.061272379010915756,
0.22362525761127472,
0.035244833678007126,
0.09684935212135315,
-0.0667663961648941,
-0.024464935064315796,
-0.007296063005924225,
0.0007540617370977998,
0.031393155455589294,
0.03623228147625923,
-0.0021297046914696693,
-0.05639170855283737,
0.03123978152871132,
0.0030188201926648617,
-0.03712552413344383,
0.18208929896354675,
0.02910759113729,
-0.11043059825897217,
-0.021061072126030922,
0.024680553004145622,
0.02387913502752781,
-0.02032681740820408,
-0.03150185942649841,
0.0024160342290997505,
0.027700237929821014,
-0.00674260500818491,
0.03818860277533531,
-0.09358254820108414,
0.04396709427237511,
0.024992788210511208,
-0.035601988434791565,
-0.03127819299697876,
-0.012997726909816265,
0.04321950674057007,
0.05892426520586014,
0.02482987940311432,
0.10961681604385376,
0.022588636726140976,
-0.04482703655958176,
-0.10832194238901138,
0.1441417634487152,
-0.09268637001514435,
-0.15348094701766968,
-0.12709754705429077,
0.03129763901233673,
-0.03550170734524727,
0.0003325588477309793,
0.04230322688817978,
-0.07586982101202011,
-0.002932046540081501,
-0.073336161673069,
0.022883636876940727,
0.027046026661992073,
-0.0030106198973953724,
-0.1294175386428833,
0.04556310921907425,
0.007503497414290905,
-0.12809504568576813,
0.007876441814005375,
-0.02735009416937828,
-0.1604527086019516,
0.05792367085814476,
-0.024530906230211258,
0.05180908367037773,
0.026499146595597267,
-0.0336463525891304,
0.021736346185207367,
-0.024757297709584236,
0.19543929398059845,
-0.02198677323758602,
0.09590510278940201,
0.14893116056919098,
-0.031616874039173126,
0.10974953323602676,
0.0384228341281414,
0.01910044625401497,
-0.03365727886557579,
0.058312442153692245,
0.048106737434864044,
-0.029682690277695656,
-0.22555723786354065,
0.0021384700667113066,
-0.043357107788324356,
-0.002144343452528119,
0.08037354797124863,
0.04434356465935707,
-0.013959611766040325,
0.04458708316087723,
-0.02612767554819584,
0.03296728432178497,
0.038520876318216324,
0.0753890797495842,
0.05354488268494606,
-0.011721487157046795,
0.05902453884482384,
-0.051314834505319595,
-0.013087314553558826,
0.06421583145856857,
0.09629105031490326,
0.1897176057100296,
-0.08505468815565109,
0.12333355844020844,
0.08251991868019104,
0.11060941219329834,
0.04173680767416954,
0.08283034712076187,
-0.06039819493889809,
0.045471448451280594,
-0.04607241228222847,
-0.08733665198087692,
-0.05976969748735428,
0.06846384704113007,
0.017085691913962364,
-0.014071928337216377,
-0.026472359895706177,
0.0020695042330771685,
0.05200152099132538,
0.19229821860790253,
-0.055159199982881546,
-0.17896053194999695,
-0.028791913762688637,
-0.03975071758031845,
-0.01645057648420334,
-0.042278461158275604,
-0.018756158649921417,
0.06118122115731239,
-0.11797145754098892,
-0.002882586093619466,
-0.03395096957683563,
0.07945822179317474,
-0.04523593187332153,
0.0022765283938497305,
0.05593925341963768,
0.11008883267641068,
0.018666787073016167,
0.06260884553194046,
-0.1861211359500885,
0.02330937422811985,
0.025913219898939133,
0.06341909617185593,
-0.07909752428531647,
0.04735160619020462,
0.014484450221061707,
-0.07170015573501587,
0.12744884192943573,
0.008398097939789295,
0.0621693916618824,
0.011694417335093021,
-0.08468794077634811,
0.026739371940493584,
0.08334720134735107,
-0.05003409460186958,
0.11939701437950134,
0.017508188262581825,
-0.004166601225733757,
-0.018492069095373154,
0.12320854514837265,
-0.06510856002569199,
-0.20097677409648895,
0.02037525549530983,
-0.05994443595409393,
0.0606834851205349,
-0.018583940342068672,
-0.02835073322057724,
0.03163754567503929,
0.1434551179409027,
-0.1129918247461319,
-0.10308941453695297,
-0.09870926290750504,
0.05834861472249031,
0.09781934320926666,
-0.06891445070505142,
0.028175411745905876,
0.013138487935066223,
0.0818951427936554,
-0.0464327298104763,
-0.10795308649539948,
0.019277730956673622,
-0.007403216324746609,
-0.07853737473487854,
-0.04553768038749695,
0.11691062897443771,
0.05583392456173897,
0.03266867250204086,
0.0038875513710081577,
0.023737499490380287,
-0.042939092963933945,
-0.0796389952301979,
0.03666602447628975,
0.17752842605113983,
0.10959891229867935,
0.05673917010426521,
-0.1497340351343155,
-0.024952035397291183,
-0.09456846863031387,
0.026376189664006233,
0.1414177417755127,
0.09603408724069595,
-0.0765380933880806,
0.1017022356390953,
0.07459273189306259,
-0.11340489983558655,
-0.1907343715429306,
-0.027857163920998573,
0.14412346482276917,
0.027991728857159615,
-0.00816873274743557,
-0.29514625668525696,
0.07559951394796371,
0.0912398248910904,
0.023426655679941177,
-0.018277622759342194,
-0.2251884639263153,
-0.11193939298391342,
0.009196112863719463,
0.06311797350645065,
0.024291733279824257,
-0.118461012840271,
-0.019087541848421097,
0.01059090718626976,
-0.021302200853824615,
0.18335509300231934,
-0.18924079835414886,
0.07730596512556076,
-0.007596577983349562,
0.06826155632734299,
0.04843815788626671,
-0.017939386889338493,
0.01002234872430563,
0.01921246573328972,
0.033479463309049606,
-0.016983898356556892,
0.022362608462572098,
0.05089186131954193,
-0.05476018786430359,
0.1578548550605774,
0.07866736501455307,
0.047908782958984375,
-0.09566237777471542,
-0.03968724235892296,
-0.07939189672470093,
-0.0036789276637136936,
-0.025009898468852043,
-0.054909687489271164,
0.0014300226466730237,
0.04865637794137001,
0.08090738952159882,
-0.012416171841323376,
-0.03349694982171059,
-0.06590324640274048,
0.044169168919324875,
0.2032536268234253,
0.09515266865491867,
-0.008101848885416985,
-0.07669935375452042,
0.012000068090856075,
-0.018243098631501198,
0.05075576901435852,
-0.11943487823009491,
0.022711586207151413,
0.07805778831243515,
0.008304654620587826,
0.0945965126156807,
0.06397540867328644,
-0.14352065324783325,
-0.005266331601887941,
0.06239261105656624,
-0.13156232237815857,
-0.09780065715312958,
-0.050004955381155014,
-0.016389274969697,
-0.15048235654830933,
-0.021785805001854897,
0.08400940150022507,
-0.06072737276554108,
-0.050050728023052216,
0.009831671603024006,
0.019238542765378952,
-0.024429941549897194,
0.07461904734373093,
0.06215885281562805,
-0.0008342221262864769,
-0.05985473096370697,
0.1289598047733307,
0.08874569088220596,
-0.07122093439102173,
0.0359768271446228,
0.18028408288955688,
-0.08489323407411575,
-0.03311314061284065,
-0.09387777745723724,
0.06871587038040161,
-0.027886144816875458,
0.004222613759338856,
-0.015248237177729607,
-0.010331512428820133,
0.05750669166445732,
0.08233770728111267,
0.010237244889140129,
0.07253113389015198,
-0.04214220494031906,
0.022917509078979492,
-0.09017723798751831,
0.08019578456878662,
-0.002519140485674143,
0.00013827093061991036,
-0.03328074887394905,
0.09064053744077682,
-0.012848297134041786,
0.05147886276245117,
-0.005625254474580288,
-0.04697701707482338,
-0.07547740638256073,
0.0006301436806097627,
0.007563701365143061,
0.007479929830878973,
-0.0339980348944664,
-0.012931601144373417,
-0.026446441188454628,
0.0358528234064579,
0.06064556539058685,
0.050782255828380585,
-0.025475719943642616,
-0.0660531148314476,
-0.029117386788129807,
0.05519808828830719,
-0.1301279366016388,
-0.005890325177460909,
0.02890033647418022,
-0.06810658425092697,
0.1121499314904213,
0.07025302946567535,
-0.06043858081102371,
0.01732255518436432,
-0.039753206074237823,
-0.03842666745185852,
0.009080342017114162,
0.01173063088208437,
-0.006775090470910072,
-0.0576944462954998,
0.01887461543083191,
-0.014699595980346203,
-0.040307845920324326,
-0.012684725224971771,
0.12852968275547028,
-0.10198856890201569,
0.04096994176506996,
0.05929887294769287,
0.014322856441140175,
-0.11542894691228867,
0.040906522423028946,
0.02116350829601288,
0.08159435540437698,
0.09434043616056442,
-0.05462212488055229,
0.05114010348916054,
-0.16700729727745056,
-0.02181718684732914,
0.013786503113806248,
0.012529105879366398,
-0.01798579841852188,
-0.07424402236938477,
0.04364403709769249,
-0.001162995002232492,
0.09070143848657608,
0.005297187250107527,
0.004681884311139584,
-0.004679338540881872,
-0.04001563787460327,
-0.03722211718559265,
-0.003252037102356553,
0.048103753477334976,
-0.001104572438634932,
-0.039336659014225006,
0.013047123327851295,
0.046083930879831314,
-0.03408490866422653,
0.07590144127607346,
0.07535571604967117,
0.07685770839452744,
0.15334057807922363,
0.015509147197008133,
0.0768071860074997,
0.04274028539657593,
-0.03432772681117058,
0.04985319450497627,
0.04545086622238159,
-0.01652362011373043,
-0.09893904626369476,
0.005553283262997866,
0.1520720273256302,
-0.08786591142416,
0.12974607944488525,
0.05477682501077652,
-0.0640084445476532,
-0.08470578491687775,
-0.13236652314662933,
-0.037442732602357864,
-0.030422566458582878,
-0.008673510514199734,
-0.11808367073535919,
0.06770577281713486,
-0.01747562736272812,
-0.015897860750555992,
-0.007721728645265102,
0.11758237332105637,
-0.005389990285038948,
-0.10558135062456131,
0.04623231291770935,
-0.021884694695472717,
0.0563969649374485,
0.06420774012804031,
0.038114991039037704,
0.06506678462028503,
0.05162264406681061,
0.06818558275699615,
0.07784578949213028,
0.10112939029932022,
-0.008889343589544296,
-0.08132508397102356,
-0.05541611835360527,
-0.029447415843605995,
0.013402272947132587,
-0.0043677231296896935,
0.12287820875644684,
0.05618797242641449,
-0.09500562399625778,
0.0028082598000764847,
0.18548843264579773,
-0.04560691863298416,
-0.07460436969995499,
-0.1401176154613495,
0.1452464908361435,
0.09216362982988358,
-0.020392004400491714,
-0.0022196341305971146,
-0.09898780286312103,
0.03927973657846451,
0.12461193650960922,
0.1647031456232071,
-0.032154280692338943,
0.017773505300283432,
0.04552295804023743,
0.017507968470454216,
0.008255558088421822,
0.1172267273068428,
0.08505405485630035,
0.20767830312252045,
-0.06384696811437607,
0.1279454380273819,
-0.017799995839595795,
0.005616375710815191,
-0.09441259503364563,
0.06968099623918533,
-0.04575827717781067,
0.011508723720908165,
-0.034111108630895615,
0.0523979626595974,
-0.06680657714605331,
-0.15687867999076843,
-0.04431690275669098,
-0.10044514387845993,
-0.08242648094892502,
-0.05231102555990219,
-0.026945555582642555,
0.001177342375740409,
0.07807707041501999,
0.01676388829946518,
0.025352153927087784,
0.16108772158622742,
0.011400546878576279,
-0.0845847949385643,
-0.00314911431632936,
0.058411091566085815,
-0.041578397154808044,
0.12093829363584518,
0.016424326226115227,
0.05653652921319008,
0.05133156105875969,
0.017991362139582634,
-0.10947930812835693,
0.03526076301932335,
-0.028426283970475197,
-0.020167043432593346,
-0.015704814344644547,
0.10634955018758774,
0.005444597452878952,
0.034813228994607925,
-0.0008654367411509156,
-0.04822494834661484,
0.017748676240444183,
0.018336832523345947,
-0.028848662972450256,
-0.09810822457075119,
0.07155466824769974,
-0.08024970442056656,
0.11818090081214905,
0.18594804406166077,
-0.014901827089488506,
0.04121530428528786,
-0.09562108665704727,
-0.027648182585835457,
0.02776351012289524,
0.02538365311920643,
-0.03851934149861336,
-0.04233735799789429,
-0.011406800709664822,
0.04757772386074066,
0.013157883659005165,
-0.17672088742256165,
-0.06513818353414536,
0.0645667016506195,
-0.026784757152199745,
-0.05527525022625923,
0.07693318277597427,
0.030620045959949493,
0.058128148317337036,
-0.011732128448784351,
-0.08886764943599701,
0.007609293330460787,
0.08009792119264603,
-0.13848960399627686,
-0.05194861814379692
] |
null | null |
transformers
|
# doc2query/stackexchange-title-body-t5-small-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/stackexchange-title-body-t5-small-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=384, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-small](https://huggingface.co/google/t5-v1_1-small) for 321k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, question_body) from StackExchange.
|
{"language": "en", "license": "apache-2.0", "datasets": ["flax-sentence-embeddings/stackexchange_title_body_jsonl"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/stackexchange-title-body-t5-small-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"dataset:flax-sentence-embeddings/stackexchange_title_body_jsonl",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_body_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# doc2query/stackexchange-title-body-t5-small-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-small for 321k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, question_body) from StackExchange.
|
[
"# doc2query/stackexchange-title-body-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-small for 321k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, question_body) from StackExchange."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_body_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# doc2query/stackexchange-title-body-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-small for 321k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, question_body) from StackExchange."
] |
[
101,
266,
32,
92
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #dataset-flax-sentence-embeddings/stackexchange_title_body_jsonl #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# doc2query/stackexchange-title-body-t5-small-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-small for 321k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 384 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, question_body) from StackExchange."
] |
[
-0.05608794093132019,
0.14600585401058197,
-0.006923680659383535,
0.09132859855890274,
0.11877860128879547,
-0.003508877707645297,
0.01541055552661419,
0.13070745766162872,
-0.05637819319963455,
0.04227509722113609,
0.04000506177544594,
0.052285950630903244,
0.032736312597990036,
0.07468099892139435,
0.055745936930179596,
-0.19257286190986633,
-0.047205157577991486,
-0.046043090522289276,
0.09528106451034546,
0.09322991222143173,
0.06591051816940308,
-0.045076120644807816,
0.07484140992164612,
-0.014545922167599201,
-0.11409753561019897,
0.03517803177237511,
-0.022605525329709053,
-0.005306093953549862,
0.031537976115942,
0.028998812660574913,
0.05195030942559242,
-0.026024658232927322,
0.02760154940187931,
-0.13541431725025177,
0.02112889662384987,
0.07463975995779037,
-0.05633276328444481,
0.06189436465501785,
0.06026771292090416,
-0.021797996014356613,
0.09722869843244553,
-0.09825697541236877,
0.018052509054541588,
0.0783093199133873,
-0.08365144580602646,
-0.09118738025426865,
-0.09960320591926575,
0.013083286583423615,
0.03775468096137047,
0.10831385105848312,
-0.022691642865538597,
0.03852079063653946,
-0.019571630284190178,
0.0672866478562355,
0.11396505683660507,
-0.14423979818820953,
-0.040921900421381,
0.0034099966287612915,
-0.008191647939383984,
0.005122019909322262,
-0.054747845977544785,
-0.0855708047747612,
-0.04870763421058655,
0.03956099972128868,
0.08694831281900406,
-0.01291253324598074,
-0.03655822202563286,
-0.011579493060708046,
-0.10754618793725967,
-0.03240048512816429,
0.06395290791988373,
-0.00941444467753172,
-0.08660824596881866,
-0.11844788491725922,
-0.05559544637799263,
0.01532162819057703,
0.019984900951385498,
-0.049547597765922546,
0.016728617250919342,
0.04391734302043915,
0.045977942645549774,
-0.098860964179039,
-0.10688227415084839,
-0.03410046920180321,
-0.03781779482960701,
-0.015660276636481285,
0.04898720979690552,
0.031939178705215454,
-0.0316803976893425,
0.0925147533416748,
-0.11849778890609741,
-0.08161091059446335,
-0.04907967522740364,
-0.07690179347991943,
-0.04849005118012428,
-0.002471375744789839,
-0.06419380754232407,
-0.17661790549755096,
-0.0019301229622215033,
0.08264054358005524,
-0.06771966069936752,
-0.001917660585604608,
-0.026001397520303726,
0.019015980884432793,
0.02681378461420536,
0.11003542691469193,
-0.04343996196985245,
0.019697437062859535,
0.08118819445371628,
0.021355101838707924,
-0.03383605554699898,
0.010878429748117924,
-0.07369162887334824,
-0.0656464695930481,
0.03242919594049454,
0.011798102408647537,
0.037548232823610306,
0.057835403829813004,
-0.049678534269332886,
-0.05008137971162796,
0.08569637686014175,
-0.1205616444349289,
-0.00976545363664627,
0.009944342076778412,
-0.08871188014745712,
0.0191086083650589,
0.13732092082500458,
-0.048674702644348145,
-0.11232754588127136,
-0.025805361568927765,
-0.030957622453570366,
0.013475767336785793,
-0.11359641700983047,
-0.06472469866275787,
0.029825598001480103,
-0.10336148738861084,
-0.05780810862779617,
-0.08918338268995285,
-0.16558270156383514,
-0.08473067730665207,
0.04609307274222374,
-0.051377397030591965,
-0.0017877473728731275,
-0.06063937768340111,
-0.037179429084062576,
-0.036033328622579575,
0.01128414273262024,
-0.049240365624427795,
0.00847931019961834,
0.05897026136517525,
-0.010904951021075249,
0.04508324712514877,
-0.04353483393788338,
0.03291022777557373,
-0.04479232057929039,
-0.02540937066078186,
-0.12094162404537201,
0.18031440675258636,
0.013386107049882412,
-0.008882547728717327,
-0.09899252653121948,
-0.002825577277690172,
-0.043949007987976074,
-0.011837530881166458,
0.05468205735087395,
0.15350131690502167,
-0.25942525267601013,
-0.01784067414700985,
0.21808144450187683,
-0.06356730312108994,
-0.03558812662959099,
0.09997796267271042,
-0.063740573823452,
0.13592413067817688,
0.12038460373878479,
0.05446229875087738,
0.1426040232181549,
-0.09061314910650253,
-0.10385088622570038,
-0.07906915247440338,
-0.10742069035768509,
0.03747176378965378,
0.05883694812655449,
-0.041465118527412415,
0.11492089182138443,
0.04180850461125374,
0.04790765792131424,
-0.004164706449955702,
-0.01626911386847496,
-0.04918522387742996,
-0.015169183723628521,
-0.024624580517411232,
0.017006859183311462,
-0.023985132575035095,
0.006428802851587534,
0.04794352129101753,
-0.12800763547420502,
0.007209588773548603,
0.03579578921198845,
-0.045373573899269104,
0.04426082968711853,
-0.0533214770257473,
-0.03229230269789696,
0.014695464633405209,
0.039035435765981674,
-0.1569974422454834,
-0.05371222645044327,
0.025705493986606598,
0.03755475953221321,
0.12401299923658371,
-0.14136609435081482,
0.06128263846039772,
0.042300742119550705,
-0.02506832219660282,
0.015357697382569313,
-0.08091948926448822,
-0.04302563890814781,
-0.07321082800626755,
-0.07979925721883774,
-0.06667037308216095,
-0.025650324299931526,
0.05679153650999069,
-0.0712677538394928,
0.04937851428985596,
0.048887837678194046,
0.01814684085547924,
0.009299005381762981,
-0.05434565246105194,
0.047703295946121216,
0.0010573903564363718,
0.032502394169569016,
-0.011056451126933098,
0.056898780167102814,
0.012225991114974022,
-0.04914399981498718,
0.007922899909317493,
-0.05961071327328682,
-0.12960055470466614,
0.05958160385489464,
0.026408670470118523,
-0.08062873035669327,
-0.13963103294372559,
-0.03664655238389969,
-0.04084665700793266,
-0.06299016624689102,
-0.06667417287826538,
0.19237570464611053,
0.029082490131258965,
0.08765427023172379,
-0.059951506555080414,
-0.04160070791840553,
0.0059494748711586,
0.007442701607942581,
0.0483013316988945,
0.03668399527668953,
0.001071036676876247,
-0.0650542601943016,
0.04062053933739662,
0.008604020811617374,
-0.03989872336387634,
0.17262105643749237,
0.024571865797042847,
-0.12000209093093872,
-0.026642022654414177,
0.042935192584991455,
0.033006373792886734,
-0.024052191525697708,
-0.029439544305205345,
0.008082380518317223,
0.022625194862484932,
-0.003253959584981203,
0.045171551406383514,
-0.07714246958494186,
0.0449586845934391,
0.020271718502044678,
-0.029628949239850044,
-0.044393230229616165,
-0.009563963860273361,
0.04479524493217468,
0.04872727021574974,
0.021976670250296593,
0.11008929461240768,
0.03994148224592209,
-0.05423002317547798,
-0.11196749657392502,
0.14724034070968628,
-0.0876903384923935,
-0.13690170645713806,
-0.12819278240203857,
0.030764006078243256,
-0.0423157773911953,
0.007491340860724449,
0.04460317641496658,
-0.08757881075143814,
0.00448291702196002,
-0.05799300968647003,
0.04786253720521927,
0.03364213928580284,
0.005242208484560251,
-0.15083272755146027,
0.06403618305921555,
0.018114393576979637,
-0.11263731867074966,
0.013110593892633915,
-0.03270982578396797,
-0.16654689610004425,
0.07396719604730606,
0.00022505394008476287,
0.05390072986483574,
0.01691225729882717,
-0.0158669576048851,
0.011088322848081589,
-0.018622757866978645,
0.20672619342803955,
-0.024201177060604095,
0.09859555959701538,
0.15870927274227142,
-0.04684610292315483,
0.10689123719930649,
0.048451390117406845,
0.021182676777243614,
-0.04628387466073036,
0.056550104171037674,
0.04264964163303375,
-0.03930524364113808,
-0.22900108993053436,
-0.0006394880474545062,
-0.04583978280425072,
-0.011780383996665478,
0.07048729807138443,
0.03420007601380348,
-0.015421740710735321,
0.04494359716773033,
-0.029713572934269905,
0.01750539056956768,
0.041099678725004196,
0.06824244558811188,
0.055918142199516296,
0.004672527313232422,
0.06691491603851318,
-0.03088257648050785,
0.001552478177472949,
0.06778961420059204,
0.04964670538902283,
0.21015116572380066,
-0.08446633070707321,
0.1402628868818283,
0.08792994171380997,
0.12017788738012314,
0.02496381662786007,
0.07761771231889725,
-0.04922521859407425,
0.05227179080247879,
-0.04556521400809288,
-0.09407807886600494,
-0.06908144801855087,
0.07588641345500946,
0.002166990889236331,
-0.014256707392632961,
-0.037391141057014465,
0.008212903514504433,
0.04016738012433052,
0.16744866967201233,
-0.06332525610923767,
-0.16083285212516785,
-0.03867829218506813,
-0.053685788065195084,
-0.01763438992202282,
-0.02383481152355671,
-0.008385577239096165,
0.05712532252073288,
-0.10029196739196777,
-0.0153973288834095,
-0.04185960814356804,
0.09058717638254166,
-0.03484975919127464,
0.008730475790798664,
0.048856932669878006,
0.09511283785104752,
0.04458121955394745,
0.04583584889769554,
-0.17924053966999054,
0.02681695483624935,
0.03648645058274269,
0.06489264965057373,
-0.0763179212808609,
0.062440644949674606,
0.03273197263479233,
-0.11102276295423508,
0.13342300057411194,
0.003071932587772608,
0.07185382395982742,
-0.010507773607969284,
-0.11267049610614777,
0.01596495695412159,
0.09531603753566742,
-0.05408092215657234,
0.13234303891658783,
0.023966046050190926,
0.00048151318333111703,
-0.01682106964290142,
0.16134671866893768,
-0.06110145524144173,
-0.19943566620349884,
0.019273197278380394,
-0.07572294026613235,
0.0522942990064621,
0.002781245857477188,
-0.020718710497021675,
0.045156314969062805,
0.1371539831161499,
-0.15552765130996704,
-0.10570696741342545,
-0.1006975993514061,
0.04188794642686844,
0.1016792580485344,
-0.09545139223337173,
0.02109280414879322,
0.02488878183066845,
0.030923087149858475,
-0.03907990828156471,
-0.12080725282430649,
0.035375356674194336,
-0.026835523545742035,
-0.06419117748737335,
-0.043571073561906815,
0.1083967462182045,
0.08690566569566727,
0.03183711692690849,
-0.00034347933251410723,
0.03306284174323082,
-0.054210126399993896,
-0.08697197586297989,
0.041198547929525375,
0.2076626420021057,
0.13017627596855164,
0.04537893086671829,
-0.12602141499519348,
-0.04350389167666435,
-0.09486513584852219,
0.023323282599449158,
0.11809854209423065,
0.12001461535692215,
-0.05826949700713158,
0.10801123827695847,
0.06563884019851685,
-0.10266144573688507,
-0.19522665441036224,
-0.033644899725914,
0.15336234867572784,
0.025099100545048714,
0.00832256767898798,
-0.27456575632095337,
0.08315759152173996,
0.08140309154987335,
0.017149794846773148,
0.027104031294584274,
-0.1923363357782364,
-0.10398688167333603,
0.00835298839956522,
0.044706664979457855,
0.04574275761842728,
-0.13029588758945465,
-0.03610561043024063,
0.03412230312824249,
-0.007656012661755085,
0.20872488617897034,
-0.2024250626564026,
0.0688927099108696,
-0.01257030013948679,
0.0472516268491745,
0.0506959892809391,
-0.02122117392718792,
-0.005188826937228441,
0.026813898235559464,
0.021783338859677315,
-0.0227712020277977,
0.03212673217058182,
0.02809756249189377,
-0.05998421460390091,
0.1537240743637085,
0.05958569794893265,
0.04200509935617447,
-0.11934560537338257,
-0.03380768001079559,
-0.09103904664516449,
-0.013085801154375076,
-0.032159674912691116,
-0.03705725818872452,
0.008377994410693645,
0.045349229127168655,
0.10162491351366043,
-0.007724490482360125,
-0.012167387641966343,
-0.05728942155838013,
0.044052157551050186,
0.28804150223731995,
0.10266301035881042,
-0.013659949414432049,
-0.09293919801712036,
0.006309971679002047,
-0.006365190725773573,
0.06058388575911522,
-0.13172049820423126,
0.01667451672255993,
0.06561345607042313,
0.012991837225854397,
0.07782413810491562,
0.05614775791764259,
-0.11636340618133545,
-0.002453719498589635,
0.05969218537211418,
-0.10470244288444519,
-0.06692440062761307,
-0.0678437277674675,
-0.03523761406540871,
-0.16378392279148102,
-0.027835635468363762,
0.0923912450671196,
-0.06207885965704918,
-0.04488810896873474,
0.006494977045804262,
0.03000509925186634,
-0.011694084852933884,
0.09642579406499863,
0.06391347199678421,
0.005895494483411312,
-0.06607519835233688,
0.0982956811785698,
0.08273601531982422,
-0.0713362991809845,
0.05875854194164276,
0.19943614304065704,
-0.08190542459487915,
-0.037470120936632156,
-0.06825201213359833,
0.029615886509418488,
-0.04681704193353653,
0.026795407757163048,
-0.00992498081177473,
0.01649557612836361,
0.06578885018825531,
0.08639012277126312,
0.02777084708213806,
0.06835354119539261,
-0.03341997042298317,
0.03542336821556091,
-0.07865532487630844,
0.07860910892486572,
0.0038689824286848307,
-0.015667149797081947,
-0.015143762342631817,
0.08149643987417221,
-0.021870914846658707,
0.05718877911567688,
-0.008198300376534462,
-0.03816424310207367,
-0.09442869573831558,
-0.002483789110556245,
-0.01712833158671856,
0.049505796283483505,
-0.054620709270238876,
-0.010523056611418724,
-0.0397086963057518,
0.04556576907634735,
0.06910421699285507,
0.05017435923218727,
-0.02419566921889782,
-0.07424932718276978,
-0.030358262360095978,
0.060549523681402206,
-0.12365313619375229,
0.005897508934140205,
0.024303331971168518,
-0.05858825519680977,
0.10273623466491699,
0.06966332346200943,
-0.07014677673578262,
0.01861131563782692,
-0.020656567066907883,
-0.05148699879646301,
-0.010228583589196205,
-0.005661179311573505,
-0.0075752451084554195,
-0.06114695966243744,
0.010143190622329712,
-0.007538873236626387,
-0.03635828569531441,
-0.014519392512738705,
0.11619052290916443,
-0.10592808574438095,
0.0325348898768425,
0.053522903472185135,
0.022799896076321602,
-0.11619149893522263,
0.023828893899917603,
0.016474071890115738,
0.07822687923908234,
0.11000730842351913,
-0.055403657257556915,
0.06473582983016968,
-0.17020903527736664,
-0.014093512669205666,
0.02670053392648697,
0.03447950631380081,
-0.05603820085525513,
-0.070189468562603,
0.04935092106461525,
-0.010040688328444958,
0.04923146590590477,
-0.00495389336720109,
0.003288715612143278,
-0.0024554634001106024,
-0.03434676676988602,
-0.04821368679404259,
-0.010817504487931728,
0.03841733932495117,
-0.00041653349762782454,
-0.04164104163646698,
0.029567454010248184,
0.060809679329395294,
-0.0339922197163105,
0.023858150467276573,
0.09594669193029404,
0.07337156683206558,
0.14463219046592712,
0.028529219329357147,
0.06594535708427429,
0.049314700067043304,
-0.032676778733730316,
0.06361723691225052,
0.06206250935792923,
-0.044587019830942154,
-0.09015348553657532,
0.030384259298443794,
0.14262482523918152,
-0.07428841292858124,
0.14217929542064667,
0.05397218465805054,
-0.061553824692964554,
-0.08224478363990784,
-0.14977578818798065,
-0.040726881474256516,
-0.027317112311720848,
-0.013861395418643951,
-0.10871807485818863,
0.07703037559986115,
-0.006148624699562788,
-0.0123739093542099,
-0.017626265063881874,
0.10062792897224426,
-0.00002653210140124429,
-0.11508184671401978,
0.05405934900045395,
-0.02372835949063301,
0.03751838579773903,
0.0754302442073822,
0.05337176099419594,
0.061593662947416306,
0.0711258202791214,
0.08257186412811279,
0.08293437212705612,
0.0903824120759964,
-0.013158264569938183,
-0.10756143182516098,
-0.06796307861804962,
-0.025681335479021072,
0.04477521404623985,
0.031119756400585175,
0.1365264654159546,
0.04250519722700119,
-0.10427924990653992,
0.003372869687154889,
0.16998933255672455,
-0.04875321313738823,
-0.0752205029129982,
-0.13075368106365204,
0.11992843449115753,
0.09316711127758026,
-0.030551671981811523,
-0.017970578745007515,
-0.09914436936378479,
0.039256829768419266,
0.14090345799922943,
0.1847306489944458,
-0.03153480589389801,
0.019440609961748123,
0.029001666232943535,
0.018886733800172806,
0.007837126962840557,
0.1109188124537468,
0.07812654972076416,
0.23725919425487518,
-0.057948458939790726,
0.12167265266180038,
-0.0054627167992293835,
-0.012808550149202347,
-0.08021146059036255,
0.0425373800098896,
-0.04351396486163139,
0.024007463827729225,
-0.035694267600774765,
0.061896972358226776,
-0.08782202750444412,
-0.1180025264620781,
-0.0522693432867527,
-0.09713860601186752,
-0.07649826258420944,
-0.05773220583796501,
-0.0280681811273098,
0.020116426050662994,
0.06948650628328323,
0.027264220640063286,
0.03184371441602707,
0.14274008572101593,
0.0019815939012914896,
-0.09858683496713638,
-0.006556395906955004,
0.058286845684051514,
-0.03564639762043953,
0.11113828420639038,
0.022919384762644768,
0.07077270746231079,
0.05267135053873062,
0.018286658450961113,
-0.11268966645002365,
0.01968265511095524,
-0.033743374049663544,
-0.04077818617224693,
0.003629044396802783,
0.14079907536506653,
0.004360937513411045,
0.040450043976306915,
0.0024700185749679804,
-0.05179635062813759,
-0.004122166894376278,
0.000041435661842115223,
-0.027206357568502426,
-0.10447287559509277,
0.08519213646650314,
-0.08321308344602585,
0.1237979456782341,
0.19802500307559967,
-0.01936526782810688,
0.04162028804421425,
-0.09393702447414398,
-0.016208050772547722,
0.015179543755948544,
0.03060801327228546,
-0.053277209401130676,
-0.06745388358831406,
-0.007237937767058611,
0.08971592038869858,
0.021145567297935486,
-0.1829102337360382,
-0.05983272194862366,
0.06136080622673035,
-0.015845004469156265,
-0.05025734007358551,
0.07916359603404999,
0.02944796532392502,
0.048857469111680984,
-0.013518096879124641,
-0.08670134842395782,
-0.0025149176362901926,
0.06580445170402527,
-0.1368766576051712,
-0.05069730430841446
] |
null | null |
transformers
|
# doc2query/yahoo_answers-t5-base-v1
This is a [doc2query](https://arxiv.org/abs/1904.08375) model based on T5 (also known as [docT5query](https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf)).
It can be used for:
- **Document expansion**: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our [BEIR](https://arxiv.org/abs/2104.08663) paper we showed that BM25+docT5query is a powerful search engine. In the [BEIR repository](https://github.com/UKPLab/beir) we have an example how to use docT5query with Pyserini.
- **Domain Specific Training Data Generation**: It can be used to generate training data to learn an embedding model. On [SBERT.net](https://www.sbert.net/examples/unsupervised_learning/query_generation/README.html) we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
```python
from transformers import T5Tokenizer, T5ForConditionalGeneration
model_name = 'doc2query/yahoo_answers-t5-base-v1'
tokenizer = T5Tokenizer.from_pretrained(model_name)
model = T5ForConditionalGeneration.from_pretrained(model_name)
text = "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."
input_ids = tokenizer.encode(text, max_length=320, truncation=True, return_tensors='pt')
outputs = model.generate(
input_ids=input_ids,
max_length=64,
do_sample=True,
top_p=0.95,
num_return_sequences=5)
print("Text:")
print(text)
print("\nGenerated Queries:")
for i in range(len(outputs)):
query = tokenizer.decode(outputs[i], skip_special_tokens=True)
print(f'{i + 1}: {query}')
```
**Note:** `model.generate()` is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned [google/t5-v1_1-base](https://huggingface.co/google/t5-v1_1-base) for 111k training steps. For the training script, see the `train_script.py` in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, answer) pairs from [Yahoo Answers](https://huggingface.co/datasets/sentence-transformers/embedding-training-data).
|
{"language": "en", "license": "apache-2.0", "datasets": ["datasets/sentence-transformers/embedding-training-data"], "widget": [{"text": "Python is an interpreted, high-level and general-purpose programming language. Python's design philosophy emphasizes code readability with its notable use of significant whitespace. Its language constructs and object-oriented approach aim to help programmers write clear, logical code for small and large-scale projects."}]}
|
text2text-generation
|
doc2query/yahoo_answers-t5-base-v1
|
[
"transformers",
"pytorch",
"t5",
"text2text-generation",
"en",
"arxiv:1904.08375",
"arxiv:2104.08663",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1904.08375",
"2104.08663"
] |
[
"en"
] |
TAGS
#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
|
# doc2query/yahoo_answers-t5-base-v1
This is a doc2query model based on T5 (also known as docT5query).
It can be used for:
- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.
- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.
## Usage
Note: 'model.generate()' is non-deterministic. It produces different queries each time you run it.
## Training
This model fine-tuned google/t5-v1_1-base for 111k training steps. For the training script, see the 'train_script.py' in this repository.
The input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces.
This model was trained on a (title, answer) pairs from Yahoo Answers.
|
[
"# doc2query/yahoo_answers-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 111k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, answer) pairs from Yahoo Answers."
] |
[
"TAGS\n#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n",
"# doc2query/yahoo_answers-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.",
"## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.",
"## Training\r\nThis model fine-tuned google/t5-v1_1-base for 111k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, answer) pairs from Yahoo Answers."
] |
[
79,
261,
32,
88
] |
[
"passage: TAGS\n#transformers #pytorch #t5 #text2text-generation #en #arxiv-1904.08375 #arxiv-2104.08663 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n# doc2query/yahoo_answers-t5-base-v1\r\n\r\nThis is a doc2query model based on T5 (also known as docT5query).\r\n\r\nIt can be used for:\r\n- Document expansion: You generate for your paragraphs 20-40 queries and index the paragraphs and the generates queries in a standard BM25 index like Elasticsearch, OpenSearch, or Lucene. The generated queries help to close the lexical gap of lexical search, as the generate queries contain synonyms. Further, it re-weights words giving important words a higher weight even if they appear seldomn in a paragraph. In our BEIR paper we showed that BM25+docT5query is a powerful search engine. In the BEIR repository we have an example how to use docT5query with Pyserini.\r\n- Domain Specific Training Data Generation: It can be used to generate training data to learn an embedding model. On URL we have an example how to use the model to generate (query, text) pairs for a given collection of unlabeled texts. These pairs can then be used to train powerful dense embedding models.## Usage\r\n\r\n\r\nNote: 'model.generate()' is non-deterministic. It produces different queries each time you run it.## Training\r\nThis model fine-tuned google/t5-v1_1-base for 111k training steps. For the training script, see the 'train_script.py' in this repository.\r\n\r\nThe input-text was truncated to 320 word pieces. Output text was generated up to 64 word pieces. \r\n\r\nThis model was trained on a (title, answer) pairs from Yahoo Answers."
] |
[
-0.03584110364317894,
0.13330486416816711,
-0.0065181623212993145,
0.0958113744854927,
0.10420619696378708,
0.020641224458813667,
0.10957223176956177,
0.11880302429199219,
-0.0699642226099968,
0.0026967714074999094,
0.07110507786273956,
0.03320774435997009,
-0.019851157441735268,
0.07509993016719818,
0.04727230593562126,
-0.2105334997177124,
-0.00823888462036848,
-0.03229453042149544,
0.04140602424740791,
0.09074421226978302,
0.074671670794487,
-0.026643311604857445,
0.054681506007909775,
-0.0026756753213703632,
-0.1329401135444641,
0.04638047516345978,
-0.028075728565454483,
-0.007381385192275047,
0.06893476843833923,
0.08115579932928085,
0.08603838831186295,
-0.04234348610043526,
0.056882597506046295,
-0.16121047735214233,
0.020132441073656082,
0.0770522952079773,
-0.033703915774822235,
0.0652705505490303,
0.01605931855738163,
-0.029847411438822746,
0.18193568289279938,
-0.08012622594833374,
0.035509590059518814,
0.05529683083295822,
-0.08405454456806183,
-0.078668013215065,
-0.07429177314043045,
0.07021025568246841,
0.06635542958974838,
0.0922912210226059,
0.006086007226258516,
0.0334567129611969,
-0.05234125629067421,
0.0629492923617363,
0.17822252213954926,
-0.11768447607755661,
-0.03255237638950348,
0.034498367458581924,
-0.04365360736846924,
0.049481187015771866,
0.009217318147420883,
-0.035746194422245026,
-0.021060943603515625,
0.006177965551614761,
0.1084117665886879,
-0.005467244889587164,
-0.0558440238237381,
-0.0026912172324955463,
-0.108273446559906,
-0.05998662859201431,
0.16199925541877747,
-0.008039357140660286,
-0.058024242520332336,
-0.07129644602537155,
-0.035579830408096313,
0.11914157122373581,
0.02384090982377529,
-0.006402555853128433,
0.02185836434364319,
0.024674836546182632,
0.0240004975348711,
-0.14785172045230865,
-0.08119548857212067,
-0.08507198840379715,
0.025201866403222084,
0.007791846990585327,
0.02066194824874401,
0.019665397703647614,
-0.08181466162204742,
0.08733311295509338,
-0.1055820882320404,
-0.016357624903321266,
-0.04053608700633049,
-0.0704704076051712,
-0.062971331179142,
0.00487256795167923,
-0.07944358885288239,
-0.15503260493278503,
-0.023332389071583748,
0.08895353227853775,
0.004998517222702503,
0.0020093428902328014,
0.01837611384689808,
0.02673913538455963,
-0.008892898447811604,
0.13793741166591644,
-0.04262721911072731,
-0.022304344922304153,
0.06211785227060318,
-0.028931716457009315,
-0.014851266518235207,
-0.001785576925612986,
-0.09567200392484665,
-0.08455516397953033,
0.05248143896460533,
-0.0052446890622377396,
-0.013417487032711506,
0.07748684287071228,
-0.05473591759800911,
-0.014033937826752663,
0.037599727511405945,
-0.09595082700252533,
0.0392613485455513,
0.028138242661952972,
-0.06982744485139847,
-0.03919200226664543,
0.19189700484275818,
-0.05033613368868828,
-0.11506230384111404,
0.010320558212697506,
-0.04080226272344589,
0.008016543462872505,
-0.11653130501508713,
-0.08421435207128525,
0.01759376935660839,
-0.1257576048374176,
-0.0454975925385952,
-0.06760463118553162,
-0.3022785186767578,
-0.017373984679579735,
0.06682842969894409,
-0.042208291590213776,
-0.012788879685103893,
-0.07303006947040558,
-0.013046960346400738,
-0.05046248063445091,
0.03354199603199959,
-0.06651752442121506,
0.012903203256428242,
0.019131995737552643,
-0.029803844168782234,
0.061816245317459106,
-0.07014236599206924,
0.01618299074470997,
-0.013345612213015556,
0.0015184947988018394,
-0.18254931271076202,
0.2032073587179184,
0.021999649703502655,
-0.005010170862078667,
-0.09382396936416626,
-0.014469314366579056,
-0.061832983046770096,
0.020033709704875946,
0.061383992433547974,
0.15553240478038788,
-0.17649802565574646,
-0.05070814862847328,
0.1870226263999939,
-0.08211743086576462,
-0.03467387333512306,
0.11231043189764023,
-0.08644657582044601,
0.11926838010549545,
0.12103208899497986,
0.0625019446015358,
0.15557628870010376,
-0.07495308667421341,
-0.032155562192201614,
-0.0038351588882505894,
-0.09864801913499832,
-0.05298309028148651,
0.05752699822187424,
-0.04617033526301384,
0.01782429777085781,
0.020484687760472298,
-0.05343959107995033,
-0.001630292972549796,
-0.0077288453467190266,
-0.03458549454808235,
-0.0373840406537056,
-0.03422659635543823,
0.022801218554377556,
-0.035092782229185104,
0.025853116065263748,
0.05648870766162872,
-0.07989230006933212,
0.19400787353515625,
0.06842374801635742,
-0.08848033100366592,
0.05080045387148857,
-0.05115543305873871,
0.02866450510919094,
-0.05977834388613701,
0.029922645539045334,
-0.16538558900356293,
-0.038177240639925,
-0.009352610446512699,
0.024737538769841194,
0.16466976702213287,
-0.05151020362973213,
0.056952521204948425,
0.03135192394256592,
-0.02417801320552826,
0.022465264424681664,
-0.0766889676451683,
-0.008537652902305126,
-0.07595603913068771,
-0.1163477897644043,
-0.08237780630588531,
-0.013789688237011433,
0.1142854318022728,
-0.15828508138656616,
0.059201449155807495,
0.006520550698041916,
0.027863919734954834,
0.033434782177209854,
-0.03208719566464424,
0.04266437515616417,
-0.0006570421974174678,
0.011592099443078041,
-0.01998775638639927,
0.06689457595348358,
0.03510008379817009,
-0.11415807902812958,
-0.02700122445821762,
-0.025763051584362984,
-0.1713971346616745,
0.033901702612638474,
-0.0016251244815066457,
-0.06445547938346863,
-0.10715938359498978,
-0.021597158163785934,
-0.015621263533830643,
-0.0921260192990303,
-0.10888100415468216,
0.17841112613677979,
0.039051417261362076,
0.12808232009410858,
-0.09313713759183884,
-0.07108499854803085,
-0.009144730865955353,
-0.06846797466278076,
0.012495541013777256,
0.07614610344171524,
0.04974622279405594,
-0.014409846626222134,
0.0338762067258358,
-0.017255449667572975,
-0.08558943122625351,
0.19259029626846313,
0.04656378924846649,
-0.08599452674388885,
-0.024290192872285843,
0.0017157429829239845,
0.023446762934327126,
0.02516242302954197,
-0.08344633877277374,
-0.014654607512056828,
-0.006866904441267252,
0.021239854395389557,
0.050413280725479126,
-0.11930213868618011,
0.030258435755968094,
0.05562812089920044,
-0.058494068682193756,
-0.06728973984718323,
-0.0442817397415638,
-0.01893753372132778,
0.05931064113974571,
0.007961017079651356,
0.060188621282577515,
0.05213252082467079,
-0.04532848298549652,
-0.14427152276039124,
0.19077056646347046,
-0.10419654846191406,
-0.09437856078147888,
-0.16292086243629456,
0.037292446941137314,
-0.05571652576327324,
0.015688259154558182,
0.01475366111844778,
-0.09028684347867966,
-0.002847496420145035,
-0.10174891352653503,
0.010002621449530125,
0.03450231999158859,
-0.0013855447759851813,
-0.037500590085983276,
0.04601877182722092,
0.043198488652706146,
-0.1359873116016388,
0.01809779554605484,
-0.03645654767751694,
-0.04276743903756142,
0.04127806797623634,
-0.09510048478841782,
0.08253660053014755,
0.08276312798261642,
-0.04957574978470802,
0.04074709862470627,
-0.0012615218292921782,
0.2361222803592682,
-0.009799222461879253,
0.0680370032787323,
0.13089260458946228,
-0.06596493721008301,
0.0788780152797699,
0.05282456800341606,
0.03338547423481941,
-0.02968384139239788,
0.03817228972911835,
0.040454667061567307,
-0.06338298320770264,
-0.2506980001926422,
-0.028288600966334343,
-0.052857257425785065,
-0.04002904146909714,
0.03162816911935806,
0.028307899832725525,
0.0051270704716444016,
0.022358108311891556,
0.0170549638569355,
0.076697438955307,
0.07504255324602127,
0.09173701703548431,
0.05339611694216728,
0.019540345296263695,
0.0310300812125206,
-0.057998739182949066,
0.01868336834013462,
0.036722611635923386,
0.09326232969760895,
0.25110286474227905,
-0.046313002705574036,
0.17545051872730255,
0.07414911687374115,
-0.021413197740912437,
0.024049075320363045,
0.06497303396463394,
-0.10381322354078293,
0.052670300006866455,
-0.04650561138987541,
-0.06748245656490326,
-0.03900471329689026,
0.08287116885185242,
-0.08694575726985931,
0.01666819117963314,
-0.04829514026641846,
0.06118802726268768,
0.07468340545892715,
0.1493632048368454,
-0.028513696044683456,
-0.1973057985305786,
-0.028529854491353035,
-0.009461824782192707,
0.00681731104850769,
0.013146132230758667,
-0.030951721593737602,
0.07085002958774567,
-0.06804007291793823,
-0.022864283993840218,
-0.0246565043926239,
0.10526608675718307,
-0.046674322336912155,
0.004080250859260559,
0.027927683666348457,
0.13223490118980408,
-0.010269283317029476,
0.05964510887861252,
-0.1671128123998642,
0.11402061581611633,
0.025202644988894463,
0.0990731492638588,
-0.05219727382063866,
0.041896820068359375,
-0.015821168199181557,
0.013843433931469917,
0.13028690218925476,
0.020755644887685776,
0.04651235044002533,
-0.037549614906311035,
-0.030928578227758408,
0.01855471171438694,
0.04717976599931717,
-0.005468760151416063,
0.12927719950675964,
-0.008131316863000393,
-0.001682863337919116,
0.026434771716594696,
0.11449349671602249,
-0.06499944627285004,
-0.21550633013248444,
-0.00437117787078023,
-0.045988909900188446,
0.006507339421659708,
-0.00521690770983696,
-0.003176208818331361,
0.04985809326171875,
0.1376122683286667,
-0.16242487728595734,
-0.13799946010112762,
-0.10297141969203949,
0.02769775129854679,
0.05627293139696121,
-0.07802008837461472,
0.0016333850799128413,
0.04104231670498848,
0.10578848421573639,
-0.029903138056397438,
-0.0671699270606041,
-0.0032179043628275394,
-0.0663096010684967,
-0.09905143082141876,
-0.05391447991132736,
0.055049601942300797,
0.09772800654172897,
0.04610107094049454,
0.04948955774307251,
0.00457333168014884,
-0.019281525164842606,
-0.098643958568573,
-0.009664075449109077,
0.1482597142457962,
0.08969104290008545,
0.04520561173558235,
-0.12308071553707123,
-0.011734338477253914,
-0.0860627293586731,
0.0019916780292987823,
0.12007419764995575,
0.14055664837360382,
-0.07849234342575073,
0.06494936347007751,
0.1363176703453064,
-0.14541533589363098,
-0.14684289693832397,
0.0017569204792380333,
0.09042803943157196,
0.01514651533216238,
-0.03732285648584366,
-0.2837640941143036,
0.0658179372549057,
0.07637009769678116,
0.026824500411748886,
-0.14636348187923431,
-0.3238693177700043,
-0.14037542045116425,
0.03349054232239723,
0.07299752533435822,
0.10209430754184723,
-0.11296264827251434,
-0.023414181545376778,
-0.02342289686203003,
0.05293401703238487,
0.18898773193359375,
-0.1996866762638092,
0.0471673458814621,
0.01980147324502468,
0.06015253812074661,
0.039870791137218475,
-0.011464297771453857,
0.02188567817211151,
0.043238136917352676,
0.050196871161460876,
-0.010485202074050903,
0.03408697992563248,
0.01162401307374239,
-0.0763864666223526,
0.16035431623458862,
-0.0065917642787098885,
0.06060546636581421,
-0.13933703303337097,
-0.048406586050987244,
-0.07151146233081818,
0.004031087271869183,
-0.033172138035297394,
-0.04829821735620499,
0.0068204510025680065,
0.014281165786087513,
0.08582325279712677,
0.009000898338854313,
0.007310168351978064,
-0.07074323296546936,
0.027845555916428566,
0.12695226073265076,
0.14749886095523834,
-0.00383324408903718,
-0.10058730095624924,
0.02500656433403492,
-0.004318248014897108,
0.07596635818481445,
-0.13953831791877747,
0.029761413112282753,
0.038203272968530655,
0.04097631573677063,
0.17107602953910828,
0.058851271867752075,
-0.11548961699008942,
0.008822849951684475,
0.049318842589855194,
-0.12062399834394455,
-0.08088734745979309,
-0.007712635677307844,
-0.01167976576834917,
-0.1351555734872818,
0.0320143848657608,
0.11819277703762054,
-0.10760417580604553,
-0.022184625267982483,
0.01673140935599804,
0.02756194956600666,
-0.04400497302412987,
0.10984795540571213,
0.0983920469880104,
0.013266329653561115,
-0.03413398563861847,
0.11322551965713501,
0.05724992975592613,
0.01354724820703268,
0.07486574351787567,
0.14685994386672974,
-0.08493844419717789,
-0.02659284882247448,
-0.0808321088552475,
0.1208353266119957,
-0.06168585643172264,
-0.043087080121040344,
-0.04752536490559578,
-0.013351608999073505,
0.06572405248880386,
0.15579959750175476,
0.027745623141527176,
0.04807048290967941,
-0.07509681582450867,
0.01678226701915264,
-0.11436493694782257,
0.09552841633558273,
0.026085983961820602,
-0.024159014225006104,
-0.061763230711221695,
0.06197071820497513,
-0.007987882010638714,
0.015109559521079063,
-0.01101948507130146,
-0.019915323704481125,
-0.11512614041566849,
0.01188212726265192,
-0.07538212090730667,
0.030973609536886215,
-0.029001198709011078,
0.021907081827521324,
-0.021192841231822968,
0.03641515597701073,
0.06605299562215805,
0.0408979132771492,
-0.023352842777967453,
-0.035723183304071426,
0.0028815900441259146,
0.07141720503568649,
-0.15019501745700836,
-0.005132872611284256,
0.043767109513282776,
-0.06507718563079834,
0.10690072178840637,
0.030197761952877045,
-0.047792304307222366,
0.009391442872583866,
-0.09919875115156174,
0.014170021750032902,
-0.003041304647922516,
-0.008522212505340576,
-0.0038507815916091204,
-0.08246456831693649,
0.0016671890625730157,
-0.016418684273958206,
0.004574124701321125,
-0.01881706900894642,
0.07058979570865631,
-0.07460011541843414,
0.0792020857334137,
0.03624357655644417,
-0.008176189847290516,
-0.09977149218320847,
0.08280137181282043,
0.027540743350982666,
0.06915875524282455,
0.10797425359487534,
-0.04442533478140831,
0.036779116839170456,
-0.13988596200942993,
-0.0026315408758819103,
0.0667666420340538,
0.03625406324863434,
-0.11013323813676834,
-0.07246953994035721,
0.02388124167919159,
-0.019036225974559784,
0.06872088462114334,
0.028504744172096252,
0.0038085782434791327,
-0.005265302024781704,
-0.013290256261825562,
-0.06636369973421097,
-0.010706920176744461,
0.06371297687292099,
-0.022786574438214302,
0.004782458767294884,
0.03427640721201897,
0.041132934391498566,
-0.007052140776067972,
0.045094698667526245,
0.1547650396823883,
0.07520338147878647,
0.0729285180568695,
0.06539718806743622,
0.0632045790553093,
0.04018668085336685,
-0.04634805768728256,
0.010734910145401955,
0.06906294822692871,
0.00856051966547966,
-0.08387120813131332,
0.08976004272699356,
0.17571501433849335,
-0.10329455137252808,
0.08630722761154175,
0.07232514023780823,
-0.09648574143648148,
-0.1129552349448204,
-0.14715078473091125,
-0.020780952647328377,
-0.051203470677137375,
-0.01792263053357601,
-0.15102232992649078,
0.034062642604112625,
-0.05747336521744728,
0.015989212319254875,
0.002236650325357914,
0.06513034552335739,
0.039063483476638794,
-0.061214033514261246,
0.011889586225152016,
-0.03575411066412926,
0.029672248288989067,
0.022157322615385056,
0.036103755235672,
0.08911958336830139,
0.027687158435583115,
0.039612527936697006,
0.09124263375997543,
0.11455368250608444,
-0.02614840306341648,
-0.04035373404622078,
-0.07012184709310532,
-0.05970616266131401,
0.031779829412698746,
0.03813334181904793,
0.14901630580425262,
0.0425068661570549,
-0.10775796324014664,
-0.0072539630346000195,
0.16826759278774261,
-0.04358448460698128,
-0.05088981240987778,
-0.13666246831417084,
0.15204647183418274,
0.02589762769639492,
-0.00499514676630497,
-0.041108570992946625,
-0.09916327148675919,
0.020022662356495857,
0.1971273571252823,
0.1508924663066864,
-0.04619883373379707,
-0.007798054255545139,
-0.01857837848365307,
0.014768926426768303,
-0.0016677789390087128,
0.11742693930864334,
0.0013483120128512383,
0.24332712590694427,
-0.02438431605696678,
0.1468982994556427,
-0.020023532211780548,
0.005053312052041292,
-0.04654452204704285,
0.04884082078933716,
-0.03604855760931969,
-0.011583521030843258,
-0.01932624727487564,
0.059248268604278564,
-0.1439155787229538,
-0.2141830027103424,
-0.07681962102651596,
-0.06157596409320831,
-0.06080962345004082,
-0.025303712114691734,
-0.014623312279582024,
-0.008595033548772335,
0.0829971507191658,
-0.005325970705598593,
0.0017788393888622522,
0.1721319854259491,
-0.01008039154112339,
-0.10386785864830017,
-0.05958816036581993,
0.047791797667741776,
0.017061274498701096,
0.13157467544078827,
0.008031695149838924,
0.08346204459667206,
0.06993856281042099,
-0.018650690093636513,
-0.1406967043876648,
0.021195154637098312,
-0.020821742713451385,
0.03489407151937485,
0.0014552220236510038,
0.07547338306903839,
-0.018853774294257164,
0.06649399548768997,
0.022548286244273186,
-0.017423249781131744,
0.0004072229203302413,
0.009301353245973587,
-0.07393267750740051,
-0.11001353710889816,
0.04600993171334267,
-0.07342861592769623,
0.1133112907409668,
0.20068411529064178,
-0.05088983103632927,
0.007981175556778908,
-0.05249396339058876,
0.005420489236712456,
0.013482226990163326,
0.011666378937661648,
-0.03940153494477272,
-0.06249585747718811,
0.0048132529482245445,
0.06315841525793076,
0.006187269929796457,
-0.1620595008134842,
-0.06574436277151108,
0.042025383561849594,
-0.02780037559568882,
0.01192491129040718,
0.06383860856294632,
0.11966017633676529,
0.08397341519594193,
-0.021701961755752563,
-0.043111465871334076,
-0.022430213168263435,
0.077828548848629,
-0.09944940358400345,
-0.06321360915899277
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# bert-base-uncased-finetuned-swag
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the swag dataset.
It achieves the following results on the evaluation set:
- Loss: 0.6045
- Accuracy: 0.7960
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.7494 | 1.0 | 4597 | 0.5942 | 0.7716 |
| 0.3499 | 2.0 | 9194 | 0.6045 | 0.7960 |
### Framework versions
- Transformers 4.15.0
- Pytorch 1.10.0+cu111
- Datasets 1.17.0
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["swag"], "metrics": ["accuracy"], "model-index": [{"name": "bert-base-uncased-finetuned-swag", "results": []}]}
|
multiple-choice
|
domdomreloaded/bert-base-uncased-finetuned-swag
|
[
"transformers",
"pytorch",
"tensorboard",
"bert",
"multiple-choice",
"generated_from_trainer",
"dataset:swag",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #bert #multiple-choice #generated_from_trainer #dataset-swag #license-apache-2.0 #endpoints_compatible #region-us
|
bert-base-uncased-finetuned-swag
================================
This model is a fine-tuned version of bert-base-uncased on the swag dataset.
It achieves the following results on the evaluation set:
* Loss: 0.6045
* Accuracy: 0.7960
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.15.0
* Pytorch 1.10.0+cu111
* Datasets 1.17.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #bert #multiple-choice #generated_from_trainer #dataset-swag #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
[
54,
98,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #bert #multiple-choice #generated_from_trainer #dataset-swag #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.15.0\n* Pytorch 1.10.0+cu111\n* Datasets 1.17.0\n* Tokenizers 0.10.3"
] |
[
-0.10837743431329727,
0.051923561841249466,
-0.0017657753778621554,
0.1133657842874527,
0.17222857475280762,
0.041135333478450775,
0.12002521753311157,
0.11124030500650406,
-0.0993083268404007,
0.017644282430410385,
0.11762633919715881,
0.18611492216587067,
0.008031107485294342,
0.06217360869050026,
-0.03735630214214325,
-0.24296651780605316,
-0.007047482766211033,
0.03759719058871269,
-0.09741503745317459,
0.13771991431713104,
0.07940202206373215,
-0.1489812582731247,
0.07936003059148788,
0.015271487645804882,
-0.21219167113304138,
0.015562402084469795,
0.007887366227805614,
-0.042227257043123245,
0.15521030128002167,
0.010150864720344543,
0.12875592708587646,
0.0018123690970242023,
0.07352796196937561,
-0.1949714571237564,
0.013598200865089893,
0.04683614522218704,
0.006978578399866819,
0.08073164522647858,
0.05379624292254448,
-0.004084058105945587,
0.129921093583107,
-0.07045944035053253,
0.050235215574502945,
0.025324860587716103,
-0.12506477534770966,
-0.2541145384311676,
-0.0925753116607666,
0.00585652980953455,
0.05300937220454216,
0.10220030695199966,
-0.000557545805349946,
0.14550764858722687,
-0.10389645397663116,
0.08288982510566711,
0.2540821135044098,
-0.29518088698387146,
-0.06763876974582672,
0.03581802919507027,
0.020692387595772743,
0.07500319182872772,
-0.1075119823217392,
-0.021495003253221512,
0.04305710271000862,
0.048272401094436646,
0.12425544857978821,
-0.039276380091905594,
-0.09811851382255554,
0.04916452243924141,
-0.14724677801132202,
-0.02835978753864765,
0.13470079004764557,
0.04741582274436951,
-0.02984723262488842,
-0.0302896648645401,
-0.0510232150554657,
-0.12891395390033722,
-0.03149854764342308,
0.0045574624091386795,
0.047333553433418274,
-0.037384502589702606,
-0.08766288310289383,
-0.015508109703660011,
-0.11365051567554474,
-0.07620983570814133,
-0.06458844989538193,
0.15767885744571686,
0.04035979509353638,
0.033816512674093246,
-0.027098655700683594,
0.11204010248184204,
0.00622744532302022,
-0.12261059135198593,
0.029362211003899574,
0.02387363463640213,
-0.027582915499806404,
-0.03002273291349411,
-0.07402720302343369,
-0.0634206086397171,
0.033968064934015274,
0.09523040056228638,
-0.0405556820333004,
0.040303878486156464,
0.05608684942126274,
0.042358268052339554,
-0.10734781622886658,
0.16992858052253723,
-0.07376296073198318,
-0.010497160255908966,
-0.009552099741995335,
0.042789001017808914,
0.02284458465874195,
0.0014619727153331041,
-0.10888262093067169,
0.0010320536093786359,
0.07644032686948776,
0.010915636084973812,
-0.05422026291489601,
0.05819978937506676,
-0.06493420153856277,
-0.02839905396103859,
0.008625863119959831,
-0.08352472633123398,
0.031728461384773254,
0.008297326974570751,
-0.08309869468212128,
-0.006898213643580675,
0.0013744840398430824,
0.02184731513261795,
-0.005142437759786844,
0.12076728045940399,
-0.09179825335741043,
0.04388914629817009,
-0.10430611670017242,
-0.11551413685083389,
0.01977294124662876,
-0.06303665041923523,
0.033434007316827774,
-0.07049703598022461,
-0.17323796451091766,
-0.0015502830501645803,
0.06049255654215813,
-0.024430537596344948,
-0.04617238789796829,
-0.03296809270977974,
-0.07673165947198868,
-0.003909954335540533,
-0.010757125914096832,
0.1745768040418625,
-0.06711389869451523,
0.11540663242340088,
0.04179719462990761,
0.05598672851920128,
-0.046005357056856155,
0.05147075653076172,
-0.09092198312282562,
0.002646747976541519,
-0.1762325018644333,
0.04398424178361893,
-0.05400530993938446,
0.08689392358064651,
-0.09013061225414276,
-0.11438710242509842,
0.014038005843758583,
-0.012336552143096924,
0.07384544610977173,
0.08783575147390366,
-0.17795780301094055,
-0.07439747452735901,
0.14901858568191528,
-0.06220683082938194,
-0.1251707226037979,
0.10888437926769257,
-0.06575846672058105,
0.02355005405843258,
0.07388780266046524,
0.15481798350811005,
0.07890675961971283,
-0.09001652151346207,
0.02815142273902893,
0.013536266051232815,
0.03754054382443428,
-0.0973786786198616,
0.08388175070285797,
0.0014093787176534534,
-0.00661367317661643,
0.03565824404358864,
-0.058728184551000595,
0.07466516643762589,
-0.10187730938196182,
-0.09425432980060577,
-0.04591255262494087,
-0.10821295529603958,
0.04124387726187706,
0.08139485865831375,
0.0705966129899025,
-0.10769221186637878,
-0.07578263431787491,
0.07882186770439148,
0.09180216491222382,
-0.058541037142276764,
0.0275631882250309,
-0.04415484517812729,
0.05768444389104843,
-0.06793690472841263,
-0.029288729652762413,
-0.18900644779205322,
-0.02373356930911541,
0.000008731612069823314,
-0.013592059724032879,
0.0024677845649421215,
0.0638934001326561,
0.08054778724908829,
0.05951831862330437,
-0.0585455484688282,
-0.01938464678823948,
-0.035673387348651886,
-0.007168919779360294,
-0.11454480141401291,
-0.20971111953258514,
-0.05021582543849945,
-0.019376255571842194,
0.11390677839517593,
-0.17907924950122833,
0.027720540761947632,
-0.021094126626849174,
0.06394582241773605,
0.0005257099983282387,
-0.019002394750714302,
-0.03649785742163658,
0.08456417173147202,
-0.020258085802197456,
-0.052163999527692795,
0.07478703558444977,
0.010416027158498764,
-0.09834134578704834,
-0.05558754503726959,
-0.06872683763504028,
0.17609506845474243,
0.13804654777050018,
-0.13177356123924255,
-0.05168267711997032,
0.013049652799963951,
-0.06232909858226776,
-0.03216354921460152,
-0.04937884956598282,
0.04903243109583855,
0.1578061878681183,
0.00003520715472404845,
0.13282246887683868,
-0.07703794538974762,
-0.041060905903577805,
0.016017695888876915,
-0.02085057832300663,
0.03560177981853485,
0.12480657547712326,
0.11372290551662445,
-0.06160155311226845,
0.13594844937324524,
0.13968579471111298,
-0.08202097564935684,
0.11461016535758972,
-0.05947622284293175,
-0.07902026921510696,
-0.03192119300365448,
0.0019115691538900137,
-0.0038340496830642223,
0.11783858388662338,
-0.15114521980285645,
0.008945680223405361,
0.03092835284769535,
0.011252735741436481,
0.01814858615398407,
-0.247089222073555,
-0.056183770298957825,
0.030526967719197273,
-0.050692178308963776,
-0.02341591939330101,
-0.021661940962076187,
0.002506972523406148,
0.09609544277191162,
-0.009234271943569183,
-0.0608270987868309,
0.044245943427085876,
0.0013350568478927016,
-0.06367415934801102,
0.21898040175437927,
-0.08512059599161148,
-0.11960260570049286,
-0.11558901518583298,
-0.07068874686956406,
-0.05875789746642113,
-0.007020663935691118,
0.06628116220235825,
-0.09043244272470474,
-0.027602579444646835,
-0.044563934206962585,
0.03187357634305954,
-0.0256159920245409,
0.030065111815929413,
0.0009705040720291436,
-0.003373436164110899,
0.07723231613636017,
-0.12687601149082184,
0.0024432942736893892,
-0.05680453032255173,
-0.07078728824853897,
0.037130482494831085,
0.06633246690034866,
0.1165824681520462,
0.12961788475513458,
0.003912313841283321,
0.00542565668001771,
-0.0160771906375885,
0.23225213587284088,
-0.05261535197496414,
-0.04817574843764305,
0.12872529029846191,
0.01662667840719223,
0.05755509063601494,
0.09559108316898346,
0.09336546808481216,
-0.08276393264532089,
0.00013144129479769617,
0.03955172747373581,
-0.03776484727859497,
-0.23717784881591797,
-0.03475979343056679,
-0.060244545340538025,
-0.05618302524089813,
0.0662347748875618,
0.03050796128809452,
0.020910372957587242,
0.07443192601203918,
0.03603949025273323,
0.06891877949237823,
-0.06001264229416847,
0.04089866578578949,
0.11974308639764786,
0.046510156244039536,
0.11356353759765625,
-0.05858370289206505,
-0.059702642261981964,
0.03967205435037613,
-0.01670221984386444,
0.24632446467876434,
-0.005545818246901035,
0.16939549148082733,
0.0795481950044632,
0.2088494598865509,
-0.02164148911833763,
0.10060588270425797,
-0.00386367691680789,
-0.0502249151468277,
-0.009783195331692696,
-0.04108255356550217,
-0.02082393318414688,
0.0070739928632974625,
-0.04601709172129631,
0.07899579405784607,
-0.09648853540420532,
-0.025811167433857918,
0.05680523067712784,
0.26945555210113525,
0.019227653741836548,
-0.3052899241447449,
-0.07126174867153168,
-0.0033491889480501413,
-0.03008638322353363,
-0.0017686306964606047,
0.01006713043898344,
0.11718690395355225,
-0.08613735437393188,
0.03731003776192665,
-0.07580853253602982,
0.09072718024253845,
-0.03386981040239334,
0.04036740958690643,
0.07524969428777695,
0.10442502051591873,
0.004521994851529598,
0.0921982079744339,
-0.32032719254493713,
0.27118584513664246,
0.0049263411201536655,
0.07702090591192245,
-0.07951220870018005,
-0.006627009250223637,
0.029198186472058296,
0.04469858855009079,
0.05683285370469093,
-0.008937823586165905,
0.01495688408613205,
-0.16667956113815308,
-0.03169262781739235,
0.033678628504276276,
0.076726995408535,
-0.03034357912838459,
0.08980970829725266,
-0.019923800602555275,
0.0027423975989222527,
0.07333969324827194,
0.026607690379023552,
-0.04821466654539108,
-0.09488266706466675,
-0.01959441788494587,
0.022375380620360374,
-0.0693914070725441,
-0.07407721132040024,
-0.11568015813827515,
-0.11595908552408218,
0.12677747011184692,
0.000954287825152278,
-0.042732566595077515,
-0.1230306401848793,
0.08702096343040466,
0.1049150824546814,
-0.08249731361865997,
0.03786811605095863,
0.005106759257614613,
0.04614196717739105,
0.03354361653327942,
-0.06895502656698227,
0.09142185747623444,
-0.06212485209107399,
-0.15260708332061768,
-0.04807545989751816,
0.09908221662044525,
0.03585914894938469,
0.06412989646196365,
-0.0298979002982378,
0.0027376131620258093,
-0.05056820437312126,
-0.10703137516975403,
0.019963307306170464,
-0.022898996248841286,
0.07939707487821579,
0.02169356681406498,
-0.03481590375304222,
0.03791239485144615,
-0.047554370015859604,
-0.02695475146174431,
0.1807083785533905,
0.23212286829948425,
-0.09318123757839203,
0.017141874879598618,
0.037165407091379166,
-0.0517769455909729,
-0.16937661170959473,
0.04015589505434036,
0.0602496899664402,
0.0002022424159804359,
0.05775773897767067,
-0.18484064936637878,
0.1481015980243683,
0.1117682158946991,
-0.010057786479592323,
0.10994353145360947,
-0.3528968393802643,
-0.11340075731277466,
0.09445006400346756,
0.1789642721414566,
0.16177408397197723,
-0.16068004071712494,
-0.03148646280169487,
-0.017499376088380814,
-0.15252685546875,
0.10118992626667023,
-0.08655820041894913,
0.1206105500459671,
-0.03559473901987076,
0.08441638201475143,
-0.0017052337061613798,
-0.06662994623184204,
0.13895578682422638,
0.03341076523065567,
0.11084216833114624,
-0.0461963452398777,
-0.024885039776563644,
0.07895539700984955,
-0.02024846524000168,
-0.00012695134500972927,
-0.09366267919540405,
0.02336745522916317,
-0.0844074934720993,
-0.02480897307395935,
-0.10526064783334732,
0.03722776472568512,
-0.03067856840789318,
-0.0576147697865963,
-0.035237859934568405,
0.014030570164322853,
0.03693464398384094,
-0.015688909217715263,
0.12342987954616547,
0.04246138408780098,
0.1263166218996048,
0.07297060638666153,
0.0520104356110096,
-0.047834645956754684,
-0.10637817531824112,
-0.03128136694431305,
-0.00725093949586153,
0.05347031354904175,
-0.14447930455207825,
0.015302437357604504,
0.13018402457237244,
0.036886073648929596,
0.12136449664831161,
0.08217132091522217,
-0.04575621336698532,
0.016993412747979164,
0.060483042150735855,
-0.15651895105838776,
-0.12552541494369507,
-0.003910716623067856,
-0.08030126243829727,
-0.10486496984958649,
0.06077058985829353,
0.06823800504207611,
-0.0516669899225235,
-0.014461238868534565,
-0.02111935243010521,
-0.007560865022242069,
-0.07874390482902527,
0.2069120556116104,
0.0636819377541542,
0.05297750607132912,
-0.1266850233078003,
0.07246637344360352,
0.041327863931655884,
-0.07770247012376785,
-0.01969185099005699,
0.045816466212272644,
-0.07266605645418167,
-0.03915956988930702,
0.10352232307195663,
0.18539081513881683,
-0.05124985799193382,
-0.04867509379982948,
-0.13935667276382446,
-0.11806543171405792,
0.08029663562774658,
0.14058271050453186,
0.11962415277957916,
0.02365310862660408,
-0.05273055657744408,
0.0048611960373818874,
-0.11121239513158798,
0.07716172188520432,
0.03379383683204651,
0.06296734511852264,
-0.14630934596061707,
0.14012674987316132,
0.016660945490002632,
0.04057488217949867,
-0.010903292335569859,
0.029694315046072006,
-0.09079927206039429,
0.01752632111310959,
-0.14702005684375763,
-0.03352629393339157,
-0.04246121644973755,
0.004689457360655069,
0.0026698072906583548,
-0.0852852538228035,
-0.06011458486318588,
0.02164079248905182,
-0.13374091684818268,
-0.027518443763256073,
0.04446123167872429,
0.05998660996556282,
-0.13237649202346802,
-0.03728212043642998,
0.025297831743955612,
-0.05940594524145126,
0.05406920611858368,
0.06843705475330353,
0.019303463399410248,
0.07169171422719955,
-0.1347396969795227,
-0.012509764172136784,
0.06707608699798584,
0.012184853665530682,
0.0686737522482872,
-0.08042769879102707,
-0.007989905774593353,
0.010196344926953316,
0.06984463334083557,
0.007285000756382942,
0.048725757747888565,
-0.15064063668251038,
-0.017546921968460083,
-0.027457868680357933,
-0.09469152241945267,
-0.0649276077747345,
0.014847219921648502,
0.11028507351875305,
0.019597569480538368,
0.20854458212852478,
-0.05906305089592934,
0.03788319602608681,
-0.214204803109169,
-0.007619957439601421,
-0.019712325185537338,
-0.11184035986661911,
-0.12260209023952484,
-0.057712145149707794,
0.06044207513332367,
-0.04389849305152893,
0.14296014606952667,
0.003863460151478648,
0.04517728090286255,
0.011849792674183846,
0.0058609540574252605,
0.03373841568827629,
0.02130523882806301,
0.25161463022232056,
0.020500680431723595,
-0.029003852978348732,
0.07747890800237656,
0.05958713963627815,
0.09360160678625107,
0.1303662359714508,
0.17685769498348236,
0.17347654700279236,
0.0029856234323233366,
0.08835282176733017,
0.039113350212574005,
-0.05536116287112236,
-0.1354047805070877,
0.040686774998903275,
-0.04778829589486122,
0.0933023989200592,
-0.022329552099108696,
0.23936094343662262,
0.058121904730796814,
-0.18378101289272308,
0.05427216738462448,
-0.03336236625909805,
-0.08421040326356888,
-0.09480854868888855,
-0.046635132282972336,
-0.07182203978300095,
-0.15466240048408508,
0.010736883617937565,
-0.10849758237600327,
0.01934322528541088,
0.16540159285068512,
0.007372854743152857,
-0.014216185547411442,
0.16319306194782257,
0.07171037793159485,
0.028204895555973053,
0.04500748962163925,
0.003688481403514743,
-0.04030860960483551,
-0.08197405189275742,
-0.061259329319000244,
-0.010194117203354836,
-0.03786878287792206,
0.030127031728625298,
-0.061865366995334625,
-0.08349208533763885,
0.045526713132858276,
-0.02592124044895172,
-0.10319487005472183,
0.015986736863851547,
0.0037570807617157698,
0.07608455419540405,
0.03887875750660896,
0.013688502833247185,
0.020579025149345398,
-0.01831981912255287,
0.22784404456615448,
-0.0755690485239029,
-0.07880140841007233,
-0.09253310412168503,
0.2235030084848404,
0.03782196715474129,
-0.021525179967284203,
0.0511300191283226,
-0.06197516620159149,
0.009901183657348156,
0.24717365205287933,
0.19514864683151245,
-0.08460409194231033,
-0.016679471358656883,
0.006104414816945791,
-0.015918735414743423,
-0.0259145125746727,
0.1114739254117012,
0.11840339004993439,
0.031399454921483994,
-0.10895687341690063,
-0.0378405936062336,
-0.06912600249052048,
-0.014058271422982216,
-0.041220858693122864,
0.056196607649326324,
0.054530154913663864,
-0.012440396472811699,
-0.039868589490652084,
0.07077246159315109,
-0.03539371117949486,
-0.12249790132045746,
0.10766475647687912,
-0.17042863368988037,
-0.16471894085407257,
-0.03402670472860336,
0.1323743462562561,
0.012562860734760761,
0.07405578345060349,
-0.035305432975292206,
0.008872640319168568,
0.08050738275051117,
-0.016410386189818382,
-0.12519031763076782,
-0.09527745842933655,
0.11109840124845505,
-0.10963108390569687,
0.17573706805706024,
-0.04322907701134682,
0.06125941500067711,
0.1292448341846466,
0.06340111047029495,
-0.0750947892665863,
0.07514356076717377,
0.04512954130768776,
-0.10086067765951157,
0.02141881175339222,
0.07199279963970184,
-0.022973913699388504,
0.04768817126750946,
0.022237610071897507,
-0.11302901804447174,
0.033507708460092545,
-0.03442526236176491,
-0.038569770753383636,
-0.047540221363306046,
-0.04567338526248932,
-0.060045335441827774,
0.1140817403793335,
0.20579829812049866,
-0.02953318879008293,
0.018114231526851654,
-0.07301058620214462,
0.018967997282743454,
0.06605225056409836,
0.0503411702811718,
-0.08222661912441254,
-0.21609057486057281,
0.012517468072474003,
0.048725247383117676,
-0.023330463096499443,
-0.2202589213848114,
-0.09248873591423035,
0.01612619124352932,
-0.07852097600698471,
-0.08486967533826828,
0.06689930707216263,
0.07573557645082474,
0.051549281924963,
-0.04625053331255913,
-0.06240624189376831,
-0.08837449550628662,
0.15548942983150482,
-0.14958052337169647,
-0.08465207368135452
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-finetuned-ner
This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the conll2003 dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0492
- Precision: 0.9530
- Recall: 0.9604
- F1: 0.9567
- Accuracy: 0.9889
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
| 0.2031 | 1.0 | 878 | 0.0560 | 0.9381 | 0.9445 | 0.9413 | 0.9858 |
| 0.0446 | 2.0 | 1756 | 0.0480 | 0.9510 | 0.9578 | 0.9544 | 0.9887 |
| 0.0263 | 3.0 | 2634 | 0.0492 | 0.9530 | 0.9604 | 0.9567 | 0.9889 |
### Framework versions
- Transformers 4.10.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.0
- Tokenizers 0.10.3
|
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["conll2003"], "metrics": ["precision", "recall", "f1", "accuracy"], "model-index": [{"name": "roberta-base-finetuned-ner", "results": [{"task": {"type": "token-classification", "name": "Token Classification"}, "dataset": {"name": "conll2003", "type": "conll2003", "args": "conll2003"}, "metrics": [{"type": "precision", "value": 0.9529566113766282, "name": "Precision"}, {"type": "recall", "value": 0.9604268983755194, "name": "Recall"}, {"type": "f1", "value": 0.9566771720212616, "name": "F1"}, {"type": "accuracy", "value": 0.988938664048357, "name": "Accuracy"}]}]}]}
|
token-classification
|
dominiqueblok/roberta-base-finetuned-ner
|
[
"transformers",
"pytorch",
"roberta",
"token-classification",
"generated_from_trainer",
"dataset:conll2003",
"license:mit",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #roberta #token-classification #generated_from_trainer #dataset-conll2003 #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us
|
roberta-base-finetuned-ner
==========================
This model is a fine-tuned version of roberta-base on the conll2003 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.0492
* Precision: 0.9530
* Recall: 0.9604
* F1: 0.9567
* Accuracy: 0.9889
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3
### Training results
### Framework versions
* Transformers 4.10.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #roberta #token-classification #generated_from_trainer #dataset-conll2003 #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3"
] |
[
61,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #roberta #token-classification #generated_from_trainer #dataset-conll2003 #license-mit #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.0\n* Tokenizers 0.10.3"
] |
[
-0.11029805988073349,
0.08156268298625946,
-0.001842824975028634,
0.12898200750350952,
0.18953482806682587,
0.03818511962890625,
0.11436928063631058,
0.10812770575284958,
-0.11132709681987762,
0.0019381481688469648,
0.1297539621591568,
0.1889902502298355,
0.00019371719099581242,
0.11394859850406647,
-0.05408824607729912,
-0.26825860142707825,
-0.008074103854596615,
0.04264483228325844,
-0.06761572510004044,
0.14140352606773376,
0.10694523900747299,
-0.14721877872943878,
0.0948275551199913,
0.013287089765071869,
-0.23542174696922302,
0.0048821112141013145,
0.02476678229868412,
-0.061082977801561356,
0.15410593152046204,
0.019597390666604042,
0.14235785603523254,
-0.0031951076816767454,
0.10471165925264359,
-0.16223379969596863,
0.0075440239161252975,
0.03808512166142464,
0.012852860614657402,
0.09177816659212112,
0.052167657762765884,
-0.011814680881798267,
0.12322508543729782,
-0.06446933001279831,
0.06144107133150101,
0.026246841996908188,
-0.13559676706790924,
-0.22358235716819763,
-0.08133282512426376,
0.031917840242385864,
0.06588377058506012,
0.1003785952925682,
0.003502500709146261,
0.15212088823318481,
-0.11441582441329956,
0.08803719282150269,
0.20272447168827057,
-0.27751821279525757,
-0.06752195209264755,
0.05082442983984947,
-0.00806509517133236,
0.044445015490055084,
-0.10945417732000351,
-0.0286771971732378,
0.05233380198478699,
0.046887949109077454,
0.1328447163105011,
-0.04004121199250221,
-0.12061984837055206,
0.03108588606119156,
-0.14143216609954834,
-0.030472857877612114,
0.1474921852350235,
0.03808018937706947,
-0.03964781016111374,
-0.022540779784321785,
-0.04585815966129303,
-0.13889050483703613,
-0.034069448709487915,
-0.011725667864084244,
0.040582772344350815,
-0.04203420504927635,
-0.09508682042360306,
0.011863826774060726,
-0.0990099087357521,
-0.07604022324085236,
-0.08686873316764832,
0.17192712426185608,
0.03616495802998543,
0.017940310761332512,
-0.022338498383760452,
0.1113552525639534,
0.01237824372947216,
-0.11518464237451553,
0.01198573224246502,
0.017634497955441475,
-0.011133711785078049,
-0.06624661386013031,
-0.057819388806819916,
-0.02324892207980156,
0.002088128821924329,
0.11915440112352371,
-0.04170014336705208,
0.023960914462804794,
0.06564576178789139,
0.034471187740564346,
-0.08836207538843155,
0.19809851050376892,
-0.05109654366970062,
-0.015836812555789948,
0.003938435576856136,
0.03156942501664162,
-0.0073197041638195515,
0.0008342976216226816,
-0.12116635590791702,
-0.015803908929228783,
0.09349431097507477,
0.01180852111428976,
-0.08180185407400131,
0.07905234396457672,
-0.05856771394610405,
-0.033030763268470764,
0.0064646708779037,
-0.08488224446773529,
0.03323082998394966,
-0.00955626554787159,
-0.09411259740591049,
-0.02333640865981579,
0.012487636879086494,
0.02173750288784504,
0.00642566941678524,
0.1120559349656105,
-0.10252894461154938,
0.03130970522761345,
-0.09526628255844116,
-0.11683952808380127,
-0.0007007765816524625,
-0.08960893750190735,
0.03980677202343941,
-0.1078004315495491,
-0.18383240699768066,
-0.0035104781854897738,
0.039794307202100754,
-0.02278977818787098,
-0.058094024658203125,
-0.054533105343580246,
-0.06924896687269211,
-0.0035137273371219635,
-0.0070259165950119495,
0.12553471326828003,
-0.06100478768348694,
0.1072458028793335,
0.041591156274080276,
0.05381953716278076,
-0.05042039602994919,
0.05386679992079735,
-0.11345899105072021,
0.002054370241239667,
-0.175493061542511,
0.04687066748738289,
-0.041458919644355774,
0.07742704451084137,
-0.08136638253927231,
-0.11587420850992203,
0.037866778671741486,
-0.0018797480734065175,
0.06966551393270493,
0.09764810651540756,
-0.15929333865642548,
-0.07716431468725204,
0.1393982321023941,
-0.06069783493876457,
-0.11874835193157196,
0.1072923094034195,
-0.07994218915700912,
0.059842851012945175,
0.07355065643787384,
0.14601169526576996,
0.07971390336751938,
-0.06310094147920609,
0.02621614560484886,
-0.0033322239760309458,
0.03254775330424309,
-0.0804702639579773,
0.070414237678051,
0.005287363193929195,
-0.021489204838871956,
0.03153061866760254,
-0.03640327602624893,
0.06522492319345474,
-0.11589682102203369,
-0.09548422694206238,
-0.029729697853326797,
-0.10555907338857651,
0.06473000347614288,
0.07310302555561066,
0.0750080868601799,
-0.10446974635124207,
-0.06555324047803879,
0.086117684841156,
0.09396208077669144,
-0.03747282177209854,
0.011087781749665737,
-0.07035421580076218,
0.061673395335674286,
-0.04418550431728363,
-0.04123817756772041,
-0.17653082311153412,
-0.027100540697574615,
-0.0017217234708368778,
0.0466388463973999,
0.01643046736717224,
0.02817418985068798,
0.06900984048843384,
0.07365615665912628,
-0.04841190576553345,
-0.01356145367026329,
-0.029407015070319176,
0.005932915024459362,
-0.13312101364135742,
-0.19565661251544952,
-0.042747803032398224,
-0.013570841401815414,
0.1431867480278015,
-0.20955221354961395,
0.03309979662299156,
-0.03600196912884712,
0.07445533573627472,
0.002725910861045122,
-0.012123559601604939,
-0.0449012853205204,
0.08471479266881943,
-0.037274591624736786,
-0.045317620038986206,
0.07250799983739853,
0.001102451584301889,
-0.08395793288946152,
-0.05567113682627678,
-0.08765440434217453,
0.18763600289821625,
0.1291963905096054,
-0.13394196331501007,
-0.09308065474033356,
-0.013533085584640503,
-0.059061579406261444,
-0.019183356314897537,
-0.051197659224271774,
0.048261966556310654,
0.1747046560049057,
-0.021965498104691505,
0.14905747771263123,
-0.06389951705932617,
-0.0387842059135437,
0.019563009962439537,
-0.0385860949754715,
0.034480419009923935,
0.12285564094781876,
0.11962128430604935,
-0.10380680859088898,
0.14301243424415588,
0.13377602398395538,
-0.0842980295419693,
0.11936717480421066,
-0.029011938720941544,
-0.05960690230131149,
-0.025173403322696686,
-0.04332445189356804,
-0.008148069493472576,
0.11163970082998276,
-0.12984400987625122,
-0.002256177132949233,
0.028269080445170403,
0.023822443559765816,
0.00546647934243083,
-0.2171861231327057,
-0.06045175716280937,
0.04385460913181305,
-0.02796700783073902,
-0.0342707522213459,
-0.012551136314868927,
0.017839612439274788,
0.11048468202352524,
0.00847567431628704,
-0.10291112214326859,
0.04163738340139389,
0.013035975396633148,
-0.07975084334611893,
0.22052215039730072,
-0.08222157508134842,
-0.11298565566539764,
-0.11470971256494522,
-0.09411576390266418,
-0.04985625669360161,
0.019743990153074265,
0.05030856654047966,
-0.08693777769804001,
-0.025233803316950798,
-0.048709653317928314,
0.026618707925081253,
-0.02020401507616043,
0.03394812345504761,
-0.0033763216342777014,
-0.0044601731933653355,
0.06088070943951607,
-0.1093766838312149,
-0.011321808211505413,
-0.06374118477106094,
-0.08574411273002625,
0.052525367587804794,
0.02946891263127327,
0.12454280257225037,
0.1516307145357132,
-0.03534330800175667,
0.013455173000693321,
-0.038328561931848526,
0.2518528401851654,
-0.0585116446018219,
-0.042675502598285675,
0.13269734382629395,
0.0031555567402392626,
0.040844012051820755,
0.1153690367937088,
0.07666266709566116,
-0.08730088174343109,
0.0002118563570547849,
0.02965639904141426,
-0.04101607948541641,
-0.20293554663658142,
-0.05242347717285156,
-0.05543353036046028,
-0.0184993427246809,
0.0911906510591507,
0.014325343072414398,
0.04692484065890312,
0.0881851315498352,
0.03969613090157509,
0.07717882096767426,
-0.04943058267235756,
0.06302007287740707,
0.12015298753976822,
0.05257780849933624,
0.1329299956560135,
-0.038788389414548874,
-0.07396278530359268,
0.03931915760040283,
-0.03157983347773552,
0.22344353795051575,
0.0033456117380410433,
0.09597520530223846,
0.0641128420829773,
0.18938066065311432,
-0.0017933582421392202,
0.08240586519241333,
0.00823528878390789,
-0.042137011885643005,
-0.016168899834156036,
-0.032007813453674316,
-0.04188873991370201,
0.012523424811661243,
-0.05557873472571373,
0.0732068121433258,
-0.144608736038208,
-0.000008603436072007753,
0.05170826613903046,
0.2286183387041092,
0.042303215712308884,
-0.3529190123081207,
-0.09910328686237335,
-0.007793753407895565,
-0.018552852794528008,
-0.02005753666162491,
0.01831805892288685,
0.07524437457323074,
-0.11456426233053207,
0.0219259113073349,
-0.05460134893655777,
0.0918484628200531,
-0.04360490292310715,
0.04598712548613548,
0.06700459122657776,
0.1102864071726799,
-0.0063063534907996655,
0.08975596725940704,
-0.30213603377342224,
0.28736603260040283,
0.0030382906552404165,
0.07234315574169159,
-0.06933944672346115,
-0.018516050651669502,
0.035988643765449524,
0.0835760235786438,
0.05453218147158623,
-0.006883673369884491,
-0.006188963074237108,
-0.23852819204330444,
-0.028356485068798065,
0.0315471813082695,
0.07448285818099976,
-0.036239515990018845,
0.09613534808158875,
-0.035197753459215164,
0.015075436793267727,
0.0797770544886589,
-0.003712765173986554,
-0.04798319190740585,
-0.07901407033205032,
-0.032145287841558456,
0.02606414444744587,
-0.03178580850362778,
-0.058560870587825775,
-0.10723087191581726,
-0.11629540473222733,
0.1396653950214386,
0.0025028272066265345,
-0.0352872796356678,
-0.12039626389741898,
0.08679604530334473,
0.07470258325338364,
-0.09110480546951294,
0.04601408541202545,
0.0032251770608127117,
0.056957509368658066,
0.04180976748466492,
-0.060848627239465714,
0.10390142351388931,
-0.06674870103597641,
-0.15945100784301758,
-0.058445073664188385,
0.08380362391471863,
0.04477810114622116,
0.06650382280349731,
0.002715609734877944,
0.020691068843007088,
-0.04528819024562836,
-0.09036342054605484,
0.023699110373854637,
-0.012151211500167847,
0.08540473878383636,
0.035370755940675735,
-0.046926748007535934,
0.0370953306555748,
-0.06350155174732208,
-0.02185267210006714,
0.19057226181030273,
0.23781560361385345,
-0.10922156274318695,
0.009800318628549576,
0.018572259694337845,
-0.06156756356358528,
-0.18747086822986603,
0.05418292060494423,
0.05130113661289215,
0.008569668978452682,
0.04840919375419617,
-0.1721748262643814,
0.1332480013370514,
0.1125454530119896,
-0.005379253067076206,
0.07669946551322937,
-0.3158424496650696,
-0.12006640434265137,
0.13760711252689362,
0.15011687576770782,
0.14463117718696594,
-0.11912882328033447,
-0.005796986632049084,
-0.01785995438694954,
-0.14078308641910553,
0.10504709929227829,
-0.06901991367340088,
0.11722078919410706,
-0.0365908220410347,
0.10102244466543198,
0.011265095323324203,
-0.0508606880903244,
0.11664079874753952,
0.04050391539931297,
0.11282970011234283,
-0.06171439215540886,
-0.056819263845682144,
0.03355389088392258,
-0.02884780243039131,
0.006008739583194256,
-0.05285852402448654,
0.04111411049962044,
-0.11558342725038528,
-0.02192152850329876,
-0.08651746809482574,
0.04994169995188713,
-0.0285338182002306,
-0.07917402684688568,
-0.043473366647958755,
0.03140459954738617,
0.04234893247485161,
-0.0220522191375494,
0.11524645239114761,
0.025720957666635513,
0.15455682575702667,
0.07140742987394333,
0.07665383070707321,
-0.06424610316753387,
-0.05542586371302605,
-0.01889175921678543,
-0.01663145422935486,
0.05590033903717995,
-0.1190478578209877,
0.025243006646633148,
0.13804759085178375,
0.03229011595249176,
0.13057482242584229,
0.0947401151061058,
-0.011758523993194103,
0.014356621541082859,
0.06101367995142937,
-0.15185976028442383,
-0.061992097645998,
-0.005672582890838385,
-0.0767684355378151,
-0.1057400032877922,
0.05829700082540512,
0.08166340738534927,
-0.07553379237651825,
-0.01539477240294218,
-0.015744827687740326,
-0.009067995473742485,
-0.06725193560123444,
0.1989792138338089,
0.07350769639015198,
0.04805328696966171,
-0.10903715342283249,
0.05889488756656647,
0.04809696972370148,
-0.016797158867120743,
-0.01043790951371193,
0.044526975601911545,
-0.08994124829769135,
-0.03868231549859047,
0.07795555889606476,
0.19037647545337677,
-0.11558373272418976,
-0.029881928116083145,
-0.13339398801326752,
-0.12785948812961578,
0.07377849519252777,
0.15502095222473145,
0.12424658983945847,
0.017733469605445862,
-0.057274650782346725,
0.00238009262830019,
-0.13591015338897705,
0.07553579658269882,
0.043508708477020264,
0.068204365670681,
-0.15228570997714996,
0.19693623483181,
-0.005201751831918955,
0.05125951021909714,
-0.0220368392765522,
0.01670171692967415,
-0.10895565152168274,
0.01996840350329876,
-0.1236487329006195,
-0.04490169137716293,
-0.038860686123371124,
0.015073996968567371,
-0.004372702445834875,
-0.07461421191692352,
-0.056310124695301056,
0.007735961116850376,
-0.1184796467423439,
-0.013279317878186703,
0.050438281148672104,
0.06018907576799393,
-0.11376083642244339,
-0.04567654803395271,
0.02031245268881321,
-0.05076492577791214,
0.06444088369607925,
0.05977058783173561,
0.028202760964632034,
0.055524781346321106,
-0.13157802820205688,
0.0014884760603308678,
0.06330815702676773,
0.005565519444644451,
0.08954499661922455,
-0.07998864352703094,
0.0004442222707439214,
-0.0020980173721909523,
0.058214083313941956,
0.026707371696829796,
0.06764809042215347,
-0.13388337194919586,
-0.002248099073767662,
-0.018644187599420547,
-0.08145398646593094,
-0.06679367274045944,
0.028787754476070404,
0.09327586740255356,
0.014814501628279686,
0.20081454515457153,
-0.07271672040224075,
0.044694822281599045,
-0.2124311774969101,
-0.009801594540476799,
-0.02400263585150242,
-0.10882104188203812,
-0.149472177028656,
-0.07264503091573715,
0.0658569261431694,
-0.0474657267332077,
0.14002269506454468,
0.03919842466711998,
0.045923180878162384,
0.021889053285121918,
-0.0024429666809737682,
0.03999517112970352,
0.010540394112467766,
0.21322201192378998,
0.04591330885887146,
-0.042674366384744644,
0.08219809830188751,
0.06752626597881317,
0.10418225079774857,
0.10190626978874207,
0.18884848058223724,
0.14331428706645966,
-0.023121362552046776,
0.07193385064601898,
0.03982013836503029,
-0.06283745914697647,
-0.14681445062160492,
-0.002815295709297061,
-0.023018989711999893,
0.06674645096063614,
-0.02562800794839859,
0.20686902105808258,
0.06626549363136292,
-0.17422537505626678,
0.043918076902627945,
-0.05381741747260094,
-0.08431451767683029,
-0.11201117932796478,
-0.029039401561021805,
-0.08163163810968399,
-0.13899841904640198,
0.0064446451142430305,
-0.11171545833349228,
0.006739189848303795,
0.12946254014968872,
0.0019605704583227634,
-0.02574295550584793,
0.13576531410217285,
0.01550283282995224,
0.03941059112548828,
0.038590073585510254,
0.00022807682398706675,
-0.01808922551572323,
-0.11608725041151047,
-0.06722425669431686,
-0.02104262076318264,
-0.021596577018499374,
0.034546274691820145,
-0.07531289756298065,
-0.065815769135952,
0.02794549986720085,
-0.02953890711069107,
-0.1024184376001358,
0.016240378841757774,
0.02216995880007744,
0.06415688991546631,
0.023136315867304802,
0.009935135021805763,
0.02130029909312725,
-0.0143855856731534,
0.22901810705661774,
-0.0709448978304863,
-0.08031116425991058,
-0.0962224006652832,
0.29744797945022583,
0.05470729246735573,
-0.006176659371703863,
0.045531611889600754,
-0.06544727832078934,
0.01751311868429184,
0.24812051653862,
0.21529896557331085,
-0.10326386243104935,
-0.00565543957054615,
-0.011989974416792393,
-0.01706484891474247,
-0.008777456358075142,
0.11231676489114761,
0.11381688714027405,
0.03511130437254906,
-0.09716164320707321,
-0.029089968651533127,
-0.06684068590402603,
-0.0009619290940463543,
-0.021688546985387802,
0.06501398235559464,
0.053171686828136444,
0.0027926682960242033,
-0.045510515570640564,
0.05513330549001694,
-0.07246638834476471,
-0.07638049125671387,
0.08079196512699127,
-0.18850797414779663,
-0.15711873769760132,
-0.022092752158641815,
0.06857742369174957,
0.019274316728115082,
0.07910121232271194,
-0.03991393372416496,
0.0014771954156458378,
0.06192180886864662,
-0.01251304242759943,
-0.10643176734447479,
-0.0993606373667717,
0.11796513199806213,
-0.06097608432173729,
0.20772919058799744,
-0.05245952680706978,
0.07543212920427322,
0.1251034438610077,
0.066413514316082,
-0.07969681918621063,
0.0615069679915905,
0.04299882799386978,
-0.06448584794998169,
0.033640678972005844,
0.06220429018139839,
-0.03175985813140869,
0.05940413475036621,
0.035149626433849335,
-0.14689862728118896,
0.032271288335323334,
-0.07425359636545181,
-0.05665145441889763,
-0.03952225670218468,
-0.043232984840869904,
-0.0440322682261467,
0.12695451080799103,
0.2226317971944809,
-0.03131372481584549,
0.006422038190066814,
-0.07573222368955612,
0.018914533779025078,
0.07111115008592606,
0.007465437054634094,
-0.07659170031547546,
-0.21444669365882874,
0.0006266350392252207,
0.06461784243583679,
-0.03549910709261894,
-0.2318125069141388,
-0.0876360684633255,
-0.0058268285356462,
-0.08523984253406525,
-0.07722523808479309,
0.08735156059265137,
0.06902565062046051,
0.05739068239927292,
-0.05640697479248047,
-0.0386657752096653,
-0.08644407987594604,
0.16154277324676514,
-0.13697363436222076,
-0.0930979773402214
] |
null | null | null |
# this is a shit model
|
{}
| null |
douglas0204/shitmodel
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
# this is a shit model
|
[
"# this is a shit model"
] |
[
"TAGS\n#region-us \n",
"# this is a shit model"
] |
[
6,
6
] |
[
"passage: TAGS\n#region-us \n# this is a shit model"
] |
[
0.013940086588263512,
0.011461131274700165,
-0.007779167499393225,
-0.030228441581130028,
0.1056884303689003,
0.14580288529396057,
0.13727068901062012,
0.02501463145017624,
0.18658553063869476,
-0.007427537348121405,
0.15598051249980927,
0.06072605773806572,
-0.007316699717193842,
0.10887505114078522,
0.04913176968693733,
-0.27698230743408203,
0.06446187198162079,
-0.003407613141462207,
-0.11258967220783234,
-0.008534684777259827,
-0.042300283908843994,
-0.050472795963287354,
0.0037369802594184875,
-0.12031009048223495,
-0.10483808070421219,
0.08560910075902939,
0.0739540234208107,
0.00692721176892519,
0.04595121741294861,
0.01896097883582115,
0.17477597296237946,
-0.018616195768117905,
-0.15339471399784088,
-0.16895706951618195,
0.0646347776055336,
0.013103517703711987,
-0.027059759944677353,
0.013457714579999447,
0.03576352819800377,
0.040716417133808136,
-0.10231050103902817,
0.16791552305221558,
-0.036824896931648254,
0.07788784801959991,
-0.24160127341747284,
-0.13497284054756165,
-0.007496243342757225,
-0.03154270350933075,
0.09992896020412445,
-0.01994086056947708,
-0.006883607245981693,
0.26961684226989746,
-0.18409858644008636,
0.008053974248468876,
0.0307560283690691,
-0.15259185433387756,
0.06572217494249344,
0.2806946337223053,
0.01916946843266487,
0.1431790441274643,
0.016994405537843704,
0.12296359241008759,
0.11132518202066422,
0.05708971992135048,
-0.20038238167762756,
-0.01125036459416151,
0.09422005712985992,
0.12415356189012527,
-0.09490840137004852,
-0.08372077345848083,
0.147261843085289,
0.08101296424865723,
-0.041822291910648346,
0.10768245905637741,
-0.07114827632904053,
-0.212776780128479,
0.031985409557819366,
-0.04073784127831459,
-0.02663087099790573,
0.18355825543403625,
0.12803755700588226,
-0.019406121224164963,
-0.10760509967803955,
0.048312511295080185,
-0.22706642746925354,
0.15734979510307312,
-0.014854973182082176,
0.10767235606908798,
-0.2763964533805847,
0.04984845221042633,
-0.3166295886039734,
-0.04552953690290451,
0.07188882678747177,
-0.1092958003282547,
-0.04552476108074188,
0.02814112789928913,
-0.015555206686258316,
0.059721510857343674,
0.0865652859210968,
-0.016331946477293968,
-0.06536826491355896,
0.0663786306977272,
-0.053524550050497055,
0.13495402038097382,
0.09558278322219849,
0.10445267707109451,
-0.013141302391886711,
0.08701162040233612,
-0.04602762311697006,
-0.10630442202091217,
0.04764191061258316,
-0.08795733004808426,
-0.06303398311138153,
0.020918317139148712,
-0.09592455625534058,
0.15480680763721466,
0.0058548422530293465,
-0.07028400897979736,
-0.11227531731128693,
0.016058500856161118,
-0.00647655688226223,
0.09118690341711044,
-0.023101316764950752,
-0.0448850654065609,
-0.03121652640402317,
0.15547719597816467,
-0.11697973310947418,
0.041247833520174026,
0.06267812103033066,
0.08280529081821442,
-0.14986760914325714,
-0.04738357663154602,
-0.10647958517074585,
0.03627941012382507,
0.056743159890174866,
-0.15195240080356598,
0.11647188663482666,
-0.04760664328932762,
-0.07385137677192688,
0.06679315119981766,
0.011389161460101604,
-0.03272191807627678,
0.03986016660928726,
-0.03912973403930664,
-0.009415417909622192,
-0.02656797505915165,
-0.02031901851296425,
-0.07198138535022736,
-0.07325592637062073,
0.024325989186763763,
-0.11420241743326187,
0.00949209462851286,
-0.3065824806690216,
-0.018268650397658348,
-0.05056656524538994,
0.11863242089748383,
-0.15571454167366028,
-0.026507824659347534,
-0.013532050885260105,
0.1319097876548767,
-0.006524599622935057,
0.05906476080417633,
-0.24565637111663818,
0.0583852082490921,
-0.04648217931389809,
0.28217899799346924,
-0.18524198234081268,
-0.0647449791431427,
0.13914543390274048,
-0.07183977961540222,
-0.07848060876131058,
0.021181175485253334,
-0.04077168554067612,
0.09228593111038208,
0.05205537751317024,
0.3056480288505554,
-0.1604083925485611,
-0.0329577662050724,
0.14349773526191711,
0.1929650753736496,
-0.20754826068878174,
-0.07733170688152313,
0.047459375113248825,
-0.16077154874801636,
-0.2154301106929779,
-0.004653514828532934,
0.022221364080905914,
0.10757173597812653,
-0.07549288123846054,
-0.08772507309913635,
0.017771193757653236,
-0.01036849059164524,
0.0017925591673702002,
0.05515982210636139,
0.06910476088523865,
0.0027930382639169693,
0.020718196406960487,
-0.026182951405644417,
0.04894488304853439,
0.1643034815788269,
0.0017237047431990504,
-0.0014952514320611954,
0.03998992592096329,
-0.10805957019329071,
0.04010758176445961,
-0.18062163889408112,
-0.23011504113674164,
-0.016727538779377937,
0.11407841742038727,
0.041735295206308365,
0.20304664969444275,
0.07584591209888458,
-0.14704392850399017,
0.0781647339463234,
0.011273568496108055,
0.08922219276428223,
0.04003211110830307,
-0.01930953562259674,
-0.07355860620737076,
0.11034025251865387,
-0.06195314601063728,
-0.042224884033203125,
-0.06602920591831207,
-0.010765215381979942,
-0.0474989153444767,
0.007737379986792803,
-0.004026423208415508,
0.014421341940760612,
0.08063115179538727,
0.012054398655891418,
0.05654945969581604,
0.004366276785731316,
0.09521648287773132,
-0.06958325207233429,
-0.04885008931159973,
0.02767246961593628,
-0.07840000838041306,
0.09540870785713196,
0.19228266179561615,
-0.10029388219118118,
0.002891725394874811,
0.019055739045143127,
0.09173893928527832,
-0.006888110190629959,
0.1414664387702942,
-0.015174868516623974,
-0.004063341300934553,
0.037029363214969635,
-0.019022155553102493,
0.008319160901010036,
0.09566084295511246,
-0.02130282111465931,
-0.019518135115504265,
-0.16164587438106537,
0.16853216290473938,
0.13673439621925354,
-0.0854671448469162,
0.15642304718494415,
0.3354072570800781,
0.12295214086771011,
0.17825695872306824,
-0.0693870261311531,
-0.015957264229655266,
-0.09834233671426773,
0.03297177702188492,
-0.06207532435655594,
0.10284418612718582,
-0.1799708604812622,
0.047114163637161255,
0.042704664170742035,
0.03626508265733719,
0.08213643729686737,
-0.06442192941904068,
-0.13387921452522278,
-0.0511791817843914,
-0.006414357572793961,
-0.12377585470676422,
0.08267630636692047,
-0.05668533593416214,
0.07543499022722244,
0.07585783302783966,
-0.042490363121032715,
0.09385909885168076,
-0.0023360580671578646,
-0.027885740622878075,
0.041712094098329544,
-0.24193717539310455,
-0.021578313782811165,
-0.1520690768957138,
-0.03996122628450394,
0.00820617564022541,
-0.008658640086650848,
-0.047621261328458786,
-0.15291613340377808,
-0.05526217818260193,
0.08269715309143066,
0.07449179142713547,
-0.25355106592178345,
-0.014988385140895844,
0.01885051280260086,
0.010122250765562057,
-0.10286292433738708,
-0.11348443478345871,
-0.059454500675201416,
-0.06350981444120407,
0.1456466019153595,
0.0019810323137789965,
-0.2164442390203476,
0.018334925174713135,
0.2899656295776367,
0.12016235291957855,
0.09650623798370361,
-0.0031947316601872444,
0.29341596364974976,
-0.1397119015455246,
-0.049566734582185745,
0.059018585830926895,
-0.10548438131809235,
0.03608400747179985,
0.15206965804100037,
0.061511434614658356,
-0.10380689054727554,
-0.07429078966379166,
-0.04825977608561516,
-0.1512712836265564,
-0.13746881484985352,
-0.03930409997701645,
-0.04715835303068161,
0.05726569518446922,
-0.06962189078330994,
0.08501334488391876,
0.10832171142101288,
0.038237035274505615,
0.0617327019572258,
-0.023518893867731094,
-0.13073547184467316,
-0.0021876245737075806,
-0.011058059521019459,
-0.11632274091243744,
-0.027475755661725998,
-0.11469569802284241,
-0.06556076556444168,
0.12013018131256104,
-0.012242840602993965,
0.08594834804534912,
0.24889087677001953,
-0.0027823750860989094,
0.10301998257637024,
0.03155200183391571,
0.13238836824893951,
0.13305310904979706,
0.02044784650206566,
-0.06411172449588776,
-0.035759806632995605,
-0.006032692268490791,
-0.03912418335676193,
0.021333636716008186,
0.049160923808813095,
-0.20590035617351532,
0.0006277475040405989,
-0.0863066166639328,
0.03520932421088219,
-0.13277769088745117,
0.12545517086982727,
-0.04530796781182289,
0.07538808882236481,
0.11475921422243118,
0.10484736412763596,
-0.0694815143942833,
0.0975537821650505,
0.006378709338605404,
-0.05821714922785759,
-0.06987565010786057,
0.054977260529994965,
0.0848456472158432,
0.09644018858671188,
0.12989871203899384,
-0.07813289016485214,
-0.06635478883981705,
-0.04026489704847336,
0.029199086129665375,
-0.12434393167495728,
0.25095629692077637,
0.029420040547847748,
-0.09703222662210464,
-0.04391190782189369,
-0.12041287124156952,
0.030132576823234558,
0.18389709293842316,
0.13256138563156128,
0.060602959245443344,
-0.19887594878673553,
-0.08262594789266586,
0.05529751628637314,
-0.011868448927998543,
0.12207748740911484,
0.041779808700084686,
-0.0802161917090416,
0.005133365280926228,
0.06236690282821655,
0.026915814727544785,
0.04775485023856163,
0.05171968787908554,
-0.00907636247575283,
-0.002274148166179657,
0.07726072520017624,
-0.042493365705013275,
0.014762192964553833,
0.057473085820674896,
0.05339665710926056,
0.060210227966308594,
-0.057138677686452866,
-0.011812273412942886,
-0.07918334007263184,
-0.03754320368170738,
0.12104196101427078,
-0.055878687649965286,
-0.038118500262498856,
-0.10365797579288483,
-0.07318949699401855,
-0.006986419670283794,
-0.21395403146743774,
0.11165095120668411,
-0.034661851823329926,
0.06654026359319687,
-0.0035133713390678167,
0.048548951745033264,
-0.021115178242325783,
0.032918717712163925,
0.00667297188192606,
0.023473849520087242,
-0.036452483385801315,
-0.13894379138946533,
0.06606088578701019,
-0.1722579002380371,
-0.06918677687644958,
0.17975583672523499,
-0.015619227662682533,
0.04153543338179588,
0.020168110728263855,
-0.06261733919382095,
0.2412548065185547,
0.3190734386444092,
-0.014381916262209415,
0.0413522906601429,
0.3157774806022644,
-0.15297725796699524,
-0.18981914222240448,
-0.07920702546834946,
-0.18701152503490448,
-0.056025367230176926,
0.11166481673717499,
-0.08918128907680511,
0.07067155092954636,
0.17189829051494598,
-0.005504935048520565,
0.1806798279285431,
-0.2755383551120758,
0.00822791550308466,
0.3015287518501282,
-0.012605583295226097,
0.5971616506576538,
-0.16025413572788239,
-0.14355649054050446,
0.04184379428625107,
-0.024917926639318466,
0.18986137211322784,
-0.013333083130419254,
0.07225723564624786,
-0.00021183188073337078,
0.07474902272224426,
0.030274299904704094,
-0.02929827943444252,
0.2284250408411026,
0.03688615560531616,
0.09580519050359726,
-0.06189369782805443,
-0.3827725648880005,
0.19244374334812164,
0.012293475680053234,
-0.022425204515457153,
0.09501434117555618,
-0.06207061931490898,
-0.18465743958950043,
0.052267689257860184,
-0.07588551938533783,
-0.0697631686925888,
0.02450120821595192,
-0.05519549548625946,
-0.048811160027980804,
0.07069693505764008,
-0.09839747846126556,
-0.051593292504549026,
0.1998232752084732,
-0.0471772775053978,
0.1649431586265564,
0.08322714269161224,
-0.042450953274965286,
-0.08548245579004288,
-0.08517284691333771,
0.02695099264383316,
-0.05391417071223259,
0.06774487346410751,
-0.11221961677074432,
-0.015278501436114311,
0.1201556921005249,
-0.03200072422623634,
0.06717386096715927,
0.10209769755601883,
-0.09528832882642746,
0.014896567910909653,
0.19987046718597412,
-0.16979622840881348,
-0.1888020932674408,
-0.03206045180559158,
-0.08427160978317261,
0.08753565698862076,
-0.06692041456699371,
0.05659256875514984,
0.1702013611793518,
0.0561530627310276,
0.03321951627731323,
0.002543969079852104,
-0.11560343205928802,
-0.03357095271348953,
0.0387704074382782,
0.0012453391682356596,
-0.09399475902318954,
0.1418764740228653,
0.026654519140720367,
-0.06185975670814514,
-0.023852327838540077,
0.08333417028188705,
-0.0195180531591177,
-0.09240548312664032,
-0.06410925835371017,
0.1136704534292221,
-0.06645584851503372,
-0.06046326830983162,
0.03566215559840202,
-0.08667878061532974,
-0.03507841378450394,
0.03161134198307991,
0.02905290760099888,
0.11846036463975906,
-0.004651972558349371,
-0.10592662543058395,
0.08308115601539612,
-0.11830955743789673,
-0.07013089209794998,
-0.024934474378824234,
-0.09858489781618118,
-0.014071443118155003,
-0.04455885663628578,
0.13004927337169647,
-0.11553560197353363,
-0.15110670030117035,
-0.30525797605514526,
0.0420653373003006,
-0.17112776637077332,
-0.08935318887233734,
-0.037765827029943466,
-0.1290460079908371,
0.09779734164476395,
-0.009334791451692581,
-0.06719326227903366,
-0.06036269664764404,
-0.173761785030365,
0.07098013162612915,
0.06500037014484406,
0.06286003440618515,
0.04945385828614235,
0.01343389879912138,
0.12316841632127762,
0.0379696823656559,
0.03986668586730957,
0.19779306650161743,
0.049982160329818726,
0.2236972600221634,
-0.004325035959482193,
-0.055518146604299545,
0.08139266818761826,
-0.04111187532544136,
0.0466487854719162,
0.1005157008767128,
-0.05123274773359299,
-0.03582863137125969,
-0.006964338012039661,
0.0979052484035492,
0.024556271731853485,
-0.11302775144577026,
-0.019663412123918533,
0.12867707014083862,
-0.18176111578941345,
-0.010274248197674751,
-0.1513649821281433,
0.1559375524520874,
0.07101383060216904,
0.07701127231121063,
0.1296728253364563,
0.09600847959518433,
-0.027446048334240913,
-0.022953303530812263,
0.01831844262778759,
-0.11548283696174622,
0.15720443427562714,
-0.07894285023212433,
-0.018691113218665123,
-0.03141088783740997,
0.2708701491355896,
-0.0685366839170456,
0.06828480213880539,
-0.0030550151132047176,
0.1451295167207718,
-0.018926091492176056,
0.007499495521187782,
0.019222984090447426,
0.04805707186460495,
-0.07926623523235321,
-0.2870822548866272,
0.05079968646168709,
-0.025880735367536545,
-0.15454250574111938,
0.19360153377056122,
-0.055040232837200165,
0.02624974213540554,
0.1260022670030594,
-0.008166976273059845,
-0.0946272760629654,
0.072113536298275,
-0.08680802583694458,
-0.0067079756408929825,
0.04317067936062813,
0.003324383171275258,
-0.015149261802434921,
0.17453059554100037,
-0.07198399305343628,
0.062452010810375214,
-0.12312505394220352,
-0.044043976813554764,
-0.19230026006698608,
-0.2016453593969345,
0.06516150385141373,
-0.09752266108989716,
0.05361660569906235,
-0.02346806786954403,
-0.017898472025990486,
0.3714015483856201,
0.04404004290699959,
-0.048020463436841965,
0.07518880814313889,
-0.17626416683197021,
-0.1050008088350296,
0.02664928324520588,
0.015955179929733276,
0.042706847190856934,
-0.1025024801492691,
-0.027231808751821518,
-0.09219931066036224,
-0.07683833688497543,
-0.1055293157696724,
0.018182065337896347,
-0.024648576974868774,
-0.07747519761323929,
-0.15549540519714355,
-0.05524164438247681,
-0.08675762265920639,
0.11849132180213928,
-0.11358922719955444,
0.18091614544391632,
0.02748355269432068,
0.026752324774861336,
0.018095463514328003,
0.04947821795940399,
0.008685036562383175,
0.008951151743531227,
-0.04349660128355026,
0.11010295152664185,
-0.11897071450948715,
0.09319599717855453,
-0.05614514276385307,
-0.0036845512222498655,
-0.11002881824970245,
0.193031445145607,
0.33378636837005615,
-0.22362609207630157,
-0.019303815439343452,
0.003564262529835105,
0.053271740674972534,
0.03296298906207085,
0.2568519413471222,
-0.001484552165493369,
0.15603747963905334,
-0.09598138183355331,
0.17575114965438843,
-0.02652011252939701,
-0.004096379037946463,
0.0035468481946736574,
0.005118393339216709,
0.12473231554031372,
-0.05194445699453354,
-0.09423397481441498,
0.23527762293815613,
-0.3273721933364868,
0.17442239820957184,
-0.022851090878248215,
-0.1479114443063736,
-0.0593649223446846,
-0.07058312743902206,
0.032767847180366516,
-0.06393226236104965,
0.15694370865821838,
-0.12256745249032974,
-0.08189106732606888,
-0.044329915195703506,
-0.015526089817285538,
-0.2814066708087921,
-0.19468088448047638,
0.13581112027168274,
0.13277317583560944,
0.015977749601006508,
0.07351697981357574,
-0.02988607808947563,
0.04775554686784744,
0.015964046120643616,
-0.048170026391744614,
0.030201083049178123,
0.034772783517837524,
-0.00543323066085577,
-0.2317710518836975,
-0.14493145048618317,
0.051323480904102325,
-0.2249934822320938,
0.030792834237217903,
0.0048810215666890144,
0.01272303145378828,
0.061819493770599365,
-0.09563292562961578,
0.05237551033496857,
0.0480591356754303,
-0.1331336796283722,
0.06570837646722794,
0.0842808410525322,
0.02923179790377617,
-0.03921728581190109,
0.044221244752407074,
0.010871395468711853,
0.08969853818416595,
-0.1107565313577652,
-0.12692946195602417,
0.11538432538509369,
-0.038741499185562134,
0.24810409545898438,
-0.043838173151016235,
-0.10943551361560822,
0.012172713875770569,
-0.09831061214208603,
0.11855032294988632,
-0.05355137959122658,
0.07799208164215088,
0.10093117505311966,
-0.016680967062711716,
-0.0034757915418595076,
-0.07476786524057388,
0.0557527095079422,
-0.07898840308189392,
-0.009957500733435154,
-0.07969004660844803
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.