sha
null | last_modified
null | library_name
stringclasses 154
values | text
stringlengths 1
900k
| metadata
stringlengths 2
348k
| pipeline_tag
stringclasses 45
values | id
stringlengths 5
122
| tags
listlengths 1
1.84k
| created_at
stringlengths 25
25
| arxiv
listlengths 0
201
| languages
listlengths 0
1.83k
| tags_str
stringlengths 17
9.34k
| text_str
stringlengths 0
389k
| text_lists
listlengths 0
722
| processed_texts
listlengths 1
723
| tokens_length
listlengths 1
723
| input_texts
listlengths 1
61
| embeddings
listlengths 768
768
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
null | null |
transformers
|
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on CroSloEngual (https://huggingface.co/EMBEDDIA/crosloengual-bert) and supports different tagsets all using IOBES formats:
1. Wikiann (LOC, PER, ORG)
2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO)
3. SSJ500k (LOC, MISC, ORG, PER)
PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date
You can select the tagset to use in the output by configuring the model. This model manages differently uppercase words.
More information about the model can be found in the paper (https://aclanthology.org/2021.bsnlp-1.12.pdf) and GitHub repository (https://github.com/EMBEDDIA/NER_FEDA).
|
{"language": ["hr", "sl", "en", "multilingual"], "license": "mit", "tags": ["CroSloEngual", "ner"]}
| null |
creat89/NER_FEDA_Sl
|
[
"transformers",
"pytorch",
"bert",
"CroSloEngual",
"ner",
"hr",
"sl",
"en",
"multilingual",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"hr",
"sl",
"en",
"multilingual"
] |
TAGS
#transformers #pytorch #bert #CroSloEngual #ner #hr #sl #en #multilingual #license-mit #endpoints_compatible #region-us
|
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on CroSloEngual (URL and supports different tagsets all using IOBES formats:
1. Wikiann (LOC, PER, ORG)
2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO)
3. SSJ500k (LOC, MISC, ORG, PER)
PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date
You can select the tagset to use in the output by configuring the model. This model manages differently uppercase words.
More information about the model can be found in the paper (URL and GitHub repository (URL
|
[] |
[
"TAGS\n#transformers #pytorch #bert #CroSloEngual #ner #hr #sl #en #multilingual #license-mit #endpoints_compatible #region-us \n"
] |
[
47
] |
[
"passage: TAGS\n#transformers #pytorch #bert #CroSloEngual #ner #hr #sl #en #multilingual #license-mit #endpoints_compatible #region-us \n"
] |
[
-0.046101897954940796,
0.003084068186581135,
-0.009912987239658833,
0.020118936896324158,
0.09581785649061203,
0.03061067871749401,
0.0955689549446106,
0.039487697184085846,
0.14023089408874512,
-0.01253522839397192,
0.1302793025970459,
0.1838635951280594,
-0.03774821758270264,
-0.10596001148223877,
-0.048574354499578476,
-0.2769118547439575,
0.03592107445001602,
0.04605446010828018,
0.0001876611349871382,
0.08818645030260086,
0.07571765035390854,
-0.05304563418030739,
0.07084546238183975,
0.0036447288002818823,
-0.049355972558259964,
0.047309886664152145,
0.005650949198752642,
-0.03486891835927963,
0.12302669882774353,
0.0599081926047802,
0.10880177468061447,
0.05217298865318298,
-0.014536687172949314,
-0.2148798406124115,
0.0236379262059927,
-0.07252994924783707,
-0.08442375808954239,
-0.0039619640447199345,
0.011876353994011879,
-0.10780103504657745,
0.12996073067188263,
-0.001768068061210215,
-0.033983200788497925,
0.05454518646001816,
-0.13404406607151031,
-0.15851350128650665,
-0.05717615410685539,
0.09509368240833282,
-0.015363861806690693,
0.035392142832279205,
0.030084362253546715,
0.1436282843351364,
-0.16497276723384857,
0.06068117916584015,
0.10042638331651688,
-0.32786643505096436,
-0.0009008516208268702,
0.1040160208940506,
0.08549179136753082,
0.013809138908982277,
-0.06858427077531815,
0.08794210106134415,
0.036245111376047134,
0.0070217144675552845,
-0.10380084812641144,
-0.0995572954416275,
0.03335106745362282,
0.07790547609329224,
-0.0662488266825676,
-0.05762188881635666,
0.10565447062253952,
-0.02641178108751774,
0.025915130972862244,
0.04022420942783356,
-0.022534463554620743,
-0.02211112156510353,
-0.004403512459248304,
0.0017171564977616072,
-0.007578739896416664,
0.0795210674405098,
0.026223603636026382,
-0.047605354338884354,
-0.10185729712247849,
0.0671001598238945,
-0.2470119446516037,
0.20362767577171326,
0.059108536690473557,
0.052988968789577484,
-0.12286603450775146,
0.03229225054383278,
-0.013097964227199554,
-0.06506649404764175,
-0.04273943230509758,
-0.05362657457590103,
0.059383559972047806,
0.03192165493965149,
-0.016021959483623505,
0.1518450230360031,
0.07467029243707657,
0.17995715141296387,
-0.034080058336257935,
0.03794283792376518,
-0.011612074449658394,
0.12994910776615143,
-0.01636532135307789,
0.05997674539685249,
0.05648627504706383,
-0.004642786458134651,
-0.00348447123542428,
-0.178485706448555,
0.04207747057080269,
0.01099359430372715,
-0.15944896638393402,
-0.06947366148233414,
-0.087864950299263,
0.12646129727363586,
-0.018535831943154335,
0.052645351737737656,
-0.045039668679237366,
0.06066381186246872,
-0.0175655335187912,
-0.027239801362156868,
0.02016652561724186,
0.013753232546150684,
0.051943469792604446,
0.17256508767604828,
-0.0384962223470211,
-0.0074758464470505714,
-0.022979259490966797,
0.12318085879087448,
-0.046956948935985565,
0.04070562869310379,
-0.03429826349020004,
-0.003276701085269451,
0.08151887357234955,
-0.11472442001104355,
0.06964597851037979,
-0.11503278464078903,
-0.10161620378494263,
0.009736756794154644,
0.021344594657421112,
-0.017635449767112732,
0.022710220888257027,
-0.051344357430934906,
-0.003219804959371686,
0.009034756571054459,
-0.06261837482452393,
-0.11346893757581711,
-0.07778596878051758,
0.09481405466794968,
-0.038483623415231705,
0.005766252521425486,
-0.19370341300964355,
0.01579875871539116,
-0.07802782952785492,
0.04110882431268692,
-0.03301183879375458,
-0.016919059678912163,
-0.04220915958285332,
0.15836010873317719,
0.01566770300269127,
-0.08032304793596268,
-0.14296548068523407,
0.06578236818313599,
-0.0657002255320549,
0.15304335951805115,
-0.11342331022024155,
-0.09540978819131851,
0.23656173050403595,
-0.10326364636421204,
-0.12686122953891754,
0.10227407515048981,
0.006417231168597937,
0.00432013301178813,
0.08028869330883026,
0.22379741072654724,
0.04549868777394295,
-0.1472753882408142,
0.06182360276579857,
0.1733553260564804,
-0.08536577224731445,
-0.12865838408470154,
0.05334661900997162,
-0.05889151245355606,
0.0050757513381540775,
0.04860455542802811,
0.00635339692234993,
0.10538871586322784,
-0.07431323081254959,
-0.04350416362285614,
0.0018023509765043855,
0.014889349229633808,
0.032540369778871536,
0.053502246737480164,
0.07121594995260239,
-0.10586408525705338,
-0.014346176758408546,
0.05915049836039543,
0.012473384849727154,
0.05546606332063675,
0.07566056400537491,
-0.07824458926916122,
0.07523985207080841,
0.08620641380548477,
-0.03238515183329582,
-0.11253771930932999,
-0.04384974390268326,
-0.06783401966094971,
0.04701196029782295,
0.06976798176765442,
0.1756046712398529,
0.06230808421969414,
-0.06800560653209686,
-0.05253079906105995,
0.022036606445908546,
0.15244317054748535,
0.002737026894465089,
-0.00018774498312268406,
-0.12074439972639084,
0.056363075971603394,
-0.029165204614400864,
-0.023172860965132713,
-0.10322974622249603,
-0.004951235838234425,
0.1568737030029297,
0.14585448801517487,
-0.04912438988685608,
0.06703396886587143,
-0.09541365504264832,
0.0676719918847084,
-0.03582797944545746,
0.020632799714803696,
0.12695221602916718,
0.003176177153363824,
-0.12083008885383606,
0.2063366174697876,
-0.022861991077661514,
0.2756877541542053,
0.19385331869125366,
-0.2664072811603546,
0.03504352644085884,
-0.0771857425570488,
-0.025529207661747932,
0.02368001639842987,
0.11642418801784515,
0.012137041427195072,
0.09690926223993301,
0.02864653244614601,
0.09156838804483414,
-0.028308026492595673,
-0.04136788100004196,
-0.014135902747511864,
-0.055842325091362,
-0.08778955787420273,
0.11056928336620331,
0.11863058805465698,
-0.18987371027469635,
0.18550541996955872,
0.2674977779388428,
0.06353795528411865,
0.17024104297161102,
-0.07885193079710007,
0.038198839873075485,
-0.03564628213644028,
0.028778357431292534,
-0.050312891602516174,
0.07006462663412094,
-0.24196560680866241,
-0.026088887825608253,
0.03120078332722187,
0.012899274006485939,
0.0565243698656559,
-0.14664432406425476,
-0.09281187504529953,
-0.020192502066493034,
-0.003125994699075818,
-0.09683786332607269,
0.09982286393642426,
-0.022991131991147995,
0.05426518991589546,
-0.03466043621301651,
-0.10500705987215042,
0.11790910363197327,
-0.002195822773501277,
-0.032241325825452805,
0.12027522921562195,
-0.1407175213098526,
-0.16423015296459198,
-0.12591402232646942,
-0.14754514396190643,
-0.010272295214235783,
-0.012885184027254581,
0.09450490027666092,
-0.04180587828159332,
-0.01674993522465229,
0.09747575223445892,
0.02213745005428791,
-0.15049733221530914,
-0.011711008846759796,
-0.08220027387142181,
0.03129400312900543,
-0.12110886722803116,
-0.0710226446390152,
-0.08760104328393936,
-0.051831867545843124,
-0.04834259673953056,
0.07303786277770996,
-0.11613848060369492,
0.05853838473558426,
0.08274015039205551,
0.039481695741415024,
0.06263338029384613,
-0.056486163288354874,
0.16199715435504913,
-0.08131183683872223,
-0.062329016625881195,
0.09548679739236832,
0.004713934846222401,
0.08722091466188431,
0.1512838751077652,
0.07543521374464035,
-0.05117873474955559,
-0.0616474375128746,
-0.02352398820221424,
-0.08395214378833771,
-0.2018536925315857,
-0.09306910634040833,
-0.14184466004371643,
0.011168778873980045,
-0.06365631520748138,
0.05325627699494362,
0.02734704129397869,
0.03850555047392845,
0.014175854623317719,
-0.09953257441520691,
-0.02947876788675785,
0.05536040663719177,
0.35729339718818665,
-0.046026941388845444,
0.06622931361198425,
-0.09026601910591125,
-0.08888570219278336,
0.08657511323690414,
0.04626443237066269,
0.10290469229221344,
0.13507691025733948,
0.02754874713718891,
0.1257908195257187,
0.18891699612140656,
0.11640173196792603,
0.03859338536858559,
0.011823236010968685,
-0.042720671743154526,
-0.00948293786495924,
-0.019532080739736557,
-0.03120400384068489,
-0.0015897133853286505,
0.111160509288311,
-0.12125701457262039,
0.0047011650167405605,
-0.17910119891166687,
0.06562753021717072,
0.0389740988612175,
0.033505238592624664,
-0.059628915041685104,
0.019772076979279518,
0.0787830725312233,
0.01838655397295952,
-0.04606233909726143,
0.11887043714523315,
0.008213793858885765,
-0.12780004739761353,
0.12549720704555511,
0.01000938843935728,
0.08338531851768494,
-0.002511385828256607,
0.09371137619018555,
-0.06000813469290733,
-0.13501881062984467,
0.056642718613147736,
0.09899468719959259,
-0.2986595928668976,
0.2900567352771759,
-0.003975317347794771,
-0.047995854169130325,
-0.020688001066446304,
-0.06404721736907959,
0.014033439569175243,
0.24491296708583832,
0.13584493100643158,
0.06061333790421486,
-0.13782411813735962,
-0.08429311960935593,
0.08871324360370636,
0.005843080580234528,
0.1483781933784485,
0.03840288892388344,
-0.05057936906814575,
-0.022661352530121803,
0.005477559752762318,
-0.030877718701958656,
0.026795830577611923,
-0.014663415029644966,
-0.1569199115037918,
0.04796839877963066,
0.01706736907362938,
0.00284274248406291,
-0.02416558563709259,
-0.046898167580366135,
-0.13261452317237854,
0.1027907282114029,
-0.14743095636367798,
-0.022756356745958328,
-0.09537965804338455,
-0.1526433527469635,
0.056947190314531326,
-0.10013826936483383,
0.01784992404282093,
-0.05335329845547676,
-0.11194346100091934,
-0.08915155380964279,
-0.08901619166135788,
0.10981336981058121,
-0.056468259543180466,
0.012069344520568848,
-0.023130988702178,
0.22176940739154816,
-0.052209559828042984,
0.029052745550870895,
0.00965945702046156,
0.007217562757432461,
-0.0511179156601429,
-0.13108868896961212,
-0.013110041618347168,
-0.046934884041547775,
0.012786016799509525,
0.0196582842618227,
-0.08206365257501602,
0.008952477015554905,
0.036491647362709045,
-0.05586567148566246,
0.19850337505340576,
0.28553906083106995,
-0.017203865572810173,
0.1408652812242508,
0.1753707081079483,
-0.09586455672979355,
-0.2800855338573456,
-0.09801515191793442,
-0.20292294025421143,
-0.024142390117049217,
0.01554806251078844,
-0.0954023152589798,
0.06825890392065048,
0.033299192786216736,
-0.0381789430975914,
0.06510796397924423,
-0.26333585381507874,
-0.07794348150491714,
0.13746924698352814,
-0.0368921160697937,
0.43193519115448,
-0.08904437720775604,
-0.11173692345619202,
-0.02658582665026188,
-0.21687684953212738,
0.09352190047502518,
0.01961575448513031,
0.058151260018348694,
-0.0193945225328207,
0.09536103904247284,
0.023044656962156296,
-0.04362506791949272,
0.14327436685562134,
0.037979282438755035,
0.04158949479460716,
-0.0914607122540474,
-0.13056661188602448,
0.100129134953022,
0.026799924671649933,
-0.05319719389081001,
-0.07586289197206497,
-0.02321806363761425,
-0.11046075075864792,
-0.04369066283106804,
-0.10053863376379013,
0.09075199067592621,
-0.02548946999013424,
-0.041978634893894196,
-0.03883076086640358,
0.018127035349607468,
-0.025886287912726402,
-0.017573196440935135,
0.2704150676727295,
-0.05029110983014107,
0.08230139315128326,
0.004314720164984465,
0.12709830701351166,
-0.14783550798892975,
-0.008171911351382732,
-0.09874749928712845,
-0.04918679967522621,
0.013149467296898365,
-0.005821107421070337,
-0.0028859812300652266,
0.16494400799274445,
-0.0016126487171277404,
0.04572661593556404,
0.0781601145863533,
0.00783754326403141,
-0.014598065987229347,
0.12848930060863495,
-0.09755370020866394,
-0.1675342470407486,
-0.005817699711769819,
-0.08751524239778519,
0.14971491694450378,
0.08877339214086533,
0.0955345556139946,
0.055527105927467346,
0.016204895451664925,
-0.0016464091604575515,
-0.036455608904361725,
-0.09143055230379105,
-0.030053269118070602,
0.05294616520404816,
-0.01795961521565914,
-0.09448853880167007,
0.019243614748120308,
-0.03133035823702812,
-0.20829443633556366,
-0.06963088363409042,
0.13842712342739105,
-0.12436096370220184,
-0.09639706462621689,
-0.12784680724143982,
0.0919928327202797,
-0.21755599975585938,
-0.05964827910065651,
-0.008717602118849754,
-0.1604822427034378,
0.040605511516332626,
0.24471047520637512,
0.07297197729349136,
0.09982682019472122,
-0.04153063893318176,
-0.03091159462928772,
0.08050335198640823,
-0.012501533143222332,
-0.0452473945915699,
-0.002638202626258135,
-0.07345890253782272,
0.04825513809919357,
0.011420303955674171,
0.1301589459180832,
-0.04895208403468132,
-0.05325543135404587,
-0.18752546608448029,
0.07700138539075851,
-0.147604301571846,
-0.0692610815167427,
-0.12691113352775574,
-0.025314273312687874,
0.027804870158433914,
-0.14026212692260742,
-0.05573628470301628,
-0.04604241997003555,
-0.1318419724702835,
0.06647725403308868,
0.0716390609741211,
0.10705170780420303,
-0.06175842881202698,
-0.028296874836087227,
0.14095300436019897,
-0.001402225811034441,
0.09704601764678955,
0.1327422857284546,
-0.06940644979476929,
0.14741447567939758,
-0.17518548667430878,
-0.07558301836252213,
0.09423751384019852,
0.03092215582728386,
0.038638632744550705,
0.1198599562048912,
-0.019444664940238,
0.10723024606704712,
0.06588846445083618,
0.08974871784448624,
-0.04277998208999634,
-0.08873608708381653,
0.03429589420557022,
0.05117180943489075,
-0.18947160243988037,
0.0073074460960924625,
-0.046876516193151474,
0.08968362957239151,
0.007447609677910805,
0.12948505580425262,
-0.021867386996746063,
0.03230380639433861,
0.005720936693251133,
0.028984609991312027,
0.019918402656912804,
-0.1457792967557907,
-0.009127207100391388,
-0.12366056442260742,
-0.013681424781680107,
0.01611066795885563,
0.27927160263061523,
-0.02623768150806427,
-0.0764162540435791,
0.050191931426525116,
0.07193838059902191,
-0.07346288114786148,
-0.019987598061561584,
0.1968526840209961,
0.06915696710348129,
-0.008869391866028309,
-0.10535459220409393,
0.037410784512758255,
-0.05285441130399704,
-0.052864495664834976,
0.11264849454164505,
0.16958968341350555,
0.08453905582427979,
0.07006988674402237,
0.04557975009083748,
0.032790061086416245,
-0.04635686054825783,
-0.17187519371509552,
0.06664387881755829,
0.01884235069155693,
-0.068048857152462,
0.20343941450119019,
0.1265532523393631,
-0.05649905651807785,
0.04243641346693039,
-0.03227315470576286,
-0.01831100322306156,
-0.17570850253105164,
-0.15195508301258087,
-0.008638161234557629,
-0.0747823417186737,
0.056934211403131485,
-0.01381763443350792,
0.06902992725372314,
0.10654284805059433,
0.08606341481208801,
-0.04615834355354309,
-0.0010784080950543284,
-0.06384188681840897,
-0.058674849569797516,
-0.030458524823188782,
-0.012111633084714413,
0.08270400017499924,
-0.0993049293756485,
-0.018647413700819016,
-0.1491028219461441,
-0.10106202960014343,
-0.06257490813732147,
0.019904769957065582,
-0.02168901264667511,
-0.05687598139047623,
-0.145304873585701,
-0.04359940066933632,
-0.04170547425746918,
0.10188447684049606,
0.03194250911474228,
0.14200304448604584,
-0.010518691502511501,
0.010889274999499321,
0.015667879953980446,
0.16553902626037598,
-0.018214000388979912,
-0.14261141419410706,
0.05627380684018135,
0.14736489951610565,
0.06894932687282562,
0.11401091516017914,
-0.02089226059615612,
0.03653022274374962,
0.00966223981231451,
0.24988725781440735,
0.29920434951782227,
-0.051639754325151443,
0.05205092206597328,
0.02470596320927143,
0.05283649265766144,
0.12386806309223175,
0.0311481524258852,
0.06846757978200912,
0.2794262170791626,
-0.08743715286254883,
-0.036531202495098114,
-0.03738490119576454,
0.02726895734667778,
-0.10126470774412155,
0.06473054736852646,
-0.0243582371622324,
-0.06823703646659851,
-0.03488235920667648,
0.1050775870680809,
-0.1941498965024948,
0.08699780702590942,
0.09645490348339081,
-0.16091620922088623,
-0.010715305805206299,
-0.028643811121582985,
0.15429970622062683,
0.06591441482305527,
0.07206166535615921,
-0.052024874836206436,
-0.12876613438129425,
0.04546315595507622,
0.029952511191368103,
-0.28354015946388245,
-0.07063569873571396,
0.10782241821289062,
0.052259352058172226,
0.05389448627829552,
-0.022733157500624657,
0.060661327093839645,
0.09383045136928558,
0.0888674333691597,
0.007678474299609661,
0.031690485775470734,
0.054214492440223694,
-0.05917581915855408,
-0.08982314169406891,
-0.11588431149721146,
0.017428414896130562,
-0.0569392554461956,
0.041838180273771286,
-0.07947971671819687,
0.09357766807079315,
0.07167968153953552,
-0.08815281838178635,
-0.040075525641441345,
0.07987511157989502,
-0.10395390540361404,
0.006217112764716148,
0.03726410120725632,
0.043377529829740524,
-0.057750046253204346,
-0.04184121638536453,
-0.04491027817130089,
0.05065533518791199,
-0.09818817675113678,
-0.09175180643796921,
0.007499593310058117,
-0.05019579827785492,
0.09278787672519684,
-0.009661100804805756,
-0.08159857988357544,
-0.05562692508101463,
-0.0343334898352623,
0.09438079595565796,
-0.12782850861549377,
0.03492181375622749,
0.04612846300005913,
0.016658952459692955,
0.030640970915555954,
-0.1980668008327484,
0.04933590069413185,
0.025749148800969124,
-0.06121642142534256,
-0.055691689252853394
] |
null | null |
transformers
|
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on LaBSE and supports different tagsets all using IOBES formats:
1. Wikiann (LOC, PER, ORG)
2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO)
3. NER-UK (LOC, MISC, ORG, PER)
4. Turku (DATE, EVT, LOC, ORG, PER, PRO, TIME)
PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date, GEOPOLIT: Geopolitical,
You can select the tagset to use in the output by configuring the model. This models manages differently uppercase words.
More information about the model can be found in the paper (https://aclanthology.org/2021.bsnlp-1.12.pdf) and GitHub repository (https://github.com/EMBEDDIA/NER_FEDA).
|
{"language": ["multilingual", "uk"], "license": "mit", "tags": ["labse", "ner"]}
| null |
creat89/NER_FEDA_Uk
|
[
"transformers",
"pytorch",
"bert",
"labse",
"ner",
"multilingual",
"uk",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"multilingual",
"uk"
] |
TAGS
#transformers #pytorch #bert #labse #ner #multilingual #uk #license-mit #endpoints_compatible #region-us
|
This is a multilingual NER system trained using a Frustratingly Easy Domain Adaptation architecture. It is based on LaBSE and supports different tagsets all using IOBES formats:
1. Wikiann (LOC, PER, ORG)
2. SlavNER 19/21 (EVT, LOC, ORG, PER, PRO)
3. NER-UK (LOC, MISC, ORG, PER)
4. Turku (DATE, EVT, LOC, ORG, PER, PRO, TIME)
PER: person, LOC: location, ORG: organization, EVT: event, PRO: product, MISC: Miscellaneous, MEDIA: media, ART: Artifact, TIME: time, DATE: date, GEOPOLIT: Geopolitical,
You can select the tagset to use in the output by configuring the model. This models manages differently uppercase words.
More information about the model can be found in the paper (URL and GitHub repository (URL
|
[] |
[
"TAGS\n#transformers #pytorch #bert #labse #ner #multilingual #uk #license-mit #endpoints_compatible #region-us \n"
] |
[
39
] |
[
"passage: TAGS\n#transformers #pytorch #bert #labse #ner #multilingual #uk #license-mit #endpoints_compatible #region-us \n"
] |
[
-0.0443883053958416,
-0.03111727349460125,
-0.00857360940426588,
0.0031851688399910927,
0.08147977292537689,
0.043694525957107544,
0.12573795020580292,
0.08403627574443817,
0.1428106427192688,
0.0166592039167881,
0.13022451102733612,
0.1919403076171875,
-0.05816154554486275,
-0.024060022085905075,
-0.001699995482340455,
-0.2465478926897049,
0.07238607108592987,
0.032879434525966644,
-0.06930974870920181,
0.08793062716722488,
0.11591591686010361,
-0.053443543612957,
0.06847774982452393,
0.023010889068245888,
-0.09131401032209396,
0.044180866330862045,
0.00856421235948801,
-0.05005517229437828,
0.16446621716022491,
0.09297752380371094,
0.13968998193740845,
0.055075887590646744,
-0.05436205863952637,
-0.21769309043884277,
0.023585772141814232,
-0.047488074749708176,
-0.0782928466796875,
0.01978924870491028,
0.026010684669017792,
-0.08995172381401062,
0.11271946877241135,
0.07029707729816437,
-0.016544904559850693,
0.07968421280384064,
-0.13348780572414398,
-0.13351322710514069,
-0.01779269427061081,
0.037381529808044434,
-0.007756373379379511,
0.04980332404375076,
0.008175903931260109,
0.15007992088794708,
-0.15026096999645233,
0.05599077418446541,
0.15754951536655426,
-0.3899775743484497,
0.01447923481464386,
0.07576495409011841,
0.1133384183049202,
0.04544880613684654,
-0.07763239741325378,
0.07757234573364258,
0.04644016921520233,
0.007212092634290457,
-0.0939955785870552,
-0.06164822354912758,
-0.005739604122936726,
0.056654419749975204,
-0.07544815540313721,
-0.0416966937482357,
0.09262708574533463,
-0.01815778762102127,
-0.002387746935710311,
-0.002206408651545644,
-0.05910000577569008,
-0.004064016509801149,
-0.03944685310125351,
-0.007126611191779375,
-0.021920183673501015,
0.0823359489440918,
0.018751218914985657,
-0.05489056557416916,
-0.10967227816581726,
0.06194803863763809,
-0.25595617294311523,
0.2427111715078354,
0.07479111850261688,
0.06254708766937256,
-0.1056901142001152,
0.03832066431641579,
-0.02702520042657852,
-0.014399474486708641,
-0.03633658215403557,
-0.09335199743509293,
0.06964270770549774,
0.03849378600716591,
-0.007195224519819021,
0.13001227378845215,
0.1022440642118454,
0.12109076976776123,
-0.046540193259716034,
0.005374683998525143,
-0.013542592525482178,
0.14933395385742188,
0.008197971619665623,
0.04179992526769638,
0.01826348342001438,
0.03091052547097206,
-0.023825962096452713,
-0.15179066359996796,
0.014598210342228413,
0.013130265288054943,
-0.1310560703277588,
-0.0552237369120121,
-0.07794905453920364,
0.09202886372804642,
-0.034504421055316925,
0.0514877513051033,
-0.08526704460382462,
0.055695272982120514,
0.045859869569540024,
-0.008324303664267063,
-0.002742441836744547,
-0.009642303921282291,
0.057389263063669205,
0.15153467655181885,
-0.03604472801089287,
-0.006093971896916628,
-0.006737000774592161,
0.0678902342915535,
-0.06822454929351807,
0.020854655653238297,
-0.02491719461977482,
-0.02823764830827713,
0.0763327032327652,
-0.10891668498516083,
0.06998088955879211,
-0.1422530859708786,
-0.07560469210147858,
0.03383829444646835,
0.06753263622522354,
0.000938736426178366,
0.04568754509091377,
-0.02461254596710205,
0.0008662194013595581,
0.009197166189551353,
-0.06239799037575722,
-0.10231123119592667,
-0.06802055984735489,
0.09807050228118896,
-0.01624288596212864,
0.03829942271113396,
-0.23163793981075287,
0.03128678351640701,
-0.09646333754062653,
0.014823595061898232,
-0.02157416008412838,
-0.03667515516281128,
-0.06932248920202255,
0.11834071576595306,
-0.03309271112084389,
-0.04585357755422592,
-0.057947032153606415,
0.07135599106550217,
-0.08752696216106415,
0.10981620848178864,
-0.06634571403265,
-0.11582419276237488,
0.21301597356796265,
-0.08838498592376709,
-0.13130277395248413,
0.08613312244415283,
-0.006711555179208517,
0.06195492669939995,
0.08480557799339294,
0.12168603390455246,
0.04166283831000328,
-0.200477734208107,
0.03912419453263283,
0.14897990226745605,
-0.12023475766181946,
-0.12409520894289017,
0.060718514025211334,
-0.04444722831249237,
-0.0810980349779129,
0.06341315805912018,
-0.03806931897997856,
0.13411220908164978,
-0.09904427826404572,
-0.05806012451648712,
-0.00140188611112535,
-0.00669087516143918,
0.09213199466466904,
0.06143280491232872,
0.08958674967288971,
-0.13917574286460876,
-0.04085119813680649,
0.07205670326948166,
-0.0006608613184653223,
0.04675920680165291,
0.08120320737361908,
-0.0824178010225296,
0.0847410187125206,
0.05885850638151169,
-0.051772162318229675,
-0.1336377114057541,
-0.048520930111408234,
-0.04500309005379677,
0.0072501893155276775,
0.09193234145641327,
0.2007310539484024,
0.051264237612485886,
-0.04814521223306656,
-0.05819845572113991,
0.019342593848705292,
0.08323124051094055,
-0.0057902829721570015,
-0.011116739362478256,
-0.11951108276844025,
0.07347844541072845,
-0.03042675368487835,
-0.04162215068936348,
-0.041854191571474075,
-0.011623700149357319,
0.0813821330666542,
0.0760451927781105,
-0.018642999231815338,
0.06348656117916107,
-0.08964437991380692,
0.07173474133014679,
-0.09526370465755463,
0.05800582468509674,
0.1377299726009369,
-0.030057352036237717,
-0.13120198249816895,
0.15623553097248077,
0.006628848612308502,
0.25780555605888367,
0.213164821267128,
-0.2230830043554306,
0.03706083819270134,
-0.03499838709831238,
-0.0634530782699585,
0.05322718620300293,
0.07701438665390015,
0.005229269154369831,
0.049468085169792175,
0.0028846219647675753,
0.07526997476816177,
-0.04768834635615349,
-0.05205206945538521,
0.002414439572021365,
-0.03395453840494156,
-0.10390084236860275,
0.11877945065498352,
0.1288987100124359,
-0.17844301462173462,
0.1703796535730362,
0.2640894055366516,
0.06314177811145782,
0.18820218741893768,
-0.04602198302745819,
0.027775730937719345,
0.03641132265329361,
0.0313747301697731,
-0.05783505737781525,
0.07349207252264023,
-0.25524038076400757,
-0.014293929561972618,
0.06156802177429199,
0.007995049469172955,
0.02929047681391239,
-0.13961714506149292,
-0.08819054067134857,
-0.032921988517045975,
-0.037935417145490646,
-0.09030507504940033,
0.09002289921045303,
-0.03018912300467491,
0.056917306035757065,
-0.02195623330771923,
-0.11476951092481613,
0.12337826937437057,
-0.0021341159008443356,
-0.021561184898018837,
0.12104422599077225,
-0.16489234566688538,
-0.1505480855703354,
-0.11319107562303543,
-0.14721378684043884,
0.018818669021129608,
-0.008170287124812603,
0.08310037106275558,
-0.045089542865753174,
-0.02086292952299118,
0.0887475535273552,
0.07209749519824982,
-0.17023879289627075,
-0.018457025289535522,
-0.12149960547685623,
0.04595242068171501,
-0.11998968571424484,
-0.10843853652477264,
-0.05337873101234436,
-0.04402138292789459,
-0.02648598700761795,
0.09071813523769379,
-0.1394423395395279,
0.03652497008442879,
0.06771333515644073,
0.001604184159077704,
0.0410662516951561,
-0.06088254228234291,
0.26162028312683105,
-0.07558297365903854,
-0.06169801577925682,
0.05665111541748047,
0.043600842356681824,
0.06513789296150208,
0.1972726732492447,
0.093253955245018,
-0.058436326682567596,
-0.05120554566383362,
-0.05329509079456329,
-0.06323692947626114,
-0.2556692361831665,
-0.09195756167173386,
-0.10270613431930542,
0.049668800085783005,
-0.07282190024852753,
0.0727544054389,
0.04566404968500137,
0.032045409083366394,
0.018570654094219208,
-0.14983013272285461,
-0.017578447237610817,
0.05642358586192131,
0.2754708528518677,
-0.05908235162496567,
0.09358623623847961,
-0.0797417089343071,
-0.08285894989967346,
0.09825413674116135,
0.03814253211021423,
0.12593138217926025,
0.1686519980430603,
0.02814156748354435,
0.12212977558374405,
0.16320949792861938,
0.1322070062160492,
0.08933224529027939,
0.013878470286726952,
-0.04795623943209648,
-0.018629005178809166,
-0.006695236079394817,
-0.0033843033015727997,
0.015060726553201675,
0.11655290424823761,
-0.1159256100654602,
0.005144062917679548,
-0.2400323450565338,
0.062336478382349014,
-0.042158905416727066,
0.04178577661514282,
-0.06062609702348709,
0.026472842320799828,
0.05651863291859627,
-0.04605207219719887,
-0.07004758715629578,
0.11602278798818588,
-0.025821147486567497,
-0.15054450929164886,
0.07482967525720596,
-0.0034107721876353025,
0.06634882092475891,
0.05211476981639862,
0.05660397559404373,
-0.0552956722676754,
-0.17889098823070526,
0.06343037635087967,
0.07627712190151215,
-0.3087402284145355,
0.3312089443206787,
0.005009760614484549,
-0.02276959829032421,
0.008814138360321522,
-0.056279901415109634,
-0.031619034707546234,
0.26214733719825745,
0.173406183719635,
0.0539097785949707,
-0.11520490795373917,
-0.1079145222902298,
0.09868872165679932,
0.019998615607619286,
0.12783850729465485,
0.02171233855187893,
-0.030857933685183525,
-0.00346596771851182,
-0.00821088720113039,
-0.02286815643310547,
0.043450091034173965,
-0.05077298730611801,
-0.1576927900314331,
0.026006555184721947,
0.03418398275971413,
0.040284112095832825,
-0.014240902848541737,
-0.06130300089716911,
-0.15386468172073364,
0.12938997149467468,
-0.1923189014196396,
-0.07854445278644562,
-0.10799367725849152,
-0.14119583368301392,
0.06029301881790161,
-0.08079545944929123,
0.03737032040953636,
-0.05422087758779526,
-0.1104201152920723,
-0.09156247228384018,
-0.13628363609313965,
0.13583992421627045,
-0.07919660210609436,
0.0019750664941966534,
-0.058712899684906006,
0.2151688188314438,
-0.007609868422150612,
0.03836429491639137,
-0.02422928623855114,
0.0048594726249575615,
-0.08909960836172104,
-0.07721433788537979,
-0.0014039709931239486,
-0.001042234362103045,
-0.001079237787052989,
-0.012428006157279015,
-0.10875118523836136,
0.05577235296368599,
0.02795126847922802,
-0.06067599728703499,
0.2176428884267807,
0.23330038785934448,
-0.004457545932382345,
0.13329167664051056,
0.17578239738941193,
-0.0639556497335434,
-0.24568775296211243,
-0.11299321800470352,
-0.16120189428329468,
-0.027651753276586533,
-0.02388039231300354,
-0.10866045206785202,
0.10339666157960892,
0.04690801724791527,
-0.04431585222482681,
0.05494004487991333,
-0.22938655316829681,
-0.08964604139328003,
0.13030868768692017,
0.003922967240214348,
0.43183434009552,
-0.09493126720190048,
-0.052524134516716,
-0.02528732270002365,
-0.2487388253211975,
0.15132242441177368,
0.05849163606762886,
0.07376794517040253,
-0.029215142130851746,
0.1021343469619751,
-0.01790623553097248,
-0.03555960953235626,
0.1276356279850006,
0.03239491954445839,
0.012368390336632729,
-0.010628709569573402,
-0.131134495139122,
0.09381087869405746,
0.06916339695453644,
-0.03546249493956566,
-0.05168119817972183,
-0.02227884717285633,
-0.07325021177530289,
-0.04207635298371315,
-0.10937604308128357,
0.10362720489501953,
-0.02157191000878811,
-0.04954204335808754,
-0.0905815064907074,
0.027918457984924316,
-0.05791556090116501,
0.0017050192691385746,
0.27153313159942627,
-0.060126569122076035,
0.13318343460559845,
0.000033295433240709826,
0.11977216601371765,
-0.14591248333454132,
-0.035371556878089905,
-0.0493711419403553,
-0.07574734836816788,
0.04204394668340683,
-0.04656074196100235,
-0.021009068936109543,
0.1637069135904312,
-0.0250858087092638,
0.02997184358537197,
0.07641483098268509,
-0.03411683812737465,
-0.005784336011856794,
0.11194778978824615,
-0.10258263349533081,
-0.1885860115289688,
-0.00017529241449665278,
-0.06049462407827377,
0.14418300986289978,
0.042490921914577484,
0.06290692836046219,
0.04058540239930153,
-0.006816755048930645,
0.00960373505949974,
-0.05394064635038376,
-0.09099522233009338,
0.004256926476955414,
0.0475616455078125,
-0.02064470387995243,
-0.09291303902864456,
-0.02463647350668907,
-0.004995990078896284,
-0.18634441494941711,
-0.05293947458267212,
0.09585949778556824,
-0.12931832671165466,
-0.098225437104702,
-0.0760321170091629,
0.09722448140382767,
-0.2217109352350235,
-0.050831541419029236,
0.038465797901153564,
-0.1431759148836136,
0.036597125232219696,
0.22441110014915466,
0.07541082054376602,
0.10419061779975891,
-0.03771096095442772,
-0.004507656209170818,
0.1247367113828659,
0.01017481368035078,
-0.11615238338708878,
0.0541955940425396,
-0.048397745937108994,
-0.022875159978866577,
-0.0016973656602203846,
0.11710071563720703,
-0.05786103010177612,
-0.043350398540496826,
-0.1492023766040802,
0.05791212618350983,
-0.07069352269172668,
-0.0788756012916565,
-0.10503106564283371,
-0.056771665811538696,
0.03274638578295708,
-0.14033359289169312,
-0.030317282304167747,
-0.08218193799257278,
-0.11144521832466125,
0.02311563491821289,
0.0737898051738739,
0.07736199349164963,
-0.06482022255659103,
-0.005842356476932764,
0.15675047039985657,
0.018646281212568283,
0.09931337088346481,
0.11407094448804855,
-0.03993496298789978,
0.156137615442276,
-0.17659854888916016,
-0.05895092338323593,
0.09457843005657196,
0.023577313870191574,
0.0684560164809227,
0.06655192375183105,
-0.035253752022981644,
0.1077815443277359,
0.08918152004480362,
0.08538956195116043,
-0.03001227416098118,
-0.10133645683526993,
0.02924814075231552,
0.039211202412843704,
-0.17104007303714752,
-0.015256964601576328,
-0.07363927364349365,
0.13485482335090637,
-0.033224448561668396,
0.14608481526374817,
-0.016236914321780205,
0.005142075475305319,
0.009375985711812973,
0.04650220274925232,
0.012539372779428959,
-0.15241321921348572,
-0.04813477024435997,
-0.10774093121290207,
-0.06520745158195496,
-0.050592709332704544,
0.3151349723339081,
-0.009899349883198738,
-0.11281129717826843,
0.07626639306545258,
0.001753230462782085,
-0.054487794637680054,
-0.018049592152237892,
0.2364557385444641,
0.09315843135118484,
0.007324576377868652,
-0.09288602322340012,
0.03554171323776245,
-0.053114790469408035,
-0.07899800688028336,
0.12460973858833313,
0.17410704493522644,
0.05794823542237282,
0.04466145858168602,
0.03346344083547592,
0.06102278456091881,
-0.09000690281391144,
-0.16606968641281128,
0.07396166026592255,
-0.004293758422136307,
-0.04836807772517204,
0.27261099219322205,
0.1962086260318756,
-0.0743870660662651,
0.07801739126443863,
0.0036508829798549414,
-0.0193620678037405,
-0.18342600762844086,
-0.07934283465147018,
-0.02032974362373352,
-0.15305669605731964,
0.06851257383823395,
-0.037224430590867996,
0.02497127652168274,
0.09734928607940674,
0.08062465488910675,
-0.03949001058936119,
0.029286885634064674,
-0.02030632458627224,
-0.050175439566373825,
-0.022342640906572342,
-0.032977908849716187,
0.0709572359919548,
-0.06436360627412796,
-0.01169878151267767,
-0.13930821418762207,
-0.08113247156143188,
-0.06149251013994217,
0.0528176985681057,
-0.09877416491508484,
-0.015431505627930164,
-0.16579455137252808,
-0.0623597726225853,
-0.06659280508756638,
0.11035335808992386,
0.002843548310920596,
0.15384504199028015,
-0.008773083798587322,
0.014617523178458214,
0.007138639222830534,
0.1472489982843399,
-0.008856387808918953,
-0.0913223922252655,
0.044436872005462646,
0.14886407554149628,
0.06375004351139069,
0.14456689357757568,
-0.005373266525566578,
0.029985705390572548,
-0.0626385435461998,
0.28298231959342957,
0.30701738595962524,
-0.041866958141326904,
0.032552797347307205,
0.05729331448674202,
0.0498838871717453,
0.09763623028993607,
0.04350404813885689,
0.09507111459970474,
0.2821650505065918,
-0.12271733582019806,
-0.0731431394815445,
-0.0281531885266304,
0.042393505573272705,
-0.1482304185628891,
0.06420999020338058,
0.020908333361148834,
-0.11077714711427689,
-0.06383457779884338,
0.07452179491519928,
-0.17101038992404938,
0.1024400144815445,
0.03765791654586792,
-0.10818939656019211,
-0.02699446864426136,
-0.03542528301477432,
0.1431645303964615,
0.060563527047634125,
0.06372226774692535,
-0.048140913248062134,
-0.10727736353874207,
0.025648077949881554,
0.02518484555184841,
-0.2877611219882965,
-0.037593159824609756,
0.13725945353507996,
0.04113301634788513,
0.04327455163002014,
-0.0027579437009990215,
0.05116603150963783,
0.0815761610865593,
0.0738169252872467,
-0.003670679870992899,
0.04136170819401741,
0.052207618951797485,
-0.029110148549079895,
-0.08380785584449768,
-0.0703936517238617,
0.05859772115945816,
-0.071839839220047,
0.06312081962823868,
-0.03919653594493866,
0.09471087157726288,
0.060945864766836166,
-0.08402280509471893,
-0.07807782292366028,
0.10731973499059677,
-0.11811380833387375,
0.03604375943541527,
0.0208913441747427,
0.0429772324860096,
-0.061462413519620895,
-0.05290345102548599,
-0.037558138370513916,
0.058417078107595444,
-0.06802675873041153,
-0.11769861727952957,
-0.018011026084423065,
-0.057942334562540054,
0.0985715314745903,
-0.022827453911304474,
-0.13333620131015778,
-0.041012488305568695,
-0.051463913172483444,
0.1082782968878746,
-0.13537821173667908,
0.030424373224377632,
0.11164722591638565,
-0.008810766041278839,
0.055523019284009933,
-0.17861251533031464,
0.04935001954436302,
0.014276864938437939,
-0.07467988133430481,
-0.060280296951532364
] |
null | null |
transformers
|
# MyModel
## Model description
This is the `BART-TL-all` model from the paper [BART-TL: Weakly-Supervised Topic Label Generation](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf). We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works.
For more details not covered here, you can read the paper or look at the open-source implementation: https://github.com/CristianViorelPopa/BART-TL-topic-label-generation.
There are two models made available from the paper:
* [BART-TL-all](https://huggingface.co/cristian-popa/bart-tl-all)
* [BART-TL-ng](https://huggingface.co/cristian-popa/bart-tl-ng)
## Intended uses & limitations
#### How to use
The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
mname = "cristian-popa/bart-tl-all"
tokenizer = AutoTokenizer.from_pretrained(mname)
model = AutoModelForSeq2SeqLM.from_pretrained(mname)
input = "site web google search website online internet social content user"
enc = tokenizer(input, return_tensors="pt", truncation=True, padding="max_length", max_length=128)
outputs = model.generate(
input_ids=enc.input_ids,
attention_mask=enc.attention_mask,
max_length=15,
min_length=1,
do_sample=False,
num_beams=25,
length_penalty=1.0,
repetition_penalty=1.5
)
decoded = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(decoded) # application programming interface
```
#### Limitations and bias
The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.
## Training data
The model was fine-tuned on 5 different StackExchange corpora (see https://archive.org/download/stackexchange for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.
## Training procedure
The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the [NETL](https://www.aclweb.org/anthology/C16-1091.pdf) method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the [paper](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf) or by following [this notebook](https://github.com/CristianViorelPopa/BART-TL-topic-label-generation/blob/main/notebooks/end_to_end_workflow.ipynb).
## Eval results
model | Top-1 Avg. | Top-3 Avg. | Top-5 Avg. | nDCG-1 | nDCG-3 | nDCG-5
------------|------------|------------|------------|--------|--------|-------
NETL (U) | 2.66 | 2.59 | 2.50 | 0.83 | 0.85 | 0.87
NETL (S) | 2.74 | 2.57 | 2.49 | 0.88 | 0.85 | 0.88
BART-TL-all | 2.64 | 2.52 | 2.43 | 0.83 | 0.84 | 0.87
BART-TL-ng | 2.62 | 2.50 | 2.33 | 0.82 | 0.84 | 0.85
### BibTeX entry and citation info
```bibtex
@inproceedings{popa-rebedea-2021-bart,
title = "{BART}-{TL}: Weakly-Supervised Topic Label Generation",
author = "Popa, Cristian and
Rebedea, Traian",
booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume",
month = apr,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2021.eacl-main.121",
pages = "1418--1425",
abstract = "We propose a novel solution for assigning labels to topic models by using multiple weak labelers. The method leverages generative transformers to learn accurate representations of the most important topic terms and candidate labels. This is achieved by fine-tuning pre-trained BART models on a large number of potential labels generated by state of the art non-neural models for topic labeling, enriched with different techniques. The proposed BART-TL model is able to generate valuable and novel labels in a weakly-supervised manner and can be improved by adding other weak labelers or distant supervision on similar tasks.",
}
```
|
{"language": ["en"], "license": "apache-2.0", "tags": ["topic labeling"], "metrics": ["ndcg"], "<!-- thumbnail": "https://raw.githubusercontent.com/JetRunner/BERT-of-Theseus/master/bert-of-theseus.png -->"}
|
text2text-generation
|
cristian-popa/bart-tl-all
|
[
"transformers",
"pytorch",
"bart",
"text2text-generation",
"topic labeling",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
MyModel
=======
Model description
-----------------
This is the 'BART-TL-all' model from the paper BART-TL: Weakly-Supervised Topic Label Generation. We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works.
For more details not covered here, you can read the paper or look at the open-source implementation: URL
There are two models made available from the paper:
* BART-TL-all
* BART-TL-ng
Intended uses & limitations
---------------------------
#### How to use
The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.
#### Limitations and bias
The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.
Training data
-------------
The model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.
Training procedure
------------------
The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.
Eval results
------------
### BibTeX entry and citation info
|
[
"#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.",
"#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------",
"### BibTeX entry and citation info"
] |
[
"TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.",
"#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------",
"### BibTeX entry and citation info"
] |
[
52,
54,
236,
11
] |
[
"passage: TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with other heuristic labels, such as n-grams from the topics, relevant sentences in the corpora and noun phrases. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------### BibTeX entry and citation info"
] |
[
-0.025374848395586014,
0.039004284888505936,
0.0012018803972750902,
-0.0021390828769654036,
0.05627782270312309,
-0.012651707045733929,
0.08196865767240524,
0.07572115957736969,
-0.014824253506958485,
0.05512076988816261,
-0.005266357213258743,
-0.028726065531373024,
0.06058561056852341,
0.09164533764123917,
0.06597661972045898,
-0.23688039183616638,
-0.02427762560546398,
-0.019137242808938026,
0.025111205875873566,
0.07413837313652039,
0.0801633968949318,
-0.07733819633722305,
0.05459797382354736,
0.03504233807325363,
-0.11093968152999878,
0.014345529489219189,
-0.014009845443069935,
-0.04866432398557663,
0.08257357776165009,
0.0835619866847992,
0.07595868408679962,
-0.0010336431441828609,
0.020149867981672287,
-0.16520194709300995,
0.012523103505373001,
0.05465930700302124,
-0.03233037516474724,
0.04172619804739952,
-0.008979108184576035,
-0.015063101425766945,
0.2015094757080078,
-0.12185148149728775,
0.07069770991802216,
0.034489769488573074,
-0.10205763578414917,
-0.1403585970401764,
-0.0698171928524971,
0.048952650278806686,
0.06267505884170532,
0.0733892172574997,
-0.03587985783815384,
0.010870910249650478,
-0.12984201312065125,
0.027630822733044624,
0.10064592212438583,
-0.19328588247299194,
-0.023060372099280357,
0.07223936915397644,
0.02025728113949299,
0.14581447839736938,
-0.045333217829465866,
0.021456744521856308,
0.04448847100138664,
-0.009807242080569267,
0.024089299142360687,
0.012852762825787067,
0.040141187608242035,
-0.01932772621512413,
-0.12328711152076721,
-0.018781736493110657,
0.19605806469917297,
0.033163152635097504,
-0.025639932602643967,
-0.09235469996929169,
-0.012363945133984089,
0.00573913287371397,
-0.005389542318880558,
-0.04006806015968323,
-0.006444129161536694,
-0.030662275850772858,
0.08311662822961807,
-0.09281859546899796,
-0.10850023478269577,
-0.011667453683912754,
0.0045842742547392845,
0.0892564207315445,
-0.013522819615900517,
0.018361195921897888,
-0.037462279200553894,
0.08351389318704605,
-0.027456682175397873,
-0.0720236599445343,
-0.005998218432068825,
-0.026426732540130615,
-0.06898880004882812,
-0.04205002263188362,
-0.02252735197544098,
0.009246132336556911,
-0.05749483406543732,
0.08528144657611847,
-0.004157711286097765,
-0.03171727433800697,
-0.05008109286427498,
0.021449534222483635,
0.042779967188835144,
0.06744495034217834,
-0.04761285334825516,
-0.05863207206130028,
0.036818116903305054,
0.0028639717493206263,
0.0028553286101669073,
-0.018386458978056908,
-0.07960375398397446,
0.06726165860891342,
-0.0008613115642219782,
0.011409942060709,
-0.007289382163435221,
0.013667714782059193,
-0.013623302802443504,
-0.0645662248134613,
0.10396476089954376,
-0.096501424908638,
-0.027812102809548378,
0.005944224074482918,
0.0001212509159813635,
0.0053077274933457375,
0.034477945417165756,
-0.018044035881757736,
-0.024628138169646263,
0.1134599968791008,
-0.06307589262723923,
-0.04186088591814041,
-0.07820724695920944,
-0.08072818070650101,
-0.010292606428265572,
-0.027942879125475883,
-0.03270455822348595,
-0.09064894914627075,
-0.20244337618350983,
-0.03909974545240402,
-0.01929166167974472,
-0.04113118350505829,
-0.006094860844314098,
-0.01768438331782818,
0.019536267966032028,
-0.015020006336271763,
-0.01866171509027481,
0.04201359301805496,
-0.047658275812864304,
0.0019734289962798357,
-0.030442634597420692,
0.07271069288253784,
-0.001739004859700799,
0.04161064699292183,
-0.09151079505681992,
-0.019997790455818176,
-0.08096830546855927,
0.09918474406003952,
-0.09743387997150421,
0.040331318974494934,
-0.0501474030315876,
-0.06316972523927689,
-0.05828740447759628,
0.008384767919778824,
-0.032446131110191345,
0.11862676590681076,
-0.2411566525697708,
-0.10155875980854034,
0.22188405692577362,
-0.13674941658973694,
-0.04416818171739578,
0.09289900958538055,
-0.058656495064496994,
0.19051308929920197,
0.13308261334896088,
0.1484132707118988,
0.11054572463035583,
-0.05419793725013733,
0.012045875191688538,
0.00807588268071413,
0.024857141077518463,
0.124969482421875,
0.029876917600631714,
-0.029520142823457718,
-0.08425776660442352,
0.03338806703686714,
-0.0417444184422493,
-0.010980161838233471,
-0.041108306497335434,
-0.1037934347987175,
0.06212481111288071,
0.005830963607877493,
0.03267896920442581,
0.005314937327057123,
0.02663646638393402,
0.01572316326200962,
-0.07575704157352448,
0.04886321723461151,
0.06014768034219742,
-0.054265134036540985,
0.0027652904391288757,
-0.023712513968348503,
0.08095517754554749,
-0.03142314404249191,
-0.005250903312116861,
-0.15482088923454285,
-0.05784992873668671,
0.022354478016495705,
-0.057889532297849655,
0.10083726793527603,
0.12827707827091217,
-0.006238583941012621,
0.036908309906721115,
-0.028443872928619385,
0.04381833225488663,
-0.04617217555642128,
0.017520533874630928,
-0.0470028780400753,
-0.14531449973583221,
-0.02311975695192814,
-0.062405284494161606,
0.17908786237239838,
-0.1412356197834015,
0.0066263931803405285,
0.06552587449550629,
0.009014279581606388,
0.01885327138006687,
-0.02184450812637806,
-0.018188288435339928,
0.04846247285604477,
-0.002647588262334466,
-0.039111748337745667,
0.040195975452661514,
0.028115347027778625,
-0.12819229066371918,
-0.02621748298406601,
-0.10436825454235077,
-0.05807815492153168,
0.08593451976776123,
0.06871113926172256,
-0.1004435196518898,
-0.08555133640766144,
-0.05906803160905838,
-0.02791197970509529,
-0.043020639568567276,
-0.05431250110268593,
0.15854454040527344,
0.011350056156516075,
0.058783020824193954,
-0.08793613314628601,
-0.026198498904705048,
0.021638697013258934,
0.007649638224393129,
0.02080201357603073,
0.0329265370965004,
0.04627590626478195,
-0.14385628700256348,
0.059971995651721954,
0.060213491320610046,
0.029332660138607025,
0.2260117083787918,
0.05588638782501221,
-0.07137778401374817,
0.014338353648781776,
-0.010641013272106647,
0.0023751931730657816,
0.12133446335792542,
-0.15119768679141998,
0.0012428713962435722,
0.017992205917835236,
-0.012768971733748913,
0.06418664008378983,
-0.11056609451770782,
0.02154443971812725,
0.04523370414972305,
-0.03928098827600479,
0.01849675178527832,
-0.0439380407333374,
-0.03175586089491844,
0.112209752202034,
0.0735037699341774,
0.021935664117336273,
-0.01748288795351982,
-0.05216015502810478,
-0.13867239654064178,
0.21151670813560486,
-0.09573441743850708,
-0.2929095923900604,
-0.0642627626657486,
0.08207657933235168,
0.004545880481600761,
0.004897106904536486,
0.03593382611870766,
-0.12028618156909943,
-0.024446604773402214,
-0.09148507565259933,
0.00596793694421649,
0.06675173342227936,
-0.04713333770632744,
-0.04167424887418747,
-0.029828494414687157,
-0.041587941348552704,
-0.11556001752614975,
-0.024811800569295883,
-0.08021716773509979,
-0.09021401405334473,
0.014055823907256126,
-0.022558089345693588,
0.07490658760070801,
0.15941175818443298,
0.028983348980545998,
-0.011348794214427471,
-0.026321861892938614,
0.1380263715982437,
-0.0872647762298584,
0.05129505693912506,
0.0549904890358448,
0.04482033848762512,
0.037481676787137985,
0.08293652534484863,
0.04003308340907097,
-0.056550282984972,
0.01725054532289505,
0.06478191912174225,
-0.06773606687784195,
-0.19421544671058655,
-0.15415558218955994,
-0.04637949541211128,
-0.025978153571486473,
0.028528593480587006,
0.028891142457723618,
0.12461499124765396,
-0.0037729176692664623,
-0.027297021821141243,
-0.03139037638902664,
0.009424341842532158,
0.08151570707559586,
0.04585656151175499,
-0.04042676091194153,
0.07826856523752213,
-0.023666907101869583,
-0.03574433550238609,
0.11165247112512589,
-0.010570734739303589,
0.26734036207199097,
0.038922760635614395,
0.11020205914974213,
0.08978895843029022,
0.007828119210898876,
0.034113384783267975,
0.0389973483979702,
-0.01658691093325615,
-0.004989570006728172,
-0.04169847071170807,
-0.11096638441085815,
-0.012294625863432884,
0.12519261240959167,
-0.0356748066842556,
-0.05450219660997391,
-0.0623164027929306,
-0.030980275943875313,
0.060690637677907944,
0.19402699172496796,
0.016241122037172318,
-0.15273137390613556,
-0.07467057555913925,
0.014522177167236805,
0.01643136702477932,
-0.026698289439082146,
0.01985657960176468,
0.0745992586016655,
-0.1383182853460312,
0.03738125413656235,
-0.047442276030778885,
0.05590251460671425,
-0.016457967460155487,
0.025810606777668,
0.0573166087269783,
-0.04152926802635193,
-0.02078746259212494,
0.1067386120557785,
-0.11537405103445053,
0.17401033639907837,
0.01359597872942686,
0.03443432226777077,
-0.10598739981651306,
-0.022150877863168716,
-0.005321873817592859,
0.01936442404985428,
0.16003969311714172,
0.03688090667128563,
-0.0187061857432127,
-0.046399299055337906,
-0.0534193329513073,
0.04812349006533623,
0.03411715850234032,
-0.08110268414020538,
0.0553300715982914,
0.035752974450588226,
0.017497962340712547,
0.019154423847794533,
-0.0239311121404171,
-0.11697561293840408,
-0.17434340715408325,
0.059314996004104614,
-0.10086710751056671,
-0.04616732895374298,
-0.039796214550733566,
-0.08101294189691544,
0.06058913841843605,
0.2858569920063019,
-0.11056961119174957,
-0.0976484939455986,
-0.0931582823395729,
0.07961542904376984,
0.025420071557164192,
-0.04420466721057892,
0.009872093796730042,
0.054034892469644547,
0.12028951197862625,
-0.02685481496155262,
-0.12529999017715454,
0.025645487010478973,
-0.06017879396677017,
-0.08150379359722137,
-0.06256409734487534,
0.1256570667028427,
0.11306334286928177,
0.02061164192855358,
0.019414573907852173,
-0.040288038551807404,
-0.01881306618452072,
-0.13894546031951904,
-0.00013835493882652372,
0.17283155024051666,
0.03272166848182678,
0.06869496405124664,
-0.10351663082838058,
0.001994180493056774,
-0.08153092116117477,
-0.018127717077732086,
0.13208991289138794,
0.20465973019599915,
-0.09127013385295868,
0.13338486850261688,
0.1468401700258255,
-0.13076111674308777,
-0.155958354473114,
0.0859539583325386,
0.03101331554353237,
0.016179265454411507,
0.0023275427520275116,
-0.19987569749355316,
0.13098886609077454,
0.09469186514616013,
-0.006818537600338459,
0.018067440018057823,
-0.21116642653942108,
-0.13393408060073853,
0.07478226721286774,
0.012571830302476883,
0.0390593521296978,
-0.10043828189373016,
-0.02606084942817688,
-0.006132635287940502,
-0.06268465518951416,
0.2116919457912445,
-0.17832006514072418,
0.0579785592854023,
0.003898933995515108,
0.027896972373127937,
0.010773023590445518,
-0.03931251913309097,
0.08484436571598053,
0.09473834931850433,
0.06972039490938187,
0.001527162967249751,
-0.019097549840807915,
0.1040133535861969,
0.0012857409892603755,
0.03482336550951004,
0.05610427260398865,
0.03022756241261959,
-0.09981384873390198,
-0.05043415725231171,
-0.12163370102643967,
0.0856333076953888,
-0.07691457122564316,
-0.08862898498773575,
-0.07717464119195938,
0.12471956759691238,
0.10682511329650879,
-0.01807301677763462,
-0.10433310270309448,
-0.10616196691989899,
0.09775178134441376,
0.02757222205400467,
0.10222077369689941,
-0.06222504749894142,
-0.07527535408735275,
0.03059513121843338,
-0.00356341153383255,
0.10012028366327286,
-0.07052880525588989,
0.01892682909965515,
0.0911131352186203,
0.0434868298470974,
0.14623934030532837,
0.03134980797767639,
-0.09305533021688461,
-0.022113433107733727,
0.06249277666211128,
-0.1434045135974884,
-0.10386054217815399,
-0.0667789876461029,
-0.05403197184205055,
-0.05982176214456558,
-0.06637430191040039,
0.04598575085401535,
-0.04939538612961769,
-0.007862349972128868,
-0.03054654225707054,
0.040938373655080795,
-0.03713683784008026,
0.0963565930724144,
0.015971768647432327,
0.0425214022397995,
-0.0581410713493824,
0.015348481014370918,
0.07869705557823181,
-0.12279696017503738,
0.011533697135746479,
0.08363792300224304,
-0.09767021238803864,
-0.03269037976861,
-0.09035198390483856,
0.10608736425638199,
-0.1637578010559082,
-0.01416875422000885,
-0.018254723399877548,
-0.08534263074398041,
-0.0008854248444549739,
0.10918448120355606,
0.023284228518605232,
0.011308975517749786,
-0.04954603686928749,
0.01783323846757412,
-0.029697829857468605,
0.0575653575360775,
-0.04117586836218834,
0.006940060760825872,
0.005735950544476509,
0.13801130652427673,
0.037868618965148926,
0.04139934480190277,
-0.03202034533023834,
-0.04953631013631821,
-0.09296920895576477,
-0.005610780790448189,
-0.08332479745149612,
0.0232798270881176,
-0.09166106581687927,
0.007772115524858236,
-0.030092796310782433,
0.06940682232379913,
0.026405690237879753,
-0.01912532187998295,
-0.038181766867637634,
0.008061409927904606,
-0.0184697937220335,
0.02613282948732376,
-0.0988495871424675,
0.03648442402482033,
0.013355836272239685,
-0.05714363604784012,
0.06967658549547195,
-0.03552120551466942,
0.05073747783899307,
-0.0042486232705414295,
-0.15376003086566925,
0.04435231536626816,
0.05423755943775177,
0.06662016361951828,
0.035429757088422775,
-0.11604003608226776,
0.013955638743937016,
0.03035620041191578,
-0.006747977342456579,
-0.013334184885025024,
-0.025433465838432312,
-0.09370122104883194,
0.09390179812908173,
-0.020380087196826935,
-0.002058945596218109,
-0.07326298952102661,
0.04332165792584419,
0.03923889622092247,
0.032391343265771866,
0.04491370916366577,
-0.06826385855674744,
0.06088077276945114,
-0.11740238964557648,
-0.05905755236744881,
0.0011629258515313268,
0.022845670580863953,
-0.0019516521133482456,
-0.05847083777189255,
0.037387289106845856,
0.014355680905282497,
0.25390881299972534,
0.062093090265989304,
0.069460429251194,
0.020268874242901802,
-0.1339385062456131,
0.06423936039209366,
-0.007607430685311556,
0.071237713098526,
0.030539395287632942,
-0.019249018281698227,
0.027528604492545128,
-0.002912512980401516,
-0.052990030497312546,
0.07733967900276184,
0.26960861682891846,
0.09317513555288315,
0.07002261281013489,
-0.01596338488161564,
0.05849657207727432,
-0.09895317256450653,
-0.11765959858894348,
-0.03474835678935051,
0.015350895933806896,
0.09132018685340881,
-0.05910055711865425,
-0.020705198869109154,
0.1265740543603897,
-0.1816927045583725,
0.0885203629732132,
0.023681990802288055,
-0.09824202954769135,
-0.10433228313922882,
-0.11110048741102219,
-0.022911371663212776,
-0.023831717669963837,
0.005452331621199846,
-0.15023021399974823,
0.009538368321955204,
-0.003261393401771784,
0.0014824530808255076,
-0.06816113740205765,
0.12427623569965363,
-0.04228925332427025,
-0.06637547165155411,
0.08012707531452179,
0.03214878961443901,
0.00829031690955162,
-0.09652137756347656,
0.012890353798866272,
0.023357631638646126,
0.05309714749455452,
0.007246872875839472,
0.0351165309548378,
-0.010146970860660076,
0.0496409609913826,
-0.011189096607267857,
-0.02576020546257496,
-0.0024872974026948214,
-0.021840723231434822,
0.08070813119411469,
0.1502762734889984,
0.06309176236391068,
-0.07360146194696426,
0.0024960616137832403,
0.1180528923869133,
-0.03477959707379341,
-0.05430341511964798,
-0.12513689696788788,
0.2332977056503296,
-0.0047600348480045795,
0.016306515783071518,
-0.0002442345430608839,
-0.0810641348361969,
0.04781652241945267,
0.2072199285030365,
0.1980423480272293,
-0.0047171469777822495,
0.007482382468879223,
-0.020065026357769966,
0.00020159027189947665,
-0.014014928601682186,
0.12610025703907013,
0.015009094029664993,
0.28122517466545105,
-0.06580141931772232,
0.04875817522406578,
-0.023796308785676956,
-0.05780494213104248,
-0.05968209728598595,
0.04789343848824501,
0.06261160224676132,
0.008001189678907394,
-0.04467308893799782,
0.05927058681845665,
-0.030398819595575333,
-0.19415849447250366,
0.06482802331447601,
-0.0885695368051529,
-0.08933869004249573,
-0.02497423067688942,
-0.08924619108438492,
-0.03111174702644348,
0.032460931688547134,
-0.03607760742306709,
-0.06338637322187424,
0.08439449220895767,
0.018250491470098495,
-0.05676942318677902,
-0.10335543751716614,
0.1096220314502716,
-0.03972944989800453,
0.16455940902233124,
-0.01261191163212061,
0.04919590428471565,
0.07728080451488495,
0.052128225564956665,
-0.0696154311299324,
0.02687336876988411,
0.014175341464579105,
-0.039005499333143234,
0.028385547921061516,
0.09246441721916199,
-0.005342214368283749,
0.07834252715110779,
0.054535411298274994,
-0.08795330673456192,
0.045932330191135406,
0.0032774799037724733,
0.0011050270404666662,
-0.09536474943161011,
0.019846906885504723,
-0.07773665338754654,
0.18985538184642792,
0.18354102969169617,
-0.031886324286460876,
0.009509732015430927,
-0.05681654438376427,
0.009294232353568077,
0.017961733043193817,
0.1746419370174408,
-0.03700302913784981,
-0.10404302924871445,
0.012524577789008617,
0.010034637525677681,
0.030655305832624435,
-0.261962354183197,
-0.07351183146238327,
0.0522366426885128,
-0.02872975543141365,
0.013980776071548462,
0.09918464720249176,
0.044647157192230225,
0.027774231508374214,
-0.046033378690481186,
-0.10648221522569656,
-0.013203945942223072,
0.10111916810274124,
-0.12471795827150345,
-0.06752505153417587
] |
null | null |
transformers
|
# MyModel
## Model description
This is the `BART-TL-ng` model from the paper [BART-TL: Weakly-Supervised Topic Label Generation](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf). We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works.
For more details not covered here, you can read the paper or look at the open-source implementation: https://github.com/CristianViorelPopa/BART-TL-topic-label-generation.
There are two models made available from the paper:
* [BART-TL-all](https://huggingface.co/cristian-popa/bart-tl-all)
* [BART-TL-ng](https://huggingface.co/cristian-popa/bart-tl-ng)
## Intended uses & limitations
#### How to use
The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.
```python
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
mname = "cristian-popa/bart-tl-ng"
tokenizer = AutoTokenizer.from_pretrained(mname)
model = AutoModelForSeq2SeqLM.from_pretrained(mname)
input = "site web google search website online internet social content user"
enc = tokenizer(input, return_tensors="pt", truncation=True, padding="max_length", max_length=128)
outputs = model.generate(
input_ids=enc.input_ids,
attention_mask=enc.attention_mask,
max_length=15,
min_length=1,
do_sample=False,
num_beams=25,
length_penalty=1.0,
repetition_penalty=1.5
)
decoded = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(decoded) # windows live messenger
```
#### Limitations and bias
The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.
## Training data
The model was fine-tuned on 5 different StackExchange corpora (see https://archive.org/download/stackexchange for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.
## Training procedure
The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the [NETL](https://www.aclweb.org/anthology/C16-1091.pdf) method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the [paper](https://www.aclweb.org/anthology/2021.eacl-main.121.pdf) or by following [this notebook](https://github.com/CristianViorelPopa/BART-TL-topic-label-generation/blob/main/notebooks/end_to_end_workflow.ipynb).
## Eval results
model | Top-1 Avg. | Top-3 Avg. | Top-5 Avg. | nDCG-1 | nDCG-3 | nDCG-5
------------|------------|------------|------------|--------|--------|-------
NETL (U) | 2.66 | 2.59 | 2.50 | 0.83 | 0.85 | 0.87
NETL (S) | 2.74 | 2.57 | 2.49 | 0.88 | 0.85 | 0.88
BART-TL-all | 2.64 | 2.52 | 2.43 | 0.83 | 0.84 | 0.87
BART-TL-ng | 2.62 | 2.50 | 2.33 | 0.82 | 0.84 | 0.85
### BibTeX entry and citation info
```bibtex
@inproceedings{popa-rebedea-2021-bart,
title = "{BART}-{TL}: Weakly-Supervised Topic Label Generation",
author = "Popa, Cristian and
Rebedea, Traian",
booktitle = "Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume",
month = apr,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2021.eacl-main.121",
pages = "1418--1425",
abstract = "We propose a novel solution for assigning labels to topic models by using multiple weak labelers. The method leverages generative transformers to learn accurate representations of the most important topic terms and candidate labels. This is achieved by fine-tuning pre-trained BART models on a large number of potential labels generated by state of the art non-neural models for topic labeling, enriched with different techniques. The proposed BART-TL model is able to generate valuable and novel labels in a weakly-supervised manner and can be improved by adding other weak labelers or distant supervision on similar tasks.",
}
```
|
{"language": ["en"], "license": "apache-2.0", "tags": ["topic labeling"], "metrics": ["ndcg"], "<!-- thumbnail": "https://raw.githubusercontent.com/JetRunner/BERT-of-Theseus/master/bert-of-theseus.png -->"}
|
text2text-generation
|
cristian-popa/bart-tl-ng
|
[
"transformers",
"pytorch",
"bart",
"text2text-generation",
"topic labeling",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
MyModel
=======
Model description
-----------------
This is the 'BART-TL-ng' model from the paper BART-TL: Weakly-Supervised Topic Label Generation. We aim to solve the topic labeling task using generative methods, rather than selection from a pool of labels as was done in previous State of the Art works.
For more details not covered here, you can read the paper or look at the open-source implementation: URL
There are two models made available from the paper:
* BART-TL-all
* BART-TL-ng
Intended uses & limitations
---------------------------
#### How to use
The model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.
#### Limitations and bias
The model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.
Training data
-------------
The model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.
Training procedure
------------------
The large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.
Eval results
------------
### BibTeX entry and citation info
|
[
"#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.",
"#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------",
"### BibTeX entry and citation info"
] |
[
"TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.",
"#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------",
"### BibTeX entry and citation info"
] |
[
52,
54,
215,
11
] |
[
"passage: TAGS\n#transformers #pytorch #bart #text2text-generation #topic labeling #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n#### How to use\n\n\nThe model takes in a topic, represented as a space-separated series of words. Such topics can be generated using LDA, as was done for gathering the fine-tuning dataset for the model.#### Limitations and bias\n\n\nThe model may not generate accurate labels for topics from domains unrelated to the ones it was fine-tuned on, such as gastronomy.\n\n\nTraining data\n-------------\n\n\nThe model was fine-tuned on 5 different StackExchange corpora (see URL for a full list of existing such corpora): English, biology, economics, law, and photography. 100 topics are extracted using LDA for each of these corpora, filtered for coherence and then used for obtaining the final model here.\n\n\nTraining procedure\n------------------\n\n\nThe large Facebook BART model is fine-tuned in a weakly-supervised manner, making use of the unsupervised candidate selection of the NETL method, along with n-grams from the topics. The dataset is a one-to-many mapping from topics to labels. More details on training and parameters can be found in the paper or by following this notebook.\n\n\nEval results\n------------### BibTeX entry and citation info"
] |
[
-0.058342840522527695,
0.030573591589927673,
0.0009910708758980036,
0.029448406770825386,
0.09407609701156616,
-0.006430593319237232,
0.06431687623262405,
0.09990748763084412,
-0.020535092800855637,
0.04680928215384483,
0.009751266799867153,
0.00364040769636631,
0.06301063299179077,
0.19909220933914185,
0.07533171772956848,
-0.24952134490013123,
-0.02668389491736889,
-0.031093338504433632,
-0.021916141733527184,
0.09100772440433502,
0.09117823094129562,
-0.0978337973356247,
0.06438127160072327,
0.0327000729739666,
-0.13304725289344788,
0.006562351249158382,
-0.01713912934064865,
-0.07628340274095535,
0.08511122316122055,
0.06718479841947556,
0.09355984628200531,
-0.023052550852298737,
0.056373778730630875,
-0.15545515716075897,
0.023145727813243866,
0.07129095494747162,
-0.01641656830906868,
0.06063896417617798,
0.0487065427005291,
-0.020678333938121796,
0.17517170310020447,
-0.125321164727211,
0.06931795924901962,
0.060781676322221756,
-0.08715667575597763,
-0.14160868525505066,
-0.08093162626028061,
0.06786641478538513,
0.08159607648849487,
0.07530500739812851,
-0.02406216226518154,
-0.01739644818007946,
-0.10106674581766129,
0.037807609885931015,
0.13328221440315247,
-0.22689566016197205,
-0.04279904440045357,
0.09441834688186646,
0.016648000106215477,
0.08622246980667114,
-0.043103452771902084,
0.025955630466341972,
0.025359584018588066,
0.008822071366012096,
0.0627066045999527,
0.00790627021342516,
0.025081414729356766,
-0.04929721727967262,
-0.12087327986955643,
-0.03505789861083031,
0.1847388595342636,
0.02671620436012745,
-0.05503256618976593,
-0.12186971306800842,
-0.02728126011788845,
0.038787972182035446,
-0.017138266935944557,
-0.048493124544620514,
0.0019641402177512646,
-0.0023294149432331324,
0.039050959050655365,
-0.12985970079898834,
-0.11857695132493973,
-0.02382560819387436,
-0.04059574753046036,
0.02126193605363369,
-0.01126590371131897,
0.03685931861400604,
-0.0706740990281105,
0.12561112642288208,
-0.04233065992593765,
-0.06131984665989876,
0.024106791242957115,
-0.031216681003570557,
-0.0751732662320137,
-0.03797106444835663,
-0.003976340871304274,
-0.007467689923942089,
-0.04018536955118179,
0.07222623378038406,
-0.0018510435475036502,
-0.031280551105737686,
-0.07724938541650772,
0.020403580740094185,
0.08658487349748611,
0.04295308142900467,
-0.08782842010259628,
-0.049130238592624664,
0.035734597593545914,
-0.016364876180887222,
-0.020161747932434082,
-0.006849395576864481,
-0.055540211498737335,
0.028979478403925896,
0.002882142784073949,
0.037599075585603714,
0.011118223890662193,
-0.026767058297991753,
-0.033568646758794785,
-0.08530446887016296,
0.1328819841146469,
-0.09543154388666153,
-0.012268774211406708,
0.0034273075871169567,
-0.02129155769944191,
0.019330838695168495,
0.03403456509113312,
-0.014767639338970184,
-0.017458485439419746,
0.06998659670352936,
-0.04462329298257828,
-0.04089457914233208,
-0.0851701870560646,
-0.05928947776556015,
0.01444813422858715,
-0.0024286224506795406,
-0.043818362057209015,
-0.0753307044506073,
-0.24844898283481598,
-0.05765001103281975,
-0.0000597610414843075,
-0.05186036601662636,
-0.014315507374703884,
0.010513955727219582,
-0.009551780298352242,
-0.013130560517311096,
-0.019312243908643723,
0.1245054081082344,
-0.04913797974586487,
0.03272130712866783,
-0.0880001038312912,
0.07720021158456802,
0.03168881684541702,
0.039653703570365906,
-0.03528591990470886,
0.0019328133203089237,
-0.1451042741537094,
0.13024908304214478,
-0.07276961952447891,
0.02961386740207672,
-0.07805461436510086,
-0.06799259781837463,
-0.06438018381595612,
0.005182104185223579,
-0.01358987856656313,
0.09826919436454773,
-0.23330287635326385,
-0.07749452441930771,
0.23801572620868683,
-0.11409841477870941,
-0.03317215293645859,
0.1059221625328064,
-0.08733032643795013,
0.16701965034008026,
0.11246955394744873,
0.15654504299163818,
0.11029225587844849,
-0.027743080630898476,
-0.013853859156370163,
0.004641373176127672,
-0.012098517268896103,
0.04941832274198532,
0.0428425669670105,
-0.0006375138182193041,
-0.05939803645014763,
0.012979860417544842,
-0.01135268248617649,
0.018503153696656227,
-0.04998841881752014,
-0.08357661962509155,
0.03001316450536251,
-0.03557678312063217,
0.02019551768898964,
-0.015422710217535496,
0.0384000800549984,
0.020128604024648666,
-0.04887813329696655,
0.04966037720441818,
0.09271354228258133,
-0.03092930279672146,
-0.01780250109732151,
-0.007291102781891823,
0.0730806514620781,
-0.0596151165664196,
-0.017882810905575752,
-0.1714581847190857,
-0.08131341636180878,
0.013912152498960495,
-0.04811101406812668,
0.053962547332048416,
0.08562042564153671,
-0.011072170920670033,
0.024481510743498802,
-0.024122506380081177,
0.04766746237874031,
-0.0510750487446785,
-0.0223967544734478,
-0.05786409601569176,
-0.11757993698120117,
-0.04671051725745201,
-0.07669480890035629,
0.1559901088476181,
-0.1263575255870819,
0.023070236667990685,
0.049314141273498535,
0.04606348276138306,
0.023175332695245743,
-0.029774855822324753,
0.03438388183712959,
0.02312796376645565,
0.018023306503891945,
-0.05012032762169838,
0.022587014362215996,
0.012333761900663376,
-0.065131276845932,
-0.0370250903069973,
-0.0767601802945137,
0.006858821492642164,
0.08353833854198456,
0.07545866072177887,
-0.09723395854234695,
-0.08698149025440216,
-0.08268161863088608,
-0.00825247447937727,
-0.10843393951654434,
-0.019677050411701202,
0.16108928620815277,
0.0163166131824255,
0.07302290946245193,
-0.08568139374256134,
-0.0369659923017025,
0.01856652833521366,
-0.001009804313071072,
0.005269660614430904,
0.034230709075927734,
0.06339166313409805,
-0.15905506908893585,
0.05357871949672699,
0.06322236359119415,
0.06911996752023697,
0.2333313673734665,
0.046343844383955,
-0.0878371149301529,
-0.003461424959823489,
-0.019974468275904655,
-0.010641857981681824,
0.11016261577606201,
-0.09302546083927155,
-0.0017491488251835108,
0.041598811745643616,
0.008428750559687614,
0.07880750298500061,
-0.10156016796827316,
0.0390683077275753,
0.04556549713015556,
-0.06217455491423607,
0.013582726009190083,
-0.052571557462215424,
0.000529274286236614,
0.11707046627998352,
0.06270211189985275,
0.04449890926480293,
-0.013917402364313602,
-0.038149669766426086,
-0.11764880269765854,
0.19318106770515442,
-0.08886654675006866,
-0.2626831829547882,
-0.09855068475008011,
0.08962839841842651,
-0.0022388736251741648,
0.0025413562543690205,
0.019458815455436707,
-0.0970102995634079,
-0.0245984997600317,
-0.07629316300153732,
0.008551460690796375,
0.03635755553841591,
-0.033122628927230835,
-0.021451212465763092,
0.005116978194564581,
-0.02688535489141941,
-0.10449887067079544,
-0.03559839725494385,
-0.06772641837596893,
-0.10486956685781479,
0.005984237417578697,
0.005970361176878214,
0.06467245519161224,
0.13154171407222748,
0.005976049229502678,
0.0037174862809479237,
-0.032312504947185516,
0.13807635009288788,
-0.11079484224319458,
0.07803452014923096,
0.1077672690153122,
0.062137700617313385,
0.028223548084497452,
0.062307510524988174,
0.03146607056260109,
-0.07076277583837509,
0.03591207414865494,
0.07118268311023712,
-0.06969510018825531,
-0.22140564024448395,
-0.1311884969472885,
-0.02737789787352085,
-0.032960519194602966,
0.038947269320487976,
0.040127336978912354,
0.04888443276286125,
0.023387465626001358,
-0.012813050299882889,
-0.055598631501197815,
0.007564835250377655,
0.03651135414838791,
0.04009876027703285,
-0.0466010719537735,
0.10182483494281769,
-0.04065719619393349,
-0.009073780849575996,
0.14718441665172577,
-0.011366254650056362,
0.25477662682533264,
-0.013088683597743511,
0.06522723287343979,
0.09513356536626816,
0.039059437811374664,
0.03197916969656944,
0.0722714215517044,
-0.012408450245857239,
0.01142103224992752,
-0.04238027334213257,
-0.06244606897234917,
-0.0186020415276289,
0.07575233280658722,
-0.03216494247317314,
-0.06200456619262695,
-0.1116180568933487,
-0.0196929182857275,
0.051307179033756256,
0.22882391512393951,
0.003063799813389778,
-0.17597733438014984,
-0.08621248602867126,
0.020625611767172813,
-0.02446790598332882,
-0.06691140681505203,
0.016353396698832512,
0.10100878775119781,
-0.17418310046195984,
0.022296587005257607,
-0.043108366429805756,
0.06633739173412323,
-0.07022139430046082,
0.004406327847391367,
0.0379958301782608,
-0.029929103329777718,
-0.027290500700473785,
0.09984713792800903,
-0.11808774620294571,
0.16601800918579102,
0.01602788455784321,
0.059623558074235916,
-0.12900419533252716,
0.008938377723097801,
0.026027007028460503,
-0.03340480849146843,
0.15321989357471466,
0.012870158068835735,
0.009208454750478268,
-0.046813931316137314,
-0.10588183999061584,
0.043973155319690704,
0.00836845114827156,
-0.08309002965688705,
0.044083449989557266,
0.02996736206114292,
0.012083927169442177,
0.004998138640075922,
-0.00795533787459135,
-0.09857282787561417,
-0.14314775168895721,
0.015880858525633812,
-0.08027873933315277,
-0.049369554966688156,
-0.04961135610938072,
-0.07697252929210663,
0.06869460642337799,
0.25060421228408813,
-0.03784763440489769,
-0.10615942627191544,
-0.12849211692810059,
0.07380979508161545,
0.06786727905273438,
-0.03141176328063011,
0.04798521474003792,
0.05887294188141823,
0.15076637268066406,
-0.0087036844342947,
-0.1267872303724289,
0.031166141852736473,
-0.09814701229333878,
-0.11663303524255753,
-0.030915148556232452,
0.10102105140686035,
0.12958934903144836,
0.030847836285829544,
0.021512309089303017,
-0.03598560020327568,
-0.018173763528466225,
-0.11852359771728516,
0.007726116105914116,
0.20454733073711395,
0.035390596836805344,
0.07373800128698349,
-0.09757405519485474,
-0.03905724361538887,
-0.05015677958726883,
-0.0028359373100101948,
0.1498568207025528,
0.16449537873268127,
-0.09239063411951065,
0.17171844840049744,
0.14819326996803284,
-0.11204651743173599,
-0.1518813818693161,
0.0655488669872284,
0.0439140759408474,
0.05710531771183014,
0.014247988350689411,
-0.23451128602027893,
0.1635744422674179,
0.12239357084035873,
-0.030424607917666435,
0.012011608108878136,
-0.28192970156669617,
-0.13155917823314667,
0.10077385604381561,
-0.00009603358193999156,
0.07728811353445053,
-0.08408793061971664,
-0.010091124102473259,
-0.020169725641608238,
-0.018051577731966972,
0.17749595642089844,
-0.19531254470348358,
0.08424545079469681,
-0.006354453042149544,
0.05830709636211395,
0.006056278012692928,
-0.04561086744070053,
0.09640072286128998,
0.07856305688619614,
0.06964194029569626,
-0.02574380300939083,
0.008902081288397312,
0.1144956573843956,
-0.02633904479444027,
0.07172442972660065,
0.10801653563976288,
0.05097726732492447,
-0.15742053091526031,
-0.07774507999420166,
-0.123231902718544,
0.08208120614290237,
-0.043931037187576294,
-0.09687887132167816,
-0.08548503369092941,
0.11645499616861343,
0.09555312246084213,
-0.01752432994544506,
-0.18009839951992035,
-0.06799182295799255,
0.07152710855007172,
0.047555893659591675,
0.1290031373500824,
-0.07513950765132904,
-0.0860796794295311,
0.014088650234043598,
-0.020328694954514503,
0.062271848320961,
-0.0745231956243515,
0.02116983011364937,
0.11253651976585388,
0.023655356839299202,
0.12333513051271439,
0.036072369664907455,
-0.08212592452764511,
0.010394391603767872,
0.05205133929848671,
-0.11706235259771347,
-0.10832182317972183,
-0.03595542162656784,
-0.002475827233865857,
-0.10464441031217575,
-0.0663393959403038,
0.1027846708893776,
-0.013562353327870369,
-0.03259022906422615,
-0.02210722491145134,
0.04942626506090164,
-0.03529943898320198,
0.12882910668849945,
0.03960306569933891,
0.03345973789691925,
-0.06276913732290268,
0.07921484857797623,
0.08270875364542007,
-0.0966816246509552,
0.03750482201576233,
0.061507292091846466,
-0.10566810518503189,
-0.04214899614453316,
-0.028298791497945786,
0.10918109863996506,
-0.12837141752243042,
-0.026688596233725548,
-0.016610419377684593,
-0.0786278247833252,
0.04312693700194359,
0.10079561173915863,
0.02844146080315113,
0.012124431319534779,
-0.07948318123817444,
-0.007384150288999081,
-0.06485842913389206,
0.052358973771333694,
0.023885084316134453,
-0.004932453390210867,
-0.04161762073636055,
0.14737378060817719,
0.014695890247821808,
0.056870099157094955,
-0.046302299946546555,
-0.03702888637781143,
-0.0625452846288681,
0.00039691445999778807,
-0.10317511856555939,
-0.00512924138456583,
-0.10965092480182648,
-0.009945610538125038,
-0.025679145008325577,
0.04858749732375145,
0.01419808529317379,
0.0005101314163766801,
-0.05660496652126312,
-0.011668460443615913,
-0.045271776616573334,
0.05150192603468895,
-0.13089726865291595,
0.017271079123020172,
0.020816335454583168,
-0.03668724000453949,
0.11176998913288116,
0.005618351045995951,
0.04666804149746895,
-0.0575108602643013,
-0.13081584870815277,
0.0058205099776387215,
0.044750042259693146,
0.043321214616298676,
0.028192181140184402,
-0.11851637065410614,
0.036951951682567596,
0.019684063270688057,
-0.039027679711580276,
0.007913587614893913,
0.01619609259068966,
-0.09821455180644989,
0.07380329072475433,
-0.013126696459949017,
0.014196646399796009,
-0.07796042412519455,
0.04416633024811745,
0.06239810958504677,
0.04790641367435455,
0.02438671700656414,
-0.06155340373516083,
0.059636130928993225,
-0.14688067138195038,
-0.046048544347286224,
-0.03676383197307587,
0.026272805407643318,
0.057479340583086014,
-0.056968722492456436,
0.057352688163518906,
0.00999210774898529,
0.24028001725673676,
0.08102734386920929,
0.028194056823849678,
0.008176000788807869,
-0.14922186732292175,
0.08177367597818375,
-0.011772839352488518,
0.08119329065084457,
0.019536703824996948,
-0.01926739513874054,
0.08310384303331375,
0.017610007897019386,
-0.011455736123025417,
0.11950110644102097,
0.2567911744117737,
0.1210012212395668,
0.03448742628097534,
-0.03109383024275303,
0.046119574457407,
-0.10541165620088577,
-0.11118148267269135,
-0.02968544140458107,
-0.03591173514723778,
0.08167903125286102,
-0.07698766887187958,
-0.03791169077157974,
0.08438604325056076,
-0.19658175110816956,
0.0854947566986084,
0.04879022389650345,
-0.08471024036407471,
-0.10416188836097717,
-0.09216365218162537,
-0.026333481073379517,
-0.013496228493750095,
-0.0101845758035779,
-0.15818408131599426,
-0.00426621874794364,
0.06318286061286926,
0.02077604830265045,
-0.04341045394539833,
0.15926404297351837,
-0.07252433896064758,
-0.051497478038072586,
0.10409259051084518,
0.02818872220814228,
0.013267066329717636,
-0.11635036021471024,
-0.004668618086725473,
-0.008388787508010864,
0.056736528873443604,
0.021655913442373276,
0.030244680121541023,
-0.010239290073513985,
0.021173642948269844,
0.01835465431213379,
-0.023499276489019394,
-0.030948497354984283,
-0.02695849910378456,
0.06716864556074142,
0.14122924208641052,
0.07459448277950287,
-0.08119167387485504,
-0.009529965929687023,
0.1708410233259201,
-0.031712837517261505,
-0.02768450416624546,
-0.12802790105342865,
0.17781271040439606,
-0.001604619319550693,
-0.03558432683348656,
-0.0112056415528059,
-0.08406542986631393,
0.04428919777274132,
0.1900353878736496,
0.2325240671634674,
-0.027795586735010147,
-0.004566521849483252,
0.0027429896872490644,
-0.008089184761047363,
0.005748538766056299,
0.12975434958934784,
0.05913783982396126,
0.29923075437545776,
-0.09908019006252289,
0.03931696340441704,
-0.0071924906224012375,
-0.06713194400072098,
-0.056258562952280045,
0.0674978569149971,
0.03791152313351631,
0.01761043816804886,
-0.016676906496286392,
0.09435518085956573,
-0.047919366508722305,
-0.18710947036743164,
0.03936079517006874,
-0.09078465402126312,
-0.12131673097610474,
-0.010458340868353844,
-0.11833801120519638,
-0.015906330198049545,
0.053012073040008545,
-0.05400494858622551,
-0.020473122596740723,
0.11533978581428528,
0.017791088670492172,
-0.07696043699979782,
-0.07084628194570541,
0.10521527379751205,
-0.03689102083444595,
0.22013181447982788,
-0.013028360903263092,
-0.010242043994367123,
0.08957763016223907,
0.008665123023092747,
-0.11609933525323868,
-0.012756678275763988,
0.010921875014901161,
-0.0012333891354501247,
0.021401172503829002,
0.11248355358839035,
-0.014983473345637321,
0.03983921930193901,
0.05257648602128029,
-0.0973421111702919,
0.024479549378156662,
-0.04133936017751694,
0.022901717573404312,
-0.11432449519634247,
0.017058394849300385,
-0.08986000716686249,
0.1699211299419403,
0.17144270241260529,
-0.039137180894613266,
0.022438153624534607,
-0.06636509299278259,
0.0494748055934906,
0.032057806849479675,
0.1347661018371582,
-0.03661283105611801,
-0.12832269072532654,
0.02062162570655346,
-0.018944986164569855,
0.014135508798062801,
-0.2570283114910126,
-0.08721671998500824,
0.06243297457695007,
-0.036938123404979706,
-0.010299976915121078,
0.08259129524230957,
0.018214790150523186,
0.04814095050096512,
-0.026854095980525017,
0.011055383831262589,
-0.03250135853886604,
0.09736742824316025,
-0.1514401137828827,
-0.05418457090854645
] |
null | null | null |
### Preprocessing
1. Normalisation and tokenisation with moses scripts
2. truecased with model docgWP.tcmodel.[LAN] and moses scripts
3. bped with model model.caesen40k.bpe and subword-nmt
- Note: no prepended tag for multilinguality
### Training Data
1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices
* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia
2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300
* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs
- Final training data size for the ca/es-en: 44M parallel sentences
- Finetuned with 1.5M real parallel data (without backtranslations)
### Model
Transformer big with guided alignments. Relevant parameters:
--beam-size 6
--normalize 0.6
--enc-depth 6 --dec-depth 6 --transformer-heads 8
--transformer-preprocess n --transformer-postprocess da
--transformer-dropout 0.1
--label-smoothing 0.1
--dim-emb 1024 --transformer-dim-ffn 4096
--transformer-dropout-attention 0.1
--transformer-dropout-ffn 0.1
--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000
--optimizer-params 0.9 0.998 1e-09
--clip-norm 5
--tied-embeddings
--exponential-smoothing
--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1
## Evaluation
### Test set
https://github.com/PLXIV/Gebiotoolkit/tree/master/gebiocorpus_v2
### ca2en
BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9)
chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7)
### es2en
BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9)
chrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)
|
{"language": ["ca", "es", "en"], "tags": ["translation"]}
|
translation
|
cristinae/marian_caes2en
|
[
"translation",
"ca",
"es",
"en",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ca",
"es",
"en"
] |
TAGS
#translation #ca #es #en #region-us
|
### Preprocessing
1. Normalisation and tokenisation with moses scripts
2. truecased with model docgWP.tcmodel.[LAN] and moses scripts
3. bped with model URL and subword-nmt
- Note: no prepended tag for multilinguality
### Training Data
1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices
* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia
2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300
* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs
- Final training data size for the ca/es-en: 44M parallel sentences
- Finetuned with 1.5M real parallel data (without backtranslations)
### Model
Transformer big with guided alignments. Relevant parameters:
--beam-size 6
--normalize 0.6
--enc-depth 6 --dec-depth 6 --transformer-heads 8
--transformer-preprocess n --transformer-postprocess da
--transformer-dropout 0.1
--label-smoothing 0.1
--dim-emb 1024 --transformer-dim-ffn 4096
--transformer-dropout-attention 0.1
--transformer-dropout-ffn 0.1
--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000
--optimizer-params 0.9 0.998 1e-09
--clip-norm 5
--tied-embeddings
--exponential-smoothing
--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1
## Evaluation
### Test set
URL
### ca2en
BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9)
chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7)
### es2en
BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9)
chrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)
|
[
"### Preprocessing\n1. Normalisation and tokenisation with moses scripts\n2. truecased with model docgWP.tcmodel.[LAN] and moses scripts\n3. bped with model URL and subword-nmt\n- Note: no prepended tag for multilinguality",
"### Training Data\n1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices\n* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia\n2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300\n* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs\n\n- Final training data size for the ca/es-en: 44M parallel sentences\n- Finetuned with 1.5M real parallel data (without backtranslations)",
"### Model\nTransformer big with guided alignments. Relevant parameters:\n\n--beam-size 6 \n\n--normalize 0.6 \n\n--enc-depth 6 --dec-depth 6 --transformer-heads 8\n\n--transformer-preprocess n --transformer-postprocess da \n\n--transformer-dropout 0.1 \n\n--label-smoothing 0.1 \n\n--dim-emb 1024 --transformer-dim-ffn 4096 \n\n--transformer-dropout-attention 0.1 \n\n--transformer-dropout-ffn 0.1 \n\n--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 \n\n--optimizer-params 0.9 0.998 1e-09 \n\n--clip-norm 5 \n\n--tied-embeddings \n\n--exponential-smoothing \n\n--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1",
"## Evaluation",
"### Test set\n\nURL",
"### ca2en\n BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9)\n\n chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7)",
"### es2en\nBLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9) \n\nchrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)"
] |
[
"TAGS\n#translation #ca #es #en #region-us \n",
"### Preprocessing\n1. Normalisation and tokenisation with moses scripts\n2. truecased with model docgWP.tcmodel.[LAN] and moses scripts\n3. bped with model URL and subword-nmt\n- Note: no prepended tag for multilinguality",
"### Training Data\n1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices\n* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia\n2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300\n* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs\n\n- Final training data size for the ca/es-en: 44M parallel sentences\n- Finetuned with 1.5M real parallel data (without backtranslations)",
"### Model\nTransformer big with guided alignments. Relevant parameters:\n\n--beam-size 6 \n\n--normalize 0.6 \n\n--enc-depth 6 --dec-depth 6 --transformer-heads 8\n\n--transformer-preprocess n --transformer-postprocess da \n\n--transformer-dropout 0.1 \n\n--label-smoothing 0.1 \n\n--dim-emb 1024 --transformer-dim-ffn 4096 \n\n--transformer-dropout-attention 0.1 \n\n--transformer-dropout-ffn 0.1 \n\n--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 \n\n--optimizer-params 0.9 0.998 1e-09 \n\n--clip-norm 5 \n\n--tied-embeddings \n\n--exponential-smoothing \n\n--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1",
"## Evaluation",
"### Test set\n\nURL",
"### ca2en\n BLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 47.8 (μ = 47.8 ± 0.9)\n\n chrF|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 69.9 (μ = 69.9 ± 0.7)",
"### es2en\nBLEU|#:1|bs:1000|rs:12345|c:mixed|e:no|tok:13a|s:exp|v:2.0.0 = 48.9 (μ = 48.9 ± 0.9) \n\nchrF2|#:1|bs:1000|rs:12345|c:mixed|e:yes|nc:6|nw:0|s:no|v:2.0.0 = 70.5 (μ = 70.5 ± 0.7)"
] |
[
15,
64,
141,
210,
3,
5,
105,
106
] |
[
"passage: TAGS\n#translation #ca #es #en #region-us \n### Preprocessing\n1. Normalisation and tokenisation with moses scripts\n2. truecased with model docgWP.tcmodel.[LAN] and moses scripts\n3. bped with model URL and subword-nmt\n- Note: no prepended tag for multilinguality### Training Data\n1. Bilingual es-ca: DOGC, Wikimatrix, OpenSubtitles, JW300, GlobalVoices\n* Bilingual es-ca: Translations using systems trained with 1. of Oscar and Wikipedia\n2. Bilingual es-en, ca-en: United Nations, Europarl, Wikimatrix, OpenSubtitles, JW300\n* Bilingual es-en, ca-en: Translations using systems trained with 1. of the missing pairs\n\n- Final training data size for the ca/es-en: 44M parallel sentences\n- Finetuned with 1.5M real parallel data (without backtranslations)### Model\nTransformer big with guided alignments. Relevant parameters:\n\n--beam-size 6 \n\n--normalize 0.6 \n\n--enc-depth 6 --dec-depth 6 --transformer-heads 8\n\n--transformer-preprocess n --transformer-postprocess da \n\n--transformer-dropout 0.1 \n\n--label-smoothing 0.1 \n\n--dim-emb 1024 --transformer-dim-ffn 4096 \n\n--transformer-dropout-attention 0.1 \n\n--transformer-dropout-ffn 0.1 \n\n--learn-rate 0.00015 --lr-warmup 8000 --lr-decay-inv-sqrt 8000 \n\n--optimizer-params 0.9 0.998 1e-09 \n\n--clip-norm 5 \n\n--tied-embeddings \n\n--exponential-smoothing \n\n--transformer-guided-alignment-layer 1 --guided-alignment-cost mse --guided-alignment-weight 0.1## Evaluation### Test set\n\nURL"
] |
[
-0.04639185965061188,
0.06822874397039413,
-0.007070061285048723,
0.03232867270708084,
0.07837794721126556,
0.017189905047416687,
0.0353710874915123,
0.12522460520267487,
0.01575244963169098,
0.10191977769136429,
0.03591351956129074,
0.018606865778565407,
0.1128823533654213,
0.12153605371713638,
0.013359157368540764,
-0.2710344195365906,
0.06310766190290451,
-0.06216515600681305,
-0.06520454585552216,
0.06705158203840256,
0.10712601989507675,
-0.0395958386361599,
0.05069415271282196,
-0.025239448994398117,
-0.04079374298453331,
0.021172652021050453,
-0.0400947667658329,
-0.06287755072116852,
0.05354087054729462,
0.07959359884262085,
0.03172596916556358,
0.022446129471063614,
0.02131400629878044,
-0.22600696980953217,
0.011241520754992962,
0.042657461017370224,
0.006942070554941893,
0.02728775143623352,
0.10687469691038132,
0.0010640494292601943,
0.04743005707859993,
-0.17806628346443176,
0.019162043929100037,
0.06755266338586807,
-0.06286301463842392,
-0.22004581987857819,
-0.1080203577876091,
0.03747222572565079,
0.06874976307153702,
0.05650030076503754,
-0.03354746848344803,
0.06365031003952026,
-0.06894998252391815,
0.05784019082784653,
0.11335919052362442,
-0.27183791995048523,
-0.03031180426478386,
0.01922469586133957,
0.040880993008613586,
-0.017878007143735886,
-0.0632321685552597,
-0.005289653316140175,
-0.01858511194586754,
0.04194530099630356,
-0.01938818395137787,
-0.026658762246370316,
0.060578301548957825,
-0.06067555770277977,
-0.09228765219449997,
0.033217672258615494,
0.09958916902542114,
0.04923150688409805,
-0.07033899426460266,
-0.1647229939699173,
-0.06320804357528687,
-0.028986506164073944,
-0.04386555030941963,
-0.03364621847867966,
0.06636550277471542,
-0.018411889672279358,
0.11089494824409485,
-0.04718326777219772,
-0.05886564403772354,
0.02143622748553753,
-0.12635909020900726,
0.16353711485862732,
0.05901259928941727,
-0.0019053351134061813,
0.010659059509634972,
0.03818389028310776,
-0.03807791322469711,
-0.17463012039661407,
-0.055256761610507965,
-0.03176164999604225,
-0.08499270677566528,
-0.007302843499928713,
0.017614144831895828,
-0.01849312148988247,
0.06683015078306198,
0.1659615933895111,
-0.06529611349105835,
0.11101874709129333,
-0.08482275903224945,
0.015052583068609238,
-0.01007780060172081,
0.16671012341976166,
-0.12164173275232315,
-0.0951574519276619,
0.042720552533864975,
0.003163771703839302,
0.047682587057352066,
0.023063823580741882,
-0.07979937642812729,
-0.04974904656410217,
-0.01444453839212656,
0.08952423185110092,
0.033178817480802536,
0.022948849946260452,
-0.047656863927841187,
-0.04308939352631569,
0.16200725734233856,
-0.1644197702407837,
0.027050958946347237,
0.029321976006031036,
-0.05865006893873215,
0.1821635365486145,
0.011805066838860512,
-0.04164494574069977,
-0.10336652398109436,
0.06723595410585403,
-0.017766881734132767,
0.027719711884856224,
-0.05254121869802475,
-0.07045333087444305,
0.06534931808710098,
0.00043338563409633934,
-0.04416409879922867,
-0.07438936084508896,
-0.07404346764087677,
-0.05469398573040962,
0.01315934956073761,
-0.08856164664030075,
0.00848545040935278,
-0.029048465192317963,
-0.06383585184812546,
0.05466829612851143,
-0.005691868253052235,
-0.013578740879893303,
-0.06212864816188812,
-0.004308071453124285,
0.007904532365500927,
0.08233530819416046,
0.018635479733347893,
0.011038310825824738,
-0.044358670711517334,
0.042708758264780045,
-0.2245938926935196,
0.1344585120677948,
-0.07774863392114639,
0.008305992931127548,
-0.1496967077255249,
-0.04850482940673828,
0.01364655140787363,
0.04089697450399399,
0.03806710243225098,
0.14934135973453522,
-0.24118946492671967,
-0.02483285963535309,
0.2545061707496643,
-0.05315317213535309,
-0.04701533913612366,
0.13841156661510468,
0.022154534235596657,
-0.028323858976364136,
0.049996696412563324,
0.11592195183038712,
0.04212046414613724,
-0.0637047067284584,
0.00785769708454609,
-0.03375551849603653,
-0.001819354249164462,
0.0779011994600296,
0.08893938362598419,
-0.07782695442438126,
0.10316482931375504,
0.026484515517950058,
-0.0014285545330494642,
0.011074979789555073,
-0.009979091584682465,
-0.03213166818022728,
0.009121894836425781,
0.012440497986972332,
0.01280944887548685,
-0.052863746881484985,
0.02090325951576233,
-0.08924135565757751,
-0.11441967636346817,
-0.007243650499731302,
0.07310321182012558,
-0.007213204633444548,
0.07064800709486008,
-0.1108989343047142,
-0.0362754724919796,
0.005381281487643719,
0.045624732971191406,
-0.11133690178394318,
-0.09177358448505402,
0.02431611530482769,
-0.039049576967954636,
0.061058685183525085,
-0.055606573820114136,
0.05086323991417885,
0.016896486282348633,
-0.059460967779159546,
-0.027949152514338493,
0.05310213193297386,
-0.029746130108833313,
-0.0038439983036369085,
-0.13541734218597412,
0.013546884059906006,
-0.025157151743769646,
0.0052035474218428135,
-0.09699995070695877,
0.00039467663737013936,
0.03939451649785042,
0.15092512965202332,
0.011298760771751404,
0.005013188812881708,
0.023426935076713562,
0.01029860507696867,
0.02532286010682583,
-0.06358210742473602,
-0.0013090592110529542,
-0.03885684534907341,
-0.11817290633916855,
0.1222761869430542,
-0.11185025423765182,
0.06498368829488754,
0.10352161526679993,
0.05254485458135605,
-0.08397696912288666,
0.01626729592680931,
0.01150272786617279,
-0.04367154836654663,
0.0035871670115739107,
-0.026421988382935524,
0.11556308716535568,
0.06392046809196472,
0.11092688888311386,
-0.07084035873413086,
-0.030065827071666718,
-0.024254705756902695,
-0.03164233639836311,
-0.06591867655515671,
0.12840797007083893,
-0.06746010482311249,
-0.19760456681251526,
0.07719383388757706,
0.1252610683441162,
0.005707701668143272,
0.15914082527160645,
-0.03771594539284706,
-0.05188274383544922,
-0.06665240228176117,
0.036699771881103516,
0.02627362497150898,
0.016712907701730728,
0.0015298951184377074,
0.04600916802883148,
0.051718227565288544,
0.073259137570858,
0.03491372987627983,
-0.06351708620786667,
0.0344332791864872,
0.03315161541104317,
-0.07083142548799515,
0.03662041574716568,
0.03551597148180008,
0.028402822092175484,
0.09900988638401031,
0.06078985705971718,
0.012843489646911621,
0.007349574938416481,
-0.050027865916490555,
-0.07008806616067886,
0.17416246235370636,
-0.17362001538276672,
-0.1776224672794342,
-0.1575307995080948,
-0.05392633005976677,
-0.0865810289978981,
-0.020117968320846558,
0.0252753384411335,
-0.08898710459470749,
-0.10387516021728516,
-0.08821197599172592,
0.09422486275434494,
-0.010246383026242256,
-0.02997087500989437,
0.01200594287365675,
0.01873069629073143,
0.010727182030677795,
-0.10598654299974442,
-0.028752947226166725,
0.010553307831287384,
-0.08121258020401001,
-0.00680981669574976,
0.005313746631145477,
0.01748739369213581,
0.08683431893587112,
0.03911207243800163,
-0.021997496485710144,
0.0016765149775892496,
0.10791300237178802,
-0.09760098159313202,
0.07878820598125458,
0.18189971148967743,
-0.004283179063349962,
0.05131930857896805,
0.16738128662109375,
0.05273696780204773,
-0.05756501108407974,
-0.01489739678800106,
0.03871419280767441,
-0.009077467955648899,
-0.2729412317276001,
-0.07973740249872208,
-0.0494382381439209,
-0.018055347725749016,
0.0058037517592310905,
0.06360701471567154,
-0.06616814434528351,
0.01126775424927473,
-0.07472560554742813,
-0.06944098323583603,
0.046184174716472626,
0.06517188996076584,
0.1654781699180603,
0.039972756057977676,
0.0887419804930687,
-0.07085038721561432,
0.0020394122693687677,
0.07242672890424728,
0.03506438434123993,
0.13311345875263214,
0.05583595857024193,
0.21598698198795319,
0.05462003871798515,
-0.002020656829699874,
0.032999005168676376,
0.031679630279541016,
0.029627306386828423,
-0.02601291425526142,
0.0160685908049345,
-0.07261408120393753,
-0.012425726279616356,
0.021038001403212547,
0.03569018095731735,
0.004229366313666105,
-0.030802834779024124,
0.028268540278077126,
0.091445192694664,
0.22053954005241394,
-0.0006523411138914526,
-0.1646880805492401,
-0.03230709955096245,
0.04858832806348801,
-0.07517330348491669,
-0.0581229105591774,
0.023464921861886978,
0.05428869649767876,
-0.10710173845291138,
0.11756806820631027,
0.009435263462364674,
0.07514332979917526,
-0.12413250654935837,
-0.0010382720502093434,
0.021277492865920067,
0.12853173911571503,
0.020455345511436462,
0.035544537007808685,
-0.2025926560163498,
0.12004154920578003,
0.02921088971197605,
0.05598365515470505,
-0.048436086624860764,
0.06675048917531967,
0.0039522950537502766,
-0.01724141836166382,
0.1374511867761612,
-0.00795641727745533,
-0.12056124955415726,
-0.06670627743005753,
-0.0548182837665081,
-0.01538938470184803,
0.13587656617164612,
-0.013185619376599789,
0.08724360167980194,
-0.002482473384588957,
-0.04468294233083725,
-0.032926276326179504,
-0.05323740467429161,
-0.1264563500881195,
-0.16243673861026764,
0.05874455347657204,
-0.0011774058220908046,
-0.013833187520503998,
-0.06193224713206291,
-0.032299213111400604,
-0.13166356086730957,
0.22738789021968842,
-0.16298775374889374,
-0.06960590183734894,
-0.1218961626291275,
-0.042365871369838715,
0.17107398808002472,
-0.10358306020498276,
0.06980295479297638,
-0.043083555996418,
0.06386744976043701,
0.013165170326828957,
-0.08150216937065125,
0.1292172521352768,
-0.08257307857275009,
-0.10489518940448761,
-0.0069463979452848434,
0.20545533299446106,
-0.003036011243239045,
0.03949975222349167,
-0.003304594662040472,
0.0046327440068125725,
0.05688925087451935,
-0.10593055933713913,
0.012195144779980183,
0.13090161979198456,
0.03415803611278534,
0.08402850478887558,
-0.09108282625675201,
-0.07666175067424774,
-0.023379312828183174,
-0.04646100848913193,
0.1836387813091278,
0.26713231205940247,
-0.04211205244064331,
0.08587144315242767,
0.07745010405778885,
-0.07747635990381241,
-0.2127947360277176,
0.0023958953097462654,
0.06203586235642433,
0.043825604021549225,
-0.04129290208220482,
-0.19014669954776764,
0.09013767540454865,
0.11196570098400116,
0.012407327070832253,
0.09436100721359253,
-0.2809896171092987,
-0.14772871136665344,
0.053411439061164856,
-0.0021632390562444925,
0.04283541813492775,
-0.11455567181110382,
-0.07282695174217224,
-0.07615899294614792,
-0.16871227324008942,
0.10550092160701752,
-0.028728656470775604,
0.08003789186477661,
0.004764099605381489,
0.013808834366500378,
0.01590392179787159,
-0.01676305942237377,
0.18276360630989075,
0.002037837402895093,
0.011778688058257103,
-0.07148563861846924,
0.045439232140779495,
0.02926402911543846,
-0.03296862915158272,
0.0757802352309227,
0.03017452359199524,
-0.03771968185901642,
-0.08863142132759094,
-0.039480626583099365,
-0.08237436413764954,
-0.012455316260457039,
-0.04206642508506775,
0.008953801356256008,
-0.011939088813960552,
0.08720453083515167,
0.06787383556365967,
-0.03459601104259491,
0.09635087102651596,
-0.0951453223824501,
-0.047637537121772766,
0.09736067056655884,
0.161001518368721,
-0.023845698684453964,
-0.023642482236027718,
0.0186733640730381,
-0.025567540898919106,
0.019728463143110275,
-0.09380586445331573,
0.07443259656429291,
0.10298959910869598,
-0.03421005234122276,
0.06371613591909409,
0.01692577451467514,
-0.09004848450422287,
-0.032879967242479324,
0.09623879194259644,
-0.028621509671211243,
-0.22651731967926025,
0.018876666203141212,
0.0232190303504467,
-0.06038730964064598,
-0.012972825206816196,
0.15059585869312286,
0.05441631004214287,
-0.03482869639992714,
0.040401216596364975,
0.06694716960191727,
-0.021507572382688522,
0.14515618979930878,
0.015424409881234169,
0.03823300451040268,
-0.11480613052845001,
0.16115431487560272,
0.03245433047413826,
-0.1058218777179718,
0.024618178606033325,
0.22683563828468323,
-0.0737217590212822,
-0.060251615941524506,
-0.02146177552640438,
0.05717041343450546,
-0.018971050158143044,
-0.045443180948495865,
-0.0022556050680577755,
-0.1727001816034317,
0.02049834653735161,
0.04218333587050438,
0.04519595205783844,
-0.019048204645514488,
0.0184429120272398,
-0.01394016481935978,
-0.028069553896784782,
0.06846868246793747,
0.0605490580201149,
0.055085670202970505,
-0.06599479168653488,
0.0308831874281168,
-0.055984169244766235,
0.03623070567846298,
-0.004360000602900982,
0.020621750503778458,
-0.10623106360435486,
0.0016430328832939267,
-0.140842005610466,
0.014025258831679821,
-0.10032504051923752,
-0.019199810922145844,
-0.01856118254363537,
0.001959915040060878,
-0.03562274947762489,
-0.0007558366050943732,
-0.082035131752491,
-0.07401154190301895,
-0.07235761731863022,
0.05996303632855415,
-0.0929722934961319,
-0.004395999014377594,
0.01919666677713394,
-0.08745104819536209,
0.12022630870342255,
0.00755872530862689,
0.00615983922034502,
0.014736875891685486,
-0.11242226511240005,
-0.03990356996655464,
0.0055046542547643185,
0.007024066057056189,
0.042377032339572906,
-0.08858808875083923,
0.038703255355358124,
0.0045117647387087345,
-0.03400271013379097,
-0.0003834737872239202,
0.02569626085460186,
-0.11025117337703705,
-0.012865918688476086,
-0.07645029574632645,
-0.003620545379817486,
-0.0683392807841301,
0.030125059187412262,
0.12003545463085175,
0.026038048788905144,
0.10362199693918228,
-0.08086249977350235,
0.06800686568021774,
-0.1076691523194313,
-0.0028189695440232754,
-0.0005942368297837675,
-0.036735616624355316,
0.12116257101297379,
-0.010917754843831062,
0.08687847852706909,
-0.020137762650847435,
0.09122630953788757,
-0.08994853496551514,
0.05573754757642746,
0.06152684614062309,
-0.07574345916509628,
-0.03411426022648811,
0.07099106907844543,
0.09522873908281326,
0.018149953335523605,
-0.020199626684188843,
0.01681886799633503,
-0.0292719304561615,
0.0011611760128289461,
0.015608781017363071,
0.10803882777690887,
0.208360493183136,
0.08701474964618683,
0.1041194424033165,
0.017838116735219955,
-0.08681899309158325,
-0.05168315768241882,
0.08111163228750229,
-0.049091894179582596,
0.06075522303581238,
-0.014675106853246689,
0.043721552938222885,
0.09260226041078568,
-0.14796395599842072,
0.08810802549123764,
0.003903869306668639,
-0.05267643555998802,
-0.10830242186784744,
-0.15784315764904022,
-0.08952310681343079,
-0.03862658143043518,
0.008218671195209026,
-0.09902508556842804,
0.06374038755893707,
0.07660307735204697,
0.10045521706342697,
-0.016208117827773094,
0.12963950634002686,
-0.08378942310810089,
-0.10027429461479187,
0.13023057579994202,
0.033973731100559235,
0.023934273049235344,
0.1179545521736145,
-0.030150070786476135,
0.003214652882888913,
0.04025633633136749,
0.047405850142240524,
0.09459249675273895,
-0.017889853566884995,
0.012914943508803844,
-0.10290756076574326,
-0.08166252821683884,
0.010950771160423756,
0.04906008765101433,
0.012747756205499172,
0.15441620349884033,
0.041457291692495346,
-0.05483483895659447,
-0.015629692003130913,
0.15782183408737183,
-0.04404256120324135,
-0.11880659312009811,
-0.11835665255784988,
0.06865295767784119,
0.05686162784695625,
0.06606364995241165,
-0.03755431994795799,
-0.09462384134531021,
-0.0319533608853817,
0.13862335681915283,
0.18152275681495667,
-0.06161707267165184,
0.0017546683084219694,
0.0348035991191864,
0.002964004408568144,
0.0087481290102005,
0.06650248914957047,
0.10184389352798462,
0.1602047085762024,
-0.037884268909692764,
0.04073890298604965,
-0.013865995220839977,
-0.054295655339956284,
-0.1160353422164917,
0.12561452388763428,
-0.006054403726011515,
0.032812800258398056,
-0.026058262214064598,
0.07730580866336823,
-0.07122334837913513,
-0.08888483047485352,
0.037139762192964554,
-0.09725192934274673,
-0.14283247292041779,
-0.021186497062444687,
0.022208301350474358,
0.029586942866444588,
0.039653971791267395,
0.0439237505197525,
-0.07034078985452652,
0.13607457280158997,
0.013954893685877323,
-0.06788217276334763,
-0.052994996309280396,
0.00626909825950861,
-0.033718738704919815,
0.18290847539901733,
0.030859310179948807,
0.07858139276504517,
0.11769293248653412,
0.04740583524107933,
-0.14530988037586212,
0.03829551860690117,
0.030913660302758217,
-0.0866454541683197,
0.06468751281499863,
0.14090277254581451,
-0.023563247174024582,
0.03498158976435661,
0.08236900717020035,
-0.0776064470410347,
0.014317837543785572,
0.005403186194598675,
-0.04980172961950302,
-0.08022542297840118,
0.07462454587221146,
-0.11007649451494217,
0.10480381548404694,
0.20232516527175903,
-0.021904218941926956,
0.007218359503895044,
-0.027249017730355263,
0.02488136664032936,
0.0019813315011560917,
0.0944223403930664,
-0.04390599951148033,
-0.15712596476078033,
0.0566747672855854,
0.022116370499134064,
0.09743818640708923,
-0.1326105147600174,
-0.07895626872777939,
-0.0006316130748018622,
0.007773600984364748,
-0.09189429879188538,
0.1145000010728836,
0.05107851326465607,
0.030498135834932327,
-0.029527099803090096,
-0.18762432038784027,
0.03377383574843407,
0.09188276529312134,
-0.13706527650356293,
-0.06224900856614113
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec-timit
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 30
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.13.3
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec-timit", "results": []}]}
|
automatic-speech-recognition
|
cristinakuo/wav2vec-timit
|
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
|
# wav2vec-timit
This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 1000
- num_epochs: 30
- mixed_precision_training: Native AMP
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.13.3
- Tokenizers 0.10.3
|
[
"# wav2vec-timit\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n",
"# wav2vec-timit\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.",
"## Model description\n\nMore information needed",
"## Intended uses & limitations\n\nMore information needed",
"## Training and evaluation data\n\nMore information needed",
"## Training procedure",
"### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP",
"### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3"
] |
[
52,
32,
6,
12,
8,
3,
117,
33
] |
[
"passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n# wav2vec-timit\n\nThis model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0001\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_steps: 1000\n- num_epochs: 30\n- mixed_precision_training: Native AMP### Framework versions\n\n- Transformers 4.11.3\n- Pytorch 1.10.0+cu111\n- Datasets 1.13.3\n- Tokenizers 0.10.3"
] |
[
-0.10877498984336853,
0.09662093222141266,
-0.0029397383332252502,
0.03550403565168381,
0.126646488904953,
0.015628598630428314,
0.08752505481243134,
0.10619714856147766,
-0.05869685858488083,
0.06267892569303513,
0.07386357337236404,
0.005418893415480852,
0.06806017458438873,
0.1374618411064148,
-0.030923208221793175,
-0.21295593678951263,
0.027059225365519524,
-0.015200169757008553,
-0.05231095105409622,
0.08767984062433243,
0.09603627771139145,
-0.08322007209062576,
0.05126743018627167,
0.033149976283311844,
-0.14410389959812164,
0.023153601214289665,
-0.04061105102300644,
-0.084200918674469,
0.10286605358123779,
0.04474921524524689,
0.0767868384718895,
0.023776929825544357,
0.09924697875976562,
-0.23043031990528107,
0.005783200729638338,
0.08554726094007492,
0.029281485825777054,
0.0674399584531784,
0.08298734575510025,
0.0031263004057109356,
0.12663237750530243,
-0.14801885187625885,
0.08211545646190643,
0.056369658559560776,
-0.06638497859239578,
-0.1968664675951004,
-0.06689172238111496,
0.10077212750911713,
0.08997321128845215,
0.10150709748268127,
-0.0076980190351605415,
0.09968359023332596,
-0.09635525941848755,
0.06922397017478943,
0.1962634027004242,
-0.25214019417762756,
-0.056458957493305206,
-0.011320199817419052,
0.04585195705294609,
0.018832135945558548,
-0.09840576350688934,
0.0054738763719797134,
0.04135631024837494,
0.045132748782634735,
0.08429259806871414,
0.0022272909991443157,
-0.08431728184223175,
-0.032898493111133575,
-0.12415239959955215,
-0.007053246721625328,
0.14903110265731812,
0.08631891012191772,
-0.0538078248500824,
-0.10097965598106384,
-0.04559721052646637,
-0.0594572015106678,
-0.032833125442266464,
-0.05044430494308472,
0.01610122062265873,
-0.020094428211450577,
-0.06229245290160179,
-0.06762432307004929,
-0.06720376014709473,
-0.06341198831796646,
-0.0014623187016695738,
0.13239160180091858,
0.036509159952402115,
0.0026749717071652412,
-0.037988729774951935,
0.0786338523030281,
0.017936212942004204,
-0.09968128055334091,
-0.006951979827135801,
-0.010479096323251724,
-0.09649249166250229,
-0.0389280691742897,
-0.06104184314608574,
-0.005859626457095146,
0.027684736996889114,
0.12012793868780136,
-0.03434640169143677,
0.10160859674215317,
0.0005570212379097939,
0.00841433648020029,
-0.026372579857707024,
0.10951117426156998,
-0.04812266305088997,
-0.04760728403925896,
-0.007061479613184929,
0.05983440577983856,
-0.011831107549369335,
-0.023357165977358818,
-0.06845317035913467,
-0.05568648874759674,
0.07931321114301682,
0.05284101516008377,
-0.04364873096346855,
0.01676325500011444,
-0.04465916380286217,
-0.03761421889066696,
0.005235984455794096,
-0.10947742313146591,
0.036769986152648926,
0.02287694625556469,
-0.04694134742021561,
0.06243867427110672,
0.03479846939444542,
0.015279448591172695,
-0.05168275162577629,
0.04955633357167244,
-0.044313330203294754,
-0.017610013484954834,
-0.038690321147441864,
-0.04941916465759277,
0.021562129259109497,
-0.04514089599251747,
-0.009941866621375084,
-0.11116272211074829,
-0.14922663569450378,
-0.034337420016527176,
0.041913192719221115,
-0.046991243958473206,
-0.04487429931759834,
-0.034798379987478256,
-0.03846399858593941,
0.03228400647640228,
-0.031715672463178635,
0.16045024991035461,
-0.05732320621609688,
0.049300629645586014,
-0.021562613546848297,
0.033254507929086685,
0.007049087435007095,
0.054298434406518936,
-0.037093281745910645,
0.027908051386475563,
-0.07389504462480545,
0.11735209822654724,
-0.10204337537288666,
-0.007187202572822571,
-0.12455746531486511,
-0.08307620882987976,
-0.012575552798807621,
-0.01210657600313425,
0.0827266126871109,
0.1234971359372139,
-0.1872541308403015,
-0.06136789172887802,
0.14040981233119965,
-0.07556524127721786,
-0.03961645066738129,
0.11929165571928024,
-0.025144293904304504,
-0.0038792213890701532,
0.07057813555002213,
0.14464373886585236,
0.12784774601459503,
-0.14638209342956543,
0.007742945570498705,
0.04272308945655823,
0.03937526419758797,
0.013993365690112114,
0.05138707160949707,
-0.03210333734750748,
-0.032464854419231415,
0.014230470173060894,
-0.005971256643533707,
0.03017762303352356,
-0.08251375705003738,
-0.07797014713287354,
-0.05702979490160942,
-0.09979292750358582,
0.060017552226781845,
-0.015603349544107914,
0.013385246507823467,
-0.07732062041759491,
-0.11469990760087967,
0.061918940395116806,
0.14060690999031067,
-0.060776494443416595,
0.022271977737545967,
-0.07892517000436783,
0.014942009933292866,
0.015281029976904392,
-0.00908727291971445,
-0.18470434844493866,
-0.057300589978694916,
0.02633463405072689,
-0.09618724882602692,
0.03881969675421715,
0.047664981335401535,
0.050085414201021194,
0.05261142924427986,
-0.042295973747968674,
-0.022628501057624817,
-0.10675814747810364,
0.020135626196861267,
-0.08070103079080582,
-0.17159989476203918,
-0.065264493227005,
-0.03426970914006233,
0.1599881947040558,
-0.2024581879377365,
-0.0014569745399057865,
0.03198186308145523,
0.15644285082817078,
0.012679724022746086,
-0.03507274016737938,
0.00043003089376725256,
0.06053066626191139,
0.023288344964385033,
-0.0885084718465805,
0.04961921274662018,
0.014070974662899971,
-0.11340519040822983,
-0.04085076227784157,
-0.08018512278795242,
0.04696926474571228,
0.10342492908239365,
0.04478681460022926,
-0.07461195439100266,
-0.017937002703547478,
-0.058719102293252945,
-0.04515514522790909,
-0.05049068480730057,
0.0002573136880528182,
0.23708635568618774,
0.027808280661702156,
0.11380542814731598,
-0.05448519438505173,
-0.03775626793503761,
0.026979031041264534,
0.008490742184221745,
-0.027983270585536957,
0.07300828397274017,
0.06415735930204391,
-0.10357458889484406,
0.05635764077305794,
0.11486604809761047,
-0.03547178953886032,
0.1411331295967102,
-0.04168357327580452,
-0.07935961335897446,
-0.010078872554004192,
-0.003901989432051778,
-0.023645054548978806,
0.07583068311214447,
-0.16241131722927094,
0.005374868866056204,
0.030248617753386497,
0.029923230409622192,
0.0635615810751915,
-0.15572987496852875,
0.025070207193493843,
0.020888345316052437,
-0.05663435533642769,
-0.009417508728802204,
-0.003792727831751108,
0.03545691817998886,
0.0775628462433815,
0.031660597771406174,
-0.021704640239477158,
0.030119502916932106,
-0.013407448306679726,
-0.08602337539196014,
0.15777365863323212,
-0.1318495273590088,
-0.20200961828231812,
-0.11423411220312119,
0.02470136433839798,
-0.07521861791610718,
-0.03119315207004547,
0.03761829808354378,
-0.10950873047113419,
-0.05098309367895126,
-0.05100974813103676,
0.03972640633583069,
-0.06676285713911057,
0.013505615293979645,
0.08643274754285812,
0.018459074199199677,
0.09957531094551086,
-0.12274344265460968,
0.001335379434749484,
-0.009818623773753643,
-0.06052667275071144,
-0.023107614368200302,
0.05731344223022461,
0.06021704524755478,
0.10516297817230225,
-0.005062054842710495,
0.013561166822910309,
-0.03601580485701561,
0.23491835594177246,
-0.06469008326530457,
0.0008565475582145154,
0.14710275828838348,
0.002869020914658904,
0.04299754276871681,
0.06616554409265518,
0.025140300393104553,
-0.10242624580860138,
0.024033596739172935,
0.058749597519636154,
-0.010694453492760658,
-0.23187269270420074,
-0.06708474457263947,
-0.046210214495658875,
-0.07788047939538956,
0.0991303026676178,
0.04130746051669121,
-0.018704378977417946,
0.04137157276272774,
-0.018470797687768936,
0.036932557821273804,
-0.009577760472893715,
0.06138044223189354,
0.10673097521066666,
0.03413818031549454,
0.08926046639680862,
-0.029926951974630356,
-0.0024278471246361732,
0.09171976894140244,
-0.010211824439466,
0.22118571400642395,
0.01984224282205105,
0.05856953561306,
0.05498114973306656,
0.15964224934577942,
0.005420234054327011,
0.024226471781730652,
0.031392574310302734,
0.0009714126936160028,
0.0002608165086712688,
-0.054014842957258224,
-0.046515610069036484,
0.023241491988301277,
0.042402006685733795,
0.0008234973065555096,
-0.11991271376609802,
0.00350527698174119,
0.0018550974782556295,
0.3238118290901184,
0.038944389671087265,
-0.25084397196769714,
-0.08257074654102325,
0.001798473298549652,
-0.0702078565955162,
-0.07576259225606918,
0.030992351472377777,
0.1152917668223381,
-0.14332018792629242,
0.09683848917484283,
-0.025034070014953613,
0.09625114500522614,
-0.03242335468530655,
0.023484228178858757,
0.011573377065360546,
0.13598033785820007,
0.017334122210741043,
0.08177676051855087,
-0.2205827534198761,
0.1954256296157837,
0.008835752494633198,
0.10631981492042542,
-0.06380481272935867,
0.036702487617731094,
0.03112797997891903,
0.04275748133659363,
0.07953710108995438,
-0.0011629883665591478,
-0.03845692798495293,
-0.11792544275522232,
-0.07551047950983047,
0.04344950616359711,
0.108644999563694,
-0.02668253518640995,
0.05816507339477539,
-0.03430456295609474,
0.005597986746579409,
0.04350857436656952,
0.013656253926455975,
-0.22417296469211578,
-0.1388792246580124,
-0.0016967797419056296,
0.08348690718412399,
0.03176585212349892,
-0.08840300142765045,
-0.09631901234388351,
-0.055360421538352966,
0.1781175434589386,
-0.016398994252085686,
-0.028621455654501915,
-0.1411036103963852,
0.04369601607322693,
0.12338346242904663,
-0.04998849704861641,
0.031894803047180176,
0.02155952900648117,
0.1597699671983719,
0.00901733711361885,
-0.11063000559806824,
0.05902382358908653,
-0.09507922828197479,
-0.1686534434556961,
-0.03251172974705696,
0.154094398021698,
0.0712839663028717,
0.028560571372509003,
0.010435422882437706,
-0.006148618180304766,
0.002105240011587739,
-0.09214220941066742,
0.03739134222269058,
0.09805293381214142,
0.013161294162273407,
0.06296665966510773,
-0.04060974344611168,
-0.054933059960603714,
-0.03876582533121109,
-0.02395431138575077,
0.1455746740102768,
0.22120486199855804,
-0.07811582833528519,
0.10536453872919083,
0.1417420655488968,
-0.06267328560352325,
-0.15346555411815643,
0.03201717138290405,
0.12117011845111847,
0.030585218220949173,
0.012844417244195938,
-0.21831820905208588,
0.09412895888090134,
0.1079513356089592,
-0.0059300372377038,
0.013644036836922169,
-0.26828864216804504,
-0.13218066096305847,
0.12653376162052155,
0.07545828074216843,
0.06067531183362007,
-0.07508036494255066,
-0.0390414334833622,
-0.0754014104604721,
-0.11583466827869415,
0.12027079612016678,
-0.14535637199878693,
0.1147933229804039,
0.007497474551200867,
0.07585561275482178,
0.012461988255381584,
-0.010392104275524616,
0.1498195379972458,
0.030621975660324097,
0.06741341948509216,
-0.028260568156838417,
0.1009482890367508,
0.02284371107816696,
-0.05028301104903221,
0.036595527082681656,
-0.06192454695701599,
0.06161421164870262,
-0.11416076123714447,
-0.04131070151925087,
-0.07655195146799088,
0.08888446539640427,
-0.03127102553844452,
-0.047379277646541595,
-0.03244289010763168,
0.026667334139347076,
0.04894683137536049,
-0.030313244089484215,
-0.007860693149268627,
-0.001455042278394103,
0.0931636393070221,
0.1144702136516571,
0.11434796452522278,
-0.05298831686377525,
-0.10008092224597931,
-0.019951460883021355,
-0.04574226215481758,
0.08912388235330582,
-0.050990086048841476,
0.03147737309336662,
0.09277984499931335,
0.026356348767876625,
0.12451112270355225,
0.025713879615068436,
-0.07144437730312347,
0.0027159203309565783,
0.019825303927063942,
-0.10165812075138092,
-0.14096058905124664,
-0.029846027493476868,
-0.004764875862747431,
-0.08509330451488495,
0.007429995108395815,
0.14764276146888733,
-0.052512120455503464,
-0.007916707545518875,
-0.024907764047384262,
0.0004920942592434585,
-0.06068827584385872,
0.1573842316865921,
0.02472110092639923,
0.06495726108551025,
-0.09986817836761475,
0.12873752415180206,
0.06284455209970474,
-0.09751614183187485,
0.09520172327756882,
0.05845721811056137,
-0.08624309301376343,
-0.03083069808781147,
0.04437083750963211,
0.1642586588859558,
0.010196591727435589,
-0.08161204308271408,
-0.036447953432798386,
-0.14473889768123627,
0.0798501968383789,
0.11588931828737259,
0.02593933418393135,
-0.007121115457266569,
-0.05358415096998215,
0.017304250970482826,
-0.11217638105154037,
0.04192770645022392,
0.04482448101043701,
0.027132371440529823,
-0.14577296376228333,
0.12550897896289825,
0.02909408137202263,
0.039647504687309265,
-0.016568679362535477,
-0.010040469467639923,
-0.08777797222137451,
-0.006784357596188784,
-0.2095743715763092,
-0.008237119764089584,
-0.03066575527191162,
0.023440882563591003,
-0.010183348320424557,
-0.04376043751835823,
-0.04483703523874283,
0.05073145404458046,
-0.06613966077566147,
-0.05256340280175209,
0.004711037501692772,
0.06949934363365173,
-0.12189662456512451,
0.012972841039299965,
0.016446638852357864,
-0.1032484769821167,
0.07911131531000137,
0.07650713622570038,
0.003618380520492792,
0.041134558618068695,
-0.12442365288734436,
-0.04300391301512718,
0.03507591038942337,
0.038430456072092056,
0.04966779053211212,
-0.11179843544960022,
-0.008124489337205887,
-0.012195953167974949,
0.05058775842189789,
0.004170175641775131,
0.09512338042259216,
-0.09710487723350525,
-0.07369830459356308,
-0.05590566247701645,
-0.04561357572674751,
-0.05868973582983017,
0.021639283746480942,
0.1362050622701645,
0.05006612837314606,
0.15065382421016693,
-0.08839187026023865,
0.043587177991867065,
-0.17376750707626343,
-0.02806163765490055,
-0.028535261750221252,
-0.008047751151025295,
-0.029429322108626366,
-0.055288635194301605,
0.09381391108036041,
-0.04043901711702347,
0.12295011430978775,
-0.048985231667757034,
0.08328862488269806,
0.027871891856193542,
-0.11258480697870255,
-0.08597704023122787,
0.02016010694205761,
0.19160039722919464,
0.08632530272006989,
0.0050036064349114895,
0.06583283096551895,
-0.028710229322314262,
0.07010706514120102,
0.07263333350419998,
0.1639476865530014,
0.17230698466300964,
-0.016960781067609787,
0.0798993781208992,
0.07871364057064056,
-0.12157124280929565,
-0.1321680247783661,
0.10364412516355515,
-0.060700081288814545,
0.12586013972759247,
-0.06613123416900635,
0.13727733492851257,
0.057976800948381424,
-0.16680939495563507,
0.0602722242474556,
-0.0372600182890892,
-0.0896816998720169,
-0.10820595175027847,
-0.04301333427429199,
-0.07251179963350296,
-0.1378588080406189,
0.01722487434744835,
-0.09031563997268677,
0.019506892189383507,
0.038445085287094116,
0.018839789554476738,
0.007833017967641354,
0.1325487196445465,
-0.04186573624610901,
-0.0383242592215538,
0.10698370635509491,
0.0007293768576346338,
-0.02932034805417061,
-0.08569017052650452,
-0.09116942435503006,
0.06656603515148163,
0.03167876973748207,
0.061824750155210495,
-0.027335109189152718,
-0.04196241870522499,
0.04941545054316521,
0.0000753947242628783,
-0.06794491410255432,
0.03565674275159836,
-0.013320892117917538,
0.043671950697898865,
0.06762328743934631,
0.05600978061556816,
-0.030638400465250015,
-0.029814528301358223,
0.22779348492622375,
-0.08677167445421219,
-0.0954098179936409,
-0.14217622578144073,
0.15771393477916718,
0.006159490440040827,
0.0006626692484132946,
0.05305861309170723,
-0.07521667331457138,
-0.045105528086423874,
0.18441922962665558,
0.12098510563373566,
-0.061361562460660934,
-0.0027861965354532003,
-0.01763669028878212,
-0.020861808210611343,
-0.052814215421676636,
0.14405715465545654,
0.12812478840351105,
0.023457789793610573,
-0.03594452515244484,
0.015393108129501343,
-0.011157380416989326,
-0.051097970455884933,
-0.08315169811248779,
0.11503266543149948,
-0.013075768947601318,
-0.01598230004310608,
-0.027144089341163635,
0.08210363984107971,
-0.01008684653788805,
-0.2013113647699356,
0.015779493376612663,
-0.12701018154621124,
-0.1718190759420395,
-0.027833426371216774,
0.04411604627966881,
0.021778492256999016,
0.04970526322722435,
-0.015040060505270958,
0.006524661090224981,
0.16223646700382233,
-0.01721321977674961,
-0.028605088591575623,
-0.10335364192724228,
0.08896081149578094,
-0.05003296583890915,
0.15703432261943817,
-0.0019414934795349836,
0.06220119073987007,
0.09069833159446716,
0.0630866289138794,
-0.11956746131181717,
0.04276566207408905,
0.03978632390499115,
-0.050480179488658905,
0.040116362273693085,
0.14336562156677246,
-0.053296755999326706,
0.10124414414167404,
0.047977056354284286,
-0.1424037665128708,
-0.030942402780056,
-0.024747811257839203,
-0.014648159965872765,
-0.05001937970519066,
0.016254372894763947,
-0.0725824236869812,
0.14058947563171387,
0.1693616509437561,
-0.04982820153236389,
-0.016285572201013565,
-0.07054025679826736,
0.052148401737213135,
0.028757669031620026,
0.06308937817811966,
-0.03308512642979622,
-0.20406474173069,
0.004705186001956463,
0.0236944742500782,
0.04122009500861168,
-0.22781158983707428,
-0.09892953932285309,
0.018664345145225525,
-0.036652300506830215,
-0.04097309708595276,
0.10311908274888992,
0.02191823348402977,
0.020190713927149773,
-0.031222449615597725,
-0.10497882217168808,
-0.0347956120967865,
0.1415175050497055,
-0.14677643775939941,
-0.03699145093560219
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-latino40
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 2.8795
- Wer: 1.0
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 32
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 100
- num_epochs: 10
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:---:|
| 5.6846 | 0.83 | 100 | 2.9086 | 1.0 |
| 2.8686 | 1.67 | 200 | 2.8922 | 1.0 |
| 2.8805 | 2.5 | 300 | 2.9326 | 1.0 |
| 2.8613 | 3.33 | 400 | 2.8698 | 1.0 |
| 2.8643 | 4.17 | 500 | 2.9027 | 1.0 |
| 2.8688 | 5.0 | 600 | 2.9544 | 1.0 |
| 2.8689 | 5.83 | 700 | 2.8914 | 1.0 |
| 2.8558 | 6.67 | 800 | 2.8762 | 1.0 |
| 2.8537 | 7.5 | 900 | 2.8982 | 1.0 |
| 2.8522 | 8.33 | 1000 | 2.8820 | 1.0 |
| 2.8468 | 9.17 | 1100 | 2.8760 | 1.0 |
| 2.8454 | 10.0 | 1200 | 2.8795 | 1.0 |
### Framework versions
- Transformers 4.12.5
- Pytorch 1.9.1
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "wav2vec2-latino40", "results": []}]}
|
automatic-speech-recognition
|
cristinakuo/wav2vec2-latino40
|
[
"transformers",
"pytorch",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us
|
wav2vec2-latino40
=================
This model is a fine-tuned version of facebook/wav2vec2-base on the None dataset.
It achieves the following results on the evaluation set:
* Loss: 2.8795
* Wer: 1.0
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0001
* train\_batch\_size: 32
* eval\_batch\_size: 8
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 100
* num\_epochs: 10
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.12.5
* Pytorch 1.9.1
* Datasets 1.16.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.9.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.9.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
52,
130,
4,
31
] |
[
"passage: TAGS\n#transformers #pytorch #wav2vec2 #automatic-speech-recognition #generated_from_trainer #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 100\n* num\\_epochs: 10\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.12.5\n* Pytorch 1.9.1\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
-0.10980920493602753,
0.08066091686487198,
-0.0021913868840783834,
0.0798632800579071,
0.14092256128787994,
0.0030789559241384268,
0.08961708098649979,
0.1398172527551651,
-0.07077814638614655,
0.040380705147981644,
0.1079588532447815,
0.14933347702026367,
0.03574487194418907,
0.12165234982967377,
-0.04340481013059616,
-0.29026296734809875,
0.01716907136142254,
0.014433713629841805,
-0.044237591326236725,
0.1325906664133072,
0.09864385426044464,
-0.12339255213737488,
0.04956413060426712,
0.011601904407143593,
-0.14805734157562256,
0.010426769964396954,
0.0026738583110272884,
-0.08487695455551147,
0.14838279783725739,
0.00794005487114191,
0.07316295057535172,
0.015556921251118183,
0.0989600419998169,
-0.2548612356185913,
0.003781266510486603,
0.041084736585617065,
0.047623373568058014,
0.061330974102020264,
0.07041196525096893,
-0.0112789086997509,
0.11298255622386932,
-0.10049547255039215,
0.0634230226278305,
0.035836685448884964,
-0.10100731998682022,
-0.29945191740989685,
-0.07114958763122559,
0.022580832242965698,
0.08290360867977142,
0.11252439022064209,
-0.017463045194745064,
0.10048104077577591,
-0.08502072840929031,
0.10130675882101059,
0.2632434368133545,
-0.28493553400039673,
-0.04188449680805206,
-0.0320056714117527,
0.030116042122244835,
0.05607079714536667,
-0.1130281314253807,
-0.0199373010545969,
0.021807949990034103,
0.051630325615406036,
0.11832799762487411,
-0.02314845658838749,
-0.11859384179115295,
0.012785177677869797,
-0.1495315581560135,
-0.03822407126426697,
0.12689848244190216,
0.044106051325798035,
-0.03479421138763428,
-0.08074644953012466,
-0.04381341114640236,
-0.14233465492725372,
-0.05427202954888344,
-0.014493022114038467,
0.04658723995089531,
-0.03640612214803696,
-0.08634860068559647,
-0.017045967280864716,
-0.08484596014022827,
-0.08799180388450623,
-0.03250779211521149,
0.163325235247612,
0.049993909895420074,
-0.0054273405112326145,
-0.009451371617615223,
0.08832554519176483,
0.0006423027953132987,
-0.12581756711006165,
-0.0027684089727699757,
0.03728927671909332,
-0.030845317989587784,
-0.006714310962706804,
-0.0775303915143013,
-0.006719156634062529,
0.025007838383316994,
0.12120231240987778,
-0.07473734021186829,
0.07411094754934311,
0.013705099932849407,
0.02501939795911312,
-0.11756092309951782,
0.20496274530887604,
-0.05834757164120674,
-0.025750350207090378,
-0.013698107562959194,
0.04614108055830002,
0.005192374344915152,
-0.01254340074956417,
-0.09562832862138748,
-0.005417653825134039,
0.1215624064207077,
0.039135344326496124,
-0.06357745081186295,
0.05863622948527336,
-0.03601355850696564,
-0.01738746650516987,
-0.04621974006295204,
-0.11743055284023285,
0.03168739378452301,
0.025128355249762535,
-0.08781567960977554,
0.000711713160853833,
0.023804007098078728,
0.02700847014784813,
-0.04845186695456505,
0.08255426585674286,
-0.05339859426021576,
0.04193750023841858,
-0.06689785420894623,
-0.11167975515127182,
0.013831880874931812,
-0.06339839100837708,
0.015696832910180092,
-0.1014920100569725,
-0.15867817401885986,
-0.01975177228450775,
0.04356376826763153,
-0.036123570054769516,
-0.03404486924409866,
-0.08703483641147614,
-0.07419129461050034,
0.024441922083497047,
-0.04096292331814766,
0.14225170016288757,
-0.06730513274669647,
0.11225902289152145,
0.029156258329749107,
0.07115056365728378,
-0.018335899338126183,
0.07524584233760834,
-0.0815892219543457,
0.002993879606947303,
-0.1414223611354828,
0.08606751263141632,
-0.07554500550031662,
0.0405312106013298,
-0.12535442411899567,
-0.1199832633137703,
0.02675686776638031,
0.0007289050845429301,
0.08981318771839142,
0.10596152395009995,
-0.17258626222610474,
-0.08019126206636429,
0.1752787083387375,
-0.06593860685825348,
-0.08045100420713425,
0.11808839440345764,
-0.039672963321208954,
0.023581838235259056,
0.07135766744613647,
0.2478012591600418,
0.04693472757935524,
-0.11165597289800644,
0.043805141001939774,
-0.013276532292366028,
0.06678491830825806,
-0.025329528376460075,
0.07152348011732101,
-0.03138262778520584,
0.010174485854804516,
0.03288539499044418,
-0.043981969356536865,
0.07483356446027756,
-0.09624520689249039,
-0.08675498515367508,
-0.03934602811932564,
-0.10191374272108078,
0.03482167422771454,
0.043031103909015656,
0.06365569680929184,
-0.0985899269580841,
-0.08746787160634995,
0.04025315120816231,
0.09528588503599167,
-0.09891946613788605,
0.05347936600446701,
-0.07100776582956314,
0.04501967132091522,
-0.001897135516628623,
-0.016613401472568512,
-0.1871185600757599,
0.03687867894768715,
0.023578641936182976,
-0.02337789349257946,
0.039804913103580475,
-0.03218350186944008,
0.07619354128837585,
0.03881413862109184,
-0.04651379957795143,
-0.04409072548151016,
-0.02065667323768139,
0.012341917492449284,
-0.09351722151041031,
-0.1993872970342636,
-0.035728778690099716,
-0.018513230606913567,
0.11530566960573196,
-0.17193715274333954,
0.014598478563129902,
-0.019322114065289497,
0.07519219815731049,
0.0037099465262144804,
-0.019692348316311836,
-0.01481874380260706,
0.08457089215517044,
-0.0085816141217947,
-0.05307632312178612,
0.06940632313489914,
0.0031947139650583267,
-0.09319141507148743,
0.01586523838341236,
-0.10217957198619843,
0.10750915110111237,
0.1332051008939743,
-0.08164609968662262,
-0.06618577986955643,
0.0011545104207471013,
-0.049083709716796875,
-0.039093658328056335,
-0.02735612727701664,
0.02902553416788578,
0.20677080750465393,
-0.005885649938136339,
0.13525806367397308,
-0.07594040781259537,
-0.02799786813557148,
0.022258900105953217,
-0.023045221343636513,
0.015822691842913628,
0.13198581337928772,
0.07947170734405518,
-0.041822366416454315,
0.11116132140159607,
0.12410301715135574,
-0.09254572540521622,
0.11285964399576187,
-0.05419359728693962,
-0.08977862447500229,
-0.010861373506486416,
-0.007626814767718315,
-0.015537419356405735,
0.08842293918132782,
-0.15356121957302094,
-0.014198613353073597,
0.029136179015040398,
0.0361868180334568,
0.026153240352869034,
-0.225873202085495,
-0.01565035991370678,
0.029006700962781906,
-0.08592138439416885,
-0.03842167928814888,
-0.0021013568621128798,
0.012155167758464813,
0.09853563457727432,
0.003302189754322171,
-0.10624036937952042,
0.009442655369639397,
-0.017023036256432533,
-0.084555484354496,
0.19169212877750397,
-0.11786629259586334,
-0.17307570576667786,
-0.09892862290143967,
-0.07620367407798767,
-0.03259239345788956,
0.003938349429517984,
0.07807767391204834,
-0.09726081788539886,
-0.028173502534627914,
-0.06308752298355103,
0.03210591897368431,
-0.041175562888383865,
0.03680180385708809,
0.0036928232293576,
-0.003402488073334098,
0.06942503899335861,
-0.10922123491764069,
-0.008241287432610989,
-0.05346428602933884,
-0.03276630863547325,
0.038681626319885254,
0.04400238022208214,
0.10152043402194977,
0.16139578819274902,
0.008259491994976997,
0.028078991919755936,
-0.04554590582847595,
0.20942853391170502,
-0.06462772190570831,
-0.04289346933364868,
0.1388622522354126,
0.0070281364023685455,
0.04439382255077362,
0.12561090290546417,
0.04757501557469368,
-0.09919650852680206,
0.0033767425920814276,
0.012305780313909054,
-0.022879192605614662,
-0.21807990968227386,
-0.0541338287293911,
-0.049014441668987274,
-0.029451046139001846,
0.10054463893175125,
0.021792670711874962,
0.01725919358432293,
0.020740719512104988,
0.04053569585084915,
0.002018396742641926,
-0.0015959605807438493,
0.058000557124614716,
0.14925150573253632,
0.024657532572746277,
0.12087169289588928,
-0.01869051344692707,
-0.04857170581817627,
0.02699047513306141,
-0.0108701903373003,
0.23927874863147736,
0.012819498777389526,
0.12436876446008682,
0.050061244517564774,
0.18832843005657196,
-0.0005413531907834113,
0.07564345002174377,
0.015274266712367535,
-0.012130046263337135,
-0.002754911780357361,
-0.05093031004071236,
-0.045556794852018356,
0.031116588041186333,
0.02848834916949272,
0.025044111534953117,
-0.13001902401447296,
-0.049802038818597794,
0.03907139599323273,
0.3274923264980316,
0.05229125916957855,
-0.29431799054145813,
-0.08614405244588852,
-0.005382025148719549,
-0.0854690670967102,
-0.022076856344938278,
0.046303000301122665,
0.08580722659826279,
-0.09947432577610016,
0.04448115825653076,
-0.047977153211832047,
0.09024563431739807,
-0.03151266276836395,
0.040820345282554626,
0.04360958933830261,
0.07487986236810684,
0.02256215550005436,
0.06914918124675751,
-0.3232811391353607,
0.29180625081062317,
-0.008358528837561607,
0.07779446989297867,
-0.06456765532493591,
-0.0008857692591845989,
0.03378576412796974,
-0.0011557098478078842,
0.09022928774356842,
-0.018672609701752663,
-0.0801580622792244,
-0.18707461655139923,
-0.060589663684368134,
0.036532677710056305,
0.127633199095726,
-0.011211197823286057,
0.10346045345067978,
-0.03243345022201538,
-0.009961643256247044,
0.07621586322784424,
-0.055092498660087585,
-0.10099312663078308,
-0.08181507140398026,
-0.00940331257879734,
0.07693351060152054,
0.03743179887533188,
-0.06578808277845383,
-0.10221552848815918,
-0.11230611056089401,
0.1218077689409256,
-0.04539094492793083,
-0.027113184332847595,
-0.1105206087231636,
0.061217136681079865,
0.130745530128479,
-0.0792306512594223,
0.05503794550895691,
0.028080426156520844,
0.06753195822238922,
0.03414563834667206,
-0.06759253889322281,
0.11237512528896332,
-0.07640185952186584,
-0.17026765644550323,
-0.033496465533971786,
0.1469498723745346,
0.036228787153959274,
0.06863132119178772,
-0.006273076869547367,
0.028511548414826393,
-0.045666828751564026,
-0.0804627537727356,
0.03256896138191223,
0.036437153816223145,
0.023420540615916252,
0.04869148135185242,
-0.05705498158931732,
-0.017071884125471115,
-0.09537269175052643,
-0.04085355997085571,
0.20932559669017792,
0.2342206984758377,
-0.093354232609272,
0.0716480240225792,
0.06755233556032181,
-0.05380915105342865,
-0.18005482852458954,
-0.0030158916488289833,
0.0848696231842041,
0.0050223711878061295,
0.015279422514140606,
-0.19609257578849792,
0.07792839407920837,
0.06992988288402557,
-0.01870499551296234,
0.07932014018297195,
-0.3264188766479492,
-0.14206287264823914,
0.1441699117422104,
0.12191656976938248,
0.08031284809112549,
-0.13511276245117188,
-0.0404108390212059,
-0.01999991573393345,
-0.091950923204422,
0.0915660560131073,
-0.04971156641840935,
0.1435217708349228,
-0.017111627385020256,
0.1045382097363472,
0.013917868956923485,
-0.049031343311071396,
0.11851494759321213,
0.024936791509389877,
0.06427741050720215,
-0.036918606609106064,
0.019268162548542023,
-0.00964928325265646,
-0.034783001989126205,
0.04311220347881317,
-0.06751948595046997,
0.037642188370227814,
-0.08364549279212952,
-0.035955946892499924,
-0.10305949300527573,
0.03376590088009834,
-0.005185771267861128,
-0.05732830613851547,
-0.018481800332665443,
0.02000368759036064,
0.06819313764572144,
-0.007115368265658617,
0.09952523559331894,
-0.033060330897569656,
0.12334724515676498,
0.11199751496315002,
0.07548270374536514,
-0.04602932929992676,
-0.06700962036848068,
-0.014362026937305927,
-0.028282882645726204,
0.05370795726776123,
-0.11295565962791443,
0.029776575043797493,
0.13810531795024872,
0.03917720168828964,
0.14451472461223602,
0.06972716003656387,
-0.04443461820483208,
0.010397059842944145,
0.04125906527042389,
-0.12701374292373657,
-0.11644574999809265,
-0.004148922394961119,
-0.02314218319952488,
-0.07639680802822113,
0.04377919062972069,
0.10687291622161865,
-0.06386600434780121,
-0.011042671278119087,
-0.021161457523703575,
0.02330395020544529,
-0.07179027050733566,
0.2252303510904312,
0.06212461739778519,
0.04856344684958458,
-0.11988171190023422,
0.08925968408584595,
0.041735388338565826,
-0.1158151850104332,
0.03984756022691727,
0.07972640544176102,
-0.08497200906276703,
-0.039511069655418396,
0.05026671290397644,
0.10819706320762634,
-0.04242098703980446,
-0.07139823585748672,
-0.11464481800794601,
-0.14505107700824738,
0.09956482797861099,
0.1452970802783966,
0.07713909447193146,
0.014507410116493702,
-0.06352374702692032,
0.013524708338081837,
-0.10527925938367844,
0.07716207951307297,
0.05278395861387253,
0.04045756161212921,
-0.13215371966362,
0.1596628576517105,
0.01660974510014057,
0.05105000361800194,
-0.01783343032002449,
0.0003049750521313399,
-0.09723072499036789,
0.044797927141189575,
-0.13372665643692017,
-0.016621513292193413,
-0.0440630204975605,
0.00647035613656044,
0.009973205626010895,
-0.07510308921337128,
-0.05170198529958725,
0.037644289433956146,
-0.1226712167263031,
-0.030972497537732124,
-0.0020897414069622755,
0.05209917202591896,
-0.12659986317157745,
-0.04508664831519127,
0.02240194007754326,
-0.07813826948404312,
0.08739304542541504,
0.09181436896324158,
-0.019976526498794556,
0.08028993755578995,
-0.14488914608955383,
-0.028371071442961693,
0.0663955956697464,
0.008208757266402245,
0.0362272784113884,
-0.13547508418560028,
-0.014229604974389076,
0.009603659622371197,
0.05940083786845207,
0.014157848432660103,
0.08902829885482788,
-0.1306706815958023,
-0.016501907259225845,
-0.042343106120824814,
-0.06023785099387169,
-0.06204197555780411,
0.02433887869119644,
0.07678435742855072,
0.046087849885225296,
0.1713658571243286,
-0.09509948641061783,
0.04948916658759117,
-0.1854909211397171,
0.005001120734959841,
-0.04503310099244118,
-0.09847234189510345,
-0.09514760971069336,
-0.052987098693847656,
0.08700865507125854,
-0.04966249689459801,
0.12958776950836182,
-0.015757031738758087,
0.07188122719526291,
0.018411675468087196,
-0.06927963346242905,
-0.010539642535150051,
0.03743515536189079,
0.25493499636650085,
0.038861505687236786,
-0.029862146824598312,
0.07857978343963623,
0.04346318542957306,
0.09011953324079514,
0.15496864914894104,
0.16722559928894043,
0.19484195113182068,
0.03936972841620445,
0.10197120159864426,
0.052975527942180634,
-0.0825294777750969,
-0.14498275518417358,
0.06355110555887222,
-0.031587496399879456,
0.11224187910556793,
-0.029109589755535126,
0.2681392729282379,
0.061866383999586105,
-0.17852698266506195,
0.06703627109527588,
-0.03827032446861267,
-0.08790795505046844,
-0.1184806376695633,
-0.027287298813462257,
-0.08511841297149658,
-0.17839950323104858,
0.004721211735159159,
-0.10451129823923111,
0.05941224470734596,
0.0790998712182045,
0.03342793136835098,
0.0002967296168208122,
0.13273613154888153,
0.017104819416999817,
-0.013799617066979408,
0.09511622041463852,
-0.005110669881105423,
-0.04267231747508049,
-0.08802526444196701,
-0.08089674264192581,
0.0317658968269825,
-0.012727725319564342,
0.04702073708176613,
-0.04121053218841553,
-0.11177753657102585,
0.03308112919330597,
-0.044365085661411285,
-0.08819011598825455,
0.01934644766151905,
0.018091760575771332,
0.08175329864025116,
0.08086381107568741,
0.037871353328228,
-0.03429818153381348,
-0.0020277153234928846,
0.2526741921901703,
-0.10593343526124954,
-0.123429074883461,
-0.10019700229167938,
0.2773582935333252,
0.04580030217766762,
-0.0008489458705298603,
0.016611211001873016,
-0.0669313296675682,
-0.01992752216756344,
0.24284401535987854,
0.175593301653862,
-0.06653659045696259,
-0.0029926213901489973,
-0.0014997737016528845,
-0.007376679219305515,
-0.033362776041030884,
0.08983216434717178,
0.16204452514648438,
0.06294775754213333,
-0.07975048571825027,
-0.03254455327987671,
-0.0467907078564167,
-0.030329039320349693,
-0.0675441175699234,
0.07937409728765488,
0.0074424175545573235,
-0.03071312978863716,
-0.035275060683488846,
0.08556422591209412,
-0.08735252916812897,
-0.12293340265750885,
0.015828529372811317,
-0.1928618997335434,
-0.1565762311220169,
-0.0158489141613245,
0.07319075614213943,
0.040658846497535706,
0.03373730182647705,
-0.02140166237950325,
0.006188822444528341,
0.08847112208604813,
-0.0029859363567084074,
-0.0775647982954979,
-0.08026209473609924,
0.09346207976341248,
-0.10640282183885574,
0.1653733253479004,
-0.03455295041203499,
0.07426184415817261,
0.10542125254869461,
0.09622615575790405,
-0.06406783312559128,
0.102670818567276,
0.052303776144981384,
-0.107117660343647,
0.05454130098223686,
0.14238789677619934,
-0.042176973074674606,
0.09960916638374329,
0.03267516940832138,
-0.1318810135126114,
0.026521943509578705,
-0.044511228799819946,
-0.0656069964170456,
-0.04928657412528992,
-0.044857949018478394,
-0.05315857380628586,
0.11852610111236572,
0.1792403608560562,
-0.05418472737073898,
0.009916502982378006,
-0.06851965188980103,
0.005978732369840145,
0.03164321184158325,
0.03913820907473564,
-0.06651060283184052,
-0.2513844966888428,
0.010636888444423676,
0.023227639496326447,
0.0053297677077353,
-0.23199740052223206,
-0.09097366780042648,
0.014225414022803307,
-0.06440398842096329,
-0.0720076858997345,
0.1059345230460167,
0.06386196613311768,
0.031837984919548035,
-0.04153471067547798,
-0.09472637623548508,
-0.03624151647090912,
0.1907484531402588,
-0.18053385615348816,
-0.07049532979726791
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-MiniLM-L-12-v2
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-MiniLM-L-2-v2
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-MiniLM-L-4-v2
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-MiniLM-L-6-v2
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-TinyBERT-L-2-v2
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-TinyBERT-L-2
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-TinyBERT-L-4
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
47
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.031222937628626823,
0.13048580288887024,
-0.006592764053493738,
0.035769447684288025,
0.11023839563131332,
0.023622313514351845,
0.1118343323469162,
0.14141945540905,
0.011887344531714916,
-0.07865098118782043,
0.14571447670459747,
0.2256961315870285,
-0.011237370781600475,
0.06962031126022339,
-0.07936780899763107,
-0.23767422139644623,
0.08903764188289642,
0.04316740483045578,
-0.02329845540225506,
0.09714217483997345,
0.11549460887908936,
-0.0445655882358551,
0.0482446625828743,
-0.005831177346408367,
-0.07852476090192795,
0.02374298870563507,
0.05234311893582344,
-0.11925916373729706,
0.07824034988880157,
0.038800518959760666,
0.08641701191663742,
0.047947775572538376,
-0.02588331513106823,
-0.18091587722301483,
0.020593609660863876,
0.014527560211718082,
-0.08456390351057053,
0.06201760843396187,
0.06459520757198334,
-0.05026666447520256,
0.028283588588237762,
0.04363081604242325,
-0.00040016864659264684,
0.05442293360829353,
-0.08365041762590408,
-0.18777050077915192,
-0.07788987457752228,
0.09138200432062149,
0.0822538360953331,
0.07590723037719727,
0.041403573006391525,
0.14720378816127777,
-0.12174759060144424,
0.06750258803367615,
0.11365879327058792,
-0.3379494249820709,
-0.00012072643585270271,
0.0726609155535698,
0.019400611519813538,
0.02270316518843174,
-0.02585512027144432,
0.032551705837249756,
0.04672111198306084,
0.0034406250342726707,
0.024993298575282097,
-0.06176041439175606,
-0.11698492616415024,
0.03534247726202011,
-0.04940676689147949,
-0.06621868163347244,
0.24049511551856995,
-0.013352715410292149,
0.029132870957255363,
-0.0047011785209178925,
-0.06106462702155113,
0.0031755072996020317,
-0.027250586077570915,
0.04743704944849014,
0.022055545821785927,
0.11091086268424988,
0.07261817902326584,
-0.016392147168517113,
-0.1279716044664383,
0.0008985644089989364,
-0.1909598708152771,
0.08758813887834549,
0.020127490162849426,
0.08088716119527817,
-0.15236561000347137,
0.055559054017066956,
0.028175966814160347,
-0.11545844376087189,
0.0005364827229641378,
-0.08256713300943375,
0.10449586063623428,
0.011525196954607964,
-0.06949394196271896,
0.04138573631644249,
0.13127945363521576,
0.2090783268213272,
0.053110383450984955,
0.030188433825969696,
-0.050601422786712646,
0.11591508984565735,
-0.023805562406778336,
0.08434237539768219,
0.03620612993836403,
-0.007659158669412136,
0.11160806566476822,
-0.11625667661428452,
0.05364016443490982,
-0.03825801610946655,
-0.1572587639093399,
-0.018107563257217407,
0.027624348178505898,
0.11380995064973831,
0.04983453080058098,
0.0565413273870945,
-0.05016535893082619,
-0.0004966393462382257,
0.14144660532474518,
-0.06383424252271652,
0.01579524390399456,
0.00790349394083023,
0.028509506955742836,
0.06099670007824898,
0.026157883927226067,
0.02817358262836933,
-0.0572095662355423,
0.09920138865709305,
-0.04989606887102127,
-0.011875666677951813,
-0.037143491208553314,
-0.008297450840473175,
0.0870564803481102,
-0.08851300179958344,
0.05059066042304039,
-0.15062427520751953,
-0.14653214812278748,
0.030236298218369484,
0.0720018520951271,
0.012049839831888676,
-0.09377799183130264,
0.034003641456365585,
-0.010921157896518707,
0.047230709344148636,
-0.08446047455072403,
-0.016658423468470573,
-0.08630052953958511,
0.07090666890144348,
-0.09199928492307663,
0.024058466777205467,
-0.15290489792823792,
0.06158958747982979,
-0.11669103801250458,
-0.017135046422481537,
-0.051425229758024216,
-0.03311094269156456,
-0.09284032881259918,
0.1942492127418518,
-0.043372999876737595,
-0.03203170374035835,
0.013681466691195965,
0.005522655323147774,
-0.06157875433564186,
0.12065905332565308,
-0.09521565586328506,
-0.06215933337807655,
0.19024963676929474,
-0.1084100529551506,
-0.1800156831741333,
0.07037068903446198,
0.001334881060756743,
0.00610776524990797,
0.0773395523428917,
0.1616002321243286,
0.09672778099775314,
-0.02310440130531788,
0.09237653762102127,
0.14067962765693665,
-0.06062621623277664,
-0.19009415805339813,
0.05171007290482521,
-0.04433976858854294,
-0.14356684684753418,
0.06429912894964218,
-0.017190508544445038,
0.0829833447933197,
0.002651303308084607,
-0.07413939386606216,
-0.052518080919981,
-0.043487370014190674,
0.021368304267525673,
0.025271417573094368,
0.08122741430997849,
-0.07590355724096298,
-0.008831245824694633,
-0.031189359724521637,
0.034308433532714844,
0.05369502305984497,
0.04838934913277626,
-0.06191778555512428,
0.061439886689186096,
0.02486778050661087,
0.033942271023988724,
-0.14579053223133087,
-0.007896673865616322,
-0.0017176179680973291,
0.03740164265036583,
-0.0015597393503412604,
0.07824065536260605,
0.036536093801259995,
-0.08376944065093994,
-0.008764523081481457,
-0.015120083466172218,
0.14143557846546173,
0.0484895296394825,
-0.026612315326929092,
-0.12957508862018585,
0.049588754773139954,
-0.0399571917951107,
0.02532694675028324,
-0.01934116706252098,
0.024379892274737358,
0.0456324964761734,
0.08728967607021332,
-0.029270634055137634,
0.09425725787878036,
-0.034797847270965576,
-0.0007632380584254861,
-0.06166345998644829,
0.0052313413470983505,
0.12030249834060669,
0.04904041439294815,
-0.06734607368707657,
0.19378910958766937,
-0.05520053952932358,
0.28220629692077637,
0.20493178069591522,
-0.18299666047096252,
0.06226447969675064,
0.0013990707229822874,
-0.010337709449231625,
-0.000578522973228246,
0.03439730405807495,
0.0335940420627594,
0.03681229054927826,
0.010316891595721245,
0.17409954965114594,
-0.052756331861019135,
-0.04641488194465637,
-0.01663317158818245,
-0.05680250748991966,
-0.013033241964876652,
0.051225002855062485,
0.14472876489162445,
-0.1932520568370819,
0.19265632331371307,
0.32094016671180725,
-0.01836542785167694,
0.08962716907262802,
-0.07507248222827911,
0.018457842990756035,
0.055129729211330414,
-0.05118462070822716,
-0.022528575733304024,
-0.02265028841793537,
-0.12742535769939423,
-0.000777439447119832,
0.09433021396398544,
0.040656160563230515,
0.04232742637395859,
-0.13055551052093506,
-0.05610394477844238,
0.00009042652527568862,
-0.00696546072140336,
-0.050355494022369385,
0.057761259377002716,
0.004052021540701389,
0.09031425416469574,
-0.02385767363011837,
-0.10479959100484848,
0.1368882954120636,
-0.0031609961297363043,
-0.07879859209060669,
0.1476660817861557,
-0.1670645922422409,
-0.2675674557685852,
-0.13479484617710114,
-0.1487317681312561,
-0.027203327044844627,
0.01467166654765606,
0.13942782580852509,
-0.060368604958057404,
-0.0525057278573513,
0.012911345809698105,
-0.09812077134847641,
-0.037766944617033005,
0.028387591242790222,
-0.0297428946942091,
0.06439930200576782,
0.004743938334286213,
-0.10729271918535233,
-0.0666976347565651,
0.01350174006074667,
-0.031026283279061317,
0.0922672301530838,
-0.08812612295150757,
0.06524661928415298,
0.12273956090211868,
0.017020998522639275,
0.037275176495313644,
-0.03228471800684929,
0.13733430206775665,
-0.03686793893575668,
-0.008670826442539692,
0.2096070498228073,
-0.03775579482316971,
0.07707484066486359,
0.150436133146286,
0.04063309356570244,
-0.04427816718816757,
0.015577026642858982,
-0.049652185291051865,
-0.07288471609354019,
-0.26840829849243164,
-0.10533773899078369,
-0.10727188736200333,
0.053441211581230164,
0.06932765990495682,
0.09347819536924362,
0.14388203620910645,
0.07078276574611664,
-0.010041584260761738,
0.018074555322527885,
0.006245976779609919,
0.08051405102014542,
0.23954537510871887,
-0.003942674491554499,
0.12337736785411835,
-0.10948646068572998,
-0.053607091307640076,
0.1174393892288208,
0.055854130536317825,
0.10559877753257751,
0.12827354669570923,
0.0634327083826065,
0.056444261223077774,
0.16488425433635712,
0.10907312482595444,
0.11669280380010605,
0.02187339775264263,
-0.006276066415011883,
-0.05460451915860176,
-0.015361282974481583,
-0.06251870095729828,
0.03777467831969261,
-0.013609152287244797,
-0.12285177409648895,
-0.04584629461169243,
-0.1615571826696396,
0.06845147162675858,
0.18546631932258606,
0.020312797278165817,
-0.14464733004570007,
0.025239532813429832,
0.09377814084291458,
-0.01945749670267105,
-0.053647808730602264,
0.10041272640228271,
-0.11533371359109879,
-0.11224431544542313,
0.11980185657739639,
-0.028893087059259415,
0.14660626649856567,
-0.016873618587851524,
0.04937293007969856,
-0.013568560592830181,
-0.13501383364200592,
0.06583409011363983,
0.14391310513019562,
-0.30061182379722595,
0.19650478661060333,
-0.007461524568498135,
-0.043495722115039825,
-0.08261296898126602,
0.003955775871872902,
0.06695891916751862,
0.24799129366874695,
0.0889478474855423,
0.01616300828754902,
-0.12551507353782654,
-0.0395657904446125,
-0.062256503850221634,
0.028937330469489098,
0.02037951722741127,
-0.0179972555488348,
-0.057268936187028885,
-0.06969304382801056,
-0.022713175043463707,
-0.00025611647288315,
0.03760894015431404,
-0.026397885754704475,
-0.15958911180496216,
0.05402941629290581,
0.09700854867696762,
0.052172884345054626,
-0.05016152188181877,
-0.03343906253576279,
-0.10161250084638596,
0.19424182176589966,
-0.08696603775024414,
-0.07252633571624756,
-0.09862473607063293,
-0.13247954845428467,
0.025107014924287796,
-0.06244604289531708,
0.061074692755937576,
-0.08559167385101318,
0.011953501962125301,
-0.047901470214128494,
-0.21219149231910706,
0.10887160897254944,
-0.13795332610607147,
-0.04415391758084297,
-0.05766300484538078,
0.12810908257961273,
-0.11148732900619507,
0.022969752550125122,
0.034964412450790405,
0.016335586085915565,
-0.10780297219753265,
-0.10891950130462646,
-0.013977473601698875,
0.030052734538912773,
0.05653700977563858,
-0.04045039415359497,
-0.09685365110635757,
-0.04719218611717224,
0.013199465349316597,
-0.03896326571702957,
0.24409520626068115,
0.17588753998279572,
-0.08934667706489563,
0.19285185635089874,
0.18291422724723816,
-0.09928718209266663,
-0.3110765814781189,
-0.15143781900405884,
-0.13517117500305176,
-0.10138122737407684,
-0.023945894092321396,
-0.15794737637043,
0.12962974607944489,
0.04799811169505119,
-0.09087679535150528,
0.08863683044910431,
-0.15966655313968658,
-0.0805826410651207,
0.19894154369831085,
-0.033684100955724716,
0.3326871395111084,
-0.14180131256580353,
-0.08361893892288208,
-0.09042175114154816,
-0.17759105563163757,
0.15467439591884613,
-0.0708114504814148,
0.052896592766046524,
-0.01295430026948452,
0.002900085411965847,
-0.007546346168965101,
-0.04754116013646126,
0.11410257965326309,
-0.025622457265853882,
0.01940036565065384,
-0.1225917786359787,
-0.012375768274068832,
0.06971857696771622,
-0.029210612177848816,
0.029483215883374214,
-0.1414121687412262,
0.02156968228518963,
-0.11909553408622742,
-0.021438248455524445,
-0.06554215401411057,
0.0683499425649643,
0.0028498645406216383,
-0.04397595301270485,
-0.019701682031154633,
-0.024387499317526817,
0.0310470312833786,
-0.016381951048970222,
0.2860105335712433,
0.023737331852316856,
0.09919518977403641,
0.11846044659614563,
0.08302575349807739,
-0.18312297761440277,
-0.011074194684624672,
-0.10109596699476242,
-0.08606860041618347,
0.06998859345912933,
-0.1615401655435562,
0.06356693059206009,
0.09814772009849548,
-0.061531227082014084,
0.061257727444171906,
0.08415394276380539,
0.019474968314170837,
-0.0573870912194252,
0.13776203989982605,
-0.17544475197792053,
0.031624019145965576,
-0.017194677144289017,
0.11635028570890427,
0.06414506584405899,
0.05926978960633278,
0.11567357182502747,
0.013719623908400536,
-0.05012055113911629,
0.025427423417568207,
0.02964070253074169,
-0.05026624724268913,
0.027212506160140038,
0.05781693384051323,
0.008329174481332302,
-0.1355985403060913,
0.10132720321416855,
0.03726547211408615,
-0.1443147212266922,
-0.025202171877026558,
0.10883472114801407,
-0.1498357653617859,
-0.13316985964775085,
0.03644883260130882,
0.08920352905988693,
-0.11446324735879898,
-0.10948441922664642,
-0.049708377569913864,
-0.13059930503368378,
0.07466436177492142,
0.111910380423069,
0.12298257648944855,
0.06841865926980972,
-0.013371058739721775,
-0.06788983941078186,
0.0383983813226223,
0.0029060887172818184,
-0.07893457263708115,
0.02773320861160755,
-0.1128721833229065,
-0.06500279903411865,
0.017772741615772247,
0.10809708386659622,
-0.053233202546834946,
-0.017664581537246704,
-0.11155769228935242,
0.03594955801963806,
-0.1270546019077301,
0.004668314475566149,
-0.08084987848997116,
-0.004947167355567217,
0.017740819603204727,
-0.06953023374080658,
-0.03667932003736496,
-0.005178960505872965,
-0.12781259417533875,
-0.015513245016336441,
-0.02598072960972786,
0.07093318551778793,
-0.10114233940839767,
-0.05152695253491402,
0.09541739523410797,
-0.012447931803762913,
0.0973988026380539,
0.0750475525856018,
-0.06654786318540573,
0.10046802461147308,
-0.1393914669752121,
-0.09427568316459656,
0.09227922558784485,
0.04359714314341545,
0.03399856761097908,
-0.008078759536147118,
0.016459334641695023,
0.1202658861875534,
-0.033905576914548874,
0.043104592710733414,
0.024574751034379005,
-0.147372767329216,
-0.03224669396877289,
0.0014853047905489802,
-0.14264920353889465,
-0.009076599031686783,
-0.09593779593706131,
0.12641161680221558,
0.019937938079237938,
0.19958613812923431,
-0.019179530441761017,
0.04851064458489418,
-0.04842802882194519,
0.01850293017923832,
-0.03834648057818413,
-0.17755559086799622,
-0.14522022008895874,
-0.07037320733070374,
-0.038313861936330795,
-0.018554124981164932,
0.251798540353775,
0.04093918949365616,
-0.042603056877851486,
0.07759085297584534,
0.06571243703365326,
0.016727644950151443,
0.0212676003575325,
0.24752330780029297,
0.06033453717827797,
-0.010926000773906708,
-0.06129152700304985,
0.01934751868247986,
0.017677469179034233,
-0.056927431374788284,
0.09427402168512344,
0.09811029583215714,
0.068744957447052,
0.06465623527765274,
0.016684463247656822,
0.026975518092513084,
-0.10663208365440369,
-0.14640776813030243,
0.015742281451821327,
0.09490249305963516,
0.01810981146991253,
0.11160247772932053,
0.12422402948141098,
-0.03371858596801758,
0.0225676279515028,
-0.06432820111513138,
-0.0060334219597280025,
-0.18783824145793915,
-0.10917981714010239,
-0.08849360793828964,
-0.12464868277311325,
0.0013375792186707258,
-0.048415862023830414,
0.0207132026553154,
0.08100679516792297,
0.04537936672568321,
-0.06281905621290207,
-0.02731252834200859,
-0.021290594711899757,
-0.044625598937273026,
0.04943249374628067,
-0.03116297721862793,
-0.023418989032506943,
-0.029619375243782997,
-0.031456053256988525,
-0.10911190509796143,
-0.04594817012548447,
-0.05295524001121521,
0.040317900478839874,
-0.013821911998093128,
0.040155891329050064,
-0.10918360203504562,
-0.08355793356895447,
-0.03505280986428261,
0.029276607558131218,
-0.006034370511770248,
0.17343491315841675,
0.011322897858917713,
0.03913239762187004,
0.096920907497406,
0.17550869286060333,
-0.07044066488742828,
-0.12284383922815323,
-0.05195518955588341,
0.19905413687229156,
0.05575999990105629,
0.06016545370221138,
0.015420405194163322,
0.01774226874113083,
-0.05175405368208885,
0.3056987226009369,
0.28822413086891174,
-0.04234142228960991,
0.03812962397933006,
-0.012957140803337097,
0.017551559954881668,
0.08646208047866821,
0.14314526319503784,
0.1225508525967598,
0.17820580303668976,
-0.07520273327827454,
-0.020763490349054337,
-0.033969681710004807,
0.0011272527044638991,
-0.16123616695404053,
0.0776604637503624,
-0.01038258709013462,
-0.08717965334653854,
-0.01985633186995983,
0.09582911431789398,
-0.111983522772789,
0.1065826490521431,
0.0007611355395056307,
-0.12746141850948334,
-0.04855658859014511,
-0.01268674898892641,
0.18558096885681152,
0.010363814420998096,
0.029743928462266922,
-0.018424078822135925,
-0.06368906795978546,
0.10191062837839127,
-0.01379233319312334,
-0.20874279737472534,
-0.07239016890525818,
0.09477964043617249,
-0.033585019409656525,
0.12148012965917587,
0.005590854678303003,
0.05478139966726303,
0.07904579490423203,
0.06767632812261581,
-0.09296553581953049,
0.06549388915300369,
0.012761259451508522,
-0.05543714389204979,
0.0046954466961324215,
-0.09866335242986679,
-0.024628590792417526,
-0.06140054389834404,
0.04757868871092796,
-0.07632169872522354,
0.04376089945435524,
-0.03289888799190521,
-0.06448434293270111,
-0.04339258000254631,
0.06421805173158646,
-0.07498079538345337,
0.05848152935504913,
0.030363943427801132,
-0.026018934324383736,
-0.07160912454128265,
-0.061227958649396896,
-0.02492934837937355,
0.023381372913718224,
-0.1795109510421753,
-0.09561535716056824,
-0.007375465705990791,
-0.031752556562423706,
0.09167668968439102,
0.045114073902368546,
-0.10676556080579758,
-0.019531425088644028,
-0.09544464945793152,
0.016059521585702896,
-0.16992658376693726,
0.04376090317964554,
0.06283412128686905,
-0.003013976849615574,
-0.006386113353073597,
-0.06914757937192917,
0.025922639295458794,
0.02013215981423855,
-0.09332647174596786,
-0.08685147762298584
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-TinyBERT-L-6
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.005313366185873747,
0.12051703780889511,
-0.004693866241723299,
0.03580747917294502,
0.08237535506486893,
0.015574881806969643,
0.09972856193780899,
0.1512490212917328,
-0.002342487219721079,
-0.035675179213285446,
0.1512344777584076,
0.1678246557712555,
-0.03263122960925102,
0.07377322763204575,
-0.06331735849380493,
-0.23472362756729126,
0.10480628162622452,
0.022721635177731514,
-0.06391183286905289,
0.0852574035525322,
0.11916844546794891,
-0.06770222634077072,
0.04979204759001732,
-0.006682221312075853,
-0.07624185085296631,
0.03358222916722298,
0.023558329790830612,
-0.11514781415462494,
0.09597253054380417,
0.031052103266119957,
0.10091041028499603,
0.054762229323387146,
-0.03489936888217926,
-0.1495836228132248,
0.026294076815247536,
0.029823211953043938,
-0.08911542594432831,
0.07207200676202774,
0.03851554915308952,
-0.03519143536686897,
0.04294948652386665,
0.0052515072748064995,
-0.0014171735383570194,
0.028817106038331985,
-0.09800101071596146,
-0.21412332355976105,
-0.07283638417720795,
0.07232868671417236,
0.027582870796322823,
0.07712103426456451,
0.024217145517468452,
0.15384386479854584,
-0.14576973021030426,
0.05114344507455826,
0.14506278932094574,
-0.3469622731208801,
-0.0045180246233940125,
0.13070425391197205,
0.07035423070192337,
0.04190012812614441,
-0.03805508464574814,
0.04657916724681854,
0.05582089722156525,
0.0025910616386681795,
0.06670897454023361,
-0.04245023429393768,
-0.1239764392375946,
0.06446674466133118,
-0.07143140584230423,
-0.07907963544130325,
0.2631944417953491,
-0.019031910225749016,
0.0633535161614418,
-0.020391980186104774,
-0.07111599296331406,
-0.02444433979690075,
0.0007090915460139513,
0.04914804920554161,
0.029526568949222565,
0.10229268670082092,
0.07079360634088516,
-0.03269055485725403,
-0.14858068525791168,
0.020561350509524345,
-0.18860138952732086,
0.06773007661104202,
0.0031633847393095493,
0.07213912159204483,
-0.1380876749753952,
0.05910053104162216,
0.022541260346770287,
-0.12431655079126358,
0.026827946305274963,
-0.08040676265954971,
0.09522116184234619,
0.02466272935271263,
-0.10294348746538162,
0.048924896866083145,
0.12230401486158371,
0.1840950846672058,
0.025208648294210434,
0.004822383169084787,
-0.026003237813711166,
0.12904393672943115,
0.010278739035129547,
0.08728154003620148,
-0.017540551722049713,
-0.016535868868231773,
0.07604997605085373,
-0.07513169944286346,
0.056044477969408035,
-0.06161702796816826,
-0.1701550930738449,
-0.03133761137723923,
0.03735613077878952,
0.09264276921749115,
0.08358346670866013,
0.05583064258098602,
-0.02714722789824009,
0.025454645976424217,
0.1282920092344284,
-0.06134769693017006,
0.04080605506896973,
-0.00510947173461318,
0.03630151227116585,
0.02500070445239544,
0.02943694032728672,
0.023852426558732986,
-0.017657775431871414,
0.06957349926233292,
-0.06750348955392838,
-0.010275444015860558,
-0.051824845373630524,
-0.060393817722797394,
0.08819574117660522,
-0.0934952050447464,
0.03543853387236595,
-0.15935862064361572,
-0.08309947699308395,
0.017079420387744904,
0.06980738043785095,
0.0021625745575875044,
-0.09810265898704529,
0.07114944607019424,
-0.04728621616959572,
0.0843905657529831,
-0.08494804799556732,
0.002027006121352315,
-0.09370028227567673,
0.05686311423778534,
-0.10129949450492859,
0.055879801511764526,
-0.17257225513458252,
0.06273247301578522,
-0.1052277535200119,
-0.0019254190847277641,
-0.05104181170463562,
-0.04737512394785881,
-0.09146487712860107,
0.16669559478759766,
-0.022341273725032806,
-0.0318073108792305,
-0.019266171380877495,
0.031578149646520615,
-0.04990849643945694,
0.08952520042657852,
-0.13634717464447021,
-0.03626799210906029,
0.14703963696956635,
-0.09130478650331497,
-0.16326141357421875,
0.0717688575387001,
0.00046021450543776155,
0.0031928219832479954,
0.0205746628344059,
0.2063581943511963,
0.0636834129691124,
-0.038732003420591354,
0.0429588183760643,
0.14875924587249756,
-0.031094709411263466,
-0.15359465777873993,
0.06786390393972397,
-0.029294706881046295,
-0.09197673946619034,
0.043804608285427094,
-0.017092304304242134,
0.06857074052095413,
0.011296656914055347,
-0.08280479162931442,
-0.07357008755207062,
-0.03364529088139534,
0.04349697753787041,
0.0174326840788126,
0.09276469796895981,
-0.07873370498418808,
-0.03898242488503456,
-0.0033968340139836073,
0.03604090213775635,
0.05958762764930725,
0.056634191423654556,
-0.03324146941304207,
0.10759710520505905,
-0.00020664279873017222,
0.019652394577860832,
-0.1466042846441269,
-0.022884350270032883,
-0.024007568135857582,
0.029566245153546333,
-0.0034570449497550726,
0.1761774867773056,
0.028720151633024216,
-0.09089623391628265,
-0.005154056008905172,
-0.005816992837935686,
0.10818760842084885,
0.07017345726490021,
-0.0435544028878212,
-0.122578464448452,
0.02524160034954548,
-0.05991243198513985,
0.011386431753635406,
-0.028770629316568375,
0.03522561118006706,
0.0614837110042572,
0.095530666410923,
-0.04097152501344681,
0.10141308605670929,
-0.024964042007923126,
0.0062868064269423485,
-0.08510880917310715,
0.002189264167100191,
0.09858357906341553,
0.033706147223711014,
-0.04600461944937706,
0.2194674164056778,
-0.10151731222867966,
0.2976343631744385,
0.2267475426197052,
-0.1751745343208313,
0.057022739201784134,
0.03313778340816498,
-0.019422493875026703,
0.019689708948135376,
0.0130884675309062,
-0.001759985345415771,
-0.028527354821562767,
-0.02610705979168415,
0.14932112395763397,
-0.04963885620236397,
-0.03732553496956825,
-0.023624418303370476,
-0.06671696156263351,
-0.03914187103509903,
0.04179641976952553,
0.09773124009370804,
-0.15121658146381378,
0.21856796741485596,
0.39029189944267273,
-0.04813329502940178,
0.11297299712896347,
-0.055986665189266205,
0.025421306490898132,
0.030753333121538162,
-0.07714325934648514,
-0.05503294616937637,
0.02333962731063366,
-0.13534876704216003,
-0.02139679528772831,
0.1016768291592598,
0.056578896939754486,
0.04646281898021698,
-0.13725902140140533,
-0.0547390952706337,
0.02008267678320408,
0.020364783704280853,
-0.05890339985489845,
0.08217864483594894,
0.018010256811976433,
0.09777913242578506,
-0.0105742197483778,
-0.10413795709609985,
0.10512387752532959,
-0.0025146319530904293,
-0.046441737562417984,
0.12223948538303375,
-0.1751604825258255,
-0.26337674260139465,
-0.10911697149276733,
-0.10594077408313751,
-0.0219988115131855,
0.016123976558446884,
0.14658363163471222,
-0.05834493041038513,
-0.04452504217624664,
-0.03016248345375061,
-0.08486530184745789,
-0.06864798069000244,
0.03474634513258934,
-0.04818902164697647,
0.05575546622276306,
0.0009074720437638462,
-0.1152714416384697,
-0.06752760708332062,
0.020496297627687454,
-0.03415825590491295,
0.091299869120121,
-0.030293690040707588,
0.07342889904975891,
0.135502427816391,
0.007198224309831858,
0.027944006025791168,
-0.030184734612703323,
0.16622070968151093,
-0.03921347111463547,
0.00924068596214056,
0.19809506833553314,
-0.0031844095792621374,
0.0713656023144722,
0.18086189031600952,
0.03603030741214752,
-0.022779395803809166,
0.01003937516361475,
-0.03807242214679718,
-0.08706989139318466,
-0.20672155916690826,
-0.11673396080732346,
-0.12385633587837219,
0.03123038075864315,
0.058827221393585205,
0.10125505924224854,
0.1326202005147934,
0.05287611857056618,
0.0016830338863655925,
0.002457578666508198,
-0.029018420726060867,
0.06340944021940231,
0.23763202130794525,
-0.03204694390296936,
0.14382527768611908,
-0.10360123962163925,
-0.058313898742198944,
0.11185727268457413,
0.07886284589767456,
0.08192996680736542,
0.10665401816368103,
0.03152311220765114,
0.0695836991071701,
0.19727741181850433,
0.10917509347200394,
0.08815993368625641,
0.02420349419116974,
-0.010355910286307335,
-0.06285016238689423,
-0.023843849077820778,
-0.023379191756248474,
0.06658007949590683,
0.039418045431375504,
-0.13914524018764496,
-0.029837213456630707,
-0.19487792253494263,
0.08332343399524689,
0.151483952999115,
0.059639498591423035,
-0.13124634325504303,
0.031023966148495674,
0.10783152282238007,
-0.01136285811662674,
-0.04613389074802399,
0.08763978630304337,
-0.02930242381989956,
-0.08054038137197495,
0.07484424859285355,
0.007519237697124481,
0.12197799980640411,
0.02354556694626808,
0.07243271917104721,
-0.04469209909439087,
-0.16604083776474,
0.056991759687662125,
0.12605127692222595,
-0.2823342978954315,
0.20038354396820068,
-0.02226925455033779,
-0.09456770122051239,
-0.07824184000492096,
0.0034344338346272707,
0.07712201029062271,
0.22976431250572205,
0.061447788029909134,
0.03248010203242302,
-0.1648344248533249,
-0.04715387150645256,
-0.03496819734573364,
0.012963905930519104,
0.017423907294869423,
-0.02331795170903206,
-0.03611692786216736,
-0.06621395796537399,
-0.002671534428372979,
0.020914889872074127,
0.1569015085697174,
-0.022213159129023552,
-0.17191611230373383,
0.06727875024080276,
0.11111561954021454,
0.02136494219303131,
-0.04995092377066612,
-0.043893180787563324,
-0.16064417362213135,
0.16429398953914642,
-0.034801822155714035,
-0.03848220035433769,
-0.09644030034542084,
-0.11785909533500671,
0.05767476186156273,
-0.043937042355537415,
0.05455517768859863,
-0.08362634479999542,
0.027540860697627068,
-0.06371035426855087,
-0.19031928479671478,
0.11500367522239685,
-0.12582644820213318,
-0.04196672886610031,
-0.053986817598342896,
0.08883348852396011,
-0.15042348206043243,
0.04660613462328911,
0.03068644367158413,
0.06134520471096039,
-0.15422704815864563,
-0.11111470311880112,
-0.0030049297492951155,
0.01256405096501112,
0.059495214372873306,
-0.03420129418373108,
-0.07899564504623413,
-0.04707137122750282,
0.06431546807289124,
-0.03068653494119644,
0.25191134214401245,
0.18924297392368317,
-0.12640511989593506,
0.17781375348567963,
0.13471996784210205,
-0.07904026657342911,
-0.3402675986289978,
-0.1414533108472824,
-0.16139845550060272,
-0.07071840018033981,
0.026064196601510048,
-0.11777465045452118,
0.0958787128329277,
0.03335252031683922,
-0.10947048664093018,
0.052748363465070724,
-0.17017929255962372,
-0.07672762125730515,
0.17991377413272858,
-0.07845299690961838,
0.32294031977653503,
-0.15600302815437317,
-0.0581873282790184,
-0.07140202820301056,
-0.1397649347782135,
0.15831077098846436,
-0.12095484882593155,
0.0640173852443695,
-0.01085357554256916,
0.0009923696052283049,
0.000617707846686244,
-0.05387040227651596,
0.15228526294231415,
-0.05112951993942261,
0.02527214027941227,
-0.1299176663160324,
-0.04767915979027748,
0.10295214504003525,
-0.06544466316699982,
0.023311404511332512,
-0.1613304764032364,
0.012056032195687294,
-0.1437864750623703,
0.010370835661888123,
-0.08332697302103043,
0.08352860808372498,
-0.009336385875940323,
-0.040623150765895844,
-0.05954933911561966,
-0.005279048811644316,
0.04284205287694931,
-0.015335856936872005,
0.29630112648010254,
0.008494893088936806,
0.13189546763896942,
0.16303826868534088,
0.022608771920204163,
-0.1718173325061798,
-0.01953105255961418,
-0.03700250759720802,
-0.0736953616142273,
0.07995044440031052,
-0.2031794935464859,
0.057182274758815765,
0.07621432095766068,
-0.06988967210054398,
0.06107276678085327,
0.08900803327560425,
0.01489779632538557,
-0.04548985883593559,
0.16442106664180756,
-0.1671867072582245,
-0.0006360075785778463,
-0.008117838762700558,
0.10618383437395096,
0.0521981455385685,
0.017339367419481277,
0.12554024159908295,
0.005745370872318745,
-0.036925509572029114,
0.026103302836418152,
0.021035362035036087,
-0.0628022700548172,
0.008186271414160728,
0.0677781030535698,
0.028965607285499573,
-0.1132744550704956,
0.07926022261381149,
0.06677495688199997,
-0.129411518573761,
-0.010128652676939964,
0.09587135165929794,
-0.1095706969499588,
-0.16171106696128845,
0.00618077302351594,
0.04380159080028534,
-0.08738097548484802,
-0.08622825890779495,
-0.024138811975717545,
-0.1249731257557869,
0.04831845685839653,
0.09633947908878326,
0.13090389966964722,
0.06443681567907333,
-0.0017022290267050266,
-0.06870454549789429,
0.05082831531763077,
0.001563939149491489,
-0.07709051668643951,
0.036532118916511536,
-0.12133266031742096,
-0.02846197411417961,
0.010334618389606476,
0.11554903537034988,
-0.06259089708328247,
-0.0042324368841946125,
-0.12426569312810898,
0.001558715826831758,
-0.11614777147769928,
-0.029476512223482132,
-0.08306972682476044,
-0.033092986792325974,
0.02484716661274433,
-0.08117660880088806,
-0.04502910003066063,
-0.01520068384706974,
-0.14415021240711212,
-0.024689702317118645,
-0.04193141311407089,
0.09420331567525864,
-0.104824498295784,
-0.05173527076840401,
0.09438075870275497,
-0.011454490944743156,
0.09049908816814423,
0.0417814776301384,
-0.05572384595870972,
0.07632939517498016,
-0.07947538793087006,
-0.10600963234901428,
0.08863865584135056,
0.03787854313850403,
0.046659957617521286,
-0.005956739187240601,
-0.011253931559622288,
0.09342904388904572,
-0.0007177229854278266,
0.042248114943504333,
0.011596323922276497,
-0.12917691469192505,
-0.02238038182258606,
-0.010726326145231724,
-0.13931554555892944,
0.015678487718105316,
-0.09275448322296143,
0.14498212933540344,
0.01790434494614601,
0.17994743585586548,
-0.005748189985752106,
0.031212467700242996,
-0.06608235090970993,
0.025236306712031364,
-0.057449471205472946,
-0.17951899766921997,
-0.12403986603021622,
-0.06836044043302536,
-0.027640171349048615,
-0.028845123946666718,
0.2805347144603729,
0.08917142450809479,
-0.08347166329622269,
0.07290034741163254,
0.07196228206157684,
0.01505149994045496,
0.02077554725110531,
0.22023816406726837,
0.06361716240644455,
-0.02245054394006729,
-0.061880506575107574,
0.019000306725502014,
0.022344766184687614,
-0.03256882727146149,
0.07464652508497238,
0.1169479712843895,
0.1142251044511795,
0.07622475177049637,
0.02704872190952301,
-0.016070278361439705,
-0.11697600036859512,
-0.1287461221218109,
0.014688584953546524,
0.119608573615551,
-0.03686606511473656,
0.059426259249448776,
0.13529466092586517,
-0.037811972200870514,
0.03956560790538788,
-0.086575448513031,
0.012540166266262531,
-0.17560049891471863,
-0.09157182276248932,
-0.07360459119081497,
-0.13344724476337433,
-0.03314107656478882,
-0.052344802767038345,
0.03984220325946808,
0.1292441338300705,
0.032210953533649445,
-0.03851942718029022,
-0.015154266729950905,
-0.015868499875068665,
-0.04258604347705841,
0.03028041310608387,
-0.023542780429124832,
0.0002562306181062013,
-0.06029721349477768,
-0.02395167201757431,
-0.10368048399686813,
-0.02776411361992359,
-0.05679041147232056,
0.04382525384426117,
-0.027991825714707375,
0.021414659917354584,
-0.12156787514686584,
-0.08556406199932098,
-0.0529513917863369,
0.020444447174668312,
-0.005759124178439379,
0.18267738819122314,
0.005337047390639782,
0.031859856098890305,
0.07273654639720917,
0.2198667675256729,
-0.09987863898277283,
-0.1095680221915245,
-0.045107651501894,
0.17264553904533386,
0.04104243591427803,
0.06176810711622238,
-0.020409779623150826,
0.0026832649018615484,
-0.09016072750091553,
0.26033926010131836,
0.33503320813179016,
-0.05959988385438919,
0.06000038608908653,
-0.001532332506030798,
0.012618477456271648,
0.0598568320274353,
0.1106472909450531,
0.1316719800233841,
0.16129449009895325,
-0.08782772719860077,
-0.0036300979554653168,
-0.04469006136059761,
0.003653900232166052,
-0.15766099095344543,
0.07017975300550461,
-0.0036263535730540752,
-0.09436394274234772,
-0.022905346006155014,
0.057690780609846115,
-0.09149443358182907,
0.07677844911813736,
0.012266093865036964,
-0.18119388818740845,
-0.05720311030745506,
0.02039954625070095,
0.21919496357440948,
-0.017752492800354958,
0.05047750845551491,
-0.03148294985294342,
-0.05256124958395958,
0.07185778021812439,
-0.028453262522816658,
-0.16809456050395966,
-0.052983757108449936,
0.11245088279247284,
-0.033219534903764725,
0.11339452862739563,
-0.01019338145852089,
0.02442827820777893,
0.09866058081388474,
0.06523024290800095,
-0.09986453503370285,
0.06829454749822617,
0.024860212579369545,
-0.06409720331430435,
-0.025689758360385895,
-0.10032090544700623,
-0.02310752309858799,
-0.06327763199806213,
0.06990473717451096,
-0.10408949106931686,
0.047875236719846725,
-0.02276439219713211,
-0.041091352701187134,
-0.03285158798098564,
0.01771629974246025,
-0.06046866253018379,
0.07364235818386078,
0.030026888474822044,
-0.036611031740903854,
-0.07146476209163666,
-0.04183752462267876,
-0.04579772427678108,
0.008587779477238655,
-0.16310185194015503,
-0.11530570685863495,
-0.004291696939617395,
-0.022359905764460564,
0.07707220315933228,
0.03829155117273331,
-0.09312435239553452,
-0.03533054143190384,
-0.05517655238509178,
0.0333234928548336,
-0.12787498533725739,
0.0459769070148468,
0.08696694672107697,
-0.015718704089522362,
-0.010699767619371414,
-0.06356548517942429,
0.019441908225417137,
0.023609060794115067,
-0.10050466656684875,
-0.07246299833059311
] |
null | null |
transformers
|
# Cross-Encoder for MS Marco
This model was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html) for more details. The training code is available here: [SBERT.net Training MS Marco](https://github.com/UKPLab/sentence-transformers/tree/master/examples/training/ms_marco)
## Usage with Transformers
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Usage with SentenceTransformers
The usage becomes easier when you have [SentenceTransformers](https://www.sbert.net/) installed. Then, you can use the pre-trained models like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
scores = model.predict([('Query', 'Paragraph1'), ('Query', 'Paragraph2') , ('Query', 'Paragraph3')])
```
## Performance
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the [TREC Deep Learning 2019](https://microsoft.github.io/TREC-2019-Deep-Learning/) and the [MS Marco Passage Reranking](https://github.com/microsoft/MSMARCO-Passage-Ranking/) dataset.
| Model-Name | NDCG@10 (TREC DL 19) | MRR@10 (MS Marco Dev) | Docs / Sec |
| ------------- |:-------------| -----| --- |
| **Version 2 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2-v2 | 69.84 | 32.56 | 9000
| cross-encoder/ms-marco-MiniLM-L-2-v2 | 71.01 | 34.85 | 4100
| cross-encoder/ms-marco-MiniLM-L-4-v2 | 73.04 | 37.70 | 2500
| cross-encoder/ms-marco-MiniLM-L-6-v2 | 74.30 | 39.01 | 1800
| cross-encoder/ms-marco-MiniLM-L-12-v2 | 74.31 | 39.02 | 960
| **Version 1 models** | | |
| cross-encoder/ms-marco-TinyBERT-L-2 | 67.43 | 30.15 | 9000
| cross-encoder/ms-marco-TinyBERT-L-4 | 68.09 | 34.50 | 2900
| cross-encoder/ms-marco-TinyBERT-L-6 | 69.57 | 36.13 | 680
| cross-encoder/ms-marco-electra-base | 71.99 | 36.41 | 340
| **Other models** | | |
| nboost/pt-tinybert-msmarco | 63.63 | 28.80 | 2900
| nboost/pt-bert-base-uncased-msmarco | 70.94 | 34.75 | 340
| nboost/pt-bert-large-msmarco | 73.36 | 36.48 | 100
| Capreolus/electra-base-msmarco | 71.23 | 36.89 | 340
| amberoad/bert-multilingual-passage-reranking-msmarco | 68.40 | 35.54 | 330
| sebastian-hofstaetter/distilbert-cat-margin_mse-T2-msmarco | 72.82 | 37.88 | 720
Note: Runtime was computed on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/ms-marco-electra-base
|
[
"transformers",
"pytorch",
"electra",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #electra #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
Cross-Encoder for MS Marco
==========================
This model was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: Given a query, encode the query will all possible passages (e.g. retrieved with ElasticSearch). Then sort the passages in a decreasing order. See URL Retrieve & Re-rank for more details. The training code is available here: URL Training MS Marco
Usage with Transformers
-----------------------
Usage with SentenceTransformers
-------------------------------
The usage becomes easier when you have SentenceTransformers installed. Then, you can use the pre-trained models like this:
Performance
-----------
In the following table, we provide various pre-trained Cross-Encoders together with their performance on the TREC Deep Learning 2019 and the MS Marco Passage Reranking dataset.
Note: Runtime was computed on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #electra #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
45
] |
[
"passage: TAGS\n#transformers #pytorch #electra #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.04024451971054077,
0.1394282877445221,
-0.006416637450456619,
0.028075577691197395,
0.15309758484363556,
0.01857059821486473,
0.09045905619859695,
0.12177535146474838,
0.011697805486619473,
-0.049882590770721436,
0.13924641907215118,
0.23919491469860077,
-0.006401306949555874,
0.08411108702421188,
-0.09589783847332001,
-0.2561156153678894,
0.09550803154706955,
0.040639109909534454,
-0.024907702580094337,
0.09694972634315491,
0.1262826919555664,
-0.0372859351336956,
0.0483354777097702,
0.008995389565825462,
-0.04211917147040367,
0.025731168687343597,
0.028592536225914955,
-0.14406731724739075,
0.09406385570764542,
0.0621630996465683,
0.09353066235780716,
0.03694679215550423,
-0.024086112156510353,
-0.1928902119398117,
0.026310071349143982,
0.002302405424416065,
-0.09785549342632294,
0.052400074899196625,
0.03380037099123001,
-0.07051724940538406,
0.08566132187843323,
0.058593396097421646,
0.008490554988384247,
0.053670577704906464,
-0.08374246954917908,
-0.14458279311656952,
-0.07852615416049957,
0.07859311252832413,
0.08631940931081772,
0.08421549201011658,
0.03726954758167267,
0.15745727717876434,
-0.11553414165973663,
0.05882599949836731,
0.09277986735105515,
-0.3072034418582916,
0.0111320149153471,
0.03397609293460846,
0.041928134858608246,
0.015585123561322689,
0.001007251557894051,
0.03173469379544258,
0.05382874235510826,
0.019103143364191055,
0.01713571324944496,
-0.038105905055999756,
-0.11122983694076538,
0.036092307418584824,
-0.05960645526647568,
-0.0862732008099556,
0.2517440617084503,
-0.010789081454277039,
0.03434799611568451,
-0.013364816084504128,
-0.06594661623239517,
-0.007940991781651974,
-0.00317173614166677,
0.0677703320980072,
0.0001087129203369841,
0.09997144341468811,
0.10118997097015381,
-0.008642055094242096,
-0.13583175837993622,
0.013400900177657604,
-0.21188800036907196,
0.13454893231391907,
0.025790706276893616,
0.07544925063848495,
-0.12801332771778107,
0.0582682304084301,
0.0406012162566185,
-0.10812640935182571,
-0.007128621451556683,
-0.0808103159070015,
0.10145684331655502,
0.00537505280226469,
-0.08779856562614441,
0.025743994861841202,
0.14844615757465363,
0.2594009041786194,
0.0794718861579895,
0.015034779906272888,
-0.016524020582437515,
0.11119507998228073,
0.0022961304057389498,
0.07593061774969101,
0.028438640758395195,
0.01706576719880104,
0.10973414778709412,
-0.1238437294960022,
0.07520311325788498,
-0.04095808416604996,
-0.16450434923171997,
-0.02486010454595089,
-0.0017762617208063602,
0.12833058834075928,
0.04751913249492645,
0.04408520087599754,
-0.0488656722009182,
0.004286855924874544,
0.15226060152053833,
-0.07314354181289673,
0.0011187534546479583,
-0.0037508595269173384,
0.000644550658762455,
0.07853707671165466,
0.04579019919037819,
0.03356843441724777,
-0.0823741927742958,
0.09596745669841766,
-0.04118073731660843,
-0.020657889544963837,
-0.01995593123137951,
-0.01896272599697113,
0.09006624668836594,
-0.13740576803684235,
0.05944223329424858,
-0.17719943821430206,
-0.15579520165920258,
0.02343272976577282,
0.0567639134824276,
0.013866236433386803,
-0.11190372705459595,
0.028464799746870995,
0.002686518244445324,
0.04053020849823952,
-0.0864289179444313,
-0.06876781582832336,
-0.09028354287147522,
0.07457564771175385,
-0.07199487090110779,
0.022805485874414444,
-0.15917567908763885,
0.07116898894309998,
-0.1058075800538063,
-0.027042923495173454,
-0.050866615027189255,
-0.017219403758645058,
-0.06962284445762634,
0.17894667387008667,
-0.051072124391794205,
-0.04391330108046532,
0.027105361223220825,
0.021444199606776237,
-0.0567220002412796,
0.1178547590970993,
-0.06839759647846222,
-0.06930509209632874,
0.17668576538562775,
-0.14144107699394226,
-0.17492343485355377,
0.06966202706098557,
0.008232333697378635,
-0.020347634330391884,
0.08187971264123917,
0.10969138890504837,
0.0805254802107811,
-0.030976276844739914,
0.05714128538966179,
0.09338152408599854,
-0.0957266092300415,
-0.19246309995651245,
0.04815148934721947,
-0.044398706406354904,
-0.09391974657773972,
0.048111721873283386,
-0.006508640479296446,
0.09925005584955215,
0.004879864864051342,
-0.06286478787660599,
-0.05159604176878929,
-0.040863148868083954,
0.04632221907377243,
0.012806168757379055,
0.04861249774694443,
-0.07744750380516052,
-0.021156389266252518,
-0.012659648433327675,
0.043738968670368195,
0.038132574409246445,
0.043341927230358124,
-0.08769343048334122,
0.07317201793193817,
0.03563517704606056,
0.029601987451314926,
-0.15311481058597565,
-0.010886318050324917,
-0.02645612694323063,
0.021045319736003876,
0.014990728348493576,
0.06638722121715546,
0.017479626461863518,
-0.06817279011011124,
0.007545228116214275,
-0.018256766721606255,
0.115231953561306,
0.03910911828279495,
-0.005711002741008997,
-0.10419747978448868,
0.045774057507514954,
-0.040243010967969894,
0.04841990768909454,
-0.04744097590446472,
0.03727738931775093,
0.029863713309168816,
0.09556417912244797,
-0.04455886408686638,
0.10975157469511032,
-0.024643883109092712,
0.013350252993404865,
-0.08191020786762238,
0.005608132109045982,
0.0900387167930603,
0.04622197523713112,
-0.08618860691785812,
0.16641731560230255,
-0.09750746935606003,
0.25317510962486267,
0.190445214509964,
-0.17965662479400635,
0.05708780884742737,
-0.005851298104971647,
-0.007946527563035488,
0.0018073483370244503,
-0.00042592413956299424,
0.025624345988035202,
0.050502557307481766,
0.02388840541243553,
0.17720672488212585,
-0.041322074830532074,
-0.025940822437405586,
-0.015320357866585255,
-0.06915087252855301,
0.0011532638454809785,
0.05418197810649872,
0.18481361865997314,
-0.18892161548137665,
0.17879219353199005,
0.25811824202537537,
-0.029792683199048042,
0.07442477345466614,
-0.06929922848939896,
0.02268984541296959,
0.07036911696195602,
-0.052214909344911575,
-0.011649871245026588,
-0.0621630996465683,
-0.09995310008525848,
0.014498937875032425,
0.09725575894117355,
0.02947993576526642,
0.04098421335220337,
-0.12630021572113037,
-0.05124010890722275,
0.016034454107284546,
-0.013077307492494583,
-0.05003990978002548,
0.06871742010116577,
0.015374417416751385,
0.08863312751054764,
-0.03990410268306732,
-0.10956624150276184,
0.13317373394966125,
0.016460346058011055,
-0.07863547652959824,
0.15531882643699646,
-0.15623947978019714,
-0.265079140663147,
-0.1546567678451538,
-0.12503203749656677,
-0.009020118042826653,
0.026540031656622887,
0.1575431078672409,
-0.04157167673110962,
-0.06726272404193878,
0.009546620771288872,
-0.07972870767116547,
0.0012169755063951015,
0.014581684954464436,
-0.02148028463125229,
0.05833166465163231,
-0.01464900467544794,
-0.10707701742649078,
-0.05607196316123009,
0.021814193576574326,
-0.014264367520809174,
0.11674091964960098,
-0.08108065277338028,
0.07857286930084229,
0.13430099189281464,
0.012132087722420692,
0.02663179486989975,
-0.026369832456111908,
0.12187351286411285,
-0.04903124272823334,
-0.0022207251749932766,
0.250207781791687,
-0.04527536779642105,
0.08283510059118271,
0.1511315256357193,
0.02071426808834076,
-0.04144661873579025,
0.02659960463643074,
-0.05726110190153122,
-0.08931757509708405,
-0.2876962423324585,
-0.13703086972236633,
-0.09963475167751312,
0.0399465374648571,
0.06808318942785263,
0.09846090525388718,
0.15532617270946503,
0.08374204486608505,
-0.02614392712712288,
-0.016173049807548523,
0.005293949972838163,
0.07990442216396332,
0.2997828423976898,
0.01686898246407509,
0.12262818962335587,
-0.1313197761774063,
-0.0659727081656456,
0.11678404361009598,
0.04156997427344322,
0.14259211719036102,
0.13028112053871155,
0.07489220798015594,
0.058195099234580994,
0.11204276978969574,
0.11276265233755112,
0.13001494109630585,
0.05837985873222351,
0.0038573709316551685,
-0.045596618205308914,
0.004708520136773586,
-0.060464538633823395,
0.04057109355926514,
-0.011596299707889557,
-0.1525481790304184,
-0.056350432336330414,
-0.1299498826265335,
0.06227819249033928,
0.18490688502788544,
0.011361589655280113,
-0.17781929671764374,
0.01701386459171772,
0.07427063584327698,
-0.04004724696278572,
-0.06352955847978592,
0.07894083857536316,
-0.12588126957416534,
-0.1232147216796875,
0.153847336769104,
-0.027396807447075844,
0.1386658102273941,
-0.036836251616477966,
0.03411301225423813,
-0.039069753140211105,
-0.1176995038986206,
0.05875086411833763,
0.14249558746814728,
-0.253315269947052,
0.194421648979187,
-0.019304737448692322,
-0.03533247485756874,
-0.08516760915517807,
0.005249192006886005,
0.06463167071342468,
0.245378315448761,
0.05297144502401352,
0.004224834498018026,
-0.16488181054592133,
-0.05632767453789711,
-0.04478565603494644,
0.038439832627773285,
0.002593998098745942,
-0.021312732249498367,
-0.04078371822834015,
-0.0646892637014389,
-0.01242074929177761,
-0.016343146562576294,
0.015083102509379387,
-0.004813728388398886,
-0.16914843022823334,
0.04277850314974785,
0.11934252083301544,
0.09082981199026108,
-0.03510992228984833,
-0.0336446650326252,
-0.09059744328260422,
0.1370229870080948,
-0.139541894197464,
-0.09630466997623444,
-0.10171578079462051,
-0.11446097493171692,
0.0366479754447937,
-0.07600505650043488,
0.10345278680324554,
-0.07974762469530106,
0.01806996949017048,
-0.05596334487199783,
-0.17790526151657104,
0.09679289162158966,
-0.13203047215938568,
-0.04479781910777092,
-0.03123793564736843,
0.10218585282564163,
-0.08275012671947479,
0.007968029007315636,
0.042733967304229736,
0.020291447639465332,
-0.11023463308811188,
-0.1100921556353569,
-0.020460499450564384,
0.03185494244098663,
0.06743910908699036,
0.01607770286500454,
-0.06943298131227493,
-0.07042031735181808,
0.012650764547288418,
-0.06748540699481964,
0.2516140937805176,
0.17416873574256897,
-0.07574023306369781,
0.1617659032344818,
0.17042982578277588,
-0.0878354087471962,
-0.27645131945610046,
-0.19003352522850037,
-0.1284037083387375,
-0.08563404530286789,
0.0072863115929067135,
-0.18424472212791443,
0.11343926191329956,
0.06691261380910873,
-0.08302415907382965,
0.10592896491289139,
-0.16273947060108185,
-0.06736753135919571,
0.2054462432861328,
-0.06705047935247421,
0.2871759235858917,
-0.12948204576969147,
-0.07449353486299515,
-0.10361569374799728,
-0.16432106494903564,
0.13637487590312958,
-0.038800567388534546,
0.0602928102016449,
-0.014459117315709591,
0.0265655517578125,
-0.007882952690124512,
-0.036956220865249634,
0.12710867822170258,
-0.014967162162065506,
0.010111716575920582,
-0.14166724681854248,
0.007903330028057098,
0.019002703949809074,
-0.024332717061042786,
0.059651266783475876,
-0.17806826531887054,
0.020020918920636177,
-0.132245272397995,
-0.03736512362957001,
-0.048099830746650696,
0.09129953384399414,
0.03351367637515068,
-0.021978409960865974,
-0.034593887627124786,
-0.05407708138227463,
0.030377250164747238,
-0.0008416110649704933,
0.2564811408519745,
0.015424934215843678,
0.10035118460655212,
0.11503051221370697,
0.09328581392765045,
-0.1466190218925476,
-0.008832348510622978,
-0.12192293256521225,
-0.07277058809995651,
0.05477006733417511,
-0.1393607258796692,
0.05706823989748955,
0.11153620481491089,
-0.07808275520801544,
0.06865844875574112,
0.07176221162080765,
0.025690356269478798,
-0.02029947191476822,
0.14105558395385742,
-0.1430954933166504,
0.04574832320213318,
-0.004595305770635605,
0.14741545915603638,
0.0640120580792427,
0.06773573160171509,
0.12599653005599976,
0.024393822997808456,
-0.05459459498524666,
0.012501217424869537,
0.024452969431877136,
-0.028911540284752846,
0.038295239210128784,
0.05621488764882088,
0.017675278708338737,
-0.13390614092350006,
0.0964098572731018,
0.049044277518987656,
-0.1492808759212494,
-0.003441132139414549,
0.10814365744590759,
-0.14798347651958466,
-0.1383768618106842,
0.02242901362478733,
0.06571763753890991,
-0.2122051864862442,
-0.1294255256652832,
-0.06758717447519302,
-0.1314913034439087,
0.08356214314699173,
0.15618059039115906,
0.11304860562086105,
0.051549751311540604,
-0.008750321343541145,
-0.06876935809850693,
0.027284245938062668,
-0.01640613190829754,
-0.04039619863033295,
0.015225742943584919,
-0.11308631300926208,
-0.06689400225877762,
0.028903523460030556,
0.10405153036117554,
-0.045516327023506165,
-0.03453921899199486,
-0.06854874640703201,
0.04426014795899391,
-0.1456802487373352,
0.0278960969299078,
-0.07803834974765778,
-0.005698648747056723,
0.022678500041365623,
-0.06405041366815567,
-0.027794290333986282,
-0.011602615006268024,
-0.12157195061445236,
-0.004956862423568964,
-0.03411935642361641,
0.0728229209780693,
-0.10907217860221863,
-0.05473417788743973,
0.09611295163631439,
-0.015284935012459755,
0.11465848982334137,
0.07213246822357178,
-0.08271103352308273,
0.07898499816656113,
-0.14159905910491943,
-0.09054803103208542,
0.1161518394947052,
0.037348922342061996,
0.010520939715206623,
-0.024960726499557495,
0.03030102141201496,
0.11350668966770172,
-0.045109786093235016,
0.029737936332821846,
-0.01893732324242592,
-0.1524437814950943,
-0.024923453107476234,
-0.014129013754427433,
-0.13219018280506134,
-0.025927329435944557,
-0.10580310225486755,
0.1152222603559494,
0.019027678295969963,
0.2183641642332077,
-0.021270370110869408,
0.046936046332120895,
-0.02855025976896286,
0.023033645004034042,
-0.04410076141357422,
-0.15587663650512695,
-0.14079056680202484,
-0.06641631573438644,
-0.04875990003347397,
-0.009176901541650295,
0.25361019372940063,
0.019943062216043472,
-0.03752049803733826,
0.07169941812753677,
0.11796827614307404,
0.0026142774149775505,
0.01600082591176033,
0.24229003489017487,
0.06534610688686371,
-0.009016556665301323,
-0.08455488830804825,
-0.007550650741904974,
0.012588622979819775,
-0.10933476686477661,
0.0960666686296463,
0.06629017740488052,
0.05052799731492996,
0.07054765522480011,
0.014023473486304283,
0.031026961281895638,
-0.10648991167545319,
-0.17085735499858856,
0.03367779403924942,
0.08086472004652023,
0.05332938954234123,
0.1047055795788765,
0.1367921382188797,
-0.0185170229524374,
0.005783441010862589,
-0.04609041288495064,
0.0018499634461477399,
-0.18199138343334198,
-0.14808903634548187,
-0.0893586128950119,
-0.12070519477128983,
0.010083816014230251,
-0.04544419050216675,
0.025865929201245308,
0.11981025338172913,
0.03441726788878441,
-0.07407727837562561,
-0.04958098381757736,
-0.013592684641480446,
-0.05714087933301926,
0.018829094246029854,
-0.043657951056957245,
-0.004283376969397068,
0.0077462331391870975,
-0.045971550047397614,
-0.09671482443809509,
-0.026886535808444023,
-0.04157080501317978,
0.04619024321436882,
-0.01343106385320425,
0.06469076126813889,
-0.13383154571056366,
-0.07984981685876846,
-0.03289703279733658,
0.04114929959177971,
-0.0024660674389451742,
0.1623561531305313,
0.012266039848327637,
0.031103331595659256,
0.10458961874246597,
0.17347536981105804,
-0.05008632689714432,
-0.15108516812324524,
-0.05923185124993324,
0.14117980003356934,
0.0859256163239479,
0.06251697987318039,
0.015807492658495903,
0.0033862146083265543,
-0.06002199277281761,
0.2585621774196625,
0.23214176297187805,
-0.01547662541270256,
0.03307679295539856,
-0.04826577007770538,
0.021224405616521835,
0.11025270074605942,
0.10972068458795547,
0.12476218491792679,
0.17299842834472656,
-0.07445333153009415,
-0.023825248703360558,
-0.05336571857333183,
0.020602496340870857,
-0.17988669872283936,
0.04538416862487793,
-0.02621188387274742,
-0.09329324215650558,
-0.0012173529248684645,
0.12378253042697906,
-0.1319345235824585,
0.11799726635217667,
-0.005860974546521902,
-0.09502888470888138,
-0.044742804020643234,
-0.01523672230541706,
0.18205033242702484,
0.04049639031291008,
0.024224212393164635,
-0.02306114137172699,
-0.09810671210289001,
0.1000199019908905,
-0.003367006778717041,
-0.22396411001682281,
-0.05375676229596138,
0.08661927282810211,
-0.012011992745101452,
0.07649670541286469,
0.0002248782111564651,
0.07127409428358078,
0.08224761486053467,
0.07496672123670578,
-0.07651036977767944,
0.09888625890016556,
0.01087881438434124,
-0.017818013206124306,
0.05433560535311699,
-0.0963621437549591,
-0.023679740726947784,
-0.08142072707414627,
0.04777707904577255,
-0.08670085668563843,
0.04945719614624977,
-0.006629862356930971,
-0.04589192196726799,
-0.02703871764242649,
0.06799701601266861,
-0.0724881961941719,
0.05630885809659958,
0.003155894111841917,
-0.03590865060687065,
-0.06303263455629349,
-0.06387297809123993,
-0.024654222652316093,
0.038610395044088364,
-0.19419945776462555,
-0.08653511106967926,
-0.01225605420768261,
-0.03622356057167053,
0.06964404881000519,
0.05298729985952377,
-0.09691957384347916,
-0.010787577368319035,
-0.13345758616924286,
0.007239778526127338,
-0.16423611342906952,
0.043623339384794235,
0.07419291883707047,
-0.006328477989882231,
-0.014007709920406342,
-0.015221481211483479,
0.020509619265794754,
0.02966417372226715,
-0.1079070121049881,
-0.08729416131973267
] |
null | null |
transformers
|
# Cross-Encoder for MS MARCO - EN-DE
This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html).
The training code is available in this repository, see `train_script.py`.
## Usage with SentenceTransformers
When you have [SentenceTransformers](https://www.sbert.net/) installed, you can use the model like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
query = 'How many people live in Berlin?'
docs = ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.']
pairs = [(query, doc) for doc in docs]
scores = model.predict(pairs)
```
## Usage with Transformers
With the transformers library, you can use the model like this:
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Performance
The performance was evaluated on three datasets:
- **TREC-DL19 EN-EN**: The original [TREC 2019 Deep Learning Track](https://microsoft.github.io/msmarco/TREC-Deep-Learning-2019.html): Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47.
- **TREC-DL19 DE-EN**: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10.
- **GermanDPR DE-DE**: The [GermanDPR](https://www.deepset.ai/germanquad) dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27.
We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search.
| Model-Name | TREC-DL19 EN-EN | TREC-DL19 DE-EN | GermanDPR DE-DE | Docs / Sec |
| ------------- |:-------------:| :-----: | :---: | :----: |
| BM25 | 45.46 | - | 35.85 | -|
| **Cross-Encoder Re-Rankers** | | | |
| [cross-encoder/msmarco-MiniLM-L6-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L6-en-de-v1) | 72.43 | 65.53 | 46.77 | 1600 |
| [cross-encoder/msmarco-MiniLM-L12-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L12-en-de-v1) | 72.94 | 66.07 | 49.91 | 900 |
| [svalabs/cross-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/cross-electra-ms-marco-german-uncased) (DE only) | - | - | 53.67 | 260 |
| [deepset/gbert-base-germandpr-reranking](https://huggingface.co/deepset/gbert-base-germandpr-reranking) (DE only) | - | - | 53.59 | 260 |
| **Bi-Encoders (re-ranking)** | | | |
| [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned) | 63.38 | 58.28 | 37.88 | 940 |
| [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch) | 65.51 | 58.69 | 38.32 | 940 |
| [svalabs/bi-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/bi-electra-ms-marco-german-uncased) (DE only) | - | - | 34.31 | 450 |
| [deepset/gbert-base-germandpr-question_encoder](https://huggingface.co/deepset/gbert-base-germandpr-question_encoder) (DE only) | - | - | 42.55 | 450 |
Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/msmarco-MiniLM-L12-en-de-v1
|
[
"transformers",
"pytorch",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
Cross-Encoder for MS MARCO - EN-DE
==================================
This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: See URL Retrieve & Re-rank.
The training code is available in this repository, see 'train\_script.py'.
Usage with SentenceTransformers
-------------------------------
When you have SentenceTransformers installed, you can use the model like this:
Usage with Transformers
-----------------------
With the transformers library, you can use the model like this:
Performance
-----------
The performance was evaluated on three datasets:
* TREC-DL19 EN-EN: The original TREC 2019 Deep Learning Track: Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47.
* TREC-DL19 DE-EN: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10.
* GermanDPR DE-DE: The GermanDPR dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27.
We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search.
Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
48
] |
[
"passage: TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
-0.011929498054087162,
0.1109679788351059,
-0.004657326731830835,
0.036296773701906204,
0.08949180692434311,
0.01468193531036377,
0.1052534207701683,
0.14302724599838257,
0.00024416513042524457,
-0.031693242490291595,
0.1409340649843216,
0.17007292807102203,
-0.03366948291659355,
0.0719553679227829,
-0.07565809041261673,
-0.24093829095363617,
0.10701024532318115,
0.03007899411022663,
-0.0497613325715065,
0.0865328311920166,
0.11767031252384186,
-0.06318635493516922,
0.0450325682759285,
-0.002566360170021653,
-0.062239885330200195,
0.03159003704786301,
0.01839093491435051,
-0.12103993445634842,
0.1024293452501297,
0.03353764861822128,
0.10411366820335388,
0.050487738102674484,
-0.03610473498702049,
-0.15597668290138245,
0.02397605963051319,
0.02407376654446125,
-0.09014851599931717,
0.06582070142030716,
0.042276762425899506,
-0.03844092786312103,
0.04500403627753258,
0.005638193804770708,
-0.001494656433351338,
0.03218536451458931,
-0.09311587363481522,
-0.19191624224185944,
-0.07325631380081177,
0.07541981339454651,
0.023964963853359222,
0.08019165694713593,
0.023596404120326042,
0.15867139399051666,
-0.14464761316776276,
0.05357449874281883,
0.13281987607479095,
-0.34267017245292664,
0.0014128641923889518,
0.12683846056461334,
0.06185078248381615,
0.03614772856235504,
-0.03367999568581581,
0.05115452781319618,
0.05234549194574356,
0.01213220227509737,
0.053060855716466904,
-0.038291096687316895,
-0.12301186472177505,
0.06110190600156784,
-0.07589741796255112,
-0.08090674132108688,
0.2470371127128601,
-0.020999861881136894,
0.06525891274213791,
-0.02700621820986271,
-0.07864753156900406,
-0.015083619393408298,
0.0020947037264704704,
0.05717647820711136,
0.023739924654364586,
0.10291964560747147,
0.07124892622232437,
-0.034733641892671585,
-0.15194028615951538,
0.028371602296829224,
-0.2115340530872345,
0.07593068480491638,
0.003470568684861064,
0.0676390752196312,
-0.12714898586273193,
0.06812293827533722,
0.01710754819214344,
-0.11910293996334076,
0.02460039034485817,
-0.07910972088575363,
0.10053464025259018,
0.02612384781241417,
-0.10773281753063202,
0.0519438311457634,
0.12186801433563232,
0.198510080575943,
0.02410656027495861,
-0.0036114167887717485,
-0.03310275822877884,
0.1262211948633194,
0.0009386635501869023,
0.09358032047748566,
-0.0214508268982172,
-0.0191908348351717,
0.07524799555540085,
-0.08410874754190445,
0.06491826474666595,
-0.06015123426914215,
-0.1755080223083496,
-0.04224881902337074,
0.02791069634258747,
0.08877983689308167,
0.0867546871304512,
0.06018778309226036,
-0.023768959566950798,
0.027124647051095963,
0.14338789880275726,
-0.06075986847281456,
0.035603079944849014,
-0.0025226115249097347,
0.03203103691339493,
0.02896995283663273,
0.04888737201690674,
0.024378491565585136,
-0.028933117166161537,
0.0786980390548706,
-0.0584336593747139,
-0.012409459799528122,
-0.042271099984645844,
-0.05858355015516281,
0.08736852556467056,
-0.1067584678530693,
0.026230912655591965,
-0.16472218930721283,
-0.0831773653626442,
0.01852068118751049,
0.06086645647883415,
0.0007838314631953835,
-0.09153826534748077,
0.07393558323383331,
-0.040153417736291885,
0.07722245901823044,
-0.08859124779701233,
0.0032021517399698496,
-0.09344978630542755,
0.057036347687244415,
-0.1038811206817627,
0.04983983933925629,
-0.18570169806480408,
0.06897924840450287,
-0.10177050530910492,
-0.0035829299595206976,
-0.048835549503564835,
-0.03398766368627548,
-0.0877050906419754,
0.1558571755886078,
-0.02481936104595661,
-0.03799108788371086,
-0.02024412341415882,
0.031119640916585922,
-0.046811964362859726,
0.08756688982248306,
-0.1204364225268364,
-0.03759213536977768,
0.1446077972650528,
-0.09290962666273117,
-0.1673811823129654,
0.06678478419780731,
0.004953649360686541,
0.0004414678260218352,
0.023817099630832672,
0.2068229466676712,
0.06968206912279129,
-0.031353238970041275,
0.03667166829109192,
0.15588927268981934,
-0.036718402057886124,
-0.1597534567117691,
0.05581526458263397,
-0.036597561091184616,
-0.07861929386854172,
0.04289300739765167,
-0.020597059279680252,
0.07218319177627563,
0.013152756728231907,
-0.0813322514295578,
-0.07919875532388687,
-0.03760673850774765,
0.056981250643730164,
0.010442853905260563,
0.09089142829179764,
-0.0736541673541069,
-0.03648189455270767,
0.012337546795606613,
0.029745422303676605,
0.05869894474744797,
0.06688299775123596,
-0.03138571232557297,
0.10668440163135529,
0.013989963568747044,
0.018267758190631866,
-0.15533185005187988,
-0.016854720190167427,
-0.027512626722455025,
0.033168140798807144,
0.0027375284116715193,
0.16843344271183014,
0.028516249731183052,
-0.08179625123739243,
-0.005133424885571003,
-0.0027407007291913033,
0.1116393432021141,
0.07179897278547287,
-0.04622254520654678,
-0.1110997349023819,
0.02547508105635643,
-0.05257298797369003,
0.013891373760998249,
-0.02169971913099289,
0.0376596562564373,
0.06746970117092133,
0.10060187429189682,
-0.054126087576150894,
0.11025400459766388,
-0.02443268522620201,
0.01063956692814827,
-0.08723245561122894,
0.0012760007521137595,
0.09992439299821854,
0.03174593299627304,
-0.06503839790821075,
0.2191265970468521,
-0.11319835484027863,
0.2770872712135315,
0.2274661362171173,
-0.1816493421792984,
0.06211845576763153,
0.010664228349924088,
-0.0253766942769289,
0.016376934945583344,
0.00973495189100504,
-0.005615867208689451,
-0.017049739137291908,
-0.014084290713071823,
0.1513899713754654,
-0.04333732649683952,
-0.03485063835978508,
-0.028493531048297882,
-0.06856098771095276,
-0.03320147842168808,
0.043046679347753525,
0.09611020982265472,
-0.14282885193824768,
0.21591053903102875,
0.3771267533302307,
-0.03934421390295029,
0.1023554727435112,
-0.05790102481842041,
0.029091954231262207,
0.03956243023276329,
-0.07205614447593689,
-0.05272269248962402,
0.0077779870480299,
-0.14636331796646118,
-0.011522377841174603,
0.11102425307035446,
0.05511642247438431,
0.054583270102739334,
-0.14314037561416626,
-0.047039732336997986,
0.022951988503336906,
0.011035983450710773,
-0.05264776572585106,
0.08396845310926437,
0.028842180967330933,
0.10083223134279251,
-0.018218370154500008,
-0.09925662726163864,
0.10900299996137619,
0.0034230707678943872,
-0.04919014126062393,
0.1237892210483551,
-0.17483505606651306,
-0.2666981816291809,
-0.12263266742229462,
-0.11647240817546844,
-0.03185277059674263,
0.02454465627670288,
0.1553153693675995,
-0.05448857694864273,
-0.04608921706676483,
-0.0322968028485775,
-0.07930576056241989,
-0.03999684005975723,
0.03336773440241814,
-0.055547602474689484,
0.051703520119190216,
0.004006838425993919,
-0.12592794001102448,
-0.0660366490483284,
0.019448257982730865,
-0.0385357066988945,
0.10267893224954605,
-0.03189206123352051,
0.0755273625254631,
0.1416429579257965,
0.008576683700084686,
0.024744180962443352,
-0.028986504301428795,
0.1633431613445282,
-0.04187033325433731,
0.017040017992258072,
0.2164611518383026,
-0.008377695456147194,
0.07641872763633728,
0.18111683428287506,
0.03418927267193794,
-0.021802285686135292,
0.013025326654314995,
-0.036622799932956696,
-0.09112220257520676,
-0.21751870214939117,
-0.13478365540504456,
-0.11826911568641663,
0.022339027374982834,
0.04935091361403465,
0.10200902074575424,
0.12801212072372437,
0.05823846161365509,
-0.0066392021253705025,
-0.0018668529810383916,
-0.03226018324494362,
0.05748061463236809,
0.2714395523071289,
-0.03477463126182556,
0.14819587767124176,
-0.10991998016834259,
-0.06027120351791382,
0.12198864668607712,
0.0809679850935936,
0.10323648899793625,
0.10058606415987015,
0.02176450751721859,
0.07683505862951279,
0.17835702002048492,
0.10593564063310623,
0.09460754692554474,
0.03239644691348076,
-0.0022547899279743433,
-0.053684331476688385,
-0.020651666447520256,
-0.021006306633353233,
0.06357185542583466,
0.03231940418481827,
-0.15682291984558105,
-0.02531340718269348,
-0.19327817857265472,
0.09269165247678757,
0.15072937309741974,
0.054205313324928284,
-0.1248549297451973,
0.032165903598070145,
0.09894848614931107,
-0.013840306550264359,
-0.04561185464262962,
0.0807991772890091,
-0.029446594417095184,
-0.08341233432292938,
0.08656627684831619,
0.009248846210539341,
0.12374022603034973,
0.0148047786206007,
0.07288803160190582,
-0.04254555702209473,
-0.17002898454666138,
0.059988267719745636,
0.129883274435997,
-0.27484795451164246,
0.20877805352210999,
-0.026468385010957718,
-0.09377090632915497,
-0.0814322680234909,
0.0007435722509399056,
0.07990670949220657,
0.2424503117799759,
0.04600871354341507,
0.034760482609272,
-0.14131808280944824,
-0.0642140805721283,
-0.04361937195062637,
0.0123197752982378,
0.023223701864480972,
-0.01777510903775692,
-0.03755996376276016,
-0.06066042184829712,
-0.0020570766646414995,
0.021542828530073166,
0.14491324126720428,
-0.019954849034547806,
-0.17723368108272552,
0.0679810494184494,
0.1139739453792572,
0.020291222259402275,
-0.05237312614917755,
-0.04039272293448448,
-0.16760264337062836,
0.15746107697486877,
-0.04734361916780472,
-0.05293496698141098,
-0.09933813661336899,
-0.11485697329044342,
0.05627376213669777,
-0.04504356533288956,
0.06802637875080109,
-0.08537643402814865,
0.018626293167471886,
-0.07503805309534073,
-0.1825282722711563,
0.10788697749376297,
-0.1172252967953682,
-0.04584076628088951,
-0.04347958415746689,
0.0893806666135788,
-0.14524905383586884,
0.04412192851305008,
0.036863263696432114,
0.05305350199341774,
-0.1468811184167862,
-0.11997350305318832,
-0.012302463874220848,
0.013916125521063805,
0.06559175997972488,
-0.026150045916438103,
-0.08259673416614532,
-0.05353347212076187,
0.05847072973847389,
-0.032503772526979446,
0.256398469209671,
0.1776508241891861,
-0.1236819177865982,
0.1738201528787613,
0.13180740177631378,
-0.07809863984584808,
-0.3454630970954895,
-0.14732776582241058,
-0.15849560499191284,
-0.07533957064151764,
0.0235506072640419,
-0.11653820425271988,
0.10535410046577454,
0.03611930459737778,
-0.10646899044513702,
0.05344188213348389,
-0.17473436892032623,
-0.08326824009418488,
0.19634291529655457,
-0.08626771718263626,
0.30544474720954895,
-0.15087944269180298,
-0.06414706259965897,
-0.07380376756191254,
-0.1340590864419937,
0.14970876276493073,
-0.12145139276981354,
0.07197150588035583,
-0.012666989117860794,
0.022509323433041573,
-0.0005715735605917871,
-0.05197785422205925,
0.1541745513677597,
-0.03773857280611992,
0.025772521272301674,
-0.1358940750360489,
-0.03740015998482704,
0.0981537252664566,
-0.06408584862947464,
0.032355599105358124,
-0.1558215320110321,
0.017573487013578415,
-0.13796880841255188,
0.006053187418729067,
-0.07947039604187012,
0.09438689798116684,
-0.003972133621573448,
-0.041665222495794296,
-0.06218144670128822,
-0.012985005043447018,
0.038268182426691055,
-0.013135287910699844,
0.26377928256988525,
0.0018693876918405294,
0.11646835505962372,
0.1741446703672409,
0.02918052300810814,
-0.16934606432914734,
-0.019878719002008438,
-0.0496620275080204,
-0.07310948520898819,
0.06996731460094452,
-0.18846379220485687,
0.05133286863565445,
0.08059290796518326,
-0.07819657772779465,
0.057422809302806854,
0.08675773441791534,
0.022930843755602837,
-0.04511556401848793,
0.16467398405075073,
-0.16118508577346802,
-0.0038538638036698103,
-0.005934640299528837,
0.10133349150419235,
0.05118110030889511,
0.021290142089128494,
0.13166667520999908,
0.017662744969129562,
-0.031207405030727386,
0.02763739600777626,
0.017550131306052208,
-0.06095219403505325,
0.019465308636426926,
0.06818114966154099,
0.027803484350442886,
-0.1139194443821907,
0.0774909257888794,
0.06724732369184494,
-0.12166350334882736,
-0.01452089287340641,
0.08962465077638626,
-0.1174732893705368,
-0.16081809997558594,
-0.000930252717807889,
0.031174184754490852,
-0.11175857484340668,
-0.0874028131365776,
-0.01669522002339363,
-0.1344173103570938,
0.05237565562129021,
0.10343516618013382,
0.13125064969062805,
0.07057765126228333,
-0.011259262450039387,
-0.069544717669487,
0.05991494655609131,
-0.003658764064311981,
-0.06904593855142593,
0.04123999923467636,
-0.11242011934518814,
-0.01196740660816431,
0.009768360294401646,
0.11199247092008591,
-0.062344446778297424,
-0.005306920036673546,
-0.11247684061527252,
0.0028180480003356934,
-0.14773620665073395,
-0.02894614078104496,
-0.07614428550004959,
-0.031011177226901054,
0.02214771695435047,
-0.08436717092990875,
-0.039100389927625656,
-0.014332986436784267,
-0.13986822962760925,
-0.023398039862513542,
-0.04736753925681114,
0.09517386555671692,
-0.11655298620462418,
-0.055166248232126236,
0.09626336395740509,
-0.014291220344603062,
0.09398853778839111,
0.044903021305799484,
-0.07022914290428162,
0.06434562802314758,
-0.09496074169874191,
-0.11316927522420883,
0.0959542915225029,
0.04407193884253502,
0.04776572808623314,
0.002675051800906658,
-0.009801552630960941,
0.0964970737695694,
-0.0004672814975492656,
0.04560285806655884,
0.029566984623670578,
-0.13350655138492584,
-0.0154465576633811,
-0.023506714031100273,
-0.1332763433456421,
0.013271442614495754,
-0.0963786169886589,
0.14655865728855133,
0.02278325706720352,
0.1790017932653427,
-0.006577861960977316,
0.03352291136980057,
-0.06583039462566376,
0.02249913476407528,
-0.054776906967163086,
-0.1789943426847458,
-0.11596626788377762,
-0.06996689736843109,
-0.029697518795728683,
-0.02017328143119812,
0.2981296181678772,
0.08490652590990067,
-0.08169146627187729,
0.07198161631822586,
0.09890054166316986,
0.014289051294326782,
0.010698960162699223,
0.22402669489383698,
0.06494433432817459,
-0.015415017493069172,
-0.06803465634584427,
0.013257022015750408,
0.018359694629907608,
-0.0557638444006443,
0.07606630772352219,
0.11115698516368866,
0.0904100313782692,
0.07412039488554001,
0.03014315851032734,
-0.010024036280810833,
-0.14307071268558502,
-0.14855130016803741,
0.009788756258785725,
0.11858081817626953,
-0.04020537808537483,
0.05838916823267937,
0.11923600733280182,
-0.046544451266527176,
0.04348647594451904,
-0.0696643590927124,
0.006750994361937046,
-0.17443858087062836,
-0.09916237741708755,
-0.07066202163696289,
-0.1291390359401703,
-0.028648626059293747,
-0.05008450523018837,
0.03779691457748413,
0.13888272643089294,
0.03173473849892616,
-0.03977445513010025,
-0.0031719948165118694,
-0.03568726405501366,
-0.04079877957701683,
0.028132043778896332,
-0.027076909318566322,
0.008340715430676937,
-0.060388870537281036,
-0.03699566051363945,
-0.10331396758556366,
-0.012213758192956448,
-0.05525130033493042,
0.046983782202005386,
-0.02489228919148445,
0.022161008790135384,
-0.12678012251853943,
-0.07989493757486343,
-0.058136142790317535,
0.021603651344776154,
-0.005209204275161028,
0.1960669904947281,
0.0022967387922108173,
0.03324297443032265,
0.073024682700634,
0.21338893473148346,
-0.09978115558624268,
-0.13632486760616302,
-0.03914077952504158,
0.16939033567905426,
0.040507927536964417,
0.05812970548868179,
-0.020286252722144127,
-0.0033802634570747614,
-0.09350753575563431,
0.2765567898750305,
0.3296546936035156,
-0.06144540011882782,
0.059222228825092316,
-0.002428764943033457,
0.015853505581617355,
0.07856421172618866,
0.11569953709840775,
0.13662907481193542,
0.16383591294288635,
-0.08296691626310349,
-0.01556343026459217,
-0.047031085938215256,
0.0007883654325269163,
-0.16509802639484406,
0.0851517915725708,
-0.007957077585160732,
-0.09604888409376144,
-0.027654562145471573,
0.06391361355781555,
-0.09459078311920166,
0.08622651547193527,
0.007731959223747253,
-0.1700391173362732,
-0.0547446571290493,
0.02429783344268799,
0.21584919095039368,
-0.01918049342930317,
0.048848070204257965,
-0.027314655482769012,
-0.052800264209508896,
0.07536984980106354,
-0.027961991727352142,
-0.17292775213718414,
-0.030522989109158516,
0.10590537637472153,
-0.03492295742034912,
0.1096210852265358,
-0.005980927962809801,
0.02427011914551258,
0.09679204970598221,
0.07734619826078415,
-0.09253719449043274,
0.08694414794445038,
0.023542344570159912,
-0.0641736388206482,
-0.014475682750344276,
-0.10951966792345047,
-0.02482822723686695,
-0.08250825107097626,
0.06349749118089676,
-0.11941738426685333,
0.054807331413030624,
0.002499351045116782,
-0.0450555719435215,
-0.03437528386712074,
0.008221501484513283,
-0.06355006247758865,
0.06977298855781555,
0.028691314160823822,
-0.038544271141290665,
-0.07007227838039398,
-0.04650425910949707,
-0.052271898835897446,
-0.0004606132279150188,
-0.1830073893070221,
-0.10946376621723175,
-0.007405308540910482,
-0.023830890655517578,
0.08056151121854782,
0.04008973389863968,
-0.08659695833921432,
-0.039661966264247894,
-0.059462614357471466,
0.037761665880680084,
-0.13403542339801788,
0.05026970058679581,
0.08200293034315109,
-0.01854803040623665,
-0.008469896391034126,
-0.052380140870809555,
0.018549039959907532,
0.021080557256937027,
-0.09995073825120926,
-0.06921897828578949
] |
null | null |
transformers
|
# Cross-Encoder for MS MARCO - EN-DE
This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the [MS Marco Passage Ranking](https://github.com/microsoft/MSMARCO-Passage-Ranking) task.
The model can be used for Information Retrieval: See [SBERT.net Retrieve & Re-rank](https://www.sbert.net/examples/applications/retrieve_rerank/README.html).
The training code is available in this repository, see `train_script.py`.
## Usage with SentenceTransformers
When you have [SentenceTransformers](https://www.sbert.net/) installed, you can use the model like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name', max_length=512)
query = 'How many people live in Berlin?'
docs = ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.']
pairs = [(query, doc) for doc in docs]
scores = model.predict(pairs)
```
## Usage with Transformers
With the transformers library, you can use the model like this:
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'How many people live in Berlin?'], ['Berlin has a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
print(scores)
```
## Performance
The performance was evaluated on three datasets:
- **TREC-DL19 EN-EN**: The original [TREC 2019 Deep Learning Track](https://microsoft.github.io/msmarco/TREC-Deep-Learning-2019.html): Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47.
- **TREC-DL19 DE-EN**: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10.
- **GermanDPR DE-DE**: The [GermanDPR](https://www.deepset.ai/germanquad) dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27.
We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search.
| Model-Name | TREC-DL19 EN-EN | TREC-DL19 DE-EN | GermanDPR DE-DE | Docs / Sec |
| ------------- |:-------------:| :-----: | :---: | :----: |
| BM25 | 45.46 | - | 35.85 | -|
| **Cross-Encoder Re-Rankers** | | | |
| [cross-encoder/msmarco-MiniLM-L6-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L6-en-de-v1) | 72.43 | 65.53 | 46.77 | 1600 |
| [cross-encoder/msmarco-MiniLM-L12-en-de-v1](https://huggingface.co/cross-encoder/msmarco-MiniLM-L12-en-de-v1) | 72.94 | 66.07 | 49.91 | 900 |
| [svalabs/cross-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/cross-electra-ms-marco-german-uncased) (DE only) | - | - | 53.67 | 260 |
| [deepset/gbert-base-germandpr-reranking](https://huggingface.co/deepset/gbert-base-germandpr-reranking) (DE only) | - | - | 53.59 | 260 |
| **Bi-Encoders (re-ranking)** | | | |
| [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-lng-aligned) | 63.38 | 58.28 | 37.88 | 940 |
| [sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch](https://huggingface.co/sentence-transformers/msmarco-distilbert-multilingual-en-de-v2-tmp-trained-scratch) | 65.51 | 58.69 | 38.32 | 940 |
| [svalabs/bi-electra-ms-marco-german-uncased](https://huggingface.co/svalabs/bi-electra-ms-marco-german-uncased) (DE only) | - | - | 34.31 | 450 |
| [deepset/gbert-base-germandpr-question_encoder](https://huggingface.co/deepset/gbert-base-germandpr-question_encoder) (DE only) | - | - | 42.55 | 450 |
Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/msmarco-MiniLM-L6-en-de-v1
|
[
"transformers",
"pytorch",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
Cross-Encoder for MS MARCO - EN-DE
==================================
This is a cross-lingual Cross-Encoder model for EN-DE that can be used for passage re-ranking. It was trained on the MS Marco Passage Ranking task.
The model can be used for Information Retrieval: See URL Retrieve & Re-rank.
The training code is available in this repository, see 'train\_script.py'.
Usage with SentenceTransformers
-------------------------------
When you have SentenceTransformers installed, you can use the model like this:
Usage with Transformers
-----------------------
With the transformers library, you can use the model like this:
Performance
-----------
The performance was evaluated on three datasets:
* TREC-DL19 EN-EN: The original TREC 2019 Deep Learning Track: Given an English query and 1000 documents (retrieved by BM25 lexical search), rank documents with according to their relevance. We compute NDCG@10. BM25 achieves a score of 45.46, a perfect re-ranker can achieve a score of 95.47.
* TREC-DL19 DE-EN: The English queries of TREC-DL19 have been translated by a German native speaker to German. We rank the German queries versus the English passages from the original TREC-DL19 setup. We compute NDCG@10.
* GermanDPR DE-DE: The GermanDPR dataset provides German queries and German passages from Wikipedia. We indexed the 2.8 Million paragraphs from German Wikipedia and retrieved for each query the top 100 most relevant passages using BM25 lexical search with Elasticsearch. We compute MRR@10. BM25 achieves a score of 35.85, a perfect re-ranker can achieve a score of 76.27.
We also check the performance of bi-encoders using the same evaluation: The retrieved documents from BM25 lexical search are re-ranked using query & passage embeddings with cosine-similarity. Bi-Encoders can also be used for end-to-end semantic search.
Note: Docs / Sec gives the number of (query, document) pairs we can re-rank within a second on a V100 GPU.
|
[] |
[
"TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
44
] |
[
"passage: TAGS\n#transformers #pytorch #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.033624548465013504,
0.11583641916513443,
-0.006466314196586609,
0.03693987429141998,
0.11971456557512283,
0.02166047692298889,
0.12048175185918808,
0.1321706920862198,
0.01429376658052206,
-0.07648304849863052,
0.13479697704315186,
0.22184574604034424,
-0.012524329125881195,
0.06779264658689499,
-0.08648756891489029,
-0.24625052511692047,
0.09145351499319077,
0.047294795513153076,
-0.014880131930112839,
0.09696296602487564,
0.11783695966005325,
-0.039152491837739944,
0.04760254546999931,
-0.001044782460667193,
-0.06468931585550308,
0.021656470373272896,
0.04621468111872673,
-0.12399723380804062,
0.08577025681734085,
0.04309383034706116,
0.08981439471244812,
0.043990809470415115,
-0.02796093188226223,
-0.18899650871753693,
0.017944956198334694,
0.004928646609187126,
-0.08463240414857864,
0.05281449854373932,
0.06530068069696426,
-0.05573006346821785,
0.03398431837558746,
0.04542854055762291,
-0.0016487948596477509,
0.05704890564084053,
-0.07702083140611649,
-0.16284798085689545,
-0.07866179198026657,
0.09533879160881042,
0.07475922256708145,
0.07963944226503372,
0.04054330289363861,
0.14942073822021484,
-0.12101521342992783,
0.07066715508699417,
0.10073971748352051,
-0.34290656447410583,
0.006926678586751223,
0.07116013020277023,
0.014098972082138062,
0.017672913148999214,
-0.019676387310028076,
0.03557991981506348,
0.04636206105351448,
0.012471802532672882,
0.013373504392802715,
-0.05803173407912254,
-0.11233073472976685,
0.033623434603214264,
-0.05398819223046303,
-0.06736930459737778,
0.23002012073993683,
-0.017524652183055878,
0.029936393722891808,
-0.010111764073371887,
-0.06318812817335129,
0.013274782337248325,
-0.021011902019381523,
0.05844603851437569,
0.015864230692386627,
0.10881900787353516,
0.07598614692687988,
-0.022992447018623352,
-0.134002223610878,
0.009127716533839703,
-0.21311087906360626,
0.09606034308671951,
0.021664461120963097,
0.07777068018913269,
-0.1399146467447281,
0.06432689726352692,
0.020169859752058983,
-0.11113383620977402,
0.0031303863506764174,
-0.08473936468362808,
0.11171188205480576,
0.014474593102931976,
-0.07460460811853409,
0.043622326105833054,
0.13258297741413116,
0.22116319835186005,
0.05043065547943115,
0.01951022446155548,
-0.058310430496931076,
0.11443714052438736,
-0.029090533033013344,
0.09178415685892105,
0.02817927859723568,
-0.011186737567186356,
0.10911470651626587,
-0.12157052755355835,
0.06358799338340759,
-0.035286471247673035,
-0.1657385379076004,
-0.02933056652545929,
0.015140064992010593,
0.11050593107938766,
0.05076056346297264,
0.06345071643590927,
-0.04509174823760986,
0.00138133205473423,
0.15608423948287964,
-0.06623833626508713,
0.012619871646165848,
0.00755408825352788,
0.022991495206952095,
0.06565660983324051,
0.04396168887615204,
0.02844872511923313,
-0.0714418962597847,
0.10865788906812668,
-0.03904176875948906,
-0.012435059994459152,
-0.031150242313742638,
-0.011157393455505371,
0.08781024068593979,
-0.09729834645986557,
0.04240095987915993,
-0.15245147049427032,
-0.14401228725910187,
0.03192957490682602,
0.0642581358551979,
0.011705230921506882,
-0.08818046003580093,
0.036137018352746964,
-0.005530418362468481,
0.03979998826980591,
-0.09016831964254379,
-0.02225935459136963,
-0.08854800462722778,
0.0716094896197319,
-0.09036583453416824,
0.01928316429257393,
-0.17267097532749176,
0.06887461990118027,
-0.11407924443483353,
-0.017700498923659325,
-0.05338679254055023,
-0.0243575070053339,
-0.09160128980875015,
0.18350261449813843,
-0.04107033088803291,
-0.04106784239411354,
0.012812543660402298,
0.008747126907110214,
-0.05867214873433113,
0.11271152645349503,
-0.07905924320220947,
-0.06208209693431854,
0.193105086684227,
-0.11312826722860336,
-0.18387918174266815,
0.06954911351203918,
0.005549502093344927,
-0.0009717301581986248,
0.07701529562473297,
0.162315234541893,
0.09539500623941422,
-0.01430345606058836,
0.08831996470689774,
0.14527755975723267,
-0.06741485744714737,
-0.1899579018354416,
0.039836782962083817,
-0.0518929660320282,
-0.1316661685705185,
0.06082112714648247,
-0.023314109072089195,
0.08668104559183121,
0.005240552127361298,
-0.07265958935022354,
-0.054311603307724,
-0.04594247043132782,
0.03554585576057434,
0.015470363199710846,
0.07424456626176834,
-0.0780770555138588,
-0.005386367905884981,
-0.02062210813164711,
0.026860764250159264,
0.04755967855453491,
0.05645165964961052,
-0.0616346038877964,
0.06374672800302505,
0.038931526243686676,
0.03387635946273804,
-0.1492423564195633,
-0.0026758909225463867,
-0.006619501393288374,
0.0394667387008667,
0.008541746996343136,
0.07236725091934204,
0.034056663513183594,
-0.07543224841356277,
-0.007346620317548513,
-0.00990211684256792,
0.13999104499816895,
0.04745945334434509,
-0.027293691411614418,
-0.11992690712213516,
0.051172077655792236,
-0.03158222511410713,
0.032270461320877075,
-0.012440373189747334,
0.0255413930863142,
0.05202852562069893,
0.0896928682923317,
-0.04332510009407997,
0.10814680904150009,
-0.0314299538731575,
0.008204556070268154,
-0.06421206146478653,
0.005224584136158228,
0.11843527108430862,
0.04398849606513977,
-0.08439838886260986,
0.19714058935642242,
-0.07091148942708969,
0.26326096057891846,
0.21244864165782928,
-0.19821707904338837,
0.06781923025846481,
-0.020836390554904938,
-0.01435175072401762,
-0.0025727238971740007,
0.030823349952697754,
0.03073660098016262,
0.04172210767865181,
0.01771463453769684,
0.17754846811294556,
-0.04812063276767731,
-0.04125499352812767,
-0.02247837372124195,
-0.06345567107200623,
-0.007959081791341305,
0.05230538174510002,
0.1404448002576828,
-0.18668782711029053,
0.1913488507270813,
0.31166794896125793,
-0.01400954183191061,
0.08600834012031555,
-0.07593902945518494,
0.024557704105973244,
0.06549926102161407,
-0.04413771256804466,
-0.022823473438620567,
-0.03249483183026314,
-0.1398867517709732,
0.004700822290033102,
0.10399795323610306,
0.039237093180418015,
0.05066521093249321,
-0.13973669707775116,
-0.048357754945755005,
0.003572310321033001,
-0.01914861798286438,
-0.04409148171544075,
0.05693091079592705,
0.01529010757803917,
0.09553027153015137,
-0.029352953657507896,
-0.1036546602845192,
0.14165401458740234,
0.0015804258873686194,
-0.08185664564371109,
0.14734046161174774,
-0.1686059981584549,
-0.2672240436077118,
-0.15006466209888458,
-0.15754730999469757,
-0.040469471365213394,
0.021722793579101562,
0.14752502739429474,
-0.05383254215121269,
-0.05828969553112984,
0.010263088159263134,
-0.0916106328368187,
-0.010198148898780346,
0.02441413700580597,
-0.03552008420228958,
0.058104515075683594,
0.005071738734841347,
-0.11582627147436142,
-0.06767059117555618,
0.013409256935119629,
-0.03647346422076225,
0.10347112268209457,
-0.09030836820602417,
0.07127439975738525,
0.12871627509593964,
0.01243792474269867,
0.03326582908630371,
-0.03339415043592453,
0.13468076288700104,
-0.037399765104055405,
-0.0030397437512874603,
0.22738684713840485,
-0.04533824697136879,
0.08281181007623672,
0.1512300968170166,
0.036025270819664,
-0.04110533371567726,
0.020001502707600594,
-0.05199878290295601,
-0.07664331048727036,
-0.2774033844470978,
-0.12388425320386887,
-0.10269982367753983,
0.04752686247229576,
0.058578431606292725,
0.09034254401922226,
0.14106637239456177,
0.07592342048883438,
-0.019996946677565575,
0.015969308093190193,
0.007639035116881132,
0.07632344216108322,
0.27253058552742004,
-0.004745368380099535,
0.12732580304145813,
-0.11762017011642456,
-0.057013943791389465,
0.1265517622232437,
0.062044817954301834,
0.1236846074461937,
0.1234099343419075,
0.051585108041763306,
0.06550563126802444,
0.14724965393543243,
0.10624245554208755,
0.12493196874856949,
0.028105320408940315,
0.004841153975576162,
-0.04727954789996147,
-0.011976541019976139,
-0.059970151633024216,
0.033797647804021835,
-0.022963635623455048,
-0.14457081258296967,
-0.042962756007909775,
-0.1554267853498459,
0.07811342179775238,
0.18483006954193115,
0.015438108704984188,
-0.13616302609443665,
0.026686720550060272,
0.08427965641021729,
-0.020005637779831886,
-0.05305152013897896,
0.09713099151849747,
-0.10846424102783203,
-0.11656232923269272,
0.13418035209178925,
-0.025590287521481514,
0.14492498338222504,
-0.0248755794018507,
0.053225770592689514,
-0.010453574359416962,
-0.13999760150909424,
0.06725247949361801,
0.15039514005184174,
-0.29247668385505676,
0.20505018532276154,
-0.010331127792596817,
-0.04482366144657135,
-0.08720868080854416,
0.0013150222366675735,
0.07004573196172714,
0.25857123732566833,
0.07095246762037277,
0.01650865562260151,
-0.10063847154378891,
-0.0516783781349659,
-0.06302125006914139,
0.02917596511542797,
0.029199955984950066,
-0.011154877953231335,
-0.05557602643966675,
-0.0649041160941124,
-0.021162277087569237,
-0.0042198156006634235,
0.03514545410871506,
-0.020032672211527824,
-0.16620703041553497,
0.05486385524272919,
0.10046247392892838,
0.049888525158166885,
-0.051651325076818466,
-0.027695588767528534,
-0.11299943178892136,
0.1882910132408142,
-0.10712026804685593,
-0.08979795128107071,
-0.09791862219572067,
-0.12910650670528412,
0.024837469682097435,
-0.06235082820057869,
0.0702935978770256,
-0.08618972450494766,
0.0010052000870928168,
-0.05893877521157265,
-0.20419736206531525,
0.10277000069618225,
-0.12839920818805695,
-0.0499797947704792,
-0.043673690408468246,
0.12797610461711884,
-0.1115511879324913,
0.02084740810096264,
0.040034614503383636,
0.003994327504187822,
-0.1022302433848381,
-0.12035572528839111,
-0.022925307974219322,
0.03245508298277855,
0.059485625475645065,
-0.026264069601893425,
-0.101990707218647,
-0.0571921281516552,
0.004007372539490461,
-0.041260719299316406,
0.2471790313720703,
0.1690734475851059,
-0.08446425944566727,
0.18276453018188477,
0.18056119978427887,
-0.09941285848617554,
-0.31816133856773376,
-0.1598256230354309,
-0.13836894929409027,
-0.10192114859819412,
-0.027664819732308388,
-0.15457311272621155,
0.13434267044067383,
0.04874061420559883,
-0.08777574449777603,
0.08117468655109406,
-0.15966583788394928,
-0.08619657903909683,
0.21040642261505127,
-0.04534730687737465,
0.3174908459186554,
-0.13340862095355988,
-0.08690694719552994,
-0.09293768554925919,
-0.1719014197587967,
0.14943470060825348,
-0.07211921364068985,
0.06140616536140442,
-0.012918728403747082,
0.027143200859427452,
-0.008217151276767254,
-0.04619397595524788,
0.1133522018790245,
-0.014066650532186031,
0.024132272228598595,
-0.1293950229883194,
-0.0024696949403733015,
0.06233421340584755,
-0.02529490925371647,
0.042755890637636185,
-0.13924568891525269,
0.023840680718421936,
-0.11120016127824783,
-0.026454707607626915,
-0.061154644936323166,
0.07823684066534042,
0.00931283738464117,
-0.0443255715072155,
-0.026689087972044945,
-0.03375519812107086,
0.0254792720079422,
-0.009869445115327835,
0.25119978189468384,
0.015602859668433666,
0.07951269298791885,
0.12697459757328033,
0.09628882259130478,
-0.18091686069965363,
-0.004001361317932606,
-0.11060860753059387,
-0.08483202010393143,
0.059636712074279785,
-0.14534153044223785,
0.05873894691467285,
0.10457886010408401,
-0.07313844561576843,
0.05897313356399536,
0.08260530233383179,
0.025774210691452026,
-0.05598508194088936,
0.13984863460063934,
-0.17353367805480957,
0.025570666417479515,
-0.01700710318982601,
0.10868968814611435,
0.06578559428453445,
0.06253332644701004,
0.12149979919195175,
0.02378910779953003,
-0.045509472489356995,
0.02385631762444973,
0.026733553037047386,
-0.05165635421872139,
0.03943267837166786,
0.05652942880988121,
0.007127806544303894,
-0.135700523853302,
0.10076140612363815,
0.03949320688843727,
-0.139295756816864,
-0.02660050429403782,
0.1049129068851471,
-0.15733122825622559,
-0.13474465906620026,
0.02149812877178192,
0.07954824715852737,
-0.14437325298786163,
-0.11576265841722488,
-0.04053666815161705,
-0.14157921075820923,
0.0704789087176323,
0.1180836483836174,
0.12505967915058136,
0.07465346157550812,
-0.01606745831668377,
-0.07044725865125656,
0.04563111439347267,
-0.0016263002762570977,
-0.06975575536489487,
0.033102910965681076,
-0.10268332809209824,
-0.04989679530262947,
0.014775268733501434,
0.10792622715234756,
-0.052679866552352905,
-0.017179174348711967,
-0.09913178533315659,
0.03468547388911247,
-0.1543911248445511,
0.004061874467879534,
-0.073458231985569,
-0.006152780260890722,
0.019225608557462692,
-0.07365509122610092,
-0.02818438597023487,
-0.0066046081483364105,
-0.12461689859628677,
-0.01529979333281517,
-0.03191344812512398,
0.07186026871204376,
-0.11214398592710495,
-0.0533626489341259,
0.09820953011512756,
-0.01407189667224884,
0.10279780626296997,
0.07704830169677734,
-0.07897642254829407,
0.08706441521644592,
-0.15200023353099823,
-0.10375725477933884,
0.09744801372289658,
0.05007246136665344,
0.03702830895781517,
0.005307680461555719,
0.015036889351904392,
0.12421702593564987,
-0.033559706062078476,
0.045524436980485916,
0.04527132213115692,
-0.15067242085933685,
-0.025562534108757973,
-0.01080042403191328,
-0.13799738883972168,
-0.008309535682201385,
-0.09745097905397415,
0.13043686747550964,
0.023049110546708107,
0.19978578388690948,
-0.025248127058148384,
0.04954942688345909,
-0.04377560690045357,
0.017911331728100777,
-0.03635789826512337,
-0.1728820949792862,
-0.13917817175388336,
-0.07161751389503479,
-0.04154269024729729,
-0.012728624045848846,
0.2695949673652649,
0.041973695158958435,
-0.03874577581882477,
0.0755695179104805,
0.08605275303125381,
0.017459535971283913,
0.013623190112411976,
0.24184226989746094,
0.05891833081841469,
-0.006889847572892904,
-0.07207736372947693,
0.0095368018373847,
0.01267328392714262,
-0.07895752787590027,
0.09543026238679886,
0.09202968329191208,
0.04860799387097359,
0.062353868037462234,
0.02349235676229,
0.03302542492747307,
-0.13147249817848206,
-0.16768448054790497,
0.017875514924526215,
0.09444275498390198,
0.01440428476780653,
0.11145492643117905,
0.11040886491537094,
-0.039926040917634964,
0.02716808021068573,
-0.049468379467725754,
-0.012980833649635315,
-0.18488742411136627,
-0.12038124352693558,
-0.08426732569932938,
-0.12213902920484543,
0.003872788278385997,
-0.04934149980545044,
0.01755918748676777,
0.09066841751337051,
0.04828392341732979,
-0.067583829164505,
-0.012107289396226406,
-0.04697847366333008,
-0.04348131641745567,
0.04423016682267189,
-0.03289409726858139,
-0.011101330630481243,
-0.024587200954556465,
-0.04467953369021416,
-0.10911267250776291,
-0.0322733148932457,
-0.04883580282330513,
0.04706401005387306,
-0.008699473924934864,
0.03918442130088806,
-0.11204695701599121,
-0.0767931342124939,
-0.04038625955581665,
0.029535427689552307,
-0.0040898569859564304,
0.1869017481803894,
0.009583257138729095,
0.038710206747055054,
0.09553047269582748,
0.16753792762756348,
-0.07017162442207336,
-0.15256629884243011,
-0.047502581030130386,
0.20098082721233368,
0.05657508969306946,
0.06080913171172142,
0.012499228119850159,
0.011208959855139256,
-0.056554656475782394,
0.3228624165058136,
0.28597143292427063,
-0.04547877982258797,
0.036143917590379715,
-0.018416233360767365,
0.021603884175419807,
0.10241707414388657,
0.14358292520046234,
0.127496138215065,
0.1839708834886551,
-0.06740538030862808,
-0.026008620858192444,
-0.03907419368624687,
0.0018298098584637046,
-0.17193500697612762,
0.08618982881307602,
-0.01088035199791193,
-0.09144339710474014,
-0.022995738312602043,
0.10249924659729004,
-0.11244610697031021,
0.115500807762146,
-0.0033514192327857018,
-0.11603543907403946,
-0.043959882110357285,
-0.0058623687364161015,
0.1857547014951706,
0.007513022515922785,
0.02933778427541256,
-0.017118019983172417,
-0.06501377373933792,
0.10304711014032364,
-0.011885383166372776,
-0.21393372118473053,
-0.04481937363743782,
0.08612168580293655,
-0.037672173231840134,
0.11618560552597046,
0.011411667801439762,
0.057896751910448074,
0.07590916007757187,
0.08023592829704285,
-0.08374308794736862,
0.08137734979391098,
0.011783070862293243,
-0.049928218126297,
0.014140583574771881,
-0.10634700208902359,
-0.023432763293385506,
-0.08135902136564255,
0.04243294522166252,
-0.08825377374887466,
0.047646839171648026,
-0.0020005095284432173,
-0.0684451311826706,
-0.04261155426502228,
0.0507601760327816,
-0.08038213849067688,
0.056206125766038895,
0.03303709998726845,
-0.028430424630641937,
-0.06526947766542435,
-0.06521723419427872,
-0.031366121023893356,
0.01531433779746294,
-0.19917620718479156,
-0.09116730839014053,
-0.0053079086355865,
-0.03789437934756279,
0.09222039580345154,
0.04602926969528198,
-0.09527811408042908,
-0.016809822991490364,
-0.09963963180780411,
0.025631384924054146,
-0.17426390945911407,
0.047836631536483765,
0.06116890534758568,
-0.003921943251043558,
-0.006005494389683008,
-0.05374446138739586,
0.02516728825867176,
0.016484463587403297,
-0.0931040346622467,
-0.08561339229345322
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-MiniLM2-L6-H768')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-MiniLM2-L6-H768')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-MiniLM2-L6-H768')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-MiniLM2-L6-H768')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["MiniLMv2"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-MiniLM2-L6-H768
|
[
"transformers",
"pytorch",
"roberta",
"text-classification",
"MiniLMv2",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #roberta #text-classification #MiniLMv2 #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #roberta #text-classification #MiniLMv2 #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
74,
30,
49,
18,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #roberta #text-classification #MiniLMv2 #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.02044294774532318,
0.15575215220451355,
-0.005729633383452892,
0.10251614451408386,
0.12102481722831726,
0.0196793582290411,
0.24352902173995972,
0.08777108788490295,
-0.061177175492048264,
-0.0498560331761837,
0.027248933911323547,
0.21822427213191986,
0.0010040567722171545,
0.04355757310986519,
-0.03316637873649597,
-0.21052862703800201,
0.0374772734940052,
0.0011222735047340393,
0.035875651985406876,
0.10555332899093628,
0.1079602912068367,
-0.05825333669781685,
0.02936483733355999,
-0.0074266353622078896,
-0.1471531242132187,
0.01978910155594349,
0.014657000079751015,
-0.05597163364291191,
0.09088216722011566,
0.03162437304854393,
0.07640507817268372,
0.044130727648735046,
0.0359363928437233,
-0.11705342680215836,
0.008199307136237621,
0.04898254573345184,
-0.004913500510156155,
0.0370866060256958,
-0.00022700687986798584,
-0.0944342240691185,
0.04685598239302635,
-0.05163112282752991,
0.06335128843784332,
0.06336657702922821,
-0.06337660551071167,
-0.08779341727495193,
0.01903083175420761,
0.006823698990046978,
0.0888151302933693,
0.0916251465678215,
-0.027888722717761993,
0.176200270652771,
-0.14434005320072174,
0.09549557417631149,
0.06898857653141022,
-0.22526176273822784,
-0.029416201636195183,
0.13454563915729523,
-0.017796028405427933,
0.07568223029375076,
-0.04063544049859047,
0.001407303148880601,
-0.005642368923872709,
0.026956800371408463,
0.06969509273767471,
-0.050709933042526245,
-0.08801715075969696,
-0.019106287509202957,
-0.10946095734834671,
0.026862667873501778,
0.18969419598579407,
-0.06000250205397606,
-0.04209117963910103,
-0.11292178183794022,
-0.028622079640626907,
-0.04354577511548996,
-0.038097579032182693,
-0.013464096933603287,
0.017812706530094147,
0.04009467735886574,
-0.04261704161763191,
-0.007979943417012691,
-0.08131571859121323,
-0.0506574884057045,
-0.08506361395120621,
0.142503559589386,
0.0439772866666317,
0.02919222228229046,
-0.14018838107585907,
0.06360571086406708,
0.014296100474894047,
-0.10294913500547409,
-0.026166347786784172,
-0.0562506727874279,
-0.13996000587940216,
0.013341892510652542,
-0.13175155222415924,
-0.11357545852661133,
0.08036861568689346,
0.14524029195308685,
0.033483244478702545,
0.057364143431186676,
0.022740498185157776,
0.06059505417943001,
0.07478686422109604,
0.07375437766313553,
-0.01647154428064823,
-0.005091042723506689,
0.011495700106024742,
-0.018717771396040916,
0.027981381863355637,
-0.05632516369223595,
-0.12219145894050598,
0.012404533103108406,
0.05071580037474632,
0.044075433164834976,
0.004659691359847784,
0.06904308497905731,
-0.05017746612429619,
-0.07942458987236023,
-0.017594285309314728,
-0.08466748893260956,
0.012220519594848156,
0.0021516575943678617,
0.008066737093031406,
0.20753417909145355,
-0.0032892751041799784,
-0.001783298677764833,
-0.0777028277516365,
-0.02729785069823265,
-0.08844969421625137,
0.029972929507493973,
-0.09422492980957031,
-0.09456250071525574,
-0.005738576874136925,
0.017432570457458496,
-0.03443697839975357,
-0.10961238294839859,
-0.10451719909906387,
0.006420790683478117,
0.04920955002307892,
-0.047186944633722305,
-0.03329421207308769,
-0.10712062567472458,
-0.025519581511616707,
0.012389824725687504,
0.0024639535695314407,
-0.0523810014128685,
-0.010154986754059792,
0.016318874433636665,
0.009967450983822346,
0.06681060791015625,
-0.07963141053915024,
-0.0034459151793271303,
-0.1686994731426239,
-0.03197946771979332,
0.0888838991522789,
0.14580626785755157,
-0.026265665888786316,
-0.0024822617415338755,
-0.0786450132727623,
-0.046546462923288345,
0.029771046712994576,
0.022189827635884285,
-0.022078340873122215,
0.1154167577624321,
-0.2562636435031891,
-0.06268885731697083,
0.18092875182628632,
-0.16840341687202454,
-0.04971366748213768,
0.11542101204395294,
-0.040555015206336975,
-0.0159752257168293,
0.11004742234945297,
0.05568857863545418,
0.1427459865808487,
-0.013338529504835606,
0.020268959924578667,
0.06892896443605423,
-0.05401294305920601,
0.020948922261595726,
0.07549042999744415,
0.07451295107603073,
-0.07352538406848907,
0.06212400272488594,
-0.042508386075496674,
-0.0092052286490798,
-0.028451910242438316,
-0.07829519361257553,
-0.041404034942388535,
-0.02517317421734333,
0.1021743044257164,
0.051869772374629974,
0.05929679423570633,
-0.04298311099410057,
-0.09496255218982697,
0.1357431411743164,
0.0936148390173912,
-0.10954777151346207,
0.036424580961465836,
-0.10238581895828247,
-0.04425984248518944,
-0.026203801855444908,
-0.00053686264436692,
-0.1986764818429947,
-0.04732391983270645,
0.019514815881848335,
-0.020605089142918587,
0.11984080821275711,
0.07286829501390457,
0.004813093226402998,
0.05050431936979294,
-0.013205836527049541,
0.0251864455640316,
0.04538772255182266,
0.019697532057762146,
-0.03238143026828766,
-0.15861572325229645,
-0.05194970220327377,
-0.022638393566012383,
0.10143778473138809,
-0.11849323660135269,
0.033251602202653885,
0.003522925078868866,
0.022640027105808258,
-0.016404464840888977,
-0.004675028845667839,
-0.027870770543813705,
0.08268404006958008,
-0.060549378395080566,
-0.011601543985307217,
0.10490240901708603,
0.03367161750793457,
-0.09695181250572205,
0.015321983024477959,
-0.1289159059524536,
-0.13752931356430054,
0.08735805749893188,
-0.10627978295087814,
-0.04266037419438362,
-0.02050606906414032,
-0.004458249546587467,
0.0010316065745428205,
-0.04309717193245888,
0.0142204063013196,
0.12857533991336823,
0.044687509536743164,
0.08479636907577515,
-0.0655670315027237,
-0.03472926840186119,
-0.016005247831344604,
-0.1163235604763031,
0.027767043560743332,
0.07834739983081818,
0.007977491244673729,
-0.2155103236436844,
0.12728577852249146,
0.016468778252601624,
-0.145059734582901,
0.12234603613615036,
0.0013666314771398902,
0.019802941009402275,
-0.05205942317843437,
-0.06094863638281822,
0.02666199766099453,
-0.015601716935634613,
-0.06755883246660233,
0.024576008319854736,
0.05809354409575462,
0.03223351016640663,
0.010222329758107662,
-0.10010291635990143,
0.021383119747042656,
0.02337442710995674,
-0.003969837911427021,
0.0544491671025753,
0.021903283894062042,
0.004492407664656639,
0.07696536183357239,
-0.02368658222258091,
0.02025931514799595,
0.06709286570549011,
-0.03731979802250862,
-0.1081533432006836,
0.21347522735595703,
-0.11439304053783417,
-0.2484542280435562,
-0.09996219724416733,
0.007433710619807243,
-0.11364053934812546,
0.013527089729905128,
0.03378510847687721,
-0.09448453783988953,
-0.008910873904824257,
-0.07073260098695755,
-0.05484704673290253,
-0.0688294991850853,
-0.03356403857469559,
-0.09563419967889786,
0.05820997431874275,
0.014633937738835812,
-0.1319122165441513,
0.003746438305824995,
0.01911775954067707,
-0.14036278426647186,
0.06801030784845352,
-0.025866447016596794,
-0.02374478429555893,
0.20698107779026031,
-0.006399189122021198,
0.023344149813055992,
-0.007683551870286465,
0.16975949704647064,
0.0035672851372510195,
-0.018430767580866814,
0.19002118706703186,
0.016030965372920036,
0.047893352806568146,
0.08078493177890778,
0.050072427839040756,
-0.037740107625722885,
0.025159437209367752,
0.02966536581516266,
-0.08309731632471085,
-0.12931333482265472,
-0.1112164705991745,
-0.07011828571557999,
-0.038261134177446365,
0.06872250139713287,
-0.0005302509525790811,
0.08435732871294022,
0.08520771563053131,
0.011725612916052341,
0.012095702812075615,
0.05306835472583771,
0.10781484842300415,
0.12893790006637573,
0.07121352851390839,
0.15692542493343353,
-0.047855451703071594,
-0.0684313252568245,
0.03605961427092552,
0.00816289521753788,
0.14798010885715485,
0.024254849180579185,
0.08398832380771637,
0.13952401280403137,
-0.02514726296067238,
0.06087009981274605,
0.025777742266654968,
-0.006576240062713623,
0.014605501666665077,
-0.0035993834026157856,
-0.07545624673366547,
-0.03985719382762909,
0.008608873002231121,
-0.04186517372727394,
0.01744694449007511,
-0.028046291321516037,
-0.011955139227211475,
0.10657468438148499,
0.1183527484536171,
0.08778827637434006,
-0.2335188090801239,
-0.061750371009111404,
0.05188007280230522,
-0.00852731429040432,
0.011962666176259518,
0.03651585057377815,
0.03392406553030014,
-0.10650632530450821,
0.05428740754723549,
-0.024489751085639,
0.08219022303819656,
-0.09053569287061691,
0.02439223602414131,
-0.03778831660747528,
0.07468104362487793,
0.06307021528482437,
0.12548880279064178,
-0.3033726215362549,
0.06063929572701454,
-0.020806297659873962,
0.05622200667858124,
-0.08107896149158478,
0.0028053312562406063,
0.015252925455570221,
0.07426614314317703,
0.08526664972305298,
0.01456899382174015,
-0.0823117196559906,
-0.045259974896907806,
-0.022170212119817734,
0.026017745956778526,
0.05752839148044586,
-0.0417073555290699,
0.08663494884967804,
-0.044031962752342224,
0.012591813690960407,
-0.021412985399365425,
0.10096929967403412,
-0.13842929899692535,
-0.18496187031269073,
0.05729715898633003,
0.07190593332052231,
-0.06036533787846565,
-0.00896046869456768,
-0.06729908287525177,
-0.0046204193495213985,
0.18266166746616364,
-0.16754329204559326,
-0.10630062222480774,
-0.11823683977127075,
0.043637026101350784,
0.03401126340031624,
-0.06585201621055603,
-0.027233652770519257,
-0.024582572281360626,
0.15820975601673126,
-0.025506220757961273,
-0.17854087054729462,
0.005970880389213562,
-0.05736926943063736,
-0.10791557282209396,
-0.02273452840745449,
0.05576005578041077,
0.09391968697309494,
0.04176970571279526,
0.04560280218720436,
0.03514290973544121,
-0.005637523718178272,
-0.13116592168807983,
-0.09910948574542999,
0.05187121406197548,
0.06251195818185806,
0.048435330390930176,
-0.04185543209314346,
-0.13902944326400757,
-0.03716715797781944,
0.03413102403283119,
0.20758028328418732,
0.15466174483299255,
-0.06234185770153999,
0.09755624085664749,
0.10137150436639786,
-0.13250549137592316,
-0.20437651872634888,
-0.10882659256458282,
0.05001349374651909,
-0.0022695756051689386,
0.05168755725026131,
-0.1302715539932251,
0.10319656878709793,
0.083262600004673,
-0.01484473142772913,
-0.11818551272153854,
-0.36828625202178955,
-0.09672967344522476,
0.06553153693675995,
0.11521992832422256,
0.1520194113254547,
-0.10205123573541641,
-0.005296474788337946,
-0.07552354037761688,
-0.044183917343616486,
0.35867756605148315,
-0.09099519997835159,
0.10859952867031097,
-0.010641813278198242,
-0.030311493203043938,
0.02501179277896881,
-0.02564670890569687,
0.0802052766084671,
0.03555964305996895,
0.0896834284067154,
-0.010414876975119114,
-0.07110468298196793,
0.11722871661186218,
-0.015368213877081871,
0.05958518385887146,
-0.10305779427289963,
0.05555308610200882,
-0.021252896636724472,
-0.04688797891139984,
-0.06845952570438385,
-0.0042302715592086315,
-0.050310663878917694,
-0.06645714491605759,
-0.06168019026517868,
0.08734352886676788,
0.062900610268116,
-0.017363497987389565,
0.17581401765346527,
-0.10376202315092087,
0.07722096145153046,
0.18952016532421112,
0.16043440997600555,
0.06037820130586624,
-0.10005547106266022,
-0.05394621193408966,
-0.01869809627532959,
0.09961840510368347,
-0.11019805818796158,
0.09213057905435562,
0.05014769732952118,
0.0008330723503604531,
0.10419861227273941,
0.10926486551761627,
-0.00010683351138141006,
0.012845735065639019,
0.06380632519721985,
-0.07594796270132065,
-0.017352456226944923,
0.0023655621334910393,
0.06880790740251541,
0.00939999707043171,
0.06447158753871918,
0.15127430856227875,
-0.05777039751410484,
-0.039313800632953644,
0.011953848414123058,
0.018955323845148087,
-0.06653337180614471,
0.09266457706689835,
0.014808623120188713,
0.06856025755405426,
-0.09951932728290558,
0.08257970213890076,
0.08482392877340317,
0.05402697995305061,
0.044281087815761566,
0.09554260969161987,
-0.0932822898030281,
-0.10082618147134781,
0.039596740156412125,
0.061615992337465286,
-0.04861472174525261,
-0.0343339629471302,
-0.07771088927984238,
-0.10094257444143295,
-0.021274495869874954,
0.04030368849635124,
0.1061166450381279,
0.07835422456264496,
-0.11980900168418884,
-0.05509219691157341,
-0.10747860372066498,
0.046549104154109955,
-0.018050633370876312,
0.014276295900344849,
-0.06947217136621475,
0.14072272181510925,
0.023404918611049652,
-0.001454650075174868,
-0.06352926790714264,
-0.08115922659635544,
-0.15791098773479462,
0.026657739654183388,
-0.04205377772450447,
0.023753900080919266,
-0.10338426381349564,
0.0004739674914162606,
0.023971648886799812,
0.022347131744027138,
-0.00845251139253378,
0.022289467975497246,
-0.02622738853096962,
-0.007509141229093075,
-0.006241797469556332,
0.08058474957942963,
-0.05678468570113182,
-0.005870653782039881,
0.008782824501395226,
-0.061331361532211304,
0.037134040147066116,
0.011782398447394371,
-0.08965231478214264,
0.06462793797254562,
-0.14839473366737366,
-0.008881854824721813,
-0.003317304654046893,
0.07068034261465073,
0.010659690015017986,
-0.08165115863084793,
0.028718244284391403,
0.06622877717018127,
0.05330555886030197,
0.040457505732774734,
0.09237898886203766,
-0.09132218360900879,
-0.011976861394941807,
-0.03474588692188263,
-0.059702664613723755,
-0.06024169921875,
0.0002460430550854653,
0.06025345250964165,
0.04950081557035446,
0.19933414459228516,
-0.08655760437250137,
0.06564944982528687,
-0.11729869991540909,
-0.03846811130642891,
0.03819186985492706,
-0.05719383805990219,
-0.09395851194858551,
-0.09235701709985733,
0.023392481729388237,
-0.02268671616911888,
0.1527499556541443,
0.002475373214110732,
0.05926771089434624,
0.010905974544584751,
0.028507882729172707,
0.06112102046608925,
0.04705872759222984,
0.18926985561847687,
0.04899565875530243,
-0.0006252169259823859,
0.03444909304380417,
0.0069234031252563,
-0.00851255189627409,
-0.005979177076369524,
0.11397765576839447,
0.04995276778936386,
-0.09392622858285904,
0.06481017917394638,
0.004640630912035704,
0.07064500451087952,
-0.04637962952256203,
0.060036927461624146,
0.014667286537587643,
0.03148027881979942,
-0.04418458044528961,
0.047379978001117706,
0.14499598741531372,
-0.09075529873371124,
0.10548415780067444,
0.02095014415681362,
-0.10149512439966202,
-0.14682315289974213,
-0.2836573123931885,
-0.07666631042957306,
-0.07261471450328827,
-0.029271677136421204,
-0.1452789604663849,
-0.009410695172846317,
0.10793043673038483,
0.051152389496564865,
-0.034224238246679306,
0.01657513529062271,
0.014843174256384373,
-0.028330761939287186,
-0.0035902399104088545,
-0.05603792145848274,
0.03995335474610329,
0.011219297535717487,
0.05928570777177811,
0.04003645479679108,
0.04911723732948303,
0.06697992980480194,
0.036973632872104645,
0.0662851557135582,
-0.019685739651322365,
-0.07325179129838943,
-0.10154934972524643,
0.02310566045343876,
-0.015174109488725662,
0.018684856593608856,
0.10841603577136993,
0.060141220688819885,
-0.08344266563653946,
-0.010309474542737007,
0.17818959057331085,
-0.06457918137311935,
-0.19236153364181519,
-0.13457652926445007,
0.3051201105117798,
0.018925899639725685,
0.053329791873693466,
0.020012937486171722,
-0.04247530922293663,
-0.05543932318687439,
0.21223171055316925,
0.13816148042678833,
-0.04326262325048447,
0.003548911539837718,
0.04428713023662567,
-0.025400912389159203,
0.0012141154147684574,
0.12842924892902374,
-0.049275193363428116,
0.2147340476512909,
-0.04683703929185867,
0.11148975044488907,
-0.05481105297803879,
-0.025474337860941887,
-0.09800951182842255,
0.07831015437841415,
-0.023340299725532532,
-0.026296326890587807,
-0.024348052218556404,
0.07491261512041092,
-0.12954653799533844,
0.11372479051351547,
0.10086604952812195,
-0.016529874876141548,
-0.11656635254621506,
-0.0038510828744620085,
0.05810235068202019,
-0.04559818655252457,
0.06644906103610992,
-0.02578437328338623,
0.020790429785847664,
0.0480940118432045,
-0.012924017384648323,
-0.02156239189207554,
-0.04710719361901283,
0.02637610211968422,
-0.031752072274684906,
0.09451976418495178,
-0.014266686514019966,
0.149828240275383,
0.10387542098760605,
0.008918853476643562,
-0.06441456079483032,
0.11729678511619568,
-0.05231069400906563,
-0.03512819483876228,
0.09054436534643173,
0.0698077604174614,
-0.03000200167298317,
0.09521470963954926,
0.0763562023639679,
-0.13672411441802979,
0.02393573336303234,
-0.006911805365234613,
-0.05196363106369972,
-0.03412194177508354,
0.009850500151515007,
-0.05475502088665962,
0.0817679911851883,
0.11484018713235855,
-0.049366556107997894,
-0.031070103868842125,
-0.048949237912893295,
0.08188489079475403,
-0.02629435621201992,
-0.030747873708605766,
-0.07633061707019806,
-0.1555677205324173,
-0.010053691454231739,
0.07328963279724121,
0.040582116693258286,
-0.17272479832172394,
-0.013014145195484161,
0.013647903688251972,
-0.03834797814488411,
-0.06634438782930374,
0.04233178123831749,
0.05373592674732208,
0.018995555117726326,
-0.06601491570472717,
-0.22447992861270905,
0.00313325016759336,
0.07953866571187973,
-0.12006951123476028,
-0.1079668253660202
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-deberta-base')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-base')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-base')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-base')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["deberta-base-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-deberta-base
|
[
"transformers",
"pytorch",
"deberta",
"text-classification",
"deberta-base-base",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #deberta #text-classification #deberta-base-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #deberta #text-classification #deberta-base-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
82,
30,
49,
18,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #deberta #text-classification #deberta-base-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.017919236794114113,
0.17484505474567413,
-0.006778021343052387,
0.09020459651947021,
0.1217234879732132,
0.022064127027988434,
0.2579192519187927,
0.0706629753112793,
-0.02560870721936226,
-0.026418372988700867,
0.03061394952237606,
0.2159518152475357,
-0.0013462663628160954,
0.05589500442147255,
-0.03857579827308655,
-0.2438502162694931,
0.061350636184215546,
0.027842139825224876,
0.022447824478149414,
0.07621462643146515,
0.1191072091460228,
-0.0333460196852684,
0.026534980162978172,
-0.026621505618095398,
-0.11767103523015976,
0.06971481442451477,
0.006729434244334698,
-0.046108637005090714,
0.11075051873922348,
0.043982140719890594,
0.12971849739551544,
0.07993969321250916,
0.007482636719942093,
-0.16267713904380798,
0.0033050712663680315,
0.04660211503505707,
-0.04120033606886864,
0.018665317445993423,
-0.010308874770998955,
-0.10247897356748581,
-0.0069606006145477295,
-0.03359629586338997,
0.09808439016342163,
0.042958714067935944,
-0.09473854303359985,
-0.13059258460998535,
0.004744038451462984,
-0.009968029335141182,
0.08492931723594666,
0.09815505146980286,
-0.044237349182367325,
0.09669717401266098,
-0.12101228535175323,
0.09154734760522842,
0.06905952095985413,
-0.22493839263916016,
-0.008864817209541798,
0.0754816010594368,
0.007986187934875488,
0.0940021201968193,
-0.049159515649080276,
-0.008431563153862953,
-0.0072252945974469185,
0.04261929169297218,
0.07692787796258926,
-0.05560019612312317,
-0.029140371829271317,
-0.01132947951555252,
-0.11610542982816696,
0.03487422317266464,
0.16286525130271912,
-0.053832922130823135,
-0.05808834731578827,
-0.08846239000558853,
-0.05754591524600983,
-0.023027431219816208,
-0.012039309367537498,
-0.021956484764814377,
0.03269496187567711,
0.04693632572889328,
-0.03188761696219444,
-0.023934174329042435,
-0.09819990396499634,
-0.030339164659380913,
-0.09693356603384018,
0.1425454169511795,
0.052155353128910065,
0.01640072651207447,
-0.1095743328332901,
0.08313541114330292,
-0.006854149512946606,
-0.12270649522542953,
-0.03743729740381241,
-0.05163949355483055,
-0.15208719670772552,
0.010412071831524372,
-0.14689771831035614,
-0.1657213568687439,
0.0636163204908371,
0.22127321362495422,
-0.006838707253336906,
0.061513639986515045,
-0.03846382722258568,
0.04549495503306389,
0.07374730706214905,
0.1022772416472435,
-0.022257618606090546,
-0.01428849808871746,
0.009882316924631596,
0.00492192292585969,
0.05656255781650543,
-0.05912761762738228,
-0.13208015263080597,
-0.00998496450483799,
0.02265138551592827,
0.054991092532873154,
-0.0008808944839984179,
0.06360004097223282,
-0.03538055717945099,
-0.04901706799864769,
0.0227778609842062,
-0.0906081423163414,
0.0026485577691346407,
0.014134993776679039,
0.029431795701384544,
0.17693224549293518,
0.020862584933638573,
-0.004859711974859238,
-0.05569184198975563,
-0.009419133886694908,
-0.0977562889456749,
0.015554077923297882,
-0.08444737643003464,
-0.11969055980443954,
0.011663319543004036,
0.003652011277154088,
-0.021681562066078186,
-0.1281861960887909,
-0.10168514400720596,
0.004108457826077938,
0.03208392485976219,
-0.03909071907401085,
0.015188808552920818,
-0.1010080948472023,
0.0065541681833565235,
0.03226438909769058,
-0.025443706661462784,
-0.12368056178092957,
0.0002779632341116667,
0.02452290616929531,
0.04530460014939308,
0.07497736811637878,
-0.09423666447401047,
0.024494370445609093,
-0.15544532239437103,
-0.028563059866428375,
-0.01667734421789646,
0.14176183938980103,
-0.02910003252327442,
-0.019413335248827934,
-0.06585237383842468,
-0.007243653293699026,
0.058851540088653564,
0.045494481921195984,
-0.0012533236294984818,
0.09859539568424225,
-0.23281124234199524,
-0.07907988131046295,
0.17719946801662445,
-0.16429319977760315,
-0.0486748069524765,
0.12165110558271408,
-0.041907504200935364,
-0.010634295642375946,
0.1148764044046402,
0.0759708508849144,
0.11857113242149353,
-0.05272052064538002,
0.01327597163617611,
0.059210531413555145,
-0.013807910494506359,
0.08182888478040695,
0.05671772360801697,
0.06395997107028961,
-0.048123303800821304,
0.049450382590293884,
-0.07918567210435867,
0.025966141372919083,
0.003685257863253355,
-0.0503033809363842,
-0.010366802103817463,
-0.026216426864266396,
0.11877400428056717,
0.009352277033030987,
0.05625995621085167,
-0.016213785856962204,
-0.12251374870538712,
0.21402542293071747,
0.08897387236356735,
-0.1363316923379898,
0.058858033269643784,
-0.06846489757299423,
-0.05604909360408783,
-0.026184428483247757,
0.0015726321144029498,
-0.18267063796520233,
-0.09330670535564423,
0.03376764804124832,
-0.054645828902721405,
0.11028531938791275,
0.07214705646038055,
0.006064467132091522,
0.07033234089612961,
-0.05192290246486664,
-0.00424180319532752,
0.01720810867846012,
0.038721177726984024,
-0.036984633654356,
-0.1573333591222763,
-0.043651033192873,
-0.03378076106309891,
0.10459751635789871,
-0.083980031311512,
0.05748550593852997,
0.020577292889356613,
0.05778523534536362,
-0.018668560311198235,
0.006940379738807678,
-0.04644136503338814,
0.08093221485614777,
-0.07215810567140579,
0.0030081323347985744,
0.0480438768863678,
0.03287649527192116,
-0.12552571296691895,
0.05625665932893753,
-0.16400094330310822,
-0.08858004957437515,
0.08336731791496277,
-0.10597740858793259,
-0.05624821409583092,
0.005034140311181545,
0.013031069189310074,
-0.018484337255358696,
-0.05029313638806343,
-0.061814866960048676,
0.1293589025735855,
0.03966318815946579,
0.10980992019176483,
-0.05287345498800278,
-0.01964675262570381,
-0.0482243075966835,
-0.09251663833856583,
0.011578804813325405,
0.05432790145277977,
-0.040208250284194946,
-0.18161581456661224,
0.14689861238002777,
0.04261213168501854,
-0.18580032885074615,
0.1192583292722702,
0.011556273326277733,
-0.013267622329294682,
-0.07316488772630692,
-0.042044542729854584,
0.012686844915151596,
0.002280525164678693,
-0.07316605001688004,
0.023419227451086044,
0.0275421142578125,
0.04828672111034393,
0.017010854557156563,
-0.08768188953399658,
0.018734991550445557,
0.018892882391810417,
-0.010525209829211235,
-0.013075338676571846,
0.011025841347873211,
-0.000451444648206234,
0.06450878083705902,
-0.007773975841701031,
-0.011740980669856071,
0.08586942404508591,
-0.03331875056028366,
-0.11060875654220581,
0.21197614073753357,
-0.13663940131664276,
-0.29283758997917175,
-0.11560633778572083,
-0.01821517013013363,
-0.11588021367788315,
0.05138146132230759,
0.08542856574058533,
-0.0784367173910141,
-0.017021579667925835,
-0.055768195539712906,
0.03742534667253494,
-0.07697836309671402,
-0.02671816572546959,
-0.13119091093540192,
0.03968813270330429,
0.002618422033265233,
-0.13116998970508575,
0.011808295734226704,
0.02306325174868107,
-0.10027164965867996,
0.05156473070383072,
-0.03342854231595993,
-0.011255438439548016,
0.17475736141204834,
0.01947406306862831,
-0.007189058233052492,
-0.010381417348980904,
0.15660187602043152,
-0.012113670818507671,
0.013772371225059032,
0.1531538963317871,
-0.01804499886929989,
0.0691995620727539,
0.10988116264343262,
0.05142567306756973,
-0.03948983550071716,
0.032541390508413315,
0.0425097830593586,
-0.05227792263031006,
-0.18176481127738953,
-0.04530734568834305,
-0.06120879203081131,
-0.03588156774640083,
0.045826781541109085,
-0.002554042264819145,
0.1386830061674118,
0.0584985613822937,
-0.02608049474656582,
0.028696592897176743,
0.07444869726896286,
0.11012503504753113,
0.20868901908397675,
0.02215469628572464,
0.14364799857139587,
-0.05212293565273285,
-0.08216919004917145,
0.025713404640555382,
0.08192639797925949,
0.10736413300037384,
0.02556353434920311,
0.13540080189704895,
0.1181240975856781,
-0.0464794747531414,
0.047419823706150055,
0.029400385916233063,
0.0036782200913876295,
0.028799811378121376,
-0.011339631862938404,
-0.07278546690940857,
-0.0010116202756762505,
0.03521711379289627,
-0.02378656528890133,
-0.025403685867786407,
-0.038471147418022156,
-0.0721053034067154,
0.09753517806529999,
0.09115509688854218,
0.11036071181297302,
-0.2238849252462387,
-0.042538877576589584,
0.06578190624713898,
-0.024620208889245987,
-0.018519602715969086,
0.025245273485779762,
0.016439203172922134,
-0.07241172343492508,
0.08101996034383774,
-0.011768458411097527,
0.06556490063667297,
-0.1087847426533699,
0.022372644394636154,
-0.012932389974594116,
0.061735186725854874,
0.040821243077516556,
0.10025244206190109,
-0.3246529996395111,
0.03476742282509804,
-0.005999711807817221,
0.029654929414391518,
-0.03695162013173103,
0.03486808016896248,
-0.008343764580786228,
0.10219413787126541,
0.03985029458999634,
-0.008391914889216423,
-0.08303898572921753,
-0.06871290504932404,
0.009689846076071262,
-0.007260762155056,
0.07025502622127533,
-0.0629740059375763,
0.08817509561777115,
-0.010449778288602829,
0.0016933290753513575,
0.006975602358579636,
0.09686606377363205,
-0.14434391260147095,
-0.2131495624780655,
0.08673153072595596,
0.05021044984459877,
0.010274029336869717,
-0.019899999722838402,
-0.063311867415905,
0.02985469065606594,
0.1368086189031601,
-0.2048020213842392,
-0.11194822192192078,
-0.1205817312002182,
0.013885971158742905,
0.06674546748399734,
-0.0719052106142044,
-0.009029691107571125,
-0.01671578735113144,
0.14161917567253113,
-0.047132838517427444,
-0.15904663503170013,
0.048942841589450836,
-0.04190973564982414,
-0.08424278348684311,
-0.06921082735061646,
0.05643795430660248,
0.09686369448900223,
0.030242890119552612,
0.029346643015742302,
0.02473868802189827,
-0.0022532748989760876,
-0.10308665782213211,
-0.0937262773513794,
0.03814380615949631,
0.046387914568185806,
0.041833728551864624,
-0.054832275956869125,
-0.07265356183052063,
-0.05505887418985367,
0.034558236598968506,
0.1707107424736023,
0.15727956593036652,
-0.06044219061732292,
0.07659310102462769,
0.1564161479473114,
-0.12999629974365234,
-0.17598319053649902,
-0.0958087146282196,
0.052657872438430786,
-0.0022943171206861734,
0.010701284743845463,
-0.1882411539554596,
0.07577592134475708,
0.10238347202539444,
-0.014260304160416126,
-0.09820341318845749,
-0.31268560886383057,
-0.10006320476531982,
0.08948606997728348,
0.07572510838508606,
0.14116118848323822,
-0.14300470054149628,
-0.022387029603123665,
-0.09737753868103027,
-0.02373063750565052,
0.32563692331314087,
-0.07908619195222855,
0.11513449996709824,
0.006708592176437378,
-0.005339004099369049,
0.017618998885154724,
-0.007111316546797752,
0.10118609666824341,
0.06391410529613495,
0.07238401472568512,
-0.0010573635809123516,
-0.03157266229391098,
0.12429160624742508,
-0.043994881212711334,
0.0791054517030716,
-0.0807863101363182,
0.03458205983042717,
-0.0036144908517599106,
-0.05679265782237053,
-0.06492902338504791,
-0.0016901958733797073,
-0.05981303006410599,
-0.07778781652450562,
-0.03170030191540718,
0.02783580683171749,
0.049517083913087845,
-0.03456871584057808,
0.19143566489219666,
-0.09633027017116547,
0.04450080916285515,
0.16865786910057068,
0.11210394650697708,
0.09611877799034119,
-0.1323700100183487,
-0.023861965164542198,
-0.05238739401102066,
0.09035442769527435,
-0.14058387279510498,
0.09721042215824127,
0.04075280949473381,
-0.019153106957674026,
0.13013513386249542,
0.10257936269044876,
-0.04201320558786392,
0.005299472715705633,
0.05364769324660301,
-0.069093719124794,
-0.013505956158041954,
-0.006708675064146519,
0.02270290069282055,
-0.013517465442419052,
0.0885765329003334,
0.18381044268608093,
-0.03942682594060898,
-0.039468929171562195,
0.01676282100379467,
0.0241100937128067,
-0.05941447243094444,
0.09033694118261337,
0.031980983912944794,
0.05016849935054779,
-0.0845252051949501,
0.1065291091799736,
0.09383998066186905,
-0.018360288813710213,
0.03850986808538437,
0.09511259198188782,
-0.07044090330600739,
-0.09541173279285431,
-0.07573738694190979,
0.06390374153852463,
-0.13005295395851135,
-0.03210396692156792,
-0.09693922102451324,
-0.10981699824333191,
-0.020465118810534477,
0.12053918093442917,
0.07337996363639832,
0.13589875400066376,
-0.09528075158596039,
-0.0058266520500183105,
-0.09304293245077133,
0.04597464203834534,
-0.042746372520923615,
0.07824336737394333,
-0.07284919917583466,
0.10182245820760727,
0.01442404929548502,
-0.009569516405463219,
-0.06037160009145737,
-0.057536300271749496,
-0.16596199572086334,
0.01353226788341999,
-0.11248540133237839,
0.013369976542890072,
-0.07564810663461685,
-0.003428440308198333,
0.02889503911137581,
-0.013047188520431519,
0.02233140915632248,
0.015427744947373867,
-0.014243711717426777,
-0.001076067448593676,
-0.007512488402426243,
0.07452798634767532,
-0.09254355728626251,
-0.0224663857370615,
-0.0007979341316968203,
-0.10577289015054703,
0.07481153309345245,
0.022001992911100388,
-0.05968869850039482,
0.06870705634355545,
-0.12014054507017136,
0.018218807876110077,
-0.013477991335093975,
0.049837224185466766,
0.008990928530693054,
-0.07807828485965729,
0.010413790121674538,
0.02911846712231636,
0.03793303668498993,
0.02012135088443756,
0.0702735185623169,
-0.10435860604047775,
-0.005537795834243298,
0.0025856862775981426,
-0.0888323113322258,
-0.06285945326089859,
-0.005251996219158173,
0.08319589495658875,
0.06082065775990486,
0.16617128252983093,
-0.0849226638674736,
0.07734136283397675,
-0.08354036509990692,
-0.024099072441458702,
0.038741666823625565,
-0.06119823828339577,
-0.026489082723855972,
-0.09731624275445938,
0.024220485240221024,
-0.0221981443464756,
0.15727193653583527,
-0.05304912477731705,
0.10377173870801926,
0.025893371552228928,
0.007442222908139229,
0.10159438848495483,
0.07511613517999649,
0.16644248366355896,
0.03194189444184303,
-0.0273452065885067,
-0.044500235468149185,
0.0070147085934877396,
-0.00298611749894917,
0.012824890203773975,
0.0338033027946949,
0.04943566024303436,
-0.044641975313425064,
0.09158478677272797,
0.04284714534878731,
0.0834619477391243,
-0.09450089186429977,
0.0035546980798244476,
0.03256475180387497,
0.003104743082076311,
-0.04372749105095863,
0.05473312735557556,
0.14064237475395203,
-0.07580794394016266,
0.09932668507099152,
0.03387482091784477,
-0.10877352207899094,
-0.1351385861635208,
-0.1863681524991989,
-0.05050908029079437,
-0.07281213998794556,
-0.04108026623725891,
-0.15000101923942566,
-0.005527449306100607,
0.08127135038375854,
0.02964053489267826,
-0.03025227040052414,
0.03608755022287369,
0.11143360286951065,
-0.04370906949043274,
-0.00046194856986403465,
-0.03292416036128998,
0.04451778531074524,
0.08353973925113678,
0.03823191672563553,
0.03840858116745949,
0.06304603070020676,
0.0344526544213295,
0.028357230126857758,
0.07804963737726212,
-0.0008654832490719855,
-0.08324922621250153,
-0.09817887842655182,
0.020858142524957657,
0.012600918300449848,
0.002442109165713191,
0.1435920149087906,
0.047566432505846024,
-0.09237156063318253,
0.0035901819355785847,
0.18544302880764008,
-0.07793515175580978,
-0.1289753019809723,
-0.15431572496891022,
0.30636870861053467,
0.04648807272315025,
0.0860166996717453,
0.0031372213270515203,
-0.08658036589622498,
-0.07398116588592529,
0.1415596306324005,
0.11029724031686783,
-0.01313875149935484,
-0.0018154880963265896,
0.07163608819246292,
-0.011961987242102623,
0.027820246294140816,
0.09145978093147278,
-0.01949695497751236,
0.2702353596687317,
-0.03983939811587334,
0.09685558080673218,
-0.009753075428307056,
-0.018274527043104172,
-0.08570487052202225,
0.09368117898702621,
-0.05075543373823166,
-0.011814014986157417,
-0.04103449732065201,
0.06783640384674072,
-0.10198454558849335,
0.027102762833237648,
0.07177355140447617,
-0.009137939661741257,
-0.09852461516857147,
-0.01163004245609045,
0.09070277959108353,
-0.06330209225416183,
0.07219487428665161,
0.00593530060723424,
-0.00828663632273674,
0.13653971254825592,
-0.027425991371273994,
-0.018974507227540016,
-0.04480079934000969,
0.029946543276309967,
-0.004020014777779579,
0.05136985331773758,
-0.023417271673679352,
0.1450430303812027,
0.11358147859573364,
0.04622943699359894,
-0.04121633619070053,
0.13102847337722778,
-0.030566634610295296,
-0.04576524347066879,
0.079828180372715,
0.0841294452548027,
-0.03201373666524887,
0.1340978443622589,
0.05817287042737007,
-0.1661582887172699,
0.050168827176094055,
-0.0027838374953716993,
-0.09595660120248795,
-0.048232778906822205,
0.009206850081682205,
-0.05545344203710556,
0.04720744490623474,
0.121429443359375,
-0.056501343846321106,
-0.027821805328130722,
-0.03406064212322235,
0.01209392212331295,
0.003256601979956031,
-0.020942598581314087,
-0.05718604102730751,
-0.15206564962863922,
0.01133764162659645,
0.1001632884144783,
0.04775266721844673,
-0.2003321647644043,
-0.020460201427340508,
0.023056350648403168,
-0.00819737371057272,
-0.09109323471784592,
0.05224178358912468,
0.020707691088318825,
0.025865281000733376,
-0.05955204740166664,
-0.24627716839313507,
-0.008364363573491573,
0.07661200314760208,
-0.13099971413612366,
-0.12398263067007065
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-base](https://huggingface.co/microsoft/deberta-v3-base)
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 92.38
- Accuracy on MNLI mismatched set: 90.04
For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-deberta-v3-base')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-base')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-base')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-base')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-deberta-v3-base
|
[
"transformers",
"pytorch",
"deberta-v2",
"text-classification",
"microsoft/deberta-v3-base",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 92.38
- Accuracy on MNLI mismatched set: 90.04
For futher evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 92.38\r\n- Accuracy on MNLI mismatched set: 90.04\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 92.38\r\n- Accuracy on MNLI mismatched set: 90.04\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
83,
45,
49,
51,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-base## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 92.38\r\n- Accuracy on MNLI mismatched set: 90.04\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.05504194274544716,
0.18498900532722473,
-0.00550383236259222,
0.10590871423482895,
0.11327389627695084,
0.027110308408737183,
0.22715409100055695,
0.06776661425828934,
0.047567833214998245,
0.009399577043950558,
-0.022405998781323433,
0.11900611966848373,
0.04960653930902481,
0.087869793176651,
-0.012860161252319813,
-0.17446796596050262,
0.022154854610562325,
-0.060964759439229965,
0.04712884873151779,
0.08188575506210327,
0.0980304479598999,
-0.0625922754406929,
0.0420386902987957,
-0.013875761069357395,
-0.09353657811880112,
0.029924092814326286,
-0.022783271968364716,
0.00029585283482447267,
0.056830354034900665,
0.08876778930425644,
0.11350555717945099,
0.03970613703131676,
0.054754313081502914,
-0.20077818632125854,
0.00981262605637312,
0.07366976886987686,
-0.018998906016349792,
0.020158149302005768,
0.05771129950881004,
-0.09196603298187256,
0.02219942770898342,
0.010708280839025974,
0.05780442804098129,
0.07142206281423569,
-0.06680011004209518,
-0.09239906817674637,
-0.04663665592670441,
0.020442496985197067,
0.08217284083366394,
0.09513447433710098,
-0.031120484694838524,
0.09341774880886078,
-0.047955457121133804,
0.08692276477813721,
0.033932022750377655,
-0.20560957491397858,
-0.01999054104089737,
0.07665058225393295,
-0.00525390449911356,
0.10038639605045319,
-0.07866141945123672,
-0.04238595440983772,
0.0256351325660944,
0.0037330882623791695,
-0.02014920301735401,
-0.016897566616535187,
0.03941161185503006,
-0.008122952654957771,
-0.13860152661800385,
-0.009864633902907372,
0.12632808089256287,
-0.01790344901382923,
-0.07836967706680298,
-0.11451000720262527,
-0.05198528617620468,
-0.01190938614308834,
0.027615707367658615,
-0.012339326553046703,
0.046395160257816315,
0.044713109731674194,
0.03577519208192825,
-0.034368909895420074,
-0.09168870747089386,
-0.04361942037940025,
-0.06804998964071274,
0.11794951558113098,
0.0334017276763916,
0.037278905510902405,
-0.0888497531414032,
0.10708896070718765,
0.04664485529065132,
-0.1272583156824112,
-0.07123062759637833,
-0.06702927500009537,
-0.1781214475631714,
0.0028527032118290663,
-0.11030040681362152,
-0.15808755159378052,
0.043982647359371185,
0.20869985222816467,
0.02005097270011902,
0.049231406301259995,
-0.01257365196943283,
0.030337277799844742,
0.11087530106306076,
0.10425414144992828,
-0.014277372509241104,
-0.02953529916703701,
-0.018479187041521072,
0.03434520214796066,
0.00937954243272543,
-0.0347900390625,
-0.07278183102607727,
0.03311194106936455,
-0.037407368421554565,
0.052623506635427475,
0.02084149606525898,
0.020093923434615135,
-0.06911692768335342,
-0.07502651959657669,
0.06647256016731262,
-0.13196279108524323,
0.013282383792102337,
0.03074638545513153,
-0.021044738590717316,
0.1381591260433197,
0.04739317670464516,
-0.01108813751488924,
-0.05282898619771004,
-0.006589648779481649,
-0.07417338341474533,
0.025174275040626526,
-0.11721331626176834,
-0.12351434677839279,
0.01577894203364849,
0.01619052328169346,
-0.04012269526720047,
-0.1093621551990509,
-0.10973720997571945,
-0.06980929523706436,
0.04586409404873848,
-0.032204337418079376,
0.04230799898505211,
-0.1013774424791336,
0.0031543162185698748,
0.018516983836889267,
0.008407749235630035,
-0.07309696823358536,
-0.013904559426009655,
0.006470143795013428,
0.006821855902671814,
0.05714721232652664,
0.0030589206144213676,
-0.00819788221269846,
-0.11710900813341141,
-0.03597593680024147,
0.03594997897744179,
0.16159865260124207,
-0.07749796658754349,
0.01715139113366604,
-0.06503398716449738,
-0.006782847456634045,
0.035939574241638184,
0.029043620452284813,
0.006536621134728193,
0.11206056922674179,
-0.2584216892719269,
-0.030810268595814705,
0.19104233384132385,
-0.18144071102142334,
-0.060482531785964966,
0.09351000189781189,
-0.04705575853586197,
-0.01418253593146801,
0.08187954127788544,
0.035303402692079544,
0.19866596162319183,
-0.08584849536418915,
-0.023834193125367165,
0.03971109166741371,
-0.025841763243079185,
0.09634270519018173,
0.07864458858966827,
0.026394125074148178,
-0.025367368012666702,
0.024069787934422493,
-0.07283791899681091,
0.00022469974646810442,
-0.027742238715291023,
-0.07841481268405914,
-0.012613152153789997,
-0.056705292314291,
0.016994521021842957,
-0.006013673730194569,
-0.0005946015007793903,
-0.0015652847941964865,
-0.10310592502355576,
0.12788431346416473,
0.10505099594593048,
-0.0985238254070282,
0.020611582323908806,
-0.11138486117124557,
-0.07536660879850388,
-0.009428991936147213,
0.00015262819943018258,
-0.18932463228702545,
-0.09250278770923615,
0.02911352552473545,
-0.11723294109106064,
0.050749171525239944,
0.07531702518463135,
0.010720492340624332,
0.058598339557647705,
0.0015055234543979168,
-0.008644422516226768,
-0.0001814083952922374,
-0.006822328083217144,
-0.0433112196624279,
-0.15571190416812897,
-0.04337691515684128,
-0.03842359781265259,
0.07450328767299652,
-0.1083345040678978,
0.030728261917829514,
0.014297927729785442,
0.07298529148101807,
0.0011833281023427844,
-0.050213128328323364,
-0.0036800713278353214,
0.04567810520529747,
-0.05741357058286667,
-0.023806100711226463,
0.02381780743598938,
0.03125442937016487,
-0.06499721854925156,
0.0068349954672157764,
-0.10206270217895508,
-0.09298504889011383,
0.0743955448269844,
0.030032623559236526,
-0.045575957745313644,
0.0054344660602509975,
-0.026768343523144722,
-0.004357573576271534,
-0.05072944983839989,
-0.029448527842760086,
0.15033066272735596,
0.03960495442152023,
0.08960851281881332,
-0.05824699252843857,
-0.04624389484524727,
-0.022326815873384476,
-0.04489469528198242,
-0.0016865329816937447,
0.07372523099184036,
-0.044137462973594666,
-0.17496323585510254,
0.08889777213335037,
-0.010667950846254826,
-0.07524282485246658,
0.11211281269788742,
0.015653088688850403,
-0.05002203583717346,
-0.08680889755487442,
0.000930127571336925,
0.02013438194990158,
0.024658076465129852,
-0.04671543836593628,
0.041982270777225494,
0.034399766474962234,
0.03163986653089523,
0.020479237660765648,
-0.10359732061624527,
0.029420889914035797,
0.04479675367474556,
-0.0278167724609375,
-0.00014570071652997285,
0.0036901403218507767,
-0.012195444665849209,
0.05434782803058624,
-0.029096171259880066,
0.053911492228507996,
0.010792218148708344,
-0.08213572949171066,
-0.14955881237983704,
0.1994115114212036,
-0.12241100519895554,
-0.2808363735675812,
-0.13221433758735657,
0.04644712060689926,
-0.06551565229892731,
0.009064863435924053,
0.05141991004347801,
-0.072455033659935,
-0.04749874025583267,
-0.09251312166452408,
-0.037831857800483704,
-0.06092586740851402,
-0.033548105508089066,
-0.0441695861518383,
0.055724505335092545,
0.02894425019621849,
-0.1518130749464035,
0.02382926642894745,
0.01567000523209572,
-0.08721353858709335,
0.023121468722820282,
0.05228535458445549,
0.03364504128694534,
0.12510274350643158,
0.026103613898158073,
0.017344849184155464,
0.014317595399916172,
0.2099318504333496,
-0.010842018760740757,
-0.009409492835402489,
0.16040030121803284,
-0.04278382658958435,
0.06758035719394684,
0.0841597244143486,
0.035817213356494904,
-0.07283150404691696,
0.04025954380631447,
0.043102845549583435,
-0.040682584047317505,
-0.2191053032875061,
-0.0374254435300827,
-0.0453595407307148,
-0.053051259368658066,
0.04747210815548897,
0.017003025859594345,
0.00003742859189515002,
0.026984430849552155,
-0.036188360303640366,
-0.033118754625320435,
0.04294206202030182,
0.0830555334687233,
0.08507087826728821,
0.015191234648227692,
0.10150206834077835,
-0.06411205232143402,
0.0007889331318438053,
0.06862052530050278,
0.02737266756594181,
0.13403761386871338,
0.018215110525488853,
0.062092628329992294,
0.14115402102470398,
0.0007196529186330736,
0.03844102472066879,
0.01684563048183918,
-0.024710746482014656,
0.042225368320941925,
0.016566842794418335,
-0.08079472184181213,
-0.03737872838973999,
0.06574597954750061,
-0.04520781338214874,
-0.00975245051085949,
-0.029636720195412636,
-0.04465973377227783,
0.07290121912956238,
0.08444289863109589,
0.06730324029922485,
-0.19123469293117523,
-0.07690318673849106,
0.04482538625597954,
-0.03502175211906433,
-0.019144471734762192,
0.00846159178763628,
0.07304035872220993,
-0.08820124715566635,
0.07147428393363953,
-0.019599756225943565,
0.08889251202344894,
-0.13368314504623413,
-0.02138609066605568,
0.031508851796388626,
0.07431261986494064,
0.035433635115623474,
0.08450385183095932,
-0.2430252879858017,
0.043314170092344284,
-0.005113723687827587,
0.04611924663186073,
-0.017762305215001106,
0.0583326481282711,
0.016640249639749527,
0.0056463307701051235,
0.09610150009393692,
0.016486236825585365,
-0.03809536620974541,
-0.06276077777147293,
-0.04973018169403076,
0.01815425045788288,
0.09234899282455444,
-0.0842365026473999,
0.1107988953590393,
0.0009532049880363047,
-0.011686963029205799,
-0.030774207785725594,
0.04680319130420685,
-0.20535685122013092,
-0.20215071737766266,
0.06971310824155807,
-0.005529276095330715,
-0.010821208357810974,
-0.03295360133051872,
-0.06344439089298248,
-0.021343041211366653,
0.18232578039169312,
-0.2676972448825836,
-0.09968461096286774,
-0.1292494386434555,
0.06743105500936508,
0.08672717958688736,
-0.06368785351514816,
-0.004118633456528187,
0.00032067415304481983,
0.13790972530841827,
-0.03186510130763054,
-0.13643941283226013,
0.0022550728172063828,
-0.07091901451349258,
-0.09933602064847946,
-0.0700153186917305,
0.0577910952270031,
0.1268482357263565,
0.016119161620736122,
0.012887459248304367,
0.014917830005288124,
-0.021414864808321,
-0.12657129764556885,
-0.07375578582286835,
0.1955774426460266,
0.005497160833328962,
0.050536543130874634,
-0.052707739174366,
-0.11646358668804169,
-0.053731419146060944,
0.00653095031157136,
0.10237399488687515,
0.14385227859020233,
-0.06381053477525711,
0.09386065602302551,
0.10036513209342957,
-0.14982110261917114,
-0.14412900805473328,
-0.06292212009429932,
0.10731718689203262,
0.03679124638438225,
0.026879504323005676,
-0.1850062757730484,
0.05401420593261719,
0.07810583710670471,
0.0007905822130851448,
-0.05665091797709465,
-0.32659003138542175,
-0.10491985827684402,
0.030602186918258667,
0.08680696785449982,
0.027519717812538147,
-0.10047727078199387,
-0.05276799947023392,
-0.06236030533909798,
-0.07109225541353226,
0.24590544402599335,
0.013673312962055206,
0.08993037790060043,
0.019197948276996613,
-0.06662249565124512,
0.0380304679274559,
-0.033092107623815536,
0.10368650406599045,
0.06475149840116501,
0.08532814681529999,
-0.029724158346652985,
0.028986984863877296,
0.048918794840574265,
-0.03748374804854393,
0.10038186609745026,
-0.03363662213087082,
0.08263049274682999,
-0.08053994923830032,
-0.0643891766667366,
-0.012862388975918293,
-0.030838094651699066,
-0.04473068192601204,
-0.033143624663352966,
-0.04451719671487808,
0.056636396795511246,
0.09044865518808365,
-0.022968539968132973,
0.042835816740989685,
-0.08179818093776703,
0.0030684301164001226,
0.17638689279556274,
0.13431280851364136,
0.08338948339223862,
-0.1430440992116928,
-0.03945200517773628,
-0.005747866816818714,
0.0784253180027008,
-0.15915971994400024,
0.06068247929215431,
0.09306389838457108,
0.00307631678879261,
0.11957995593547821,
0.06650403887033463,
-0.061954572796821594,
0.0008464406710118055,
0.07332736253738403,
-0.08233359456062317,
-0.08449359983205795,
-0.026530278846621513,
-0.014443869702517986,
-0.05248107761144638,
0.006567076314240694,
0.1620473712682724,
-0.032912857830524445,
-0.01715313270688057,
-0.004230120684951544,
0.03040088340640068,
-0.02396770939230919,
0.12579575181007385,
0.007509858347475529,
0.05384809151291847,
-0.08447986841201782,
0.10498879849910736,
0.07405369728803635,
-0.055420421063899994,
0.025117255747318268,
0.08802884072065353,
-0.09397705644369125,
-0.07642081379890442,
-0.07873471081256866,
0.07749859988689423,
-0.10678935796022415,
-0.015011227689683437,
-0.07683262974023819,
-0.06368281692266464,
0.008625335060060024,
0.11357968300580978,
0.07616094499826431,
0.12236165255308151,
-0.03562367334961891,
-0.009489047341048717,
-0.044686999171972275,
0.08360257744789124,
0.01730869710445404,
0.013971706852316856,
-0.0729127898812294,
0.09206555783748627,
0.027562865987420082,
-0.022792065516114235,
-0.022118747234344482,
-0.05011564493179321,
-0.11403799802064896,
0.0007474387530237436,
-0.07230600714683533,
0.03707944229245186,
-0.044822100549936295,
0.009820621460676193,
0.02886822819709778,
-0.01765349693596363,
0.013824311085045338,
0.044652070850133896,
-0.01898236945271492,
-0.041413240134716034,
-0.026295199990272522,
0.08962386846542358,
-0.11560934036970139,
-0.012414404191076756,
0.0022809773217886686,
-0.08495457470417023,
0.07077190279960632,
0.02228407934308052,
-0.045053694397211075,
0.009669739753007889,
-0.1592082679271698,
0.013810086995363235,
-0.009634409099817276,
0.09126240015029907,
0.0067637208849191666,
-0.08613287657499313,
0.05600525811314583,
0.0405578538775444,
-0.01700415089726448,
0.012183732353150845,
0.11135251075029373,
-0.10732860863208771,
-0.04200209677219391,
0.01903790980577469,
-0.010048873722553253,
-0.0683654174208641,
-0.02090873382985592,
0.1011577770113945,
0.07878413051366806,
0.1601315140724182,
-0.05765629932284355,
0.09059855341911316,
-0.16924983263015747,
-0.04424883425235748,
0.06704110652208328,
0.014047033153474331,
-0.0403677336871624,
-0.08153467625379562,
0.04822923615574837,
-0.014816482551395893,
0.156684011220932,
-0.010644150897860527,
0.09035526216030121,
0.028607718646526337,
-0.005412376951426268,
0.05004942789673805,
0.056467726826667786,
0.1252620816230774,
0.032749395817518234,
0.018169699236750603,
-0.001147633302025497,
-0.019088109955191612,
-0.02833385206758976,
-0.08665604144334793,
0.05277206003665924,
0.07217884808778763,
0.0028929272666573524,
0.05970098823308945,
0.0585295744240284,
0.022957148030400276,
-0.05473305657505989,
0.04166874289512634,
-0.04177645593881607,
0.031668052077293396,
-0.05330437049269676,
0.030936282128095627,
0.11086554825305939,
-0.09851198643445969,
0.10417468845844269,
0.04388988018035889,
-0.08361519873142242,
-0.11755797266960144,
-0.1959228366613388,
-0.058607012033462524,
-0.06164124235510826,
-0.00793041754513979,
-0.14971981942653656,
0.03239493817090988,
0.02468680776655674,
0.030236756429076195,
-0.01650729961693287,
0.02879367209970951,
0.03597574308514595,
-0.05494645982980728,
0.009652252309024334,
-0.013384037651121616,
0.04743177816271782,
0.06195243448019028,
0.029018133878707886,
0.08591712266206741,
0.05816486477851868,
0.034173645079135895,
0.05000169947743416,
0.09236400574445724,
0.01564439758658409,
-0.02632337436079979,
-0.07658162713050842,
0.01973372884094715,
-0.010552431456744671,
0.015115015208721161,
0.130109965801239,
0.08645438402891159,
-0.036448486149311066,
-0.004318690858781338,
0.17335094511508942,
-0.06728760898113251,
-0.11852625757455826,
-0.16631029546260834,
0.23736071586608887,
0.03406469523906708,
0.07335697114467621,
0.01075048092752695,
-0.0871499553322792,
-0.04426463693380356,
0.16756469011306763,
0.11053695529699326,
-0.0447482094168663,
0.00004068798807566054,
0.05820777267217636,
-0.011789142154157162,
0.047675687819719315,
0.07236865907907486,
-0.010301457718014717,
0.285607248544693,
-0.015102813020348549,
0.08420159667730331,
-0.027701037004590034,
-0.025477083399891853,
-0.010551589541137218,
0.09621260315179825,
-0.06445971131324768,
0.0005277849268168211,
-0.04907277226448059,
0.10025610029697418,
-0.05446722358465195,
-0.027208935469388962,
0.1188337504863739,
-0.030321670696139336,
-0.09363868832588196,
0.003559492528438568,
0.10779121518135071,
-0.02672160416841507,
0.04499704763293266,
-0.012052743695676327,
-0.008704240433871746,
0.20586338639259338,
-0.0336524173617363,
-0.05189778655767441,
-0.08592905104160309,
-0.02672574482858181,
-0.08246854692697525,
0.09190065413713455,
0.006385164801031351,
0.14535409212112427,
0.09763486683368683,
0.04096727445721626,
-0.09245656430721283,
0.08173714578151703,
-0.026385648176074028,
-0.09223491698503494,
0.0661483034491539,
0.10716433823108673,
-0.025122305378317833,
0.16293202340602875,
0.06686834245920181,
-0.10190567374229431,
0.03913125395774841,
0.006544200703501701,
-0.04689357057213783,
-0.073078453540802,
0.0356152281165123,
-0.07678929716348648,
0.09748276323080063,
0.14592063426971436,
-0.03842494636774063,
0.0007231294875964522,
-0.006602349691092968,
0.0662732943892479,
-0.027565699070692062,
0.05327728018164635,
-0.030983202159404755,
-0.16287203133106232,
0.022207386791706085,
0.028720468282699585,
0.04467839002609253,
-0.14724937081336975,
-0.050950564444065094,
0.02882760763168335,
-0.024051189422607422,
-0.025614513084292412,
0.08672451972961426,
0.03650067001581192,
0.040117595344781876,
-0.04832733795046806,
-0.2077750712633133,
0.03914104029536247,
0.07718996703624725,
-0.09839004278182983,
-0.10816533118486404
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large)
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 92.20
- Accuracy on MNLI mismatched set: 90.49
For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-deberta-v3-large')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-large')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-large')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-large')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-large"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-deberta-v3-large
|
[
"transformers",
"pytorch",
"deberta-v2",
"text-classification",
"microsoft/deberta-v3-large",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-large #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 92.20
- Accuracy on MNLI mismatched set: 90.49
For futher evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 92.20\r\n- Accuracy on MNLI mismatched set: 90.49\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-large #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 92.20\r\n- Accuracy on MNLI mismatched set: 90.49\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
84,
46,
49,
51,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-large #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-large## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 92.20\r\n- Accuracy on MNLI mismatched set: 90.49\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.057665105909109116,
0.18749360740184784,
-0.005360424518585205,
0.11260697990655899,
0.124428391456604,
0.03597496077418327,
0.21414512395858765,
0.0684213638305664,
0.05462667718529701,
0.013975394889712334,
-0.021411990746855736,
0.12714605033397675,
0.05493280291557312,
0.08521971851587296,
-0.03299973905086517,
-0.17481763660907745,
0.030826937407255173,
-0.06417763978242874,
0.033448584377765656,
0.07868829369544983,
0.10413143038749695,
-0.05694631114602089,
0.04402574896812439,
-0.02547227218747139,
-0.11947084218263626,
0.0278790220618248,
-0.009689955040812492,
0.00522234570235014,
0.051854293793439865,
0.09441712498664856,
0.12076503783464432,
0.03237863630056381,
0.06096956133842468,
-0.19429568946361542,
0.008188631385564804,
0.06436579674482346,
-0.029312923550605774,
0.02400778792798519,
0.06048731133341789,
-0.07863526046276093,
0.013365150429308414,
-0.0184487272053957,
0.05757369101047516,
0.0826721042394638,
-0.0772474929690361,
-0.07403121143579483,
-0.0579853430390358,
0.022782428190112114,
0.07859579473733902,
0.09012099355459213,
-0.03662097081542015,
0.10435523092746735,
-0.059308234602212906,
0.09157734364271164,
0.03107433207333088,
-0.21792584657669067,
-0.027574265375733376,
0.10919419676065445,
-0.010593520477414131,
0.06392355263233185,
-0.09540204703807831,
-0.03552809730172157,
0.03054763562977314,
0.003580765100196004,
-0.021235376596450806,
-0.032983433455228806,
0.037567514926195145,
-0.025145642459392548,
-0.13379289209842682,
0.004098758567124605,
0.15568944811820984,
-0.005732879042625427,
-0.07971493899822235,
-0.10764528810977936,
-0.04420115798711777,
0.011245384812355042,
0.0212345402687788,
-0.013802890665829182,
0.049024973064661026,
0.03860615938901901,
0.042909663170576096,
-0.027878105640411377,
-0.08155112713575363,
-0.041080545634031296,
-0.06712549179792404,
0.11067065596580505,
0.023882165551185608,
0.019579702988266945,
-0.07150612026453018,
0.08721789717674255,
0.013381178490817547,
-0.12713360786437988,
-0.07692398875951767,
-0.06587696075439453,
-0.20380225777626038,
-0.01908659003674984,
-0.11178566515445709,
-0.1266910284757614,
0.04743029549717903,
0.22238828241825104,
0.016788771376013756,
0.05589614436030388,
-0.01910209096968174,
0.02586151286959648,
0.11089242994785309,
0.11586755514144897,
-0.01817133091390133,
-0.03449416905641556,
-0.013699338771402836,
0.03577441722154617,
0.0001439151237718761,
-0.03508862853050232,
-0.07775447517633438,
0.021693643182516098,
-0.023777026683092117,
0.05819425731897354,
0.012987148016691208,
0.025885554030537605,
-0.07576637715101242,
-0.0728289857506752,
0.04582131281495094,
-0.1463281810283661,
0.02260354720056057,
0.030153706669807434,
-0.021976225078105927,
0.13877995312213898,
0.03682136535644531,
-0.010269691236317158,
-0.0543697252869606,
-0.006753763649612665,
-0.08732923120260239,
0.0291579682379961,
-0.10947098582983017,
-0.12894666194915771,
0.017647719010710716,
0.01578080840408802,
-0.03457295149564743,
-0.10102622210979462,
-0.10997718572616577,
-0.0799730196595192,
0.04159274324774742,
-0.042273376137018204,
0.03651728481054306,
-0.10499851405620575,
-0.010253130458295345,
0.018152723088860512,
0.018452053889632225,
-0.07126293331384659,
-0.0051747336983680725,
0.003612604457885027,
0.013888021931052208,
0.04588911309838295,
-0.01944245956838131,
-0.00347140827216208,
-0.11644599586725235,
-0.03008832223713398,
0.018221072852611542,
0.17437005043029785,
-0.08743807673454285,
0.022183049470186234,
-0.06788342446088791,
-0.012129595503211021,
0.04933780059218407,
0.031438734382390976,
0.006341559812426567,
0.11363165825605392,
-0.27857905626296997,
-0.02079509198665619,
0.1852819323539734,
-0.17586204409599304,
-0.058162715286016464,
0.08968862146139145,
-0.04986872151494026,
0.003432365832850337,
0.09094945341348648,
0.04772567376494408,
0.1907980591058731,
-0.0633150264620781,
-0.04464982450008392,
0.03979748114943504,
-0.028503358364105225,
0.09688862413167953,
0.07944066822528839,
0.029896317049860954,
-0.03609561547636986,
0.022328805178403854,
-0.08807741850614548,
-0.0077741811983287334,
-0.02598041668534279,
-0.07691919803619385,
-0.019325541332364082,
-0.05756010860204697,
0.016163425520062447,
-0.0025257959496229887,
-0.0017187452176585793,
-0.002616995247080922,
-0.09822903573513031,
0.15441814064979553,
0.10775012522935867,
-0.08834459632635117,
0.013333420269191265,
-0.11338664591312408,
-0.0627516582608223,
0.0017005131812766194,
0.007857901975512505,
-0.1924450397491455,
-0.10430307686328888,
0.023818427696824074,
-0.11876226216554642,
0.06154824048280716,
0.0645974799990654,
-0.00046903264592401683,
0.050695259124040604,
-0.008530653081834316,
-0.005592285189777613,
-0.011149031110107899,
-0.002284345217049122,
-0.04571639373898506,
-0.1634732484817505,
-0.046855852007865906,
-0.033279091119766235,
0.07354672253131866,
-0.12968425452709198,
0.038164619356393814,
0.020890727639198303,
0.07344438880681992,
-0.008698543533682823,
-0.0475444421172142,
-0.005303744226694107,
0.04697900637984276,
-0.05418785661458969,
-0.031500641256570816,
0.02776838093996048,
0.022325146943330765,
-0.07177898287773132,
-0.01577751711010933,
-0.09356905519962311,
-0.08638206869363785,
0.0749434307217598,
0.024311508983373642,
-0.06713926792144775,
0.00669783353805542,
-0.013919364660978317,
-0.0008909713942557573,
-0.052165985107421875,
-0.037982694804668427,
0.1494857668876648,
0.03555867448449135,
0.08326481282711029,
-0.05364582687616348,
-0.04386705160140991,
-0.021648945286870003,
-0.03292682021856308,
-0.0075813839212059975,
0.08220168948173523,
-0.039504263550043106,
-0.21701526641845703,
0.09629640728235245,
0.001731791766360402,
-0.07103391736745834,
0.10204152762889862,
0.007804511114954948,
-0.03915666416287422,
-0.09736821055412292,
0.003968504257500172,
0.019769337028265,
0.03015170246362686,
-0.052374277263879776,
0.042239103466272354,
0.02941642515361309,
0.024998938664793968,
0.012696952559053898,
-0.08994367718696594,
0.024478083476424217,
0.03306983411312103,
-0.021085407584905624,
-0.018499024212360382,
-0.0017380345379933715,
-0.005166277289390564,
0.055426955223083496,
-0.028111081570386887,
0.05027680844068527,
0.0182187519967556,
-0.07955874502658844,
-0.14706432819366455,
0.21186062693595886,
-0.1267666071653366,
-0.29290419816970825,
-0.13089533150196075,
0.046468522399663925,
-0.0774347186088562,
0.0051574790850281715,
0.03937140479683876,
-0.05986529216170311,
-0.05219099670648575,
-0.07977934181690216,
-0.03320649638772011,
-0.05274564027786255,
-0.027074161916971207,
-0.036772120743989944,
0.06054796278476715,
0.03860192373394966,
-0.15082554519176483,
0.02644861675798893,
0.016771407797932625,
-0.0918116644024849,
0.025218816474080086,
0.04909722879528999,
0.034891873598098755,
0.12289142608642578,
0.023938274011015892,
0.015967542305588722,
0.010422807186841965,
0.22915177047252655,
-0.01910499483346939,
-0.007999210618436337,
0.13789662718772888,
-0.042449429631233215,
0.056360840797424316,
0.08899103105068207,
0.04169352725148201,
-0.09107819199562073,
0.034382421523332596,
0.05126716569066048,
-0.03418070077896118,
-0.22638624906539917,
-0.0401419922709465,
-0.052265699952840805,
-0.05004623159766197,
0.05254204198718071,
0.012379410676658154,
-0.011889515444636345,
0.02845088578760624,
-0.026638058945536613,
-0.017692891880869865,
0.040743667632341385,
0.08135311305522919,
0.1022188737988472,
0.01662721112370491,
0.10964755713939667,
-0.05655030161142349,
0.012636630795896053,
0.0697571188211441,
0.012473252601921558,
0.12328147143125534,
0.03019139915704727,
0.05531429126858711,
0.13001295924186707,
-0.00199308292940259,
0.0321027971804142,
0.025434719398617744,
-0.022264394909143448,
0.03257638216018677,
0.019793175160884857,
-0.0775173008441925,
-0.038814958184957504,
0.06517823785543442,
-0.029327796772122383,
-0.0013128501595929265,
-0.041973937302827835,
-0.04573383927345276,
0.08372483402490616,
0.06903909891843796,
0.07332003116607666,
-0.19415488839149475,
-0.08053550869226456,
0.04872570186853409,
-0.025881577283143997,
-0.02876596711575985,
0.005629075225442648,
0.06838619709014893,
-0.08527381718158722,
0.08008579909801483,
-0.01814793050289154,
0.09516307711601257,
-0.12457230687141418,
-0.016709117218852043,
0.015517441555857658,
0.07965568453073502,
0.02490548975765705,
0.09043360501527786,
-0.2550288140773773,
0.04072898253798485,
0.00717624556273222,
0.047693051397800446,
-0.008393471129238605,
0.05171489343047142,
0.02945207804441452,
0.02867080457508564,
0.10762064903974533,
0.02067841775715351,
-0.03271696716547012,
-0.09061522781848907,
-0.05023936927318573,
0.018165867775678635,
0.09380891174077988,
-0.07836149632930756,
0.10808750987052917,
-0.007234653457999229,
-0.0056988815777003765,
-0.01972959190607071,
0.041364751756191254,
-0.21893079578876495,
-0.19074194133281708,
0.07411524653434753,
-0.011108919978141785,
0.00011937936506001279,
-0.02815171703696251,
-0.06802035868167877,
-0.027272144332528114,
0.20286639034748077,
-0.2734784185886383,
-0.08458000421524048,
-0.13907702267169952,
0.0876479223370552,
0.0947323814034462,
-0.061831433326005936,
-0.008452769368886948,
0.004895468708127737,
0.13755019009113312,
-0.03773697838187218,
-0.13174059987068176,
0.011345510371029377,
-0.06858566403388977,
-0.10028401762247086,
-0.0618094727396965,
0.06277436017990112,
0.12373785674571991,
0.02303234487771988,
0.0250016450881958,
0.020958080887794495,
0.0052667343989014626,
-0.1396549791097641,
-0.06933476775884628,
0.1692645400762558,
0.011821463704109192,
0.07224257290363312,
-0.06350373476743698,
-0.1261928379535675,
-0.060596950352191925,
0.02175183594226837,
0.10319159924983978,
0.16373462975025177,
-0.06260588020086288,
0.06989205628633499,
0.08469809591770172,
-0.14647924900054932,
-0.15073134005069733,
-0.03442138060927391,
0.09728588908910751,
0.03737456351518631,
0.03593964874744415,
-0.18439321219921112,
0.06819945573806763,
0.08834658563137054,
0.0033898521214723587,
-0.04533231258392334,
-0.3093365728855133,
-0.10703980922698975,
0.04386529698967934,
0.07531942427158356,
0.03564247861504555,
-0.10586807876825333,
-0.05846872180700302,
-0.06543514877557755,
-0.04251181706786156,
0.25640934705734253,
0.0031355072278529406,
0.09846021980047226,
0.020927825942635536,
-0.07630974799394608,
0.03578653931617737,
-0.02363801933825016,
0.11309444159269333,
0.08084964752197266,
0.09372159093618393,
-0.02151012234389782,
0.007526206783950329,
0.0659996047616005,
-0.03831157833337784,
0.09619487076997757,
-0.017913738265633583,
0.08665180951356888,
-0.08355838060379028,
-0.057030245661735535,
-0.012653544545173645,
-0.00773481372743845,
-0.05056479945778847,
-0.03332303836941719,
-0.0456065833568573,
0.07295481115579605,
0.08371299505233765,
-0.018338747322559357,
0.05189519375562668,
-0.07958797365427017,
-0.0008753683650866151,
0.1818644106388092,
0.13421708345413208,
0.08429650217294693,
-0.13534115254878998,
-0.02595195733010769,
-0.007799254730343819,
0.07737915962934494,
-0.13849106431007385,
0.0631863996386528,
0.09942299872636795,
0.009675749577581882,
0.134225994348526,
0.07184412330389023,
-0.05271662026643753,
-0.009729314595460892,
0.06637807935476303,
-0.07509785890579224,
-0.08206158131361008,
-0.028724312782287598,
-0.03602799028158188,
-0.059088852256536484,
0.0010326291667297482,
0.16228578984737396,
-0.03392717242240906,
-0.01251340750604868,
0.004647531546652317,
0.03306873142719269,
-0.02317374013364315,
0.1279098093509674,
0.011217324994504452,
0.059141725301742554,
-0.09118495881557465,
0.10162673890590668,
0.07630746811628342,
-0.04885835200548172,
0.029999015852808952,
0.09039470553398132,
-0.09347172826528549,
-0.08237362653017044,
-0.07664462178945541,
0.08065587282180786,
-0.11686337739229202,
-0.012881292961537838,
-0.07209533452987671,
-0.07654780894517899,
0.013541803695261478,
0.13123466074466705,
0.08874864131212234,
0.12391982972621918,
-0.03176678717136383,
-0.01613692007958889,
-0.05146809294819832,
0.0788135975599289,
0.006876029074192047,
0.014342489652335644,
-0.08744140714406967,
0.09091003984212875,
0.031344179064035416,
-0.027464235201478004,
-0.025579849258065224,
-0.04624859616160393,
-0.13640174269676208,
-0.009153662249445915,
-0.057192154228687286,
0.04208645969629288,
-0.04431177303195,
0.018927663564682007,
0.02675640769302845,
-0.019677074626088142,
0.009271005168557167,
0.03907862678170204,
-0.016868602484464645,
-0.04957958310842514,
-0.01973636820912361,
0.09832814335823059,
-0.11968693137168884,
-0.016389477998018265,
0.006816430017352104,
-0.08819790184497833,
0.06730537116527557,
0.009467188268899918,
-0.03805196285247803,
0.01764019951224327,
-0.16711702942848206,
0.011368012987077236,
-0.0017038426594808698,
0.08446966111660004,
0.01649138517677784,
-0.11123771220445633,
0.05102556571364403,
0.030675696209073067,
0.0002710152475629002,
0.02728324383497238,
0.12810318171977997,
-0.11076090484857559,
-0.04538813605904579,
0.004113869741559029,
-0.011708800680935383,
-0.06045517325401306,
-0.017462842166423798,
0.11436296254396439,
0.07116929441690445,
0.17041409015655518,
-0.05567152425646782,
0.08613038063049316,
-0.1765887290239334,
-0.039721257984638214,
0.05763218551874161,
0.001283330493606627,
-0.03540359064936638,
-0.0763227716088295,
0.051812853664159775,
-0.020768260583281517,
0.15189769864082336,
-0.01785372383892536,
0.09527633339166641,
0.038822516798973083,
0.00433873338624835,
0.033376555889844894,
0.060359273105859756,
0.13976849615573883,
0.04243898391723633,
0.017001481726765633,
0.0028528545517474413,
-0.006474299356341362,
-0.02161298505961895,
-0.0928155854344368,
0.07097221165895462,
0.06980274617671967,
-0.03997572883963585,
0.05717054009437561,
0.06018413230776787,
0.006740938872098923,
-0.04455997422337532,
0.0706065371632576,
-0.039364349097013474,
0.02398986741900444,
-0.051083970814943314,
0.019606031477451324,
0.10718338936567307,
-0.10580746084451675,
0.10925860702991486,
0.03248471021652222,
-0.0947871059179306,
-0.11583490669727325,
-0.18595130741596222,
-0.06794553995132446,
-0.04599566012620926,
-0.011483184061944485,
-0.16359439492225647,
0.033263932913541794,
0.03364426642656326,
0.022364115342497826,
-0.010881992056965828,
0.031191887333989143,
-0.0006552398554049432,
-0.04614849016070366,
0.020301055163145065,
-0.008696205914020538,
0.027063438668847084,
0.04934412240982056,
0.02049436792731285,
0.08580917865037918,
0.06293024867773056,
0.04137958213686943,
0.04419529810547829,
0.0972491055727005,
0.009964424185454845,
-0.0365554504096508,
-0.07838965952396393,
0.022900940850377083,
0.0012353123165667057,
0.008685890585184097,
0.10989752411842346,
0.08343306183815002,
-0.04763570800423622,
-0.003344073425978422,
0.17858655750751495,
-0.06960365921258926,
-0.109405018389225,
-0.1677924394607544,
0.23614457249641418,
0.03791625425219536,
0.07592961937189102,
0.0038883064407855272,
-0.09718609601259232,
-0.03092035837471485,
0.16161982715129852,
0.11521202325820923,
-0.04700302705168724,
-0.004755087662488222,
0.04495500773191452,
-0.009700769558548927,
0.05009165778756142,
0.0812082290649414,
-0.011078368872404099,
0.24896250665187836,
-0.015392433851957321,
0.07938766479492188,
-0.03243725746870041,
-0.03668249398469925,
0.0007912452565506101,
0.09442874044179916,
-0.04320375993847847,
0.0045418692752718925,
-0.04999171569943428,
0.11236628144979477,
-0.054178789258003235,
-0.03703644871711731,
0.12046990543603897,
-0.03526090085506439,
-0.10279686003923416,
-0.008671681396663189,
0.07478763163089752,
-0.024839574471116066,
0.04309184104204178,
-0.016671204939484596,
-0.009069737046957016,
0.18809691071510315,
-0.033953920006752014,
-0.06159922853112221,
-0.09707266837358475,
-0.023511532694101334,
-0.06726019829511642,
0.08833751082420349,
0.004294530022889376,
0.1456068754196167,
0.10218770056962967,
0.034757621586322784,
-0.08349652588367462,
0.07914342731237411,
-0.02416873723268509,
-0.0700109675526619,
0.07606734335422516,
0.12370195984840393,
-0.020728524774312973,
0.17198072373867035,
0.06389904022216797,
-0.1258181631565094,
0.03290615603327751,
0.002535546664148569,
-0.0544767901301384,
-0.07193843275308609,
0.041357822716236115,
-0.08097375184297562,
0.09928510338068008,
0.16572722792625427,
-0.03551780804991722,
0.0014993146760389209,
-0.013682534918189049,
0.06789954006671906,
-0.023382846266031265,
0.05415410175919533,
-0.04138858616352081,
-0.16572260856628418,
0.030495984479784966,
0.014189633540809155,
0.03612922504544258,
-0.14817483723163605,
-0.047600679099559784,
0.01902780868113041,
-0.01818658970296383,
-0.027880078181624413,
0.09874292463064194,
0.028974974527955055,
0.042395126074552536,
-0.052695002406835556,
-0.22470088303089142,
0.0348617359995842,
0.08085587620735168,
-0.08967912197113037,
-0.10403793305158615
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-small](https://huggingface.co/microsoft/deberta-v3-small)
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 91.65
- Accuracy on MNLI mismatched set: 87.55
For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-deberta-v3-small')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-small')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-small')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-small')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-small"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-deberta-v3-small
|
[
"transformers",
"pytorch",
"deberta-v2",
"text-classification",
"microsoft/deberta-v3-small",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-small #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 91.65
- Accuracy on MNLI mismatched set: 87.55
For futher evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 91.65\r\n- Accuracy on MNLI mismatched set: 87.55\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-small #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 91.65\r\n- Accuracy on MNLI mismatched set: 87.55\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
88,
46,
49,
51,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-small #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-small## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 91.65\r\n- Accuracy on MNLI mismatched set: 87.55\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.056961070746183395,
0.1961870789527893,
-0.005332413595169783,
0.09397640079259872,
0.12969684600830078,
0.030759280547499657,
0.20943328738212585,
0.0907621681690216,
0.017969153821468353,
0.015267309732735157,
-0.017602231353521347,
0.10468098521232605,
0.06298787891864777,
0.08545204997062683,
-0.03651503100991249,
-0.17780393362045288,
0.04796379432082176,
-0.03407283499836922,
0.03159286826848984,
0.09035217016935349,
0.10395898669958115,
-0.04964560642838478,
0.03876478970050812,
-0.011030765250325203,
-0.1415451467037201,
0.03361905366182327,
0.0051561384461820126,
-0.0020623458549380302,
0.053033627569675446,
0.09050386399030685,
0.13333773612976074,
0.02632060833275318,
0.05859701335430145,
-0.17763419449329376,
0.009237413294613361,
0.07520344108343124,
-0.013635381124913692,
0.03653927147388458,
0.0638813003897667,
-0.052451565861701965,
0.004688677377998829,
-0.02969725802540779,
0.05727693811058998,
0.0856747031211853,
-0.07996667921543121,
-0.12344317138195038,
-0.06564246118068695,
0.02343943901360035,
0.07945317775011063,
0.09610721468925476,
-0.03643164783716202,
0.10329966992139816,
-0.07307448983192444,
0.09137709438800812,
0.07578229159116745,
-0.22034852206707,
-0.02919360250234604,
0.12271309643983841,
-0.01669551245868206,
0.0544452928006649,
-0.09719231724739075,
-0.03607233986258507,
0.019358912482857704,
0.0038723223842680454,
0.01262687798589468,
-0.025430399924516678,
-0.006691235117614269,
-0.026973122730851173,
-0.11965502053499222,
-0.00014310423284769058,
0.13069888949394226,
-0.0028886974323540926,
-0.07842600345611572,
-0.10936353355646133,
-0.05831489711999893,
-0.022318540140986443,
0.016667451709508896,
-0.016764741390943527,
0.04643956199288368,
0.03739266097545624,
0.002302109729498625,
-0.032362136989831924,
-0.07539698481559753,
-0.04761741682887077,
-0.06762035936117172,
0.09118301421403885,
0.021787207573652267,
0.028133036568760872,
-0.07599619776010513,
0.09970089793205261,
0.005289139691740274,
-0.1182696744799614,
-0.0789690762758255,
-0.054308827966451645,
-0.21623556315898895,
-0.022545037791132927,
-0.10969150811433792,
-0.11233719438314438,
0.032436005771160126,
0.20078672468662262,
0.00407967297360301,
0.063546322286129,
-0.0010119822109118104,
0.023612026125192642,
0.06472960859537125,
0.1400364637374878,
-0.033656924962997437,
-0.010959330014884472,
-0.024638643488287926,
0.030695592984557152,
-0.0041329036466777325,
-0.024394407868385315,
-0.06526073068380356,
-0.024274786934256554,
0.007041584700345993,
0.05273788794875145,
-0.000617657438851893,
0.023635249584913254,
-0.06414564698934555,
-0.04882162809371948,
0.0041771549731493,
-0.14638164639472961,
0.034152802079916,
0.027090927585959435,
-0.04794420674443245,
0.13472896814346313,
0.051494523882865906,
-0.021505294367671013,
-0.07091271877288818,
0.0069885896518826485,
-0.08569799363613129,
0.04107192903757095,
-0.11503090709447861,
-0.13300921022891998,
0.027024706825613976,
0.00036001738044433296,
-0.04776831343770027,
-0.0877830907702446,
-0.11901956051588058,
-0.068480484187603,
0.036347195506095886,
-0.041209787130355835,
0.04326841980218887,
-0.08692765235900879,
-0.021107451990246773,
0.02897862158715725,
0.02264774963259697,
-0.06347545981407166,
-0.005577965173870325,
0.0141081428155303,
-0.0005748371477238834,
0.05079640820622444,
-0.017347725108265877,
0.014431588351726532,
-0.12141559273004532,
-0.023558735847473145,
-0.014693436212837696,
0.17323268949985504,
-0.08072654157876968,
-0.00003345928780618124,
-0.08439966291189194,
-0.03274482861161232,
0.0396089106798172,
0.01695961505174637,
0.023262349888682365,
0.11895835399627686,
-0.23950646817684174,
-0.021843325346708298,
0.18407565355300903,
-0.15927079319953918,
-0.03525390475988388,
0.0889754593372345,
-0.05120646953582764,
-0.009058183059096336,
0.09082705527544022,
0.060031957924366,
0.19853036105632782,
-0.0685625895857811,
-0.046013642102479935,
0.05430838465690613,
-0.025606682524085045,
0.06836490333080292,
0.07663080096244812,
0.03238728269934654,
-0.03465449437499046,
0.01956859417259693,
-0.08739784359931946,
0.0023523413110524416,
-0.03580824285745621,
-0.06834175437688828,
-0.03907546401023865,
-0.0403694212436676,
0.018120774999260902,
0.008914916776120663,
0.011326510459184647,
-0.009339330717921257,
-0.10252520442008972,
0.1758558452129364,
0.10552716255187988,
-0.0841434970498085,
0.019238483160734177,
-0.11075163632631302,
-0.06033647805452347,
-0.0031039307359606028,
0.004072209820151329,
-0.1965838223695755,
-0.11076483875513077,
0.024568570777773857,
-0.1334027647972107,
0.04687068611383438,
0.051188234239816666,
0.012907381169497967,
0.04356514289975166,
-0.02970122918486595,
-0.004830125253647566,
-0.024140704423189163,
-0.010315688326954842,
-0.049343518912792206,
-0.13780340552330017,
-0.06606730818748474,
-0.023271184414625168,
0.09432083368301392,
-0.1415734440088272,
0.04146553575992584,
0.026056276634335518,
0.0861760675907135,
-0.004940844606608152,
-0.06198602914810181,
0.0006747650913894176,
0.03693617880344391,
-0.043233055621385574,
-0.044995155185461044,
0.02960134483873844,
0.04101593792438507,
-0.08540904521942139,
0.0017956276424229145,
-0.11070018261671066,
-0.05321377515792847,
0.07073202729225159,
0.028861263766884804,
-0.08048144727945328,
-0.0030741558875888586,
-0.012171059846878052,
-0.004922389984130859,
-0.07879284769296646,
-0.0442563071846962,
0.10113686323165894,
0.029502592980861664,
0.10207447409629822,
-0.0669952854514122,
-0.043319184333086014,
-0.031170563772320747,
-0.03509499877691269,
0.0033279391936957836,
0.09502648562192917,
-0.048851050436496735,
-0.21788381040096283,
0.10462065786123276,
0.000007496769285353366,
-0.08834182471036911,
0.11559999734163284,
-0.0033718696795403957,
-0.03926501050591469,
-0.09929623454809189,
0.0036536515690386295,
0.015375284478068352,
0.03322527930140495,
-0.03470364585518837,
0.055137235671281815,
0.033362049609422684,
0.026268688961863518,
-0.00538782449439168,
-0.11924227327108383,
0.022927885875105858,
0.04354250803589821,
-0.0400046780705452,
-0.016311755403876305,
0.0017426759004592896,
0.01467554084956646,
0.07017359882593155,
-0.02757679671049118,
0.04390476271510124,
0.022832009941339493,
-0.0770452469587326,
-0.14205944538116455,
0.22206856310367584,
-0.1199897825717926,
-0.2605627179145813,
-0.1535775065422058,
0.0587504543364048,
-0.06399208307266235,
-0.0035633051302284002,
0.03243519738316536,
-0.06811096519231796,
-0.044749531894922256,
-0.08056598156690598,
-0.04170610383152962,
-0.06297550350427628,
-0.026499466970562935,
-0.04852226376533508,
0.0514201857149601,
0.060543373227119446,
-0.15224848687648773,
0.020836113020777702,
0.014202283695340157,
-0.09774819761514664,
0.022519497200846672,
0.025742901489138603,
0.020679831504821777,
0.13922634720802307,
-0.0005475696525536478,
0.020450416952371597,
-0.0017292104894295335,
0.2370370626449585,
-0.021612582728266716,
0.003287105355411768,
0.1335442066192627,
-0.019259387627243996,
0.06566687673330307,
0.08157089352607727,
0.043811701238155365,
-0.06757771968841553,
0.03084348328411579,
0.05638407915830612,
-0.04038724675774574,
-0.23690764605998993,
-0.02760658785700798,
-0.04851305112242699,
-0.05192003399133682,
0.052688419818878174,
0.014444472268223763,
0.006107872352004051,
0.04240432754158974,
-0.038990601897239685,
-0.010389072820544243,
0.06150193512439728,
0.08314985036849976,
0.11515584588050842,
0.036032963544130325,
0.11449538171291351,
-0.06641249358654022,
0.02337203361093998,
0.05747813358902931,
-0.0010139275109395385,
0.14964117109775543,
0.013843987137079239,
0.06882432848215103,
0.10457455366849899,
0.019567999988794327,
0.037688400596380234,
0.023961113765835762,
0.0010024991352111101,
0.028486931696534157,
0.02710648439824581,
-0.08003382384777069,
-0.023219913244247437,
0.03872571885585785,
-0.039589010179042816,
0.02129344455897808,
-0.054115258157253265,
-0.03512758016586304,
0.06833846867084503,
0.12211459130048752,
0.08409421890974045,
-0.2301958054304123,
-0.0886494368314743,
0.030209293588995934,
-0.036933816969394684,
-0.030981486663222313,
-0.0060872784815728664,
0.08728189021348953,
-0.09017068147659302,
0.07668162137269974,
-0.03618237376213074,
0.10074330121278763,
-0.10408969223499298,
-0.014148456044495106,
0.03471336513757706,
0.07646972686052322,
0.014898359775543213,
0.08425787836313248,
-0.2505313456058502,
0.07498232275247574,
0.008051376789808273,
0.06272706389427185,
0.0032530322205275297,
0.033703722059726715,
0.03904540091753006,
0.04015763849020004,
0.09559804201126099,
0.02999647706747055,
-0.03591477498412132,
-0.10999628156423569,
-0.0336206778883934,
0.002025442197918892,
0.09012547880411148,
-0.0427987277507782,
0.10987574607133865,
-0.022830365225672722,
-0.0014936849474906921,
-0.0023441307712346315,
0.04861946776509285,
-0.2045290768146515,
-0.1693422794342041,
0.056700605899095535,
0.029933782294392586,
0.006762049160897732,
-0.035537052899599075,
-0.0720948651432991,
0.005997186526656151,
0.1846836507320404,
-0.23794373869895935,
-0.08160409331321716,
-0.14408522844314575,
0.09477245807647705,
0.09690804034471512,
-0.06397052854299545,
0.00017520775145385414,
-0.013649235479533672,
0.1489323377609253,
-0.02142726257443428,
-0.1405046433210373,
0.013837188482284546,
-0.06044485419988632,
-0.10533513873815536,
-0.054327525198459625,
0.0592300184071064,
0.11367049813270569,
0.033552054315805435,
0.02613644115626812,
0.014130808413028717,
0.014584802091121674,
-0.1273113191127777,
-0.05459503456950188,
0.15321394801139832,
0.002254319842904806,
0.10390106588602066,
-0.05416392907500267,
-0.11052535474300385,
-0.04752272367477417,
0.025736380368471146,
0.10757150501012802,
0.16233095526695251,
-0.0800897628068924,
0.04041724652051926,
0.07141215354204178,
-0.1516539603471756,
-0.15825630724430084,
-0.00425605708733201,
0.09616869688034058,
0.037083376199007034,
0.03072933293879032,
-0.18486237525939941,
0.07317405939102173,
0.10363899916410446,
-0.004574956838041544,
-0.07986216247081757,
-0.34693461656570435,
-0.10520916432142258,
0.04198843240737915,
0.08313769847154617,
0.060070645064115524,
-0.10958565771579742,
-0.05167347937822342,
-0.06506418436765671,
-0.08158134669065475,
0.2552858293056488,
-0.011383550241589546,
0.1138865202665329,
0.004456536844372749,
-0.06698188185691833,
0.03861719369888306,
-0.028944261372089386,
0.11940091103315353,
0.06652441620826721,
0.10273183882236481,
-0.02294248528778553,
0.022330382838845253,
0.07942374795675278,
-0.05146781727671623,
0.10527573525905609,
-0.004938979633152485,
0.07758183032274246,
-0.10695798695087433,
-0.03981383144855499,
-0.033583950251340866,
0.0010570435551926494,
-0.04799271374940872,
-0.023285062983632088,
-0.06632781028747559,
0.07830964773893356,
0.09005264937877655,
-0.02042083442211151,
0.05805366858839989,
-0.0709240511059761,
0.04380493611097336,
0.1671580672264099,
0.11670037358999252,
0.09546897560358047,
-0.15661010146141052,
-0.014498712494969368,
-0.0010534921893849969,
0.07967856526374817,
-0.132485032081604,
0.06265733391046524,
0.11045993119478226,
0.027785461395978928,
0.1395520716905594,
0.07954127341508865,
-0.044443730264902115,
-0.007938376627862453,
0.052884820848703384,
-0.060271210968494415,
-0.11546752601861954,
-0.029542604461312294,
-0.019196825101971626,
-0.061455484479665756,
-0.00043220765655860305,
0.14659729599952698,
-0.05325944349169731,
-0.001288223429583013,
0.01698659174144268,
0.026382988318800926,
-0.02045580744743347,
0.13848403096199036,
0.0190631952136755,
0.06719701737165451,
-0.08813715726137161,
0.12404534220695496,
0.08111557364463806,
-0.05381942167878151,
0.03241436555981636,
0.08617100864648819,
-0.09342825412750244,
-0.061714768409729004,
-0.05896897614002228,
0.07398626953363419,
-0.12855491042137146,
-0.028941912576556206,
-0.08085085451602936,
-0.09971784800291061,
0.028459830209612846,
0.13796161115169525,
0.08680640161037445,
0.12122590094804764,
-0.03262415528297424,
-0.014813854359090328,
-0.06274110078811646,
0.0715794563293457,
0.030078357085585594,
0.02865665592253208,
-0.10288037359714508,
0.10995189100503922,
0.030523695051670074,
-0.01837310567498207,
-0.024026336148381233,
-0.049234695732593536,
-0.14208920300006866,
0.0006098106387071311,
-0.0974268987774849,
0.03894205391407013,
-0.03128518909215927,
0.004905003122985363,
0.016475951299071312,
-0.023368461057543755,
0.006408669985830784,
0.04357844591140747,
-0.02703693136572838,
-0.047953732311725616,
-0.021202903240919113,
0.08244408667087555,
-0.12416809797286987,
-0.01840076968073845,
0.011747940443456173,
-0.08701442182064056,
0.05822761356830597,
0.027990076690912247,
-0.04509776458144188,
0.004477526526898146,
-0.16352279484272003,
-0.00909185316413641,
-0.003301572985947132,
0.059549976140260696,
0.012296210043132305,
-0.1166386753320694,
0.053857821971178055,
0.02099844254553318,
0.014055428095161915,
0.03278478980064392,
0.12304199486970901,
-0.11628343909978867,
-0.056314315646886826,
-0.016660217195749283,
0.00044560778769664466,
-0.07505793869495392,
0.015817973762750626,
0.13673697412014008,
0.08051448315382004,
0.16762056946754456,
-0.06189186871051788,
0.07298401743173599,
-0.19466251134872437,
-0.03820746764540672,
0.050433285534381866,
-0.01597246713936329,
-0.03462592139840126,
-0.062134359031915665,
0.05127105861902237,
-0.014583575539290905,
0.11440566927194595,
-0.021732686087489128,
0.08002149313688278,
0.05129130929708481,
0.000621872313786298,
0.011416231282055378,
0.04960817098617554,
0.14353249967098236,
0.04175664857029915,
-0.0038538945373147726,
0.053394611924886703,
-0.017675647512078285,
-0.004553397186100483,
-0.05275626853108406,
0.09570778906345367,
0.08556406199932098,
-0.04584071412682533,
0.07026151567697525,
0.04271816462278366,
0.011462559923529625,
-0.059924717992544174,
0.06624244153499603,
-0.030110515654087067,
0.049340542405843735,
-0.054501358419656754,
0.04801312834024429,
0.12171947211027145,
-0.11846218258142471,
0.12039345502853394,
0.015215769410133362,
-0.09585969150066376,
-0.12975963950157166,
-0.17491841316223145,
-0.0667714923620224,
-0.06470268219709396,
-0.012252843007445335,
-0.15192881226539612,
0.05179985985159874,
0.0718541368842125,
0.03516662120819092,
-0.010590044781565666,
0.04866155609488487,
-0.008129975758492947,
-0.05039401724934578,
0.04115501046180725,
-0.0190975870937109,
0.02007700875401497,
0.038339391350746155,
0.009153026156127453,
0.11093249917030334,
0.055144380778074265,
0.05472366511821747,
0.05135929211974144,
0.0981246754527092,
0.008300126530230045,
-0.034893184900283813,
-0.0797475129365921,
0.02014467678964138,
0.008098161779344082,
0.015479384921491146,
0.09278547763824463,
0.07891716808080673,
-0.03576907142996788,
-0.016950344666838646,
0.2205762267112732,
-0.08238531649112701,
-0.09898889064788818,
-0.16472327709197998,
0.24246849119663239,
0.034275513142347336,
0.056518424302339554,
0.015396383590996265,
-0.09635913372039795,
-0.032805293798446655,
0.14195358753204346,
0.09264257550239563,
-0.03852815926074982,
-0.011131549254059792,
0.028302785009145737,
-0.019546370953321457,
0.04360827058553696,
0.09876316040754318,
0.0009901679586619139,
0.21663962304592133,
-0.032520487904548645,
0.10163549333810806,
-0.026748700067400932,
-0.039204709231853485,
-0.01982319913804531,
0.09718060493469238,
-0.039680689573287964,
0.011060677468776703,
-0.06099942699074745,
0.10252456367015839,
-0.0558469295501709,
-0.05870893970131874,
0.11199069023132324,
-0.03100733831524849,
-0.10865721106529236,
-0.022357655689120293,
0.05840349197387695,
-0.028117772191762924,
0.0430804044008255,
-0.019487028941512108,
0.016798168420791626,
0.1654404103755951,
-0.02948765829205513,
-0.06529358774423599,
-0.12318524718284607,
0.001640239148400724,
-0.07495937496423721,
0.10043440759181976,
0.013007666915655136,
0.12557460367679596,
0.10131639987230301,
0.017857175320386887,
-0.09460242092609406,
0.07473164796829224,
-0.016450976952910423,
-0.05922538787126541,
0.05306198075413704,
0.13161322474479675,
-0.014576971530914307,
0.1667606383562088,
0.04486963897943497,
-0.12680228054523468,
0.02564382553100586,
0.0007472023135051131,
-0.053585659712553024,
-0.08150466531515121,
0.02220173180103302,
-0.061192091554403305,
0.10534800589084625,
0.19582420587539673,
-0.03958282992243767,
0.011014478281140327,
-0.02779238112270832,
0.0411464124917984,
-0.009685077704489231,
0.03688124567270279,
-0.017235610634088516,
-0.16092664003372192,
0.04031168669462204,
0.012002391740679741,
0.04712655395269394,
-0.15618866682052612,
-0.051108330488204956,
0.012334921397268772,
-0.03568067401647568,
-0.02219984494149685,
0.11052946746349335,
0.03266215696930885,
0.04590963199734688,
-0.04853643476963043,
-0.2415795475244522,
0.02950061485171318,
0.10446979850530624,
-0.09330163151025772,
-0.09403061866760254
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class. This model is based on [microsoft/deberta-v3-xsmall](https://huggingface.co/microsoft/deberta-v3-xsmall)
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 91.64
- Accuracy on MNLI mismatched set: 87.77
For futher evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-deberta-v3-xsmall')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-deberta-v3-xsmall')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-deberta-v3-xsmall')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-deberta-v3-xsmall')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["microsoft/deberta-v3-xsmall"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-deberta-v3-xsmall
|
[
"transformers",
"pytorch",
"deberta-v2",
"text-classification",
"microsoft/deberta-v3-xsmall",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-xsmall #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
- Accuracy on SNLI-test dataset: 91.64
- Accuracy on MNLI mismatched set: 87.77
For futher evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 91.64\r\n- Accuracy on MNLI mismatched set: 87.77\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-xsmall #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\r\n- Accuracy on SNLI-test dataset: 91.64\r\n- Accuracy on MNLI mismatched set: 87.77\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
89,
47,
49,
51,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #deberta-v2 #text-classification #microsoft/deberta-v3-xsmall #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class. This model is based on microsoft/deberta-v3-xsmall## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\r\n- Accuracy on SNLI-test dataset: 91.64\r\n- Accuracy on MNLI mismatched set: 87.77\r\n\nFor futher evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.05526004731655121,
0.20002509653568268,
-0.004549776203930378,
0.09243170917034149,
0.1255296766757965,
0.04860931262373924,
0.2057001292705536,
0.10026445984840393,
0.023124130442738533,
0.010743226855993271,
0.0033932949882000685,
0.10624440014362335,
0.0735110193490982,
0.06944567710161209,
-0.03670136258006096,
-0.15651534497737885,
0.06273180991411209,
-0.031661514192819595,
0.03551706671714783,
0.07941190898418427,
0.09823884814977646,
-0.03796064481139183,
0.027161527425050735,
-0.011348406784236431,
-0.16056472063064575,
0.03263018652796745,
0.013877912424504757,
0.008690397255122662,
0.04654226079583168,
0.07941348105669022,
0.134487122297287,
0.022073298692703247,
0.054972339421510696,
-0.15428295731544495,
0.0071214670315384865,
0.08016258478164673,
-0.008101080544292927,
0.040968116372823715,
0.06426200270652771,
-0.042856525629758835,
-0.013149209320545197,
-0.021201087161898613,
0.049097828567028046,
0.09118204563856125,
-0.08311184495687485,
-0.12524905800819397,
-0.07852311432361603,
0.041662704199552536,
0.0712871104478836,
0.10549604147672653,
-0.034251656383275986,
0.11924458295106888,
-0.04591577500104904,
0.07998860627412796,
0.07225118577480316,
-0.2156272679567337,
-0.02652299962937832,
0.1276988387107849,
-0.009356200695037842,
0.03971582651138306,
-0.0813535749912262,
-0.03941688686609268,
0.021172143518924713,
0.009966623038053513,
-0.00391641678288579,
-0.0237180944532156,
0.006062181666493416,
-0.029210276901721954,
-0.10915707796812057,
0.006141735706478357,
0.12676866352558136,
0.00768245430663228,
-0.07872988283634186,
-0.10875414311885834,
-0.058593858033418655,
-0.03962946683168411,
0.03087957203388214,
-0.024735471233725548,
0.032105784863233566,
0.0497613251209259,
-0.005708195734769106,
-0.05533894896507263,
-0.06467992067337036,
-0.039844073355197906,
-0.056803010404109955,
0.09126632660627365,
0.023043472319841385,
0.026669776067137718,
-0.07527591288089752,
0.09551551938056946,
-0.0213076863437891,
-0.11828793585300446,
-0.07102097570896149,
-0.06846754252910614,
-0.2110780030488968,
-0.01225515641272068,
-0.11614567041397095,
-0.1259959638118744,
0.032758407294750214,
0.17687439918518066,
0.007095970679074526,
0.06652242690324783,
-0.002010607160627842,
0.028787173330783844,
0.05354767292737961,
0.13736560940742493,
-0.060415077954530716,
-0.002238245913758874,
-0.021257642656564713,
0.05231703072786331,
-0.007573797833174467,
-0.037986647337675095,
-0.061677105724811554,
-0.04001437500119209,
0.02109338343143463,
0.04897123947739601,
0.02537170611321926,
0.014566423371434212,
-0.07527897506952286,
-0.05470074713230133,
-0.019534463062882423,
-0.13625620305538177,
0.040359847247600555,
0.010849490761756897,
-0.043869972229003906,
0.13095544278621674,
0.056962668895721436,
-0.011650481261312962,
-0.07216348499059677,
-0.008361739106476307,
-0.09025966376066208,
0.028068073093891144,
-0.12218091636896133,
-0.13411946594715118,
0.017059730365872383,
0.0033256823662668467,
-0.050030454993247986,
-0.08235623687505722,
-0.11587823182344437,
-0.06394140422344208,
0.0480373352766037,
-0.03380927816033363,
0.03946206346154213,
-0.08020124584436417,
-0.043792419135570526,
0.02135917916893959,
0.030846720561385155,
-0.08625948429107666,
0.006595269776880741,
0.006464182864874601,
-0.004632262513041496,
0.03593510761857033,
-0.037830982357263565,
0.011226363480091095,
-0.1057412400841713,
-0.01779843121767044,
-0.01367349922657013,
0.17019301652908325,
-0.07706235349178314,
-0.03032475896179676,
-0.08232045918703079,
-0.028791768476366997,
0.038733985275030136,
0.002514935564249754,
0.04955456778407097,
0.1231609359383583,
-0.25279495120048523,
-0.016558513045310974,
0.17411406338214874,
-0.1440533995628357,
-0.02526102215051651,
0.08771003037691116,
-0.054725196212530136,
-0.014178600162267685,
0.07762700319290161,
0.04953781142830849,
0.18506020307540894,
-0.0654539242386818,
-0.0630742758512497,
0.028384022414684296,
-0.039159536361694336,
0.08868248760700226,
0.05894562229514122,
0.024817630648612976,
-0.05215814337134361,
0.01748264580965042,
-0.07780946791172028,
0.004641425330191851,
-0.04488604515790939,
-0.06902945041656494,
-0.056538451462984085,
-0.04126434028148651,
0.027539560571312904,
0.0118724899366498,
0.0017194912070408463,
-0.010538462549448013,
-0.11391661316156387,
0.1652051955461502,
0.10532593727111816,
-0.08411567658185959,
0.017406493425369263,
-0.11201956123113632,
-0.056013334542512894,
-0.014112266711890697,
-0.003420748980715871,
-0.18552462756633759,
-0.12056156992912292,
0.03247095271945,
-0.13150183856487274,
0.048213325440883636,
0.05443410947918892,
0.011858492158353329,
0.041721902787685394,
-0.04069984331727028,
-0.016422860324382782,
-0.03822660818696022,
0.00253804842941463,
-0.053086087107658386,
-0.145975723862648,
-0.0744340792298317,
-0.013810910284519196,
0.11950114369392395,
-0.1476273387670517,
0.03934638574719429,
-0.008963657543063164,
0.10053218901157379,
-0.02334304340183735,
-0.0712960809469223,
-0.0078027271665632725,
0.03558935597538948,
-0.044864337891340256,
-0.05433737486600876,
0.02244669385254383,
0.047151777893304825,
-0.10827724635601044,
-0.012608900666236877,
-0.12684570252895355,
-0.07478852570056915,
0.07438035309314728,
0.03776266798377037,
-0.09108209609985352,
0.023321187123656273,
-0.003115458181127906,
-0.017943192273378372,
-0.07709351181983948,
-0.05175177380442619,
0.09172356873750687,
0.03298046439886093,
0.09771651774644852,
-0.0593714602291584,
-0.025719748809933662,
-0.02357889898121357,
-0.04015195742249489,
-0.010305040515959263,
0.08047933876514435,
-0.06983973830938339,
-0.2463240772485733,
0.10257625579833984,
-0.009722452610731125,
-0.0948178693652153,
0.12154147028923035,
0.007542720530182123,
-0.024805154651403427,
-0.11668193340301514,
-0.0011946127051487565,
0.024205509573221207,
0.03602546453475952,
-0.013188350014388561,
0.07781384885311127,
0.03337604179978371,
0.021276824176311493,
-0.004435725975781679,
-0.10908447206020355,
0.021862894296646118,
0.033628687262535095,
-0.032307546585798264,
-0.008865191601216793,
0.0007872581481933594,
0.021584955975413322,
0.06619110703468323,
-0.024662980809807777,
0.04763133451342583,
0.03605271130800247,
-0.06721260398626328,
-0.13504284620285034,
0.2060798853635788,
-0.12730136513710022,
-0.22790519893169403,
-0.16253462433815002,
0.08094564825296402,
-0.08800404518842697,
-0.007628193125128746,
0.02798990160226822,
-0.05362257733941078,
-0.03763335198163986,
-0.06899643689393997,
-0.05992531031370163,
-0.08606456965208054,
-0.03155779838562012,
-0.046935420483350754,
0.038899872452020645,
0.07190898060798645,
-0.15101571381092072,
0.014210899360477924,
0.02578127197921276,
-0.09533428400754929,
0.020432092249393463,
0.020389767363667488,
0.009096344001591206,
0.12877605855464935,
-0.012108997441828251,
0.019529975950717926,
0.008782971650362015,
0.24189293384552002,
-0.02311779372394085,
0.013831663876771927,
0.1305152177810669,
-0.017054205760359764,
0.07600722461938858,
0.0714179053902626,
0.036474671214818954,
-0.05457156151533127,
0.024647213518619537,
0.05605526268482208,
-0.02884596399962902,
-0.22931574285030365,
-0.020762525498867035,
-0.04652857035398483,
-0.06584543734788895,
0.047503963112831116,
0.005538444500416517,
0.029352053999900818,
0.04730710759758949,
-0.04970710352063179,
-0.009375541470944881,
0.05643647909164429,
0.09424328804016113,
0.09038220345973969,
0.04567413777112961,
0.11692193150520325,
-0.056956756860017776,
0.03082604520022869,
0.05201992392539978,
-0.007143000606447458,
0.1600838154554367,
0.013410461135208607,
0.04721095785498619,
0.10276498645544052,
0.03679201751947403,
0.0310200247913599,
0.014912590384483337,
0.00809069350361824,
0.028331810608506203,
0.02830437757074833,
-0.07428167015314102,
-0.02321471832692623,
0.04102614149451256,
-0.016134409233927727,
0.042444996535778046,
-0.0498187318444252,
-0.037808772176504135,
0.05511840060353279,
0.13065864145755768,
0.08739162236452103,
-0.2263941913843155,
-0.08992251008749008,
0.02907826192677021,
-0.02196105755865574,
-0.05251423269510269,
-0.024110721424221992,
0.0708085373044014,
-0.09102767705917358,
0.07846494764089584,
-0.04659658670425415,
0.09251465648412704,
-0.09061575680971146,
-0.008865991607308388,
0.019277319312095642,
0.09133949875831604,
0.019473060965538025,
0.08203797042369843,
-0.24349677562713623,
0.06946094334125519,
0.01937716081738472,
0.07636678218841553,
-0.00872777309268713,
0.030601568520069122,
0.03617939352989197,
0.010129506699740887,
0.08710088580846786,
0.041659388691186905,
-0.027197865769267082,
-0.1150759756565094,
-0.023102223873138428,
-0.007418039720505476,
0.09859456866979599,
-0.017858855426311493,
0.11254573613405228,
-0.02774149365723133,
0.001497284509241581,
0.014587260782718658,
0.09557417035102844,
-0.2257792353630066,
-0.17292897403240204,
0.07721123844385147,
0.05216822773218155,
0.005891559179872274,
-0.03831576928496361,
-0.06998211145401001,
0.024822402745485306,
0.19740532338619232,
-0.2103462517261505,
-0.07233606278896332,
-0.13952100276947021,
0.09111177921295166,
0.12416928261518478,
-0.05513986572623253,
0.00009735400089994073,
-0.02422177419066429,
0.14852957427501678,
-0.031116874888539314,
-0.14463263750076294,
0.01451297476887703,
-0.06030525267124176,
-0.11267613619565964,
-0.05059532821178436,
0.04382174462080002,
0.13375182449817657,
0.03136204183101654,
0.028089629486203194,
0.019533617421984673,
0.009650579653680325,
-0.1235002875328064,
-0.03506326675415039,
0.15866219997406006,
0.02247747592628002,
0.10125023126602173,
-0.048202112317085266,
-0.09401606768369675,
-0.04505863040685654,
0.014439413323998451,
0.08238668739795685,
0.1870661973953247,
-0.07968001812696457,
0.026350706815719604,
0.06339296698570251,
-0.15366311371326447,
-0.13537529110908508,
-0.01084535475820303,
0.10304450243711472,
0.04391127824783325,
0.02795068547129631,
-0.16729827225208282,
0.04932452738285065,
0.10431630909442902,
-0.005228186491876841,
-0.08625398576259613,
-0.331917405128479,
-0.09810144454240799,
0.04879440367221832,
0.07571657001972198,
0.0806494727730751,
-0.10175281763076782,
-0.046115290373563766,
-0.07832036912441254,
-0.0973736122250557,
0.2644710838794708,
-0.0183671023696661,
0.12955538928508759,
-0.002068682573735714,
-0.07628535479307175,
0.038372885435819626,
-0.03725653514266014,
0.1319662630558014,
0.05388110503554344,
0.09742793440818787,
-0.013299467042088509,
0.02761622704565525,
0.0932011678814888,
-0.060825806111097336,
0.10788195580244064,
0.012473334558308125,
0.08270631730556488,
-0.10545656085014343,
-0.03285276144742966,
-0.051310889422893524,
0.01307655405253172,
-0.05470138043165207,
-0.023084498941898346,
-0.05630824714899063,
0.08486324548721313,
0.09255378693342209,
-0.021846063435077667,
0.049153704196214676,
-0.0644868016242981,
0.06753507256507874,
0.1795922964811325,
0.11086799949407578,
0.10491730272769928,
-0.1357363611459732,
-0.00796203501522541,
-0.015356073155999184,
0.07556792348623276,
-0.08878416568040848,
0.061512693762779236,
0.10372717678546906,
0.01874486543238163,
0.13325223326683044,
0.08214081823825836,
-0.06845542788505554,
-0.013451945036649704,
0.03447776287794113,
-0.06679690629243851,
-0.12127379328012466,
-0.03262878581881523,
0.009781312197446823,
-0.04673761501908302,
-0.008575957268476486,
0.15609432756900787,
-0.053851038217544556,
-0.01126149669289589,
0.01770956814289093,
0.017896676436066628,
-0.030258458107709885,
0.12499350309371948,
0.03066054917871952,
0.06378744542598724,
-0.07536640018224716,
0.13217280805110931,
0.07460890710353851,
-0.041762061417102814,
0.06199248135089874,
0.07489340752363205,
-0.0834018811583519,
-0.052167411893606186,
-0.05468985065817833,
0.0977640375494957,
-0.14572970569133759,
-0.03307158499956131,
-0.0768420472741127,
-0.08394142985343933,
0.016596071422100067,
0.13418643176555634,
0.06883121281862259,
0.11296132951974869,
-0.027256835252046585,
-0.0220254547894001,
-0.08510271459817886,
0.05331804230809212,
0.035991642624139786,
0.03789135441184044,
-0.10309403389692307,
0.10163548588752747,
0.016543392091989517,
-0.017702657729387283,
-0.019248230382800102,
-0.03800003230571747,
-0.15475058555603027,
-0.013253603130578995,
-0.09909892827272415,
0.017674067988991737,
-0.02689705230295658,
-0.011389706283807755,
0.019581230357289314,
-0.030131153762340546,
0.015352029353380203,
0.04410799220204353,
-0.01754850707948208,
-0.06508079916238785,
-0.016204819083213806,
0.07038749009370804,
-0.11688005924224854,
-0.020178450271487236,
0.012178920209407806,
-0.08747567236423492,
0.04597383737564087,
0.04459499195218086,
-0.04297042265534401,
-0.005474344361573458,
-0.14277306199073792,
-0.008529802784323692,
-0.025870036333799362,
0.05463242530822754,
0.016924383118748665,
-0.14239075779914856,
0.04644409939646721,
0.011865769512951374,
0.01711282506585121,
0.03389603644609451,
0.13610851764678955,
-0.11131532490253448,
-0.08498310297727585,
-0.009332544170320034,
0.0073852394707500935,
-0.08044186234474182,
0.027333354577422142,
0.15503734350204468,
0.0781305581331253,
0.1578414887189865,
-0.05805593356490135,
0.0735042467713356,
-0.1932593137025833,
-0.03789258003234863,
0.04128594323992729,
-0.01986934058368206,
0.026910986751317978,
-0.052011556923389435,
0.05672163516283035,
-0.017574578523635864,
0.0958341732621193,
-0.045539963990449905,
0.07110320031642914,
0.04999064281582832,
-0.01866411790251732,
-0.004324854351580143,
0.03556132689118385,
0.1433001607656479,
0.03574147820472717,
-0.00927751138806343,
0.05565795674920082,
-0.015413232147693634,
-0.011262943968176842,
-0.030906792730093002,
0.08786176145076752,
0.06970210373401642,
-0.04754186049103737,
0.06267664581537247,
0.04087492451071739,
-0.001255349605344236,
-0.041630636900663376,
0.0709548071026802,
-0.0212703887373209,
0.052719008177518845,
-0.04906558245420456,
0.03644232824444771,
0.12547452747821808,
-0.12472984194755554,
0.12890584766864777,
-0.004883239511400461,
-0.10244381427764893,
-0.12320643663406372,
-0.1686878800392151,
-0.06349381804466248,
-0.05562512204051018,
-0.0075285779312253,
-0.1425994634628296,
0.07038996368646622,
0.07254988700151443,
0.030373357236385345,
-0.015587267465889454,
0.04138612002134323,
0.004951649811118841,
-0.04910730943083763,
0.03400575742125511,
-0.02207789197564125,
0.01334572583436966,
0.05089576542377472,
-0.0008714591385796666,
0.11947233229875565,
0.06427314877510071,
0.05739249289035797,
0.05161244049668312,
0.098675936460495,
0.010708424262702465,
-0.016627291217446327,
-0.09647484123706818,
0.02619153819978237,
-0.0020465936977416277,
0.021808089688420296,
0.09357680380344391,
0.06404224783182144,
-0.026813378557562828,
-0.024197036400437355,
0.21261854469776154,
-0.09218671917915344,
-0.10322973132133484,
-0.15621735155582428,
0.22996847331523895,
0.017834262922406197,
0.04141052067279816,
0.011291940696537495,
-0.10143448412418365,
-0.02846521884202957,
0.12315940111875534,
0.0824773907661438,
-0.046393368393182755,
-0.014142340049147606,
0.041037920862436295,
-0.028326867148280144,
0.02931460738182068,
0.08926624059677124,
-0.001772263552993536,
0.18433497846126556,
-0.04629528895020485,
0.127704456448555,
-0.032595861703157425,
-0.02758432924747467,
-0.03689383715391159,
0.1284128874540329,
-0.039467182010412216,
0.007822422310709953,
-0.04500308260321617,
0.08230483531951904,
-0.03602297604084015,
-0.051367033272981644,
0.10093633085489273,
-0.025097988545894623,
-0.11398974806070328,
-0.024396510794758797,
0.06002479046583176,
-0.05242688208818436,
0.05477088689804077,
-0.025026924908161163,
0.0375724621117115,
0.17063480615615845,
-0.03723656386137009,
-0.049455832690000534,
-0.12045363336801529,
-0.009596114978194237,
-0.06487730890512466,
0.09438472986221313,
0.024705907329916954,
0.12418009340763092,
0.1092207059264183,
0.012137599289417267,
-0.10241204500198364,
0.08259478211402893,
-0.01986747793853283,
-0.04321520775556564,
0.044761773198843,
0.15041036903858185,
-0.013709241524338722,
0.16318008303642273,
0.03208295628428459,
-0.12901747226715088,
0.02778017334640026,
0.002916376106441021,
-0.05903704836964607,
-0.06797663867473602,
0.023460211232304573,
-0.05931134149432182,
0.11819195747375488,
0.1875811517238617,
-0.04592861607670784,
0.007415512576699257,
-0.03134014084935188,
0.04082135483622551,
0.00736593222245574,
0.056211818009614944,
-0.018419325351715088,
-0.1616811901330948,
0.046633459627628326,
0.04239862784743309,
0.0557604543864727,
-0.16134314239025116,
-0.05228674039244652,
0.010217325761914253,
-0.05349377915263176,
-0.01762446202337742,
0.09871968626976013,
0.011879968456923962,
0.04144349321722984,
-0.04619990661740303,
-0.21636387705802917,
0.01770223118364811,
0.09774301201105118,
-0.08837727457284927,
-0.09243623167276382
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-distilroberta-base')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-distilroberta-base')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-distilroberta-base')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-distilroberta-base')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["distilroberta-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-distilroberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"distilroberta-base",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #distilroberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #distilroberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
83,
30,
49,
18,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #distilroberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.016179563477635384,
0.17483745515346527,
-0.006616861559450626,
0.09181816130876541,
0.12407670170068741,
0.014483985491096973,
0.25409168004989624,
0.08052258938550949,
-0.024745415896177292,
-0.026574302464723587,
0.02368866838514805,
0.21927763521671295,
-0.00322065781801939,
0.06625080853700638,
-0.03953150287270546,
-0.23719273507595062,
0.04993871971964836,
0.010431649163365364,
0.022437317296862602,
0.0974813774228096,
0.10483920574188232,
-0.04197899252176285,
0.02976999618113041,
-0.015116767026484013,
-0.1281358003616333,
0.049260880798101425,
0.015758223831653595,
-0.04626785218715668,
0.10696873068809509,
0.043159373104572296,
0.12747281789779663,
0.07645203918218613,
0.011554154567420483,
-0.1631617695093155,
0.008877493441104889,
0.04404466226696968,
-0.0367804691195488,
0.04261625185608864,
0.008047709241509438,
-0.11447547376155853,
-0.016085002571344376,
-0.016497448086738586,
0.09156777709722519,
0.054332684725522995,
-0.09889556467533112,
-0.12051784247159958,
-0.011081195436418056,
0.00017126153397839516,
0.07707807421684265,
0.09557634592056274,
-0.03899674490094185,
0.1306856870651245,
-0.12225210666656494,
0.0997372418642044,
0.1056051105260849,
-0.23877760767936707,
-0.020944200456142426,
0.07414770126342773,
0.014419755898416042,
0.08777645230293274,
-0.07266884297132492,
-0.0044306120835244656,
-0.01774299331009388,
0.03498782590031624,
0.08532581478357315,
-0.05648383870720863,
-0.053472768515348434,
-0.0006906320340931416,
-0.1286039501428604,
0.030365698039531708,
0.18127663433551788,
-0.0454140342772007,
-0.059429511427879333,
-0.08828768879175186,
-0.056107547134160995,
-0.03005872666835785,
-0.02063913084566593,
0.007782226428389549,
0.019559595733880997,
0.04391113296151161,
-0.055060334503650665,
-0.026878872886300087,
-0.10843238979578018,
-0.027044519782066345,
-0.08467471599578857,
0.16180631518363953,
0.044600095599889755,
0.023999322205781937,
-0.10822272300720215,
0.07826224714517593,
0.006206526421010494,
-0.10930239409208298,
-0.02193615585565567,
-0.06736277043819427,
-0.16911758482456207,
0.011127190664410591,
-0.13218165934085846,
-0.17250461876392365,
0.05156683921813965,
0.21048834919929504,
0.002937544835731387,
0.051607150584459305,
-0.018387233838438988,
0.0541221983730793,
0.0679842084646225,
0.09305501729249954,
-0.020400486886501312,
0.012726773507893085,
-0.0036299999337643385,
0.0016314326785504818,
0.03340854123234749,
-0.058135442435741425,
-0.11504131555557251,
0.009309695102274418,
-0.0017239813460037112,
0.04937438666820526,
0.00951436161994934,
0.07073435187339783,
-0.0335627906024456,
-0.049755703657865524,
-0.0006959459860809147,
-0.09152618050575256,
0.009183530695736408,
0.011584545485675335,
0.03514039143919945,
0.15902100503444672,
0.031651418656110764,
-0.002812829101458192,
-0.048847224563360214,
-0.004587486851960421,
-0.10236294567584991,
0.032688502222299576,
-0.08541063964366913,
-0.1171928197145462,
0.012499981559813023,
0.013864015229046345,
-0.040647804737091064,
-0.11918497830629349,
-0.08753523230552673,
-0.0008315534214489162,
0.046627987176179886,
-0.03458356112241745,
0.016639618203043938,
-0.08538159728050232,
0.0013399961171671748,
0.03875163197517395,
-0.015709562227129936,
-0.09363697469234467,
0.004699713084846735,
0.03450348228216171,
0.03244491666555405,
0.07970036566257477,
-0.06699655205011368,
0.016386419534683228,
-0.1646519899368286,
-0.03142491728067398,
0.006780876778066158,
0.1217116266489029,
-0.03425361216068268,
-0.0066752079874277115,
-0.06473144143819809,
-0.009321007877588272,
0.04376450181007385,
0.037563640624284744,
-0.01458753366023302,
0.1011132001876831,
-0.22586117684841156,
-0.08066738396883011,
0.20015248656272888,
-0.1569576859474182,
-0.05629594251513481,
0.1268291175365448,
-0.036664269864559174,
0.00030231449636630714,
0.0937829241156578,
0.07872416824102402,
0.12591122090816498,
-0.05994310975074768,
0.004796607885509729,
0.05270844325423241,
-0.0270254984498024,
0.060454316437244415,
0.07076749950647354,
0.0695013478398323,
-0.052496034651994705,
0.04716191813349724,
-0.06020314618945122,
0.03215261548757553,
-0.011810760013759136,
-0.06451310962438583,
-0.02296626754105091,
-0.021850228309631348,
0.11763324588537216,
0.014268080703914165,
0.05659034103155136,
-0.021617628633975983,
-0.11923425644636154,
0.16562096774578094,
0.08534129709005356,
-0.12463154643774033,
0.06476061046123505,
-0.07605014741420746,
-0.05157604068517685,
-0.033458955585956573,
-0.004958768375217915,
-0.19925834238529205,
-0.0758775919675827,
0.030906612053513527,
-0.0458175390958786,
0.11143489181995392,
0.06717057526111603,
0.010290004312992096,
0.07028773427009583,
-0.04784758388996124,
-0.008522887714207172,
0.03808912634849548,
0.024599183350801468,
-0.0386495366692543,
-0.15975509583950043,
-0.05049788951873779,
-0.038394197821617126,
0.10305216163396835,
-0.0857989490032196,
0.05634041875600815,
0.012331915087997913,
0.053104277700185776,
-0.011447243392467499,
-0.0051216729916632175,
-0.029532384127378464,
0.06958496570587158,
-0.07152968645095825,
0.009618018753826618,
0.06093680486083031,
0.03378266096115112,
-0.11768080294132233,
0.06514627486467361,
-0.1600673645734787,
-0.08432389050722122,
0.08296851068735123,
-0.10564347356557846,
-0.05126577988266945,
0.001399049418978393,
-0.006671771872788668,
-0.010683183558285236,
-0.06793859601020813,
-0.034471992403268814,
0.10708066076040268,
0.049419913440942764,
0.11128227412700653,
-0.0504717156291008,
-0.02233082242310047,
-0.03787035495042801,
-0.111070416867733,
0.017316877841949463,
0.06695768237113953,
-0.016547420993447304,
-0.18737031519412994,
0.1343526840209961,
0.03924483433365822,
-0.173817977309227,
0.12946970760822296,
0.01175601501017809,
-0.007131066173315048,
-0.0784919485449791,
-0.04020204395055771,
0.02160249836742878,
-0.008550150319933891,
-0.04664032161235809,
0.023650972172617912,
0.0448274239897728,
0.0415160208940506,
0.006467767059803009,
-0.09996951371431351,
0.007487115915864706,
0.013942837715148926,
-0.002898495877161622,
-0.004599038045853376,
0.020723408088088036,
0.004215818364173174,
0.06908963620662689,
-0.02101384662091732,
-0.0039139119908213615,
0.07265464216470718,
-0.037437908351421356,
-0.11171500384807587,
0.20069730281829834,
-0.11854097247123718,
-0.29734691977500916,
-0.13242457807064056,
-0.039392098784446716,
-0.10575996339321136,
0.03792177140712738,
0.08238454908132553,
-0.08204706013202667,
-0.021595755591988564,
-0.06242850050330162,
0.020970208570361137,
-0.0703187882900238,
-0.023407788947224617,
-0.12867413461208344,
0.05095246434211731,
-0.003342570038512349,
-0.12006861716508865,
0.015183577314019203,
0.015035636723041534,
-0.10446028411388397,
0.06700155884027481,
-0.031813845038414,
0.0019646359141916037,
0.17564554512500763,
0.02088848501443863,
0.009827089495956898,
-0.011283481493592262,
0.14933401346206665,
-0.006647629197686911,
0.014341549016535282,
0.15318730473518372,
-0.0034220018424093723,
0.07998388260602951,
0.10431496053934097,
0.05772854760289192,
-0.025488581508398056,
0.03595352545380592,
0.03833400830626488,
-0.059422567486763,
-0.18204456567764282,
-0.05274856463074684,
-0.07324843853712082,
-0.027607547119259834,
0.05821212753653526,
-0.000189896221854724,
0.12514004111289978,
0.07010465860366821,
-0.027163533493876457,
-0.0014743686188012362,
0.06147759407758713,
0.11883123964071274,
0.16548371315002441,
0.027780737727880478,
0.1472623497247696,
-0.06051899120211601,
-0.07565368711948395,
0.030213316902518272,
0.05498189106583595,
0.13718605041503906,
0.027757035568356514,
0.14371085166931152,
0.125014066696167,
-0.003406201722100377,
0.059519775211811066,
0.04479513689875603,
0.0038814737927168608,
0.022032257169485092,
-0.01044598501175642,
-0.0775943249464035,
-0.0037867010105401278,
0.028578635305166245,
-0.020195385441184044,
-0.010159359313547611,
-0.03715430945158005,
-0.06333133578300476,
0.10174635052680969,
0.1005399227142334,
0.09545137733221054,
-0.20800025761127472,
-0.04152628779411316,
0.05912453681230545,
-0.028573619201779366,
-0.006625225767493248,
0.02329978719353676,
0.03102215938270092,
-0.09489990025758743,
0.06610623002052307,
-0.03649982810020447,
0.07221157103776932,
-0.07735342532396317,
0.013820097781717777,
-0.01966692879796028,
0.05486895889043808,
0.035253167152404785,
0.10927782952785492,
-0.2927652895450592,
0.07265476882457733,
-0.00901646539568901,
0.023563170805573463,
-0.037343576550483704,
0.029314659535884857,
-0.002276182873174548,
0.08374850451946259,
0.05558665096759796,
-0.004939447622746229,
-0.08275876939296722,
-0.07078203558921814,
0.003186519956216216,
-0.014674581587314606,
0.08165452629327774,
-0.07017985731363297,
0.10152842849493027,
-0.02961593121290207,
-0.008105854503810406,
0.0011123274452984333,
0.09578943997621536,
-0.12093859165906906,
-0.21786056458950043,
0.0880807414650917,
0.053528252989053726,
-0.011638030409812927,
-0.02493259124457836,
-0.07641366124153137,
0.016655921936035156,
0.14715206623077393,
-0.1796661764383316,
-0.09044169634580612,
-0.13549792766571045,
0.0122912572696805,
0.08402104675769806,
-0.0789397582411766,
-0.015380834229290485,
-0.027268212288618088,
0.14540453255176544,
-0.05334625765681267,
-0.15198127925395966,
0.039407700300216675,
-0.051500555127859116,
-0.09440787881612778,
-0.06473853439092636,
0.07421325147151947,
0.07745064049959183,
0.046454787254333496,
0.01871923729777336,
0.03982684761285782,
-0.023802436888217926,
-0.1094416007399559,
-0.08747740834951401,
0.027760369703173637,
0.05152600258588791,
0.05364903062582016,
-0.053973857313394547,
-0.07973649352788925,
-0.05835186690092087,
0.022482963278889656,
0.18966326117515564,
0.1650882065296173,
-0.0721210464835167,
0.09591861814260483,
0.13214534521102905,
-0.1299150288105011,
-0.17268408834934235,
-0.11367371678352356,
0.04817552492022514,
-0.011846001259982586,
0.025218447670340538,
-0.16305403411388397,
0.07960549741983414,
0.07957201451063156,
-0.01888785883784294,
-0.13352219760417938,
-0.3094058334827423,
-0.09645522385835648,
0.06238507851958275,
0.08344234526157379,
0.1311528980731964,
-0.14314928650856018,
-0.021699998527765274,
-0.07430125027894974,
-0.03848343715071678,
0.3309873044490814,
-0.07007944583892822,
0.12766224145889282,
0.006164290010929108,
-0.008706865832209587,
0.016811218112707138,
-0.02183225378394127,
0.09028959274291992,
0.05097613111138344,
0.0674244835972786,
-0.0019208004232496023,
-0.05256814882159233,
0.12754115462303162,
-0.039938367903232574,
0.0733877494931221,
-0.06719009578227997,
0.04116196557879448,
-0.012987282127141953,
-0.04798605293035507,
-0.08007510751485825,
-0.005557004362344742,
-0.055010173469781876,
-0.06858979910612106,
-0.0363316610455513,
0.04096995294094086,
0.05727098882198334,
-0.030181707814335823,
0.20965923368930817,
-0.09514881670475006,
0.07091905176639557,
0.17341426014900208,
0.11355318874120712,
0.08429756760597229,
-0.14683124423027039,
-0.040883950889110565,
-0.05153617262840271,
0.07043055444955826,
-0.15260078012943268,
0.09326495230197906,
0.040754154324531555,
-0.003324153833091259,
0.12823991477489471,
0.10343579947948456,
-0.03638547658920288,
0.006390796974301338,
0.06517700105905533,
-0.0864969864487648,
-0.01623670570552349,
-0.012498071417212486,
0.046885937452316284,
-0.03388489410281181,
0.06833715736865997,
0.17238618433475494,
-0.036461807787418365,
-0.0408623032271862,
0.016877571120858192,
0.029122615233063698,
-0.049503110349178314,
0.08536682277917862,
0.04029679298400879,
0.05833930894732475,
-0.09374278038740158,
0.09707432985305786,
0.09584259986877441,
-0.005094690248370171,
0.020100897178053856,
0.11768631637096405,
-0.07548633962869644,
-0.09394325315952301,
-0.0530935637652874,
0.06317728757858276,
-0.12944823503494263,
-0.017804672941565514,
-0.0976978987455368,
-0.09668396413326263,
-0.01939958892762661,
0.12659905850887299,
0.0766846314072609,
0.11842494457960129,
-0.09932445734739304,
-0.019686317071318626,
-0.09794259071350098,
0.051842499524354935,
-0.023298628628253937,
0.06088927015662193,
-0.07922925800085068,
0.12406975775957108,
0.014904997311532497,
0.002170486841350794,
-0.05888098478317261,
-0.06754309684038162,
-0.16016803681850433,
0.01884085312485695,
-0.06090020015835762,
0.014263663440942764,
-0.08026371151208878,
-0.0065395524725317955,
0.020119376480579376,
-0.025683296844363213,
0.007969677448272705,
0.008470220491290092,
-0.030647197738289833,
0.002187131904065609,
-0.01459111925214529,
0.0719708725810051,
-0.09134095907211304,
-0.013293493539094925,
0.018665945157408714,
-0.09429667890071869,
0.06254291534423828,
0.028111321851611137,
-0.06703388690948486,
0.07451833039522171,
-0.1378915011882782,
0.0023480693344026804,
-0.024436915293335915,
0.045437440276145935,
0.009103174321353436,
-0.07701527327299118,
0.007307255640625954,
0.025044677779078484,
0.05201352760195732,
0.03025892935693264,
0.057038065046072006,
-0.11172782629728317,
0.004029986448585987,
0.0020826321560889482,
-0.08905141055583954,
-0.07296968251466751,
0.0011386439437046647,
0.07145696133375168,
0.0589909553527832,
0.17313775420188904,
-0.07683166861534119,
0.08571616560220718,
-0.11118802428245544,
-0.014618984423577785,
0.02810252644121647,
-0.0758822113275528,
-0.036692064255476,
-0.10213347524404526,
0.019412590190768242,
-0.017477290704846382,
0.13867177069187164,
-0.03479471802711487,
0.09077233076095581,
0.024683751165866852,
0.020439982414245605,
0.09082754701375961,
0.06152648478746414,
0.1565525084733963,
0.03633545711636543,
-0.0259388480335474,
-0.00901868287473917,
0.0023429151624441147,
-0.010662316344678402,
0.0004242085851728916,
0.04680642858147621,
0.043853629380464554,
-0.020808227360248566,
0.08123642206192017,
0.03400968387722969,
0.08004342764616013,
-0.07147060334682465,
0.008197580464184284,
0.014643549919128418,
-0.0009162737987935543,
-0.05534913018345833,
0.06244191154837608,
0.14821477234363556,
-0.07661014795303345,
0.10842666774988174,
0.034275300800800323,
-0.10361023247241974,
-0.13479052484035492,
-0.19935575127601624,
-0.05684429779648781,
-0.06645174324512482,
-0.034187037497758865,
-0.14763310551643372,
0.015209724195301533,
0.09013374894857407,
0.0318148173391819,
-0.014901946298778057,
0.02488572895526886,
0.09696076065301895,
-0.040128953754901886,
-0.0032125350553542376,
-0.04694400727748871,
0.04197174683213234,
0.07110978662967682,
0.046558964997529984,
0.03762049973011017,
0.053181588649749756,
0.04261656105518341,
0.04404380917549133,
0.0778457447886467,
0.0023964443244040012,
-0.09400436282157898,
-0.11946669220924377,
0.016113445162773132,
0.003664941294118762,
0.01995585858821869,
0.11948590725660324,
0.044670626521110535,
-0.08036690950393677,
-0.0045856027863919735,
0.2301207333803177,
-0.08689501136541367,
-0.1251082569360733,
-0.15396207571029663,
0.2857726216316223,
0.03647658973932266,
0.06389530748128891,
0.001611840445548296,
-0.07280007004737854,
-0.06554117798805237,
0.1745193898677826,
0.12345030903816223,
-0.0024509101640433073,
0.0023756285663694143,
0.07530868798494339,
-0.012673580087721348,
0.026622604578733444,
0.08768334984779358,
-0.01438157632946968,
0.26289913058280945,
-0.06071364879608154,
0.1048322319984436,
-0.01602211594581604,
-0.021167518571019173,
-0.08362694829702377,
0.0885547325015068,
-0.04866098240017891,
-0.007364971097558737,
-0.04332225024700165,
0.07167846709489822,
-0.10192621499300003,
0.03093337081372738,
0.050756193697452545,
-0.019278760999441147,
-0.1002311110496521,
-0.015285230241715908,
0.11093827337026596,
-0.050110913813114166,
0.07738878577947617,
0.005788525100797415,
-0.012631089426577091,
0.11618766188621521,
-0.018669895827770233,
-0.03229374438524246,
-0.051501404494047165,
0.041841406375169754,
-0.053959742188453674,
0.055369436740875244,
-0.021204764023423195,
0.13008840382099152,
0.11406723409891129,
0.019049206748604774,
-0.048471953719854355,
0.1211332231760025,
-0.03307358920574188,
-0.05321257933974266,
0.0821930542588234,
0.11083729565143585,
-0.024354765191674232,
0.14556267857551575,
0.050124071538448334,
-0.16575747728347778,
0.05721302703022957,
-0.016143830493092537,
-0.07299494743347168,
-0.0363951250910759,
0.014887064695358276,
-0.05211925879120827,
0.06375987082719803,
0.1320514678955078,
-0.059226058423519135,
-0.025236736983060837,
-0.043585821986198425,
0.01620788499712944,
-0.006757156923413277,
-0.0285655464977026,
-0.06140536814928055,
-0.15907734632492065,
0.0056280698627233505,
0.08307977020740509,
0.04432760551571846,
-0.21461695432662964,
-0.027090881019830704,
0.019541483372449875,
-0.015807973220944405,
-0.09448376297950745,
0.05615435168147087,
0.042489759624004364,
0.030411502346396446,
-0.048695389181375504,
-0.245537668466568,
0.007639718241989613,
0.07803911715745926,
-0.13509754836559296,
-0.11017920821905136
] |
null | null |
transformers
|
# Cross-Encoder for Natural Language Inference
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
The model was trained on the [SNLI](https://nlp.stanford.edu/projects/snli/) and [MultiNLI](https://cims.nyu.edu/~sbowman/multinli/) datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see [SBERT.net - Pretrained Cross-Encoder](https://www.sbert.net/docs/pretrained_cross-encoders.html#nli).
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('cross-encoder/nli-roberta-base')
scores = model.predict([('A man is eating pizza', 'A man eats something'), ('A black race car starts up in front of a crowd of people.', 'A man is driving down a lonely road.')])
#Convert scores to labels
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(axis=1)]
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('cross-encoder/nli-roberta-base')
tokenizer = AutoTokenizer.from_pretrained('cross-encoder/nli-roberta-base')
features = tokenizer(['A man is eating pizza', 'A black race car starts up in front of a crowd of people.'], ['A man eats something', 'A man is driving down a lonely road.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = model(**features).logits
label_mapping = ['contradiction', 'entailment', 'neutral']
labels = [label_mapping[score_max] for score_max in scores.argmax(dim=1)]
print(labels)
```
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
```python
from transformers import pipeline
classifier = pipeline("zero-shot-classification", model='cross-encoder/nli-roberta-base')
sent = "Apple just announced the newest iPhone X"
candidate_labels = ["technology", "sports", "politics"]
res = classifier(sent, candidate_labels)
print(res)
```
|
{"language": "en", "license": "apache-2.0", "tags": ["roberta-base"], "datasets": ["multi_nli", "snli"], "metrics": ["accuracy"], "pipeline_tag": "zero-shot-classification"}
|
zero-shot-classification
|
cross-encoder/nli-roberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"roberta-base",
"zero-shot-classification",
"en",
"dataset:multi_nli",
"dataset:snli",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #roberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Natural Language Inference
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
The model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.
## Performance
For evaluation results, see URL - Pretrained Cross-Encoder.
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
## Zero-Shot Classification
This model can also be used for zero-shot-classification:
|
[
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #roberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.",
"## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):",
"## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
81,
30,
49,
18,
14,
33,
21
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #roberta-base #zero-shot-classification #en #dataset-multi_nli #dataset-snli #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Natural Language Inference\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThe model was trained on the SNLI and MultiNLI datasets. For a given sentence pair, it will output three scores corresponding to the labels: contradiction, entailment, neutral.## Performance\nFor evaluation results, see URL - Pretrained Cross-Encoder.## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):## Zero-Shot Classification\nThis model can also be used for zero-shot-classification:"
] |
[
-0.012450331822037697,
0.18565769493579865,
-0.006619107909500599,
0.08923697471618652,
0.12173096835613251,
0.007555744145065546,
0.23860366642475128,
0.0946941077709198,
-0.03857089951634407,
-0.020934268832206726,
0.027760099619627,
0.2214883416891098,
-0.0005910868640057743,
0.05267477408051491,
-0.04104902222752571,
-0.2249462902545929,
0.0505533404648304,
0.010125595144927502,
0.03220982849597931,
0.1024373322725296,
0.10679016262292862,
-0.03439415246248245,
0.031699955463409424,
-0.019713222980499268,
-0.12865905463695526,
0.04705250263214111,
0.01927412860095501,
-0.052999526262283325,
0.09766140580177307,
0.040177322924137115,
0.11399038881063461,
0.07511471956968307,
-0.0033349243458360434,
-0.16176249086856842,
0.010577261447906494,
0.044718094170093536,
-0.0414261594414711,
0.03746254742145538,
0.0037557773757725954,
-0.12255161255598068,
0.006816508714109659,
-0.006379482802003622,
0.08544141054153442,
0.05841675028204918,
-0.100577712059021,
-0.12125720828771591,
-0.012571580708026886,
0.0007140889065340161,
0.06652866303920746,
0.09600599110126495,
-0.04178692400455475,
0.1290944516658783,
-0.13174787163734436,
0.1046469658613205,
0.10754507780075073,
-0.24696797132492065,
-0.02518063224852085,
0.09047453850507736,
0.018305305391550064,
0.09359800070524216,
-0.07329710572957993,
0.0006153449648991227,
-0.025271685793995857,
0.03697926178574562,
0.08716819435358047,
-0.06203833222389221,
-0.07390651851892471,
-0.006042410619556904,
-0.12024009972810745,
0.024691374972462654,
0.1711689978837967,
-0.05308464169502258,
-0.052414704114198685,
-0.08973576128482819,
-0.044284265488386154,
-0.03676522150635719,
-0.024898335337638855,
0.01421323698014021,
0.017205171287059784,
0.046667974442243576,
-0.05754328519105911,
-0.028465790674090385,
-0.09630981087684631,
-0.023670313879847527,
-0.07510897517204285,
0.14011380076408386,
0.040735889226198196,
0.022895149886608124,
-0.10597661882638931,
0.06976048648357391,
-0.02050093561410904,
-0.10978848487138748,
-0.026790235191583633,
-0.05872846022248268,
-0.16605323553085327,
0.010712525807321072,
-0.13065214455127716,
-0.17338524758815765,
0.06294272840023041,
0.21598485112190247,
-0.004371569491922855,
0.05843447893857956,
-0.03290529176592827,
0.06020529940724373,
0.05529670789837837,
0.0960676372051239,
-0.02164079248905182,
-0.005296574905514717,
0.004179676529020071,
0.003346445271745324,
0.032136645168066025,
-0.05887006223201752,
-0.11428090184926987,
0.007645449135452509,
0.01293103862553835,
0.05218927934765816,
0.01327122189104557,
0.06948358565568924,
-0.029193222522735596,
-0.044936228543519974,
-0.009873587638139725,
-0.0875915139913559,
0.007795531302690506,
0.014495502226054668,
0.031724900007247925,
0.1569458693265915,
0.031711380928754807,
-0.009118339978158474,
-0.05874954164028168,
-0.011500927619636059,
-0.10075683891773224,
0.035394199192523956,
-0.07876450568437576,
-0.11124494671821594,
0.01730329543352127,
-0.0074829114601016045,
-0.04367668181657791,
-0.1113424226641655,
-0.08893245458602905,
-0.009352643974125385,
0.04121452569961548,
-0.042052045464515686,
0.010894590057432652,
-0.09175154566764832,
-0.00673984969034791,
0.0372416190803051,
-0.015526141040027142,
-0.08178561180830002,
0.0020215539261698723,
0.03847956284880638,
0.023425810039043427,
0.07876661419868469,
-0.06705465167760849,
0.020072977989912033,
-0.17164160311222076,
-0.03310557082295418,
-0.0010835685534402728,
0.13588696718215942,
-0.03681298717856407,
-0.0009214031742885709,
-0.07224419713020325,
-0.009640789590775967,
0.043723903596401215,
0.03753501549363136,
-0.019805893301963806,
0.10028775781393051,
-0.2174084186553955,
-0.07815679162740707,
0.2078258991241455,
-0.15413279831409454,
-0.0536031499505043,
0.12620580196380615,
-0.03625909239053726,
-0.00011715140863088891,
0.1027451753616333,
0.09397595375776291,
0.1261587291955948,
-0.039543334394693375,
0.00725755887106061,
0.056654706597328186,
-0.022766239941120148,
0.0521070621907711,
0.07045575231313705,
0.06330736726522446,
-0.05274278670549393,
0.05019250884652138,
-0.06546486914157867,
0.02483208291232586,
-0.005907400045543909,
-0.06826876848936081,
-0.024227667599916458,
-0.024844413623213768,
0.11584315448999405,
0.021924156695604324,
0.061033040285110474,
-0.02400381863117218,
-0.12372925877571106,
0.14643016457557678,
0.07588348537683487,
-0.1250528246164322,
0.06585909426212311,
-0.07270309329032898,
-0.024557748809456825,
-0.025992048904299736,
0.003319599200040102,
-0.2118838131427765,
-0.06400881707668304,
0.025655049830675125,
-0.022052984684705734,
0.12084250152111053,
0.051451925188302994,
0.011525008827447891,
0.058986298739910126,
-0.05141885206103325,
0.0012017883127555251,
0.043840549886226654,
0.021123921498656273,
-0.035444580018520355,
-0.15632383525371552,
-0.04418105632066727,
-0.038777485489845276,
0.10818906873464584,
-0.09663290530443192,
0.054174698889255524,
0.021981481462717056,
0.050402410328388214,
-0.006965021137148142,
-0.010302817448973656,
-0.02123831957578659,
0.07732733339071274,
-0.06794185191392899,
0.009771956130862236,
0.06980151683092117,
0.0350063294172287,
-0.12334111332893372,
0.06959979236125946,
-0.1572062373161316,
-0.10307637602090836,
0.07777225971221924,
-0.10368896275758743,
-0.0595441572368145,
-0.004109621047973633,
-0.0010901902569457889,
-0.0052638971246778965,
-0.06212881952524185,
-0.028930626809597015,
0.11709123849868774,
0.044428709894418716,
0.11677329242229462,
-0.058479633182287216,
-0.02022840641438961,
-0.03941849246621132,
-0.1090300977230072,
0.026229269802570343,
0.0846758484840393,
-0.03289739042520523,
-0.18519560992717743,
0.14462050795555115,
0.03408856317400932,
-0.17737914621829987,
0.1293485313653946,
0.005879031028598547,
-0.0031871211249381304,
-0.06671623885631561,
-0.026162147521972656,
0.02710885740816593,
-0.02462318167090416,
-0.05220670625567436,
0.022138455882668495,
0.04719655588269234,
0.039878956973552704,
0.004247005563229322,
-0.0967043787240982,
0.009573903866112232,
0.009243948385119438,
-0.0066325729712843895,
0.010803228244185448,
0.025998078286647797,
0.008302241563796997,
0.07369326055049896,
-0.022184044122695923,
-0.00197369116358459,
0.07530448585748672,
-0.036185361444950104,
-0.10855858027935028,
0.2090742588043213,
-0.12123976647853851,
-0.29193076491355896,
-0.13374346494674683,
-0.046876322478055954,
-0.10550013184547424,
0.03204580768942833,
0.08204121142625809,
-0.07689680904150009,
-0.01193046011030674,
-0.06913482397794724,
0.004393591079860926,
-0.06569338589906693,
-0.03360237553715706,
-0.14162077009677887,
0.05315481498837471,
0.002271217992529273,
-0.11402598768472672,
0.012300659902393818,
0.017375318333506584,
-0.11220746487379074,
0.06886997073888779,
-0.03351445868611336,
-0.004102560691535473,
0.1780775487422943,
0.01737040840089321,
0.008323193527758121,
-0.01722915843129158,
0.14847148954868317,
-0.011598735116422176,
0.0070030828937888145,
0.1525987982749939,
-0.005505564622581005,
0.07806174457073212,
0.10200884938240051,
0.05584898591041565,
-0.025205958634614944,
0.022778287529945374,
0.038980718702077866,
-0.07139378041028976,
-0.18761035799980164,
-0.05153791978955269,
-0.07656782120466232,
-0.023912549018859863,
0.058784499764442444,
0.0059951250441372395,
0.12992975115776062,
0.0690356120467186,
-0.021815136075019836,
0.0007329290965572,
0.06570333242416382,
0.11630966514348984,
0.16486039757728577,
0.03420660272240639,
0.15011534094810486,
-0.05943819135427475,
-0.06557590514421463,
0.027526261284947395,
0.0575735904276371,
0.13567616045475006,
0.01709792949259281,
0.14106470346450806,
0.12881769239902496,
-0.008604228496551514,
0.0622728168964386,
0.026579810306429863,
-0.007510789670050144,
0.023269684985280037,
-0.011807480826973915,
-0.08216328918933868,
-0.015526183880865574,
0.025179650634527206,
-0.020070306956768036,
-0.014619782567024231,
-0.029411379247903824,
-0.07155580818653107,
0.10113569349050522,
0.10026796162128448,
0.09475132077932358,
-0.20711542665958405,
-0.039506491273641586,
0.0557674840092659,
-0.031290240585803986,
-0.005813129711896181,
0.029256368055939674,
0.03115158900618553,
-0.09995120763778687,
0.07046851515769958,
-0.03600506857037544,
0.07880963385105133,
-0.07761731743812561,
0.012502390891313553,
-0.03771689906716347,
0.05046102777123451,
0.034374356269836426,
0.10782507061958313,
-0.2936159074306488,
0.06921129673719406,
-0.00794208887964487,
0.02203403227031231,
-0.03863126039505005,
0.02527143433690071,
0.006815529894083738,
0.0898805633187294,
0.07078266888856888,
-0.0014410163275897503,
-0.09332704544067383,
-0.05730877444148064,
-0.0034667770378291607,
-0.018515843898057938,
0.06967348605394363,
-0.06149952858686447,
0.09431260824203491,
-0.025228282436728477,
-0.006380427163094282,
-0.014646177180111408,
0.08775563538074493,
-0.10262458771467209,
-0.21019777655601501,
0.08087897300720215,
0.059506915509700775,
-0.013380584307014942,
-0.018973233178257942,
-0.06711675971746445,
0.015813827514648438,
0.1510452777147293,
-0.17426945269107819,
-0.09025180339813232,
-0.13491782546043396,
0.017078332602977753,
0.07379380613565445,
-0.07790034264326096,
-0.020112087950110435,
-0.03486156091094017,
0.13486869633197784,
-0.050726428627967834,
-0.152598574757576,
0.0359918475151062,
-0.04495254531502724,
-0.09107216447591782,
-0.05870461463928223,
0.0776471197605133,
0.07092888653278351,
0.05132799595594406,
0.021823175251483917,
0.04548642784357071,
-0.0171645637601614,
-0.10621954500675201,
-0.0885140523314476,
0.022299306467175484,
0.05113700032234192,
0.06444106251001358,
-0.057388000190258026,
-0.09148228168487549,
-0.060147810727357864,
0.025848114863038063,
0.20653387904167175,
0.15422925353050232,
-0.07067760825157166,
0.10784995555877686,
0.13928671181201935,
-0.13079266250133514,
-0.1883605420589447,
-0.11151424050331116,
0.04821036756038666,
-0.010691069066524506,
0.03179192543029785,
-0.1638483852148056,
0.07557666301727295,
0.08533897250890732,
-0.017466308549046516,
-0.1276526302099228,
-0.3185724914073944,
-0.09157223999500275,
0.05842962861061096,
0.08790219575166702,
0.1381182074546814,
-0.1497616022825241,
-0.03312225267291069,
-0.06271783262491226,
-0.02897387556731701,
0.3255256712436676,
-0.08629051595926285,
0.13215389847755432,
0.0007825204520486295,
-0.012758033350110054,
0.01575522869825363,
-0.0195584949105978,
0.08022560179233551,
0.049682386219501495,
0.06443493813276291,
-0.0022013415582478046,
-0.045003172010183334,
0.13290157914161682,
-0.04270981252193451,
0.06555693596601486,
-0.08682785928249359,
0.034320149570703506,
-0.027255650609731674,
-0.03942519798874855,
-0.08806643635034561,
-0.0024810596369206905,
-0.054281752556562424,
-0.05716060474514961,
-0.04055728018283844,
0.048742007464170456,
0.05758865550160408,
-0.025887634605169296,
0.2156621664762497,
-0.11639522761106491,
0.09311468154191971,
0.1888318657875061,
0.1134427934885025,
0.08856187760829926,
-0.15146325528621674,
-0.04203873500227928,
-0.04933140426874161,
0.07479318976402283,
-0.16137710213661194,
0.08695916086435318,
0.0392037108540535,
-0.0013083107769489288,
0.13433799147605896,
0.09533225744962692,
-0.03444722667336464,
0.009931126609444618,
0.06690029799938202,
-0.0792231410741806,
-0.010253879241645336,
-0.010765978135168552,
0.04333420470356941,
-0.029997024685144424,
0.05569550022482872,
0.1724521964788437,
-0.03874122351408005,
-0.039347365498542786,
0.016536908224225044,
0.026186419650912285,
-0.048673149198293686,
0.08235491812229156,
0.03646804764866829,
0.059354428201913834,
-0.10045748203992844,
0.09188644587993622,
0.1002320870757103,
0.0039356364868581295,
0.029999110847711563,
0.1323169320821762,
-0.0779738798737526,
-0.0970979705452919,
-0.05326148867607117,
0.06478368490934372,
-0.11777961254119873,
-0.01663161814212799,
-0.09164499491453171,
-0.0945427194237709,
-0.012141851708292961,
0.12607525289058685,
0.07872182875871658,
0.11170949786901474,
-0.09641902148723602,
-0.03670111671090126,
-0.09147769957780838,
0.05137421190738678,
-0.034467797726392746,
0.05260249599814415,
-0.061976488679647446,
0.11316274106502533,
0.01803116500377655,
0.0004827326338272542,
-0.05757958069443703,
-0.07269540429115295,
-0.16554270684719086,
0.02317528799176216,
-0.05414706841111183,
0.022160746157169342,
-0.08197957277297974,
-0.0036054474767297506,
0.015162445604801178,
-0.014068846590816975,
0.005122476723045111,
0.0060508958995342255,
-0.03442186862230301,
-0.00146681338082999,
-0.019080210477113724,
0.0775470957159996,
-0.0888189747929573,
-0.015499530360102654,
0.014998024329543114,
-0.08853261917829514,
0.06202443316578865,
0.02140459604561329,
-0.07560200989246368,
0.07974190264940262,
-0.14136558771133423,
-0.004071236588060856,
-0.021847251802682877,
0.04319392517209053,
0.00724010169506073,
-0.06812774389982224,
0.015139240771532059,
0.023973088711500168,
0.04937394708395004,
0.03863552585244179,
0.05129775404930115,
-0.11559862643480301,
-0.001832915237173438,
-0.0041874004527926445,
-0.08880443871021271,
-0.06782576441764832,
0.015198949724435806,
0.07358159124851227,
0.05603847652673721,
0.17545641958713531,
-0.08211732655763626,
0.08861978352069855,
-0.10476278513669968,
-0.018140899017453194,
0.030521614477038383,
-0.08419627696275711,
-0.05313688516616821,
-0.10013669729232788,
0.01900581270456314,
-0.017195412889122963,
0.13121715188026428,
-0.02993791364133358,
0.08263762295246124,
0.025417737662792206,
0.024023108184337616,
0.09215352684259415,
0.05519764497876167,
0.1491793543100357,
0.04666512459516525,
-0.03225264325737953,
-0.0018034680979326367,
0.0037583094090223312,
-0.012539030984044075,
0.004546519368886948,
0.0545489527285099,
0.04932969808578491,
-0.01606171578168869,
0.07904919236898422,
0.029038414359092712,
0.09077579528093338,
-0.06462176144123077,
0.028090033680200577,
0.017449606209993362,
0.008699202910065651,
-0.05558589845895767,
0.06645425409078598,
0.15993380546569824,
-0.07975988835096359,
0.11584963649511337,
0.037023209035396576,
-0.10352065414190292,
-0.13718436658382416,
-0.21220919489860535,
-0.0602843277156353,
-0.05901140347123146,
-0.03146836906671524,
-0.14729511737823486,
0.01310656312853098,
0.1004943698644638,
0.032974135130643845,
-0.015864668413996696,
0.022646350786089897,
0.08327309042215347,
-0.04495495557785034,
0.005938476882874966,
-0.04723614454269409,
0.050899773836135864,
0.06836778670549393,
0.051162924617528915,
0.04944939911365509,
0.04774516075849533,
0.05007452890276909,
0.049260515719652176,
0.08545883744955063,
0.001274654408916831,
-0.10227260738611221,
-0.11640217900276184,
0.020349618047475815,
0.008513372391462326,
0.02770891971886158,
0.12480239570140839,
0.04735568165779114,
-0.08447862416505814,
-0.002136517083272338,
0.22758834064006805,
-0.0805298313498497,
-0.11833160370588303,
-0.156891867518425,
0.29159459471702576,
0.03818436712026596,
0.05357228219509125,
-0.0035360364709049463,
-0.06761692464351654,
-0.05248383432626724,
0.1833062618970871,
0.12296327203512192,
-0.004452371504157782,
0.0039054020307958126,
0.07213283330202103,
-0.011482469737529755,
0.03249801695346832,
0.09326603263616562,
-0.02214781939983368,
0.2506294846534729,
-0.05969396233558655,
0.10301662981510162,
-0.020685408264398575,
-0.03065490908920765,
-0.09430119395256042,
0.08438673615455627,
-0.042060308158397675,
-0.015744652599096298,
-0.04170462116599083,
0.07591334730386734,
-0.11736249923706055,
0.03543869033455849,
0.06528057903051376,
-0.02677137404680252,
-0.10847145318984985,
-0.024805670604109764,
0.09372296184301376,
-0.0398162417113781,
0.06957661360502243,
0.01040432695299387,
-0.011880357749760151,
0.09260989725589752,
-0.02058948017656803,
-0.02488364651799202,
-0.05536671355366707,
0.03933794051408768,
-0.03530619665980339,
0.05548027530312538,
-0.02345281094312668,
0.11604415625333786,
0.11743488162755966,
0.016561206430196762,
-0.044704265892505646,
0.12358115613460541,
-0.03522425517439842,
-0.047284085303545,
0.08167476952075958,
0.10866093635559082,
-0.023611299693584442,
0.1418391466140747,
0.05594569817185402,
-0.14776618778705597,
0.05313095450401306,
-0.021721797063946724,
-0.07074429094791412,
-0.032888688147068024,
0.011423781514167786,
-0.04935833811759949,
0.06298163533210754,
0.14100544154644012,
-0.054499056190252304,
-0.028164897114038467,
-0.04636334255337715,
0.012559326365590096,
-0.010281362570822239,
-0.04133826121687889,
-0.067349374294281,
-0.15483511984348297,
0.0026553256902843714,
0.09059789031744003,
0.04062337055802345,
-0.23069506883621216,
-0.02104339376091957,
0.01721298135817051,
-0.009184980764985085,
-0.09905694425106049,
0.060342442244291306,
0.04011520370841026,
0.023741453886032104,
-0.048332829028367996,
-0.2520994544029236,
0.011945041827857494,
0.0776926577091217,
-0.1302192211151123,
-0.10313306003808975
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the [GLUE QNLI](https://arxiv.org/abs/1804.07461) dataset, which transformed the [SQuAD dataset](https://rajpurkar.github.io/SQuAD-explorer/) into an NLI task.
## Performance
For performance results of this model, see [SBERT.net Pre-trained Cross-Encoder][https://www.sbert.net/docs/pretrained_cross-encoders.html].
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Query1', 'Paragraph1'), ('Query2', 'Paragraph2')])
#e.g.
scores = model.predict([('How many people live in Berlin?', 'Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.'), ('What is the size of New York?', 'New York City is famous for the Metropolitan Museum of Art.')])
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'What is the size of New York?'], ['Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = torch.nn.functional.sigmoid(model(**features).logits)
print(scores)
```
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/qnli-distilroberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"arxiv:1804.07461",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1804.07461"
] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.
## Performance
For performance results of this model, see [URL Pre-trained Cross-Encoder][URL
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.",
"## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.",
"## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):"
] |
[
60,
34,
50,
23,
14,
33
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):"
] |
[
-0.06683853268623352,
0.16760697960853577,
-0.0020268643274903297,
0.10583819448947906,
0.1362825334072113,
-0.009169505909085274,
0.14919030666351318,
0.10903897881507874,
-0.04821023717522621,
-0.0023559401743113995,
0.08864589780569077,
0.13430660963058472,
-0.0054634143598377705,
0.034476324915885925,
0.02518564648926258,
-0.15910016000270844,
0.019955988973379135,
-0.006365683861076832,
-0.08433472365140915,
0.13067057728767395,
0.0988072082400322,
-0.0931401327252388,
0.061112914234399796,
0.013440467417240143,
-0.10522468388080597,
0.039205994457006454,
-0.023723365738987923,
-0.0640815943479538,
0.10041163116693497,
0.0811297595500946,
0.10844796150922775,
0.05113250017166138,
0.06541751325130463,
-0.13173463940620422,
0.013813051395118237,
0.07765104621648788,
-0.026058517396450043,
0.06262119859457016,
0.01992044597864151,
-0.010152021422982216,
-0.0005655037239193916,
0.019790975376963615,
0.037233348935842514,
0.026268666610121727,
-0.06624831259250641,
-0.12578657269477844,
-0.0661386027932167,
0.03116391785442829,
0.11600460857152939,
0.14586001634597778,
-0.03655621036887169,
0.14105455577373505,
-0.11890484392642975,
0.12211491912603378,
0.06447269767522812,
-0.2988509237766266,
-0.039884645491838455,
0.025126058608293533,
0.08559348434209824,
0.07851284742355347,
-0.05456576868891716,
0.01201643142849207,
0.04576081037521362,
0.04329817742109299,
0.014182036742568016,
-0.05634068697690964,
-0.12855534255504608,
0.08308294415473938,
-0.16502605378627777,
-0.05542537197470665,
0.1729709357023239,
-0.046488940715789795,
-0.06271598488092422,
-0.08167596906423569,
-0.06165126711130142,
-0.024913186207413673,
0.0036036984529346228,
-0.044971540570259094,
0.021064020693302155,
0.015857728198170662,
-0.11956077069044113,
-0.033619631081819534,
-0.08863332867622375,
-0.11785409599542618,
-0.057395122945308685,
0.08514110743999481,
0.04625026509165764,
0.042303647845983505,
-0.1135537326335907,
0.12018485367298126,
0.08160609751939774,
-0.10745636373758316,
-0.05947639420628548,
-0.02424311637878418,
-0.08919071406126022,
0.00890941172838211,
-0.10131756961345673,
-0.11463272571563721,
0.05269194021821022,
0.1965171843767166,
0.051153093576431274,
0.01824604906141758,
-0.011637482792139053,
0.04784025624394417,
0.034574560821056366,
0.21728280186653137,
-0.14064449071884155,
-0.008191355504095554,
0.035948220640420914,
0.0076000625267624855,
-0.012466652318835258,
-0.03425968810915947,
-0.13695009052753448,
-0.0024551786482334137,
0.06762012839317322,
0.09113042056560516,
0.03413432464003563,
0.07016292959451675,
0.010430342517793179,
-0.024981321766972542,
0.032159142196178436,
-0.08805420994758606,
0.0058694747276604176,
-0.015758614987134933,
-0.08880799263715744,
-0.0153798284009099,
0.09653092920780182,
0.01814018003642559,
-0.07191339880228043,
-0.05967153236269951,
-0.09024703502655029,
-0.010409367270767689,
-0.07021861523389816,
-0.11322527378797531,
0.009601966477930546,
0.0034246263094246387,
-0.018639525398612022,
-0.1537676304578781,
-0.1704186499118805,
-0.05199958756566048,
0.03720729053020477,
-0.012784007005393505,
-0.016444895416498184,
-0.0595741905272007,
-0.0375097431242466,
0.021855896338820457,
-0.009328732267022133,
0.05659256502985954,
-0.01174902357161045,
0.0787515714764595,
0.0033956000115722418,
0.07190149277448654,
0.00853913463652134,
0.050659723579883575,
-0.1500507891178131,
-0.0033963483292609453,
-0.020674046128988266,
0.12588021159172058,
-0.046421781182289124,
0.03524399548768997,
-0.11872247606515884,
-0.07679155468940735,
0.047609224915504456,
0.010270624421536922,
0.05977817624807358,
0.12338727712631226,
-0.1544087529182434,
-0.01592903584241867,
0.1283807009458542,
-0.06736008822917938,
-0.11897040903568268,
0.09780126065015793,
-0.04853251576423645,
0.06412322074174881,
0.04699956998229027,
0.10580247640609741,
0.1538478434085846,
-0.07015594840049744,
-0.031635187566280365,
0.05567150563001633,
-0.0678265392780304,
0.006915123667567968,
0.06452841311693192,
0.043177440762519836,
-0.03597727790474892,
0.018048442900180817,
-0.09559491276741028,
0.06975944340229034,
-0.040747616440057755,
-0.09144311398267746,
-0.04836862534284592,
-0.09867190569639206,
0.05501418933272362,
0.03788171708583832,
0.08055569231510162,
0.03680695220828056,
-0.11269854754209518,
0.08248145133256912,
0.08990166336297989,
-0.07771731168031693,
0.030897170305252075,
-0.1289188414812088,
0.03713952749967575,
-0.09294251352548599,
0.025099698454141617,
-0.21614325046539307,
-0.026741234585642815,
0.00985931046307087,
0.00942805502563715,
0.06489617377519608,
0.01667157933115959,
0.0288349911570549,
0.038976266980171204,
0.016633568331599236,
-0.041141342371702194,
-0.019100312143564224,
-0.007467746268957853,
-0.09864334017038345,
-0.11688781529664993,
-0.039594508707523346,
-0.03511656075716019,
0.12005649507045746,
-0.09536267817020416,
0.02839803695678711,
-0.03473607823252678,
0.057819388806819916,
-0.023539695888757706,
0.02971918135881424,
0.0070484415628015995,
0.0396498441696167,
-0.06503662467002869,
-0.030652714893221855,
0.03127564862370491,
0.014797977171838284,
-0.11238161474466324,
0.08407531678676605,
-0.07162512093782425,
-0.040841154754161835,
0.07799077033996582,
-0.0519942082464695,
-0.06232871487736702,
0.07276387512683868,
-0.03809581324458122,
-0.019486282020807266,
-0.052465710788965225,
-0.024708691984415054,
0.24112772941589355,
0.022290779277682304,
0.1262621432542801,
-0.08571378141641617,
-0.04610699042677879,
-0.024333016946911812,
-0.059227995574474335,
0.03396028280258179,
0.091547392308712,
0.022995205596089363,
-0.12426958233118057,
0.08966395258903503,
0.09900747239589691,
-0.11247168481349945,
0.09898383170366287,
-0.024950487539172173,
-0.041310057044029236,
-0.04107385873794556,
0.00023245986085385084,
0.013472733087837696,
0.02641969732940197,
-0.06136687472462654,
-0.0008174444083124399,
0.03396273031830788,
-0.0007232127827592194,
0.03863777220249176,
-0.13767066597938538,
0.013969301246106625,
0.018980583176016808,
-0.030771510675549507,
-0.035520996898412704,
0.05423792451620102,
-0.013325664214789867,
0.0696108415722847,
0.005106380209326744,
-0.02996974252164364,
0.05415669083595276,
-0.028570914641022682,
-0.12829607725143433,
0.22488225996494293,
-0.09603536128997803,
-0.22774852812290192,
-0.13665509223937988,
0.024677246809005737,
-0.02399257756769657,
-0.00945186335593462,
0.07520220428705215,
-0.10349368304014206,
-0.02139909379184246,
-0.0711468756198883,
-0.04644950106739998,
-0.016865303739905357,
-0.02589220553636551,
-0.07672198116779327,
0.004928263369947672,
0.027016986161470413,
-0.15316550433635712,
0.042099665850400925,
0.0055327038280665874,
-0.08691573143005371,
0.03167159855365753,
-0.0018628028919920325,
0.11148902028799057,
0.1663927584886551,
-0.019008580595254898,
0.0245186910033226,
-0.015546257607638836,
0.2309579700231552,
-0.08556707948446274,
-0.016719231382012367,
0.16801564395427704,
0.017369689419865608,
0.04717768356204033,
0.09610187262296677,
0.009077681228518486,
-0.054042570292949677,
0.04009239003062248,
0.00898192822933197,
-0.06670462340116501,
-0.1967061311006546,
-0.08733046054840088,
-0.05435071885585785,
-0.0019963670056313276,
0.058085277676582336,
0.04592158645391464,
0.044658634811639786,
0.03303653001785278,
0.016391027718782425,
0.06455487012863159,
-0.042085062712430954,
0.07370206713676453,
0.0832008495926857,
0.00898972898721695,
0.1295560598373413,
-0.059773411601781845,
-0.08394615352153778,
0.03776200860738754,
0.0699373334646225,
0.20093883574008942,
-0.03501866012811661,
0.07046154886484146,
0.09120561927556992,
0.10920114815235138,
0.0713454857468605,
0.0436578243970871,
-0.02143164537847042,
0.014635814353823662,
0.0036580278538167477,
-0.059523992240428925,
-0.012287215329706669,
0.06339888274669647,
-0.016284968703985214,
-0.019208626821637154,
-0.05971582978963852,
0.0047233146615326405,
0.055831752717494965,
0.17679768800735474,
0.02101622335612774,
-0.23761917650699615,
-0.0456165075302124,
0.0186733640730381,
-0.04656938090920448,
-0.013333170674741268,
0.04500093683600426,
0.05143100395798683,
-0.14000338315963745,
-0.06039545685052872,
-0.05044795200228691,
0.125654399394989,
-0.026891158893704414,
-0.005294478498399258,
0.021189045161008835,
-0.001703764428384602,
0.02745729498565197,
0.13703376054763794,
-0.2667916417121887,
0.11674847453832626,
0.009572298265993595,
0.05008187144994736,
-0.053387969732284546,
0.0274791456758976,
0.02256331779062748,
0.10295333713293076,
0.12725406885147095,
0.0030123982578516006,
0.010877483524382114,
-0.09620069712400436,
-0.04922156780958176,
0.037676043808460236,
0.05891343578696251,
0.0051406691782176495,
0.08297355473041534,
-0.020884307101368904,
0.007325014099478722,
-0.005753623321652412,
0.05122080445289612,
-0.08993537724018097,
-0.1444888710975647,
0.031164053827524185,
0.02360234037041664,
0.07756015658378601,
-0.049443237483501434,
-0.05392201244831085,
-0.02922237105667591,
0.1137341856956482,
-0.13104788959026337,
-0.08347545564174652,
-0.10555845499038696,
0.042328525334596634,
0.08906083554029465,
-0.08547109365463257,
0.003047098172828555,
-0.012066217139363289,
0.05198335275053978,
0.0011178147979080677,
-0.10362903773784637,
0.01988779567182064,
-0.06434579938650131,
-0.09776798635721207,
-0.053710877895355225,
0.005933969747275114,
0.027272263541817665,
0.0423470102250576,
0.02857867442071438,
0.033049196004867554,
-0.11053433269262314,
-0.12496538460254669,
-0.06572171300649643,
-0.014415023848414421,
-0.000938532582949847,
0.048962172120809555,
-0.04123404994606972,
0.02246498130261898,
-0.05426478013396263,
0.019125107675790787,
0.18611574172973633,
0.027716651558876038,
-0.12143387645483017,
0.07646087557077408,
0.1571614146232605,
-0.07852497696876526,
-0.2750796377658844,
-0.08864019811153412,
0.03706253319978714,
0.032724834978580475,
-0.07439777255058289,
-0.14852474629878998,
0.10069870948791504,
0.026417113840579987,
-0.028291059657931328,
-0.03525562956929207,
-0.3468773663043976,
-0.10672072321176529,
0.11028289794921875,
0.06790085136890411,
0.017017818987369537,
-0.14152270555496216,
-0.026377499103546143,
-0.02542470395565033,
-0.053526412695646286,
0.11842072755098343,
-0.15185602009296417,
0.10908640921115875,
-0.0077660842798650265,
0.053971562534570694,
0.0011061321711167693,
-0.05203355476260185,
0.08123216032981873,
0.023336492478847504,
0.051261600106954575,
-0.028821000829339027,
-0.023190610110759735,
0.08404416590929031,
-0.034598756581544876,
0.11941490322351456,
-0.05809007212519646,
0.12276197224855423,
-0.12252192944288254,
0.00017233568360097706,
-0.05280499905347824,
0.009558195248246193,
-0.016701793298125267,
-0.07055538147687912,
-0.0733523741364479,
0.04851476475596428,
0.051897041499614716,
-0.0176771879196167,
0.0731092169880867,
-0.036747559905052185,
0.033179160207509995,
0.16419610381126404,
0.12259941548109055,
0.013128727674484253,
-0.12156126648187637,
-0.013373380526900291,
-0.031431205570697784,
0.14170706272125244,
-0.1881362348794937,
0.08219043910503387,
0.10742058604955673,
0.0394715778529644,
0.1096542626619339,
0.07886325567960739,
-0.03203693777322769,
0.02590557187795639,
0.0464790053665638,
-0.12411253899335861,
-0.10097771883010864,
-0.02735777758061886,
0.01979765295982361,
-0.0677456334233284,
0.04362274333834648,
0.13313435018062592,
-0.08146915584802628,
-0.030428923666477203,
0.030553290620446205,
0.0063253468833863735,
-0.05826461687684059,
0.15519841015338898,
0.1079331636428833,
0.05097516253590584,
-0.0825340524315834,
0.08648589998483658,
0.0829111784696579,
-0.014632360078394413,
0.060698501765728,
0.06954222172498703,
-0.110160693526268,
-0.10051184147596359,
-0.07892830669879913,
0.1455468237400055,
-0.13397765159606934,
-0.0733928233385086,
-0.1096741259098053,
-0.03487095609307289,
0.03385604918003082,
0.0701722726225853,
0.050564516335725784,
0.06091950461268425,
-0.08757323026657104,
-0.043117884546518326,
-0.10297101736068726,
0.10449744015932083,
0.015996556729078293,
0.014125491492450237,
-0.0862927958369255,
0.09464609622955322,
0.025278368964791298,
0.12302318960428238,
-0.061422429978847504,
-0.053639598190784454,
-0.1109188124537468,
0.051465753465890884,
-0.1108650267124176,
0.005700241308659315,
-0.00804917886853218,
-0.00954259093850851,
-0.011127916164696217,
-0.01819438487291336,
-0.03459872305393219,
0.05927010253071785,
-0.04982389509677887,
-0.027976613491773605,
-0.017083333805203438,
0.05314554274082184,
-0.14227302372455597,
-0.03162020817399025,
0.012343761511147022,
-0.030025340616703033,
0.06660739332437515,
0.03452412784099579,
-0.05025063455104828,
0.023977013304829597,
-0.0691557377576828,
-0.029691459611058235,
-0.039703283458948135,
0.041693612933158875,
0.04182566702365875,
-0.07525695115327835,
0.03390820696949959,
0.0129158990457654,
0.008893415331840515,
0.01731681637465954,
0.04977913200855255,
-0.09505146741867065,
0.03201073408126831,
-0.019522910937666893,
0.033274196088314056,
-0.10332247614860535,
0.06459403783082962,
0.012841559015214443,
0.11758334934711456,
0.16421903669834137,
-0.0752158910036087,
0.09582366794347763,
-0.12311281263828278,
-0.022933954373002052,
0.0294441320002079,
-0.02162560448050499,
-0.09729430824518204,
-0.09793496131896973,
0.014731362462043762,
-0.06328652054071426,
0.15613850951194763,
0.048196278512477875,
0.0923430398106575,
0.006857549771666527,
0.09729024022817612,
0.06867080181837082,
0.008483966812491417,
0.1500026285648346,
0.028926605358719826,
0.01021699421107769,
0.00305820326320827,
0.046558089554309845,
-0.011440226808190346,
0.06908919662237167,
0.014021030627191067,
0.10226263105869293,
0.0831887423992157,
0.10516121983528137,
0.021369928494095802,
-0.012843471020460129,
-0.08477047830820084,
-0.05878866836428642,
-0.02031048759818077,
0.06092534959316254,
-0.06337404251098633,
0.1027132049202919,
0.13144825398921967,
-0.08360836654901505,
0.058809250593185425,
0.024886129423975945,
-0.08495844900608063,
-0.09705867618322372,
-0.051588866859674454,
-0.05258803442120552,
-0.1261771321296692,
0.011857743375003338,
-0.12238503992557526,
0.012698633596301079,
-0.002562653971835971,
0.014114049263298512,
-0.021884020417928696,
0.08085934072732925,
0.014131204225122929,
-0.04646634683012962,
0.010727196000516415,
-0.040407732129096985,
0.08823232352733612,
0.022766225039958954,
0.061805061995983124,
0.03990910202264786,
-0.013083919882774353,
0.054742276668548584,
0.04084904491901398,
0.02544950507581234,
0.05082608386874199,
-0.07746461033821106,
-0.04864777997136116,
-0.02018146961927414,
0.048653069883584976,
0.013535905629396439,
0.13341858983039856,
0.05444905906915665,
-0.022763529792428017,
-0.01780889183282852,
0.23645195364952087,
-0.10533805936574936,
-0.12192191928625107,
-0.1653653234243393,
0.26937630772590637,
0.023285314440727234,
0.028923669829964638,
-0.017611967399716377,
-0.0957862064242363,
-0.04759334772825241,
0.28645163774490356,
0.14203692972660065,
-0.08690868318080902,
-0.006080891005694866,
0.03610460087656975,
-0.009376497939229012,
0.013545478694140911,
0.10087932646274567,
0.07736768573522568,
0.2724824845790863,
-0.0824165940284729,
0.01594119518995285,
-0.07852651923894882,
-0.04358866810798645,
-0.04872411862015724,
0.11386265605688095,
-0.014412621967494488,
-0.04961789771914482,
-0.051188498735427856,
0.0750098004937172,
-0.13491606712341309,
-0.018695546314120293,
0.0046454379335045815,
-0.05705263838171959,
-0.09829151630401611,
-0.01922590471804142,
0.08067891746759415,
-0.015223979949951172,
0.047055572271347046,
-0.01515207439661026,
0.03697894513607025,
0.13114656507968903,
-0.012056717649102211,
-0.06255386769771576,
-0.0382801815867424,
0.05431593582034111,
-0.07418250292539597,
0.13452599942684174,
-0.02140597626566887,
0.11451972275972366,
0.12813307344913483,
0.039515167474746704,
-0.10494865477085114,
0.08975453674793243,
0.01995856687426567,
-0.10715493559837341,
0.044772952795028687,
0.04817403852939606,
0.019841190427541733,
0.058998994529247284,
0.01528418343514204,
-0.07097680121660233,
0.060279808938503265,
-0.06948189437389374,
-0.0052827694453299046,
-0.15087740123271942,
0.010146057233214378,
-0.03500804305076599,
0.12241935729980469,
0.14494851231575012,
-0.06073346361517906,
-0.0038054739125072956,
-0.046093132346868515,
0.023543426766991615,
-0.008556652814149857,
-0.028448930010199547,
-0.019037554040551186,
-0.12259189784526825,
0.005629850085824728,
0.010194981470704079,
-0.00324035226367414,
-0.21996444463729858,
-0.0404871441423893,
-0.00033505918690934777,
-0.05897333845496178,
0.024518771097064018,
0.11740656942129135,
0.12440529465675354,
0.039086584001779556,
-0.020194856449961662,
-0.13804538547992706,
-0.004218795336782932,
0.08220471441745758,
-0.12070762366056442,
-0.10985631495714188
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the [GLUE QNLI](https://arxiv.org/abs/1804.07461) dataset, which transformed the [SQuAD dataset](https://rajpurkar.github.io/SQuAD-explorer/) into an NLI task.
## Performance
For performance results of this model, see [SBERT.net Pre-trained Cross-Encoder][https://www.sbert.net/docs/pretrained_cross-encoders.html].
## Usage
Pre-trained models can be used like this:
```python
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Query1', 'Paragraph1'), ('Query2', 'Paragraph2')])
#e.g.
scores = model.predict([('How many people live in Berlin?', 'Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.'), ('What is the size of New York?', 'New York City is famous for the Metropolitan Museum of Art.')])
```
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
```python
from transformers import AutoTokenizer, AutoModelForSequenceClassification
import torch
model = AutoModelForSequenceClassification.from_pretrained('model_name')
tokenizer = AutoTokenizer.from_pretrained('model_name')
features = tokenizer(['How many people live in Berlin?', 'What is the size of New York?'], ['Berlin had a population of 3,520,031 registered inhabitants in an area of 891.82 square kilometers.', 'New York City is famous for the Metropolitan Museum of Art.'], padding=True, truncation=True, return_tensors="pt")
model.eval()
with torch.no_grad():
scores = torch.nn.functional.sigmoid(model(**features).logits)
print(scores)
```
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/qnli-electra-base
|
[
"transformers",
"pytorch",
"electra",
"text-classification",
"arxiv:1804.07461",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1804.07461"
] |
[] |
TAGS
#transformers #pytorch #electra #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
Given a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.
## Performance
For performance results of this model, see [URL Pre-trained Cross-Encoder][URL
## Usage
Pre-trained models can be used like this:
## Usage with Transformers AutoModel
You can use the model also directly with Transformers library (without SentenceTransformers library):
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.",
"## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):"
] |
[
"TAGS\n#transformers #pytorch #electra #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.",
"## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL",
"## Usage\n\nPre-trained models can be used like this:",
"## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):"
] |
[
57,
34,
50,
23,
14,
33
] |
[
"passage: TAGS\n#transformers #pytorch #electra #text-classification #arxiv-1804.07461 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nGiven a question and paragraph, can the question be answered by the paragraph? The models have been trained on the GLUE QNLI dataset, which transformed the SQuAD dataset into an NLI task.## Performance\nFor performance results of this model, see [URL Pre-trained Cross-Encoder][URL## Usage\n\nPre-trained models can be used like this:## Usage with Transformers AutoModel\nYou can use the model also directly with Transformers library (without SentenceTransformers library):"
] |
[
-0.08139800280332565,
0.16041189432144165,
-0.0008130564237944782,
0.10608351975679398,
0.15591081976890564,
-0.011612229980528355,
0.12859240174293518,
0.09285445511341095,
-0.023193156346678734,
-0.002844154601916671,
0.09732834994792938,
0.15097904205322266,
-0.00794843677431345,
0.047209952026605606,
0.0248067919164896,
-0.16479501128196716,
0.03332924470305443,
0.0017449820879846811,
-0.09433584660291672,
0.12972529232501984,
0.10567570477724075,
-0.10027066618204117,
0.05378144234418869,
0.03262406960129738,
-0.10094267129898071,
0.034859634935855865,
-0.032263610512018204,
-0.08513280004262924,
0.1250745803117752,
0.08892641216516495,
0.10992085933685303,
0.05206308886408806,
0.08884402364492416,
-0.12824943661689758,
0.013646877370774746,
0.06721924990415573,
-0.035412028431892395,
0.06213344633579254,
0.007376992143690586,
0.01839068904519081,
-0.0016585626872256398,
0.05138326808810234,
0.04991662874817848,
0.019196275621652603,
-0.06860174238681793,
-0.11976032704114914,
-0.05249357968568802,
0.028090858832001686,
0.12185889482498169,
0.13583965599536896,
-0.03791302070021629,
0.1663469523191452,
-0.141435444355011,
0.10164438188076019,
0.040629565715789795,
-0.28557613492012024,
-0.027670511975884438,
0.013138815760612488,
0.10049527138471603,
0.05972932651638985,
-0.03860277310013771,
0.0063815913163125515,
0.05883136764168739,
0.041739050298929214,
0.02147262915968895,
-0.03436869755387306,
-0.10826406627893448,
0.09091263264417648,
-0.17723731696605682,
-0.06637513637542725,
0.17933069169521332,
-0.027695955708622932,
-0.06187807396054268,
-0.08125351369380951,
-0.06469547003507614,
-0.037126533687114716,
0.022455966100096703,
-0.06079183518886566,
0.010594941675662994,
0.008495462127029896,
-0.10168193280696869,
-0.019162356853485107,
-0.10004673898220062,
-0.11831222474575043,
-0.06680125743150711,
0.1075991690158844,
0.056893013417720795,
0.033461134880781174,
-0.08893516659736633,
0.12554068863391876,
0.07643282413482666,
-0.09495841711759567,
-0.06644738465547562,
-0.021701984107494354,
-0.09872200340032578,
-0.0042851888574659824,
-0.11015177518129349,
-0.09532145410776138,
0.05226999148726463,
0.21868574619293213,
0.07125167548656464,
0.008366395719349384,
0.026703793555498123,
0.046773988753557205,
0.0516340397298336,
0.2127561718225479,
-0.1531628668308258,
0.02549305371940136,
0.03257477283477783,
0.009071044623851776,
-0.012134011834859848,
-0.04220334067940712,
-0.156703919172287,
-0.010980642400681973,
0.07549683004617691,
0.08833886682987213,
0.012023506686091423,
0.06494312733411789,
0.00677122687920928,
-0.03589637205004692,
0.05761205032467842,
-0.08330338448286057,
-0.005388041026890278,
-0.02135605923831463,
-0.10772255808115005,
-0.017310362309217453,
0.09681075811386108,
0.0329165942966938,
-0.06894959509372711,
-0.060483358800411224,
-0.09771394729614258,
-0.019486168399453163,
-0.06476064771413803,
-0.11602319777011871,
0.005267797503620386,
0.0018403249559924006,
-0.005286057014018297,
-0.17412731051445007,
-0.17291617393493652,
-0.0404776856303215,
0.037816789001226425,
-0.003412348683923483,
-0.041513774544000626,
-0.062021076679229736,
-0.017630161717534065,
0.009625636972486973,
-0.017713600769639015,
0.02181921899318695,
-0.01290993019938469,
0.08329292386770248,
0.022692644968628883,
0.06923427432775497,
-0.0011288134846836329,
0.05495980381965637,
-0.1454368680715561,
-0.011578001081943512,
-0.014261730946600437,
0.11030662059783936,
-0.032760899513959885,
0.020392777398228645,
-0.11172357946634293,
-0.09672924131155014,
0.07478238642215729,
0.02518833801150322,
0.06091412529349327,
0.12131917476654053,
-0.13158130645751953,
-0.023685414344072342,
0.1157577782869339,
-0.08388425409793854,
-0.12303067743778229,
0.1010332778096199,
-0.04843950644135475,
0.05426306650042534,
0.04906899482011795,
0.07917223125696182,
0.13783039152622223,
-0.07077278196811676,
-0.038807645440101624,
0.024996019899845123,
-0.07197368144989014,
-0.00534677691757679,
0.0662134662270546,
0.049276139587163925,
-0.052924495190382004,
0.015095002017915249,
-0.11117643117904663,
0.08332133293151855,
-0.04751208797097206,
-0.09073854237794876,
-0.04353620111942291,
-0.10225661098957062,
0.10020554065704346,
0.03785790503025055,
0.07416708767414093,
0.04055711254477501,
-0.12684866786003113,
0.10639164596796036,
0.10764288902282715,
-0.08344073593616486,
0.015291821211576462,
-0.15174749493598938,
0.048125259578228,
-0.1080242246389389,
0.010176953859627247,
-0.20498065650463104,
-0.04084979370236397,
0.00762198306620121,
-0.03515848144888878,
0.07704205811023712,
0.0425899364054203,
0.011598322540521622,
0.056836869567632675,
0.022345248609781265,
-0.05848044529557228,
-0.05022143945097923,
0.0015535458223894238,
-0.08317416906356812,
-0.09390532970428467,
-0.05188380926847458,
-0.03076985292136669,
0.12782451510429382,
-0.12303172051906586,
0.030375568196177483,
-0.06593065708875656,
0.04928303509950638,
-0.041594602167606354,
0.041681885719299316,
-0.008600437082350254,
0.04327097907662392,
-0.07353381812572479,
-0.0273295845836401,
0.028205396607518196,
0.0183856263756752,
-0.11378183960914612,
0.047134820371866226,
-0.09562186151742935,
-0.034481946378946304,
0.07185937464237213,
-0.06827595084905624,
-0.06740511208772659,
0.06567925214767456,
-0.02990131638944149,
-0.019254470244050026,
-0.07681123912334442,
-0.03079714998602867,
0.24600520730018616,
0.0182985570281744,
0.1291915625333786,
-0.08430833369493484,
-0.03867870196700096,
-0.023686926811933517,
-0.05832396820187569,
0.03359166532754898,
0.07285743951797485,
0.09185261279344559,
-0.12377593666315079,
0.08792320638895035,
0.08277741819620132,
-0.12243568897247314,
0.09036372601985931,
-0.0181779395788908,
-0.036820460110902786,
-0.040781170129776,
-0.012042832560837269,
0.005123232491314411,
0.02291913516819477,
-0.07384143769741058,
0.013077628798782825,
0.03172018751502037,
-0.003997522406280041,
0.037272077053785324,
-0.1399163156747818,
0.00778957549482584,
0.026867931708693504,
-0.03185276687145233,
-0.04662272334098816,
0.054373349994421005,
-0.0196001548320055,
0.06130257621407509,
0.0036240865010768175,
-0.054306965321302414,
0.0679377019405365,
-0.016273682937026024,
-0.126485675573349,
0.2272968739271164,
-0.09432802349328995,
-0.23742002248764038,
-0.14099344611167908,
0.049827829003334045,
-0.022502955049276352,
0.00009721218521008268,
0.0777333676815033,
-0.10286743938922882,
-0.03366447612643242,
-0.06213495135307312,
-0.04504316672682762,
-0.0036299435887485743,
-0.027430865913629532,
-0.06945542991161346,
-0.010447622276842594,
0.02490893192589283,
-0.15674988925457,
0.04706132784485817,
0.007023976184427738,
-0.08626218140125275,
0.04185148701071739,
0.0029891638550907373,
0.1067970022559166,
0.1960018426179886,
-0.02413005568087101,
0.02651907689869404,
-0.00726644042879343,
0.23054414987564087,
-0.08604064583778381,
-0.021201562136411667,
0.18044254183769226,
0.01872771978378296,
0.043656788766384125,
0.10356315970420837,
0.008468960411846638,
-0.057315099984407425,
0.049287889152765274,
-0.004549901932477951,
-0.06690163165330887,
-0.19151751697063446,
-0.11045792698860168,
-0.03948434069752693,
-0.021297533065080643,
0.0694139152765274,
0.04318515211343765,
0.08081472665071487,
0.04348184913396835,
0.018074190244078636,
0.03763613477349281,
-0.05827733129262924,
0.06848836690187454,
0.13207189738750458,
0.02116619050502777,
0.1282666027545929,
-0.04784670099616051,
-0.09669385850429535,
0.03337545320391655,
0.060967229306697845,
0.19964778423309326,
-0.01974724791944027,
0.07059352844953537,
0.08233581483364105,
0.11265186965465546,
0.08237448334693909,
0.0635192021727562,
0.004001186229288578,
0.010697315447032452,
0.0038907343987375498,
-0.03967567905783653,
-0.009146814234554768,
0.06405694037675858,
0.008191139437258244,
-0.02752971649169922,
-0.07786457985639572,
0.002938255900517106,
0.03741166740655899,
0.16207660734653473,
0.026552166789770126,
-0.2654740810394287,
-0.053377170115709305,
-0.0003798084217123687,
-0.05867813527584076,
-0.017580129206180573,
0.0393492616713047,
0.03652188926935196,
-0.1541680246591568,
-0.05263804644346237,
-0.048492707312107086,
0.11309026926755905,
-0.020219622179865837,
-0.00605420209467411,
0.020747799426317215,
0.006953320931643248,
0.031044241040945053,
0.14369075000286102,
-0.2502667307853699,
0.12488118559122086,
-0.007276750635355711,
0.045106690376996994,
-0.05646171793341637,
0.028047095984220505,
0.02348932810127735,
0.10600898414850235,
0.10934498906135559,
-0.006311903707683086,
-0.007974495179951191,
-0.1315472275018692,
-0.03186040744185448,
0.04728011414408684,
0.030823886394500732,
-0.005594308953732252,
0.08862561732530594,
-0.027782952412962914,
0.01050530094653368,
0.0032611035276204348,
0.053817085921764374,
-0.09527373313903809,
-0.14110241830348969,
0.03590252622961998,
0.02212880179286003,
0.11458145081996918,
-0.046901166439056396,
-0.07091386616230011,
-0.04997522383928299,
0.08300955593585968,
-0.1333480328321457,
-0.08625634759664536,
-0.09907760471105576,
0.05868001654744148,
0.09242206811904907,
-0.09274252504110336,
0.04083194583654404,
-0.0013828100636601448,
0.06681860238313675,
0.0022394759580492973,
-0.07944504171609879,
0.027794383466243744,
-0.07028324902057648,
-0.10587794333696365,
-0.03641572967171669,
-0.016785068437457085,
0.048428975045681,
0.030425358563661575,
0.048514340072870255,
0.03301828354597092,
-0.10775123536586761,
-0.12688679993152618,
-0.07203947752714157,
-0.034519847482442856,
-0.007783565204590559,
0.057889729738235474,
-0.03256085515022278,
0.02026546187698841,
-0.04618046432733536,
-0.0024980579037219286,
0.19505828619003296,
0.039525579661130905,
-0.11683136224746704,
0.062615305185318,
0.17444665729999542,
-0.06722389161586761,
-0.2509252727031708,
-0.11400186270475388,
0.04470352455973625,
0.0421157069504261,
-0.03936753794550896,
-0.16495324671268463,
0.10933112353086472,
0.04266039654612541,
-0.029854783788323402,
-0.035246819257736206,
-0.3328753709793091,
-0.10231165587902069,
0.13586847484111786,
0.0615118071436882,
0.0058561060577631,
-0.11881396919488907,
-0.014703931286931038,
-0.033282458782196045,
-0.03160635009407997,
0.12155454605817795,
-0.09806844592094421,
0.10469990223646164,
-0.0024012746289372444,
0.06324615329504013,
0.0019245477160438895,
-0.047895826399326324,
0.0962444469332695,
0.027136772871017456,
0.040699537843465805,
-0.03716219589114189,
-0.028053686022758484,
0.04991461709141731,
-0.025721397250890732,
0.12406432628631592,
-0.0566718615591526,
0.11353502422571182,
-0.12097983062267303,
-0.022193288430571556,
-0.04509788751602173,
0.02509419061243534,
-0.005740065593272448,
-0.07350139319896698,
-0.07065974920988083,
0.04291513189673424,
0.04880080372095108,
-0.008129912428557873,
0.050113365054130554,
-0.028356170281767845,
0.03973136097192764,
0.15196964144706726,
0.13187557458877563,
0.004879108630120754,
-0.12759815156459808,
-0.017691902816295624,
-0.026403378695249557,
0.14904849231243134,
-0.14711865782737732,
0.08906105905771255,
0.11451581865549088,
0.02583991177380085,
0.12089399993419647,
0.08053065091371536,
-0.03350910171866417,
0.041785646229982376,
0.041522953659296036,
-0.10840277373790741,
-0.09087944775819778,
-0.024622449651360512,
0.04816613718867302,
-0.057327330112457275,
0.05016898363828659,
0.12405502051115036,
-0.08611401915550232,
-0.040513891726732254,
0.025530533865094185,
0.0004927467089146376,
-0.04107546806335449,
0.1607077717781067,
0.11660023033618927,
0.05957872048020363,
-0.07770217210054398,
0.07625927031040192,
0.09618300199508667,
-0.019504429772496223,
0.06401363760232925,
0.04222135618329048,
-0.10588940978050232,
-0.1103757917881012,
-0.07960686087608337,
0.1384643018245697,
-0.19781872630119324,
-0.09469494968652725,
-0.11905554682016373,
-0.03372616693377495,
0.03942948579788208,
0.0783766359090805,
0.060027673840522766,
0.04003910347819328,
-0.0882609635591507,
-0.03351844474673271,
-0.10916462540626526,
0.0971139594912529,
0.03896551579236984,
0.00898434966802597,
-0.10211482644081116,
0.09407038986682892,
0.03678252920508385,
0.13014233112335205,
-0.06049257144331932,
-0.06628801673650742,
-0.0947970300912857,
0.054597947746515274,
-0.11082018911838531,
0.006630706135183573,
-0.009182666428387165,
-0.003364246105775237,
-0.0013000423787161708,
-0.022072691470384598,
-0.02563532255589962,
0.05659262463450432,
-0.05722765251994133,
-0.016132229939103127,
-0.003746338654309511,
0.05047748237848282,
-0.13491074740886688,
-0.03542096167802811,
0.01846247911453247,
-0.030056245625019073,
0.07684505730867386,
0.016060970723628998,
-0.04867076128721237,
0.03700822964310646,
-0.06371543556451797,
-0.026278939098119736,
-0.029950659722089767,
0.052899401634931564,
0.029590507969260216,
-0.10082758218050003,
0.0479368157684803,
0.010732253082096577,
0.003260029247030616,
0.004641746170818806,
0.025631263852119446,
-0.09639104455709457,
0.02236383967101574,
-0.021108347922563553,
0.029366178438067436,
-0.10731998085975647,
0.0533750057220459,
-0.0046320208348333836,
0.10615136474370956,
0.1785527467727661,
-0.07071241736412048,
0.08788717538118362,
-0.10957670211791992,
-0.015601745806634426,
0.015768053010106087,
-0.006543227005749941,
-0.09936949610710144,
-0.09540324658155441,
0.00806848332285881,
-0.056448034942150116,
0.15897256135940552,
0.029129894450306892,
0.10313519090414047,
0.004442616365849972,
0.10726684331893921,
0.06152825430035591,
0.008106076158583164,
0.17940543591976166,
0.038186486810445786,
0.02091521956026554,
-0.008555014617741108,
0.04223451390862465,
-0.01013815589249134,
0.03904852271080017,
0.015664000064134598,
0.07562847435474396,
0.0323067307472229,
0.09869390726089478,
0.024889444932341576,
-0.026704302057623863,
-0.09224668890237808,
-0.07329224795103073,
-0.01579851657152176,
0.05143291503190994,
-0.03889162838459015,
0.1086818054318428,
0.1208462193608284,
-0.07150524109601974,
0.045061156153678894,
0.029461825266480446,
-0.0804935172200203,
-0.10145817697048187,
-0.07401122152805328,
-0.05860727280378342,
-0.14500443637371063,
0.014304226264357567,
-0.11121705174446106,
0.00839908141642809,
0.03192901238799095,
0.012442899867892265,
-0.03689589723944664,
0.07212062925100327,
0.01241342257708311,
-0.04368603974580765,
-0.011692216619849205,
-0.051914338022470474,
0.08563107252120972,
0.019892370328307152,
0.050665464252233505,
0.026905560865998268,
0.012094538658857346,
0.052553921937942505,
0.02468538098037243,
0.0013424432836472988,
0.058701109141111374,
-0.06811324506998062,
-0.04319026693701744,
-0.027799414470791817,
0.04126004874706268,
0.014343496412038803,
0.11326438188552856,
0.040522441267967224,
-0.020508119836449623,
-0.013906619511544704,
0.21895933151245117,
-0.09283414483070374,
-0.1323097199201584,
-0.17311827838420868,
0.2407122254371643,
0.04464066028594971,
0.03343944624066353,
-0.0007162546389736235,
-0.0947662889957428,
-0.055648159235715866,
0.25857871770858765,
0.10800667107105255,
-0.0696985051035881,
-0.007954142056405544,
0.021195219829678535,
-0.01281147450208664,
0.013393421657383442,
0.09773919731378555,
0.07339742034673691,
0.2580370306968689,
-0.09211987257003784,
0.009003457613289356,
-0.0889134630560875,
-0.031529951840639114,
-0.04620738700032234,
0.086461141705513,
-0.016097228974103928,
-0.048674214631319046,
-0.05010990798473358,
0.08224178850650787,
-0.12877905368804932,
-0.004731197375804186,
-0.0053450921550393105,
-0.0320814773440361,
-0.10141699016094208,
-0.025547554716467857,
0.07512250542640686,
-0.011001619510352612,
0.04268690571188927,
-0.026301449164748192,
0.022380536422133446,
0.137967050075531,
-0.0008905127760954201,
-0.07643347233533859,
-0.03970276564359665,
0.07141333073377609,
-0.03271233290433884,
0.1386837214231491,
-0.023238688707351685,
0.13844414055347443,
0.12417110800743103,
0.04076496511697769,
-0.08828112483024597,
0.11496991664171219,
0.027894401922822,
-0.08750739693641663,
0.07264227420091629,
0.02900698035955429,
0.023894550278782845,
0.05439811572432518,
0.00248232064768672,
-0.10540831834077835,
0.0592973455786705,
-0.07049877196550369,
-0.0013822696637362242,
-0.1395513415336609,
0.030802344903349876,
-0.032772623002529144,
0.12444650381803513,
0.11269434541463852,
-0.06748658418655396,
-0.02025596797466278,
-0.052902814000844955,
0.043637823313474655,
0.0024602997582405806,
-0.029597505927085876,
-0.013550655916333199,
-0.13677170872688293,
0.012596950866281986,
-0.04356306418776512,
-0.0036949694622308016,
-0.18841518461704254,
-0.03744067624211311,
-0.010731018148362637,
-0.06564662605524063,
0.02741948887705803,
0.11450342833995819,
0.13345317542552948,
0.04000673070549965,
-0.029460368677973747,
-0.06212691590189934,
-0.032210007309913635,
0.08422917127609253,
-0.13290034234523773,
-0.11524412781000137
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')])
```
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/quora-distilroberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
## Usage and Performance
Pre-trained models can be used like this:
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
48,
34,
93,
39
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.028236351907253265,
0.02298957295715809,
-0.0028521628119051456,
0.0646425411105156,
0.11765683442354202,
0.023299384862184525,
0.16609442234039307,
0.12721982598304749,
-0.02589987777173519,
-0.002497537527233362,
0.06902404129505157,
0.09041360765695572,
0.03389200568199158,
0.01736462675035,
-0.0756693035364151,
-0.25257617235183716,
0.05000140145421028,
0.001953172730281949,
0.005523595958948135,
0.09553001075983047,
0.1380188912153244,
-0.06342750787734985,
0.06385538727045059,
0.03808078169822693,
-0.1387544721364975,
0.029387259855866432,
0.007435559760779142,
-0.07278154790401459,
0.1284724920988083,
0.06276270002126694,
0.12536683678627014,
0.04794098809361458,
0.06633620709180832,
-0.1390024721622467,
0.010272445157170296,
0.04691065847873688,
0.030382536351680756,
0.042663365602493286,
0.05961441248655319,
-0.0883123055100441,
-0.015024879947304726,
0.13564015924930573,
0.11921955645084381,
0.049067553132772446,
-0.11800447851419449,
-0.1697719544172287,
-0.09260506927967072,
-0.007557870354503393,
0.19177691638469696,
0.0677613914012909,
-0.019196776673197746,
0.17920920252799988,
-0.16551508009433746,
0.042527250945568085,
0.11477108299732208,
-0.29275941848754883,
-0.004422552417963743,
0.15177641808986664,
0.027819789946079254,
0.0426139235496521,
-0.035739149898290634,
-0.025486495345830917,
0.04099881276488304,
0.06852063536643982,
-0.023120788857340813,
-0.02526349201798439,
-0.027673279866576195,
0.019225681200623512,
-0.13329489529132843,
-0.05995437130331993,
0.16933511197566986,
-0.03704995661973953,
-0.05897475406527519,
-0.1055462658405304,
-0.039407879114151,
0.11085079610347748,
-0.008900068700313568,
-0.11178149282932281,
-0.016698947176337242,
0.05211774632334709,
-0.08724465221166611,
0.030446620658040047,
-0.07607933133840561,
-0.10109981149435043,
-0.07195515185594559,
0.023931683972477913,
0.07561454176902771,
0.05545526742935181,
-0.13705310225486755,
0.11136855185031891,
0.0016995988553389907,
-0.09120044857263565,
-0.040101051330566406,
-0.03573271259665489,
-0.09327655285596848,
-0.02613293193280697,
-0.13934822380542755,
-0.1651875376701355,
0.0202079638838768,
0.2592458128929138,
0.08788236975669861,
0.014212816022336483,
-0.03806806728243828,
0.053692687302827835,
0.04510349780321121,
0.17679710686206818,
-0.05372918024659157,
0.027468273416161537,
-0.01085987500846386,
-0.013765137642621994,
-0.021306518465280533,
-0.020566361024975777,
-0.1032397598028183,
-0.008217532187700272,
0.1410859078168869,
0.09234005957841873,
0.03564830124378204,
-0.013417183421552181,
-0.0345475859940052,
-0.04595864564180374,
-0.012750633992254734,
-0.08189065009355545,
0.014671438373625278,
-0.019546320661902428,
-0.08976012468338013,
0.0931757315993309,
0.008910393342375755,
0.08085379004478455,
-0.09844629466533661,
0.016391411423683167,
-0.03052634932100773,
-0.008689015172421932,
-0.051119543612003326,
-0.10875950753688812,
0.003899436676874757,
0.014876306988298893,
-0.007105946075171232,
-0.0993533730506897,
-0.1569775640964508,
-0.04772467911243439,
0.08535834401845932,
-0.07518336921930313,
-0.023914759978652,
-0.057448431849479675,
0.0000682943282299675,
-0.0035771564580500126,
-0.005692343693226576,
0.004734512884169817,
0.01771661452949047,
0.0641326829791069,
0.0076678963378071785,
0.07278640568256378,
0.10819133371114731,
0.07143276929855347,
-0.13799850642681122,
0.02069621905684471,
-0.07119812816381454,
0.11011669784784317,
0.004979265388101339,
-0.013975284993648529,
-0.14423952996730804,
-0.05839769169688225,
0.07755563408136368,
0.024960514158010483,
0.011379744857549667,
0.1003107875585556,
-0.1067274734377861,
-0.02742200717329979,
0.18494172394275665,
-0.11589451879262924,
-0.09805846959352493,
0.17407652735710144,
-0.07393323630094528,
0.09218185395002365,
0.11173079162836075,
0.04231793060898781,
0.04665884003043175,
-0.03837313503026962,
0.08602046966552734,
0.029852157458662987,
-0.05481569096446037,
0.09542881697416306,
0.08757112920284271,
-0.003998757805675268,
-0.12087751924991608,
0.008103582076728344,
-0.055464938282966614,
0.008033440448343754,
-0.05064034461975098,
-0.09141690284013748,
-0.042799897491931915,
-0.039185527712106705,
0.06163324788212776,
0.03553221747279167,
0.0880303904414177,
-0.02030434086918831,
-0.11953816562891006,
0.0649106502532959,
0.08500145375728607,
-0.08635890483856201,
0.0024070849176496267,
-0.10713905841112137,
0.030678845942020416,
-0.10741345584392548,
0.009620863944292068,
-0.15626586973667145,
0.04919950291514397,
0.027935834601521492,
-0.03920014202594757,
-0.007673514541238546,
0.12937958538532257,
0.008290670812129974,
0.009866510517895222,
-0.00487419543787837,
0.022477887570858,
0.05470797419548035,
-0.004190203733742237,
-0.09365541487932205,
-0.05990401655435562,
-0.03416970372200012,
-0.011483747512102127,
0.11308112740516663,
-0.034315500408411026,
-0.010413381271064281,
-0.13609528541564941,
0.09713054448366165,
-0.020865246653556824,
0.0609685517847538,
-0.01808066852390766,
0.0062762899324297905,
-0.026004375889897346,
-0.03020760416984558,
0.026992889121174812,
-0.0161923598498106,
-0.06696552038192749,
0.0824466347694397,
-0.13160282373428345,
-0.05768928676843643,
0.07812556624412537,
-0.058859750628471375,
-0.06345855444669724,
0.10349206626415253,
-0.012653677724301815,
-0.010400879196822643,
-0.07582449167966843,
0.0031032441183924675,
0.23192745447158813,
0.0235385000705719,
0.1441880762577057,
-0.10394033044576645,
-0.03910656273365021,
-0.018137700855731964,
-0.09496079385280609,
0.026489682495594025,
0.06276388466358185,
0.05044737830758095,
-0.16057021915912628,
0.0668400451540947,
0.08755689114332199,
-0.0982009693980217,
-0.049597419798374176,
-0.048767510801553726,
-0.04040800780057907,
-0.018693367019295692,
0.020596902817487717,
0.015745369717478752,
-0.046255916357040405,
0.034495946019887924,
0.02687368355691433,
0.055701132863759995,
0.05161583796143532,
0.0025433097034692764,
-0.14435702562332153,
0.018420690670609474,
0.002069329610094428,
-0.011670002713799477,
-0.007494900841265917,
0.03560759127140045,
0.02058662660419941,
0.07043254375457764,
0.04449900612235069,
-0.06807035207748413,
0.05159986391663551,
-0.033485181629657745,
-0.13951453566551208,
0.21684391796588898,
-0.055628452450037,
-0.22624246776103973,
-0.10680501163005829,
0.015781333670020103,
0.008800704032182693,
0.011640649288892746,
0.07580499351024628,
-0.05282660946249962,
-0.06239672750234604,
-0.039476968348026276,
0.07172691822052002,
-0.07547223567962646,
-0.07435284554958344,
-0.04170399531722069,
0.012975737452507019,
-0.012076806277036667,
-0.1478412002325058,
0.019435778260231018,
-0.017133019864559174,
-0.13254156708717346,
0.07526431232690811,
-0.11449142545461655,
-0.03394071012735367,
0.16663315892219543,
-0.02926144190132618,
0.0007279141573235393,
-0.03429313376545906,
0.20161108672618866,
-0.03790539130568504,
-0.011347824707627296,
0.20231664180755615,
-0.05146447941660881,
0.062159910798072815,
0.09932482987642288,
0.011152914725244045,
-0.05724319815635681,
0.05240379646420479,
0.0005114402738399804,
-0.02567918971180916,
-0.2063896358013153,
-0.06296171993017197,
-0.013676065020263195,
0.012855847366154194,
0.1107834056019783,
0.0007182872504927218,
0.09203709661960602,
0.036990851163864136,
-0.08321081846952438,
0.0317850187420845,
0.051255226135253906,
0.10255010426044464,
0.14584149420261383,
0.03280441462993622,
0.1406385451555252,
-0.03090754896402359,
-0.12735994160175323,
0.023079652339220047,
0.07208462059497833,
0.1642562747001648,
0.05984221771359444,
-0.001967767719179392,
0.06608805805444717,
0.04556852951645851,
0.06684515625238419,
0.03215453773736954,
-0.05517565459012985,
0.03261186555027962,
-0.06754516810178757,
-0.010860487818717957,
0.01439973246306181,
0.0679946020245552,
-0.015071535483002663,
-0.020678622648119926,
-0.09019079059362411,
-0.03930370509624481,
-0.012169619090855122,
0.17837707698345184,
0.13425955176353455,
-0.22493694722652435,
-0.06657751649618149,
0.0727817565202713,
-0.11972156912088394,
-0.022309724241495132,
0.06736302375793457,
-0.0696304440498352,
-0.21775352954864502,
-0.03360950946807861,
-0.04842714965343475,
0.14884597063064575,
-0.01835242100059986,
0.002515877829864621,
-0.11317767202854156,
-0.02717144414782524,
0.035474423319101334,
0.13648757338523865,
-0.3211652338504791,
0.050579994916915894,
-0.011529350653290749,
0.05429583042860031,
-0.09626295417547226,
0.01913764327764511,
0.026257775723934174,
0.0498470813035965,
0.06445223838090897,
-0.0350845642387867,
-0.037249721586704254,
-0.06635495275259018,
-0.014014074578881264,
0.04204874113202095,
0.01779092289507389,
-0.014398159459233284,
0.06714478880167007,
-0.08821114897727966,
0.02323475293815136,
0.023213429376482964,
-0.008164915256202221,
-0.047604676336050034,
-0.09148906171321869,
0.04724789038300514,
-0.09038347005844116,
0.0661279559135437,
-0.024140935391187668,
-0.05117615684866905,
-0.028319090604782104,
0.07918522506952286,
-0.1357451230287552,
-0.07783983647823334,
-0.08871682733297348,
0.07960803061723709,
0.11344430595636368,
-0.05970853194594383,
-0.002664614701643586,
-0.016676858067512512,
0.10650132596492767,
-0.017847442999482155,
-0.1516508162021637,
0.04764583334326744,
-0.08915083855390549,
-0.16881871223449707,
-0.041135936975479126,
0.01868947222828865,
0.030808664858341217,
0.06020243465900421,
0.08865536749362946,
0.08448202908039093,
-0.12201088666915894,
-0.07252929359674454,
-0.04164522513747215,
-0.0001835663861129433,
0.05557839572429657,
0.04091138020157814,
-0.02533305063843727,
0.001882239943370223,
-0.07327264547348022,
0.02687269076704979,
0.16272500157356262,
0.09315472841262817,
-0.0806574672460556,
0.11199698597192764,
0.16080211102962494,
-0.03340577706694603,
-0.22554153203964233,
-0.11893472820520401,
0.021587781608104706,
0.031612880527973175,
0.013462396338582039,
-0.10714894533157349,
0.11854009330272675,
0.030479401350021362,
-0.06616199761629105,
-0.17188070714473724,
-0.3671005964279175,
-0.07443568855524063,
0.1713097244501114,
0.031541381031274796,
0.19470879435539246,
-0.10560496896505356,
0.014806372113525867,
-0.07986968010663986,
-0.1159261167049408,
0.10894657671451569,
0.0050005074590444565,
0.08501661568880081,
-0.034093260765075684,
0.05420592054724693,
0.010197254829108715,
-0.025416484102606773,
0.08226087689399719,
0.0289011113345623,
0.04439155384898186,
-0.033352259546518326,
0.01492379605770111,
-0.019460739567875862,
-0.016228822991251945,
0.1672912836074829,
-0.04890361800789833,
0.12779323756694794,
-0.1419142484664917,
-0.06276766210794449,
-0.016046419739723206,
0.024127203971147537,
-0.027066685259342194,
-0.05535364896059036,
-0.06470940262079239,
0.018597234040498734,
0.07261050492525101,
-0.00032456679036840796,
0.04953699931502342,
0.002024851506575942,
0.10348399728536606,
0.08021503686904907,
0.09067286550998688,
0.013471505604684353,
-0.07162142544984818,
-0.017887728288769722,
-0.023417463526129723,
0.11423425376415253,
-0.1455373764038086,
0.10131525248289108,
0.06285625696182251,
-0.00002391975067439489,
0.14184269309043884,
0.06915212422609329,
-0.060422394424676895,
0.02636248804628849,
-0.012382125481963158,
-0.04637305438518524,
-0.13348804414272308,
0.010848729871213436,
0.10435190796852112,
-0.04518750309944153,
-0.0019641646649688482,
0.11644396930932999,
-0.08119875192642212,
-0.0706375241279602,
0.008771682158112526,
0.057382337749004364,
-0.03926582634449005,
0.10512572526931763,
0.05246136337518692,
0.057231150567531586,
-0.07779885828495026,
0.10765685886144638,
0.1493096947669983,
-0.018206464126706123,
0.06928034871816635,
0.017544802278280258,
-0.10551436990499496,
-0.06108799949288368,
0.0024780798703432083,
0.19576038420200348,
-0.1448131948709488,
-0.06071742996573448,
-0.10743575543165207,
-0.08942824602127075,
0.027615470811724663,
0.031815871596336365,
0.09662008285522461,
0.050208594650030136,
-0.0582425519824028,
-0.06624390184879303,
-0.11565201729536057,
0.07845304906368256,
0.10126746445894241,
0.01305184792727232,
-0.0042015789076685905,
0.02217823825776577,
0.043310754001140594,
0.07663457095623016,
-0.06105278059840202,
-0.08271121978759766,
-0.12199990451335907,
0.045144807547330856,
-0.17047615349292755,
0.009044838137924671,
-0.036601509898900986,
0.0003022434248123318,
0.045126281678676605,
-0.07448507100343704,
-0.014146518893539906,
0.019420986995100975,
-0.0656830221414566,
-0.024723190814256668,
-0.04790743812918663,
0.06352365761995316,
-0.12028785049915314,
-0.01956821046769619,
0.09564586728811264,
-0.04970070347189903,
0.051322758197784424,
0.0371239073574543,
-0.03554849326610565,
0.05435511842370033,
-0.06133165955543518,
-0.0969357043504715,
-0.06096852943301201,
0.05549168586730957,
0.03747888654470444,
-0.1624082773923874,
0.030587302520871162,
0.03868907317519188,
0.039308175444602966,
0.03676174208521843,
0.07743558287620544,
-0.11606483161449432,
-0.016286034137010574,
-0.08082305639982224,
-0.08908569067716599,
-0.08358822017908096,
0.01325599942356348,
0.008200830779969692,
0.08032640069723129,
0.20025207102298737,
-0.05918624997138977,
0.11930246651172638,
-0.13045009970664978,
-0.02069983445107937,
-0.03156894072890282,
-0.048923369497060776,
-0.10179747641086578,
-0.09497866779565811,
-0.0031806458719074726,
-0.02177494578063488,
0.11698029935359955,
0.0460670180618763,
0.12382645159959793,
0.03951376676559448,
0.05483991280198097,
0.10191228985786438,
0.04061557725071907,
0.14771799743175507,
0.10198737680912018,
0.017651734873652458,
-0.07955329865217209,
0.026712190359830856,
-0.013219586573541164,
-0.0833050087094307,
0.006515745539218187,
0.16688743233680725,
-0.06644535064697266,
0.047414086759090424,
0.012016004882752895,
-0.004906527232378721,
-0.011575120501220226,
-0.035505589097738266,
0.008810104802250862,
0.006349038332700729,
0.026974722743034363,
0.09378200769424438,
0.18293404579162598,
-0.09469182789325714,
0.04320510849356651,
-0.050525445491075516,
-0.055529188364744186,
-0.16978317499160767,
-0.033468253910541534,
-0.10318079590797424,
-0.16978172957897186,
0.0065807802602648735,
-0.08645619451999664,
0.005117142107337713,
0.13488566875457764,
0.041238997131586075,
-0.05799603834748268,
-0.025628315284848213,
-0.03854294866323471,
-0.051959339529275894,
-0.019285572692751884,
-0.05013048276305199,
0.0983944982290268,
0.09431729465723038,
0.011599581688642502,
0.05918845534324646,
0.04862096533179283,
0.09595529735088348,
0.011941192671656609,
0.011575278826057911,
0.020041348412632942,
-0.10389022529125214,
-0.10060551017522812,
0.029499050229787827,
0.014725682325661182,
-0.011605012230575085,
0.11036495119333267,
0.0414777547121048,
0.024291569367051125,
-0.00839686393737793,
0.21794293820858002,
-0.06551621854305267,
-0.1474355012178421,
-0.20090630650520325,
0.3832744359970093,
0.060175586491823196,
0.049961160868406296,
0.041826892644166946,
-0.09354113042354584,
-0.08771362900733948,
0.21245501935482025,
0.11224941909313202,
-0.034066617488861084,
-0.04883113503456116,
0.0555717758834362,
-0.01236236747354269,
0.03185252845287323,
0.10699848085641861,
0.028393233194947243,
0.2274254858493805,
-0.12071922421455383,
0.1020331010222435,
-0.02440096065402031,
-0.08203203231096268,
-0.09326554834842682,
0.07510117441415787,
-0.010312275029718876,
-0.06498714536428452,
-0.004234660416841507,
0.14585183560848236,
-0.02664933353662491,
0.021413059905171394,
-0.02548215165734291,
-0.051007241010665894,
-0.1135861948132515,
0.0017591587966307998,
0.03279903903603554,
-0.04903649911284447,
0.07186000794172287,
-0.019432999193668365,
0.035942111164331436,
0.12832286953926086,
0.009199274703860283,
-0.048081398010253906,
-0.05993684381246567,
0.09315922111272812,
-0.007312708999961615,
0.0295732282102108,
0.029766615480184555,
0.1741655021905899,
0.09672340750694275,
0.023585068061947823,
-0.02040170319378376,
0.06222079321742058,
0.04055801033973694,
0.031598370522260666,
0.11429018527269363,
0.11499851942062378,
0.0022774036042392254,
0.03847866132855415,
0.08489995449781418,
-0.1526382565498352,
0.03130566328763962,
-0.14398004114627838,
-0.021081620827317238,
-0.11639705300331116,
0.05685395747423172,
-0.026789942756295204,
0.10026440769433975,
0.15534526109695435,
-0.06755910068750381,
0.02107861079275608,
-0.022147351875901222,
0.06704210489988327,
0.017395958304405212,
-0.04797833785414696,
-0.023424401879310608,
-0.1702391356229782,
0.01641293615102768,
-0.045574892312288284,
0.009130828082561493,
-0.21578000485897064,
-0.008812090381979942,
-0.010528406128287315,
-0.0485992357134819,
-0.007692213635891676,
0.10061061382293701,
0.07568053901195526,
0.0011062530102208257,
-0.06103714928030968,
-0.18307805061340332,
-0.04356975853443146,
0.07539752870798111,
-0.0829671174287796,
-0.10309106856584549
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')])
```
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/quora-roberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
## Usage and Performance
Pre-trained models can be used like this:
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
48,
34,
93,
39
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.028236351907253265,
0.02298957295715809,
-0.0028521628119051456,
0.0646425411105156,
0.11765683442354202,
0.023299384862184525,
0.16609442234039307,
0.12721982598304749,
-0.02589987777173519,
-0.002497537527233362,
0.06902404129505157,
0.09041360765695572,
0.03389200568199158,
0.01736462675035,
-0.0756693035364151,
-0.25257617235183716,
0.05000140145421028,
0.001953172730281949,
0.005523595958948135,
0.09553001075983047,
0.1380188912153244,
-0.06342750787734985,
0.06385538727045059,
0.03808078169822693,
-0.1387544721364975,
0.029387259855866432,
0.007435559760779142,
-0.07278154790401459,
0.1284724920988083,
0.06276270002126694,
0.12536683678627014,
0.04794098809361458,
0.06633620709180832,
-0.1390024721622467,
0.010272445157170296,
0.04691065847873688,
0.030382536351680756,
0.042663365602493286,
0.05961441248655319,
-0.0883123055100441,
-0.015024879947304726,
0.13564015924930573,
0.11921955645084381,
0.049067553132772446,
-0.11800447851419449,
-0.1697719544172287,
-0.09260506927967072,
-0.007557870354503393,
0.19177691638469696,
0.0677613914012909,
-0.019196776673197746,
0.17920920252799988,
-0.16551508009433746,
0.042527250945568085,
0.11477108299732208,
-0.29275941848754883,
-0.004422552417963743,
0.15177641808986664,
0.027819789946079254,
0.0426139235496521,
-0.035739149898290634,
-0.025486495345830917,
0.04099881276488304,
0.06852063536643982,
-0.023120788857340813,
-0.02526349201798439,
-0.027673279866576195,
0.019225681200623512,
-0.13329489529132843,
-0.05995437130331993,
0.16933511197566986,
-0.03704995661973953,
-0.05897475406527519,
-0.1055462658405304,
-0.039407879114151,
0.11085079610347748,
-0.008900068700313568,
-0.11178149282932281,
-0.016698947176337242,
0.05211774632334709,
-0.08724465221166611,
0.030446620658040047,
-0.07607933133840561,
-0.10109981149435043,
-0.07195515185594559,
0.023931683972477913,
0.07561454176902771,
0.05545526742935181,
-0.13705310225486755,
0.11136855185031891,
0.0016995988553389907,
-0.09120044857263565,
-0.040101051330566406,
-0.03573271259665489,
-0.09327655285596848,
-0.02613293193280697,
-0.13934822380542755,
-0.1651875376701355,
0.0202079638838768,
0.2592458128929138,
0.08788236975669861,
0.014212816022336483,
-0.03806806728243828,
0.053692687302827835,
0.04510349780321121,
0.17679710686206818,
-0.05372918024659157,
0.027468273416161537,
-0.01085987500846386,
-0.013765137642621994,
-0.021306518465280533,
-0.020566361024975777,
-0.1032397598028183,
-0.008217532187700272,
0.1410859078168869,
0.09234005957841873,
0.03564830124378204,
-0.013417183421552181,
-0.0345475859940052,
-0.04595864564180374,
-0.012750633992254734,
-0.08189065009355545,
0.014671438373625278,
-0.019546320661902428,
-0.08976012468338013,
0.0931757315993309,
0.008910393342375755,
0.08085379004478455,
-0.09844629466533661,
0.016391411423683167,
-0.03052634932100773,
-0.008689015172421932,
-0.051119543612003326,
-0.10875950753688812,
0.003899436676874757,
0.014876306988298893,
-0.007105946075171232,
-0.0993533730506897,
-0.1569775640964508,
-0.04772467911243439,
0.08535834401845932,
-0.07518336921930313,
-0.023914759978652,
-0.057448431849479675,
0.0000682943282299675,
-0.0035771564580500126,
-0.005692343693226576,
0.004734512884169817,
0.01771661452949047,
0.0641326829791069,
0.0076678963378071785,
0.07278640568256378,
0.10819133371114731,
0.07143276929855347,
-0.13799850642681122,
0.02069621905684471,
-0.07119812816381454,
0.11011669784784317,
0.004979265388101339,
-0.013975284993648529,
-0.14423952996730804,
-0.05839769169688225,
0.07755563408136368,
0.024960514158010483,
0.011379744857549667,
0.1003107875585556,
-0.1067274734377861,
-0.02742200717329979,
0.18494172394275665,
-0.11589451879262924,
-0.09805846959352493,
0.17407652735710144,
-0.07393323630094528,
0.09218185395002365,
0.11173079162836075,
0.04231793060898781,
0.04665884003043175,
-0.03837313503026962,
0.08602046966552734,
0.029852157458662987,
-0.05481569096446037,
0.09542881697416306,
0.08757112920284271,
-0.003998757805675268,
-0.12087751924991608,
0.008103582076728344,
-0.055464938282966614,
0.008033440448343754,
-0.05064034461975098,
-0.09141690284013748,
-0.042799897491931915,
-0.039185527712106705,
0.06163324788212776,
0.03553221747279167,
0.0880303904414177,
-0.02030434086918831,
-0.11953816562891006,
0.0649106502532959,
0.08500145375728607,
-0.08635890483856201,
0.0024070849176496267,
-0.10713905841112137,
0.030678845942020416,
-0.10741345584392548,
0.009620863944292068,
-0.15626586973667145,
0.04919950291514397,
0.027935834601521492,
-0.03920014202594757,
-0.007673514541238546,
0.12937958538532257,
0.008290670812129974,
0.009866510517895222,
-0.00487419543787837,
0.022477887570858,
0.05470797419548035,
-0.004190203733742237,
-0.09365541487932205,
-0.05990401655435562,
-0.03416970372200012,
-0.011483747512102127,
0.11308112740516663,
-0.034315500408411026,
-0.010413381271064281,
-0.13609528541564941,
0.09713054448366165,
-0.020865246653556824,
0.0609685517847538,
-0.01808066852390766,
0.0062762899324297905,
-0.026004375889897346,
-0.03020760416984558,
0.026992889121174812,
-0.0161923598498106,
-0.06696552038192749,
0.0824466347694397,
-0.13160282373428345,
-0.05768928676843643,
0.07812556624412537,
-0.058859750628471375,
-0.06345855444669724,
0.10349206626415253,
-0.012653677724301815,
-0.010400879196822643,
-0.07582449167966843,
0.0031032441183924675,
0.23192745447158813,
0.0235385000705719,
0.1441880762577057,
-0.10394033044576645,
-0.03910656273365021,
-0.018137700855731964,
-0.09496079385280609,
0.026489682495594025,
0.06276388466358185,
0.05044737830758095,
-0.16057021915912628,
0.0668400451540947,
0.08755689114332199,
-0.0982009693980217,
-0.049597419798374176,
-0.048767510801553726,
-0.04040800780057907,
-0.018693367019295692,
0.020596902817487717,
0.015745369717478752,
-0.046255916357040405,
0.034495946019887924,
0.02687368355691433,
0.055701132863759995,
0.05161583796143532,
0.0025433097034692764,
-0.14435702562332153,
0.018420690670609474,
0.002069329610094428,
-0.011670002713799477,
-0.007494900841265917,
0.03560759127140045,
0.02058662660419941,
0.07043254375457764,
0.04449900612235069,
-0.06807035207748413,
0.05159986391663551,
-0.033485181629657745,
-0.13951453566551208,
0.21684391796588898,
-0.055628452450037,
-0.22624246776103973,
-0.10680501163005829,
0.015781333670020103,
0.008800704032182693,
0.011640649288892746,
0.07580499351024628,
-0.05282660946249962,
-0.06239672750234604,
-0.039476968348026276,
0.07172691822052002,
-0.07547223567962646,
-0.07435284554958344,
-0.04170399531722069,
0.012975737452507019,
-0.012076806277036667,
-0.1478412002325058,
0.019435778260231018,
-0.017133019864559174,
-0.13254156708717346,
0.07526431232690811,
-0.11449142545461655,
-0.03394071012735367,
0.16663315892219543,
-0.02926144190132618,
0.0007279141573235393,
-0.03429313376545906,
0.20161108672618866,
-0.03790539130568504,
-0.011347824707627296,
0.20231664180755615,
-0.05146447941660881,
0.062159910798072815,
0.09932482987642288,
0.011152914725244045,
-0.05724319815635681,
0.05240379646420479,
0.0005114402738399804,
-0.02567918971180916,
-0.2063896358013153,
-0.06296171993017197,
-0.013676065020263195,
0.012855847366154194,
0.1107834056019783,
0.0007182872504927218,
0.09203709661960602,
0.036990851163864136,
-0.08321081846952438,
0.0317850187420845,
0.051255226135253906,
0.10255010426044464,
0.14584149420261383,
0.03280441462993622,
0.1406385451555252,
-0.03090754896402359,
-0.12735994160175323,
0.023079652339220047,
0.07208462059497833,
0.1642562747001648,
0.05984221771359444,
-0.001967767719179392,
0.06608805805444717,
0.04556852951645851,
0.06684515625238419,
0.03215453773736954,
-0.05517565459012985,
0.03261186555027962,
-0.06754516810178757,
-0.010860487818717957,
0.01439973246306181,
0.0679946020245552,
-0.015071535483002663,
-0.020678622648119926,
-0.09019079059362411,
-0.03930370509624481,
-0.012169619090855122,
0.17837707698345184,
0.13425955176353455,
-0.22493694722652435,
-0.06657751649618149,
0.0727817565202713,
-0.11972156912088394,
-0.022309724241495132,
0.06736302375793457,
-0.0696304440498352,
-0.21775352954864502,
-0.03360950946807861,
-0.04842714965343475,
0.14884597063064575,
-0.01835242100059986,
0.002515877829864621,
-0.11317767202854156,
-0.02717144414782524,
0.035474423319101334,
0.13648757338523865,
-0.3211652338504791,
0.050579994916915894,
-0.011529350653290749,
0.05429583042860031,
-0.09626295417547226,
0.01913764327764511,
0.026257775723934174,
0.0498470813035965,
0.06445223838090897,
-0.0350845642387867,
-0.037249721586704254,
-0.06635495275259018,
-0.014014074578881264,
0.04204874113202095,
0.01779092289507389,
-0.014398159459233284,
0.06714478880167007,
-0.08821114897727966,
0.02323475293815136,
0.023213429376482964,
-0.008164915256202221,
-0.047604676336050034,
-0.09148906171321869,
0.04724789038300514,
-0.09038347005844116,
0.0661279559135437,
-0.024140935391187668,
-0.05117615684866905,
-0.028319090604782104,
0.07918522506952286,
-0.1357451230287552,
-0.07783983647823334,
-0.08871682733297348,
0.07960803061723709,
0.11344430595636368,
-0.05970853194594383,
-0.002664614701643586,
-0.016676858067512512,
0.10650132596492767,
-0.017847442999482155,
-0.1516508162021637,
0.04764583334326744,
-0.08915083855390549,
-0.16881871223449707,
-0.041135936975479126,
0.01868947222828865,
0.030808664858341217,
0.06020243465900421,
0.08865536749362946,
0.08448202908039093,
-0.12201088666915894,
-0.07252929359674454,
-0.04164522513747215,
-0.0001835663861129433,
0.05557839572429657,
0.04091138020157814,
-0.02533305063843727,
0.001882239943370223,
-0.07327264547348022,
0.02687269076704979,
0.16272500157356262,
0.09315472841262817,
-0.0806574672460556,
0.11199698597192764,
0.16080211102962494,
-0.03340577706694603,
-0.22554153203964233,
-0.11893472820520401,
0.021587781608104706,
0.031612880527973175,
0.013462396338582039,
-0.10714894533157349,
0.11854009330272675,
0.030479401350021362,
-0.06616199761629105,
-0.17188070714473724,
-0.3671005964279175,
-0.07443568855524063,
0.1713097244501114,
0.031541381031274796,
0.19470879435539246,
-0.10560496896505356,
0.014806372113525867,
-0.07986968010663986,
-0.1159261167049408,
0.10894657671451569,
0.0050005074590444565,
0.08501661568880081,
-0.034093260765075684,
0.05420592054724693,
0.010197254829108715,
-0.025416484102606773,
0.08226087689399719,
0.0289011113345623,
0.04439155384898186,
-0.033352259546518326,
0.01492379605770111,
-0.019460739567875862,
-0.016228822991251945,
0.1672912836074829,
-0.04890361800789833,
0.12779323756694794,
-0.1419142484664917,
-0.06276766210794449,
-0.016046419739723206,
0.024127203971147537,
-0.027066685259342194,
-0.05535364896059036,
-0.06470940262079239,
0.018597234040498734,
0.07261050492525101,
-0.00032456679036840796,
0.04953699931502342,
0.002024851506575942,
0.10348399728536606,
0.08021503686904907,
0.09067286550998688,
0.013471505604684353,
-0.07162142544984818,
-0.017887728288769722,
-0.023417463526129723,
0.11423425376415253,
-0.1455373764038086,
0.10131525248289108,
0.06285625696182251,
-0.00002391975067439489,
0.14184269309043884,
0.06915212422609329,
-0.060422394424676895,
0.02636248804628849,
-0.012382125481963158,
-0.04637305438518524,
-0.13348804414272308,
0.010848729871213436,
0.10435190796852112,
-0.04518750309944153,
-0.0019641646649688482,
0.11644396930932999,
-0.08119875192642212,
-0.0706375241279602,
0.008771682158112526,
0.057382337749004364,
-0.03926582634449005,
0.10512572526931763,
0.05246136337518692,
0.057231150567531586,
-0.07779885828495026,
0.10765685886144638,
0.1493096947669983,
-0.018206464126706123,
0.06928034871816635,
0.017544802278280258,
-0.10551436990499496,
-0.06108799949288368,
0.0024780798703432083,
0.19576038420200348,
-0.1448131948709488,
-0.06071742996573448,
-0.10743575543165207,
-0.08942824602127075,
0.027615470811724663,
0.031815871596336365,
0.09662008285522461,
0.050208594650030136,
-0.0582425519824028,
-0.06624390184879303,
-0.11565201729536057,
0.07845304906368256,
0.10126746445894241,
0.01305184792727232,
-0.0042015789076685905,
0.02217823825776577,
0.043310754001140594,
0.07663457095623016,
-0.06105278059840202,
-0.08271121978759766,
-0.12199990451335907,
0.045144807547330856,
-0.17047615349292755,
0.009044838137924671,
-0.036601509898900986,
0.0003022434248123318,
0.045126281678676605,
-0.07448507100343704,
-0.014146518893539906,
0.019420986995100975,
-0.0656830221414566,
-0.024723190814256668,
-0.04790743812918663,
0.06352365761995316,
-0.12028785049915314,
-0.01956821046769619,
0.09564586728811264,
-0.04970070347189903,
0.051322758197784424,
0.0371239073574543,
-0.03554849326610565,
0.05435511842370033,
-0.06133165955543518,
-0.0969357043504715,
-0.06096852943301201,
0.05549168586730957,
0.03747888654470444,
-0.1624082773923874,
0.030587302520871162,
0.03868907317519188,
0.039308175444602966,
0.03676174208521843,
0.07743558287620544,
-0.11606483161449432,
-0.016286034137010574,
-0.08082305639982224,
-0.08908569067716599,
-0.08358822017908096,
0.01325599942356348,
0.008200830779969692,
0.08032640069723129,
0.20025207102298737,
-0.05918624997138977,
0.11930246651172638,
-0.13045009970664978,
-0.02069983445107937,
-0.03156894072890282,
-0.048923369497060776,
-0.10179747641086578,
-0.09497866779565811,
-0.0031806458719074726,
-0.02177494578063488,
0.11698029935359955,
0.0460670180618763,
0.12382645159959793,
0.03951376676559448,
0.05483991280198097,
0.10191228985786438,
0.04061557725071907,
0.14771799743175507,
0.10198737680912018,
0.017651734873652458,
-0.07955329865217209,
0.026712190359830856,
-0.013219586573541164,
-0.0833050087094307,
0.006515745539218187,
0.16688743233680725,
-0.06644535064697266,
0.047414086759090424,
0.012016004882752895,
-0.004906527232378721,
-0.011575120501220226,
-0.035505589097738266,
0.008810104802250862,
0.006349038332700729,
0.026974722743034363,
0.09378200769424438,
0.18293404579162598,
-0.09469182789325714,
0.04320510849356651,
-0.050525445491075516,
-0.055529188364744186,
-0.16978317499160767,
-0.033468253910541534,
-0.10318079590797424,
-0.16978172957897186,
0.0065807802602648735,
-0.08645619451999664,
0.005117142107337713,
0.13488566875457764,
0.041238997131586075,
-0.05799603834748268,
-0.025628315284848213,
-0.03854294866323471,
-0.051959339529275894,
-0.019285572692751884,
-0.05013048276305199,
0.0983944982290268,
0.09431729465723038,
0.011599581688642502,
0.05918845534324646,
0.04862096533179283,
0.09595529735088348,
0.011941192671656609,
0.011575278826057911,
0.020041348412632942,
-0.10389022529125214,
-0.10060551017522812,
0.029499050229787827,
0.014725682325661182,
-0.011605012230575085,
0.11036495119333267,
0.0414777547121048,
0.024291569367051125,
-0.00839686393737793,
0.21794293820858002,
-0.06551621854305267,
-0.1474355012178421,
-0.20090630650520325,
0.3832744359970093,
0.060175586491823196,
0.049961160868406296,
0.041826892644166946,
-0.09354113042354584,
-0.08771362900733948,
0.21245501935482025,
0.11224941909313202,
-0.034066617488861084,
-0.04883113503456116,
0.0555717758834362,
-0.01236236747354269,
0.03185252845287323,
0.10699848085641861,
0.028393233194947243,
0.2274254858493805,
-0.12071922421455383,
0.1020331010222435,
-0.02440096065402031,
-0.08203203231096268,
-0.09326554834842682,
0.07510117441415787,
-0.010312275029718876,
-0.06498714536428452,
-0.004234660416841507,
0.14585183560848236,
-0.02664933353662491,
0.021413059905171394,
-0.02548215165734291,
-0.051007241010665894,
-0.1135861948132515,
0.0017591587966307998,
0.03279903903603554,
-0.04903649911284447,
0.07186000794172287,
-0.019432999193668365,
0.035942111164331436,
0.12832286953926086,
0.009199274703860283,
-0.048081398010253906,
-0.05993684381246567,
0.09315922111272812,
-0.007312708999961615,
0.0295732282102108,
0.029766615480184555,
0.1741655021905899,
0.09672340750694275,
0.023585068061947823,
-0.02040170319378376,
0.06222079321742058,
0.04055801033973694,
0.031598370522260666,
0.11429018527269363,
0.11499851942062378,
0.0022774036042392254,
0.03847866132855415,
0.08489995449781418,
-0.1526382565498352,
0.03130566328763962,
-0.14398004114627838,
-0.021081620827317238,
-0.11639705300331116,
0.05685395747423172,
-0.026789942756295204,
0.10026440769433975,
0.15534526109695435,
-0.06755910068750381,
0.02107861079275608,
-0.022147351875901222,
0.06704210489988327,
0.017395958304405212,
-0.04797833785414696,
-0.023424401879310608,
-0.1702391356229782,
0.01641293615102768,
-0.045574892312288284,
0.009130828082561493,
-0.21578000485897064,
-0.008812090381979942,
-0.010528406128287315,
-0.0485992357134819,
-0.007692213635891676,
0.10061061382293701,
0.07568053901195526,
0.0011062530102208257,
-0.06103714928030968,
-0.18307805061340332,
-0.04356975853443146,
0.07539752870798111,
-0.0829671174287796,
-0.10309106856584549
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [Quora Duplicate Questions](https://www.quora.com/q/quoradata/First-Quora-Dataset-Release-Question-Pairs) dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Question 1', 'Question 2'), ('Question 3', 'Question 4')])
```
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/quora-roberta-large
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.
Note: The model is not suitable to estimate the similarity of questions, e.g. the two questions "How to learn Java" and "How to learn Python" will result in a rahter low score, as these are not duplicates.
## Usage and Performance
Pre-trained models can be used like this:
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
52,
34,
93,
39
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the Quora Duplicate Questions dataset. The model will predict a score between 0 and 1 how likely the two given questions are duplicates.\n\nNote: The model is not suitable to estimate the similarity of questions, e.g. the two questions \"How to learn Java\" and \"How to learn Python\" will result in a rahter low score, as these are not duplicates.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.03287089988589287,
0.03220580890774727,
-0.002792539307847619,
0.0679178386926651,
0.11583749204874039,
0.01822967268526554,
0.18687804043293,
0.13437679409980774,
-0.03370516374707222,
0.01239317748695612,
0.06613951176404953,
0.07354146987199783,
0.02668229676783085,
0.02977769263088703,
-0.06534553319215775,
-0.2713264226913452,
0.04993590712547302,
0.006637313403189182,
-0.016340354457497597,
0.082265205681324,
0.14126195013523102,
-0.07298634946346283,
0.06217064708471298,
0.05221877619624138,
-0.1548863649368286,
0.03254738822579384,
0.011856010183691978,
-0.08051485568284988,
0.14062024652957916,
0.05658416822552681,
0.14231140911579132,
0.05355057120323181,
0.06941576302051544,
-0.11361847072839737,
0.011748974211513996,
0.057355280965566635,
0.028690189123153687,
0.06483281403779984,
0.07381010800600052,
-0.0848153606057167,
-0.029585467651486397,
0.11678961664438248,
0.12436320632696152,
0.041707661002874374,
-0.12498147040605545,
-0.17967434227466583,
-0.08553016185760498,
-0.009838024154305458,
0.1716412454843521,
0.05830838158726692,
-0.026011858135461807,
0.18556058406829834,
-0.16587524116039276,
0.04901837557554245,
0.1095905676484108,
-0.2921205163002014,
-0.0037483195774257183,
0.1480834037065506,
0.040460679680109024,
0.02314128540456295,
-0.03290810436010361,
-0.013682027347385883,
0.038183312863111496,
0.06915600597858429,
-0.008306341245770454,
-0.013202316127717495,
-0.038243960589170456,
0.013941487297415733,
-0.1394719034433365,
-0.0653618797659874,
0.17320235073566437,
-0.03870536759495735,
-0.0455210916697979,
-0.09346166998147964,
-0.06613437831401825,
0.11054395139217377,
0.005591654684394598,
-0.09999769181013107,
-0.019150730222463608,
0.037361808121204376,
-0.09903206676244736,
0.01328632514923811,
-0.09017538279294968,
-0.0847623348236084,
-0.0869031473994255,
0.04317522048950195,
0.07581494003534317,
0.0557035431265831,
-0.12159845978021622,
0.1135704517364502,
0.010577685199677944,
-0.10575850307941437,
-0.016711916774511337,
-0.03581656888127327,
-0.11436302214860916,
-0.034662242978811264,
-0.1445568948984146,
-0.1624370664358139,
0.009622065350413322,
0.2746860086917877,
0.09734425693750381,
-0.004394448362290859,
-0.03641950711607933,
0.060459788888692856,
0.052444856613874435,
0.18199503421783447,
-0.07806859910488129,
0.05308889225125313,
-0.01046827994287014,
0.015972938388586044,
-0.01670544221997261,
-0.03589500114321709,
-0.11638379096984863,
-0.0042594014666974545,
0.15442943572998047,
0.08017696440219879,
0.06319478154182434,
-0.0013094873866066337,
-0.023120921105146408,
-0.03826361522078514,
-0.015064915642142296,
-0.08397867530584335,
0.03448371961712837,
-0.03227211534976959,
-0.09382329881191254,
0.0662597268819809,
0.02178693376481533,
0.08284146338701248,
-0.0862513929605484,
0.011881799437105656,
-0.04020761698484421,
-0.00654161861166358,
-0.058926839381456375,
-0.13009408116340637,
0.005502915941178799,
0.02374870888888836,
-0.013810358941555023,
-0.10695427656173706,
-0.15747015178203583,
-0.042498767375946045,
0.08429151773452759,
-0.07375969737768173,
-0.04254097491502762,
-0.04171816632151604,
-0.006774070207029581,
0.0047939070500433445,
-0.005850460380315781,
-0.0019866847433149815,
0.01938837766647339,
0.05480531230568886,
0.007294788025319576,
0.08186715096235275,
0.09258009493350983,
0.07079054415225983,
-0.14123468101024628,
0.02159246988594532,
-0.07554320245981216,
0.09733321517705917,
0.0007424764335155487,
-0.03692271560430527,
-0.13090606033802032,
-0.05251488462090492,
0.08022475242614746,
0.03617674112319946,
0.022776678204536438,
0.08334081619977951,
-0.12080594152212143,
-0.030719386413693428,
0.16126753389835358,
-0.1115150973200798,
-0.08656304329633713,
0.176836296916008,
-0.07481200248003006,
0.08404294401407242,
0.08925269544124603,
0.06167721003293991,
0.01441936008632183,
-0.030046794563531876,
0.06590458005666733,
0.017781605944037437,
-0.04565657302737236,
0.10647022724151611,
0.0802680179476738,
0.0016104582464322448,
-0.10714733600616455,
0.0021788794547319412,
-0.04201694577932358,
0.0033173300325870514,
-0.06894306093454361,
-0.09312140196561813,
-0.04831736162304878,
-0.029916822910308838,
0.07975649833679199,
0.026474790647625923,
0.08946885913610458,
-0.024270592257380486,
-0.12763433158397675,
0.09345117956399918,
0.08358923345804214,
-0.07853730767965317,
0.0015502659371122718,
-0.08790002018213272,
0.05514185503125191,
-0.12374123185873032,
-0.0019612046889960766,
-0.16382460296154022,
0.005857852753251791,
0.035480793565511703,
-0.04692165181040764,
0.015140080824494362,
0.14238341152668,
-0.00022190056915860623,
0.04730154946446419,
0.0001127191717387177,
0.004371917340904474,
0.04459569230675697,
0.007726526353508234,
-0.09755449742078781,
-0.07664702832698822,
-0.04875147342681885,
-0.0175436083227396,
0.1276908814907074,
-0.032725993543863297,
-0.00023919263912830502,
-0.13663747906684875,
0.12270230799913406,
-0.02795683778822422,
0.06726392358541489,
-0.010458721779286861,
0.012501533143222332,
-0.04388565570116043,
-0.02504308894276619,
0.029530491679906845,
-0.014848093502223492,
-0.049829524010419846,
0.09627728164196014,
-0.15836934745311737,
-0.0409531369805336,
0.08836714923381805,
-0.07706864178180695,
-0.06798715144395828,
0.0993131622672081,
-0.018489625304937363,
-0.009376898407936096,
-0.09485068172216415,
-0.0156826451420784,
0.20844614505767822,
0.015332008711993694,
0.1537424772977829,
-0.10160452872514725,
-0.021846767514944077,
-0.02373281493782997,
-0.11354660987854004,
0.01935139298439026,
0.04426613077521324,
0.04160051792860031,
-0.14409592747688293,
0.09184277057647705,
0.09137747436761856,
-0.11941323429346085,
-0.029019704088568687,
-0.03746531158685684,
-0.03288352116942406,
-0.036266423761844635,
-0.008530855178833008,
0.00992781762033701,
-0.027668938040733337,
0.05548539385199547,
0.022097064182162285,
0.06508249789476395,
0.0435728020966053,
0.014114669524133205,
-0.15526345372200012,
0.02620646171271801,
0.008247812278568745,
-0.0009513023542240262,
0.00020343544019851834,
0.03980361670255661,
0.03240915760397911,
0.08016364276409149,
0.0351797454059124,
-0.08251238614320755,
0.04375332593917847,
-0.0245799720287323,
-0.11935722082853317,
0.20220446586608887,
-0.06418857723474503,
-0.23171773552894592,
-0.12543325126171112,
0.04153398051857948,
-0.0254649817943573,
0.0284960325807333,
0.07483630627393723,
-0.047969311475753784,
-0.05940685793757439,
-0.04744051769375801,
0.07056485861539841,
-0.06553135067224503,
-0.057584840804338455,
-0.06564519554376602,
0.006420007906854153,
-0.0010493986774235964,
-0.15580697357654572,
0.011212955228984356,
-0.01985001191496849,
-0.13604773581027985,
0.08270909637212753,
-0.07899542897939682,
-0.010255329310894012,
0.16837617754936218,
-0.04011749476194382,
-0.0020551157649606466,
-0.035932037979364395,
0.21144583821296692,
-0.03162882849574089,
0.012489858083426952,
0.20609085261821747,
-0.04364995285868645,
0.07578480243682861,
0.12185834348201752,
0.014143186621367931,
-0.04346136003732681,
0.055697668343782425,
0.004085979890078306,
-0.020722251385450363,
-0.1827675998210907,
-0.07326171547174454,
-0.013312481343746185,
0.0015584429493173957,
0.1118616834282875,
-0.002338045509532094,
0.11703769117593765,
0.03614424914121628,
-0.08995484560728073,
0.03120020404458046,
0.033848587423563004,
0.10977817326784134,
0.15997794270515442,
0.03603929281234741,
0.16397495567798615,
-0.01753595843911171,
-0.132488414645195,
0.01706407032907009,
0.07142373919487,
0.15959052741527557,
0.04072384163737297,
-0.018242504447698593,
0.07191752642393112,
0.03405921161174774,
0.07868117094039917,
0.02640828862786293,
-0.03679972514510155,
0.04370824247598648,
-0.06622111797332764,
-0.014968479983508587,
0.026352783665060997,
0.07653508335351944,
0.005359221249818802,
-0.018195437267422676,
-0.09236548840999603,
-0.0386410690844059,
-0.004821833688765764,
0.1912795752286911,
0.14512018859386444,
-0.2419639676809311,
-0.061507634818553925,
0.06962853670120239,
-0.11014198511838913,
-0.018605468794703484,
0.04699371010065079,
-0.057219963520765305,
-0.20587919652462006,
-0.036483559757471085,
-0.044423479586839676,
0.12873513996601105,
-0.02396460250020027,
0.017407871782779694,
-0.1088668704032898,
-0.0037940931506454945,
0.023492345586419106,
0.1306726485490799,
-0.30602124333381653,
0.07735942304134369,
-0.005571290384978056,
0.028852391988039017,
-0.08847087621688843,
0.02636065147817135,
0.037499554455280304,
0.05410102382302284,
0.027914877980947495,
-0.03412196412682533,
-0.06794535368680954,
-0.1019541546702385,
-0.01447890605777502,
0.03329499810934067,
0.02434415929019451,
-0.01893431507050991,
0.07295048981904984,
-0.08791227638721466,
0.018951719626784325,
0.031694427132606506,
0.025779711082577705,
-0.061407942324876785,
-0.08968307077884674,
0.053266994655132294,
-0.09155987203121185,
0.05666026845574379,
-0.04244745522737503,
-0.0647258535027504,
-0.0575680211186409,
0.07236741483211517,
-0.09864015132188797,
-0.06886366009712219,
-0.09187743812799454,
0.10358817875385284,
0.1201443299651146,
-0.058892492204904556,
0.005045982543379068,
-0.03299126774072647,
0.1276741623878479,
-0.030271658673882484,
-0.14077235758304596,
0.04787242412567139,
-0.08695114403963089,
-0.18228569626808167,
-0.05778321251273155,
0.016378920525312424,
0.005391437094658613,
0.06597761064767838,
0.08555072546005249,
0.1060669869184494,
-0.1281929761171341,
-0.05918307974934578,
-0.024490704759955406,
0.010349417105317116,
0.0530129075050354,
0.05980363488197327,
-0.03403376042842865,
-0.0053240396082401276,
-0.05730501562356949,
0.03151492029428482,
0.16266945004463196,
0.13333921134471893,
-0.09335033595561981,
0.09398265928030014,
0.16118700802326202,
-0.035691626369953156,
-0.23807929456233978,
-0.10470245033502579,
0.02114677056670189,
0.04097786173224449,
0.01596149243414402,
-0.10623084008693695,
0.10559507459402084,
0.026080582290887833,
-0.06810668110847473,
-0.19921305775642395,
-0.3535882830619812,
-0.08426278084516525,
0.16295690834522247,
0.009544314816594124,
0.16835252940654755,
-0.09334618598222733,
0.01882992312312126,
-0.07200830429792404,
-0.05674244090914726,
0.11677592247724533,
-0.010123214684426785,
0.09212194383144379,
-0.022549862042069435,
0.045421380549669266,
0.014774376526474953,
-0.025502415373921394,
0.09999138861894608,
0.0214613676071167,
0.058467257767915726,
-0.03684123978018761,
0.012596921063959599,
0.002864339156076312,
-0.027969928458333015,
0.169255331158638,
-0.034150782972574234,
0.12560446560382843,
-0.12425871938467026,
-0.0488470233976841,
-0.02793426811695099,
0.04721225053071976,
-0.029558653011918068,
-0.06160027161240578,
-0.07574033737182617,
0.022044742479920387,
0.06429511308670044,
-0.0031263066921383142,
0.03413941711187363,
0.007824328728020191,
0.11757781356573105,
0.11530592292547226,
0.07052170485258102,
0.002290953416377306,
-0.07644207030534744,
0.005038187373429537,
-0.02865632250905037,
0.11116921156644821,
-0.13136543333530426,
0.09488803893327713,
0.0625145360827446,
-0.00515682203695178,
0.14277641475200653,
0.07113510370254517,
-0.0691724419593811,
0.025240028277039528,
-0.0022865214850753546,
-0.06102057173848152,
-0.12716534733772278,
0.001014350214973092,
0.1120416522026062,
-0.0400361493229866,
-0.01710464432835579,
0.12777604162693024,
-0.07729045301675797,
-0.06502936780452728,
0.01161304023116827,
0.041276223957538605,
-0.02173643931746483,
0.1007012128829956,
0.06882841140031815,
0.06178441271185875,
-0.06406088918447495,
0.10418111085891724,
0.1749107986688614,
-0.009909868240356445,
0.07344341278076172,
0.015463700518012047,
-0.08871257305145264,
-0.06942931562662125,
-0.01776329055428505,
0.2067471742630005,
-0.15562395751476288,
-0.05347033962607384,
-0.12376556545495987,
-0.07888049632310867,
0.016117596998810768,
0.047817569226026535,
0.09953563660383224,
0.034959256649017334,
-0.06096327304840088,
-0.05343448743224144,
-0.12025012075901031,
0.09239010512828827,
0.11192873865365982,
0.04453551769256592,
-0.013860038481652737,
0.06599844992160797,
0.032309625297784805,
0.07947007566690445,
-0.06409940868616104,
-0.08243674784898758,
-0.12620995938777924,
0.028330039232969284,
-0.15713593363761902,
-0.010111291892826557,
-0.03577524796128273,
-0.016892528161406517,
0.04599938169121742,
-0.08631021529436111,
-0.013367350213229656,
0.012637665495276451,
-0.07286768406629562,
-0.026227736845612526,
-0.048548243939876556,
0.06766729056835175,
-0.13450081646442413,
-0.02712484821677208,
0.09844990819692612,
-0.051080815494060516,
0.05167121812701225,
0.006130293942987919,
-0.02681487426161766,
0.021283909678459167,
-0.040365420281887054,
-0.10448890179395676,
-0.06326239556074142,
0.05370960384607315,
0.04407564178109169,
-0.15046574175357819,
0.034633342176675797,
0.023813530802726746,
0.04863613843917847,
0.032495491206645966,
0.09285463392734528,
-0.10681762546300888,
-0.011162840761244297,
-0.0654686763882637,
-0.09288402646780014,
-0.08509011566638947,
0.015860581770539284,
0.015865439549088478,
0.07959801703691483,
0.19732770323753357,
-0.06674476712942123,
0.11078502982854843,
-0.14607058465480804,
-0.014212640933692455,
-0.045948781073093414,
-0.06775068491697311,
-0.09055338054895401,
-0.10398133844137192,
0.0018127302173525095,
-0.020655889064073563,
0.1290983110666275,
0.07949729263782501,
0.1304897964000702,
0.04222028702497482,
0.07883203029632568,
0.09613931179046631,
0.047242846339941025,
0.10612136870622635,
0.09710146486759186,
0.024568289518356323,
-0.07696165144443512,
0.03487970307469368,
-0.005151093006134033,
-0.075327567756176,
-0.009145429357886314,
0.13885553181171417,
-0.04313445836305618,
0.047403838485479355,
0.027900487184524536,
-0.025592822581529617,
-0.03286461904644966,
-0.061549026519060135,
0.010117235593497753,
0.023258639499545097,
0.01573597453534603,
0.05198691040277481,
0.17294037342071533,
-0.09616678953170776,
0.05143878236413002,
-0.042103033512830734,
-0.05420548841357231,
-0.15789414942264557,
-0.03333789110183716,
-0.09685321897268295,
-0.15734343230724335,
-0.0025480675976723433,
-0.0885271281003952,
0.015913022682070732,
0.1391577422618866,
0.028112351894378662,
-0.05023058131337166,
-0.016297034919261932,
-0.042800422757864,
-0.0447847880423069,
-0.03685012832283974,
-0.04912824183702469,
0.10199691355228424,
0.07438888400793076,
-0.011180942878127098,
0.043773967772722244,
0.08056911081075668,
0.07660676538944244,
0.012079463340342045,
0.021409275010228157,
0.01567625068128109,
-0.10064703226089478,
-0.10546135157346725,
0.030458437278866768,
-0.00639837933704257,
-0.010613745078444481,
0.09491019695997238,
0.03778057545423508,
0.02429492399096489,
-0.0063088382594287395,
0.22515292465686798,
-0.0718308836221695,
-0.15340447425842285,
-0.19897988438606262,
0.3925504982471466,
0.03735268488526344,
0.04824819415807724,
0.031650569289922714,
-0.10381948202848434,
-0.1141526848077774,
0.1846715360879898,
0.13957232236862183,
-0.03757867217063904,
-0.04579593613743782,
0.05021177977323532,
-0.022542808204889297,
0.01426080334931612,
0.09667779505252838,
0.024123458191752434,
0.19574320316314697,
-0.12143624573945999,
0.11033306270837784,
-0.023489724844694138,
-0.07747054100036621,
-0.09914333373308182,
0.08757908642292023,
-0.0006054493715055287,
-0.055196214467287064,
-0.0055865035392344,
0.12537860870361328,
-0.011684657074511051,
0.016954666003584862,
-0.0414222776889801,
-0.06764676421880722,
-0.12190880626440048,
0.021816249936819077,
0.04498741030693054,
-0.05891931802034378,
0.08162081986665726,
-0.031079348176717758,
0.04032259061932564,
0.12032638490200043,
-0.006586451083421707,
-0.03429294750094414,
-0.05201276019215584,
0.1021626740694046,
-0.03672444075345993,
0.04110773652791977,
0.01947266049683094,
0.17124368250370026,
0.11389374732971191,
0.019412698224186897,
-0.027855420485138893,
0.05599062889814377,
0.044213857501745224,
0.01930294930934906,
0.11120204627513885,
0.11952421814203262,
0.011368711479008198,
0.03814197704195976,
0.08243228495121002,
-0.18457508087158203,
0.04268735274672508,
-0.14795227348804474,
-0.030769649893045425,
-0.10342152416706085,
0.042879872024059296,
-0.01862787827849388,
0.10443770885467529,
0.15344755351543427,
-0.06863529980182648,
0.013661391101777554,
-0.024796264246106148,
0.05178695544600487,
0.019096791744232178,
-0.04969056695699692,
-0.027796681970357895,
-0.1725776195526123,
0.029365232214331627,
-0.05421414598822594,
0.02013411559164524,
-0.20382514595985413,
-0.0325913205742836,
0.001093793660402298,
-0.05562740936875343,
-0.00424076896160841,
0.10651307553052902,
0.07583669573068619,
-0.001469014910981059,
-0.06099686771631241,
-0.162979394197464,
-0.047733474522829056,
0.06599317491054535,
-0.0958397164940834,
-0.09633159637451172
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')])
```
The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`.
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/stsb-TinyBERT-L-4
|
[
"transformers",
"pytorch",
"jax",
"bert",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
47,
34,
38,
85
] |
[
"passage: TAGS\n#transformers #pytorch #jax #bert #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.1018068939447403,
0.05181725695729256,
-0.0023968915920704603,
0.059157323092222214,
0.1179395318031311,
-0.01356672216206789,
0.1925230473279953,
0.06266329437494278,
0.0032281309831887484,
-0.024624381214380264,
0.1129741221666336,
0.10616011172533035,
0.016521207988262177,
0.10064561665058136,
-0.042440690100193024,
-0.2364850789308548,
0.06007177755236626,
-0.0070038302801549435,
-0.028153659775853157,
0.13106979429721832,
0.10826661437749863,
-0.07041341066360474,
0.06480579078197479,
0.03029913827776909,
-0.1302100270986557,
0.031012598425149918,
0.006991867907345295,
-0.05731632187962532,
0.10829401761293411,
0.03067704662680626,
0.10081019997596741,
0.0939921960234642,
0.07745488733053207,
-0.12296551465988159,
0.007382601965218782,
0.0063626691699028015,
0.017311513423919678,
0.025654563680291176,
0.05087742954492569,
-0.10948707908391953,
-0.10683263838291168,
0.09693643450737,
0.08571042865514755,
0.01270565576851368,
-0.12037516385316849,
-0.08612044155597687,
-0.025205207988619804,
-0.04198529198765755,
0.16773280501365662,
0.08882227540016174,
-0.05072407051920891,
0.14051233232021332,
-0.16546614468097687,
0.041112933307886124,
0.03895944729447365,
-0.323636531829834,
0.003483750158920884,
0.08764201402664185,
0.0034883744083344936,
0.03945263475179672,
-0.027819111943244934,
-0.005329758860170841,
0.041045285761356354,
0.05586545914411545,
0.03135641664266586,
-0.038039617240428925,
0.03505055606365204,
0.014455686323344707,
-0.17962168157100677,
-0.02957088127732277,
0.2132486253976822,
-0.012625647708773613,
-0.05105971544981003,
-0.09227420389652252,
-0.03575388714671135,
0.05454649776220322,
-0.027917219325900078,
-0.09403304010629654,
-0.0002995362738147378,
0.06360745429992676,
-0.09977148473262787,
0.062051158398389816,
-0.11694594472646713,
-0.12564818561077118,
-0.12393181025981903,
0.12397140264511108,
0.058516498655080795,
0.04854318127036095,
-0.10782363265752792,
0.1164080873131752,
-0.03153989091515541,
-0.07901397347450256,
-0.02859816700220108,
-0.07697615772485733,
-0.12585236132144928,
-0.013163987547159195,
-0.19046393036842346,
-0.15082679688930511,
0.018758419901132584,
0.22904662787914276,
0.060466669499874115,
0.004972934722900391,
0.033935729414224625,
0.041198570281267166,
0.03478030487895012,
0.1743755042552948,
-0.09926558285951614,
0.021922988817095757,
-0.048428550362586975,
-0.002462996169924736,
-0.013277100399136543,
-0.004015984013676643,
-0.1436050832271576,
-0.03460471332073212,
0.08038260042667389,
0.03557009622454643,
-0.04887658357620239,
0.06058119982481003,
-0.038236744701862335,
-0.04881091043353081,
0.05196266248822212,
-0.07350710779428482,
-0.014168184250593185,
-0.00074140535434708,
-0.08883697539567947,
0.07080439478158951,
-0.01790766790509224,
0.06728566437959671,
-0.0643407553434372,
-0.027974655851721764,
-0.04908320680260658,
-0.008619080297648907,
-0.05475703999400139,
-0.1105925589799881,
-0.02645830437541008,
0.048080120235681534,
-0.0027558491565287113,
-0.11305185407400131,
-0.15955372154712677,
-0.058441177010536194,
0.05148078501224518,
-0.03036997653543949,
0.0027453897055238485,
-0.12507066130638123,
-0.018036365509033203,
-0.005363141186535358,
-0.026156703010201454,
-0.017081469297409058,
0.0020510340109467506,
0.032484348863363266,
0.012856436893343925,
0.1034550592303276,
0.006090695038437843,
0.05886368453502655,
-0.17560699582099915,
-0.0024692874867469072,
-0.07560712844133377,
0.1525801718235016,
0.018734389916062355,
0.028917210176587105,
-0.08698704093694687,
-0.04454522207379341,
-0.021243851631879807,
0.053984384983778,
0.040969934314489365,
0.13965754210948944,
-0.16391904652118683,
-0.061851002275943756,
0.2513390779495239,
-0.1284327507019043,
-0.11598508805036545,
0.1371479332447052,
-0.03622976318001747,
0.01943778805434704,
0.13422444462776184,
0.11409961432218552,
0.056645311415195465,
-0.013185767456889153,
0.08900745213031769,
0.060256052762269974,
-0.08443780988454819,
0.002127943327650428,
0.1069052442908287,
0.026767587289214134,
-0.10531603544950485,
0.041462890803813934,
-0.07566383481025696,
0.03522966429591179,
-0.05421165004372597,
-0.06971017271280289,
-0.02012300305068493,
-0.06686417013406754,
0.06848528981208801,
0.013206487521529198,
0.08170952647924423,
-0.021763892844319344,
-0.09114915132522583,
0.08406414836645126,
0.08524670451879501,
-0.08920326083898544,
0.02498677745461464,
-0.11104071140289307,
0.012608169578015804,
-0.0720713883638382,
-0.00818286556750536,
-0.2049560546875,
0.030590128153562546,
0.0010281912982463837,
0.026924526318907738,
0.04258265718817711,
0.0865471139550209,
0.03300182521343231,
0.027052683755755424,
0.008815101347863674,
-0.017389707267284393,
0.10040396451950073,
-0.005394225008785725,
-0.08699692040681839,
-0.09920375794172287,
-0.00590997189283371,
-0.018526380881667137,
0.16438403725624084,
-0.013949048705399036,
0.024582549929618835,
-0.10159825533628464,
0.05533580854535103,
-0.050071485340595245,
0.058535899966955185,
-0.0007597642834298313,
0.03356035426259041,
-0.02695932239294052,
0.0003881471639033407,
0.05768725648522377,
0.01757374405860901,
-0.1148795336484909,
0.15499082207679749,
-0.15193228423595428,
-0.04112286865711212,
0.11559657007455826,
-0.10884633660316467,
-0.06290720403194427,
0.013559277169406414,
-0.014183246530592442,
0.014825847931206226,
-0.047694068402051926,
0.0007997894426807761,
0.18626055121421814,
0.022305170074105263,
0.12479870021343231,
-0.08444536477327347,
-0.024937346577644348,
-0.01686849072575569,
-0.08763117343187332,
0.02107910066843033,
0.05984365940093994,
0.034147195518016815,
-0.1403716504573822,
0.08710502088069916,
0.0974455326795578,
-0.08408372104167938,
-0.0122535340487957,
-0.018786951899528503,
-0.05176491662859917,
-0.0221885833889246,
-0.023430481553077698,
-0.03204912319779396,
-0.05046255141496658,
-0.033675484359264374,
-0.03023233637213707,
0.034682922065258026,
0.04521404951810837,
0.0034213250037282705,
-0.155630424618721,
0.0080204326659441,
0.003638196038082242,
-0.0203767828643322,
-0.014345481060445309,
0.03083779849112034,
-0.013281100429594517,
0.07333807647228241,
0.02015344798564911,
-0.14261353015899658,
0.05891278013586998,
-0.03858984634280205,
-0.15263532102108002,
0.22569985687732697,
-0.07110138237476349,
-0.2786460220813751,
-0.12235891819000244,
0.028318103402853012,
-0.021125204861164093,
0.06356321275234222,
0.06646061688661575,
-0.05676727741956711,
-0.03817734122276306,
-0.07428191602230072,
-0.027466019615530968,
-0.03168352693319321,
0.009854322299361229,
-0.07029229402542114,
-0.0037583066150546074,
0.004941050428897142,
-0.08933887630701065,
-0.00434132432565093,
-0.06283523142337799,
-0.07909105718135834,
0.043802373111248016,
-0.12476077675819397,
-0.024092787876725197,
0.2181084007024765,
0.01601947657763958,
0.031246930360794067,
-0.05494377017021179,
0.18576189875602722,
-0.009880712255835533,
-0.03940189629793167,
0.16586916148662567,
-0.01927039958536625,
0.03855458274483681,
0.15495267510414124,
0.011838171631097794,
-0.08688342571258545,
0.07673905789852142,
-0.029861317947506905,
-0.027374960482120514,
-0.16827046871185303,
-0.08797821402549744,
-0.019058354198932648,
0.0029649899806827307,
0.08240267634391785,
-0.01751703955233097,
0.08579690009355545,
0.0695563480257988,
-0.021018480882048607,
-0.0005222564213909209,
0.04430143162608147,
0.07364543527364731,
0.16303817927837372,
0.0025723474100232124,
0.16941216588020325,
-0.014031201601028442,
-0.12389104068279266,
0.02420184761285782,
0.018687864765524864,
0.08982378989458084,
0.0533696673810482,
0.012395033612847328,
0.08919225633144379,
0.021671243011951447,
0.07004024088382721,
0.006178485695272684,
-0.0004188550228718668,
0.006742182187736034,
-0.037925828248262405,
-0.0244216937571764,
0.005592566914856434,
0.0764646902680397,
0.001343207317404449,
-0.02597714215517044,
-0.13366010785102844,
-0.009858713485300541,
0.044687867164611816,
0.1371791809797287,
0.17257051169872284,
-0.3082663416862488,
-0.05881521478295326,
0.036572180688381195,
-0.06716003268957138,
0.0197786632925272,
0.10509262979030609,
-0.024768933653831482,
-0.1274256408214569,
0.02793874219059944,
-0.003028233302757144,
0.09369175136089325,
0.004122328013181686,
0.026334088295698166,
-0.0601830892264843,
0.008279175497591496,
0.02421797625720501,
0.12616759538650513,
-0.36846140027046204,
0.09926177561283112,
-0.026310620829463005,
0.020479802042245865,
-0.04275868460536003,
-0.016107743605971336,
0.03927360102534294,
0.15804648399353027,
0.09322899580001831,
-0.03330657631158829,
-0.05638880655169487,
-0.031805459409952164,
-0.066896952688694,
0.027491994202136993,
0.03170914947986603,
-0.05940031260251999,
0.0503549762070179,
-0.04554425925016403,
-0.007709940429776907,
0.024181507527828217,
0.026194728910923004,
-0.020046330988407135,
-0.06001001596450806,
-0.00879651214927435,
0.007586304564028978,
0.072652667760849,
-0.007978993467986584,
-0.06681016832590103,
-0.13694220781326294,
0.052377160638570786,
-0.10020110011100769,
-0.06997738033533096,
-0.0902373418211937,
0.11111807078123093,
0.07158495485782623,
-0.05788290500640869,
-0.00672223512083292,
0.011055811308324337,
0.12987114489078522,
0.016508178785443306,
-0.13435445725917816,
0.08141613006591797,
-0.07861921191215515,
-0.10619295388460159,
-0.059240348637104034,
0.06913714110851288,
0.04407327249646187,
0.04800542816519737,
0.07065581530332565,
0.048081379383802414,
-0.09482857584953308,
-0.0784943625330925,
-0.03114073909819126,
0.010291383601725101,
0.05988830700516701,
0.020789729431271553,
-0.015861110761761665,
-0.036075446754693985,
-0.07026886940002441,
0.05532866343855858,
0.1666422188282013,
0.049988165497779846,
-0.07837219536304474,
0.10538347065448761,
0.19119104743003845,
-0.03443188592791557,
-0.2710075378417969,
-0.09393182396888733,
0.052779968827962875,
0.05966762825846672,
-0.0036762533709406853,
-0.09317929297685623,
0.12003692984580994,
0.01758931390941143,
-0.03927377238869667,
-0.20792117714881897,
-0.30292636156082153,
-0.1059204638004303,
0.19534169137477875,
0.045353010296821594,
0.19527271389961243,
-0.06996535509824753,
0.0006368850008584559,
-0.0701228529214859,
-0.02188224345445633,
0.13054832816123962,
-0.03208127245306969,
0.07262752205133438,
0.029194189235568047,
0.026614254340529442,
-0.00354999047704041,
-0.012914572842419147,
0.07330986857414246,
0.11169873923063278,
0.04423724114894867,
-0.0230934526771307,
0.008428706787526608,
0.01551474817097187,
-0.02156813070178032,
0.1388540118932724,
-0.03433184698224068,
0.10813695192337036,
-0.037260331213474274,
-0.09222746640443802,
-0.02682902291417122,
-0.007109030149877071,
0.020407337695360184,
-0.045286476612091064,
-0.003708033822476864,
0.01889204792678356,
0.04637684300541878,
-0.008154493756592274,
0.03212818130850792,
-0.06161528453230858,
0.023317275568842888,
0.08996660262346268,
0.1535702794790268,
-0.007848595269024372,
-0.030906114727258682,
-0.018208878114819527,
-0.044997502118349075,
0.10206405073404312,
-0.1243642196059227,
0.09742600470781326,
0.08768989145755768,
-0.023522846400737762,
0.1322346031665802,
0.10066074132919312,
-0.028441043570637703,
0.007724082097411156,
0.03397630527615547,
-0.08349581062793732,
-0.07828566431999207,
-0.0177594106644392,
-0.010412252508103848,
-0.009639780037105083,
0.06921340525150299,
0.12561507523059845,
-0.04973725974559784,
-0.049510978162288666,
-0.0014399120118469,
0.03787189722061157,
-0.06970417499542236,
0.1286851465702057,
0.04328149929642677,
0.02771693840622902,
-0.10920175164937973,
0.08079849183559418,
0.06557505577802658,
-0.02886749431490898,
0.05063258111476898,
0.002709042513743043,
-0.14908482134342194,
-0.07961821556091309,
-0.038368768990039825,
0.16730418801307678,
-0.1231829896569252,
-0.1041891947388649,
-0.12432868778705597,
-0.1195027232170105,
0.055146776139736176,
0.11373137682676315,
0.1308000683784485,
0.07971426844596863,
-0.08254364877939224,
-0.06629341840744019,
-0.06873215734958649,
0.07043080776929855,
0.0590834766626358,
0.0014476168435066938,
-0.07249955087900162,
0.1078842356801033,
0.02953767031431198,
0.06995809078216553,
-0.07426763325929642,
-0.10544809699058533,
-0.11137775331735611,
0.06761300563812256,
-0.09696733206510544,
-0.0013616761425510049,
-0.07044665515422821,
-0.020865771919488907,
0.03833087161183357,
-0.06195796653628349,
-0.02672228403389454,
0.009703190997242928,
-0.06699907779693604,
0.02723001316189766,
-0.046083513647317886,
0.06712814420461655,
-0.11217520385980606,
-0.023237034678459167,
0.026498185470700264,
-0.03590751066803932,
0.04223775491118431,
0.09461784362792969,
-0.03911161795258522,
0.061090435832738876,
-0.09148291498422623,
-0.04641120508313179,
-0.001979498891159892,
0.056505050510168076,
0.030138403177261353,
-0.12308505177497864,
0.007712612394243479,
0.08656618744134903,
0.012758440338075161,
0.027216410264372826,
0.05956273898482323,
-0.06229201331734657,
0.024123365059494972,
-0.0008471736800856888,
-0.04460861161351204,
-0.08894117176532745,
-0.018898092210292816,
0.025442427024245262,
0.1234932616353035,
0.17448166012763977,
-0.06430473178625107,
0.04541806876659393,
-0.12680687010288239,
0.005586212035268545,
-0.026076631620526314,
-0.03762659430503845,
-0.11931457370519638,
-0.10948051512241364,
0.021586015820503235,
0.007657730486243963,
0.18232405185699463,
0.048868268728256226,
0.09762267023324966,
0.02667461894452572,
0.07704909890890121,
0.14989763498306274,
0.0837806686758995,
0.15277248620986938,
0.08505940437316895,
-0.001308452570810914,
-0.06153569743037224,
0.06000564247369766,
0.04609448090195656,
0.01871381513774395,
0.01030363142490387,
0.08698059618473053,
-0.031070802360773087,
0.09347410500049591,
0.04217350855469704,
0.07338280230760574,
-0.029209233820438385,
-0.07380496710538864,
-0.06661934405565262,
-0.027354905381798744,
0.01160237193107605,
0.1700151115655899,
0.1219412311911583,
-0.09097740799188614,
0.03803998976945877,
0.0007353593246079981,
-0.07200080156326294,
-0.1551312804222107,
-0.1218455508351326,
-0.0787975862622261,
-0.1789652556180954,
-0.020515741780400276,
-0.09347008168697357,
-0.03238793835043907,
0.1722450852394104,
0.061212293803691864,
-0.020552337169647217,
0.036514975130558014,
-0.026509378105401993,
-0.026977956295013428,
-0.006232324056327343,
-0.08245831727981567,
0.07446432113647461,
0.06176842749118805,
-0.01726594753563404,
0.07555916905403137,
0.043883468955755234,
0.059370145201683044,
-0.02057473734021187,
0.01889934577047825,
0.0017426462145522237,
-0.08713501691818237,
-0.08469713479280472,
-0.014220773242413998,
0.0037208078429102898,
-0.013231090269982815,
0.024186650291085243,
0.06992467492818832,
-0.03062121942639351,
0.011872920207679272,
0.19517837464809418,
-0.08884906023740768,
-0.13992168009281158,
-0.23437249660491943,
0.35079628229141235,
0.05717453733086586,
0.07900752127170563,
0.04559033736586571,
-0.0946241095662117,
-0.06920360773801804,
0.25321730971336365,
0.13183432817459106,
-0.08776503056287766,
-0.027379212900996208,
0.06537112593650818,
-0.011876639910042286,
0.06274634599685669,
0.11160000413656235,
0.006347394082695246,
0.2118706852197647,
-0.09449732303619385,
0.045286647975444794,
-0.042395275086164474,
-0.07067117840051651,
0.00304989330470562,
0.048596229404211044,
-0.025210218504071236,
-0.051842279732227325,
-0.03354787826538086,
0.15831102430820465,
-0.08447630703449249,
0.04331846535205841,
-0.02344190701842308,
-0.005120161455124617,
-0.08395712822675705,
-0.014423919841647148,
0.03069530427455902,
0.005818168167024851,
0.08256194740533829,
-0.038627512753009796,
0.0474165640771389,
0.11462542414665222,
-0.0019665148574858904,
-0.062286101281642914,
-0.04901717230677605,
0.07197322696447372,
-0.0521366111934185,
0.00411691190674901,
-0.005958066321909428,
0.1981751173734665,
0.0844266265630722,
0.07160070538520813,
-0.01448103692382574,
0.06654112040996552,
-0.011041480116546154,
0.03469720855355263,
0.12156840413808823,
0.045243531465530396,
0.0045079803094267845,
0.04379255324602127,
0.045405611395835876,
-0.16603533923625946,
0.06943680346012115,
-0.10327289998531342,
-0.07379655539989471,
-0.11981610208749771,
0.04639599844813347,
-0.020899711176753044,
0.11346854269504547,
0.10301072150468826,
-0.050007306039333344,
0.015899445861577988,
0.013804791495203972,
0.018075525760650635,
-0.039431195706129074,
-0.09157073497772217,
-0.009393940679728985,
-0.18247821927070618,
0.005740893539041281,
-0.007659159600734711,
-0.0004097088531125337,
-0.2470398247241974,
-0.013012047857046127,
-0.011412807740271091,
-0.031997185200452805,
0.00288157700560987,
0.08741876482963562,
0.06337825208902359,
0.041867293417453766,
-0.07839862257242203,
-0.06604918837547302,
-0.00466717267408967,
0.09431437402963638,
-0.11357368528842926,
-0.16285744309425354
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')])
```
The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`.
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/stsb-distilroberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
52,
34,
38,
85
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.08450348675251007,
0.06658493727445602,
-0.0020891379099339247,
0.06090733781456947,
0.1219165176153183,
-0.010885688476264477,
0.18726801872253418,
0.06976161152124405,
0.004681415390223265,
-0.032464902848005295,
0.10880614817142487,
0.10593046247959137,
0.009075884707272053,
0.1311509907245636,
-0.0318918414413929,
-0.2397564798593521,
0.05535270273685455,
-0.021333403885364532,
-0.023506149649620056,
0.1251000612974167,
0.10538756102323532,
-0.08173905313014984,
0.06480519473552704,
0.036619383841753006,
-0.1481119692325592,
0.0361047126352787,
0.004630628507584333,
-0.06398891657590866,
0.1055711954832077,
0.029057493433356285,
0.12339451909065247,
0.09259660542011261,
0.07925070077180862,
-0.12724295258522034,
0.012720810249447823,
0.016518067568540573,
0.009311544708907604,
0.03629666194319725,
0.036708857864141464,
-0.10701856762170792,
-0.09100929647684097,
0.08599001914262772,
0.08524163067340851,
0.0009422770235687494,
-0.11783528327941895,
-0.1288551539182663,
-0.02296801283955574,
-0.0451493114233017,
0.1464175432920456,
0.0878152996301651,
-0.045602668076753616,
0.1485908329486847,
-0.1755804717540741,
0.04145662114024162,
0.054525889456272125,
-0.3068389296531677,
0.010614708065986633,
0.0938933864235878,
0.023070018738508224,
0.04588997736573219,
-0.04204950109124184,
-0.006081583444029093,
0.028909359127283096,
0.04380388930439949,
0.06281910091638565,
-0.04001569375395775,
-0.010886033065617085,
0.037939924746751785,
-0.1775822639465332,
-0.03491581231355667,
0.21934248507022858,
-0.03001006878912449,
-0.03818020224571228,
-0.10346876084804535,
-0.030659932643175125,
0.05167389288544655,
-0.022362971678376198,
-0.08230966329574585,
0.004080281592905521,
0.05453440919518471,
-0.09994271397590637,
0.043615926057100296,
-0.12259428203105927,
-0.1212959885597229,
-0.11897437274456024,
0.1481919288635254,
0.0519876554608345,
0.05614864081144333,
-0.11817987263202667,
0.11007988452911377,
-0.018013888970017433,
-0.08047500997781754,
-0.03508051112294197,
-0.07759242504835129,
-0.13628479838371277,
-0.017040323466062546,
-0.1888146996498108,
-0.1461862325668335,
0.019468307495117188,
0.23470254242420197,
0.05425926297903061,
-0.003253627335652709,
0.05223825201392174,
0.052410710602998734,
0.05260057747364044,
0.16987985372543335,
-0.09921816736459732,
0.019514871761202812,
-0.040489714592695236,
-0.006623496767133474,
-0.014924317598342896,
-0.020644033327698708,
-0.15668311715126038,
-0.02293773740530014,
0.06836400926113129,
0.03340352699160576,
-0.03544469177722931,
0.060046419501304626,
-0.03591883182525635,
-0.04250994324684143,
0.04778998717665672,
-0.07110997289419174,
0.007199421525001526,
0.001498668105341494,
-0.08984625339508057,
0.04597581550478935,
0.004727287217974663,
0.05564960837364197,
-0.04794371500611305,
-0.03951052203774452,
-0.05796322971582413,
-0.0041662254370749,
-0.07994621992111206,
-0.13172027468681335,
-0.017280934378504753,
0.0645439475774765,
0.004875025246292353,
-0.12934610247612,
-0.15153756737709045,
-0.06250552833080292,
0.06014043092727661,
-0.025576326996088028,
-0.0060349274426698685,
-0.10992851108312607,
-0.024137191474437714,
0.014359449967741966,
-0.03408343344926834,
-0.012038946151733398,
-0.006700094789266586,
0.027012227103114128,
-0.010106818750500679,
0.10634177178144455,
0.011723324656486511,
0.056721992790699005,
-0.17293746769428253,
-0.014487387612462044,
-0.06987504661083221,
0.13998807966709137,
0.00621027173474431,
0.036297235637903214,
-0.07619620859622955,
-0.03155680000782013,
-0.03006790578365326,
0.06660110503435135,
0.037052832543849945,
0.1293592005968094,
-0.16096548736095428,
-0.06480851024389267,
0.2344948649406433,
-0.12800095975399017,
-0.11015582084655762,
0.12573368847370148,
-0.049440640956163406,
0.03731274977326393,
0.10660473257303238,
0.12351691722869873,
0.0702773854136467,
-0.034878578037023544,
0.08157497644424438,
0.06925155967473984,
-0.0970594733953476,
-0.01633760891854763,
0.10197000205516815,
0.03484668955206871,
-0.08979984372854233,
0.031350795179605484,
-0.06290501356124878,
0.05259009823203087,
-0.04869356006383896,
-0.06925348937511444,
-0.013377794064581394,
-0.054921574890613556,
0.07530007511377335,
0.020727040246129036,
0.08574032783508301,
-0.02007264457643032,
-0.10619869083166122,
0.11279048770666122,
0.08791369199752808,
-0.0893058329820633,
0.029166527092456818,
-0.10053735971450806,
0.013915755786001682,
-0.07938611507415771,
-0.007064095698297024,
-0.20703130960464478,
0.006485046353191137,
0.0018438141560181975,
0.014914322644472122,
0.034115370362997055,
0.10435596108436584,
0.030128762125968933,
0.022408952936530113,
0.029981357976794243,
-0.018031882122159004,
0.08979107439517975,
0.009407839737832546,
-0.08929740637540817,
-0.10686875879764557,
-0.02606169506907463,
-0.03660851716995239,
0.15328125655651093,
-0.013654633425176144,
0.02513095922768116,
-0.09524213522672653,
0.0469084158539772,
-0.029759293422102928,
0.05308831110596657,
-0.004855465143918991,
0.04144696891307831,
-0.04124278575181961,
0.0014224101323634386,
0.04660613834857941,
0.017619729042053223,
-0.09673509001731873,
0.1709372103214264,
-0.18291769921779633,
-0.02089911699295044,
0.12491315603256226,
-0.11515885591506958,
-0.06447350978851318,
0.02345173992216587,
-0.014730798080563545,
0.02202417515218258,
-0.05534713342785835,
0.014598351903259754,
0.16470010578632355,
0.014831818640232086,
0.12296251952648163,
-0.083440862596035,
-0.026016926392912865,
-0.018848687410354614,
-0.09721343219280243,
0.018556935712695122,
0.04981476813554764,
0.039000771939754486,
-0.1162211075425148,
0.10053393989801407,
0.10247835516929626,
-0.10287507623434067,
0.015802966430783272,
0.004366625100374222,
-0.04628602787852287,
-0.03741034120321274,
-0.04338941350579262,
-0.03472225368022919,
-0.027288775891065598,
-0.029719965532422066,
-0.042659372091293335,
0.03510940074920654,
0.029411355033516884,
0.013905813917517662,
-0.15196870267391205,
0.005906535778194666,
0.013755157589912415,
-0.011594702489674091,
-0.01026098895817995,
0.04376592859625816,
-0.008289550431072712,
0.061599910259246826,
0.028126584365963936,
-0.14367526769638062,
0.058363452553749084,
-0.0351489894092083,
-0.14141881465911865,
0.22615239024162292,
-0.07566404342651367,
-0.27953410148620605,
-0.12290982156991959,
0.021696176379919052,
0.006087016314268112,
0.062644362449646,
0.0708455741405487,
-0.045471686869859695,
-0.03838431462645531,
-0.0820717066526413,
-0.03139931708574295,
-0.024082103744149208,
0.0024445243179798126,
-0.0878162756562233,
0.00009271603630622849,
-0.003847202518954873,
-0.09421621263027191,
-0.0031321351416409016,
-0.05890399217605591,
-0.06755509972572327,
0.04214369133114815,
-0.09165877103805542,
-0.007889453321695328,
0.2206796407699585,
0.01076422818005085,
0.03367805480957031,
-0.05330446735024452,
0.17513544857501984,
-0.015779828652739525,
-0.03460876643657684,
0.1611550748348236,
-0.012880779802799225,
0.04445143789052963,
0.17388220131397247,
0.019242996349930763,
-0.0683070495724678,
0.07065898925065994,
-0.03215143457055092,
-0.04168960079550743,
-0.1452464610338211,
-0.09278552234172821,
-0.038062673062086105,
-0.0233317743986845,
0.07220902293920517,
-0.01197992917150259,
0.08731245249509811,
0.05724746733903885,
-0.003093191422522068,
-0.01159081980586052,
0.03088546171784401,
0.0780472680926323,
0.1442079395055771,
0.003124864771962166,
0.17355042695999146,
-0.016894925385713577,
-0.11764701455831528,
0.027372056618332863,
0.02583908475935459,
0.10140083730220795,
0.05113178864121437,
0.009328491054475307,
0.10008900612592697,
0.027490992099046707,
0.07921173423528671,
-0.004589700140058994,
0.004766227677464485,
0.0007489954587072134,
-0.03577183932065964,
-0.029451249167323112,
0.016168830916285515,
0.09111353009939194,
0.01197222899645567,
-0.02815566398203373,
-0.12441352754831314,
-0.03104766085743904,
0.04662104323506355,
0.11217736452817917,
0.16801334917545319,
-0.30731672048568726,
-0.05213860049843788,
0.04639245942234993,
-0.06299179047346115,
0.023163730278611183,
0.09142906963825226,
-0.004638497717678547,
-0.13518525660037994,
0.00644878763705492,
-0.004350210539996624,
0.08985988050699234,
0.0018690075958147645,
0.02098994515836239,
-0.06347128748893738,
0.02823813259601593,
0.014463197439908981,
0.12384771555662155,
-0.34352806210517883,
0.10411574691534042,
-0.029895853251218796,
0.011633627116680145,
-0.04604942351579666,
-0.006390930153429508,
0.04763137176632881,
0.14670772850513458,
0.08665851503610611,
-0.02811373583972454,
-0.09105808287858963,
-0.06061112508177757,
-0.04584410414099693,
0.022072266787290573,
0.025318568572402,
-0.05517273768782616,
0.05771936476230621,
-0.04354587942361832,
-0.005889877211302519,
0.02616376057267189,
0.05868706852197647,
-0.025132019072771072,
-0.07593091577291489,
-0.007170701399445534,
0.013278568163514137,
0.05211564898490906,
-0.015402347780764103,
-0.06917586177587509,
-0.13652382791042328,
0.042136695235967636,
-0.07789728045463562,
-0.06028033047914505,
-0.103256456553936,
0.09486180543899536,
0.07285741716623306,
-0.06623293459415436,
0.0039231302216649055,
0.016175266355276108,
0.133241206407547,
0.004705662373453379,
-0.1416292041540146,
0.08322334289550781,
-0.07802871614694595,
-0.08273158967494965,
-0.06269396841526031,
0.03027677722275257,
0.04064171016216278,
0.04588952288031578,
0.0636964812874794,
0.06164630129933357,
-0.11224881559610367,
-0.07974357903003693,
-0.02778312936425209,
0.016498364508152008,
0.05221463739871979,
0.024097010493278503,
-0.005767281167209148,
-0.055050041526556015,
-0.03973664715886116,
0.06736788898706436,
0.17233264446258545,
0.0396297462284565,
-0.09726333618164062,
0.12583638727664948,
0.17738917469978333,
-0.04895620420575142,
-0.283265620470047,
-0.09362755715847015,
0.04195278882980347,
0.07143981009721756,
0.014062013477087021,
-0.09430427849292755,
0.1031544879078865,
0.012528144754469395,
-0.048478029668331146,
-0.2168145477771759,
-0.3028346598148346,
-0.10955806821584702,
0.18833613395690918,
0.03733060136437416,
0.1766844391822815,
-0.07329234480857849,
0.00888361781835556,
-0.06110344082117081,
0.008218829520046711,
0.149373859167099,
-0.06190621480345726,
0.0771213248372078,
0.031203610822558403,
0.02545219659805298,
-0.0002090671769110486,
-0.01993192359805107,
0.07104034721851349,
0.08380923420190811,
0.060775671154260635,
-0.02233022265136242,
-0.005947250407189131,
0.030198819935321808,
-0.038693733513355255,
0.13508112728595734,
-0.038966890424489975,
0.09695424884557724,
-0.060185179114341736,
-0.08268783241510391,
-0.033626738935709,
0.0095839723944664,
0.021442634984850883,
-0.047682344913482666,
-0.014281190931797028,
0.0293685682117939,
0.04048173502087593,
-0.010960451327264309,
0.06327590346336365,
-0.06340527534484863,
0.045411452651023865,
0.11049935221672058,
0.13265644013881683,
-0.01526104100048542,
-0.03578300401568413,
0.0051394738256931305,
-0.047414444386959076,
0.11430126428604126,
-0.164499893784523,
0.0911836102604866,
0.07006113976240158,
-0.022437749430537224,
0.1343110352754593,
0.0996335968375206,
-0.04331466183066368,
0.018353044986724854,
0.052883587777614594,
-0.0847473293542862,
-0.06420569121837616,
-0.015110892243683338,
-0.007178271654993296,
-0.005094901192933321,
0.0529211089015007,
0.11877772212028503,
-0.05779349431395531,
-0.04995092377066612,
-0.008740345947444439,
0.035886529833078384,
-0.06572326272726059,
0.11153262853622437,
0.06726455688476562,
0.032411862164735794,
-0.09644010663032532,
0.06804357469081879,
0.07756434381008148,
-0.02221454121172428,
0.06189196556806564,
0.018371189013123512,
-0.13287536799907684,
-0.09099080413579941,
-0.03262708708643913,
0.17415082454681396,
-0.1126992255449295,
-0.09605299681425095,
-0.1218004897236824,
-0.10991249978542328,
0.04737558588385582,
0.11134099960327148,
0.13007406890392303,
0.07032902538776398,
-0.07839404791593552,
-0.06061587482690811,
-0.07010961323976517,
0.0824027732014656,
0.05524514988064766,
-0.002843055408447981,
-0.07459040731191635,
0.11347991228103638,
0.031216993927955627,
0.0675557404756546,
-0.08114106953144073,
-0.10859313607215881,
-0.128158301115036,
0.06127750873565674,
-0.06947355717420578,
-0.005555060692131519,
-0.07801594585180283,
-0.010933276265859604,
0.03555882349610329,
-0.05734606087207794,
-0.0364067368209362,
0.013651168905198574,
-0.08039981126785278,
0.025344787165522575,
-0.037574876099824905,
0.06731393933296204,
-0.10970719158649445,
-0.029593832790851593,
0.023546205833554268,
-0.03511421009898186,
0.04349852725863457,
0.07940305769443512,
-0.02931799553334713,
0.04794476926326752,
-0.0984886959195137,
-0.04244762659072876,
-0.0008963823784142733,
0.05560130998492241,
0.04657924920320511,
-0.11562971025705338,
0.011889439076185226,
0.08023174852132797,
0.02496957965195179,
0.01808847486972809,
0.04983857646584511,
-0.06669441610574722,
0.03893890976905823,
-0.005539896432310343,
-0.04568590968847275,
-0.08155433088541031,
0.0019124216632917523,
0.029221396893262863,
0.11990660429000854,
0.17286483943462372,
-0.05838547274470329,
0.0494680181145668,
-0.12196874618530273,
0.002273800317198038,
-0.03104289248585701,
-0.03879985958337784,
-0.12439896911382675,
-0.10816622525453568,
0.023254498839378357,
0.005873893387615681,
0.19243818521499634,
0.08622796088457108,
0.09015563130378723,
0.02026992477476597,
0.08280457556247711,
0.15640299022197723,
0.08169982582330704,
0.13479216396808624,
0.0753636434674263,
0.003552020760253072,
-0.04583922401070595,
0.06687968969345093,
0.04342857003211975,
0.0072935158386826515,
0.0038177764508873224,
0.09528568387031555,
0.00667149294167757,
0.10526926070451736,
0.042451679706573486,
0.0500829704105854,
-0.032434239983558655,
-0.06250517815351486,
-0.0641176775097847,
-0.01888967491686344,
-0.011569607071578503,
0.12127377092838287,
0.14277249574661255,
-0.08079589903354645,
0.04679444059729576,
0.015447227284312248,
-0.06474620848894119,
-0.1609332412481308,
-0.1257009506225586,
-0.07168443500995636,
-0.18344774842262268,
-0.027470940724015236,
-0.09683065861463547,
-0.02620367519557476,
0.16843217611312866,
0.04557392746210098,
-0.009155435487627983,
0.02763553522527218,
0.0004341208259575069,
-0.02509470470249653,
-0.020050279796123505,
-0.08659908920526505,
0.07982605695724487,
0.04354038089513779,
-0.010250295512378216,
0.07036767154932022,
0.038146670907735825,
0.051697567105293274,
-0.016725361347198486,
0.016275659203529358,
0.014421421103179455,
-0.0931367501616478,
-0.08636751025915146,
-0.014829674735665321,
0.006926656700670719,
-0.012881503440439701,
0.022557280957698822,
0.07137629389762878,
-0.03945473954081535,
0.001135995495133102,
0.21051138639450073,
-0.09168606251478195,
-0.13335897028446198,
-0.22504109144210815,
0.3292695879936218,
0.06128848344087601,
0.08006542921066284,
0.01782870665192604,
-0.09772481769323349,
-0.08380256593227386,
0.23835910856723785,
0.15601664781570435,
-0.07657447457313538,
-0.014831600710749626,
0.07228817790746689,
-0.016094770282506943,
0.04914749786257744,
0.11798988282680511,
-0.0018936273409053683,
0.22898456454277039,
-0.08471135795116425,
0.04600175470113754,
-0.03982498124241829,
-0.06657011061906815,
-0.0016606205608695745,
0.04384338855743408,
-0.028883112594485283,
-0.0646972507238388,
-0.0373566634953022,
0.14134474098682404,
-0.09122554212808609,
0.049046847969293594,
-0.0002345513494219631,
-0.022823620587587357,
-0.07716405391693115,
-0.006424167193472385,
0.06062676012516022,
0.00020406620751600713,
0.08504018932580948,
-0.038146086037158966,
0.03794025629758835,
0.11565134674310684,
-0.0017569564515724778,
-0.06979916244745255,
-0.058996688574552536,
0.06873651593923569,
-0.06620217114686966,
0.02624395117163658,
-0.020865874364972115,
0.19483453035354614,
0.09904284030199051,
0.06483505666255951,
-0.03644286096096039,
0.0626656711101532,
-0.013008199632167816,
0.011695616878569126,
0.12002301216125488,
0.03363015130162239,
0.0020754639990627766,
0.06408432871103287,
0.044256992638111115,
-0.1768614798784256,
0.07748190313577652,
-0.14381177723407745,
-0.06964220106601715,
-0.10966911166906357,
0.02179321087896824,
-0.012797866016626358,
0.1110919713973999,
0.09294867515563965,
-0.04715340584516525,
0.012934789061546326,
0.018140526488423347,
0.015916945412755013,
-0.041642192751169205,
-0.078044094145298,
-0.03142290934920311,
-0.16770023107528687,
0.007813815027475357,
-0.03266200050711632,
0.002701889956369996,
-0.22878091037273407,
-0.011267242021858692,
-0.008542446419596672,
-0.033176716417074203,
0.008387149311602116,
0.08422781527042389,
0.08410125225782394,
0.03710673004388809,
-0.07843427360057831,
-0.06219158694148064,
0.0013762605376541615,
0.09299024194478989,
-0.12327927350997925,
-0.15334023535251617
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')])
```
The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`.
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/stsb-roberta-base
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
52,
34,
38,
85
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.08450348675251007,
0.06658493727445602,
-0.0020891379099339247,
0.06090733781456947,
0.1219165176153183,
-0.010885688476264477,
0.18726801872253418,
0.06976161152124405,
0.004681415390223265,
-0.032464902848005295,
0.10880614817142487,
0.10593046247959137,
0.009075884707272053,
0.1311509907245636,
-0.0318918414413929,
-0.2397564798593521,
0.05535270273685455,
-0.021333403885364532,
-0.023506149649620056,
0.1251000612974167,
0.10538756102323532,
-0.08173905313014984,
0.06480519473552704,
0.036619383841753006,
-0.1481119692325592,
0.0361047126352787,
0.004630628507584333,
-0.06398891657590866,
0.1055711954832077,
0.029057493433356285,
0.12339451909065247,
0.09259660542011261,
0.07925070077180862,
-0.12724295258522034,
0.012720810249447823,
0.016518067568540573,
0.009311544708907604,
0.03629666194319725,
0.036708857864141464,
-0.10701856762170792,
-0.09100929647684097,
0.08599001914262772,
0.08524163067340851,
0.0009422770235687494,
-0.11783528327941895,
-0.1288551539182663,
-0.02296801283955574,
-0.0451493114233017,
0.1464175432920456,
0.0878152996301651,
-0.045602668076753616,
0.1485908329486847,
-0.1755804717540741,
0.04145662114024162,
0.054525889456272125,
-0.3068389296531677,
0.010614708065986633,
0.0938933864235878,
0.023070018738508224,
0.04588997736573219,
-0.04204950109124184,
-0.006081583444029093,
0.028909359127283096,
0.04380388930439949,
0.06281910091638565,
-0.04001569375395775,
-0.010886033065617085,
0.037939924746751785,
-0.1775822639465332,
-0.03491581231355667,
0.21934248507022858,
-0.03001006878912449,
-0.03818020224571228,
-0.10346876084804535,
-0.030659932643175125,
0.05167389288544655,
-0.022362971678376198,
-0.08230966329574585,
0.004080281592905521,
0.05453440919518471,
-0.09994271397590637,
0.043615926057100296,
-0.12259428203105927,
-0.1212959885597229,
-0.11897437274456024,
0.1481919288635254,
0.0519876554608345,
0.05614864081144333,
-0.11817987263202667,
0.11007988452911377,
-0.018013888970017433,
-0.08047500997781754,
-0.03508051112294197,
-0.07759242504835129,
-0.13628479838371277,
-0.017040323466062546,
-0.1888146996498108,
-0.1461862325668335,
0.019468307495117188,
0.23470254242420197,
0.05425926297903061,
-0.003253627335652709,
0.05223825201392174,
0.052410710602998734,
0.05260057747364044,
0.16987985372543335,
-0.09921816736459732,
0.019514871761202812,
-0.040489714592695236,
-0.006623496767133474,
-0.014924317598342896,
-0.020644033327698708,
-0.15668311715126038,
-0.02293773740530014,
0.06836400926113129,
0.03340352699160576,
-0.03544469177722931,
0.060046419501304626,
-0.03591883182525635,
-0.04250994324684143,
0.04778998717665672,
-0.07110997289419174,
0.007199421525001526,
0.001498668105341494,
-0.08984625339508057,
0.04597581550478935,
0.004727287217974663,
0.05564960837364197,
-0.04794371500611305,
-0.03951052203774452,
-0.05796322971582413,
-0.0041662254370749,
-0.07994621992111206,
-0.13172027468681335,
-0.017280934378504753,
0.0645439475774765,
0.004875025246292353,
-0.12934610247612,
-0.15153756737709045,
-0.06250552833080292,
0.06014043092727661,
-0.025576326996088028,
-0.0060349274426698685,
-0.10992851108312607,
-0.024137191474437714,
0.014359449967741966,
-0.03408343344926834,
-0.012038946151733398,
-0.006700094789266586,
0.027012227103114128,
-0.010106818750500679,
0.10634177178144455,
0.011723324656486511,
0.056721992790699005,
-0.17293746769428253,
-0.014487387612462044,
-0.06987504661083221,
0.13998807966709137,
0.00621027173474431,
0.036297235637903214,
-0.07619620859622955,
-0.03155680000782013,
-0.03006790578365326,
0.06660110503435135,
0.037052832543849945,
0.1293592005968094,
-0.16096548736095428,
-0.06480851024389267,
0.2344948649406433,
-0.12800095975399017,
-0.11015582084655762,
0.12573368847370148,
-0.049440640956163406,
0.03731274977326393,
0.10660473257303238,
0.12351691722869873,
0.0702773854136467,
-0.034878578037023544,
0.08157497644424438,
0.06925155967473984,
-0.0970594733953476,
-0.01633760891854763,
0.10197000205516815,
0.03484668955206871,
-0.08979984372854233,
0.031350795179605484,
-0.06290501356124878,
0.05259009823203087,
-0.04869356006383896,
-0.06925348937511444,
-0.013377794064581394,
-0.054921574890613556,
0.07530007511377335,
0.020727040246129036,
0.08574032783508301,
-0.02007264457643032,
-0.10619869083166122,
0.11279048770666122,
0.08791369199752808,
-0.0893058329820633,
0.029166527092456818,
-0.10053735971450806,
0.013915755786001682,
-0.07938611507415771,
-0.007064095698297024,
-0.20703130960464478,
0.006485046353191137,
0.0018438141560181975,
0.014914322644472122,
0.034115370362997055,
0.10435596108436584,
0.030128762125968933,
0.022408952936530113,
0.029981357976794243,
-0.018031882122159004,
0.08979107439517975,
0.009407839737832546,
-0.08929740637540817,
-0.10686875879764557,
-0.02606169506907463,
-0.03660851716995239,
0.15328125655651093,
-0.013654633425176144,
0.02513095922768116,
-0.09524213522672653,
0.0469084158539772,
-0.029759293422102928,
0.05308831110596657,
-0.004855465143918991,
0.04144696891307831,
-0.04124278575181961,
0.0014224101323634386,
0.04660613834857941,
0.017619729042053223,
-0.09673509001731873,
0.1709372103214264,
-0.18291769921779633,
-0.02089911699295044,
0.12491315603256226,
-0.11515885591506958,
-0.06447350978851318,
0.02345173992216587,
-0.014730798080563545,
0.02202417515218258,
-0.05534713342785835,
0.014598351903259754,
0.16470010578632355,
0.014831818640232086,
0.12296251952648163,
-0.083440862596035,
-0.026016926392912865,
-0.018848687410354614,
-0.09721343219280243,
0.018556935712695122,
0.04981476813554764,
0.039000771939754486,
-0.1162211075425148,
0.10053393989801407,
0.10247835516929626,
-0.10287507623434067,
0.015802966430783272,
0.004366625100374222,
-0.04628602787852287,
-0.03741034120321274,
-0.04338941350579262,
-0.03472225368022919,
-0.027288775891065598,
-0.029719965532422066,
-0.042659372091293335,
0.03510940074920654,
0.029411355033516884,
0.013905813917517662,
-0.15196870267391205,
0.005906535778194666,
0.013755157589912415,
-0.011594702489674091,
-0.01026098895817995,
0.04376592859625816,
-0.008289550431072712,
0.061599910259246826,
0.028126584365963936,
-0.14367526769638062,
0.058363452553749084,
-0.0351489894092083,
-0.14141881465911865,
0.22615239024162292,
-0.07566404342651367,
-0.27953410148620605,
-0.12290982156991959,
0.021696176379919052,
0.006087016314268112,
0.062644362449646,
0.0708455741405487,
-0.045471686869859695,
-0.03838431462645531,
-0.0820717066526413,
-0.03139931708574295,
-0.024082103744149208,
0.0024445243179798126,
-0.0878162756562233,
0.00009271603630622849,
-0.003847202518954873,
-0.09421621263027191,
-0.0031321351416409016,
-0.05890399217605591,
-0.06755509972572327,
0.04214369133114815,
-0.09165877103805542,
-0.007889453321695328,
0.2206796407699585,
0.01076422818005085,
0.03367805480957031,
-0.05330446735024452,
0.17513544857501984,
-0.015779828652739525,
-0.03460876643657684,
0.1611550748348236,
-0.012880779802799225,
0.04445143789052963,
0.17388220131397247,
0.019242996349930763,
-0.0683070495724678,
0.07065898925065994,
-0.03215143457055092,
-0.04168960079550743,
-0.1452464610338211,
-0.09278552234172821,
-0.038062673062086105,
-0.0233317743986845,
0.07220902293920517,
-0.01197992917150259,
0.08731245249509811,
0.05724746733903885,
-0.003093191422522068,
-0.01159081980586052,
0.03088546171784401,
0.0780472680926323,
0.1442079395055771,
0.003124864771962166,
0.17355042695999146,
-0.016894925385713577,
-0.11764701455831528,
0.027372056618332863,
0.02583908475935459,
0.10140083730220795,
0.05113178864121437,
0.009328491054475307,
0.10008900612592697,
0.027490992099046707,
0.07921173423528671,
-0.004589700140058994,
0.004766227677464485,
0.0007489954587072134,
-0.03577183932065964,
-0.029451249167323112,
0.016168830916285515,
0.09111353009939194,
0.01197222899645567,
-0.02815566398203373,
-0.12441352754831314,
-0.03104766085743904,
0.04662104323506355,
0.11217736452817917,
0.16801334917545319,
-0.30731672048568726,
-0.05213860049843788,
0.04639245942234993,
-0.06299179047346115,
0.023163730278611183,
0.09142906963825226,
-0.004638497717678547,
-0.13518525660037994,
0.00644878763705492,
-0.004350210539996624,
0.08985988050699234,
0.0018690075958147645,
0.02098994515836239,
-0.06347128748893738,
0.02823813259601593,
0.014463197439908981,
0.12384771555662155,
-0.34352806210517883,
0.10411574691534042,
-0.029895853251218796,
0.011633627116680145,
-0.04604942351579666,
-0.006390930153429508,
0.04763137176632881,
0.14670772850513458,
0.08665851503610611,
-0.02811373583972454,
-0.09105808287858963,
-0.06061112508177757,
-0.04584410414099693,
0.022072266787290573,
0.025318568572402,
-0.05517273768782616,
0.05771936476230621,
-0.04354587942361832,
-0.005889877211302519,
0.02616376057267189,
0.05868706852197647,
-0.025132019072771072,
-0.07593091577291489,
-0.007170701399445534,
0.013278568163514137,
0.05211564898490906,
-0.015402347780764103,
-0.06917586177587509,
-0.13652382791042328,
0.042136695235967636,
-0.07789728045463562,
-0.06028033047914505,
-0.103256456553936,
0.09486180543899536,
0.07285741716623306,
-0.06623293459415436,
0.0039231302216649055,
0.016175266355276108,
0.133241206407547,
0.004705662373453379,
-0.1416292041540146,
0.08322334289550781,
-0.07802871614694595,
-0.08273158967494965,
-0.06269396841526031,
0.03027677722275257,
0.04064171016216278,
0.04588952288031578,
0.0636964812874794,
0.06164630129933357,
-0.11224881559610367,
-0.07974357903003693,
-0.02778312936425209,
0.016498364508152008,
0.05221463739871979,
0.024097010493278503,
-0.005767281167209148,
-0.055050041526556015,
-0.03973664715886116,
0.06736788898706436,
0.17233264446258545,
0.0396297462284565,
-0.09726333618164062,
0.12583638727664948,
0.17738917469978333,
-0.04895620420575142,
-0.283265620470047,
-0.09362755715847015,
0.04195278882980347,
0.07143981009721756,
0.014062013477087021,
-0.09430427849292755,
0.1031544879078865,
0.012528144754469395,
-0.048478029668331146,
-0.2168145477771759,
-0.3028346598148346,
-0.10955806821584702,
0.18833613395690918,
0.03733060136437416,
0.1766844391822815,
-0.07329234480857849,
0.00888361781835556,
-0.06110344082117081,
0.008218829520046711,
0.149373859167099,
-0.06190621480345726,
0.0771213248372078,
0.031203610822558403,
0.02545219659805298,
-0.0002090671769110486,
-0.01993192359805107,
0.07104034721851349,
0.08380923420190811,
0.060775671154260635,
-0.02233022265136242,
-0.005947250407189131,
0.030198819935321808,
-0.038693733513355255,
0.13508112728595734,
-0.038966890424489975,
0.09695424884557724,
-0.060185179114341736,
-0.08268783241510391,
-0.033626738935709,
0.0095839723944664,
0.021442634984850883,
-0.047682344913482666,
-0.014281190931797028,
0.0293685682117939,
0.04048173502087593,
-0.010960451327264309,
0.06327590346336365,
-0.06340527534484863,
0.045411452651023865,
0.11049935221672058,
0.13265644013881683,
-0.01526104100048542,
-0.03578300401568413,
0.0051394738256931305,
-0.047414444386959076,
0.11430126428604126,
-0.164499893784523,
0.0911836102604866,
0.07006113976240158,
-0.022437749430537224,
0.1343110352754593,
0.0996335968375206,
-0.04331466183066368,
0.018353044986724854,
0.052883587777614594,
-0.0847473293542862,
-0.06420569121837616,
-0.015110892243683338,
-0.007178271654993296,
-0.005094901192933321,
0.0529211089015007,
0.11877772212028503,
-0.05779349431395531,
-0.04995092377066612,
-0.008740345947444439,
0.035886529833078384,
-0.06572326272726059,
0.11153262853622437,
0.06726455688476562,
0.032411862164735794,
-0.09644010663032532,
0.06804357469081879,
0.07756434381008148,
-0.02221454121172428,
0.06189196556806564,
0.018371189013123512,
-0.13287536799907684,
-0.09099080413579941,
-0.03262708708643913,
0.17415082454681396,
-0.1126992255449295,
-0.09605299681425095,
-0.1218004897236824,
-0.10991249978542328,
0.04737558588385582,
0.11134099960327148,
0.13007406890392303,
0.07032902538776398,
-0.07839404791593552,
-0.06061587482690811,
-0.07010961323976517,
0.0824027732014656,
0.05524514988064766,
-0.002843055408447981,
-0.07459040731191635,
0.11347991228103638,
0.031216993927955627,
0.0675557404756546,
-0.08114106953144073,
-0.10859313607215881,
-0.128158301115036,
0.06127750873565674,
-0.06947355717420578,
-0.005555060692131519,
-0.07801594585180283,
-0.010933276265859604,
0.03555882349610329,
-0.05734606087207794,
-0.0364067368209362,
0.013651168905198574,
-0.08039981126785278,
0.025344787165522575,
-0.037574876099824905,
0.06731393933296204,
-0.10970719158649445,
-0.029593832790851593,
0.023546205833554268,
-0.03511421009898186,
0.04349852725863457,
0.07940305769443512,
-0.02931799553334713,
0.04794476926326752,
-0.0984886959195137,
-0.04244762659072876,
-0.0008963823784142733,
0.05560130998492241,
0.04657924920320511,
-0.11562971025705338,
0.011889439076185226,
0.08023174852132797,
0.02496957965195179,
0.01808847486972809,
0.04983857646584511,
-0.06669441610574722,
0.03893890976905823,
-0.005539896432310343,
-0.04568590968847275,
-0.08155433088541031,
0.0019124216632917523,
0.029221396893262863,
0.11990660429000854,
0.17286483943462372,
-0.05838547274470329,
0.0494680181145668,
-0.12196874618530273,
0.002273800317198038,
-0.03104289248585701,
-0.03879985958337784,
-0.12439896911382675,
-0.10816622525453568,
0.023254498839378357,
0.005873893387615681,
0.19243818521499634,
0.08622796088457108,
0.09015563130378723,
0.02026992477476597,
0.08280457556247711,
0.15640299022197723,
0.08169982582330704,
0.13479216396808624,
0.0753636434674263,
0.003552020760253072,
-0.04583922401070595,
0.06687968969345093,
0.04342857003211975,
0.0072935158386826515,
0.0038177764508873224,
0.09528568387031555,
0.00667149294167757,
0.10526926070451736,
0.042451679706573486,
0.0500829704105854,
-0.032434239983558655,
-0.06250517815351486,
-0.0641176775097847,
-0.01888967491686344,
-0.011569607071578503,
0.12127377092838287,
0.14277249574661255,
-0.08079589903354645,
0.04679444059729576,
0.015447227284312248,
-0.06474620848894119,
-0.1609332412481308,
-0.1257009506225586,
-0.07168443500995636,
-0.18344774842262268,
-0.027470940724015236,
-0.09683065861463547,
-0.02620367519557476,
0.16843217611312866,
0.04557392746210098,
-0.009155435487627983,
0.02763553522527218,
0.0004341208259575069,
-0.02509470470249653,
-0.020050279796123505,
-0.08659908920526505,
0.07982605695724487,
0.04354038089513779,
-0.010250295512378216,
0.07036767154932022,
0.038146670907735825,
0.051697567105293274,
-0.016725361347198486,
0.016275659203529358,
0.014421421103179455,
-0.0931367501616478,
-0.08636751025915146,
-0.014829674735665321,
0.006926656700670719,
-0.012881503440439701,
0.022557280957698822,
0.07137629389762878,
-0.03945473954081535,
0.001135995495133102,
0.21051138639450073,
-0.09168606251478195,
-0.13335897028446198,
-0.22504109144210815,
0.3292695879936218,
0.06128848344087601,
0.08006542921066284,
0.01782870665192604,
-0.09772481769323349,
-0.08380256593227386,
0.23835910856723785,
0.15601664781570435,
-0.07657447457313538,
-0.014831600710749626,
0.07228817790746689,
-0.016094770282506943,
0.04914749786257744,
0.11798988282680511,
-0.0018936273409053683,
0.22898456454277039,
-0.08471135795116425,
0.04600175470113754,
-0.03982498124241829,
-0.06657011061906815,
-0.0016606205608695745,
0.04384338855743408,
-0.028883112594485283,
-0.0646972507238388,
-0.0373566634953022,
0.14134474098682404,
-0.09122554212808609,
0.049046847969293594,
-0.0002345513494219631,
-0.022823620587587357,
-0.07716405391693115,
-0.006424167193472385,
0.06062676012516022,
0.00020406620751600713,
0.08504018932580948,
-0.038146086037158966,
0.03794025629758835,
0.11565134674310684,
-0.0017569564515724778,
-0.06979916244745255,
-0.058996688574552536,
0.06873651593923569,
-0.06620217114686966,
0.02624395117163658,
-0.020865874364972115,
0.19483453035354614,
0.09904284030199051,
0.06483505666255951,
-0.03644286096096039,
0.0626656711101532,
-0.013008199632167816,
0.011695616878569126,
0.12002301216125488,
0.03363015130162239,
0.0020754639990627766,
0.06408432871103287,
0.044256992638111115,
-0.1768614798784256,
0.07748190313577652,
-0.14381177723407745,
-0.06964220106601715,
-0.10966911166906357,
0.02179321087896824,
-0.012797866016626358,
0.1110919713973999,
0.09294867515563965,
-0.04715340584516525,
0.012934789061546326,
0.018140526488423347,
0.015916945412755013,
-0.041642192751169205,
-0.078044094145298,
-0.03142290934920311,
-0.16770023107528687,
0.007813815027475357,
-0.03266200050711632,
0.002701889956369996,
-0.22878091037273407,
-0.011267242021858692,
-0.008542446419596672,
-0.033176716417074203,
0.008387149311602116,
0.08422781527042389,
0.08410125225782394,
0.03710673004388809,
-0.07843427360057831,
-0.06219158694148064,
0.0013762605376541615,
0.09299024194478989,
-0.12327927350997925,
-0.15334023535251617
] |
null | null |
transformers
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using [SentenceTransformers](https://sbert.net) [Cross-Encoder](https://www.sbert.net/examples/applications/cross-encoder/README.html) class.
## Training Data
This model was trained on the [STS benchmark dataset](http://ixa2.si.ehu.eus/stswiki/index.php/STSbenchmark). The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
```
from sentence_transformers import CrossEncoder
model = CrossEncoder('model_name')
scores = model.predict([('Sentence 1', 'Sentence 2'), ('Sentence 3', 'Sentence 4')])
```
The model will predict scores for the pairs `('Sentence 1', 'Sentence 2')` and `('Sentence 3', 'Sentence 4')`.
You can use this model also without sentence_transformers and by just using Transformers ``AutoModel`` class
|
{"license": "apache-2.0"}
|
text-classification
|
cross-encoder/stsb-roberta-large
|
[
"transformers",
"pytorch",
"jax",
"roberta",
"text-classification",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
|
# Cross-Encoder for Quora Duplicate Questions Detection
This model was trained using SentenceTransformers Cross-Encoder class.
## Training Data
This model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.
## Usage and Performance
Pre-trained models can be used like this:
The model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.
You can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class
|
[
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
"TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.",
"## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.",
"## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
52,
34,
38,
85
] |
[
"passage: TAGS\n#transformers #pytorch #jax #roberta #text-classification #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# Cross-Encoder for Quora Duplicate Questions Detection\nThis model was trained using SentenceTransformers Cross-Encoder class.## Training Data\nThis model was trained on the STS benchmark dataset. The model will predict a score between 0 and 1 how for the semantic similarity of two sentences.## Usage and Performance\n\nPre-trained models can be used like this:\n\n\nThe model will predict scores for the pairs '('Sentence 1', 'Sentence 2')' and '('Sentence 3', 'Sentence 4')'.\n\nYou can use this model also without sentence_transformers and by just using Transformers ''AutoModel'' class"
] |
[
-0.08450348675251007,
0.06658493727445602,
-0.0020891379099339247,
0.06090733781456947,
0.1219165176153183,
-0.010885688476264477,
0.18726801872253418,
0.06976161152124405,
0.004681415390223265,
-0.032464902848005295,
0.10880614817142487,
0.10593046247959137,
0.009075884707272053,
0.1311509907245636,
-0.0318918414413929,
-0.2397564798593521,
0.05535270273685455,
-0.021333403885364532,
-0.023506149649620056,
0.1251000612974167,
0.10538756102323532,
-0.08173905313014984,
0.06480519473552704,
0.036619383841753006,
-0.1481119692325592,
0.0361047126352787,
0.004630628507584333,
-0.06398891657590866,
0.1055711954832077,
0.029057493433356285,
0.12339451909065247,
0.09259660542011261,
0.07925070077180862,
-0.12724295258522034,
0.012720810249447823,
0.016518067568540573,
0.009311544708907604,
0.03629666194319725,
0.036708857864141464,
-0.10701856762170792,
-0.09100929647684097,
0.08599001914262772,
0.08524163067340851,
0.0009422770235687494,
-0.11783528327941895,
-0.1288551539182663,
-0.02296801283955574,
-0.0451493114233017,
0.1464175432920456,
0.0878152996301651,
-0.045602668076753616,
0.1485908329486847,
-0.1755804717540741,
0.04145662114024162,
0.054525889456272125,
-0.3068389296531677,
0.010614708065986633,
0.0938933864235878,
0.023070018738508224,
0.04588997736573219,
-0.04204950109124184,
-0.006081583444029093,
0.028909359127283096,
0.04380388930439949,
0.06281910091638565,
-0.04001569375395775,
-0.010886033065617085,
0.037939924746751785,
-0.1775822639465332,
-0.03491581231355667,
0.21934248507022858,
-0.03001006878912449,
-0.03818020224571228,
-0.10346876084804535,
-0.030659932643175125,
0.05167389288544655,
-0.022362971678376198,
-0.08230966329574585,
0.004080281592905521,
0.05453440919518471,
-0.09994271397590637,
0.043615926057100296,
-0.12259428203105927,
-0.1212959885597229,
-0.11897437274456024,
0.1481919288635254,
0.0519876554608345,
0.05614864081144333,
-0.11817987263202667,
0.11007988452911377,
-0.018013888970017433,
-0.08047500997781754,
-0.03508051112294197,
-0.07759242504835129,
-0.13628479838371277,
-0.017040323466062546,
-0.1888146996498108,
-0.1461862325668335,
0.019468307495117188,
0.23470254242420197,
0.05425926297903061,
-0.003253627335652709,
0.05223825201392174,
0.052410710602998734,
0.05260057747364044,
0.16987985372543335,
-0.09921816736459732,
0.019514871761202812,
-0.040489714592695236,
-0.006623496767133474,
-0.014924317598342896,
-0.020644033327698708,
-0.15668311715126038,
-0.02293773740530014,
0.06836400926113129,
0.03340352699160576,
-0.03544469177722931,
0.060046419501304626,
-0.03591883182525635,
-0.04250994324684143,
0.04778998717665672,
-0.07110997289419174,
0.007199421525001526,
0.001498668105341494,
-0.08984625339508057,
0.04597581550478935,
0.004727287217974663,
0.05564960837364197,
-0.04794371500611305,
-0.03951052203774452,
-0.05796322971582413,
-0.0041662254370749,
-0.07994621992111206,
-0.13172027468681335,
-0.017280934378504753,
0.0645439475774765,
0.004875025246292353,
-0.12934610247612,
-0.15153756737709045,
-0.06250552833080292,
0.06014043092727661,
-0.025576326996088028,
-0.0060349274426698685,
-0.10992851108312607,
-0.024137191474437714,
0.014359449967741966,
-0.03408343344926834,
-0.012038946151733398,
-0.006700094789266586,
0.027012227103114128,
-0.010106818750500679,
0.10634177178144455,
0.011723324656486511,
0.056721992790699005,
-0.17293746769428253,
-0.014487387612462044,
-0.06987504661083221,
0.13998807966709137,
0.00621027173474431,
0.036297235637903214,
-0.07619620859622955,
-0.03155680000782013,
-0.03006790578365326,
0.06660110503435135,
0.037052832543849945,
0.1293592005968094,
-0.16096548736095428,
-0.06480851024389267,
0.2344948649406433,
-0.12800095975399017,
-0.11015582084655762,
0.12573368847370148,
-0.049440640956163406,
0.03731274977326393,
0.10660473257303238,
0.12351691722869873,
0.0702773854136467,
-0.034878578037023544,
0.08157497644424438,
0.06925155967473984,
-0.0970594733953476,
-0.01633760891854763,
0.10197000205516815,
0.03484668955206871,
-0.08979984372854233,
0.031350795179605484,
-0.06290501356124878,
0.05259009823203087,
-0.04869356006383896,
-0.06925348937511444,
-0.013377794064581394,
-0.054921574890613556,
0.07530007511377335,
0.020727040246129036,
0.08574032783508301,
-0.02007264457643032,
-0.10619869083166122,
0.11279048770666122,
0.08791369199752808,
-0.0893058329820633,
0.029166527092456818,
-0.10053735971450806,
0.013915755786001682,
-0.07938611507415771,
-0.007064095698297024,
-0.20703130960464478,
0.006485046353191137,
0.0018438141560181975,
0.014914322644472122,
0.034115370362997055,
0.10435596108436584,
0.030128762125968933,
0.022408952936530113,
0.029981357976794243,
-0.018031882122159004,
0.08979107439517975,
0.009407839737832546,
-0.08929740637540817,
-0.10686875879764557,
-0.02606169506907463,
-0.03660851716995239,
0.15328125655651093,
-0.013654633425176144,
0.02513095922768116,
-0.09524213522672653,
0.0469084158539772,
-0.029759293422102928,
0.05308831110596657,
-0.004855465143918991,
0.04144696891307831,
-0.04124278575181961,
0.0014224101323634386,
0.04660613834857941,
0.017619729042053223,
-0.09673509001731873,
0.1709372103214264,
-0.18291769921779633,
-0.02089911699295044,
0.12491315603256226,
-0.11515885591506958,
-0.06447350978851318,
0.02345173992216587,
-0.014730798080563545,
0.02202417515218258,
-0.05534713342785835,
0.014598351903259754,
0.16470010578632355,
0.014831818640232086,
0.12296251952648163,
-0.083440862596035,
-0.026016926392912865,
-0.018848687410354614,
-0.09721343219280243,
0.018556935712695122,
0.04981476813554764,
0.039000771939754486,
-0.1162211075425148,
0.10053393989801407,
0.10247835516929626,
-0.10287507623434067,
0.015802966430783272,
0.004366625100374222,
-0.04628602787852287,
-0.03741034120321274,
-0.04338941350579262,
-0.03472225368022919,
-0.027288775891065598,
-0.029719965532422066,
-0.042659372091293335,
0.03510940074920654,
0.029411355033516884,
0.013905813917517662,
-0.15196870267391205,
0.005906535778194666,
0.013755157589912415,
-0.011594702489674091,
-0.01026098895817995,
0.04376592859625816,
-0.008289550431072712,
0.061599910259246826,
0.028126584365963936,
-0.14367526769638062,
0.058363452553749084,
-0.0351489894092083,
-0.14141881465911865,
0.22615239024162292,
-0.07566404342651367,
-0.27953410148620605,
-0.12290982156991959,
0.021696176379919052,
0.006087016314268112,
0.062644362449646,
0.0708455741405487,
-0.045471686869859695,
-0.03838431462645531,
-0.0820717066526413,
-0.03139931708574295,
-0.024082103744149208,
0.0024445243179798126,
-0.0878162756562233,
0.00009271603630622849,
-0.003847202518954873,
-0.09421621263027191,
-0.0031321351416409016,
-0.05890399217605591,
-0.06755509972572327,
0.04214369133114815,
-0.09165877103805542,
-0.007889453321695328,
0.2206796407699585,
0.01076422818005085,
0.03367805480957031,
-0.05330446735024452,
0.17513544857501984,
-0.015779828652739525,
-0.03460876643657684,
0.1611550748348236,
-0.012880779802799225,
0.04445143789052963,
0.17388220131397247,
0.019242996349930763,
-0.0683070495724678,
0.07065898925065994,
-0.03215143457055092,
-0.04168960079550743,
-0.1452464610338211,
-0.09278552234172821,
-0.038062673062086105,
-0.0233317743986845,
0.07220902293920517,
-0.01197992917150259,
0.08731245249509811,
0.05724746733903885,
-0.003093191422522068,
-0.01159081980586052,
0.03088546171784401,
0.0780472680926323,
0.1442079395055771,
0.003124864771962166,
0.17355042695999146,
-0.016894925385713577,
-0.11764701455831528,
0.027372056618332863,
0.02583908475935459,
0.10140083730220795,
0.05113178864121437,
0.009328491054475307,
0.10008900612592697,
0.027490992099046707,
0.07921173423528671,
-0.004589700140058994,
0.004766227677464485,
0.0007489954587072134,
-0.03577183932065964,
-0.029451249167323112,
0.016168830916285515,
0.09111353009939194,
0.01197222899645567,
-0.02815566398203373,
-0.12441352754831314,
-0.03104766085743904,
0.04662104323506355,
0.11217736452817917,
0.16801334917545319,
-0.30731672048568726,
-0.05213860049843788,
0.04639245942234993,
-0.06299179047346115,
0.023163730278611183,
0.09142906963825226,
-0.004638497717678547,
-0.13518525660037994,
0.00644878763705492,
-0.004350210539996624,
0.08985988050699234,
0.0018690075958147645,
0.02098994515836239,
-0.06347128748893738,
0.02823813259601593,
0.014463197439908981,
0.12384771555662155,
-0.34352806210517883,
0.10411574691534042,
-0.029895853251218796,
0.011633627116680145,
-0.04604942351579666,
-0.006390930153429508,
0.04763137176632881,
0.14670772850513458,
0.08665851503610611,
-0.02811373583972454,
-0.09105808287858963,
-0.06061112508177757,
-0.04584410414099693,
0.022072266787290573,
0.025318568572402,
-0.05517273768782616,
0.05771936476230621,
-0.04354587942361832,
-0.005889877211302519,
0.02616376057267189,
0.05868706852197647,
-0.025132019072771072,
-0.07593091577291489,
-0.007170701399445534,
0.013278568163514137,
0.05211564898490906,
-0.015402347780764103,
-0.06917586177587509,
-0.13652382791042328,
0.042136695235967636,
-0.07789728045463562,
-0.06028033047914505,
-0.103256456553936,
0.09486180543899536,
0.07285741716623306,
-0.06623293459415436,
0.0039231302216649055,
0.016175266355276108,
0.133241206407547,
0.004705662373453379,
-0.1416292041540146,
0.08322334289550781,
-0.07802871614694595,
-0.08273158967494965,
-0.06269396841526031,
0.03027677722275257,
0.04064171016216278,
0.04588952288031578,
0.0636964812874794,
0.06164630129933357,
-0.11224881559610367,
-0.07974357903003693,
-0.02778312936425209,
0.016498364508152008,
0.05221463739871979,
0.024097010493278503,
-0.005767281167209148,
-0.055050041526556015,
-0.03973664715886116,
0.06736788898706436,
0.17233264446258545,
0.0396297462284565,
-0.09726333618164062,
0.12583638727664948,
0.17738917469978333,
-0.04895620420575142,
-0.283265620470047,
-0.09362755715847015,
0.04195278882980347,
0.07143981009721756,
0.014062013477087021,
-0.09430427849292755,
0.1031544879078865,
0.012528144754469395,
-0.048478029668331146,
-0.2168145477771759,
-0.3028346598148346,
-0.10955806821584702,
0.18833613395690918,
0.03733060136437416,
0.1766844391822815,
-0.07329234480857849,
0.00888361781835556,
-0.06110344082117081,
0.008218829520046711,
0.149373859167099,
-0.06190621480345726,
0.0771213248372078,
0.031203610822558403,
0.02545219659805298,
-0.0002090671769110486,
-0.01993192359805107,
0.07104034721851349,
0.08380923420190811,
0.060775671154260635,
-0.02233022265136242,
-0.005947250407189131,
0.030198819935321808,
-0.038693733513355255,
0.13508112728595734,
-0.038966890424489975,
0.09695424884557724,
-0.060185179114341736,
-0.08268783241510391,
-0.033626738935709,
0.0095839723944664,
0.021442634984850883,
-0.047682344913482666,
-0.014281190931797028,
0.0293685682117939,
0.04048173502087593,
-0.010960451327264309,
0.06327590346336365,
-0.06340527534484863,
0.045411452651023865,
0.11049935221672058,
0.13265644013881683,
-0.01526104100048542,
-0.03578300401568413,
0.0051394738256931305,
-0.047414444386959076,
0.11430126428604126,
-0.164499893784523,
0.0911836102604866,
0.07006113976240158,
-0.022437749430537224,
0.1343110352754593,
0.0996335968375206,
-0.04331466183066368,
0.018353044986724854,
0.052883587777614594,
-0.0847473293542862,
-0.06420569121837616,
-0.015110892243683338,
-0.007178271654993296,
-0.005094901192933321,
0.0529211089015007,
0.11877772212028503,
-0.05779349431395531,
-0.04995092377066612,
-0.008740345947444439,
0.035886529833078384,
-0.06572326272726059,
0.11153262853622437,
0.06726455688476562,
0.032411862164735794,
-0.09644010663032532,
0.06804357469081879,
0.07756434381008148,
-0.02221454121172428,
0.06189196556806564,
0.018371189013123512,
-0.13287536799907684,
-0.09099080413579941,
-0.03262708708643913,
0.17415082454681396,
-0.1126992255449295,
-0.09605299681425095,
-0.1218004897236824,
-0.10991249978542328,
0.04737558588385582,
0.11134099960327148,
0.13007406890392303,
0.07032902538776398,
-0.07839404791593552,
-0.06061587482690811,
-0.07010961323976517,
0.0824027732014656,
0.05524514988064766,
-0.002843055408447981,
-0.07459040731191635,
0.11347991228103638,
0.031216993927955627,
0.0675557404756546,
-0.08114106953144073,
-0.10859313607215881,
-0.128158301115036,
0.06127750873565674,
-0.06947355717420578,
-0.005555060692131519,
-0.07801594585180283,
-0.010933276265859604,
0.03555882349610329,
-0.05734606087207794,
-0.0364067368209362,
0.013651168905198574,
-0.08039981126785278,
0.025344787165522575,
-0.037574876099824905,
0.06731393933296204,
-0.10970719158649445,
-0.029593832790851593,
0.023546205833554268,
-0.03511421009898186,
0.04349852725863457,
0.07940305769443512,
-0.02931799553334713,
0.04794476926326752,
-0.0984886959195137,
-0.04244762659072876,
-0.0008963823784142733,
0.05560130998492241,
0.04657924920320511,
-0.11562971025705338,
0.011889439076185226,
0.08023174852132797,
0.02496957965195179,
0.01808847486972809,
0.04983857646584511,
-0.06669441610574722,
0.03893890976905823,
-0.005539896432310343,
-0.04568590968847275,
-0.08155433088541031,
0.0019124216632917523,
0.029221396893262863,
0.11990660429000854,
0.17286483943462372,
-0.05838547274470329,
0.0494680181145668,
-0.12196874618530273,
0.002273800317198038,
-0.03104289248585701,
-0.03879985958337784,
-0.12439896911382675,
-0.10816622525453568,
0.023254498839378357,
0.005873893387615681,
0.19243818521499634,
0.08622796088457108,
0.09015563130378723,
0.02026992477476597,
0.08280457556247711,
0.15640299022197723,
0.08169982582330704,
0.13479216396808624,
0.0753636434674263,
0.003552020760253072,
-0.04583922401070595,
0.06687968969345093,
0.04342857003211975,
0.0072935158386826515,
0.0038177764508873224,
0.09528568387031555,
0.00667149294167757,
0.10526926070451736,
0.042451679706573486,
0.0500829704105854,
-0.032434239983558655,
-0.06250517815351486,
-0.0641176775097847,
-0.01888967491686344,
-0.011569607071578503,
0.12127377092838287,
0.14277249574661255,
-0.08079589903354645,
0.04679444059729576,
0.015447227284312248,
-0.06474620848894119,
-0.1609332412481308,
-0.1257009506225586,
-0.07168443500995636,
-0.18344774842262268,
-0.027470940724015236,
-0.09683065861463547,
-0.02620367519557476,
0.16843217611312866,
0.04557392746210098,
-0.009155435487627983,
0.02763553522527218,
0.0004341208259575069,
-0.02509470470249653,
-0.020050279796123505,
-0.08659908920526505,
0.07982605695724487,
0.04354038089513779,
-0.010250295512378216,
0.07036767154932022,
0.038146670907735825,
0.051697567105293274,
-0.016725361347198486,
0.016275659203529358,
0.014421421103179455,
-0.0931367501616478,
-0.08636751025915146,
-0.014829674735665321,
0.006926656700670719,
-0.012881503440439701,
0.022557280957698822,
0.07137629389762878,
-0.03945473954081535,
0.001135995495133102,
0.21051138639450073,
-0.09168606251478195,
-0.13335897028446198,
-0.22504109144210815,
0.3292695879936218,
0.06128848344087601,
0.08006542921066284,
0.01782870665192604,
-0.09772481769323349,
-0.08380256593227386,
0.23835910856723785,
0.15601664781570435,
-0.07657447457313538,
-0.014831600710749626,
0.07228817790746689,
-0.016094770282506943,
0.04914749786257744,
0.11798988282680511,
-0.0018936273409053683,
0.22898456454277039,
-0.08471135795116425,
0.04600175470113754,
-0.03982498124241829,
-0.06657011061906815,
-0.0016606205608695745,
0.04384338855743408,
-0.028883112594485283,
-0.0646972507238388,
-0.0373566634953022,
0.14134474098682404,
-0.09122554212808609,
0.049046847969293594,
-0.0002345513494219631,
-0.022823620587587357,
-0.07716405391693115,
-0.006424167193472385,
0.06062676012516022,
0.00020406620751600713,
0.08504018932580948,
-0.038146086037158966,
0.03794025629758835,
0.11565134674310684,
-0.0017569564515724778,
-0.06979916244745255,
-0.058996688574552536,
0.06873651593923569,
-0.06620217114686966,
0.02624395117163658,
-0.020865874364972115,
0.19483453035354614,
0.09904284030199051,
0.06483505666255951,
-0.03644286096096039,
0.0626656711101532,
-0.013008199632167816,
0.011695616878569126,
0.12002301216125488,
0.03363015130162239,
0.0020754639990627766,
0.06408432871103287,
0.044256992638111115,
-0.1768614798784256,
0.07748190313577652,
-0.14381177723407745,
-0.06964220106601715,
-0.10966911166906357,
0.02179321087896824,
-0.012797866016626358,
0.1110919713973999,
0.09294867515563965,
-0.04715340584516525,
0.012934789061546326,
0.018140526488423347,
0.015916945412755013,
-0.041642192751169205,
-0.078044094145298,
-0.03142290934920311,
-0.16770023107528687,
0.007813815027475357,
-0.03266200050711632,
0.002701889956369996,
-0.22878091037273407,
-0.011267242021858692,
-0.008542446419596672,
-0.033176716417074203,
0.008387149311602116,
0.08422781527042389,
0.08410125225782394,
0.03710673004388809,
-0.07843427360057831,
-0.06219158694148064,
0.0013762605376541615,
0.09299024194478989,
-0.12327927350997925,
-0.15334023535251617
] |
null | null |
transformers
|
### Kw2Poem
|
{"language": "vi", "tags": ["gpt"], "widget": [{"text": "<s> n\u00fai nh\u00e0 xe [SEP] "}]}
|
text-generation
|
crylake/kw2poem-generation
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"gpt",
"vi",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"vi"
] |
TAGS
#transformers #pytorch #gpt2 #text-generation #gpt #vi #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
### Kw2Poem
|
[
"### Kw2Poem"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #gpt #vi #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"### Kw2Poem"
] |
[
52,
7
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #gpt #vi #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Kw2Poem"
] |
[
0.00957817304879427,
0.06301134824752808,
-0.007064964156597853,
0.0036542851012200117,
0.12309610843658447,
0.031439412385225296,
0.07243029028177261,
0.1607857048511505,
0.00964074395596981,
-0.01801552250981331,
0.18630695343017578,
0.17791466414928436,
0.03313320875167847,
0.0681326687335968,
-0.027575921267271042,
-0.31554916501045227,
0.05325678735971451,
0.0723961889743805,
-0.020475471392273903,
0.12803497910499573,
0.04001649096608162,
-0.06150275841355324,
0.08209304511547089,
0.003992442507296801,
-0.16980701684951782,
-0.008797045797109604,
0.025250831618905067,
-0.11566734313964844,
0.09016840904951096,
0.05655558779835701,
0.06918779015541077,
0.04472552239894867,
-0.03827078640460968,
-0.1263580322265625,
0.02615487389266491,
0.006352890748530626,
-0.04630299657583237,
0.043258532881736755,
0.09118277579545975,
-0.06698725372552872,
0.09973156452178955,
0.022308319807052612,
-0.07163960486650467,
0.02657143771648407,
-0.13541530072689056,
-0.071018747985363,
-0.035225287079811096,
0.07012996822595596,
0.06396427005529404,
0.06126224249601364,
-0.010328901931643486,
0.1294802576303482,
-0.072997085750103,
0.08923482149839401,
0.18938976526260376,
-0.3316987156867981,
-0.0015500655863434076,
0.07680265605449677,
0.03494839370250702,
-0.01640024594962597,
-0.031431008130311966,
0.08453582972288132,
0.01143955159932375,
-0.0011958361137658358,
-0.019267888739705086,
-0.09979125112295151,
-0.11628061532974243,
0.042645879089832306,
-0.10252959281206131,
-0.055888764560222626,
0.2122832089662552,
-0.05855756998062134,
0.0569487027823925,
0.02094966731965542,
-0.09387509524822235,
-0.014127666130661964,
0.0008141568978317082,
0.023698817938566208,
-0.07356588542461395,
0.04334239289164543,
-0.004188459366559982,
-0.06054595112800598,
-0.12845057249069214,
-0.04309607669711113,
-0.1624227911233902,
0.23481477797031403,
0.03394542261958122,
0.0395079143345356,
-0.15551385283470154,
0.13015125691890717,
-0.06904776394367218,
-0.09651973098516464,
-0.0034290305338799953,
-0.07348773628473282,
0.09329099208116531,
0.026412852108478546,
-0.018657071515917778,
-0.05315980315208435,
0.093925341963768,
0.09602594375610352,
0.031985338777303696,
0.024359751492738724,
-0.002628166461363435,
0.09244594722986221,
0.06021767109632492,
0.09955921769142151,
-0.002758112270385027,
-0.01708148792386055,
0.05911704897880554,
-0.12917882204055786,
-0.0078049651347100735,
-0.06700736284255981,
-0.14689482748508453,
-0.07990694791078568,
0.057311221957206726,
0.07181021571159363,
0.044495828449726105,
0.10255920886993408,
-0.03267213702201843,
-0.029797419905662537,
0.0361289456486702,
-0.0445629246532917,
-0.008850942365825176,
-0.0107669522985816,
0.03428369387984276,
0.1568555235862732,
0.00325220450758934,
0.026154393330216408,
-0.12736763060092926,
0.08676701784133911,
-0.07274826616048813,
-0.0031747606117278337,
-0.02124539203941822,
-0.03635423257946968,
0.020667510107159615,
-0.10058806091547012,
0.030601467937231064,
-0.1443287432193756,
-0.1689336895942688,
0.013190992176532745,
-0.02336847223341465,
-0.013140843249857426,
-0.0766136571764946,
-0.030397942289710045,
-0.04604308679699898,
0.05602826178073883,
-0.05154234915971756,
0.019728047773241997,
-0.05774667114019394,
0.11128174513578415,
-0.052658792585134506,
0.07410987466573715,
-0.13703618943691254,
0.053574539721012115,
-0.10183457285165787,
0.006429322995245457,
-0.13172046840190887,
0.04558270052075386,
-0.028972763568162918,
0.04250296205282211,
-0.019867192953824997,
-0.039371710270643234,
-0.06422477960586548,
0.03419045731425285,
-0.032960034906864166,
0.18798483908176422,
-0.07311984151601791,
-0.13018974661827087,
0.2925363779067993,
-0.06353839486837387,
-0.1162010058760643,
0.08763570338487625,
0.031029900535941124,
0.03783457726240158,
0.07800211757421494,
0.2249654084444046,
0.06717246025800705,
-0.00396937970072031,
0.03586290031671524,
0.1031804159283638,
-0.08144260197877884,
-0.02689230628311634,
0.03230758756399155,
-0.024571599438786507,
-0.13738372921943665,
0.03794455528259277,
-0.01866414211690426,
0.052874647080898285,
-0.04736626520752907,
-0.013525504618883133,
-0.0316149927675724,
0.0037264577113091946,
0.08108885586261749,
-0.03282852843403816,
0.10934285819530487,
-0.04816563427448273,
-0.027070369571447372,
0.0007734297541901469,
0.025037771090865135,
-0.020882830023765564,
0.047117553651332855,
-0.035543374717235565,
0.12087835371494293,
-0.020327946171164513,
0.06259502470493317,
-0.1060953140258789,
-0.02786347270011902,
0.004772518761456013,
0.08213009685277939,
0.026698121801018715,
0.0887259989976883,
0.04245677590370178,
-0.0168427973985672,
-0.012565472163259983,
-0.030297597870230675,
0.08050557971000671,
-0.01142654474824667,
-0.08747817575931549,
-0.07889477163553238,
0.0820138081908226,
-0.029122885316610336,
-0.0694442018866539,
-0.027604790404438972,
0.011847486719489098,
0.12812773883342743,
0.09897224605083466,
-0.02164166420698166,
0.03927278891205788,
-0.005955390632152557,
0.0067389062605798244,
-0.06489384919404984,
0.008974163793027401,
0.08441240340471268,
-0.010470270179212093,
-0.0728292241692543,
0.2262585312128067,
-0.13634425401687622,
0.15140721201896667,
0.19395232200622559,
-0.2116062343120575,
0.03191465511918068,
-0.02867828868329525,
-0.024682335555553436,
-0.002660174388438463,
0.020991845056414604,
-0.029479147866368294,
0.1510322242975235,
-0.0025584178511053324,
0.13806121051311493,
-0.05763198435306549,
-0.0770176500082016,
0.00790970865637064,
-0.05813254415988922,
0.017832832410931587,
0.05987262353301048,
0.1267615258693695,
-0.1311156302690506,
0.1655711531639099,
0.25164952874183655,
0.026769690215587616,
0.20776516199111938,
0.019149042665958405,
-0.016153262928128242,
0.04002733156085014,
-0.021725956350564957,
-0.038566626608371735,
-0.021408110857009888,
-0.20528195798397064,
-0.013087124563753605,
0.07204267382621765,
0.03846413642168045,
0.07889289408922195,
-0.1345205157995224,
-0.0437425896525383,
-0.04111592099070549,
-0.01434281188994646,
0.0031131766736507416,
0.09778230637311935,
0.028985416516661644,
0.11504949629306793,
0.01989189349114895,
0.05871374532580376,
0.0693468376994133,
0.029436497017741203,
-0.07359284907579422,
0.2007758468389511,
-0.1341300904750824,
-0.326452374458313,
-0.12633070349693298,
-0.08797509968280792,
-0.03790261223912239,
0.034858908504247665,
0.10500060766935349,
-0.11347244679927826,
-0.020820388570427895,
0.024372363463044167,
0.14073021709918976,
-0.12031327188014984,
0.04537583515048027,
-0.05039072781801224,
0.021589763462543488,
-0.11016356945037842,
-0.0658922791481018,
-0.05893188714981079,
0.013843301683664322,
-0.09759803861379623,
0.17724177241325378,
-0.06949371099472046,
0.059028707444667816,
0.1800614595413208,
0.031153270974755287,
0.03705417737364769,
-0.029824350029230118,
0.17289897799491882,
-0.09951750189065933,
0.02500886283814907,
0.15677721798419952,
-0.06731083244085312,
0.09062217175960541,
0.046989936381578445,
-0.012154776602983475,
-0.07390722632408142,
0.016875628381967545,
-0.011683935299515724,
-0.08091358840465546,
-0.20660856366157532,
-0.08553552627563477,
-0.0886504277586937,
0.15322765707969666,
0.05269729718565941,
0.04272379353642464,
0.0866679698228836,
0.11853060126304626,
-0.03408195450901985,
0.07567588984966278,
-0.009533594362437725,
0.09122149646282196,
0.1803288757801056,
-0.0285966657102108,
0.10404042154550552,
-0.04416600614786148,
-0.11093810200691223,
0.10502474009990692,
0.08413784950971603,
0.09985124319791794,
0.030600035563111305,
-0.003252935828641057,
0.0469202995300293,
0.09457972645759583,
0.14892789721488953,
0.08599001169204712,
0.019359536468982697,
-0.01143420860171318,
-0.039638202637434006,
-0.030968688428401947,
-0.035622384399175644,
0.05220633000135422,
0.05053401365876198,
-0.17141735553741455,
-0.03554258495569229,
-0.07563856244087219,
0.08663869649171829,
0.0934019461274147,
0.10105573385953903,
-0.1376226544380188,
-0.008649363182485104,
0.100326769053936,
-0.028425011783838272,
-0.133919358253479,
0.08279795199632645,
0.01092482265084982,
-0.1892797350883484,
0.08940976858139038,
-0.04272071644663811,
0.10588239878416061,
-0.1050056740641594,
0.06997676193714142,
-0.07266444712877274,
-0.08001580089330673,
0.015113139525055885,
0.12805187702178955,
-0.2630266845226288,
0.1980765014886856,
0.00470179645344615,
-0.05748160555958748,
-0.11423757672309875,
-0.02530396357178688,
0.02106613852083683,
0.07345574349164963,
0.11278049647808075,
0.03998386114835739,
0.0015202027279883623,
-0.07409339398145676,
-0.04619338735938072,
0.05795834958553314,
0.1431449055671692,
-0.1010911613702774,
0.02264595218002796,
-0.01898161880671978,
-0.0020718886516988277,
-0.04349042475223541,
-0.05938585847616196,
-0.026471013203263283,
-0.1162632405757904,
0.09133803844451904,
0.004934979602694511,
0.10479361563920975,
-0.013762879185378551,
-0.031141597777605057,
-0.1040981113910675,
0.23426604270935059,
-0.12257760763168335,
-0.10206079483032227,
-0.09111861139535904,
-0.012589845806360245,
0.03955908864736557,
-0.09625139832496643,
0.0342625267803669,
-0.07532605528831482,
0.014382464811205864,
-0.055122461169958115,
-0.17879384756088257,
0.11824013292789459,
-0.08392655104398727,
-0.13032107055187225,
-0.020929044112563133,
0.17684286832809448,
-0.03928419202566147,
0.023285819217562675,
0.003155027749016881,
0.023544492200016975,
-0.07451646775007248,
-0.13398806750774384,
0.05609554797410965,
0.02090577594935894,
0.0189196839928627,
0.034923769533634186,
-0.02513517625629902,
0.06144357845187187,
-0.04819108545780182,
-0.05300992727279663,
0.3036767840385437,
0.2120986431837082,
-0.03889444097876549,
0.1935625970363617,
0.06893247365951538,
-0.08998647332191467,
-0.367197185754776,
-0.10181158781051636,
-0.07947426289319992,
-0.0382821150124073,
-0.05219694599509239,
-0.209986612200737,
0.07809457182884216,
0.03275468572974205,
-0.013884881511330605,
0.12455035746097565,
-0.27253809571266174,
-0.09908903390169144,
0.10868138074874878,
0.025725973770022392,
0.35784900188446045,
-0.10534874349832535,
-0.10031116753816605,
-0.03133375570178032,
-0.14740042388439178,
0.16990341246128082,
-0.04684435948729515,
0.1021876186132431,
-0.0563608855009079,
0.12966550886631012,
0.03599758818745613,
-0.031546905636787415,
0.09999095648527145,
0.01583133079111576,
-0.02859887108206749,
-0.1090945228934288,
-0.017750374972820282,
0.055275458842515945,
0.004516797140240669,
-0.0057978457771241665,
-0.047339312732219696,
0.03101438842713833,
-0.15586735308170319,
-0.02265102230012417,
-0.10903351753950119,
0.08096453547477722,
0.015084374696016312,
-0.0857151448726654,
-0.014661803841590881,
-0.06186206638813019,
-0.028775228187441826,
0.005849520210176706,
0.15451213717460632,
-0.004929624032229185,
0.0800783708691597,
0.04938990995287895,
0.09341436624526978,
-0.05828000232577324,
0.027915148064494133,
-0.07094196230173111,
-0.060766562819480896,
0.07760199159383774,
-0.11375044286251068,
0.0411730594933033,
0.1161332055926323,
-0.010778318159282207,
0.0207026619464159,
0.10265020281076431,
-0.03212396800518036,
-0.00035324503551237285,
0.11143036186695099,
-0.28819185495376587,
-0.01519784890115261,
-0.08012323826551437,
0.0003487319918349385,
0.09877925366163254,
0.06377357989549637,
0.19824156165122986,
0.00480060838162899,
-0.0733577311038971,
0.00582854775711894,
0.010526592843234539,
-0.043945059180259705,
0.08654814213514328,
0.02875732257962227,
0.01958903670310974,
-0.12828795611858368,
0.07599804550409317,
-0.02586091123521328,
-0.14057785272598267,
0.03576691821217537,
0.1520865559577942,
-0.13470661640167236,
-0.1196122094988823,
0.02517283894121647,
0.07615108788013458,
-0.15355950593948364,
-0.009285707026720047,
-0.01853850856423378,
-0.1146107092499733,
0.08270766586065292,
0.16739656031131744,
0.06393714994192123,
0.09372229129076004,
-0.016709135845303535,
0.00033211661502718925,
-0.017629390582442284,
0.0060826437547802925,
0.015956709161400795,
0.038097936660051346,
-0.0988348126411438,
0.06918919086456299,
-0.029006076976656914,
0.18566696345806122,
-0.08008981496095657,
-0.05143186077475548,
-0.15967725217342377,
-0.001775190932676196,
-0.08343244343996048,
-0.05541709065437317,
-0.10140618681907654,
-0.06321018189191818,
-0.023288492113351822,
-0.034263819456100464,
-0.038036808371543884,
-0.041667092591524124,
-0.1316547989845276,
0.01385523658245802,
-0.07998481392860413,
0.024233905598521233,
-0.031159712001681328,
-0.0043153963051736355,
0.12203552573919296,
-0.05407611280679703,
0.13436287641525269,
0.12337575852870941,
-0.04349830374121666,
0.1346055269241333,
-0.08672069013118744,
-0.05363473296165466,
0.10277332365512848,
0.019689666107296944,
0.009308061562478542,
0.07256065309047699,
0.02191895991563797,
0.012116474099457264,
0.05955912545323372,
0.06519178301095963,
0.09907940775156021,
-0.10559206455945969,
0.02268049865961075,
-0.03747428208589554,
-0.1432260125875473,
-0.057319775223731995,
-0.03375370427966118,
0.022990288212895393,
0.02676946483552456,
0.07909193634986877,
-0.05677521973848343,
0.1060420423746109,
-0.04767235741019249,
0.04007749632000923,
0.0009903513127937913,
-0.16581644117832184,
-0.037122491747140884,
-0.06677180528640747,
0.030188927426934242,
0.01579337939620018,
0.2170153260231018,
-0.004767094273120165,
-0.031621865928173065,
0.02309555560350418,
0.04177442193031311,
-0.008157406933605671,
0.008448606356978416,
0.20548230409622192,
0.10310740768909454,
-0.06846706569194794,
-0.14684386551380157,
0.08422907441854477,
0.0026786397211253643,
-0.06211854889988899,
0.16172944009304047,
-0.01823771744966507,
0.028763005509972572,
0.08257352560758591,
-0.02413180097937584,
0.015303188003599644,
-0.14920642971992493,
-0.10885327309370041,
-0.049236614257097244,
0.07946305721998215,
-0.00982586108148098,
0.12070965021848679,
0.11676641553640366,
-0.04379433020949364,
0.0015108027728274465,
-0.012010033242404461,
-0.05686352401971817,
-0.17215904593467712,
-0.21417103707790375,
-0.08673834055662155,
-0.13437552750110626,
0.0188327394425869,
-0.10698898136615753,
0.05854099616408348,
-0.0021438240073621273,
0.056589897722005844,
-0.049702372401952744,
0.1177777424454689,
0.04731317237019539,
-0.12980766594409943,
0.08827393501996994,
-0.017468633130192757,
0.02003621682524681,
0.0049504381604492664,
0.004185068421065807,
-0.08732683956623077,
0.025415578857064247,
0.0007953525637276471,
0.0564325787127018,
-0.06978806853294373,
-0.0034391884692013264,
-0.12339784950017929,
-0.10002655535936356,
-0.05990864336490631,
0.05073666572570801,
-0.02586827427148819,
0.08469163626432419,
0.01148642785847187,
-0.022830944508314133,
0.04015327990055084,
0.1898038536310196,
-0.06600670516490936,
-0.0655871033668518,
-0.021236835047602654,
0.19605761766433716,
0.01429776381701231,
0.08746463060379028,
-0.011506977491080761,
0.01829349622130394,
-0.04844636097550392,
0.3601461946964264,
0.2312781661748886,
-0.04722781479358673,
0.03223119303584099,
0.014052394777536392,
0.03352605551481247,
0.1355055868625641,
0.11694815009832382,
0.08269927650690079,
0.23108479380607605,
-0.10545637458562851,
-0.08099020272493362,
-0.0423525795340538,
-0.01121626514941454,
-0.08901078999042511,
0.06734571605920792,
0.03955891728401184,
-0.08915092796087265,
-0.05525268241763115,
0.08274523913860321,
-0.23970729112625122,
0.11162406951189041,
-0.08171556890010834,
-0.162613183259964,
-0.059820499271154404,
0.026959683746099472,
0.09690212458372116,
0.015219182707369328,
0.09199827164411545,
-0.014955630525946617,
-0.07619146257638931,
0.05853145197033882,
0.05077424272894859,
-0.2382902354001999,
0.03595710173249245,
0.05467145889997482,
-0.05854099243879318,
0.007449338678270578,
-0.018502935767173767,
0.06915213912725449,
0.06619628518819809,
0.04410887509584427,
-0.05260109901428223,
0.015458744950592518,
-0.0034508355893194675,
-0.04801314324140549,
-0.02780556119978428,
0.10057107359170914,
0.034579407423734665,
-0.1206040009856224,
0.09023909270763397,
-0.10645681619644165,
0.027654634788632393,
0.06223444640636444,
-0.0004995701601728797,
-0.020608041435480118,
0.021370284259319305,
-0.0963246300816536,
0.05579786375164986,
0.07245944440364838,
-0.015751751139760017,
-0.01680879481136799,
-0.06924273818731308,
-0.020133692771196365,
-0.03015086241066456,
-0.04436968266963959,
-0.06642002612352371,
-0.11278419196605682,
-0.10139625519514084,
0.056709371507167816,
0.02185703255236149,
-0.15141497552394867,
0.015565968118607998,
-0.11661290377378464,
0.0685729831457138,
-0.17312651872634888,
0.12360571324825287,
0.024642599746584892,
-0.00669059669598937,
0.0005565152387134731,
-0.09654258191585541,
0.029081624001264572,
0.0706893652677536,
-0.11944233626127243,
-0.09647279977798462
] |
null | null |
transformers
|
#Rick Dialogpt model
|
{"tags": ["conversational"]}
|
text-generation
|
crystalgate/DialoGPT-small-rick
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
#Rick Dialogpt model
|
[] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
51
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
-0.009697278961539268,
0.03208012506365776,
-0.007204889785498381,
0.004809224978089333,
0.16726240515708923,
0.014898733235895634,
0.09765533357858658,
0.13672804832458496,
-0.007841327227652073,
-0.031050153076648712,
0.14490588009357452,
0.20411323010921478,
-0.006439372431486845,
0.0661218985915184,
-0.07572533935308456,
-0.2683109939098358,
0.05759621039032936,
0.046649303287267685,
0.016515716910362244,
0.1200079694390297,
0.08573378622531891,
-0.05473608896136284,
0.08714032918214798,
-0.014583407901227474,
-0.150366872549057,
0.017733458429574966,
0.043394338339567184,
-0.12260226160287857,
0.11910516023635864,
0.05462685227394104,
0.07063519209623337,
0.014929565601050854,
-0.07541623711585999,
-0.1631229966878891,
0.03031250834465027,
0.01425902172923088,
-0.0594632662832737,
0.04757995903491974,
0.059961482882499695,
-0.10165371745824814,
0.10819483548402786,
0.09530027210712433,
-0.013078106567263603,
0.06798283755779266,
-0.16849711537361145,
-0.020869607105851173,
-0.01446688175201416,
0.009899779222905636,
0.05550243332982063,
0.09964893013238907,
-0.03413357585668564,
0.10497362166643143,
-0.09214533120393753,
0.11017382889986038,
0.10932035744190216,
-0.32057443261146545,
-0.005767723545432091,
0.09167823940515518,
0.039358653128147125,
0.07352814823389053,
-0.04467793554067612,
0.06258884817361832,
0.018015462905168533,
0.017986174672842026,
-0.014015024527907372,
-0.07283061742782593,
-0.11612214148044586,
0.04717336222529411,
-0.08668071031570435,
-0.059868961572647095,
0.2244078367948532,
-0.05464440956711769,
0.06881742179393768,
-0.05281897634267807,
-0.10522868484258652,
-0.04308144748210907,
-0.029833965003490448,
0.00475557055324316,
-0.07660607248544693,
0.08692064881324768,
0.00869679357856512,
-0.09547875821590424,
-0.1376667022705078,
-0.02496783249080181,
-0.1776352822780609,
0.16140350699424744,
0.02465328387916088,
0.05232657864689827,
-0.2027255892753601,
0.09623090922832489,
0.017906051129102707,
-0.08045592904090881,
0.022091427817940712,
-0.10046248883008957,
0.029131146147847176,
0.013760408386588097,
-0.04754498973488808,
-0.061387211084365845,
0.0843690037727356,
0.11199145019054413,
-0.01731434464454651,
0.025486016646027565,
-0.039331406354904175,
0.08100687712430954,
0.03553595021367073,
0.09077847748994827,
0.007288969587534666,
-0.028338588774204254,
0.025842782109975815,
-0.13719046115875244,
-0.003647835226729512,
-0.07116208970546722,
-0.16572439670562744,
-0.021088803187012672,
0.02994808368384838,
0.08289173990488052,
0.015449047088623047,
0.11682453751564026,
-0.03272046521306038,
-0.025152435526251793,
0.03602350503206253,
-0.047656361013650894,
-0.012649794109165668,
0.016648368909955025,
0.013163427822291851,
0.12399329990148544,
-0.0022096503525972366,
0.03235051408410072,
-0.13653022050857544,
0.031423524022102356,
-0.06793295592069626,
-0.003740974934771657,
-0.03486552834510803,
-0.040637075901031494,
0.009043924510478973,
-0.06862333416938782,
0.003486064961180091,
-0.15030112862586975,
-0.15063877403736115,
0.007587034720927477,
-0.007836631499230862,
-0.04107699543237686,
-0.06370922178030014,
-0.06952770054340363,
-0.013550350442528725,
0.04251532256603241,
-0.07093454152345657,
-0.011352915316820145,
-0.06403283774852753,
0.11004766076803207,
-0.03197755664587021,
0.07921615242958069,
-0.11953279376029968,
0.08390819281339645,
-0.11260783672332764,
-0.02386913076043129,
-0.060801517218351364,
0.09317506104707718,
-0.0006014376995153725,
0.09549830108880997,
-0.006563255097717047,
-0.017931854352355003,
-0.07981178909540176,
0.06445012241601944,
-0.042872510850429535,
0.21701598167419434,
-0.0615808479487896,
-0.11181682348251343,
0.28781595826148987,
-0.052628401666879654,
-0.1370542049407959,
0.11647392809391022,
0.008682746440172195,
0.05777018144726753,
0.10703510791063309,
0.19733482599258423,
-0.015276194550096989,
0.004040541127324104,
0.09471915662288666,
0.11263324320316315,
-0.11276852339506149,
-0.033160366117954254,
0.013019153848290443,
-0.04081077128648758,
-0.10867965966463089,
0.04689536616206169,
0.09810488671064377,
0.07090286910533905,
-0.04786505550146103,
-0.03377414867281914,
-0.01366397924721241,
0.0052589005790650845,
0.08885077387094498,
-0.007157256826758385,
0.10962837189435959,
-0.05819983780384064,
-0.03796621412038803,
-0.029282379895448685,
-0.012126247398555279,
-0.03951939567923546,
0.03137664496898651,
-0.043376367539167404,
0.10821941494941711,
-0.011204327456653118,
0.06364280730485916,
-0.16185984015464783,
-0.07691477984189987,
-0.017002692446112633,
0.1581239402294159,
0.024538565427064896,
0.09859629720449448,
0.0552486926317215,
-0.040398042649030685,
-0.0012767292791977525,
0.012792680412530899,
0.15581141412258148,
-0.022091681137681007,
-0.065607450902462,
-0.052166227251291275,
0.08642971515655518,
-0.05641226842999458,
0.04504093527793884,
-0.05937713757157326,
0.012367865070700645,
0.05064384639263153,
0.10342344641685486,
-0.00018274025933351368,
0.03323284164071083,
-0.008164864964783192,
0.002145637758076191,
-0.058205123990774155,
0.007405933458358049,
0.10799351334571838,
0.00036868182360194623,
-0.07365862280130386,
0.22074243426322937,
-0.17796069383621216,
0.1765957772731781,
0.1893044263124466,
-0.299345999956131,
0.017949223518371582,
-0.10759581625461578,
-0.04561871662735939,
0.014407722279429436,
0.05567655712366104,
-0.0454222597181797,
0.1703362911939621,
-0.009871348738670349,
0.18874616920948029,
-0.04946064203977585,
-0.04464937001466751,
-0.0200483538210392,
-0.05118836089968681,
-0.0024189651012420654,
0.07781197130680084,
0.10685696452856064,
-0.13992026448249817,
0.1964332014322281,
0.1621224284172058,
0.048237916082143784,
0.19945049285888672,
0.015346456319093704,
-0.011589210480451584,
0.0909530371427536,
0.005220826715230942,
-0.058739423751831055,
-0.07409929484128952,
-0.2594851851463318,
-0.030033592134714127,
0.07992640137672424,
0.0422382652759552,
0.1212305948138237,
-0.11349532753229141,
-0.038956157863140106,
-0.01763172075152397,
-0.023146281018853188,
0.021672505885362625,
0.0914369598031044,
0.06075398623943329,
0.13201528787612915,
-0.001710098935291171,
-0.007300339173525572,
0.10524573177099228,
0.01783694699406624,
-0.09354141354560852,
0.18308524787425995,
-0.13652534782886505,
-0.37097251415252686,
-0.13911493122577667,
-0.18057456612586975,
-0.05449081212282181,
0.05712554603815079,
0.11679314076900482,
-0.12011238187551498,
-0.018752124160528183,
0.01578843593597412,
0.10931742936372757,
-0.08449502289295197,
0.0021454424131661654,
-0.06880278885364532,
0.0321490578353405,
-0.10310184955596924,
-0.09194442629814148,
-0.055416494607925415,
-0.031392451375722885,
-0.08001253753900528,
0.1423761546611786,
-0.10777941346168518,
0.04476889222860336,
0.20262959599494934,
0.04653622955083847,
0.05625178664922714,
-0.044105201959609985,
0.19377262890338898,
-0.11264272034168243,
-0.01661740615963936,
0.19215328991413116,
-0.048360925167798996,
0.07476246356964111,
0.1232115849852562,
-0.006348740309476852,
-0.08765771239995956,
0.03011748194694519,
-0.02085109055042267,
-0.07988511025905609,
-0.23219464719295502,
-0.13938382267951965,
-0.12429051846265793,
0.09477275609970093,
0.028005298227071762,
0.056365787982940674,
0.17219258844852448,
0.06577219814062119,
-0.038416244089603424,
0.006410336587578058,
0.02959546446800232,
0.08237514644861221,
0.23417828977108002,
-0.06035616248846054,
0.1364797055721283,
-0.03420931473374367,
-0.14982740581035614,
0.08169995993375778,
0.0713929831981659,
0.10213395953178406,
0.06678459793329239,
0.0804823637008667,
0.0149586396291852,
0.06188136339187622,
0.1311223804950714,
0.08191446959972382,
0.019586285576224327,
-0.02480296604335308,
-0.03388110175728798,
-0.025523077696561813,
-0.05937909707427025,
0.040128443390131,
0.06589099019765854,
-0.16763372719287872,
-0.039227183908224106,
-0.09338314831256866,
0.09657008945941925,
0.0873042419552803,
0.06609832495450974,
-0.1842060089111328,
-0.008006223477423191,
0.08488986641168594,
-0.03854905813932419,
-0.13727426528930664,
0.09535189718008041,
0.01523482333868742,
-0.15144726634025574,
0.03139317408204079,
-0.04061909019947052,
0.12188644707202911,
-0.07804752141237259,
0.09809603542089462,
-0.08108244836330414,
-0.07448557764291763,
0.02123199962079525,
0.1261177361011505,
-0.30527687072753906,
0.20240111649036407,
-0.0024993624538183212,
-0.06486981362104416,
-0.1243603527545929,
-0.0032166161108762026,
0.002410882618278265,
0.07357452809810638,
0.10519039630889893,
-0.007196315098553896,
0.001897757756523788,
-0.06300821900367737,
-0.01829923689365387,
0.032471053302288055,
0.13080233335494995,
-0.0401318334043026,
-0.021158374845981598,
-0.050194524228572845,
-0.001653497340157628,
-0.03173094615340233,
-0.06934895366430283,
0.02002747356891632,
-0.19509181380271912,
0.08751901984214783,
0.04166261479258537,
0.09648149460554123,
0.029994789510965347,
0.004265148192644119,
-0.09651939570903778,
0.24698667228221893,
-0.07148019969463348,
-0.10072879493236542,
-0.10919588059186935,
-0.046813901513814926,
0.03569883480668068,
-0.05628936365246773,
0.04309194162487984,
-0.0788632407784462,
0.028997479006648064,
-0.06352769583463669,
-0.19235502183437347,
0.12410202622413635,
-0.09027006477117538,
-0.04412810131907463,
-0.02371402643620968,
0.2110891044139862,
-0.05598580464720726,
0.010335659608244896,
0.02930437959730625,
0.01208863127976656,
-0.11645778268575668,
-0.09678568691015244,
0.031018631532788277,
-0.007351789623498917,
0.050603240728378296,
0.041841957718133926,
-0.05915454775094986,
-0.017138581722974777,
-0.052199993282556534,
-0.022926922887563705,
0.3496883809566498,
0.14231905341148376,
-0.043836336582899094,
0.19347235560417175,
0.12347975373268127,
-0.07452994585037231,
-0.3159443140029907,
-0.1066238060593605,
-0.10937739163637161,
-0.04680149629712105,
-0.07012093812227249,
-0.2002030611038208,
0.06474938243627548,
0.00662544509395957,
-0.013415241613984108,
0.12749312818050385,
-0.2561831772327423,
-0.07571036368608475,
0.15906259417533875,
-0.017980827018618584,
0.3745945692062378,
-0.1168576180934906,
-0.10926306992769241,
-0.03950892388820648,
-0.14175476133823395,
0.16968177258968353,
-0.01989765651524067,
0.11221715062856674,
-0.009765521623194218,
0.14388824999332428,
0.05548359826207161,
-0.023479344323277473,
0.08544106781482697,
0.004999885335564613,
-0.03290518373250961,
-0.10304180532693863,
-0.05676887184381485,
0.007092386484146118,
0.02477436140179634,
0.018026655539870262,
-0.041834570467472076,
0.02227151393890381,
-0.11731979995965958,
-0.04657655209302902,
-0.08982590585947037,
0.04431166127324104,
0.03899754583835602,
-0.07325074821710587,
-0.002380647463724017,
-0.07165111601352692,
-0.012272949330508709,
0.022334342822432518,
0.20356793701648712,
-0.08029330521821976,
0.16448934376239777,
0.09239562600851059,
0.12419285625219345,
-0.14376309514045715,
-0.00019283240544609725,
-0.0762530043721199,
-0.05611240118741989,
0.07737895101308823,
-0.09433035552501678,
0.058893077075481415,
0.10901971161365509,
-0.04567738622426987,
0.08828683942556381,
0.10377411544322968,
0.008936077356338501,
0.003213887568563223,
0.10916902124881744,
-0.2667325437068939,
-0.0296600554138422,
-0.07532413303852081,
0.000883326749317348,
0.09092561900615692,
0.08562852442264557,
0.18840822577476501,
0.025361526757478714,
-0.04293036088347435,
-0.002770674182102084,
0.028597986325621605,
-0.039021048694849014,
0.051667019724845886,
0.001123449532315135,
0.01947369985282421,
-0.1530752182006836,
0.072522833943367,
0.01490565575659275,
-0.15215420722961426,
0.021316176280379295,
0.16572684049606323,
-0.11656328290700912,
-0.1283872276544571,
-0.06520111113786697,
0.08313824236392975,
-0.11755692958831787,
-0.01578943058848381,
-0.03279297426342964,
-0.13145680725574493,
0.07992171496152878,
0.12629036605358124,
0.05557859688997269,
0.0972496047616005,
-0.06061713397502899,
-0.020469192415475845,
-0.018721895292401314,
-0.014099318534135818,
-0.012384648434817791,
-0.007667020428925753,
-0.055978111922740936,
0.0590752474963665,
-0.026677248999476433,
0.1425808072090149,
-0.09221141785383224,
-0.1037059873342514,
-0.16142144799232483,
0.0374140702188015,
-0.11013076454401016,
-0.08825794607400894,
-0.08821134269237518,
-0.050188567489385605,
0.002360827289521694,
-0.019856395199894905,
-0.04037635400891304,
-0.05829505994915962,
-0.12300454825162888,
0.0338277705013752,
-0.040771447122097015,
0.024727050215005875,
-0.07512269169092178,
0.015856385231018066,
0.08507686108350754,
-0.03285100311040878,
0.15655414760112762,
0.1450488418340683,
-0.1006515845656395,
0.10741901397705078,
-0.14806775748729706,
-0.09138492494821548,
0.11116421222686768,
0.015329592861235142,
0.0449691042304039,
0.09723787009716034,
0.013362943194806576,
0.0635865181684494,
0.032776717096567154,
0.05308786407113075,
0.027619892731308937,
-0.11959987878799438,
0.06483134627342224,
-0.03626115620136261,
-0.14700546860694885,
-0.049338050186634064,
-0.05282869189977646,
0.01647452637553215,
0.013054544106125832,
0.09622690081596375,
-0.05301849544048309,
0.10698331147432327,
-0.04055701196193695,
0.0346808135509491,
0.017554637044668198,
-0.1730053424835205,
-0.03816922754049301,
-0.08538098633289337,
0.03681723028421402,
0.014741539023816586,
0.25266793370246887,
0.030072299763560295,
0.012416383251547813,
0.032671261578798294,
0.08285367488861084,
0.03899408504366875,
0.010228337720036507,
0.17482228577136993,
0.1162426546216011,
-0.06621865928173065,
-0.10445023328065872,
0.0729617029428482,
0.016332454979419708,
0.01286179106682539,
0.13617953658103943,
0.008365051820874214,
0.005795429926365614,
0.08649782836437225,
-0.016865963116288185,
0.009968153201043606,
-0.10052056610584259,
-0.13426925241947174,
-0.022176474332809448,
0.05151832848787308,
-0.04655967652797699,
0.11727844923734665,
0.1406494379043579,
-0.01806013658642769,
0.03222079202532768,
-0.021771740168333054,
-0.05699979141354561,
-0.1683429479598999,
-0.1429590880870819,
-0.06883849948644638,
-0.13416796922683716,
0.00897989235818386,
-0.11180389672517776,
0.05395037308335304,
0.06001098081469536,
0.06750501692295074,
-0.06899319589138031,
0.10220931470394135,
0.04626858979463577,
-0.11440542340278625,
0.06264589726924896,
-0.0296088308095932,
0.09430401772260666,
-0.02759445086121559,
-0.019505485892295837,
-0.09039592742919922,
0.014574515633285046,
0.011419114656746387,
0.06245238706469536,
-0.04707273095846176,
0.007463190704584122,
-0.14696238934993744,
-0.08972041308879852,
-0.0523175448179245,
0.0718572810292244,
-0.050409089773893356,
0.14282815158367157,
0.00775480642914772,
-0.0170906875282526,
0.039554283022880554,
0.22787313163280487,
-0.07476283609867096,
-0.04778539761900902,
-0.05269690603017807,
0.20717895030975342,
0.02975541539490223,
0.1171872541308403,
-0.022938819602131844,
-0.006106364540755749,
-0.0919521227478981,
0.3764844834804535,
0.30030161142349243,
-0.09031439572572708,
0.011794124729931355,
0.02137952297925949,
0.04502861574292183,
0.1316293478012085,
0.1216534823179245,
0.10318691283464432,
0.3006802201271057,
-0.07452366501092911,
-0.04653361067175865,
-0.012629742734134197,
-0.023858042433857918,
-0.09059546142816544,
0.1021224707365036,
0.04839762672781944,
-0.06382183730602264,
-0.03313443064689636,
0.0954432487487793,
-0.25862133502960205,
0.1277991235256195,
-0.12311873584985733,
-0.17578600347042084,
-0.06654827296733856,
0.009760108776390553,
0.10465722531080246,
0.015642458572983742,
0.0946015790104866,
0.007128213066607714,
-0.11252258718013763,
0.06305865943431854,
0.03397420793771744,
-0.22762253880500793,
0.0006893770187161863,
0.06642123311758041,
-0.07006710022687912,
-0.0024247700348496437,
-0.026499588042497635,
0.05657242611050606,
0.0656052976846695,
0.054629553109407425,
-0.00971333310008049,
0.03816632181406021,
0.0034184439573436975,
-0.0585215799510479,
0.016623929142951965,
0.05121519789099693,
0.02472509816288948,
-0.09763528406620026,
0.06927435845136642,
-0.1574270874261856,
0.04766253009438515,
-0.0030655991286039352,
-0.04124255105853081,
0.006064958870410919,
0.008823691867291927,
-0.06491616368293762,
0.05165379121899605,
0.07916834205389023,
-0.0016257909592241049,
-0.0062433634884655476,
-0.057178743183612823,
-0.02632102556526661,
-0.027755750343203545,
-0.09291748702526093,
-0.10495562851428986,
-0.14682936668395996,
-0.11640441417694092,
0.09368976950645447,
-0.01011267676949501,
-0.1848134547472,
0.022154374048113823,
-0.08606051653623581,
0.08319322764873505,
-0.1670055389404297,
0.08040720224380493,
0.07041648775339127,
0.013038921169936657,
-0.0031511052511632442,
-0.02002427540719509,
0.054132770746946335,
0.086809903383255,
-0.10407156497240067,
-0.07400695979595184
] |
null | null |
spacy
|
NER Model for 'Ministerratsprotokolle'
| Feature | Description |
| --- | --- |
| **Name** | `de_MRP_NER` |
| **Version** | `0.0.0` |
| **spaCy** | `>=3.1.0,<3.2.0` |
| **Default Pipeline** | `tok2vec`, `ner` |
| **Components** | `tok2vec`, `ner` |
| **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
| **Sources** | n/a |
| **License** | `cc-by` |
| **Author** | [Peter Andorfer]() |
### Label Scheme
<details>
<summary>View label scheme (4 labels for 1 components)</summary>
| Component | Labels |
| --- | --- |
| **`ner`** | `GPE`, `LOC`, `ORG`, `PER` |
</details>
### Accuracy
| Type | Score |
| --- | --- |
| `ENTS_F` | 88.04 |
| `ENTS_P` | 90.53 |
| `ENTS_R` | 85.69 |
| `TOK2VEC_LOSS` | 40077.56 |
| `NER_LOSS` | 77727.57 |
|
{"language": ["de"], "license": "cc-by-4.0", "tags": ["spacy", "token-classification"]}
|
token-classification
|
csae8092/de_MRP_NER
|
[
"spacy",
"token-classification",
"de",
"license:cc-by-4.0",
"model-index",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#spacy #token-classification #de #license-cc-by-4.0 #model-index #region-us
|
NER Model for 'Ministerratsprotokolle'
### Label Scheme
View label scheme (4 labels for 1 components)
### Accuracy
|
[
"### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)",
"### Accuracy"
] |
[
"TAGS\n#spacy #token-classification #de #license-cc-by-4.0 #model-index #region-us \n",
"### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)",
"### Accuracy"
] |
[
30,
16,
5
] |
[
"passage: TAGS\n#spacy #token-classification #de #license-cc-by-4.0 #model-index #region-us \n### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)### Accuracy"
] |
[
-0.06385856866836548,
0.13472512364387512,
-0.003094917628914118,
0.0397038497030735,
0.07022737711668015,
0.06061854213476181,
0.2767327129840851,
0.05958044156432152,
0.2079005092382431,
0.02886129729449749,
0.061364416033029556,
0.09853294491767883,
0.04757606238126755,
0.21973487734794617,
-0.09127046167850494,
-0.21529531478881836,
0.0958523154258728,
-0.005406170152127743,
0.06667861342430115,
0.11700994521379471,
0.08378114551305771,
-0.12813547253608704,
0.06597858667373657,
-0.04171180725097656,
-0.22465991973876953,
0.014053160324692726,
0.03370482847094536,
-0.1148187518119812,
0.0720188170671463,
-0.04196806624531746,
0.17148704826831818,
0.08092565089464188,
0.07878386229276657,
-0.20359699428081512,
-0.0030305241234600544,
-0.039836589246988297,
-0.09770330041646957,
0.0686204805970192,
0.05432229861617088,
0.016963504254817963,
-0.03868686407804489,
-0.030729399994015694,
0.05363672226667404,
0.06719394773244858,
-0.12571488320827484,
-0.17105752229690552,
-0.06533747911453247,
0.1414593756198883,
0.10574507713317871,
-0.04827113449573517,
-0.0008316130843013525,
0.04904069006443024,
-0.11670505255460739,
0.03754619136452675,
0.09219346940517426,
-0.3793019652366638,
0.02676801197230816,
0.25279706716537476,
-0.03578133136034012,
0.10048729926347733,
-0.04890355467796326,
0.13123422861099243,
0.13417236506938934,
-0.011546103283762932,
-0.006435838062316179,
-0.018642647191882133,
0.11277974396944046,
0.03329147771000862,
-0.09689734876155853,
-0.06436038017272949,
0.5035092234611511,
0.08564742654561996,
-0.05668892711400986,
-0.10495588183403015,
-0.027417441830039024,
-0.12387033551931381,
-0.08367842435836792,
-0.03049219585955143,
0.08829320967197418,
0.04531635716557503,
0.09365047514438629,
0.11988046765327454,
-0.10377410054206848,
-0.06064224988222122,
-0.15473124384880066,
0.163921520113945,
0.004786196164786816,
0.09316974133253098,
-0.13639798760414124,
0.024215595796704292,
-0.10437284409999847,
-0.06111231446266174,
-0.00911126658320427,
-0.07900924980640411,
-0.045119304209947586,
-0.03997853025794029,
0.028874948620796204,
0.17723388969898224,
0.08720173686742783,
0.10498438775539398,
-0.030816076323390007,
0.010638272389769554,
-0.031197812408208847,
0.07542527467012405,
0.10660627484321594,
0.14602580666542053,
-0.01859539933502674,
-0.005969756282866001,
-0.018562190234661102,
-0.0687531903386116,
0.049070023000240326,
-0.03357533738017082,
-0.17074260115623474,
0.00444003427401185,
0.055237870663404465,
0.11786045134067535,
-0.08242696523666382,
-0.03722166642546654,
-0.12658122181892395,
-0.03671668469905853,
0.15735626220703125,
-0.0961843952536583,
0.021880265325307846,
-0.002741612959653139,
-0.007103442680090666,
0.07422768324613571,
-0.12446694821119308,
-0.010836279951035976,
0.06269129365682602,
0.01369499135762453,
-0.11603740602731705,
-0.002407974563539028,
-0.013717913068830967,
-0.08804995566606522,
0.045516520738601685,
-0.10310220718383789,
0.03914336487650871,
-0.053654979914426804,
-0.1452307552099228,
-0.0009558024466969073,
-0.030147681012749672,
-0.07168180495500565,
-0.009538908489048481,
-0.013986019417643547,
-0.07069491595029831,
-0.01845092885196209,
0.002758969785645604,
-0.07804422825574875,
-0.10382562130689621,
0.0034123361110687256,
-0.03263712301850319,
0.06887941062450409,
-0.12672613561153412,
0.02941019833087921,
-0.06512124091386795,
0.05166752263903618,
-0.1610855907201767,
0.02406146377325058,
-0.09226509183645248,
0.06466101109981537,
-0.07589632272720337,
-0.09269633889198303,
0.061340827494859695,
-0.011624468490481377,
-0.13421355187892914,
0.14594314992427826,
-0.22205208241939545,
-0.0526152066886425,
0.16571009159088135,
-0.19176805019378662,
-0.12826240062713623,
0.03924340754747391,
-0.006549354176968336,
0.05590920150279999,
0.08420353382825851,
0.12143457680940628,
-0.02370697259902954,
-0.1314009130001068,
-0.025085048750042915,
0.11224989593029022,
-0.04297644644975662,
-0.0785822868347168,
0.10934149473905563,
-0.023989448323845863,
-0.0004887625109404325,
0.04426784813404083,
-0.014842969365417957,
-0.1327861100435257,
-0.05040987953543663,
-0.06960400938987732,
-0.019735239446163177,
0.032347410917282104,
0.03897624462842941,
0.02424178458750248,
0.019212277606129646,
-0.04772020876407623,
0.012773526832461357,
0.03171473369002342,
0.05064195767045021,
0.03200134262442589,
-0.058171194046735764,
-0.06322992593050003,
0.12199145555496216,
-0.12139715999364853,
-0.0672859326004982,
-0.1207939088344574,
-0.1298864334821701,
0.05583832412958145,
0.023324623703956604,
0.05183057114481926,
0.15762446820735931,
-0.01785740628838539,
0.007647615857422352,
0.0006017946288920939,
-0.00009712190512800589,
-0.0336846299469471,
0.08433259278535843,
-0.02969747968018055,
-0.20240919291973114,
-0.048181142657995224,
-0.058493755757808685,
0.07329977303743362,
-0.045227501541376114,
0.01508280634880066,
0.21874359250068665,
0.04697594419121742,
0.019380390644073486,
0.06796710193157196,
0.003552958369255066,
0.04556804150342941,
-0.04112231731414795,
-0.03648826852440834,
0.08591944724321365,
-0.06811971217393875,
-0.07636383175849915,
-0.009264483116567135,
-0.08540673553943634,
0.10322486609220505,
0.16760317981243134,
-0.10279545933008194,
-0.037930283695459366,
-0.0717015340924263,
0.004292863886803389,
0.007437978405505419,
-0.0956098735332489,
-0.003497655736282468,
-0.052054256200790405,
-0.04657944664359093,
0.041597481817007065,
-0.1100899875164032,
-0.01715380698442459,
0.041845791041851044,
-0.02108917385339737,
-0.14812292158603668,
0.06613586097955704,
0.0584745816886425,
-0.185817688703537,
0.1686919927597046,
0.30281689763069153,
0.16255336999893188,
0.056260641664266586,
-0.06412261724472046,
-0.0391269214451313,
-0.06661314517259598,
-0.009954454377293587,
-0.07246392965316772,
0.21468143165111542,
-0.13147565722465515,
-0.036326006054878235,
0.06489772349596024,
0.045830368995666504,
0.0034093984868377447,
-0.218227356672287,
-0.009527834132313728,
-0.03230443596839905,
-0.07123538851737976,
-0.14049088954925537,
-0.031165169551968575,
-0.0004003677167929709,
0.11683294177055359,
0.040234919637441635,
-0.2255738228559494,
0.10721158981323242,
-0.0462210513651371,
-0.08177534490823746,
0.1540534794330597,
-0.10037878900766373,
-0.19356489181518555,
-0.16068489849567413,
-0.0701606422662735,
-0.07641502469778061,
0.05050945654511452,
-0.009261485189199448,
-0.06828142702579498,
-0.05475330352783203,
-0.001720659201964736,
-0.08436471223831177,
-0.16805703938007355,
-0.047997571527957916,
-0.05655145272612572,
0.09935460984706879,
-0.12468291819095612,
-0.07338842749595642,
-0.09220466017723083,
-0.039462942630052567,
0.04420270398259163,
0.09570270031690598,
-0.14521010220050812,
0.0699942484498024,
0.2794201076030731,
-0.03050101548433304,
0.0736946314573288,
-0.040954940021038055,
0.11923367530107498,
-0.06064528971910477,
0.02051813155412674,
0.11583442986011505,
0.0643831118941307,
0.030563564971089363,
0.277820348739624,
0.10009436309337616,
-0.15038803219795227,
-0.0595962256193161,
-0.07407922297716141,
-0.09704037010669708,
-0.16420117020606995,
-0.10058043897151947,
-0.046560510993003845,
0.004016268998384476,
0.036852769553661346,
0.044215861707925797,
0.05284356325864792,
0.04906097427010536,
0.015118678100407124,
0.0204082690179348,
0.06604676693677902,
0.06324142217636108,
0.15622709691524506,
-0.04276072233915329,
0.09965383261442184,
-0.06289228051900864,
-0.025226188823580742,
0.0894511267542839,
0.12093429267406464,
0.1966082602739334,
0.20855025947093964,
0.09481177479028702,
0.11651195585727692,
0.04046018421649933,
0.1559753566980362,
0.04380505904555321,
0.15898334980010986,
-0.013063622638583183,
-0.02398163080215454,
-0.07263662666082382,
0.0375189334154129,
0.07477036118507385,
-0.0328284353017807,
-0.0816185399889946,
-0.06038087233901024,
-0.10970363020896912,
0.07390500605106354,
0.017855538055300713,
0.2346818596124649,
-0.2395005077123642,
0.022497406229376793,
0.10488820821046829,
0.0983944833278656,
-0.054567813873291016,
0.10608436167240143,
-0.02306150086224079,
-0.09193559736013412,
0.06983160227537155,
0.00089328020112589,
0.10323382169008255,
-0.09055528044700623,
-0.008957095444202423,
-0.005206885281950235,
-0.06979008764028549,
0.00707263732329011,
0.07507967203855515,
-0.12020961195230484,
0.29240480065345764,
0.0426936149597168,
-0.028578734025359154,
-0.06788825243711472,
-0.008076646365225315,
-0.0007517561316490173,
0.25961899757385254,
0.22250711917877197,
0.04293669015169144,
-0.2164505124092102,
-0.2013159841299057,
-0.032605595886707306,
-0.026546606793999672,
0.1415209025144577,
-0.04416179656982422,
0.01667141355574131,
0.021114172413945198,
0.014406885951757431,
-0.017347004264593124,
0.022742174565792084,
-0.07771234214305878,
-0.03618045151233673,
0.031241199001669884,
0.09718374162912369,
-0.050463031977415085,
-0.03520617634057999,
-0.05626020208001137,
-0.18548449873924255,
0.11872873455286026,
-0.09660688787698746,
-0.09831982105970383,
-0.08909325301647186,
-0.035723503679037094,
0.07262872904539108,
0.00034884762135334313,
-0.003604019060730934,
-0.051312677562236786,
0.11295121163129807,
0.0037790383212268353,
-0.09101162105798721,
0.10814189165830612,
-0.044469308108091354,
-0.05510375276207924,
-0.05687025561928749,
0.15071026980876923,
-0.05432809889316559,
-0.0062538031488657,
0.08159001916646957,
0.08229915052652359,
-0.007715534884482622,
-0.1229311153292656,
0.10220275074243546,
-0.007388601545244455,
0.061741188168525696,
0.2304355949163437,
-0.1039058044552803,
-0.13919441401958466,
-0.029209161177277565,
0.043926987797021866,
0.07515872269868851,
0.2951124608516693,
-0.10481518507003784,
0.10352829098701477,
0.09339039772748947,
-0.04291057214140892,
-0.19017571210861206,
-0.034780971705913544,
-0.17624826729297638,
-0.001757705700583756,
-0.015936270356178284,
-0.06501487642526627,
0.1285591572523117,
0.0648246556520462,
-0.07398983091115952,
0.11633309721946716,
-0.2414579540491104,
-0.08161857724189758,
0.1729169636964798,
0.04181941971182823,
0.16291682422161102,
-0.0755450427532196,
-0.11555517464876175,
-0.07985201478004456,
-0.19725216925144196,
0.14310023188591003,
0.0005399496294558048,
0.08442310988903046,
-0.04212398827075958,
-0.005843115039169788,
-0.0065605719573795795,
0.001037825015373528,
0.24139390885829926,
0.12887810170650482,
0.11884041875600815,
0.03965285047888756,
-0.17397071421146393,
0.24620196223258972,
0.011072013527154922,
0.021278368309140205,
0.11777789145708084,
0.0003030449151992798,
-0.08528976142406464,
-0.013182105496525764,
0.01791825145483017,
0.015533043071627617,
-0.08250680565834045,
-0.0635489895939827,
-0.09078391641378403,
0.011602002196013927,
-0.08558391034603119,
-0.07019177079200745,
0.2477424442768097,
-0.055860526859760284,
0.05654608830809593,
0.17641717195510864,
0.0071913874708116055,
-0.10680430382490158,
-0.056847911328077316,
-0.07235399633646011,
-0.08549944311380386,
0.05462358519434929,
-0.1716119945049286,
0.04486902058124542,
0.11605849117040634,
0.03764433413743973,
0.1458284556865692,
0.11573556065559387,
-0.025558974593877792,
-0.03858017176389694,
0.11410942673683167,
-0.0871037021279335,
-0.16330817341804504,
0.01956498622894287,
-0.19243831932544708,
0.03317764028906822,
0.07748282700777054,
0.07294271886348724,
0.007732516620308161,
-0.025997038930654526,
0.022729454562067986,
0.038057610392570496,
-0.06932798027992249,
0.08340159058570862,
0.010080228559672832,
0.0530225895345211,
-0.15094584226608276,
0.16154132783412933,
0.08157075196504593,
0.015968071296811104,
-0.08788306266069412,
-0.08414801955223083,
-0.1479039192199707,
-0.05553516745567322,
-0.043264977633953094,
0.11175774782896042,
-0.14274948835372925,
-0.13104818761348724,
-0.10925167798995972,
-0.19520559906959534,
0.016354035586118698,
0.0734907016158104,
0.14293748140335083,
0.12293282151222229,
0.02050197124481201,
-0.10516117513179779,
0.03622541204094887,
0.05650341510772705,
-0.09742678701877594,
0.024806447327136993,
-0.21353350579738617,
0.04280906170606613,
-0.023841651156544685,
0.08768045157194138,
-0.09157887101173401,
-0.0540555939078331,
-0.1144741028547287,
0.008720414713025093,
-0.07954091578722,
0.07619575411081314,
-0.05949022248387337,
0.00260352180339396,
0.01619233749806881,
0.005654802080243826,
-0.050586551427841187,
0.0011580080026760697,
-0.09682515263557434,
0.051886383444070816,
0.02870391309261322,
0.17768704891204834,
-0.09211369603872299,
-0.02339390479028225,
0.06155183166265488,
-0.036514125764369965,
0.07221446186304092,
0.023940272629261017,
0.025443673133850098,
0.04981645569205284,
-0.14346326887607574,
0.01962137594819069,
0.09380312263965607,
0.02325369603931904,
0.07448268681764603,
-0.12157528847455978,
-0.007448112592101097,
0.05225950852036476,
-0.08137289434671402,
0.08750279992818832,
-0.08615986257791519,
-0.11477933824062347,
-0.14613686501979828,
-0.1110873743891716,
-0.11810288578271866,
-0.03426399827003479,
0.030515458434820175,
0.22328586876392365,
0.035684701055288315,
0.05442927032709122,
0.03418609872460365,
-0.0015771803446114063,
-0.04475044459104538,
-0.019530853256583214,
-0.07041628658771515,
-0.08664001524448395,
-0.07220181077718735,
-0.01607312634587288,
-0.013425714336335659,
-0.031976502388715744,
0.36454659700393677,
0.05732744559645653,
-0.011970520950853825,
0.05629134178161621,
0.19290320575237274,
-0.017126457765698433,
0.0449921153485775,
0.2833068370819092,
0.061492159962654114,
-0.031086483970284462,
0.018635911867022514,
0.037992510944604874,
-0.005333461798727512,
0.015527223236858845,
0.1400713324546814,
0.10933893918991089,
-0.08730760216712952,
0.05873912200331688,
0.0780143216252327,
-0.03330259770154953,
-0.03762426599860191,
0.10519281774759293,
0.07444004714488983,
0.042111918330192566,
0.03479110077023506,
-0.07125592976808548,
0.12115585058927536,
-0.1791415512561798,
0.09288692474365234,
-0.006179723888635635,
-0.08739826083183289,
-0.17860956490039825,
-0.10152742266654968,
-0.08148526400327682,
-0.07028895616531372,
0.02150050736963749,
-0.10600727796554565,
-0.03871189057826996,
0.17212988436222076,
0.020535040646791458,
-0.023896463215351105,
-0.023610737174749374,
-0.26378002762794495,
0.006225386634469032,
0.10995002835988998,
0.013883452862501144,
-0.006323289126157761,
-0.07754497975111008,
-0.011102365329861641,
0.013827439397573471,
-0.09627052396535873,
-0.05095994099974632,
-0.029241446405649185,
0.06731701642274857,
-0.043965455144643784,
-0.15324409306049347,
-0.05358757823705673,
-0.0366339348256588,
-0.006860018707811832,
-0.013555695302784443,
-0.02414824068546295,
0.02549726888537407,
-0.017940975725650787,
0.04175939783453941,
0.198669895529747,
-0.04961869493126869,
0.0036279840860515833,
-0.0596187561750412,
0.23645314574241638,
-0.04969308525323868,
0.08447844535112381,
0.071415014564991,
-0.07115527242422104,
0.002547794720157981,
0.08144151419401169,
0.1945222020149231,
-0.011031047441065311,
-0.012736010365188122,
0.017325596883893013,
-0.0033969159703701735,
0.045961786061525345,
0.01874379627406597,
-0.07100801914930344,
0.12705321609973907,
-0.060855668038129807,
0.07399261742830276,
-0.10312953591346741,
-0.00028696656227111816,
-0.05199654400348663,
-0.03880583867430687,
0.1205589696764946,
-0.08417676389217377,
-0.12917931377887726,
0.18815116584300995,
-0.04628484323620796,
0.02685202844440937,
0.2880057990550995,
-0.0838184803724289,
-0.08413383364677429,
0.0018586480291560292,
-0.0019251869525760412,
-0.010824977420270443,
0.04467420279979706,
-0.14644813537597656,
0.013314886018633842,
0.06302494555711746,
0.017491688951849937,
-0.19339530169963837,
-0.05360937491059303,
0.03869786486029625,
-0.01130244042724371,
0.09033926576375961,
-0.00819533970206976,
0.2109563946723938,
0.09916825592517853,
-0.00021478664712049067,
-0.07667802274227142,
0.0918726995587349,
-0.0004323174653109163,
0.036239538341760635,
-0.0003783024149015546,
0.023019962012767792,
-0.009633944369852543,
-0.055316261947155,
0.09364426881074905,
-0.0695633813738823,
0.010846465826034546,
0.017939193174242973,
-0.1260785311460495,
-0.07138343900442123,
0.03084569238126278,
-0.1096898540854454,
0.08493559807538986,
0.052276045083999634,
-0.027786660939455032,
-0.029946936294436455,
0.018320240080356598,
0.07753024995326996,
0.10255879908800125,
-0.10511687397956848,
0.027729913592338562,
0.01991782896220684,
0.01782114803791046,
0.10816572606563568,
-0.015320327132940292,
-0.1585107147693634,
-0.028531674295663834,
-0.05843463912606239,
0.030427468940615654,
-0.07957793027162552,
0.10330179333686829,
0.13538327813148499,
0.05102115124464035,
-0.05644628405570984,
-0.19549737870693207,
0.03676711767911911,
0.0685453936457634,
-0.0640287771821022,
-0.08285749703645706
] |
null | null |
spacy
|
Regensburger Reichstag von 1576
| Feature | Description |
| --- | --- |
| **Name** | `de_RTA_NER` |
| **Version** | `0.0.0` |
| **spaCy** | `>=3.1.0,<3.2.0` |
| **Default Pipeline** | `tok2vec`, `ner` |
| **Components** | `tok2vec`, `ner` |
| **Vectors** | 0 keys, 0 unique vectors (0 dimensions) |
| **Sources** | n/a |
| **License** | `https://creativecommons.org/licenses/by-nc/4.0/` |
| **Author** | [n/a](https://reichstagsakten-1576.uni-graz.at) |
### Label Scheme
<details>
<summary>View label scheme (4 labels for 1 components)</summary>
| Component | Labels |
| --- | --- |
| **`ner`** | `DATE`, `LOC`, `PER`, `TIME` |
</details>
### Accuracy
| Type | Score |
| --- | --- |
| `ENTS_F` | 86.86 |
| `ENTS_P` | 86.30 |
| `ENTS_R` | 87.43 |
| `TOK2VEC_LOSS` | 43588.74 |
| `NER_LOSS` | 95573.96 |
|
{"language": ["de"], "license": "cc-by-nc-4.0", "tags": ["spacy", "token-classification"]}
|
token-classification
|
csae8092/de_RTA_NER
|
[
"spacy",
"token-classification",
"de",
"license:cc-by-nc-4.0",
"model-index",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"de"
] |
TAGS
#spacy #token-classification #de #license-cc-by-nc-4.0 #model-index #region-us
|
Regensburger Reichstag von 1576
### Label Scheme
View label scheme (4 labels for 1 components)
### Accuracy
|
[
"### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)",
"### Accuracy"
] |
[
"TAGS\n#spacy #token-classification #de #license-cc-by-nc-4.0 #model-index #region-us \n",
"### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)",
"### Accuracy"
] |
[
32,
16,
5
] |
[
"passage: TAGS\n#spacy #token-classification #de #license-cc-by-nc-4.0 #model-index #region-us \n### Label Scheme\n\n\n\nView label scheme (4 labels for 1 components)### Accuracy"
] |
[
-0.06110462173819542,
0.14928708970546722,
-0.0028338648844510317,
0.038839101791381836,
0.06316200643777847,
0.0646965429186821,
0.2631303369998932,
0.06015896797180176,
0.2008504867553711,
0.03368014469742775,
0.06354188919067383,
0.1015159860253334,
0.05020608380436897,
0.21348270773887634,
-0.09233225882053375,
-0.19019414484500885,
0.09389578551054001,
0.0021890117786824703,
0.06711212545633316,
0.12087611854076385,
0.07616858929395676,
-0.12225784361362457,
0.06479385495185852,
-0.040295619517564774,
-0.2286669909954071,
0.010682852938771248,
0.036758389323949814,
-0.11454617977142334,
0.06631971895694733,
-0.04869807884097099,
0.17253954708576202,
0.08619572222232819,
0.07923801243305206,
-0.20459318161010742,
-0.004246303346008062,
-0.03376244381070137,
-0.10251011699438095,
0.07040094584226608,
0.04750203713774681,
0.025502661243081093,
-0.02476944588124752,
-0.0324251651763916,
0.04968176409602165,
0.06787426024675369,
-0.12092024832963943,
-0.17420317232608795,
-0.056716226041316986,
0.14966116845607758,
0.09733416885137558,
-0.04583374410867691,
-0.0024321943055838346,
0.05968613177537918,
-0.11434418708086014,
0.03933097794651985,
0.0816856399178505,
-0.36326414346694946,
0.02808782458305359,
0.2629767060279846,
-0.046797014772892,
0.10794921219348907,
-0.04286104440689087,
0.12650255858898163,
0.1369887739419937,
-0.01352559681981802,
-0.0023801145143806934,
-0.01901690475642681,
0.11377502977848053,
0.03303033858537674,
-0.1014999970793724,
-0.0666455402970314,
0.4941031336784363,
0.09218547493219376,
-0.04978824034333229,
-0.09308982640504837,
-0.028182676061987877,
-0.1282017081975937,
-0.08176214247941971,
-0.03292077034711838,
0.08395244181156158,
0.041040655225515366,
0.0940835028886795,
0.12142830342054367,
-0.09550467878580093,
-0.06849364936351776,
-0.14241454005241394,
0.15228673815727234,
0.0019091127905994654,
0.09312522411346436,
-0.12582635879516602,
0.024333614856004715,
-0.10191629827022552,
-0.06097405403852463,
-0.010760028846561909,
-0.07445398718118668,
-0.047203291207551956,
-0.03998209163546562,
0.036510732024908066,
0.2015213966369629,
0.08162354677915573,
0.08816076815128326,
-0.03832700476050377,
0.010092937387526035,
-0.04223581776022911,
0.07210152596235275,
0.09304200112819672,
0.1339450627565384,
-0.027612878009676933,
0.01115256268531084,
-0.017841944471001625,
-0.06511490046977997,
0.04480023309588432,
-0.03720713406801224,
-0.16685254871845245,
0.0025890308897942305,
0.05136943608522415,
0.11517107486724854,
-0.08015115559101105,
-0.03272959589958191,
-0.1186399981379509,
-0.03685823827981949,
0.1764293760061264,
-0.08968512713909149,
0.019757559522986412,
0.010118530131876469,
-0.0032089592423290014,
0.06719992309808731,
-0.11482031643390656,
-0.015261069871485233,
0.061030495911836624,
0.007496617268770933,
-0.1205148696899414,
-0.005136375781148672,
-0.010322950780391693,
-0.08767126500606537,
0.05129232257604599,
-0.09417334944009781,
0.028626572340726852,
-0.06181781738996506,
-0.1451798677444458,
0.0002215041604358703,
-0.03896224498748779,
-0.06725562363862991,
-0.013918896205723286,
-0.016590874642133713,
-0.08126276731491089,
-0.013168801553547382,
0.006633809767663479,
-0.08834993839263916,
-0.1028381809592247,
0.001417609048075974,
-0.028458813205361366,
0.07016754895448685,
-0.11654146015644073,
0.026977211236953735,
-0.06332800537347794,
0.05549534782767296,
-0.14475572109222412,
0.020050646737217903,
-0.09326419979333878,
0.0570247508585453,
-0.07852895557880402,
-0.10106373578310013,
0.05096474662423134,
-0.02017456851899624,
-0.13295289874076843,
0.14724178612232208,
-0.2153913974761963,
-0.05250290781259537,
0.16301921010017395,
-0.18967191874980927,
-0.13751102983951569,
0.0379890538752079,
-0.0036378768272697926,
0.04848620668053627,
0.0791160985827446,
0.10962922126054764,
-0.0026807901449501514,
-0.12871235609054565,
-0.03102375566959381,
0.11038882285356522,
-0.041480742394924164,
-0.08714690059423447,
0.10417705029249191,
-0.015529789961874485,
-0.021610897034406662,
0.04044807329773903,
-0.018335646018385887,
-0.1302129179239273,
-0.05070408806204796,
-0.07474495470523834,
-0.025225086137652397,
0.032680902630090714,
0.04096785932779312,
0.025761742144823074,
0.007920372299849987,
-0.046142857521772385,
0.008001288399100304,
0.0034461612813174725,
0.05432545393705368,
0.020896684378385544,
-0.052211694419384,
-0.06544879823923111,
0.11389130353927612,
-0.12881024181842804,
-0.06618030369281769,
-0.11233045160770416,
-0.13547652959823608,
0.057874150574207306,
0.01331031508743763,
0.051483966410160065,
0.1475301831960678,
-0.026410100981593132,
0.008838362991809845,
0.002357106190174818,
-0.00048384544788859785,
-0.03306188806891441,
0.08313758671283722,
-0.016073917970061302,
-0.20119942724704742,
-0.0413375161588192,
-0.05324540287256241,
0.08237910270690918,
-0.06827397644519806,
0.017301950603723526,
0.21818594634532928,
0.03783643618226051,
0.011769582517445087,
0.06608058512210846,
0.004398121498525143,
0.05121941119432449,
-0.0392167754471302,
-0.034536831080913544,
0.0866045132279396,
-0.06819479167461395,
-0.08293548971414566,
-0.006530133541673422,
-0.0914611741900444,
0.08176618069410324,
0.16157542169094086,
-0.10185445845127106,
-0.038921061903238297,
-0.0576113797724247,
0.0074025653302669525,
0.0017722424818202853,
-0.08556929975748062,
-0.0013578023063018918,
-0.05503223091363907,
-0.05103278532624245,
0.039725158363580704,
-0.11120767891407013,
-0.014453868381679058,
0.048862479627132416,
-0.01609986647963524,
-0.1501091569662094,
0.06784731149673462,
0.0806848406791687,
-0.18246974050998688,
0.16650362312793732,
0.3054373264312744,
0.14630679786205292,
0.050898145884275436,
-0.056826647371053696,
-0.03993844613432884,
-0.07074804604053497,
-0.010058927349746227,
-0.06517490744590759,
0.2051897794008255,
-0.12545111775398254,
-0.025860246270895004,
0.06273581832647324,
0.046096645295619965,
0.005877739749848843,
-0.21709392964839935,
-0.011881912127137184,
-0.032071635127067566,
-0.07879411429166794,
-0.15016750991344452,
-0.03440835326910019,
0.003392346203327179,
0.11854957789182663,
0.034024305641651154,
-0.22931793332099915,
0.10609198361635208,
-0.04737434163689613,
-0.09509562700986862,
0.16003082692623138,
-0.10406653583049774,
-0.20107436180114746,
-0.16413523256778717,
-0.06666447967290878,
-0.0779612585902214,
0.04569277912378311,
-0.01195988804101944,
-0.0711042508482933,
-0.054942209273576736,
-0.00915894191712141,
-0.09555239230394363,
-0.16586478054523468,
-0.04950440302491188,
-0.04515838623046875,
0.09681696444749832,
-0.11735314875841141,
-0.07337035983800888,
-0.08632862567901611,
-0.03518729656934738,
0.034721218049526215,
0.09551652520895004,
-0.14903251826763153,
0.08019978553056717,
0.29893213510513306,
-0.029134975746273994,
0.07034244388341904,
-0.03767318278551102,
0.11524485051631927,
-0.06462199985980988,
0.015605909749865532,
0.10782861709594727,
0.06749647855758667,
0.030609358102083206,
0.2699412405490875,
0.10237937420606613,
-0.14466895163059235,
-0.053586822003126144,
-0.07584010064601898,
-0.10207401216030121,
-0.15878231823444366,
-0.10914254933595657,
-0.047386735677719116,
0.004906166810542345,
0.03935186192393303,
0.04487299174070358,
0.06150816008448601,
0.04912425950169563,
0.01224262360483408,
0.008874048478901386,
0.0661998763680458,
0.06143568083643913,
0.15972647070884705,
-0.033401161432266235,
0.09321679174900055,
-0.0627286359667778,
-0.01842958852648735,
0.08260802924633026,
0.12441718578338623,
0.18458519876003265,
0.20124678313732147,
0.11431397497653961,
0.11447865515947342,
0.050407566130161285,
0.16001783311367035,
0.040465738624334335,
0.1573100984096527,
-0.012489864602684975,
-0.013086059130728245,
-0.07840822637081146,
0.042192794382572174,
0.07250076532363892,
-0.036377135664224625,
-0.07762982696294785,
-0.05500749126076698,
-0.11353377997875214,
0.07224692404270172,
-0.0010787948267534375,
0.23510460555553436,
-0.25581225752830505,
0.022347936406731606,
0.0959784984588623,
0.11728867143392563,
-0.05414513126015663,
0.1044125109910965,
-0.01692565716803074,
-0.08377499878406525,
0.07198834419250488,
0.00019440639880485833,
0.10272329300642014,
-0.0847085490822792,
-0.01359903160482645,
-0.0002808326098602265,
-0.06344937533140182,
0.004871460143476725,
0.07742992788553238,
-0.1053648516535759,
0.28623726963996887,
0.04499635845422745,
-0.033125706017017365,
-0.06514368206262589,
-0.015923339873552322,
-0.005727894604206085,
0.2691815495491028,
0.22284995019435883,
0.04262508824467659,
-0.21850423514842987,
-0.21181504428386688,
-0.03627162054181099,
-0.022259915247559547,
0.143771693110466,
-0.03929362818598747,
0.020101236179471016,
0.031348682940006256,
0.012494059279561043,
-0.006826250813901424,
0.02048342488706112,
-0.08275233954191208,
-0.03255539387464523,
0.033041760325431824,
0.1218532994389534,
-0.05907479301095009,
-0.03297272324562073,
-0.05490074306726456,
-0.18863923847675323,
0.12267446517944336,
-0.11230623722076416,
-0.08918671309947968,
-0.08554235845804214,
-0.038757938891649246,
0.07661259919404984,
-0.0028275330550968647,
-0.00882504042237997,
-0.05145658925175667,
0.11679480969905853,
0.005665498785674572,
-0.09478648751974106,
0.1060485690832138,
-0.04567953944206238,
-0.05723625421524048,
-0.05143343284726143,
0.14832210540771484,
-0.04571009427309036,
-0.004954999778419733,
0.08285273611545563,
0.07894118875265121,
-0.0044600991532206535,
-0.12646590173244476,
0.09886225312948227,
-0.0099251763895154,
0.061704132705926895,
0.23459993302822113,
-0.10100562125444412,
-0.13306733965873718,
-0.035452552139759064,
0.049234360456466675,
0.07172832638025284,
0.2987561523914337,
-0.11066237837076187,
0.09777367115020752,
0.09469505399465561,
-0.04181015491485596,
-0.18932203948497772,
-0.02026222087442875,
-0.17316946387290955,
-0.00574108911678195,
-0.020051151514053345,
-0.06167563423514366,
0.12931960821151733,
0.06526342779397964,
-0.07319558411836624,
0.12036222219467163,
-0.23816736042499542,
-0.08278535306453705,
0.16413938999176025,
0.04290027171373367,
0.14626993238925934,
-0.06810710579156876,
-0.10985154658555984,
-0.07606552541255951,
-0.2023620456457138,
0.15823830664157867,
0.008313131518661976,
0.08137176930904388,
-0.04466485604643822,
-0.01031293161213398,
-0.005764672998338938,
0.0016942763468250632,
0.24252279102802277,
0.13000307977199554,
0.11575871706008911,
0.0410989448428154,
-0.1560717523097992,
0.2534436583518982,
0.011263403110206127,
0.020132796838879585,
0.11813158541917801,
0.00565307354554534,
-0.08344988524913788,
-0.014656865037977695,
0.009873991832137108,
0.016077779233455658,
-0.08585617691278458,
-0.0618777871131897,
-0.09066876024007797,
0.011224602349102497,
-0.07304469496011734,
-0.07567139714956284,
0.23793651163578033,
-0.057095106691122055,
0.06352704018354416,
0.1880899965763092,
0.005279981531202793,
-0.10675178468227386,
-0.049573712050914764,
-0.060629893094301224,
-0.08667755872011185,
0.05587523430585861,
-0.17109231650829315,
0.05056922882795334,
0.11474195867776871,
0.03820998594164848,
0.14317600429058075,
0.11862756311893463,
-0.02345028892159462,
-0.04831225052475929,
0.11662789434194565,
-0.08561655133962631,
-0.14909638464450836,
0.014233003370463848,
-0.2027336061000824,
0.027250483632087708,
0.07886254042387009,
0.07179734855890274,
-0.0018246157560497522,
-0.016222795471549034,
0.024982329457998276,
0.035005856305360794,
-0.06894157826900482,
0.07725698500871658,
0.024850904941558838,
0.05899198725819588,
-0.14741376042366028,
0.15153467655181885,
0.08140043914318085,
0.01461124885827303,
-0.09131469577550888,
-0.0851060003042221,
-0.15023864805698395,
-0.05527853965759277,
-0.03901746869087219,
0.10751599073410034,
-0.1442699134349823,
-0.13946859538555145,
-0.10302101820707321,
-0.18507981300354004,
0.019559036940336227,
0.0825081393122673,
0.14364397525787354,
0.11854439228773117,
0.022650552913546562,
-0.10586965084075928,
0.020916512235999107,
0.05989677459001541,
-0.11545216292142868,
0.025937216356396675,
-0.21120066940784454,
0.05260515213012695,
-0.021542245522141457,
0.09287124127149582,
-0.0909704640507698,
-0.050593018531799316,
-0.10576430708169937,
0.01215311698615551,
-0.06243650242686272,
0.0779610201716423,
-0.06788360327482224,
0.005036062560975552,
0.018205055966973305,
0.015125119127333164,
-0.05468454957008362,
0.002838734071701765,
-0.0881517305970192,
0.05076843872666359,
0.02580079436302185,
0.17738373577594757,
-0.09735121577978134,
-0.020686699077486992,
0.05904453992843628,
-0.035131461918354034,
0.06808776408433914,
0.0172868724912405,
0.0209550429135561,
0.058924052864313126,
-0.15162110328674316,
0.031100193038582802,
0.09912572801113129,
0.03231531009078026,
0.0649973601102829,
-0.12496054172515869,
-0.0009505844791419804,
0.05332842841744423,
-0.09204840660095215,
0.08818327635526657,
-0.08908951282501221,
-0.10885732620954514,
-0.148936465382576,
-0.11772937327623367,
-0.11350806802511215,
-0.031287409365177155,
0.03786865621805191,
0.21015797555446625,
0.03687160462141037,
0.0513434074819088,
0.03282148018479347,
-0.004413660150021315,
-0.04382277652621269,
-0.029208706691861153,
-0.06632871180772781,
-0.0825536772608757,
-0.06193843483924866,
-0.01185196079313755,
-0.008238728158175945,
-0.028696352615952492,
0.35530662536621094,
0.04186559468507767,
-0.019017742946743965,
0.05373983830213547,
0.17581380903720856,
-0.01903216727077961,
0.03825109824538231,
0.2920352816581726,
0.06690358370542526,
-0.034791313111782074,
0.0385039821267128,
0.03505776450037956,
-0.005083543714135885,
0.026611438021063805,
0.13967959582805634,
0.09902851283550262,
-0.08417165279388428,
0.05233171209692955,
0.08394066244363785,
-0.03554443269968033,
-0.024233177304267883,
0.11782915145158768,
0.05672597512602806,
0.04202960804104805,
0.03877565264701843,
-0.07952756434679031,
0.1245117262005806,
-0.17514373362064362,
0.08291848003864288,
-0.001869208994321525,
-0.09116273373365402,
-0.18642351031303406,
-0.11220096051692963,
-0.08741098642349243,
-0.07877695560455322,
0.025062382221221924,
-0.10904214531183243,
-0.04557626694440842,
0.17665907740592957,
0.019587984308600426,
-0.0223850104957819,
-0.018440401181578636,
-0.24619726836681366,
0.011170857585966587,
0.10871850699186325,
0.015622579492628574,
-0.009099767543375492,
-0.07968004047870636,
-0.005052314605563879,
0.024557968601584435,
-0.10349047183990479,
-0.0413210503757,
-0.02892024628818035,
0.06758610904216766,
-0.03424226865172386,
-0.15372827649116516,
-0.057961177080869675,
-0.038767993450164795,
-0.004906147718429565,
-0.01300619263201952,
-0.023529192432761192,
0.025360742583870888,
-0.02226540446281433,
0.04073664918541908,
0.18833975493907928,
-0.048294443637132645,
0.008473499678075314,
-0.06769437342882156,
0.23921804130077362,
-0.055905360728502274,
0.07767180353403091,
0.07663410156965256,
-0.07191310822963715,
0.0003459442814346403,
0.0792202353477478,
0.18814904987812042,
-0.012818766757845879,
-0.008260911330580711,
0.015916144475340843,
-0.003277712734416127,
0.04123755916953087,
0.024818619713187218,
-0.06559041887521744,
0.12435705959796906,
-0.05413991957902908,
0.05450914055109024,
-0.10693985968828201,
0.006451353430747986,
-0.044598665088415146,
-0.04943462461233139,
0.11579426378011703,
-0.08825626224279404,
-0.13360080122947693,
0.18862922489643097,
-0.04589084908366203,
0.02840382792055607,
0.2960452437400818,
-0.07968685775995255,
-0.07816416025161743,
-0.0003751165058929473,
-0.01642579771578312,
-0.023035269230604172,
0.034523699432611465,
-0.15269382297992706,
0.0059292674995958805,
0.05508046597242355,
0.01407061330974102,
-0.18646740913391113,
-0.05368676781654358,
0.02799980342388153,
0.001904219388961792,
0.09223173558712006,
-0.0017765635857358575,
0.21953818202018738,
0.09995222836732864,
-0.0035242121666669846,
-0.08077842742204666,
0.10203532874584198,
0.0035727459471672773,
0.0376279316842556,
-0.001758163794875145,
0.02231457643210888,
-0.016706448048353195,
-0.04367652162909508,
0.09478902816772461,
-0.07720743119716644,
0.013284144923090935,
0.016123885288834572,
-0.1176999881863594,
-0.07376014441251755,
0.01919964887201786,
-0.10536607354879379,
0.08765166252851486,
0.05095658078789711,
-0.02849440462887287,
-0.03391394391655922,
0.02172129787504673,
0.07538878917694092,
0.09047824889421463,
-0.09260491281747818,
0.036674220114946365,
0.014517667703330517,
0.0216269101947546,
0.10169968754053116,
-0.01017606258392334,
-0.16540388762950897,
-0.02997240051627159,
-0.051250360906124115,
0.02315935492515564,
-0.07713884115219116,
0.10741882771253586,
0.12350884824991226,
0.04750902205705643,
-0.05969291180372238,
-0.19981643557548523,
0.03945329040288925,
0.0740835890173912,
-0.0726441964507103,
-0.08894511312246323
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# roberta-base-bne-finetuned-amazon_reviews_multi
This model is a fine-tuned version of [BSC-TeMU/roberta-base-bne](https://huggingface.co/BSC-TeMU/roberta-base-bne) on the amazon_reviews_multi dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2303
- Accuracy: 0.9325
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.1942 | 1.0 | 1250 | 0.1751 | 0.932 |
| 0.0935 | 2.0 | 2500 | 0.2303 | 0.9325 |
### Framework versions
- Transformers 4.10.2
- Pytorch 1.9.0+cu102
- Datasets 1.12.1
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["amazon_reviews_multi"], "metrics": ["accuracy"], "model-index": [{"name": "roberta-base-bne-finetuned-amazon_reviews_multi", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "amazon_reviews_multi", "type": "amazon_reviews_multi", "args": "es"}, "metrics": [{"type": "accuracy", "value": 0.9325, "name": "Accuracy"}]}]}]}
|
text-classification
|
csalamea/roberta-base-bne-finetuned-amazon_reviews_multi
|
[
"transformers",
"pytorch",
"tensorboard",
"roberta",
"text-classification",
"generated_from_trainer",
"dataset:amazon_reviews_multi",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
roberta-base-bne-finetuned-amazon\_reviews\_multi
=================================================
This model is a fine-tuned version of BSC-TeMU/roberta-base-bne on the amazon\_reviews\_multi dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2303
* Accuracy: 0.9325
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.10.2
* Pytorch 1.9.0+cu102
* Datasets 1.12.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
71,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.10.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.12.1\n* Tokenizers 0.10.3"
] |
[
-0.08748505264520645,
0.10135051608085632,
-0.002402752870693803,
0.12601561844348907,
0.1706485152244568,
0.03687635064125061,
0.1622617095708847,
0.11569225788116455,
-0.08599758893251419,
-0.005865089595317841,
0.11542874574661255,
0.1548892855644226,
0.02126530557870865,
0.12217552959918976,
-0.05617023631930351,
-0.24667197465896606,
-0.0027746891137212515,
0.031354084610939026,
-0.04765567183494568,
0.14976318180561066,
0.10777784883975983,
-0.12571990489959717,
0.10260860621929169,
0.006307870149612427,
-0.16979162395000458,
-0.014389348216354847,
0.024533210322260857,
-0.0696016252040863,
0.13328847289085388,
0.02715180069208145,
0.12315018475055695,
-0.001376794883981347,
0.07599072903394699,
-0.19718699157238007,
0.018568968400359154,
0.04323352873325348,
0.007994817569851875,
0.09346025437116623,
0.02965611405670643,
-0.017832037061452866,
0.11385732144117355,
-0.05450711026787758,
0.07709876447916031,
0.017052244395017624,
-0.12038940191268921,
-0.23559437692165375,
-0.09075293689966202,
0.04976227506995201,
0.05859314277768135,
0.09819751977920532,
-0.005986036732792854,
0.15179461240768433,
-0.08127760142087936,
0.08983530849218369,
0.23404285311698914,
-0.28485214710235596,
-0.0688856989145279,
0.021996930241584778,
0.030939679592847824,
0.08619936555624008,
-0.09751075506210327,
-0.033980995416641235,
0.04781979322433472,
0.05363927781581879,
0.12734320759773254,
-0.0339665561914444,
-0.09865715354681015,
0.014300784096121788,
-0.14419035613536835,
-0.03703422099351883,
0.2110384702682495,
0.05224103480577469,
-0.04936090111732483,
-0.033593520522117615,
-0.033585041761398315,
-0.13613587617874146,
-0.043771665543317795,
0.00799341220408678,
0.06177709624171257,
-0.047818057239055634,
-0.08494832366704941,
-0.01645578444004059,
-0.1082940623164177,
-0.04342750832438469,
-0.06920549273490906,
0.12287094444036484,
0.039220601320266724,
0.02422584593296051,
-0.04213975742459297,
0.0993783175945282,
0.016896851360797882,
-0.11286630481481552,
0.016236115247011185,
0.007824072614312172,
-0.005469383206218481,
-0.04410327225923538,
-0.055700261145830154,
-0.07245267927646637,
0.01179128885269165,
0.1487048864364624,
-0.03140841796994209,
0.029435962438583374,
0.03786361217498779,
0.058199744671583176,
-0.08600983023643494,
0.2011789083480835,
-0.03808872774243355,
-0.019710825756192207,
0.0025984770618379116,
0.06378598511219025,
0.0366097018122673,
-0.01270147506147623,
-0.1338830590248108,
0.021989855915308,
0.08388182520866394,
0.012604576535522938,
-0.05425940081477165,
0.05256333574652672,
-0.07947677373886108,
-0.051174573600292206,
-0.00332809635438025,
-0.07323624938726425,
0.01931113190948963,
-0.012507437728345394,
-0.06862229853868484,
-0.023805459961295128,
0.019980812445282936,
0.038462378084659576,
-0.004955208860337734,
0.12952286005020142,
-0.08990225940942764,
0.027578171342611313,
-0.08297199010848999,
-0.09065180271863937,
0.030825134366750717,
-0.08147244155406952,
0.04217829927802086,
-0.10889053344726562,
-0.18562471866607666,
-0.022130921483039856,
0.06331305205821991,
-0.01936176046729088,
-0.08627636730670929,
-0.031135743483901024,
-0.06069953739643097,
0.007166009396314621,
-0.012592478655278683,
0.14891761541366577,
-0.07661006599664688,
0.10016986727714539,
0.018237467855215073,
0.05291328206658363,
-0.04553815349936485,
0.04563181847333908,
-0.10529284179210663,
-0.0009554275893606246,
-0.14402805268764496,
0.03092854842543602,
-0.038881681859493256,
0.07289519160985947,
-0.08744402229785919,
-0.10114515572786331,
0.01812109909951687,
0.0007858672179281712,
0.043990358710289,
0.08458218723535538,
-0.16604909300804138,
-0.0736570879817009,
0.1392926573753357,
-0.05869176238775253,
-0.13591888546943665,
0.12868312001228333,
-0.07782956212759018,
0.054878707975149155,
0.0784146636724472,
0.16067740321159363,
0.06189722940325737,
-0.06682360172271729,
0.03472161293029785,
-0.00491720624268055,
0.042379315942525864,
-0.06637439131736755,
0.09732348471879959,
0.008913968689739704,
-0.017364542931318283,
0.027644045650959015,
-0.05057794228196144,
0.04496327042579651,
-0.08496037125587463,
-0.09797971695661545,
-0.03988826647400856,
-0.10726158320903778,
0.06420964747667313,
0.06216416135430336,
0.06328140944242477,
-0.10556590557098389,
-0.07531622052192688,
0.05133743956685066,
0.09141959249973297,
-0.03835410252213478,
0.019969282671809196,
-0.054281849414110184,
0.07662713527679443,
-0.04022379219532013,
-0.0186300165951252,
-0.17294596135616302,
-0.020581988617777824,
0.014201127924025059,
-0.021982483565807343,
0.04035462439060211,
0.01553034782409668,
0.052458200603723526,
0.034605711698532104,
-0.06715890020132065,
-0.0013398109003901482,
-0.05249806120991707,
-0.004969918634742498,
-0.11456017941236496,
-0.21327568590641022,
-0.022570785135030746,
-0.015485681593418121,
0.15145157277584076,
-0.20356561243534088,
0.03070143051445484,
-0.05047479271888733,
0.07912503182888031,
0.038957469165325165,
-0.008191968314349651,
-0.023654483258724213,
0.06938077509403229,
-0.029641080647706985,
-0.041676975786685944,
0.07319847494363785,
0.015923120081424713,
-0.11104981601238251,
-0.004017536528408527,
-0.07411255687475204,
0.18232916295528412,
0.12914419174194336,
-0.09189614653587341,
-0.07028195261955261,
0.02698177844285965,
-0.04606953263282776,
-0.033419687300920486,
-0.07677464187145233,
0.04015912488102913,
0.1786910742521286,
0.0123989786952734,
0.1393493115901947,
-0.09585434198379517,
-0.046626247465610504,
0.026534689590334892,
-0.04766058921813965,
0.029939081519842148,
0.1388590931892395,
0.11947977542877197,
-0.08978938311338425,
0.1408824771642685,
0.1801275759935379,
-0.08513115346431732,
0.13522493839263916,
-0.04518233612179756,
-0.057722315192222595,
-0.027640100568532944,
-0.046595100313425064,
-0.014140930958092213,
0.10802013427019119,
-0.12211117148399353,
0.008783064782619476,
0.038905952125787735,
0.01179005578160286,
0.007891636341810226,
-0.220276340842247,
-0.04456881061196327,
0.03768514469265938,
-0.03834724798798561,
-0.016397174447774887,
-0.0017061536200344563,
0.01423916406929493,
0.1071462482213974,
0.013389743864536285,
-0.07625076174736023,
0.04752439633011818,
0.007178463041782379,
-0.08612769842147827,
0.21150490641593933,
-0.07518668472766876,
-0.1776520013809204,
-0.13706344366073608,
-0.0545949712395668,
-0.05295511707663536,
-0.003149912226945162,
0.0614364892244339,
-0.07413402199745178,
-0.03180314600467682,
-0.07279524952173233,
-0.0024537749122828245,
-0.0038665372412651777,
0.008661623112857342,
-0.011396964080631733,
0.014962404035031796,
0.05599657818675041,
-0.09647130221128464,
-0.012222180142998695,
-0.04970064014196396,
-0.02174299955368042,
0.039092689752578735,
0.040774259716272354,
0.10311493277549744,
0.15333402156829834,
-0.014502475969493389,
-0.008167930878698826,
-0.02158825471997261,
0.23083464801311493,
-0.08114238828420639,
-0.043072085827589035,
0.14223170280456543,
-0.01715271733701229,
0.0431506372988224,
0.13593509793281555,
0.07226701825857162,
-0.08730296790599823,
0.015815844759345055,
0.02846592850983143,
-0.03927796334028244,
-0.268171101808548,
-0.033646851778030396,
-0.0507233627140522,
-0.007956546731293201,
0.0783345103263855,
0.019632715731859207,
0.003466376569122076,
0.06746618449687958,
0.036469560116529465,
0.07697193324565887,
-0.024515310302376747,
0.0725059062242508,
0.12220457196235657,
0.047124069184064865,
0.13204236328601837,
-0.051729895174503326,
-0.05593999847769737,
0.06652961671352386,
-0.005725420080125332,
0.2295539677143097,
0.018175311386585236,
0.14277383685112,
0.07377737015485764,
0.1386025846004486,
0.008008481003344059,
0.04504062235355377,
0.017918335273861885,
-0.019810272380709648,
-0.03349725902080536,
-0.024137189611792564,
-0.03777434304356575,
0.026318218559026718,
-0.04787589609622955,
0.05510532483458519,
-0.1216469258069992,
-0.011227552779018879,
0.05953656882047653,
0.26264089345932007,
0.02338678576052189,
-0.31383731961250305,
-0.10186455398797989,
0.016122763976454735,
-0.05420726165175438,
-0.0018863070290535688,
0.028928078711032867,
0.06222778186202049,
-0.13198141753673553,
0.041389141231775284,
-0.07491769641637802,
0.10316328704357147,
-0.08767863363027573,
0.04334351792931557,
0.0641695037484169,
0.07428254187107086,
0.0013433737913146615,
0.08148038387298584,
-0.29981619119644165,
0.27460289001464844,
-0.008198552764952183,
0.048941873013973236,
-0.06314488500356674,
-0.027548687532544136,
0.035845763981342316,
0.054587073624134064,
0.05386405438184738,
0.0038666280452162027,
-0.0410676933825016,
-0.15771129727363586,
-0.040473781526088715,
0.02706184983253479,
0.06241348013281822,
-0.02723764255642891,
0.08145689219236374,
-0.03775765746831894,
0.004442242439836264,
0.05066261440515518,
-0.000723798293620348,
-0.05325597897171974,
-0.09176760166883469,
-0.0010839176829904318,
0.02408144436776638,
-0.048524193465709686,
-0.06867502629756927,
-0.1267763376235962,
-0.07732057571411133,
0.11763031780719757,
-0.018511952832341194,
-0.04903564229607582,
-0.09150055795907974,
0.06172432377934456,
0.08089186996221542,
-0.08031509816646576,
0.03645249828696251,
-0.002867340575903654,
0.09249304234981537,
0.028154658153653145,
-0.04747135564684868,
0.09116265922784805,
-0.0564410425722599,
-0.19131174683570862,
-0.06531649827957153,
0.1091921254992485,
0.02420763298869133,
0.06942029297351837,
-0.02535104751586914,
0.010509409941732883,
-0.06197625771164894,
-0.08357580006122589,
0.01257217675447464,
0.00821640994399786,
0.07106112688779831,
0.04573948308825493,
-0.036858171224594116,
0.012068760581314564,
-0.08304876834154129,
-0.06253399699926376,
0.18849623203277588,
0.2214498221874237,
-0.08738681674003601,
0.04031470790505409,
0.036602821201086044,
-0.07414983212947845,
-0.15455880761146545,
0.011297655291855335,
0.05963510647416115,
0.0006671959417872131,
0.06797772645950317,
-0.14696437120437622,
0.11902517825365067,
0.09333305060863495,
-0.02068314328789711,
0.11209964007139206,
-0.3217496871948242,
-0.13395251333713531,
0.10134933888912201,
0.1372978538274765,
0.13894343376159668,
-0.13172198832035065,
-0.02082870528101921,
-0.04572247341275215,
-0.146324023604393,
0.1378229707479477,
-0.10358641296625137,
0.12806963920593262,
-0.03092869371175766,
0.1106255054473877,
0.0037294209469109774,
-0.047299932688474655,
0.12950772047042847,
0.019552117213606834,
0.09933143854141235,
-0.057583749294281006,
-0.028160307556390762,
0.03231542930006981,
-0.04104400798678398,
0.016753770411014557,
-0.09342203289270401,
0.0307865459471941,
-0.09329255670309067,
-0.030915401875972748,
-0.07106336951255798,
0.023760778829455376,
-0.040977418422698975,
-0.048031218349933624,
-0.04165596887469292,
0.03756199777126312,
0.034384556114673615,
-0.012077606283128262,
0.1650484949350357,
0.024534137919545174,
0.13603870570659637,
0.07593537122011185,
0.09001367539167404,
-0.062306176871061325,
-0.09832610934972763,
-0.04641509801149368,
-0.030273811891674995,
0.048144929111003876,
-0.17123320698738098,
0.02887239120900631,
0.13160167634487152,
0.011140162125229836,
0.15611597895622253,
0.06286188215017319,
-0.03973785787820816,
0.007328100968152285,
0.060210712254047394,
-0.1473669558763504,
-0.09190437942743301,
-0.007199518382549286,
-0.055258866399526596,
-0.13529729843139648,
0.03206604719161987,
0.12203475832939148,
-0.06728495657444,
-0.03177840635180473,
-0.006757003255188465,
0.017898136749863625,
-0.05075324699282646,
0.17833282053470612,
0.07896671444177628,
0.047494325786828995,
-0.10676582902669907,
0.1078370213508606,
0.07093336433172226,
-0.0688825398683548,
0.004336613230407238,
0.05840630084276199,
-0.09213374555110931,
-0.056954748928546906,
0.0604483038187027,
0.1810775101184845,
-0.08093660324811935,
-0.05632931366562843,
-0.14352649450302124,
-0.12352213263511658,
0.08281289786100388,
0.1308985948562622,
0.11646679043769836,
0.011451567523181438,
-0.03895536810159683,
-0.02378356270492077,
-0.08104820549488068,
0.10755844414234161,
0.07420925796031952,
0.06614638864994049,
-0.1541292667388916,
0.07461540400981903,
0.02957092598080635,
0.05183961242437363,
-0.01501353271305561,
0.037376768887043,
-0.10405530780553818,
0.014517159201204777,
-0.14131805300712585,
0.0029443365056067705,
-0.01665409840643406,
0.023352298885583878,
-0.0036167106591165066,
-0.06881476938724518,
-0.07280328869819641,
0.013769056648015976,
-0.12243233621120453,
-0.023349132388830185,
0.037886541336774826,
0.07895298302173615,
-0.09635411202907562,
-0.03891822323203087,
0.04241982474923134,
-0.04902099817991257,
0.07320401817560196,
0.04164178669452667,
0.014727015048265457,
0.06638604402542114,
-0.12203842401504517,
0.031979624181985855,
0.03355097770690918,
0.020873717963695526,
0.047238755971193314,
-0.131159245967865,
-0.0029366749804466963,
0.0008752434514462948,
0.06524120271205902,
0.023719217628240585,
0.07833414524793625,
-0.15566109120845795,
-0.011630808934569359,
0.008436531759798527,
-0.08374901115894318,
-0.04931286349892616,
0.013993371278047562,
0.0644172802567482,
0.03146012872457504,
0.2301100641489029,
-0.06705774366855621,
0.03818197548389435,
-0.19638380408287048,
0.011850639246404171,
-0.02603587694466114,
-0.12201353162527084,
-0.14749230444431305,
-0.07459408044815063,
0.04430370405316353,
-0.06755319982767105,
0.16799087822437286,
0.03966196998953819,
0.07443663477897644,
0.029320290312170982,
0.019833926111459732,
-0.0012724585831165314,
0.01750226877629757,
0.15605953335762024,
0.017472129315137863,
-0.0385797917842865,
0.0553094781935215,
0.029253395274281502,
0.10055885463953018,
0.09190146625041962,
0.1844479888677597,
0.17639435827732086,
0.029132237657904625,
0.08518511056900024,
0.03540175408124924,
-0.019055966287851334,
-0.1243986189365387,
0.04176418483257294,
-0.01574995554983616,
0.11440835893154144,
-0.013763918541371822,
0.21950466930866241,
0.07746719568967819,
-0.16513484716415405,
0.046616800129413605,
-0.06316568702459335,
-0.0753798633813858,
-0.1084527000784874,
-0.0666743814945221,
-0.09666218608617783,
-0.150067538022995,
0.0005522034480236471,
-0.11988969147205353,
0.0009473717072978616,
0.11587756127119064,
0.0018456995021551847,
-0.043480828404426575,
0.09077279269695282,
0.009991518221795559,
0.0034860013984143734,
0.08070904016494751,
0.009907604195177555,
-0.04668661952018738,
-0.08920586854219437,
-0.06421113014221191,
-0.021299738436937332,
-0.012637236155569553,
0.021962594240903854,
-0.05690223351120949,
-0.05890936031937599,
0.0226567555218935,
-0.028140811249613762,
-0.10631848871707916,
0.021962102502584457,
0.009738780558109283,
0.07327093183994293,
0.032759103924036026,
0.01137351430952549,
0.026509786024689674,
0.007413798477500677,
0.26563242077827454,
-0.06026383861899376,
-0.07014074176549911,
-0.12706468999385834,
0.22582225501537323,
0.02476755902171135,
-0.04121285676956177,
0.03648244962096214,
-0.06837393343448639,
0.0009176467428915203,
0.24895620346069336,
0.2247450053691864,
-0.09457942843437195,
-0.01954338513314724,
0.011463322676718235,
-0.007764757145196199,
-0.013498657383024693,
0.10854029655456543,
0.10903426259756088,
0.006476501002907753,
-0.08121499419212341,
-0.014552807435393333,
-0.05886773392558098,
-0.0011637939605861902,
-0.016025392338633537,
0.06969863176345825,
0.03950197994709015,
-0.0072053768672049046,
-0.04185018688440323,
0.07962869107723236,
-0.0905042514204979,
-0.11997628957033157,
0.032596368342638016,
-0.20977969467639923,
-0.18324455618858337,
-0.02352200448513031,
0.09226783365011215,
0.017664752900600433,
0.058241747319698334,
-0.01702686958014965,
-0.012324018403887749,
0.07180946320295334,
-0.016307998448610306,
-0.10270724445581436,
-0.10217396169900894,
0.1081206202507019,
-0.09479250013828278,
0.20147524774074554,
-0.04637618735432625,
0.05569562315940857,
0.12050214409828186,
0.06346482783555984,
-0.07389548420906067,
0.05974552035331726,
0.044635169208049774,
-0.024630477651953697,
0.04403224587440491,
0.0851675420999527,
-0.025177715346217155,
0.06843596696853638,
0.06200204789638519,
-0.10778521746397018,
0.0037345902528613806,
-0.06963280588388443,
-0.04284678027033806,
-0.05494605004787445,
-0.01213314663618803,
-0.07727371901273727,
0.1269443780183792,
0.22313793003559113,
-0.05046144500374794,
-0.014237524941563606,
-0.06106938049197197,
0.02333267591893673,
0.06654650717973709,
0.01524131279438734,
-0.05319321155548096,
-0.2105739861726761,
0.00934837106615305,
0.06471182405948639,
-0.008181453682482243,
-0.26207032799720764,
-0.07192134112119675,
-0.0029018190689384937,
-0.06448803097009659,
-0.06932234764099121,
0.08557683229446411,
0.07956942170858383,
0.03840881958603859,
-0.06028938665986061,
-0.04749587923288345,
-0.07239669561386108,
0.15115003287792206,
-0.15023668110370636,
-0.09064730256795883
] |
null | null |
transformers
|
## BERT-base uncased model fine-tuned on SQuAD v1
This model was fine-tuned from the HuggingFace [BERT](https://www.aclweb.org/anthology/N19-1423/) base uncased checkpoint on [SQuAD1.1](https://rajpurkar.github.io/SQuAD-explorer).
This model is case-insensitive: it does not make a difference between english and English.
## Details
| Dataset | Split | # samples |
| -------- | ----- | --------- |
| SQuAD1.1 | train | 90.6K |
| SQuAD1.1 | eval | 11.1k |
### Fine-tuning
- Python: `3.7.5`
- Machine specs:
`CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz`
`Memory: 32 GiB`
`GPUs: 2 GeForce GTX 1070, each with 8GiB memory`
`GPU driver: 418.87.01, CUDA: 10.1`
- script:
```shell
# after install https://github.com/huggingface/transformers
cd examples/question-answering
mkdir -p data
wget -O data/train-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v1.1.json
wget -O data/dev-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v1.1.json
python run_squad.py \
--model_type bert \
--model_name_or_path bert-base-uncased \
--do_train \
--do_eval \
--do_lower_case \
--train_file train-v1.1.json \
--predict_file dev-v1.1.json \
--per_gpu_train_batch_size 12 \
--per_gpu_eval_batch_size=16 \
--learning_rate 3e-5 \
--num_train_epochs 2.0 \
--max_seq_length 320 \
--doc_stride 128 \
--data_dir data \
--output_dir data/bert-base-uncased-squad-v1 2>&1 | tee train-energy-bert-base-squad-v1.log
```
It took about 2 hours to finish.
### Results
**Model size**: `418M`
| Metric | # Value | # Original ([Table 2](https://www.aclweb.org/anthology/N19-1423.pdf))|
| ------ | --------- | --------- |
| **EM** | **80.9** | **80.8** |
| **F1** | **88.2** | **88.5** |
Note that the above results didn't involve any hyperparameter search.
## Example Usage
```python
from transformers import pipeline
qa_pipeline = pipeline(
"question-answering",
model="csarron/bert-base-uncased-squad-v1",
tokenizer="csarron/bert-base-uncased-squad-v1"
)
predictions = qa_pipeline({
'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.",
'question': "What day was the game played on?"
})
print(predictions)
# output:
# {'score': 0.8730505704879761, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'}
```
> Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp)
> Made with ❤️ in New York.
|
{"language": "en", "license": "mit", "tags": ["question-answering", "bert", "bert-base"], "datasets": ["squad"], "metrics": ["squad"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}], "model-index": [{"name": "csarron/bert-base-uncased-squad-v1", "results": [{"task": {"type": "question-answering", "name": "Question Answering"}, "dataset": {"name": "squad", "type": "squad", "config": "plain_text", "split": "validation"}, "metrics": [{"type": "exact_match", "value": 80.9104, "name": "Exact Match", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiNDJlOWQ0OTE0ZjRhMTQwNDY5MjVhYmZiN2RmYzY0OWJiOWUyNjcyMWU5N2I3YmU0OThjZTVjNTc2MjM2Yzg5NiIsInZlcnNpb24iOjF9.cuJ34B-ngUur5wKGhfhVP8FM6NX4IFrIJEdXypbLQJw1i8M5Bb2EeIs-0M5n35YIx2PfqSQcnVj_jP8vLUk4Dg"}, {"type": "f1", "value": 88.2302, "name": "F1", "verified": true, "verifyToken": "eyJhbGciOiJFZERTQSIsInR5cCI6IkpXVCJ9.eyJoYXNoIjoiYmE4NzFmNDA3MDRiODk3ZDg5NWYyNjczOGE5YjdkZWQ0ZmEzNWU5YjFjMzc1ODA2OGRjYzU0Y2M5MmU0NGNhYSIsInZlcnNpb24iOjF9.phmkVWF3I-rl2xrHW0EW9OQqzfuefoqNjWplOpFdzJuW8d2C4sJ8snW0Ikw9kQqZaBCdwdkmsf5VTgOupHb8Dw"}]}]}]}
|
question-answering
|
csarron/bert-base-uncased-squad-v1
|
[
"transformers",
"pytorch",
"jax",
"safetensors",
"bert",
"question-answering",
"bert-base",
"en",
"dataset:squad",
"license:mit",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #pytorch #jax #safetensors #bert #question-answering #bert-base #en #dataset-squad #license-mit #model-index #endpoints_compatible #has_space #region-us
|
BERT-base uncased model fine-tuned on SQuAD v1
----------------------------------------------
This model was fine-tuned from the HuggingFace BERT base uncased checkpoint on SQuAD1.1.
This model is case-insensitive: it does not make a difference between english and English.
Details
-------
Dataset: SQuAD1.1, Split: train, # samples: 90.6K
Dataset: SQuAD1.1, Split: eval, # samples: 11.1k
### Fine-tuning
* Python: '3.7.5'
* Machine specs:
'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'
'Memory: 32 GiB'
'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'
'GPU driver: 418.87.01, CUDA: 10.1'
* script:
It took about 2 hours to finish.
### Results
Model size: '418M'
Metric: EM, # Value: 80.9, # Original (Table 2): 80.8
Metric: F1, # Value: 88.2, # Original (Table 2): 88.5
Note that the above results didn't involve any hyperparameter search.
Example Usage
-------------
>
> Created by Qingqing Cao | GitHub | Twitter
>
>
>
>
> Made with ️ in New York.
>
>
>
|
[
"# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.",
"### Results\n\n\nModel size: '418M'\n\n\nMetric: EM, # Value: 80.9, # Original (Table 2): 80.8\nMetric: F1, # Value: 88.2, # Original (Table 2): 88.5\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
"TAGS\n#transformers #pytorch #jax #safetensors #bert #question-answering #bert-base #en #dataset-squad #license-mit #model-index #endpoints_compatible #has_space #region-us \n",
"# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.",
"### Results\n\n\nModel size: '418M'\n\n\nMetric: EM, # Value: 80.9, # Original (Table 2): 80.8\nMetric: F1, # Value: 88.2, # Original (Table 2): 88.5\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
62,
27,
95,
107
] |
[
"passage: TAGS\n#transformers #pytorch #jax #safetensors #bert #question-answering #bert-base #en #dataset-squad #license-mit #model-index #endpoints_compatible #has_space #region-us \n# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.### Results\n\n\nModel size: '418M'\n\n\nMetric: EM, # Value: 80.9, # Original (Table 2): 80.8\nMetric: F1, # Value: 88.2, # Original (Table 2): 88.5\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
-0.12932617962360382,
0.09527844190597534,
-0.0013612484326586127,
0.06507304310798645,
0.10248512774705887,
0.04070766642689705,
0.06183772161602974,
0.12735673785209656,
0.09208894520998001,
0.16582350432872772,
0.10345950722694397,
0.029340939596295357,
0.08973577618598938,
0.1347506195306778,
-0.02124198153614998,
-0.08617104589939117,
0.04407426714897156,
-0.002348623936995864,
-0.010493583045899868,
0.1197686716914177,
0.07891575247049332,
-0.0649980753660202,
0.11728579550981522,
-0.008787203580141068,
-0.0997900515794754,
-0.021268481388688087,
0.029858607798814774,
-0.033813219517469406,
0.09710841625928879,
0.03504723682999611,
0.02357751876115799,
0.059879980981349945,
0.03427023068070412,
-0.10387212038040161,
0.0220886692404747,
0.12788905203342438,
-0.04108845070004463,
0.04301576688885689,
0.03201032429933548,
0.022163953632116318,
0.06649830937385559,
-0.01784614659845829,
0.01861419901251793,
0.05478544533252716,
-0.13121047616004944,
-0.17107513546943665,
-0.16496899724006653,
0.03450269252061844,
0.09538573771715164,
0.07215755432844162,
-0.019059259444475174,
0.1349763572216034,
-0.03178787603974342,
0.06628777831792831,
0.2513565421104431,
-0.25441160798072815,
-0.04934193566441536,
0.06875239312648773,
0.05772488936781883,
0.07069052010774612,
-0.008942680433392525,
0.0030745866242796183,
0.037165652960538864,
0.04026574268937111,
0.018368229269981384,
-0.011301624588668346,
0.037891414016485214,
0.013231619261205196,
-0.1438325196504593,
-0.055574920028448105,
0.09523468464612961,
0.044664040207862854,
-0.05262686312198639,
-0.03419601172208786,
-0.1115698590874672,
-0.2100444734096527,
-0.03550101816654205,
0.015954680740833282,
-0.02222253754734993,
0.016859620809555054,
-0.03507425636053085,
-0.0001171863914350979,
-0.06001238152384758,
-0.09210889041423798,
-0.04184725135564804,
0.0303509384393692,
0.051970552653074265,
0.08323269337415695,
0.009682801552116871,
0.06922407448291779,
-0.17697985470294952,
-0.09222684800624847,
-0.04422200098633766,
0.002766762161627412,
-0.07759823650121689,
0.021901553496718407,
0.018398014828562737,
0.03636583313345909,
0.04918463155627251,
0.1884605437517166,
-0.07648123055696487,
0.03903239220380783,
0.08343237638473511,
-0.026919430121779442,
-0.03417009860277176,
0.13793939352035522,
-0.14397084712982178,
-0.08262433856725693,
0.03556523472070694,
0.05724022909998894,
-0.019722728058695793,
-0.04963428154587746,
-0.011491496115922928,
0.012771762907505035,
0.09927424043416977,
0.012383983470499516,
0.08180970698595047,
0.056715212762355804,
-0.030198555439710617,
-0.010293140076100826,
0.208576500415802,
-0.066241055727005,
0.01800600253045559,
0.05172057822346687,
-0.07481939345598221,
-0.03429774194955826,
0.013649354688823223,
-0.02496403641998768,
-0.0826924666762352,
-0.015320496633648872,
-0.10423612594604492,
-0.049669377505779266,
-0.04095999151468277,
-0.09072230756282806,
0.04646654427051544,
-0.09507085382938385,
-0.0520758181810379,
-0.12032681703567505,
-0.15330259501934052,
-0.046840932220220566,
0.045259781181812286,
-0.07434764504432678,
0.01927323266863823,
0.05593917518854141,
-0.05355292558670044,
0.06662744283676147,
-0.022739820182323456,
0.02561947889626026,
-0.07218604534864426,
0.09486978501081467,
0.014563975855708122,
0.012159562669694424,
-0.03955208882689476,
0.03204038366675377,
-0.10713828355073929,
0.07743111997842789,
-0.17043885588645935,
0.017071088775992393,
-0.08562745153903961,
0.05940651521086693,
-0.1303657740354538,
-0.0423809252679348,
0.008841387927532196,
-0.02623651549220085,
0.10586827248334885,
0.10929633677005768,
-0.11141173541545868,
-0.0180997122079134,
0.08645106852054596,
-0.043725159019231796,
-0.09467173367738724,
0.17957501113414764,
0.03210274130105972,
-0.07073090970516205,
0.08281669020652771,
0.12327904999256134,
0.04736149311065674,
-0.17229290306568146,
-0.07397118955850601,
-0.03778606280684471,
0.030503613874316216,
-0.0380096398293972,
0.08641120791435242,
-0.03312172740697861,
0.045339446514844894,
0.022093959152698517,
-0.018005773425102234,
-0.024212105199694633,
-0.087443508207798,
-0.03774942085146904,
-0.06327178329229355,
-0.07364396750926971,
-0.0759449377655983,
0.006780123338103294,
-0.005224363878369331,
-0.0646403580904007,
-0.14615841209888458,
-0.13747233152389526,
0.08642414212226868,
-0.0337863489985466,
-0.015177717432379723,
-0.09014178812503815,
0.17376376688480377,
0.0041328370571136475,
0.015759138390421867,
-0.0973181426525116,
-0.090190090239048,
0.08734408766031265,
-0.08121494948863983,
-0.02467106282711029,
-0.048713475465774536,
0.016061602160334587,
0.028067341074347496,
0.01211546454578638,
0.0012994634453207254,
0.004790047183632851,
-0.039438702166080475,
-0.08580918610095978,
-0.08651985973119736,
-0.02045409195125103,
-0.015500565059483051,
0.06863606721162796,
-0.07683993875980377,
0.02783958986401558,
0.07245545834302902,
0.09130562841892242,
-0.04443483427166939,
-0.08367706090211868,
-0.06577539443969727,
-0.02606881782412529,
-0.029845556244254112,
-0.0854300707578659,
0.010224880650639534,
0.011755453422665596,
0.001415895065292716,
0.08473777025938034,
-0.18132515251636505,
0.030505485832691193,
0.09651084989309311,
0.10809248685836792,
-0.056114379316568375,
0.020022228360176086,
-0.024860672652721405,
-0.03609909862279892,
-0.02343451790511608,
-0.030398186296224594,
0.06134729087352753,
0.024503111839294434,
0.07866086810827255,
-0.09249512106180191,
-0.014115246944129467,
0.06663544476032257,
0.021277667954564095,
-0.036545757204294205,
0.1199679970741272,
0.124179907143116,
-0.17760860919952393,
0.10457179695367813,
0.007183239329606295,
0.02590322680771351,
0.1096678227186203,
0.0007817743462510407,
-0.05227208137512207,
-0.037927597761154175,
0.029676642268896103,
-0.00868771132081747,
0.17837980389595032,
-0.05755097046494484,
0.012424414046108723,
0.027332276105880737,
0.010513942688703537,
0.036651451140642166,
-0.13346660137176514,
0.006395295727998018,
-0.03637048974633217,
-0.08158035576343536,
-0.05590479075908661,
0.016145136207342148,
-0.003693752223625779,
0.07553227245807648,
0.03879564628005028,
-0.028251830488443375,
-0.0032108912710100412,
0.011166098527610302,
-0.07539361715316772,
0.21376265585422516,
-0.09589456021785736,
-0.17223761975765228,
-0.07606903463602066,
-0.01859196461737156,
-0.009305106475949287,
-0.018327180296182632,
0.038644663989543915,
-0.09579513221979141,
-0.056125544011592865,
-0.06396704167127609,
0.08249752223491669,
0.03345334157347679,
0.02200450748205185,
0.0077803898602724075,
-0.03610680624842644,
0.03229643777012825,
-0.118328757584095,
-0.010133707895874977,
-0.017586831003427505,
-0.10040958970785141,
0.04639869183301926,
-0.026295650750398636,
0.04173017665743828,
0.14622443914413452,
0.0452050119638443,
0.0015140373725444078,
0.027688605710864067,
0.2774498760700226,
-0.06704521179199219,
0.07738552987575531,
0.16200071573257446,
0.12889213860034943,
0.07562616467475891,
0.15063956379890442,
-0.002533361781388521,
-0.09068287163972855,
0.00244868453592062,
0.06264893710613251,
-0.07677565515041351,
-0.21540533006191254,
-0.04452720656991005,
-0.019984522834420204,
0.09027890861034393,
0.07620132714509964,
0.06188764423131943,
-0.09501797705888748,
0.03200150281190872,
-0.08787587285041809,
0.07669968158006668,
0.005339606199413538,
0.02209189534187317,
-0.010765799321234226,
0.06733442097902298,
0.07942456007003784,
0.007119955029338598,
0.06614141166210175,
0.1135859563946724,
0.023987429216504097,
0.1170167624950409,
-0.07698395103216171,
0.19237540662288666,
0.006574857980012894,
0.18717753887176514,
0.025094442069530487,
0.00577850267291069,
-0.021351328119635582,
0.020077720284461975,
0.010450529865920544,
-0.03229645639657974,
-0.05047442764043808,
0.007041223347187042,
0.02773929014801979,
-0.07420109212398529,
-0.04454195499420166,
0.09565652906894684,
0.04478701949119568,
0.3267490267753601,
0.10555431246757507,
-0.2863011956214905,
-0.1254839301109314,
-0.02379368618130684,
-0.03326467424631119,
-0.12474375218153,
-0.0035115585196763277,
0.06531558185815811,
-0.07023275643587112,
-0.0002598424907773733,
-0.04390999302268028,
0.08272963762283325,
0.015896636992692947,
0.0015774033963680267,
0.14153090119361877,
0.08048204332590103,
-0.023524004966020584,
0.0024177886079996824,
-0.19370588660240173,
0.16873973608016968,
0.04334590956568718,
0.05687454715371132,
-0.04651058837771416,
0.00499389972537756,
-0.029942290857434273,
0.013174455612897873,
0.060067590326070786,
0.030757341533899307,
-0.018587857484817505,
-0.1730903834104538,
-0.2010306864976883,
0.01880086399614811,
0.0698627382516861,
-0.030739570036530495,
0.07550223916769028,
-0.012653414160013199,
-0.002842299174517393,
-0.02818796969950199,
-0.01779618114233017,
-0.06289578974246979,
-0.07896735519170761,
0.09204927086830139,
-0.0646408274769783,
-0.06364371627569199,
-0.09798726439476013,
0.001398950582370162,
-0.06287092715501785,
0.11378558725118637,
-0.0728965476155281,
-0.05949185788631439,
-0.124782994389534,
0.043033916503190994,
0.15005597472190857,
-0.09903857111930847,
0.05495788902044296,
-0.06900211423635483,
0.0258889552205801,
-0.01744743436574936,
-0.10786356776952744,
0.11907324939966202,
-0.09382147341966629,
-0.20237857103347778,
-0.06414800137281418,
0.11716429889202118,
0.022805094718933105,
0.019682085141539574,
0.005316729657351971,
0.023569585755467415,
-0.06737852841615677,
-0.08101378381252289,
0.024897318333387375,
-0.05043058097362518,
0.04812312126159668,
0.06835107505321503,
0.05658997967839241,
-0.010469536297023296,
-0.035923928022384644,
-0.03820358216762543,
-0.05829101428389549,
0.27774739265441895,
-0.062495507299900055,
0.013879083096981049,
0.11518797278404236,
0.0075990138575434685,
-0.23922105133533478,
-0.0016476581804454327,
0.01800476387143135,
-0.00012216651521157473,
-0.08546695113182068,
-0.11977250128984451,
0.03824513778090477,
0.12445946037769318,
-0.08318526297807693,
0.17846983671188354,
-0.20790702104568481,
-0.11226828396320343,
0.04439050331711769,
-0.0008177252602763474,
0.10079948604106903,
-0.20941157639026642,
-0.05367864668369293,
0.01868799887597561,
-0.19879794120788574,
0.1039763018488884,
-0.07879771292209625,
0.08610179275274277,
-0.015094593167304993,
0.05863004922866821,
0.002865842543542385,
-0.09638675302267075,
0.10485311597585678,
-0.0163038931787014,
0.03271135315299034,
-0.032476458698511124,
0.026555394753813744,
0.026328830048441887,
-0.08129409700632095,
0.08705757558345795,
-0.035646941512823105,
0.08425428718328476,
-0.07042881846427917,
-0.007827936671674252,
-0.09780098497867584,
0.03795607015490532,
-0.044521354138851166,
-0.01841883920133114,
-0.11000855267047882,
0.05181330814957619,
-0.012078003026545048,
-0.03644152358174324,
0.07850531488656998,
0.04626961052417755,
0.07334157824516296,
0.14747683703899384,
-0.05192997679114342,
-0.09210719913244247,
-0.06993280351161957,
-0.012951300479471684,
-0.010630782693624496,
0.05954656004905701,
-0.15827077627182007,
0.043712686747312546,
0.07327432930469513,
0.017811408266425133,
0.07086441665887833,
-0.00670740008354187,
-0.12166193127632141,
0.0001820592297008261,
0.0484573133289814,
-0.16202224791049957,
-0.11536379903554916,
0.01999085023999214,
-0.04161614552140236,
-0.04615718126296997,
0.026548651978373528,
0.12576673924922943,
-0.007774354424327612,
-0.021777305752038956,
-0.013110833242535591,
0.09747102111577988,
0.013260024599730968,
0.21484580636024475,
0.050696227699518204,
0.07128537446260452,
-0.13175718486309052,
0.08107136189937592,
0.0773545578122139,
-0.06263001263141632,
0.030007489025592804,
0.11144791543483734,
-0.13466878235340118,
-0.06305393576622009,
0.01989286206662655,
-0.031073221936821938,
-0.07784814387559891,
0.012524311430752277,
-0.11539239436388016,
-0.03730837255716324,
0.06213248148560524,
0.08425556868314743,
0.017399102449417114,
0.052630722522735596,
0.09923559427261353,
0.0013254339573904872,
-0.05080118775367737,
0.049693763256073,
0.046267468482255936,
0.04266437515616417,
-0.09634417295455933,
0.09012728184461594,
-0.004267234355211258,
0.08602356165647507,
-0.008259821683168411,
0.06455008685588837,
-0.1345210075378418,
-0.015798933804035187,
-0.13249854743480682,
0.0060732197016477585,
-0.03859938681125641,
-0.017760921269655228,
-0.023370856419205666,
-0.04220357537269592,
-0.07364487648010254,
0.033379871398210526,
-0.06896509975194931,
-0.06256455183029175,
-0.047020796686410904,
-0.007466315291821957,
-0.12909282743930817,
-0.005067525897175074,
0.07694672048091888,
-0.06463082879781723,
0.11302319169044495,
0.07883470505475998,
0.030646968632936478,
0.030827077105641365,
-0.07003810256719589,
0.0005729474942199886,
-0.03760702535510063,
0.0477917306125164,
0.04510211944580078,
-0.014299989677965641,
0.02804255299270153,
-0.006830951198935509,
0.03241370990872383,
0.0063338205218315125,
0.049419380724430084,
-0.11992432922124863,
-0.052759770303964615,
-0.05979601666331291,
0.0033262036740779877,
-0.03818093612790108,
0.012415079399943352,
0.07944846153259277,
0.05271385237574577,
0.09385941922664642,
-0.053359854966402054,
0.010163498111069202,
-0.21470162272453308,
-0.015684720128774643,
0.009073643945157528,
-0.048723284155130386,
-0.017766118049621582,
0.049573808908462524,
0.11256293952465057,
-0.0294329896569252,
0.1386604756116867,
0.029574772343039513,
-0.023953964933753014,
0.03662450984120369,
0.012155307456851006,
0.03533734381198883,
0.028888585045933723,
0.059271231293678284,
0.024946298450231552,
0.019903771579265594,
0.023763328790664673,
0.012679914012551308,
0.04626740142703056,
0.0600520595908165,
0.10694878548383713,
0.11977499723434448,
0.15467426180839539,
0.08992307633161545,
0.05354926735162735,
-0.1643214225769043,
-0.07271148264408112,
0.023060232400894165,
-0.18472512066364288,
0.10327903181314468,
-0.03754295036196709,
0.07302463054656982,
0.11120422184467316,
-0.1365637332201004,
0.02942361868917942,
-0.060693975538015366,
-0.08106401562690735,
-0.11766768246889114,
-0.07531900703907013,
-0.07759025692939758,
-0.13369013369083405,
0.017119938507676125,
-0.04789062589406967,
0.05972621217370033,
0.09250277280807495,
0.05700019747018814,
-0.003426644951105118,
0.09932531416416168,
0.07109908759593964,
-0.03786993399262428,
0.05934887006878853,
0.0565064400434494,
-0.03790393844246864,
0.11956938356161118,
-0.07211889326572418,
0.06485005468130112,
-0.0018039783462882042,
0.10646461695432663,
0.016450515016913414,
-0.04287562891840935,
0.118175208568573,
0.005369361490011215,
-0.09227381646633148,
-0.03193233534693718,
-0.0031115449965000153,
-0.020456036552786827,
0.08403573930263519,
0.03995326906442642,
0.06060263141989708,
0.014632970094680786,
0.18239018321037292,
-0.02117457240819931,
0.024229977279901505,
-0.14063924551010132,
0.09944065660238266,
-0.053309399634599686,
0.018946899101138115,
0.017884422093629837,
-0.10395421087741852,
-0.020231151953339577,
0.17358942329883575,
0.14569272100925446,
-0.06285636872053146,
-0.040539514273405075,
0.0033664791844785213,
-0.01698381081223488,
-0.029102787375450134,
0.09110759943723679,
0.07068585604429245,
0.16642719507217407,
-0.026012800633907318,
-0.03267159312963486,
-0.015093696303665638,
-0.06282123178243637,
-0.09358847886323929,
0.07317916303873062,
-0.005398460663855076,
0.04963167384266853,
-0.04629317671060562,
0.03400987759232521,
-0.005078617483377457,
-0.17134995758533478,
0.0022273699287325144,
-0.10637053847312927,
-0.11648929119110107,
0.010992459021508694,
-0.004520716145634651,
-0.007831824012100697,
0.07725109905004501,
-0.01251812744885683,
0.033222075551748276,
0.02468113601207733,
-0.03642919287085533,
-0.14312398433685303,
-0.13234375417232513,
0.09422384947538376,
0.06893356144428253,
0.25350576639175415,
-0.020365510135889053,
0.04684649780392647,
0.1550900936126709,
-0.008533555082976818,
-0.10625056177377701,
0.06462197750806808,
0.05231251195073128,
-0.14851787686347961,
-0.010933483019471169,
0.06213158741593361,
-0.007628923282027245,
0.05762145668268204,
0.04692940413951874,
-0.03998645395040512,
-0.017307454720139503,
0.02902400679886341,
0.015097782015800476,
-0.12131704390048981,
-0.01181893702596426,
-0.07924219965934753,
0.14206863939762115,
0.164865180850029,
-0.05862019211053848,
0.021843407303094864,
-0.06718625873327255,
0.03355111554265022,
-0.001337337540462613,
0.08674121648073196,
0.002132061868906021,
-0.26697319746017456,
0.05938104912638664,
-0.007998572662472725,
0.03831733390688896,
-0.17423830926418304,
-0.01004110835492611,
-0.003022170625627041,
-0.019566942006349564,
-0.07184203714132309,
0.14447835087776184,
0.056312188506126404,
0.020911935716867447,
-0.06863939017057419,
-0.021335572004318237,
-0.0742020234465599,
0.11285985261201859,
-0.1407293826341629,
-0.12538515031337738
] |
null | null |
transformers
|
## MobileBERT fine-tuned on SQuAD v1
[MobileBERT](https://arxiv.org/abs/2004.02984) is a thin version of BERT_LARGE, while equipped with bottleneck structures and a carefully designed balance
between self-attentions and feed-forward networks.
This model was fine-tuned from the HuggingFace checkpoint `google/mobilebert-uncased` on [SQuAD1.1](https://rajpurkar.github.io/SQuAD-explorer).
## Details
| Dataset | Split | # samples |
| -------- | ----- | --------- |
| SQuAD1.1 | train | 90.6K |
| SQuAD1.1 | eval | 11.1k |
### Fine-tuning
- Python: `3.7.5`
- Machine specs:
`CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz`
`Memory: 32 GiB`
`GPUs: 2 GeForce GTX 1070, each with 8GiB memory`
`GPU driver: 418.87.01, CUDA: 10.1`
- script:
```shell
# after install https://github.com/huggingface/transformers
cd examples/question-answering
mkdir -p data
wget -O data/train-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v1.1.json
wget -O data/dev-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v1.1.json
export SQUAD_DIR=`pwd`/data
python run_squad.py \
--model_type mobilebert \
--model_name_or_path google/mobilebert-uncased \
--do_train \
--do_eval \
--do_lower_case \
--train_file $SQUAD_DIR/train-v1.1.json \
--predict_file $SQUAD_DIR/dev-v1.1.json \
--per_gpu_train_batch_size 16 \
--per_gpu_eval_batch_size 16 \
--learning_rate 4e-5 \
--num_train_epochs 5.0 \
--max_seq_length 320 \
--doc_stride 128 \
--warmup_steps 1400 \
--output_dir $SQUAD_DIR/mobilebert-uncased-warmup-squad_v1 2>&1 | tee train-mobilebert-warmup-squad_v1.log
```
It took about 3 hours to finish.
### Results
**Model size**: `95M`
| Metric | # Value | # Original ([Table 5](https://arxiv.org/pdf/2004.02984.pdf))|
| ------ | --------- | --------- |
| **EM** | **82.6** | **82.9** |
| **F1** | **90.0** | **90.0** |
Note that the above results didn't involve any hyperparameter search.
## Example Usage
```python
from transformers import pipeline
qa_pipeline = pipeline(
"question-answering",
model="csarron/mobilebert-uncased-squad-v1",
tokenizer="csarron/mobilebert-uncased-squad-v1"
)
predictions = qa_pipeline({
'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.",
'question': "What day was the game played on?"
})
print(predictions)
# output:
# {'score': 0.7754058241844177, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'}
```
> Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp)
> Made with ❤️ in New York.
|
{"language": "en", "license": "mit", "tags": ["question-answering", "mobilebert"], "datasets": ["squad"], "metrics": ["squad"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}]}
|
question-answering
|
csarron/mobilebert-uncased-squad-v1
|
[
"transformers",
"pytorch",
"safetensors",
"mobilebert",
"question-answering",
"en",
"dataset:squad",
"arxiv:2004.02984",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2004.02984"
] |
[
"en"
] |
TAGS
#transformers #pytorch #safetensors #mobilebert #question-answering #en #dataset-squad #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us
|
MobileBERT fine-tuned on SQuAD v1
---------------------------------
MobileBERT is a thin version of BERT\_LARGE, while equipped with bottleneck structures and a carefully designed balance
between self-attentions and feed-forward networks.
This model was fine-tuned from the HuggingFace checkpoint 'google/mobilebert-uncased' on SQuAD1.1.
Details
-------
Dataset: SQuAD1.1, Split: train, # samples: 90.6K
Dataset: SQuAD1.1, Split: eval, # samples: 11.1k
### Fine-tuning
* Python: '3.7.5'
* Machine specs:
'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'
'Memory: 32 GiB'
'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'
'GPU driver: 418.87.01, CUDA: 10.1'
* script:
It took about 3 hours to finish.
### Results
Model size: '95M'
Metric: EM, # Value: 82.6, # Original (Table 5): 82.9
Metric: F1, # Value: 90.0, # Original (Table 5): 90.0
Note that the above results didn't involve any hyperparameter search.
Example Usage
-------------
>
> Created by Qingqing Cao | GitHub | Twitter
>
>
>
>
> Made with ️ in New York.
>
>
>
|
[
"# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3 hours to finish.",
"### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 82.6, # Original (Table 5): 82.9\nMetric: F1, # Value: 90.0, # Original (Table 5): 90.0\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
"TAGS\n#transformers #pytorch #safetensors #mobilebert #question-answering #en #dataset-squad #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n",
"# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3 hours to finish.",
"### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 82.6, # Original (Table 5): 82.9\nMetric: F1, # Value: 90.0, # Original (Table 5): 90.0\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
56,
27,
95,
107
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #mobilebert #question-answering #en #dataset-squad #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n# samples: 90.6K\nDataset: SQuAD1.1, Split: eval, # samples: 11.1k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3 hours to finish.### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 82.6, # Original (Table 5): 82.9\nMetric: F1, # Value: 90.0, # Original (Table 5): 90.0\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
-0.1281723529100418,
0.08570630848407745,
-0.0007291804067790508,
0.058129169046878815,
0.13153044879436493,
0.02889830246567726,
0.055642496794462204,
0.1125917062163353,
0.028485996648669243,
0.15931548178195953,
0.09990927577018738,
0.07868606597185135,
0.08968527615070343,
0.13260599970817566,
-0.023596402257680893,
-0.07598183304071426,
0.03766999766230583,
0.00540190702304244,
0.02326218970119953,
0.13577726483345032,
0.06668034940958023,
-0.08778902143239975,
0.12380661070346832,
0.006110794376581907,
-0.10720532387495041,
-0.01668737269937992,
0.022113298997282982,
-0.03610038012266159,
0.10813000798225403,
0.0002679723547771573,
0.035003796219825745,
0.053422000259160995,
0.03375904634594917,
-0.10011233389377594,
0.02191036008298397,
0.1293012499809265,
-0.023997837677598,
0.03726348653435707,
0.0698721781373024,
0.002741941250860691,
0.0740412250161171,
-0.022071989253163338,
0.016623208299279213,
0.0493050143122673,
-0.11501078307628632,
-0.21026688814163208,
-0.1677536815404892,
0.03447209671139717,
0.09948986023664474,
0.07222483307123184,
-0.013021879829466343,
0.16908496618270874,
-0.02829725854098797,
0.07727405428886414,
0.2995966076850891,
-0.25542306900024414,
-0.05047981068491936,
0.07663514465093613,
0.0649174377322197,
0.0335833840072155,
-0.008733865804970264,
-0.00558442110195756,
0.022315863519906998,
0.026988821104168892,
0.013991720043122768,
-0.0003156281018164009,
-0.028195297345519066,
0.0020319544710218906,
-0.1470910906791687,
-0.03017105907201767,
0.07265578955411911,
0.029706785455346107,
-0.03695593774318695,
-0.05276186391711235,
-0.11573825776576996,
-0.17500729858875275,
-0.0404442623257637,
0.012291195802390575,
-0.015016064047813416,
-0.005138043779879808,
-0.02790842577815056,
0.013943352736532688,
-0.0468619242310524,
-0.0932098925113678,
-0.05141409859061241,
0.03996976464986801,
0.06724973767995834,
0.06818445026874542,
0.01520901545882225,
0.07512777298688889,
-0.1344144344329834,
-0.0852445662021637,
-0.03591204062104225,
-0.004272595047950745,
-0.054465122520923615,
0.023538779467344284,
0.031145447865128517,
0.023193122819066048,
0.04457341879606247,
0.1584451049566269,
-0.047634486109018326,
0.04557894542813301,
0.0634578987956047,
-0.03687847778201103,
-0.07176114618778229,
0.13696007430553436,
-0.13393165171146393,
-0.10100404918193817,
0.039891745895147324,
0.07395326346158981,
0.021401159465312958,
-0.024462493136525154,
-0.009353662841022015,
0.022815696895122528,
0.09428108483552933,
0.03622260317206383,
0.09403444081544876,
0.06257787346839905,
-0.036091167479753494,
-0.0052993036806583405,
0.2115626484155655,
-0.061565618962049484,
0.005831829737871885,
0.047650039196014404,
-0.06899625062942505,
-0.0157284215092659,
0.010910937562584877,
-0.034126799553632736,
-0.09318246692419052,
-0.017250049859285355,
-0.10285595059394836,
-0.04104690998792648,
-0.04929700866341591,
-0.08563736826181412,
0.03660774230957031,
-0.09507399052381516,
-0.06429053097963333,
-0.11651943624019623,
-0.1718679964542389,
-0.0368206724524498,
0.040904451161623,
-0.06794393062591553,
0.036658186465501785,
0.04344703257083893,
-0.06590315699577332,
0.04145476222038269,
-0.02024911344051361,
0.0347946360707283,
-0.07209000736474991,
0.10292915999889374,
0.03053976595401764,
0.016056643798947334,
-0.004664186388254166,
0.03721780329942703,
-0.11739791929721832,
0.0700545459985733,
-0.18013177812099457,
0.008250212296843529,
-0.07943408936262131,
0.05128416046500206,
-0.12314870208501816,
-0.07076875120401382,
-0.001000970951281488,
-0.03582946956157684,
0.09961288422346115,
0.12101729959249496,
-0.08869898319244385,
-0.03366914764046669,
0.10395041108131409,
-0.05375559255480766,
-0.10695614665746689,
0.16707858443260193,
0.03487902879714966,
-0.06567992269992828,
0.09279084950685501,
0.11923854798078537,
0.01427658274769783,
-0.17230567336082458,
-0.09733448177576065,
-0.049640778452157974,
0.005916693713515997,
-0.05898142606019974,
0.07115781307220459,
-0.045078303664922714,
0.045956891030073166,
0.028009792789816856,
-0.021049970760941505,
-0.016271909698843956,
-0.08914881944656372,
-0.03156176581978798,
-0.06911934167146683,
-0.07593926042318344,
-0.05653279647231102,
0.028567517176270485,
-0.021402940154075623,
-0.07829848676919937,
-0.1466924101114273,
-0.13491366803646088,
0.09685006737709045,
-0.04715485870838165,
-0.016939274966716766,
-0.10246147960424423,
0.19299627840518951,
-0.00618429621681571,
0.006170296110212803,
-0.10496208816766739,
-0.11436451971530914,
0.08220595866441727,
-0.12115510553121567,
-0.04876110702753067,
-0.06052029877901077,
0.01727294735610485,
0.057163652032613754,
-0.004294142127037048,
-0.007466957438737154,
0.007648894097656012,
-0.0301689263433218,
-0.06709866225719452,
-0.09125685691833496,
-0.007296395022422075,
-0.0030296319164335728,
0.08431290090084076,
-0.0856182873249054,
0.029514366760849953,
0.050681665539741516,
0.10191141068935394,
-0.04173027351498604,
-0.08968625962734222,
-0.05828815698623657,
-0.01856853999197483,
-0.03836235776543617,
-0.09503632038831711,
0.03256338834762573,
0.007338738068938255,
0.03438485413789749,
0.0745897889137268,
-0.2134910672903061,
0.02690223604440689,
0.11342931538820267,
0.06320236623287201,
-0.04176324978470802,
0.03831552341580391,
-0.022973692044615746,
-0.050698172301054,
-0.030209219083189964,
-0.0013064505765214562,
0.02755667082965374,
0.035415809601545334,
0.09186898916959763,
-0.10780512541532516,
-0.017602333799004555,
0.05697731673717499,
0.027748124673962593,
-0.022255197167396545,
0.12205840647220612,
0.15334376692771912,
-0.18240001797676086,
0.09751949459314346,
0.013116379268467426,
0.001293693669140339,
0.09467928111553192,
0.013034231029450893,
-0.03866375610232353,
-0.02766496129333973,
0.03197164833545685,
-0.0030305134132504463,
0.16470380127429962,
-0.044636406004428864,
0.027084486559033394,
0.030735762789845467,
-0.01670413464307785,
0.047287192195653915,
-0.14480258524417877,
0.014055841602385044,
-0.050755392760038376,
-0.07440098375082016,
-0.053076814860105515,
0.03994777053594589,
0.010386054404079914,
0.07374117523431778,
0.01863345317542553,
-0.03932831063866615,
0.0015412588836625218,
0.017800409346818924,
-0.06252558529376984,
0.21975776553153992,
-0.09289627522230148,
-0.20707975327968597,
-0.06715793162584305,
-0.05012746527791023,
-0.0439496673643589,
-0.024305136874318123,
0.03428393229842186,
-0.10480619221925735,
-0.060947809368371964,
-0.06252685189247131,
0.07711748033761978,
0.05261329561471939,
0.005283801816403866,
0.00417301757261157,
-0.026435611769557,
0.03452940285205841,
-0.12087853252887726,
0.0064681158401072025,
-0.032104481011629105,
-0.11673321574926376,
0.03435057774186134,
-0.02594999223947525,
0.04384561628103256,
0.15907885134220123,
0.040934041142463684,
-0.0058520641177892685,
0.017080971971154213,
0.2309456467628479,
-0.05171231925487518,
0.0636562928557396,
0.19525891542434692,
0.1372450739145279,
0.08158185333013535,
0.13251665234565735,
-0.00923473946750164,
-0.08812344074249268,
-0.009057295508682728,
0.0698154866695404,
-0.0677889958024025,
-0.22707077860832214,
-0.056107353419065475,
-0.018411843106150627,
0.09428385645151138,
0.058355145156383514,
0.05123889818787575,
-0.08617939800024033,
0.03039383888244629,
-0.0904838889837265,
0.030735397711396217,
0.00956042855978012,
0.035126883536577225,
0.030641967430710793,
0.06873127818107605,
0.09426084905862808,
-0.01068099308758974,
0.04511015862226486,
0.0947004035115242,
0.010015550069510937,
0.14250726997852325,
-0.0810835137963295,
0.16522741317749023,
0.008300893008708954,
0.1753038465976715,
0.013530018739402294,
0.021893661469221115,
-0.01621248945593834,
0.015208140946924686,
0.00942090805619955,
-0.038081757724285126,
-0.050812017172575,
-0.011833248659968376,
0.037510547786951065,
-0.06472085416316986,
-0.018620213493704796,
0.11899451166391373,
0.05507862940430641,
0.27987438440322876,
0.09625507891178131,
-0.29798170924186707,
-0.09379517287015915,
-0.037213586270809174,
-0.030533239245414734,
-0.12500028312206268,
0.0019429032690823078,
0.02439350076019764,
-0.0769483670592308,
0.03657359257340431,
-0.06980489194393158,
0.07449367642402649,
0.006943781394511461,
-0.006429658737033606,
0.12749989330768585,
0.10870843380689621,
-0.006578861735761166,
0.020624175667762756,
-0.1773792803287506,
0.17109785974025726,
0.03335776925086975,
0.06511598825454712,
-0.04550167918205261,
-0.0033497512340545654,
-0.025622954592108727,
0.00887361727654934,
0.06360466033220291,
0.012896018102765083,
0.0033115334808826447,
-0.140901580452919,
-0.1992248296737671,
0.005682358518242836,
0.08233020454645157,
0.009668082930147648,
0.0808180421590805,
-0.028004685416817665,
-0.01823950931429863,
-0.02792060375213623,
-0.06394444406032562,
-0.07124938815832138,
-0.07706892490386963,
0.08425886183977127,
-0.07859651744365692,
-0.056871041655540466,
-0.10622875392436981,
-0.020317038521170616,
-0.08686867356300354,
0.12209145724773407,
-0.08225228637456894,
-0.07095641642808914,
-0.1013856828212738,
0.0200240109115839,
0.15064887702465057,
-0.09309768676757812,
0.07915309816598892,
-0.081821970641613,
0.029298007488250732,
-0.00005200473970035091,
-0.1097412034869194,
0.11141075193881989,
-0.11384148895740509,
-0.21799024939537048,
-0.054595697671175,
0.12195868790149689,
0.003964398987591267,
0.007741871755570173,
-0.00441429577767849,
0.02720395289361477,
-0.07571899890899658,
-0.07730959355831146,
0.016544535756111145,
-0.0325239934027195,
0.08348660171031952,
0.10715679824352264,
0.041884277015924454,
-0.05047295615077019,
-0.042925167828798294,
-0.043793682008981705,
-0.04999789968132973,
0.28265380859375,
-0.0481351763010025,
0.00926921796053648,
0.10793404281139374,
0.003547027939930558,
-0.25197210907936096,
0.0009037457639351487,
0.014119905419647694,
-0.022735752165317535,
-0.06953179836273193,
-0.11491650342941284,
0.052741799503564835,
0.11640908569097519,
-0.08863838016986847,
0.16266939043998718,
-0.22235481441020966,
-0.12469402700662613,
0.05224720761179924,
0.01629801094532013,
0.12703457474708557,
-0.19542984664440155,
-0.043366797268390656,
0.004029137082397938,
-0.16769176721572876,
0.10334132611751556,
-0.06091926619410515,
0.09582805633544922,
-0.03310491517186165,
0.01410752348601818,
-0.0023097428493201733,
-0.1168244257569313,
0.08990463614463806,
-0.02854992263019085,
0.04087362438440323,
-0.027594804763793945,
0.03812519088387489,
0.056707713752985,
-0.06615842878818512,
0.08156998455524445,
-0.013484043069183826,
0.09649258106946945,
-0.037955593317747116,
-0.010178606025874615,
-0.10893367230892181,
0.050075385719537735,
-0.03372151032090187,
-0.007498697843402624,
-0.11171559244394302,
0.04828931763768196,
-0.010799835436046124,
-0.03058622032403946,
0.07617335021495819,
0.05343976989388466,
0.0794675424695015,
0.11737023293972015,
-0.027566751465201378,
-0.06896141171455383,
-0.0881236270070076,
-0.03030143305659294,
-0.007797908969223499,
0.03603983297944069,
-0.09966365247964859,
0.033539071679115295,
0.0722324550151825,
0.016332218423485756,
0.04284796491265297,
0.0024117755237966776,
-0.10813689976930618,
-0.0038608673494309187,
0.060640327632427216,
-0.1737251728773117,
-0.12167509645223618,
0.020879363641142845,
-0.023205138742923737,
-0.04747404158115387,
0.050773587077856064,
0.10199900716543198,
0.00639977864921093,
-0.028274646028876305,
-0.004613880068063736,
0.10485051572322845,
0.0384281724691391,
0.21457092463970184,
0.061469461768865585,
0.06322800368070602,
-0.141827791929245,
0.0894114151597023,
0.06735550612211227,
-0.05811115726828575,
0.004670657217502594,
0.09585600346326828,
-0.1471938043832779,
-0.04481268674135208,
0.05564812943339348,
-0.005988295190036297,
-0.06881692260503769,
-0.01187820266932249,
-0.13564494252204895,
-0.03324778005480766,
0.061516884714365005,
0.08317626267671585,
0.030020352452993393,
0.052598338574171066,
0.09150780737400055,
-0.008144129998981953,
-0.07207709550857544,
0.07137487083673477,
0.0580412894487381,
0.055953413248062134,
-0.1128140389919281,
0.09627275913953781,
0.0236677099019289,
0.07107191532850266,
-0.010865801945328712,
0.07522078603506088,
-0.10224664956331253,
-0.01184140332043171,
-0.1198311373591423,
0.0034043455962091684,
-0.042027588933706284,
-0.010304766707122326,
-0.03033776581287384,
-0.025352830067276955,
-0.07378429919481277,
0.033859193325042725,
-0.07028207182884216,
-0.05243883281946182,
-0.03154654800891876,
-0.016927173361182213,
-0.1425936371088028,
-0.00009854932432062924,
0.08165675401687622,
-0.07249492406845093,
0.10469141602516174,
0.0783856138586998,
0.03225061297416687,
0.04668701812624931,
-0.050853896886110306,
-0.03174346685409546,
-0.030845602974295616,
0.06681918352842331,
0.015550102107226849,
-0.04061933234333992,
0.033559445291757584,
0.019686130806803703,
0.03179789334535599,
0.004230717197060585,
0.04821373149752617,
-0.12702491879463196,
-0.04484424740076065,
-0.07396277785301208,
-0.0032591246999800205,
-0.01785336807370186,
0.013548623770475388,
0.07078985869884491,
0.03773192688822746,
0.1296970695257187,
-0.056063808500766754,
-0.003697392763569951,
-0.2067391723394394,
-0.012151308357715607,
-0.007883132435381413,
-0.08597986400127411,
-0.04033525660634041,
0.05562194436788559,
0.1184370219707489,
-0.0066884104162454605,
0.1383175402879715,
0.026844263076782227,
-0.04270349442958832,
0.02684815227985382,
0.04740207642316818,
0.030365651473402977,
0.023847846314311028,
0.06245436146855354,
0.02349221706390381,
0.0297593604773283,
0.05056219920516014,
0.007757210172712803,
0.0591050460934639,
0.04905487224459648,
0.15195126831531525,
0.13920824229717255,
0.13268937170505524,
0.08602630347013474,
0.0540211945772171,
-0.1908472180366516,
-0.09493941068649292,
0.038596488535404205,
-0.18510907888412476,
0.11085380613803864,
-0.030573217198252678,
0.079927459359169,
0.10857335478067398,
-0.11250822246074677,
0.01929665543138981,
-0.04850829020142555,
-0.07133303582668304,
-0.1307152807712555,
-0.025466151535511017,
-0.08851084858179092,
-0.13747437298297882,
0.02306281588971615,
-0.0349297933280468,
0.052755262702703476,
0.11643386632204056,
0.049224864691495895,
-0.009821922518312931,
0.1349688023328781,
0.0654420256614685,
-0.045275963842868805,
0.03503873571753502,
0.04409567639231682,
-0.04400062561035156,
0.1207452043890953,
-0.07290308177471161,
0.048154424875974655,
-0.004736609756946564,
0.12080845981836319,
0.022443482652306557,
-0.07319579273462296,
0.1231035515666008,
-0.0019160052761435509,
-0.08430390805006027,
-0.03154836967587471,
0.015265172347426414,
-0.010493306443095207,
0.10382641851902008,
0.025247404351830482,
0.06459442526102066,
0.012375871650874615,
0.16448810696601868,
-0.013946505263447762,
-0.006885244511067867,
-0.09986419230699539,
0.1234905868768692,
-0.022958559915423393,
0.01940230093896389,
0.026480359956622124,
-0.08674603700637817,
-0.012049615383148193,
0.18189582228660583,
0.1332205981016159,
-0.040074028074741364,
-0.050708696246147156,
0.014535804279148579,
-0.016546206548810005,
-0.04235183075070381,
0.07896623015403748,
0.08299049735069275,
0.17539039254188538,
-0.0300071332603693,
-0.0511050783097744,
-0.010319283232092857,
-0.048886075615882874,
-0.09443975239992142,
0.061108872294425964,
-0.002485140459612012,
0.028373125940561295,
-0.038449086248874664,
0.032238930463790894,
-0.000900645274668932,
-0.144352525472641,
-0.011073320172727108,
-0.10436812788248062,
-0.1364927589893341,
0.007674311753362417,
0.04124966636300087,
-0.016291802749037743,
0.06499557942152023,
-0.009521308355033398,
0.018702669069170952,
0.023437708616256714,
-0.032494254410266876,
-0.15516406297683716,
-0.1355285346508026,
0.08219451457262039,
0.049585483968257904,
0.25397977232933044,
-0.024880874902009964,
0.05850524082779884,
0.16060583293437958,
-0.03746908903121948,
-0.10895667970180511,
0.07481741160154343,
0.03347165510058403,
-0.14713945984840393,
0.023371858522295952,
0.05064348131418228,
-0.006386316381394863,
0.04198093339800835,
0.02564511075615883,
-0.011862942948937416,
-0.007177518215030432,
0.047512806951999664,
-0.008516724221408367,
-0.10942783206701279,
0.004614336881786585,
-0.08490052819252014,
0.13803359866142273,
0.1545625925064087,
-0.04673907905817032,
0.03756178542971611,
-0.06723552197217941,
0.0291450172662735,
-0.001575352973304689,
0.10009229928255081,
0.02550787851214409,
-0.29713162779808044,
0.05295230448246002,
-0.04489642754197121,
0.042872972786426544,
-0.15767212212085724,
-0.0098408292979002,
0.0008992935181595385,
-0.03541203588247299,
-0.11183057725429535,
0.1330677717924118,
0.04871894046664238,
0.018283400684595108,
-0.06487303972244263,
0.014872150495648384,
-0.08986767381429672,
0.10973016172647476,
-0.15408168733119965,
-0.12841732800006866
] |
null | null |
transformers
|
## MobileBERT fine-tuned on SQuAD v2
[MobileBERT](https://arxiv.org/abs/2004.02984) is a thin version of BERT_LARGE, while equipped with bottleneck structures and a carefully designed balance
between self-attentions and feed-forward networks.
This model was fine-tuned from the HuggingFace checkpoint `google/mobilebert-uncased` on [SQuAD2.0](https://rajpurkar.github.io/SQuAD-explorer).
## Details
| Dataset | Split | # samples |
| -------- | ----- | --------- |
| SQuAD2.0 | train | 130k |
| SQuAD2.0 | eval | 12.3k |
### Fine-tuning
- Python: `3.7.5`
- Machine specs:
`CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz`
`Memory: 32 GiB`
`GPUs: 2 GeForce GTX 1070, each with 8GiB memory`
`GPU driver: 418.87.01, CUDA: 10.1`
- script:
```shell
# after install https://github.com/huggingface/transformers
cd examples/question-answering
mkdir -p data
wget -O data/train-v2.0.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v2.0.json
wget -O data/dev-v2.0.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v2.0.json
export SQUAD_DIR=`pwd`/data
python run_squad.py \
--model_type mobilebert \
--model_name_or_path google/mobilebert-uncased \
--do_train \
--do_eval \
--do_lower_case \
--version_2_with_negative \
--train_file $SQUAD_DIR/train-v2.0.json \
--predict_file $SQUAD_DIR/dev-v2.0.json \
--per_gpu_train_batch_size 16 \
--per_gpu_eval_batch_size 16 \
--learning_rate 4e-5 \
--num_train_epochs 5.0 \
--max_seq_length 320 \
--doc_stride 128 \
--warmup_steps 1400 \
--save_steps 2000 \
--output_dir $SQUAD_DIR/mobilebert-uncased-warmup-squad_v2 2>&1 | tee train-mobilebert-warmup-squad_v2.log
```
It took about 3.5 hours to finish.
### Results
**Model size**: `95M`
| Metric | # Value | # Original ([Table 5](https://arxiv.org/pdf/2004.02984.pdf))|
| ------ | --------- | --------- |
| **EM** | **75.2** | **76.2** |
| **F1** | **78.8** | **79.2** |
Note that the above results didn't involve any hyperparameter search.
## Example Usage
```python
from transformers import pipeline
qa_pipeline = pipeline(
"question-answering",
model="csarron/mobilebert-uncased-squad-v2",
tokenizer="csarron/mobilebert-uncased-squad-v2"
)
predictions = qa_pipeline({
'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.",
'question': "What day was the game played on?"
})
print(predictions)
# output:
# {'score': 0.71434086561203, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'}
```
> Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp)
> Made with ❤️ in New York.
|
{"language": "en", "license": "mit", "tags": ["question-answering", "mobilebert"], "datasets": ["squad_v2"], "metrics": ["squad_v2"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}]}
|
question-answering
|
csarron/mobilebert-uncased-squad-v2
|
[
"transformers",
"pytorch",
"onnx",
"safetensors",
"mobilebert",
"question-answering",
"en",
"dataset:squad_v2",
"arxiv:2004.02984",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2004.02984"
] |
[
"en"
] |
TAGS
#transformers #pytorch #onnx #safetensors #mobilebert #question-answering #en #dataset-squad_v2 #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us
|
MobileBERT fine-tuned on SQuAD v2
---------------------------------
MobileBERT is a thin version of BERT\_LARGE, while equipped with bottleneck structures and a carefully designed balance
between self-attentions and feed-forward networks.
This model was fine-tuned from the HuggingFace checkpoint 'google/mobilebert-uncased' on SQuAD2.0.
Details
-------
Dataset: SQuAD2.0, Split: train, # samples: 130k
Dataset: SQuAD2.0, Split: eval, # samples: 12.3k
### Fine-tuning
* Python: '3.7.5'
* Machine specs:
'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'
'Memory: 32 GiB'
'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'
'GPU driver: 418.87.01, CUDA: 10.1'
* script:
It took about 3.5 hours to finish.
### Results
Model size: '95M'
Metric: EM, # Value: 75.2, # Original (Table 5): 76.2
Metric: F1, # Value: 78.8, # Original (Table 5): 79.2
Note that the above results didn't involve any hyperparameter search.
Example Usage
-------------
>
> Created by Qingqing Cao | GitHub | Twitter
>
>
>
>
> Made with ️ in New York.
>
>
>
|
[
"# samples: 130k\nDataset: SQuAD2.0, Split: eval, # samples: 12.3k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3.5 hours to finish.",
"### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 75.2, # Original (Table 5): 76.2\nMetric: F1, # Value: 78.8, # Original (Table 5): 79.2\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
"TAGS\n#transformers #pytorch #onnx #safetensors #mobilebert #question-answering #en #dataset-squad_v2 #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n",
"# samples: 130k\nDataset: SQuAD2.0, Split: eval, # samples: 12.3k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3.5 hours to finish.",
"### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 75.2, # Original (Table 5): 76.2\nMetric: F1, # Value: 78.8, # Original (Table 5): 79.2\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
63,
26,
95,
106
] |
[
"passage: TAGS\n#transformers #pytorch #onnx #safetensors #mobilebert #question-answering #en #dataset-squad_v2 #arxiv-2004.02984 #license-mit #endpoints_compatible #region-us \n# samples: 130k\nDataset: SQuAD2.0, Split: eval, # samples: 12.3k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 3.5 hours to finish.### Results\n\n\nModel size: '95M'\n\n\nMetric: EM, # Value: 75.2, # Original (Table 5): 76.2\nMetric: F1, # Value: 78.8, # Original (Table 5): 79.2\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
-0.1348331719636917,
0.06482125073671341,
-0.0014147701440379024,
0.06846626847982407,
0.12749825417995453,
0.030476873740553856,
0.07556270807981491,
0.12080681324005127,
0.036692678928375244,
0.15933430194854736,
0.11493188887834549,
0.05672044679522514,
0.10260391235351562,
0.12507805228233337,
-0.029870033264160156,
-0.06451567262411118,
0.021088002249598503,
-0.01894855685532093,
0.009883414953947067,
0.12182489037513733,
0.07178754359483719,
-0.05509971082210541,
0.12925876677036285,
-0.008578754030168056,
-0.11878717690706253,
-0.02193247154355049,
0.039938703179359436,
-0.037402622401714325,
0.09524746984243393,
0.024694446474313736,
0.01329679973423481,
0.05307583138346672,
0.031040508300065994,
-0.10887689888477325,
0.025260739028453827,
0.13255377113819122,
-0.04389951750636101,
0.02870817296206951,
0.05680600181221962,
0.01113729178905487,
0.07766890525817871,
-0.01704060472548008,
0.01566094346344471,
0.06325140595436096,
-0.11228108406066895,
-0.1691054403781891,
-0.1785532832145691,
0.0019119702046737075,
0.07045222818851471,
0.07205919176340103,
-0.00417752331122756,
0.159361332654953,
-0.012000841088593006,
0.09209232777357101,
0.2546135485172272,
-0.2456066757440567,
-0.05730119347572327,
0.07755997776985168,
0.06694239377975464,
0.055682044476270676,
-0.0016493768198415637,
-0.003205580171197653,
0.015256049111485481,
0.045784372836351395,
0.013953780755400658,
-0.015520048327744007,
0.014174536801874638,
-0.0012585826916620135,
-0.13256551325321198,
-0.04595695808529854,
0.07159682363271713,
0.0425277017056942,
-0.040034741163253784,
-0.05335162580013275,
-0.12929567694664001,
-0.19027842581272125,
-0.02746642753481865,
0.010770496912300587,
-0.03878052905201912,
-0.0017314390279352665,
-0.045077428221702576,
-0.0067637222819030285,
-0.045885514467954636,
-0.09043674916028976,
-0.036997199058532715,
0.02699427865445614,
0.05244852229952812,
0.08181840926408768,
-0.007198757492005825,
0.07145199924707413,
-0.14432387053966522,
-0.09506437182426453,
-0.025633305311203003,
-0.009152333252131939,
-0.08169374614953995,
0.02605416625738144,
0.04428299888968468,
-0.005838743411004543,
0.04107629507780075,
0.15827420353889465,
-0.0756428986787796,
0.0501379631459713,
0.061086505651474,
-0.04298532009124756,
-0.050424348562955856,
0.1408945620059967,
-0.1301269382238388,
-0.11197088658809662,
0.040574487298727036,
0.05276830121874809,
-0.022335905581712723,
-0.04079902917146683,
-0.017672916874289513,
0.02624223381280899,
0.0794583410024643,
0.01769668608903885,
0.07901889085769653,
0.04687368869781494,
-0.033657144755125046,
-0.016714414581656456,
0.20920021831989288,
-0.057883478701114655,
0.014124120585620403,
0.05103038251399994,
-0.07178646326065063,
-0.01674412377178669,
0.020453309640288353,
-0.041247446089982986,
-0.08401820063591003,
-0.03955862671136856,
-0.09400021284818649,
-0.0515160895884037,
-0.040960583835840225,
-0.07933327555656433,
0.04499272629618645,
-0.07496897876262665,
-0.05790998414158821,
-0.13436418771743774,
-0.16269788146018982,
-0.04720320925116539,
0.041711095720529556,
-0.07363785058259964,
0.04282372444868088,
0.050076693296432495,
-0.06806588917970657,
0.051332298666238785,
-0.009975780732929707,
0.02994430810213089,
-0.06768705695867538,
0.11284131556749344,
0.04012124985456467,
0.00847654789686203,
-0.00893340166658163,
0.03750036656856537,
-0.09667719155550003,
0.06026172637939453,
-0.1816948801279068,
0.01857839524745941,
-0.08067227900028229,
0.06629940122365952,
-0.11262200027704239,
-0.053286630660295486,
-0.007275879383087158,
-0.04713151976466179,
0.10793141275644302,
0.11355511099100113,
-0.10911983996629715,
-0.030858850106596947,
0.09205437451601028,
-0.038860175758600235,
-0.10065856575965881,
0.1669652909040451,
0.03425586223602295,
-0.048420388251543045,
0.08131548762321472,
0.11273399740457535,
0.060543328523635864,
-0.18551842868328094,
-0.08912960439920425,
-0.04549495130777359,
-0.01306143682450056,
-0.07019457221031189,
0.07148536294698715,
-0.019573615863919258,
0.03382408991456032,
0.026583347469568253,
0.015403376892209053,
-0.00665037939324975,
-0.08712419867515564,
-0.02743547223508358,
-0.07597742229700089,
-0.06695222854614258,
-0.09308359771966934,
0.025030553340911865,
-0.025997286662459373,
-0.06496132910251617,
-0.13205105066299438,
-0.1531618982553482,
0.0884876623749733,
-0.03446047753095627,
-0.024294409900903702,
-0.09503377974033356,
0.17841362953186035,
0.007844019681215286,
0.02558610588312149,
-0.09224912524223328,
-0.11786813288927078,
0.08336535096168518,
-0.11645854264497757,
-0.03619533032178879,
-0.0810890793800354,
0.013332301750779152,
0.05141584575176239,
0.0060492572374641895,
0.010488479398190975,
-0.00045097761903889477,
-0.038483861833810806,
-0.07371237874031067,
-0.08706994354724884,
-0.0025193863548338413,
-0.014986516907811165,
0.05740660801529884,
-0.07987397164106369,
0.03559320420026779,
0.08772893249988556,
0.10393025726079941,
-0.033877357840538025,
-0.09221091121435165,
-0.06881581246852875,
-0.034234486520290375,
-0.031748272478580475,
-0.09717584401369095,
0.008375917561352253,
0.01791330799460411,
0.03939003124833107,
0.09181573987007141,
-0.20963048934936523,
0.03842736780643463,
0.09461328387260437,
0.09233810752630234,
-0.04869237169623375,
0.024879785254597664,
-0.021061638370156288,
-0.04723455384373665,
-0.04114806279540062,
-0.010948087088763714,
0.03749584034085274,
0.03328258916735649,
0.09700644761323929,
-0.10765132308006287,
-0.015010151080787182,
0.08089877665042877,
0.028543680906295776,
-0.03131740540266037,
0.1362035572528839,
0.13764113187789917,
-0.17782799899578094,
0.09005396068096161,
0.01817108504474163,
0.039103999733924866,
0.11150537431240082,
0.016077082604169846,
-0.03561709076166153,
-0.05347372964024544,
0.03240615874528885,
-0.0004040376516059041,
0.17523877322673798,
-0.03929611295461655,
0.03522917628288269,
0.02535278908908367,
-0.0007227034657262266,
0.05142279341816902,
-0.12031596899032593,
0.009594643488526344,
-0.04006841033697128,
-0.07719156891107559,
-0.036157846450805664,
0.027170758694410324,
0.014480270445346832,
0.06661684066057205,
0.014588924124836922,
-0.053454943001270294,
0.012162400409579277,
0.02064933069050312,
-0.06787529587745667,
0.21314795315265656,
-0.11286848038434982,
-0.1915399432182312,
-0.07094243168830872,
-0.05911662429571152,
-0.041433777660131454,
-0.029216717928647995,
0.032878417521715164,
-0.10690892487764359,
-0.04690791666507721,
-0.05461728200316429,
0.08041893690824509,
0.04633680731058121,
0.02125677280128002,
0.002876556944102049,
-0.013798462226986885,
0.02894367277622223,
-0.11821004748344421,
0.000486676930449903,
-0.011212819255888462,
-0.11579986661672592,
0.04483799263834953,
-0.019701629877090454,
0.02930363640189171,
0.16491924226284027,
0.03916148096323013,
0.0036076658871024847,
0.02743173949420452,
0.239219531416893,
-0.06583308428525925,
0.08032689988613129,
0.16792848706245422,
0.16190990805625916,
0.07864197343587875,
0.132242813706398,
-0.001750884810462594,
-0.10026516765356064,
0.011645302176475525,
0.06375906616449356,
-0.08152353018522263,
-0.23826782405376434,
-0.05428393930196762,
-0.006162494886666536,
0.1147189736366272,
0.04669208452105522,
0.06412835419178009,
-0.06057362258434296,
0.0204917062073946,
-0.06766138970851898,
0.05642900615930557,
0.030629243701696396,
0.024660352617502213,
0.007690927013754845,
0.06506771594285965,
0.07981478422880173,
0.0036779153160750866,
0.06977567076683044,
0.12005521357059479,
0.027814548462629318,
0.12022709846496582,
-0.07321034371852875,
0.15584062039852142,
0.00012641778448596597,
0.16988520324230194,
0.008701481856405735,
-0.0030554889235645533,
-0.0017406785627827048,
0.018089009448885918,
0.017327893525362015,
-0.03799505531787872,
-0.03523990139365196,
-0.005917562637478113,
0.04574687406420708,
-0.08235104382038116,
-0.043926119804382324,
0.12684576213359833,
0.057203829288482666,
0.29495662450790405,
0.08559902757406235,
-0.27912768721580505,
-0.10359197109937668,
-0.039050642400979996,
-0.04867742583155632,
-0.13135594129562378,
0.010564575903117657,
0.07138548791408539,
-0.08911094814538956,
0.00949674379080534,
-0.06849472969770432,
0.08223137259483337,
0.007352953776717186,
0.002886299742385745,
0.1708400547504425,
0.0947020873427391,
-0.014461755752563477,
-0.00701071647927165,
-0.1915665864944458,
0.16142480075359344,
0.04890453815460205,
0.06213477626442909,
-0.042807117104530334,
0.005097328685224056,
-0.028616761788725853,
0.01615128107368946,
0.05213978886604309,
0.017008261755108833,
-0.015444912016391754,
-0.1449717879295349,
-0.19390732049942017,
0.002401783363893628,
0.08573709428310394,
0.006111904978752136,
0.0556677021086216,
-0.010607185773551464,
-0.016208332031965256,
-0.03444278612732887,
-0.07309979945421219,
-0.06998365372419357,
-0.06117863208055496,
0.08949273824691772,
-0.06029640510678291,
-0.08626436442136765,
-0.09719620645046234,
0.005781920161098242,
-0.07849756628274918,
0.12409244477748871,
-0.09111499786376953,
-0.05115395039319992,
-0.1178598552942276,
0.021216977387666702,
0.13457083702087402,
-0.07800544798374176,
0.05396252125501633,
-0.08349763602018356,
0.008259491994976997,
-0.019733255729079247,
-0.10515950620174408,
0.10955928266048431,
-0.10117296874523163,
-0.20965059101581573,
-0.06121477857232094,
0.11762060225009918,
0.02683568000793457,
0.0033774988260120153,
-0.009682273492217064,
0.009889221750199795,
-0.07606826722621918,
-0.08948666602373123,
0.016378726810216904,
-0.03482753410935402,
0.06635941565036774,
0.08401720225811005,
0.06251971423625946,
-0.03782954439520836,
-0.03013736568391323,
-0.020831899717450142,
-0.08156714588403702,
0.2861442565917969,
-0.045257698744535446,
0.004515094216912985,
0.11972567439079285,
0.009480423294007778,
-0.23643815517425537,
0.014848031103610992,
0.013037081807851791,
-0.010838235728442669,
-0.05731187388300896,
-0.11363107711076736,
0.04471229016780853,
0.12186188995838165,
-0.08061345666646957,
0.177566796541214,
-0.24130786955356598,
-0.11332201212644577,
0.025751888751983643,
-0.009498639963567257,
0.11684077978134155,
-0.18273937702178955,
-0.05599500238895416,
0.010425654239952564,
-0.17193861305713654,
0.08758438378572464,
-0.10144858062267303,
0.09202231466770172,
-0.011393621563911438,
0.04051385447382927,
-0.0034467962104827166,
-0.10318537056446075,
0.08977499604225159,
-0.011953071691095829,
0.04142184555530548,
-0.026326585561037064,
0.047118179500103,
0.049301162362098694,
-0.08018404990434647,
0.0762656107544899,
-0.008490418083965778,
0.1032635048031807,
-0.0807107537984848,
-0.0045647406950592995,
-0.09575337916612625,
0.06319194287061691,
-0.034005049616098404,
0.0036333089228719473,
-0.11007029563188553,
0.05066094174981117,
-0.017046483233571053,
-0.0361638106405735,
0.06617727875709534,
0.04584668576717377,
0.04376015067100525,
0.16472257673740387,
-0.047864221036434174,
-0.08122091740369797,
-0.0646846741437912,
-0.025044599547982216,
-0.002922005718573928,
0.043508049100637436,
-0.11412105709314346,
0.03890938311815262,
0.06889235228300095,
0.0025144501123577356,
0.033639635890722275,
-0.005509801208972931,
-0.11023804545402527,
-0.006482742261141539,
0.0479883998632431,
-0.17342379689216614,
-0.09715636819601059,
0.009011887945234776,
-0.025444040074944496,
-0.028121475130319595,
0.03979288786649704,
0.1278742104768753,
0.011957037262618542,
-0.011896476149559021,
-0.012359658256173134,
0.10895814746618271,
0.015564044937491417,
0.21250635385513306,
0.05453051999211311,
0.058698419481515884,
-0.13887713849544525,
0.09257350116968155,
0.07872849702835083,
-0.06607192009687424,
0.015396300703287125,
0.12810003757476807,
-0.1401439756155014,
-0.05803163722157478,
0.03277469426393509,
-0.027342375367879868,
-0.06558695435523987,
-0.00017469850718043745,
-0.11368627101182938,
-0.01413644663989544,
0.06158934533596039,
0.12229891866445541,
0.012485681101679802,
0.055413179099559784,
0.0889621376991272,
-0.00279720826074481,
-0.07900621742010117,
0.06112065538764,
0.042381804436445236,
0.036552801728248596,
-0.10627780109643936,
0.07612224668264389,
0.011221976019442081,
0.10241568833589554,
-0.013680223375558853,
0.07153818011283875,
-0.11272856593132019,
-0.01952824927866459,
-0.1011662483215332,
-0.013392903842031956,
-0.03870587423443794,
-0.012457243166863918,
-0.02432967536151409,
-0.030372900888323784,
-0.07373017817735672,
0.034489743411540985,
-0.05381782352924347,
-0.05298604816198349,
-0.030663490295410156,
-0.010751071386039257,
-0.1287216991186142,
-0.010115094482898712,
0.07274964451789856,
-0.071330726146698,
0.11637204140424728,
0.08879709243774414,
0.025012675672769547,
0.020578008145093918,
-0.044556792825460434,
-0.009849697351455688,
-0.03921699523925781,
0.05847616121172905,
0.034982483834028244,
-0.029393797740340233,
0.04927792772650719,
0.007790009491145611,
0.04141681268811226,
0.014301362447440624,
0.05990172177553177,
-0.1288014054298401,
-0.02681107632815838,
-0.07664389163255692,
-0.0006017145933583379,
-0.035163599997758865,
0.008696618489921093,
0.0648256316781044,
0.04504961520433426,
0.10366618633270264,
-0.05499552935361862,
0.01007228996604681,
-0.20486584305763245,
-0.015757201239466667,
0.007604541257023811,
-0.07557160407304764,
-0.01936223916709423,
0.06769633293151855,
0.11954747885465622,
-0.01569502428174019,
0.14435000717639923,
0.03363832086324692,
-0.03140869736671448,
0.038460295647382736,
0.018929725512862206,
0.012130686081945896,
0.022173769772052765,
0.016558963805437088,
0.024685552343726158,
0.037998802959918976,
0.04189207777380943,
0.016971372067928314,
0.05063599720597267,
0.031988419592380524,
0.13192886114120483,
0.12721692025661469,
0.13379476964473724,
0.09398150444030762,
0.05600174143910408,
-0.17644278705120087,
-0.08785908669233322,
0.03692740947008133,
-0.21091385185718536,
0.10223986208438873,
-0.05952521413564682,
0.04713758826255798,
0.10146670788526535,
-0.10837873071432114,
0.03011152893304825,
-0.033397335559129715,
-0.08545713871717453,
-0.12700052559375763,
-0.0363888218998909,
-0.07634913921356201,
-0.14513979852199554,
0.017944911494851112,
-0.029961377382278442,
0.06056534871459007,
0.0862906277179718,
0.05556677281856537,
-0.010644372552633286,
0.12834478914737701,
0.06456902623176575,
-0.025848396122455597,
0.05730942636728287,
0.05232679843902588,
-0.028290536254644394,
0.10821722447872162,
-0.07394231855869293,
0.04306493327021599,
0.00023573647195007652,
0.11424470692873001,
0.029095470905303955,
-0.050496943295001984,
0.10935445129871368,
-0.004281180910766125,
-0.08561962842941284,
-0.04095667600631714,
0.002849854528903961,
-0.040309786796569824,
0.10641604661941528,
0.042498864233493805,
0.04758431017398834,
0.01770045794546604,
0.17011375725269318,
-0.02489110454916954,
0.020001256838440895,
-0.13337738811969757,
0.10221792012453079,
-0.028469432145357132,
0.029426055029034615,
0.00783777516335249,
-0.09318839013576508,
-0.035255637019872665,
0.1573617160320282,
0.12456817924976349,
-0.05204317718744278,
-0.04881466552615166,
0.024862607941031456,
-0.017035435885190964,
-0.04018436744809151,
0.08199281245470047,
0.06230364739894867,
0.18624477088451385,
-0.010487762279808521,
-0.0632333904504776,
-0.005669459700584412,
-0.06465958058834076,
-0.08288388699293137,
0.04772495850920677,
-0.013835538178682327,
0.04348573833703995,
-0.022935904562473297,
0.04936932027339935,
0.003533120732754469,
-0.14932803809642792,
0.015919514000415802,
-0.08627418428659439,
-0.1107056513428688,
0.024906247854232788,
0.0002245309588033706,
-0.03202281519770622,
0.06339973956346512,
-0.007371486164629459,
0.026014108210802078,
0.03481810539960861,
-0.04214536398649216,
-0.14475028216838837,
-0.13627216219902039,
0.0790550708770752,
0.021123366430401802,
0.2686983346939087,
-0.03309524059295654,
0.062195658683776855,
0.163852721452713,
-0.026210501790046692,
-0.10252328217029572,
0.07071689516305923,
0.037577830255031586,
-0.15975546836853027,
-0.015147584490478039,
0.06202227622270584,
-0.00675982004031539,
0.06121225282549858,
0.028044072911143303,
-0.016561077907681465,
-0.016836773604154587,
0.024781635031104088,
0.027185121551156044,
-0.11046166718006134,
-0.000894260301720351,
-0.06741277873516083,
0.13456477224826813,
0.14954042434692383,
-0.045499350875616074,
0.04774820804595947,
-0.07158298790454865,
0.03987516835331917,
-0.01164600346237421,
0.11438732594251633,
0.016446711495518684,
-0.3007405698299408,
0.05721544101834297,
-0.03360287472605705,
0.05472428724169731,
-0.15194830298423767,
-0.02172429859638214,
0.014130598865449429,
-0.030473848804831505,
-0.09200019389390945,
0.14946943521499634,
0.04485694319009781,
0.017781635746359825,
-0.07117582112550735,
0.013169560581445694,
-0.09068875759840012,
0.1158563569188118,
-0.16052468121051788,
-0.12758013606071472
] |
null | null |
transformers
|
## RoBERTa-base fine-tuned on SQuAD v1
This model was fine-tuned from the HuggingFace [RoBERTa](https://arxiv.org/abs/1907.11692) base checkpoint on [SQuAD1.1](https://rajpurkar.github.io/SQuAD-explorer).
This model is case-sensitive: it makes a difference between english and English.
## Details
| Dataset | Split | # samples |
| -------- | ----- | --------- |
| SQuAD1.1 | train | 96.8K |
| SQuAD1.1 | eval | 11.8k |
### Fine-tuning
- Python: `3.7.5`
- Machine specs:
`CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz`
`Memory: 32 GiB`
`GPUs: 2 GeForce GTX 1070, each with 8GiB memory`
`GPU driver: 418.87.01, CUDA: 10.1`
- script:
```shell
# after install https://github.com/huggingface/transformers
cd examples/question-answering
mkdir -p data
wget -O data/train-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/train-v1.1.json
wget -O data/dev-v1.1.json https://rajpurkar.github.io/SQuAD-explorer/dataset/dev-v1.1.json
python run_energy_squad.py \
--model_type roberta \
--model_name_or_path roberta-base \
--do_train \
--do_eval \
--train_file train-v1.1.json \
--predict_file dev-v1.1.json \
--per_gpu_train_batch_size 12 \
--per_gpu_eval_batch_size 16 \
--learning_rate 3e-5 \
--num_train_epochs 2.0 \
--max_seq_length 320 \
--doc_stride 128 \
--data_dir data \
--output_dir data/roberta-base-squad-v1 2>&1 | tee train-roberta-base-squad-v1.log
```
It took about 2 hours to finish.
### Results
**Model size**: `477M`
| Metric | # Value |
| ------ | --------- |
| **EM** | **83.0** |
| **F1** | **90.4** |
Note that the above results didn't involve any hyperparameter search.
## Example Usage
```python
from transformers import pipeline
qa_pipeline = pipeline(
"question-answering",
model="csarron/roberta-base-squad-v1",
tokenizer="csarron/roberta-base-squad-v1"
)
predictions = qa_pipeline({
'context': "The game was played on February 7, 2016 at Levi's Stadium in the San Francisco Bay Area at Santa Clara, California.",
'question': "What day was the game played on?"
})
print(predictions)
# output:
# {'score': 0.8625259399414062, 'start': 23, 'end': 39, 'answer': 'February 7, 2016'}
```
> Created by [Qingqing Cao](https://awk.ai/) | [GitHub](https://github.com/csarron) | [Twitter](https://twitter.com/sysnlp)
> Made with ❤️ in New York.
|
{"language": "en", "license": "mit", "tags": ["question-answering", "roberta", "roberta-base"], "datasets": ["squad"], "metrics": ["squad"], "widget": [{"text": "Which name is also used to describe the Amazon rainforest in English?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}, {"text": "How many square kilometers of rainforest is covered in the basin?", "context": "The Amazon rainforest (Portuguese: Floresta Amaz\u00f4nica or Amaz\u00f4nia; Spanish: Selva Amaz\u00f3nica, Amazon\u00eda or usually Amazonia; French: For\u00eat amazonienne; Dutch: Amazoneregenwoud), also known in English as Amazonia or the Amazon Jungle, is a moist broadleaf forest that covers most of the Amazon basin of South America. This basin encompasses 7,000,000 square kilometres (2,700,000 sq mi), of which 5,500,000 square kilometres (2,100,000 sq mi) are covered by the rainforest. This region includes territory belonging to nine nations. The majority of the forest is contained within Brazil, with 60% of the rainforest, followed by Peru with 13%, Colombia with 10%, and with minor amounts in Venezuela, Ecuador, Bolivia, Guyana, Suriname and French Guiana. States or departments in four nations contain \"Amazonas\" in their names. The Amazon represents over half of the planet's remaining rainforests, and comprises the largest and most biodiverse tract of tropical rainforest in the world, with an estimated 390 billion individual trees divided into 16,000 species."}]}
|
question-answering
|
csarron/roberta-base-squad-v1
|
[
"transformers",
"pytorch",
"jax",
"safetensors",
"roberta",
"question-answering",
"roberta-base",
"en",
"dataset:squad",
"arxiv:1907.11692",
"license:mit",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1907.11692"
] |
[
"en"
] |
TAGS
#transformers #pytorch #jax #safetensors #roberta #question-answering #roberta-base #en #dataset-squad #arxiv-1907.11692 #license-mit #endpoints_compatible #region-us
|
RoBERTa-base fine-tuned on SQuAD v1
-----------------------------------
This model was fine-tuned from the HuggingFace RoBERTa base checkpoint on SQuAD1.1.
This model is case-sensitive: it makes a difference between english and English.
Details
-------
Dataset: SQuAD1.1, Split: train, # samples: 96.8K
Dataset: SQuAD1.1, Split: eval, # samples: 11.8k
### Fine-tuning
* Python: '3.7.5'
* Machine specs:
'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'
'Memory: 32 GiB'
'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'
'GPU driver: 418.87.01, CUDA: 10.1'
* script:
It took about 2 hours to finish.
### Results
Model size: '477M'
Note that the above results didn't involve any hyperparameter search.
Example Usage
-------------
>
> Created by Qingqing Cao | GitHub | Twitter
>
>
>
>
> Made with ️ in New York.
>
>
>
|
[
"# samples: 96.8K\nDataset: SQuAD1.1, Split: eval, # samples: 11.8k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.",
"### Results\n\n\nModel size: '477M'\n\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
"TAGS\n#transformers #pytorch #jax #safetensors #roberta #question-answering #roberta-base #en #dataset-squad #arxiv-1907.11692 #license-mit #endpoints_compatible #region-us \n",
"# samples: 96.8K\nDataset: SQuAD1.1, Split: eval, # samples: 11.8k",
"### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.",
"### Results\n\n\nModel size: '477M'\n\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
64,
27,
95,
65
] |
[
"passage: TAGS\n#transformers #pytorch #jax #safetensors #roberta #question-answering #roberta-base #en #dataset-squad #arxiv-1907.11692 #license-mit #endpoints_compatible #region-us \n# samples: 96.8K\nDataset: SQuAD1.1, Split: eval, # samples: 11.8k### Fine-tuning\n\n\n* Python: '3.7.5'\n* Machine specs:\n\n\n'CPU: Intel(R) Core(TM) i7-6800K CPU @ 3.40GHz'\n\n\n'Memory: 32 GiB'\n\n\n'GPUs: 2 GeForce GTX 1070, each with 8GiB memory'\n\n\n'GPU driver: 418.87.01, CUDA: 10.1'\n* script:\n\n\nIt took about 2 hours to finish.### Results\n\n\nModel size: '477M'\n\n\n\nNote that the above results didn't involve any hyperparameter search.\n\n\nExample Usage\n-------------\n\n\n\n> \n> Created by Qingqing Cao | GitHub | Twitter\n> \n> \n> \n\n\n\n> \n> Made with ️ in New York.\n> \n> \n>"
] |
[
-0.12511669099330902,
0.08033295720815659,
0.0019898004829883575,
0.09229458123445511,
0.1241801381111145,
0.0372810885310173,
0.06103558838367462,
0.11210767179727554,
0.08790256828069687,
0.15083837509155273,
0.12080906331539154,
0.005513670854270458,
0.11062385141849518,
0.07094967365264893,
-0.020809201523661613,
-0.055015310645103455,
0.07802557945251465,
0.0058367252349853516,
-0.03937360271811485,
0.11245864629745483,
0.05890525504946709,
-0.07193668186664581,
0.14353559911251068,
0.01708204858005047,
-0.14267095923423767,
-0.03910968452692032,
0.0027513992972671986,
-0.02401791885495186,
0.10150570422410965,
0.06479541212320328,
0.03067050687968731,
0.037357281893491745,
0.058527640998363495,
-0.07764778286218643,
0.02674601413309574,
0.08750122785568237,
-0.019844358786940575,
0.04380087926983833,
0.016718478873372078,
0.11056305468082428,
0.07935472577810287,
0.02526266686618328,
-0.033494096249341965,
0.05109177902340889,
-0.07219788432121277,
-0.07746578007936478,
-0.13002565503120422,
0.0421563982963562,
0.09030601382255554,
0.06259012967348099,
0.0025327084586024284,
0.15781839191913605,
-0.046115074306726456,
0.026457538828253746,
0.20671707391738892,
-0.23454728722572327,
-0.07349208742380142,
0.07540369778871536,
0.102747343480587,
0.031172743067145348,
-0.0004600492538884282,
-0.029989758506417274,
0.04059353843331337,
0.009943129494786263,
-0.08794315159320831,
-0.005039576906710863,
-0.06306574493646622,
0.0196515042334795,
-0.07946407794952393,
-0.03540051355957985,
0.09555551409721375,
0.04006609320640564,
-0.014046581462025642,
-0.015940312296152115,
-0.11723894625902176,
-0.1892227679491043,
0.008109575137495995,
0.04782066121697426,
-0.05414870008826256,
0.01991982012987137,
0.004263496957719326,
0.02871181257069111,
-0.03715760260820389,
-0.15082497894763947,
-0.027636803686618805,
0.006126039661467075,
0.09185240417718887,
0.08717487007379532,
0.058489035815000534,
0.07142117619514465,
-0.11532342433929443,
-0.05882543697953224,
-0.05904599279165268,
-0.021439990028738976,
-0.09162328392267227,
0.060554083436727524,
0.018693430349230766,
0.10993645340204239,
0.07199402898550034,
0.17090792953968048,
-0.04285818710923195,
0.032326605170965195,
0.11752071231603622,
-0.03693123534321785,
-0.09242262691259384,
0.07903185486793518,
-0.18633343279361725,
-0.0550951287150383,
0.04635453596711159,
0.046942565590143204,
-0.015975231304764748,
-0.01013822853565216,
-0.05387534573674202,
-0.032813090831041336,
0.04820476099848747,
-0.0071702697314321995,
0.04553172364830971,
0.03875013813376427,
-0.045253705233335495,
0.0033331005834043026,
0.24754223227500916,
-0.00763206509873271,
-0.010957153514027596,
0.053292497992515564,
-0.08705583214759827,
-0.06002170965075493,
-0.008156083524227142,
-0.01931033656001091,
-0.060257527977228165,
-0.07482248544692993,
-0.10467620939016342,
-0.05596429482102394,
-0.07085558772087097,
-0.0681738331913948,
0.035996414721012115,
-0.10007650405168533,
-0.014087649993598461,
-0.11745559424161911,
-0.19888950884342194,
-0.010803153738379478,
0.06995881348848343,
-0.11074025183916092,
0.012133174575865269,
0.06954643875360489,
-0.03539124131202698,
0.048242200165987015,
-0.016835786402225494,
0.09176085889339447,
-0.06114637851715088,
0.09504581242799759,
0.060890913009643555,
0.0359007902443409,
-0.06572005152702332,
0.02077685482800007,
-0.10984192788600922,
0.08759152889251709,
-0.09512759745121002,
-0.013207699172198772,
-0.061888258904218674,
0.06689146161079407,
-0.10570560395717621,
-0.029922939836978912,
-0.00878959521651268,
-0.02714766189455986,
0.1265120953321457,
0.10358907282352448,
-0.07845049351453781,
0.026599399745464325,
0.04275236278772354,
-0.019014161080121994,
-0.09372232109308243,
0.2047242522239685,
0.03612635284662247,
-0.05605677142739296,
0.026818227022886276,
0.07597742974758148,
-0.00478793028742075,
-0.14531858265399933,
-0.07063446938991547,
-0.015051867812871933,
0.0007983158575370908,
-0.11781530827283859,
0.09988489001989365,
0.014427870512008667,
0.030393710359930992,
0.0360155925154686,
-0.020405836403369904,
0.02032954804599285,
-0.08551383763551712,
-0.04193012788891792,
-0.09364071488380432,
-0.11558616906404495,
-0.11212100833654404,
0.018420027568936348,
0.01317423116415739,
-0.031230665743350983,
-0.13267022371292114,
-0.19822393357753754,
0.108881376683712,
-0.04496333748102188,
-0.019832564517855644,
-0.11805010586977005,
0.18342095613479614,
-0.007990180514752865,
0.013897573575377464,
-0.05748526751995087,
-0.04004311189055443,
0.08797436207532883,
-0.029886268079280853,
-0.0474105142056942,
-0.04384811222553253,
-0.0005638161092065275,
0.039642587304115295,
0.008677268400788307,
0.025555090978741646,
0.0335683599114418,
-0.04327473044395447,
-0.08873657137155533,
-0.05981065332889557,
-0.003827781416475773,
0.006992635317146778,
0.07947976142168045,
-0.09589002281427383,
0.027093978598713875,
0.043423283845186234,
0.07513155788183212,
-0.06706930696964264,
-0.05478530004620552,
-0.053690288215875626,
-0.018272899091243744,
-0.08302430808544159,
-0.08566063642501831,
0.028892314061522484,
0.00534641882404685,
0.02174641564488411,
0.08916537463665009,
-0.11509377509355545,
-0.018907733261585236,
0.10824289172887802,
0.14130403101444244,
-0.05108734220266342,
-0.0017057252116501331,
-0.006014936603605747,
-0.10383453220129013,
0.019973112270236015,
-0.022486049681901932,
0.05275988206267357,
0.005689411889761686,
0.08618691563606262,
-0.0910477414727211,
0.043556202203035355,
0.054551612585783005,
0.04429495334625244,
-0.02007310464978218,
0.12531161308288574,
0.19901205599308014,
-0.12131654471158981,
0.11471742391586304,
-0.035755306482315063,
-0.018425974994897842,
0.07511133700609207,
0.03355373442173004,
-0.05960497260093689,
-0.003382425056770444,
0.06652391701936722,
-0.004454551264643669,
0.2018878012895584,
-0.08638107031583786,
0.03406859189271927,
0.027198564261198044,
-0.0364646278321743,
0.04740689694881439,
-0.15912029147148132,
0.0060183703899383545,
-0.0766647532582283,
-0.06641428172588348,
-0.02427137829363346,
-0.008818353526294231,
-0.04270834103226662,
0.054251834750175476,
0.049719348549842834,
-0.08830530941486359,
0.03144653141498566,
0.02722499892115593,
-0.032417796552181244,
0.19478629529476166,
-0.06447888910770416,
-0.17093664407730103,
-0.09211451560258865,
-0.04901131987571716,
0.0033918346744030714,
0.003964693751186132,
-0.004340191371738911,
-0.10574639588594437,
-0.08041159808635712,
-0.03640498220920563,
0.05086902901530266,
0.07833154499530792,
0.011693866923451424,
0.05860462412238121,
0.019833015277981758,
0.010167497210204601,
-0.06187431886792183,
0.05483057349920273,
-0.021275650709867477,
-0.0767352506518364,
0.053011804819107056,
-0.06795322149991989,
0.06328283995389938,
0.10066501051187515,
0.06260813772678375,
-0.045625340193510056,
0.06242063269019127,
0.19418513774871826,
-0.04049207270145416,
0.07347402721643448,
0.17389319837093353,
0.09094569087028503,
0.025887642055749893,
0.14873231947422028,
-0.024499299004673958,
-0.08302769064903259,
0.0008359103812836111,
0.01466398872435093,
-0.09926797449588776,
-0.1955271065235138,
-0.031239181756973267,
-0.05715440958738327,
0.08077002316713333,
0.05198691785335541,
0.058936748653650284,
-0.14919129014015198,
0.07356786727905273,
-0.06528079509735107,
0.10455609112977982,
-0.03696320578455925,
0.019746161997318268,
-0.027643907815217972,
0.06367212533950806,
0.047492511570453644,
-0.006610110402107239,
0.06932542473077774,
0.08948501944541931,
0.12045704573392868,
0.14302316308021545,
-0.07024543732404709,
0.24662064015865326,
-0.043091677129268646,
0.259014755487442,
0.05838152766227722,
0.020290188491344452,
-0.03471509367227554,
0.0006411668146029115,
-0.004206801764667034,
0.010167845524847507,
-0.08642220497131348,
0.05286344140768051,
-0.0038494945038110018,
-0.07576483488082886,
0.03027421422302723,
0.17603209614753723,
0.028710227459669113,
0.3026382327079773,
0.11073609441518784,
-0.27055028080940247,
-0.13840560615062714,
-0.023689229041337967,
-0.03465264290571213,
-0.16146674752235413,
0.03150816261768341,
0.10258857905864716,
-0.03568821772933006,
-0.08290619403123856,
-0.06370226293802261,
0.0812639445066452,
0.024654017761349678,
-0.017327532172203064,
0.13018231093883514,
0.10912411659955978,
-0.0009388708858750761,
0.024454697966575623,
-0.16978655755519867,
0.1553334891796112,
0.04372698441147804,
0.04458607733249664,
-0.08061715960502625,
0.014444938860833645,
-0.026292147114872932,
-0.0176131222397089,
0.05844777449965477,
0.0002727328974287957,
-0.04820246249437332,
-0.13236001133918762,
-0.24677640199661255,
0.02975097857415676,
0.024892905727028847,
0.04185381904244423,
0.05057726055383682,
-0.02294851839542389,
0.010208538733422756,
-0.030934102833271027,
-0.023952048271894455,
-0.04057185351848602,
-0.09576497226953506,
0.07212257385253906,
-0.03861511871218681,
-0.03230797126889229,
-0.09196159243583679,
0.02196114882826805,
-0.07865983992815018,
0.17134541273117065,
-0.1518375724554062,
-0.05053498223423958,
-0.06666160374879837,
0.05308578908443451,
0.17036136984825134,
-0.0744728147983551,
0.049510031938552856,
-0.08705455809831619,
-0.02973516471683979,
0.00902554951608181,
-0.06262864917516708,
0.12830311059951782,
-0.1057542935013771,
-0.20205965638160706,
-0.08982089906930923,
0.05073537304997444,
-0.007667015306651592,
0.03443003445863724,
-0.009770984761416912,
0.003097095061093569,
-0.1204366609454155,
-0.049665067344903946,
-0.028278980404138565,
-0.08083072304725647,
0.06325042992830276,
0.08110218495130539,
0.01416532602161169,
-0.01849510334432125,
-0.013932867906987667,
-0.11190661042928696,
-0.05248258262872696,
0.20970696210861206,
-0.027542611584067345,
0.0011010878952220082,
0.13777825236320496,
0.04194314405322075,
-0.23418192565441132,
0.0065557630732655525,
-0.002301809610798955,
0.034975651651620865,
-0.15125438570976257,
-0.13100066781044006,
0.027029965072870255,
0.150539368391037,
-0.08454665541648865,
0.21434111893177032,
-0.1551101803779602,
-0.11628198623657227,
0.03139668330550194,
-0.014003731310367584,
0.10807280242443085,
-0.1819053441286087,
-0.02874867245554924,
0.014214541763067245,
-0.20172472298145294,
0.08854002505540848,
-0.00259042507968843,
0.09018535166978836,
-0.053911324590444565,
0.034253861755132675,
-0.016581688076257706,
-0.1306401640176773,
0.12408355623483658,
-0.03623967245221138,
0.030617985874414444,
-0.04651077836751938,
0.026393504813313484,
0.05095502734184265,
-0.05539441108703613,
0.07239196449518204,
-0.0570080392062664,
0.08435958623886108,
-0.04568242281675339,
-0.027311792597174644,
-0.06917482614517212,
-0.01485166884958744,
-0.00266720331273973,
-0.011463218368589878,
-0.11992849409580231,
0.0794827789068222,
-0.09165371209383011,
-0.018978308886289597,
0.09243986010551453,
0.06480126082897186,
-0.02737390249967575,
0.1479605734348297,
-0.05159126594662666,
-0.07133834064006805,
0.046789031475782394,
-0.00289475847966969,
0.014812774024903774,
0.08224822580814362,
-0.12524281442165375,
0.014033900573849678,
0.08335968852043152,
-0.03036402352154255,
0.019021442160010338,
-0.007445489056408405,
-0.07257555425167084,
0.04784915968775749,
0.04531199485063553,
-0.15284332633018494,
-0.08199446648359299,
0.024116355925798416,
-0.03212409093976021,
-0.04062014818191528,
-0.024483639746904373,
0.05469662696123123,
-0.020036380738019943,
-0.02064058743417263,
-0.009202237240970135,
0.10542626678943634,
0.01933528482913971,
0.20677368342876434,
0.03864389657974243,
0.05326426774263382,
-0.1357565075159073,
0.04147309437394142,
0.06236249953508377,
-0.04085729643702507,
0.015650272369384766,
0.10731633007526398,
-0.151676207780838,
-0.0619879812002182,
-0.026816826313734055,
-0.06956535577774048,
-0.10000719875097275,
-0.027059605345129967,
-0.128329798579216,
-0.02235296741127968,
0.03737792372703552,
-0.015455869026482105,
0.03303945064544678,
0.03658042848110199,
0.11329052597284317,
-0.023293767124414444,
-0.09780408442020416,
0.040934883058071136,
0.00021820818074047565,
0.0589040108025074,
-0.09443497657775879,
0.02195035107433796,
0.010961705818772316,
0.07474999129772186,
-0.0006352117634378374,
0.09432482719421387,
-0.1353064328432083,
0.018234169110655785,
-0.19460001587867737,
0.019177168607711792,
-0.0722452700138092,
-0.023983508348464966,
-0.009800449945032597,
-0.032140132039785385,
-0.08423460274934769,
0.03833796828985214,
-0.04931497201323509,
-0.022340063005685806,
-0.05051322281360626,
-0.021595532074570656,
-0.06236673891544342,
-0.0062759872525930405,
0.09809009730815887,
-0.05501163750886917,
0.09967754036188126,
0.08458871394395828,
0.03015897609293461,
-0.006599153857678175,
-0.0967446118593216,
-0.002036202931776643,
-0.042262688279151917,
0.08061495423316956,
0.017470696941018105,
-0.001912513398565352,
0.048478417098522186,
0.026227768510580063,
0.05060058459639549,
-0.025658614933490753,
0.1094214916229248,
-0.09858562797307968,
-0.08258114010095596,
-0.10390038788318634,
-0.009037399664521217,
-0.019261043518781662,
-0.003923955839127302,
0.09572352468967438,
0.053531575947999954,
0.09135178476572037,
-0.011255043558776379,
-0.0074316831305623055,
-0.12908023595809937,
0.003583968849852681,
0.008582305163145065,
-0.04731256887316704,
-0.0483899861574173,
0.01617048680782318,
0.10602815449237823,
0.003197113983333111,
0.1649995744228363,
-0.010968852788209915,
-0.06925515085458755,
0.01575944945216179,
0.027072738856077194,
0.05333217605948448,
0.020296916365623474,
0.0925002470612526,
0.0488501638174057,
0.0602470263838768,
0.013809454627335072,
0.03261339291930199,
0.06401683390140533,
0.08154385536909103,
0.0799485519528389,
0.13307936489582062,
0.14818289875984192,
0.1316453516483307,
0.038804661482572556,
-0.19494791328907013,
-0.02414006181061268,
0.07603884488344193,
-0.257575660943985,
0.05336753651499748,
-0.03908374533057213,
0.05981828272342682,
0.10352659225463867,
-0.1122567430138588,
-0.043688077479600906,
-0.09422563761472702,
-0.0944831594824791,
-0.0839943066239357,
-0.02210305444896221,
-0.09246519953012466,
-0.12575215101242065,
0.044706761837005615,
0.006127920933067799,
-0.009363592602312565,
0.131329745054245,
0.04864474758505821,
-0.019439680501818657,
0.11931996047496796,
0.13070684671401978,
-0.01802360638976097,
-0.019541461020708084,
0.06929042935371399,
-0.0387980081140995,
0.10683926939964294,
-0.13577552139759064,
0.04999830201268196,
-0.050973113626241684,
0.10991673916578293,
0.018798300996422768,
-0.028212809935212135,
0.11582256108522415,
0.029997780919075012,
-0.05081027373671532,
-0.05795849859714508,
0.011383144184947014,
-0.010886295698583126,
0.12010978907346725,
0.01860242336988449,
0.0991823598742485,
0.018175166100263596,
0.09196852892637253,
-0.002361285500228405,
0.03892827406525612,
-0.14392554759979248,
0.044723086059093475,
-0.09725549072027206,
-0.004538796842098236,
0.02490386553108692,
-0.06712735444307327,
-0.019337791949510574,
0.2266186773777008,
0.12285266071557999,
-0.12832112610340118,
-0.08526937663555145,
0.007128546014428139,
-0.02024785988032818,
-0.09512249380350113,
0.10485120117664337,
0.12377376109361649,
0.1605490893125534,
-0.033363793045282364,
-0.1296795904636383,
-0.0054885284043848515,
-0.07179886847734451,
-0.1096532791852951,
0.002744883531704545,
-0.014071911573410034,
0.036331307142972946,
-0.05776062607765198,
-0.01828037016093731,
-0.02075415477156639,
-0.10121777653694153,
0.06690993160009384,
-0.0630887821316719,
-0.11967804282903671,
0.028874170035123825,
0.004759683273732662,
-0.019093530252575874,
0.03328077495098114,
-0.061331916600465775,
0.07537327706813812,
-0.02189522050321102,
-0.034466084092855453,
-0.1554774045944214,
-0.09291493147611618,
0.10461296886205673,
0.12922148406505585,
0.22783328592777252,
-0.03200850635766983,
0.09445347636938095,
0.1615506112575531,
-0.011533587239682674,
-0.16101686656475067,
0.06943288445472717,
0.029604170471429825,
-0.17241813242435455,
0.006542276591062546,
0.01661098562180996,
-0.011806783266365528,
0.06984207779169083,
0.04075060412287712,
0.032504938542842865,
-0.01129087433218956,
0.033277466893196106,
0.06496579200029373,
-0.13294020295143127,
0.016417672857642174,
-0.09708263725042343,
0.14337758719921112,
0.14947152137756348,
-0.06567158550024033,
-0.0007600170793011785,
-0.0518001914024353,
0.05922393873333931,
-0.03952344134449959,
0.08298414945602417,
-0.008808977901935577,
-0.2958131730556488,
0.047758832573890686,
-0.007834800519049168,
0.021065223962068558,
-0.15410815179347992,
-0.011737802997231483,
-0.05935872718691826,
-0.05107859894633293,
-0.094682477414608,
0.12004440277814865,
0.07197826355695724,
0.03033703751862049,
-0.06493119895458221,
-0.008059477433562279,
-0.09701120853424072,
0.06263981014490128,
-0.12052959948778152,
-0.17217421531677246
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# distilbert-base-uncased-finetuned-emotion
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2175
- Accuracy: 0.923
- F1: 0.9233
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|
| 0.8352 | 1.0 | 250 | 0.3079 | 0.91 | 0.9086 |
| 0.247 | 2.0 | 500 | 0.2175 | 0.923 | 0.9233 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "args": "default"}, "metrics": [{"type": "accuracy", "value": 0.923, "name": "Accuracy"}, {"type": "f1", "value": 0.9232542847906783, "name": "F1"}]}]}]}
|
text-classification
|
cscottp27/distilbert-base-uncased-finetuned-emotion
|
[
"transformers",
"pytorch",
"tensorboard",
"distilbert",
"text-classification",
"generated_from_trainer",
"dataset:emotion",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
distilbert-base-uncased-finetuned-emotion
=========================================
This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2175
* Accuracy: 0.923
* F1: 0.9233
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 64
* eval\_batch\_size: 64
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.10.0+cu111
* Datasets 1.16.1
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
67,
98,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #distilbert #text-classification #generated_from_trainer #dataset-emotion #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3"
] |
[
-0.10365526378154755,
0.11108539253473282,
-0.0026109113823622465,
0.1317654550075531,
0.16546793282032013,
0.045472968369722366,
0.1148209348320961,
0.12493137270212173,
-0.08185860514640808,
0.032128069549798965,
0.10837704688310623,
0.1617085337638855,
0.02285127155482769,
0.09674810618162155,
-0.05789001286029816,
-0.2760312557220459,
-0.013312868773937225,
0.05033736675977707,
-0.015638014301657677,
0.13228055834770203,
0.09348677843809128,
-0.12362140417098999,
0.09650623053312302,
0.003427819348871708,
-0.17494015395641327,
0.003667527576908469,
0.0020847665145993233,
-0.04351482167840004,
0.1478479653596878,
0.0203389972448349,
0.10677221417427063,
0.008356831967830658,
0.08522238582372665,
-0.2250649482011795,
0.018061498180031776,
0.03874968737363815,
0.0002561478759162128,
0.08761061728000641,
0.03720410540699959,
-0.015301057137548923,
0.15303263068199158,
-0.06405466049909592,
0.05445051193237305,
0.021798426285386086,
-0.11284452676773071,
-0.2199874073266983,
-0.08077996224164963,
0.041833750903606415,
0.06329819560050964,
0.11930888146162033,
-0.019842874258756638,
0.12837383151054382,
-0.09601787477731705,
0.09610112756490707,
0.23596636950969696,
-0.2448965162038803,
-0.06858085840940475,
0.020751558244228363,
0.014530576765537262,
0.043537914752960205,
-0.11989867687225342,
-0.039740413427352905,
0.050211407244205475,
0.05117671564221382,
0.11863920837640762,
-0.03312069922685623,
-0.0999940037727356,
0.00910378061234951,
-0.12911130487918854,
-0.04672069475054741,
0.16751150786876678,
0.0594206377863884,
-0.02597641386091709,
-0.05425291508436203,
-0.056162502616643906,
-0.1675000786781311,
-0.031035728752613068,
-0.016450364142656326,
0.054442595690488815,
-0.01592334546148777,
-0.06372737884521484,
0.010441784746944904,
-0.12011896073818207,
-0.04583032429218292,
-0.06421378999948502,
0.10641534626483917,
0.021976949647068977,
0.006818883586674929,
-0.017207900062203407,
0.10352788120508194,
0.0009800537955015898,
-0.12149646133184433,
0.0210917666554451,
0.02099326066672802,
0.026333261281251907,
-0.03047778643667698,
-0.06905151158571243,
-0.055389244109392166,
-0.004919437691569328,
0.10190358012914658,
-0.06666407734155655,
0.045935261994600296,
0.04524286091327667,
0.037766341120004654,
-0.06852026283740997,
0.19527418911457062,
-0.03294919803738594,
-0.03212658688426018,
-0.012501158751547337,
0.06016761064529419,
0.020768651738762856,
-0.006436762399971485,
-0.12171690911054611,
0.02350432053208351,
0.08950570225715637,
-0.00007678163092350587,
-0.09379716217517853,
0.08152900636196136,
-0.07605183124542236,
-0.01949211396276951,
-0.021253660321235657,
-0.0765935555100441,
0.028752928599715233,
0.02075762115418911,
-0.07241667807102203,
0.0030450790654867887,
0.030792895704507828,
0.008529079146683216,
-0.015946075320243835,
0.0917603150010109,
-0.0787631943821907,
0.025490665808320045,
-0.0946962907910347,
-0.1045740470290184,
0.02897842787206173,
-0.09482478350400925,
0.033921513706445694,
-0.09247273951768875,
-0.19394199550151825,
-0.024185366928577423,
0.0686044692993164,
-0.021704984828829765,
-0.0471469946205616,
-0.07331191003322601,
-0.06313978135585785,
0.01850917749106884,
-0.002557029016315937,
0.09711819887161255,
-0.06576484441757202,
0.09101833403110504,
0.027181919664144516,
0.08311154693365097,
-0.031909260898828506,
0.056395988911390305,
-0.11463139951229095,
0.004114741925150156,
-0.13587668538093567,
0.049997471272945404,
-0.047578345984220505,
0.07202952355146408,
-0.06358253955841064,
-0.11173929274082184,
0.015572934411466122,
-0.006378253921866417,
0.06443363428115845,
0.10948903113603592,
-0.19403833150863647,
-0.09464851766824722,
0.16610102355480194,
-0.07067948579788208,
-0.10680301487445831,
0.12891265749931335,
-0.06742480397224426,
0.06976816058158875,
0.06998448073863983,
0.17836694419384003,
0.05879383906722069,
-0.07357979565858841,
-0.016340071335434914,
0.011599463410675526,
0.050163306295871735,
-0.031123390421271324,
0.052572306245565414,
0.02624650113284588,
0.03453371301293373,
0.03880535066127777,
-0.01196232344955206,
0.07060033828020096,
-0.09326792508363724,
-0.10062386095523834,
-0.03427024930715561,
-0.09134482592344284,
0.04848787933588028,
0.09255674481391907,
0.06272212415933609,
-0.10717325657606125,
-0.073724165558815,
0.028808284550905228,
0.09422781318426132,
-0.06488901376724243,
0.028829196467995644,
-0.0596344880759716,
0.0622246228158474,
0.0027249802369624376,
-0.015573904849588871,
-0.17582817375659943,
0.013070456683635712,
0.006194745190441608,
0.027181124314665794,
0.006623989902436733,
0.03696899861097336,
0.06610392779111862,
0.04349454119801521,
-0.05583404377102852,
-0.024697057902812958,
-0.04591572657227516,
-0.002911053132265806,
-0.11217767000198364,
-0.222952201962471,
-0.01753520965576172,
-0.024884168058633804,
0.17654204368591309,
-0.21061822772026062,
0.04510215297341347,
-0.006178705487400293,
0.055832020938396454,
0.014611254446208477,
-0.01919422671198845,
-0.033304933458566666,
0.06556599587202072,
-0.054479777812957764,
-0.042237140238285065,
0.07778333127498627,
0.01036946102976799,
-0.08646371960639954,
-0.037160664796829224,
-0.10661646723747253,
0.14299476146697998,
0.13010098040103912,
-0.11318360269069672,
-0.07107231020927429,
-0.016780277714133263,
-0.06628571450710297,
-0.01899905502796173,
-0.03863293677568436,
0.03865957632660866,
0.19597330689430237,
-0.007075474597513676,
0.13832490146160126,
-0.06229453533887863,
-0.02452687919139862,
0.02397817187011242,
-0.04459702596068382,
0.005227315239608288,
0.13486889004707336,
0.12143418937921524,
-0.06046079099178314,
0.15041641891002655,
0.13522854447364807,
-0.08931641280651093,
0.16353555023670197,
-0.03590400516986847,
-0.05880224332213402,
-0.025102846324443817,
-0.04856446385383606,
-0.018963899463415146,
0.10565771162509918,
-0.18460632860660553,
-0.011765911243855953,
0.02307305857539177,
0.0011658030562102795,
0.006093000527471304,
-0.2260960191488266,
-0.05179370567202568,
0.04856545850634575,
-0.04331237077713013,
-0.006302523892372847,
-0.010050495155155659,
0.00567513657733798,
0.1050773411989212,
-0.0037724007852375507,
-0.08540020883083344,
0.030277138575911522,
-0.001458055805414915,
-0.08583482354879379,
0.20442630350589752,
-0.0918363556265831,
-0.17275619506835938,
-0.1108991801738739,
-0.07250522077083588,
-0.047569986432790756,
0.00643935427069664,
0.07166393846273422,
-0.11759302020072937,
-0.019171783700585365,
-0.07828955352306366,
0.0264158733189106,
0.011945844627916813,
0.020029593259096146,
0.028738269582390785,
-0.0024876517709344625,
0.047073788940906525,
-0.10855920612812042,
-0.019495608285069466,
-0.06457715481519699,
-0.04853179305791855,
0.054971616715192795,
0.019171450287103653,
0.11925762891769409,
0.16973093152046204,
-0.005927021149545908,
0.011735835112631321,
-0.03866672143340111,
0.22641333937644958,
-0.072452612221241,
-0.019722480326890945,
0.13654112815856934,
-0.012618577107787132,
0.05264660716056824,
0.11521118134260178,
0.06805716454982758,
-0.09154286980628967,
0.014333990402519703,
0.04578052833676338,
-0.037165895104408264,
-0.22064699232578278,
-0.04134365916252136,
-0.04843695089221001,
0.025485754013061523,
0.06965936720371246,
0.021163061261177063,
0.0463443286716938,
0.07631208002567291,
0.04127596318721771,
0.04955020546913147,
-0.04838881269097328,
0.05194804072380066,
0.1304904669523239,
0.018573161214590073,
0.10156048089265823,
-0.03687027096748352,
-0.05285428836941719,
0.05776938423514366,
-0.019366687163710594,
0.21297587454319,
0.001861072494648397,
0.14540186524391174,
0.05809827521443367,
0.16863800585269928,
-0.03068048506975174,
0.07258651405572891,
-0.014496046118438244,
-0.04065464809536934,
-0.03182343393564224,
-0.028970252722501755,
-0.06382753700017929,
0.03464755788445473,
-0.05723574012517929,
0.08380259573459625,
-0.13994887471199036,
0.01241863239556551,
0.06259757280349731,
0.2786177396774292,
0.027485299855470657,
-0.31776177883148193,
-0.11263156682252884,
0.005084906704723835,
-0.03786880522966385,
-0.005435932893306017,
0.02235650084912777,
0.09276892989873886,
-0.09591041505336761,
0.0349600613117218,
-0.06057644262909889,
0.08381333947181702,
-0.07126171886920929,
0.06418641656637192,
0.046352777630090714,
0.0721132755279541,
0.010117967613041401,
0.0875658169388771,
-0.2863155007362366,
0.26768237352371216,
-0.010248835198581219,
0.05953718349337578,
-0.08695551007986069,
-0.0006002221489325166,
0.06266553699970245,
0.06651800870895386,
0.06717213243246078,
-0.007731396239250898,
0.0020543483551591635,
-0.182090163230896,
-0.03915363922715187,
0.029793573543429375,
0.06323503702878952,
-0.03606581315398216,
0.08719656616449356,
-0.02526232600212097,
0.008744700811803341,
0.07756873965263367,
0.03433313965797424,
-0.04849676042795181,
-0.10168127715587616,
-0.010243501514196396,
0.033136118203401566,
-0.053475815802812576,
-0.05309470370411873,
-0.128390833735466,
-0.10907188057899475,
0.14396999776363373,
-0.003665628144517541,
-0.023246966302394867,
-0.10303903371095657,
0.0824100524187088,
0.04088747873902321,
-0.088455930352211,
0.027437539771199226,
0.009255582466721535,
0.08022436499595642,
0.020462913438677788,
-0.06966151297092438,
0.106208935379982,
-0.07627078890800476,
-0.1729302853345871,
-0.06904780864715576,
0.09420742839574814,
0.05349254980683327,
0.07726606726646423,
-0.0060472674667835236,
-0.010492062196135521,
-0.04896660894155502,
-0.08469203859567642,
0.03883951157331467,
0.030602402985095978,
0.060137249529361725,
0.015432341024279594,
-0.05169805884361267,
0.006301424000412226,
-0.07002141326665878,
-0.037205908447504044,
0.19957755506038666,
0.23401984572410583,
-0.088113933801651,
0.030145883560180664,
0.032210662961006165,
-0.0741310715675354,
-0.1937704235315323,
0.04973391443490982,
0.059615232050418854,
0.009831813164055347,
0.041455693542957306,
-0.19466006755828857,
0.12152144312858582,
0.08430802822113037,
-0.01133981067687273,
0.09791947156190872,
-0.30057990550994873,
-0.11297563463449478,
0.13992652297019958,
0.1440286636352539,
0.12114910036325455,
-0.14069239795207977,
-0.002414488699287176,
-0.030839765444397926,
-0.12448009848594666,
0.11540801078081131,
-0.08373696357011795,
0.1239585131406784,
-0.024087201803922653,
0.11903462558984756,
0.008925092406570911,
-0.046338435262441635,
0.11534544825553894,
0.021625714376568794,
0.09919611364603043,
-0.07103241235017776,
-0.0303041934967041,
0.022858066484332085,
-0.040134724229574203,
0.02944108098745346,
-0.09900438040494919,
0.017833665013313293,
-0.11792459338903427,
-0.03237378969788551,
-0.08853314816951752,
0.03513661399483681,
-0.040174245834350586,
-0.07397525757551193,
-0.050041165202856064,
0.02816113457083702,
0.07622185349464417,
-0.004667900502681732,
0.08376561850309372,
0.01950220949947834,
0.11442000418901443,
0.09925757348537445,
0.09666679054498672,
-0.05519923195242882,
-0.07141568511724472,
-0.022381464019417763,
-0.009976433590054512,
0.048478029668331146,
-0.14822426438331604,
0.01608099974691868,
0.13894839584827423,
0.019480889663100243,
0.1673259288072586,
0.08592981100082397,
-0.0385739766061306,
0.017456399276852608,
0.06034190207719803,
-0.1507411003112793,
-0.08740292489528656,
-0.02028888277709484,
-0.07070736587047577,
-0.12298166751861572,
0.031328827142715454,
0.08302098512649536,
-0.07266725599765778,
-0.00016929450794123113,
-0.015392606146633625,
0.01725717820227146,
-0.04057107865810394,
0.16434220969676971,
0.048006441444158554,
0.029739059507846832,
-0.10324610769748688,
0.07677920162677765,
0.02069034054875374,
-0.10905935615301132,
0.029330000281333923,
0.07366786152124405,
-0.07665736973285675,
-0.05674770846962929,
0.06659182161092758,
0.21424585580825806,
-0.060438938438892365,
-0.049563173204660416,
-0.1491101235151291,
-0.12783226370811462,
0.08513522893190384,
0.1479852944612503,
0.1134083941578865,
0.008781511336565018,
-0.08621153980493546,
0.024828435853123665,
-0.11682089418172836,
0.08967143297195435,
0.05856127291917801,
0.04151647537946701,
-0.13298381865024567,
0.12173733115196228,
0.00987596157938242,
0.04105108231306076,
-0.020841378718614578,
0.010980355553328991,
-0.09201997518539429,
0.008116669952869415,
-0.11870959401130676,
-0.027281617745757103,
-0.04016470909118652,
0.011824891902506351,
0.0021165378857403994,
-0.04331228882074356,
-0.0449797622859478,
0.003677732776850462,
-0.11604902893304825,
-0.015674734488129616,
0.03586804121732712,
0.07664954662322998,
-0.113495834171772,
-0.03808976709842682,
0.028106754645705223,
-0.06515103578567505,
0.09211961925029755,
0.06284593045711517,
0.013682783581316471,
0.05738198012113571,
-0.16451308131217957,
0.02591823786497116,
0.09183235466480255,
0.015367215499281883,
0.05428994446992874,
-0.08173952251672745,
-0.01165669783949852,
-0.010551336221396923,
0.039947234094142914,
0.016846122220158577,
0.08105680346488953,
-0.12655404210090637,
0.018837958574295044,
0.005064732860773802,
-0.08740627020597458,
-0.06847129762172699,
0.03200390934944153,
0.08091627061367035,
0.009806549176573753,
0.19735103845596313,
-0.07809913903474808,
0.046744346618652344,
-0.21774250268936157,
0.007651552092283964,
0.00039223834755830467,
-0.10069233924150467,
-0.12886367738246918,
-0.07508815824985504,
0.05666074901819229,
-0.05609523504972458,
0.1320609599351883,
0.04614754393696785,
0.009946395643055439,
0.010845249518752098,
-0.009018459357321262,
0.023153048008680344,
0.003420208115130663,
0.18353857100009918,
0.035507675260305405,
-0.05026087164878845,
0.06071584299206734,
0.05424007400870323,
0.11871401220560074,
0.12723422050476074,
0.19787128269672394,
0.1401710957288742,
0.025031769648194313,
0.10930082201957703,
0.03284158185124397,
-0.03587397560477257,
-0.14979085326194763,
0.030222611501812935,
-0.0520317442715168,
0.11461912095546722,
-0.017922502011060715,
0.24583737552165985,
0.06318626552820206,
-0.15751786530017853,
0.0627346932888031,
-0.0622653029859066,
-0.080485038459301,
-0.10323148965835571,
-0.062278345227241516,
-0.07988351583480835,
-0.14291183650493622,
0.0025185495615005493,
-0.1344880908727646,
0.005178164690732956,
0.0950811505317688,
0.010578589513897896,
-0.041072338819503784,
0.13897496461868286,
0.01453063078224659,
0.020731834694743156,
0.08987818658351898,
0.008688630536198616,
-0.06495719403028488,
-0.13327042758464813,
-0.0563986711204052,
-0.012872141785919666,
-0.01658868044614792,
0.040306270122528076,
-0.050967987626791,
-0.06254465132951736,
0.0255191158503294,
-0.017901567742228508,
-0.1019640564918518,
0.008450163528323174,
0.00674017146229744,
0.06198660656809807,
0.045322615653276443,
0.0007392247207462788,
0.02236340567469597,
0.0022965685930103064,
0.19070175290107727,
-0.07466679811477661,
-0.02867077849805355,
-0.10520850121974945,
0.22406406700611115,
0.02177613228559494,
-0.014602554962038994,
0.03250817582011223,
-0.0718652755022049,
-0.005044568330049515,
0.24970677495002747,
0.20883415639400482,
-0.08701977133750916,
-0.005430365912616253,
0.0030840749386698008,
0.002857531188055873,
-0.046956535428762436,
0.0954923927783966,
0.15175102651119232,
0.02246778830885887,
-0.09816353768110275,
-0.02384945936501026,
-0.058102697134017944,
-0.023486129939556122,
-0.016953278332948685,
0.05730053782463074,
0.062264759093523026,
0.012224104255437851,
-0.04443337395787239,
0.050532374531030655,
-0.08826623111963272,
-0.10048242658376694,
0.07598186284303665,
-0.21897344291210175,
-0.15388427674770355,
-0.017186596989631653,
0.09836289286613464,
0.028858967125415802,
0.07270942628383636,
-0.017048347741365433,
-0.0037730636540800333,
0.1151106208562851,
-0.02014276571571827,
-0.11828415095806122,
-0.07104312628507614,
0.09810949116945267,
-0.13000449538230896,
0.20294146239757538,
-0.06483820080757141,
0.0402655303478241,
0.12451666593551636,
0.07169801741838455,
-0.05337971821427345,
0.07340917736291885,
0.048565737903118134,
-0.055177778005599976,
0.006028510630130768,
0.10180963575839996,
-0.0313725620508194,
0.07693233340978622,
0.04867885634303093,
-0.15307050943374634,
0.025778576731681824,
-0.0403456836938858,
-0.06710763275623322,
-0.04520198702812195,
-0.00809670053422451,
-0.06513381749391556,
0.12103945016860962,
0.22164419293403625,
-0.02433830127120018,
-0.002707727486267686,
-0.07023309171199799,
0.006103829480707645,
0.04813467711210251,
0.00954868271946907,
-0.05603432282805443,
-0.20643669366836548,
0.012327268719673157,
0.06683766096830368,
-0.013608798384666443,
-0.2591656446456909,
-0.10369356721639633,
0.0030699747148901224,
-0.06895381212234497,
-0.09046507626771927,
0.061003975570201874,
0.06755086034536362,
0.059811607003211975,
-0.04745471104979515,
-0.05766627937555313,
-0.06149698793888092,
0.1691356599330902,
-0.13927510380744934,
-0.08477252721786499
] |
null | null |
transformers
|
# BanglaBERT
This repository contains the pretrained discriminator checkpoint of the model **BanglaBERT**. This is an [ELECTRA](https://openreview.net/pdf?id=r1xMH1BtvB) discriminator model pretrained with the Replaced Token Detection (RTD) objective. Finetuned models using this checkpoint achieve state-of-the-art results on many of the NLP tasks in bengali.
For finetuning on different downstream tasks such as `Sentiment classification`, `Named Entity Recognition`, `Natural Language Inference` etc., refer to the scripts in the official GitHub [repository](https://github.com/csebuetnlp/banglabert).
**Note**: This model was pretrained using a specific normalization pipeline available [here](https://github.com/csebuetnlp/normalizer). All finetuning scripts in the official GitHub repository uses this normalization by default. If you need to adapt the pretrained model for a different task make sure the text units are normalized using this pipeline before tokenizing to get best results. A basic example is given below:
## Using this model as a discriminator in `transformers` (tested on 4.11.0.dev0)
```python
from transformers import AutoModelForPreTraining, AutoTokenizer
from normalizer import normalize # pip install git+https://github.com/csebuetnlp/normalizer
import torch
model = AutoModelForPreTraining.from_pretrained("csebuetnlp/banglabert")
tokenizer = AutoTokenizer.from_pretrained("csebuetnlp/banglabert")
original_sentence = "আমি কৃতজ্ঞ কারণ আপনি আমার জন্য অনেক কিছু করেছেন।"
fake_sentence = "আমি হতাশ কারণ আপনি আমার জন্য অনেক কিছু করেছেন।"
fake_sentence = normalize(fake_sentence) # this normalization step is required before tokenizing the text
fake_tokens = tokenizer.tokenize(fake_sentence)
fake_inputs = tokenizer.encode(fake_sentence, return_tensors="pt")
discriminator_outputs = model(fake_inputs).logits
predictions = torch.round((torch.sign(discriminator_outputs) + 1) / 2)
[print("%7s" % token, end="") for token in fake_tokens]
print("\n" + "-" * 50)
[print("%7s" % int(prediction), end="") for prediction in predictions.squeeze().tolist()[1:-1]]
print("\n" + "-" * 50)
```
## Benchmarks
* Zero-shot cross-lingual transfer-learning
| Model | Params | SC (macro-F1) | NLI (accuracy) | NER (micro-F1) | QA (EM/F1) | BangLUE score |
|----------------|-----------|-----------|-----------|-----------|-----------|-----------|
|[mBERT](https://huggingface.co/bert-base-multilingual-cased) | 180M | 27.05 | 62.22 | 39.27 | 59.01/64.18 | 50.35 |
|[XLM-R (base)](https://huggingface.co/xlm-roberta-base) | 270M | 42.03 | 72.18 | 45.37 | 55.03/61.83 | 55.29 |
|[XLM-R (large)](https://huggingface.co/xlm-roberta-large) | 550M | 49.49 | 78.13 | 56.48 | 71.13/77.70 | 66.59 |
|[BanglishBERT](https://huggingface.co/csebuetnlp/banglishbert) | 110M | 48.39 | 75.26 | 55.56 | 72.87/78.63 | 66.14 |
* Supervised fine-tuning
| Model | Params | SC (macro-F1) | NLI (accuracy) | NER (micro-F1) | QA (EM/F1) | BangLUE score |
|----------------|-----------|-----------|-----------|-----------|-----------|-----------|
|[mBERT](https://huggingface.co/bert-base-multilingual-cased) | 180M | 67.59 | 75.13 | 68.97 | 67.12/72.64 | 70.29 |
|[XLM-R (base)](https://huggingface.co/xlm-roberta-base) | 270M | 69.54 | 78.46 | 73.32 | 68.09/74.27 | 72.82 |
|[XLM-R (large)](https://huggingface.co/xlm-roberta-large) | 550M | 70.97 | 82.40 | 78.39 | 73.15/79.06 | 76.79 |
|[sahajBERT](https://huggingface.co/neuropark/sahajBERT) | 18M | 71.12 | 76.92 | 70.94 | 65.48/70.69 | 71.03 |
|[BanglishBERT](https://huggingface.co/csebuetnlp/banglishbert) | 110M | 70.61 | 80.95 | 76.28 | 72.43/78.40 | 75.73 |
|[BanglaBERT](https://huggingface.co/csebuetnlp/banglabert) | 110M | 72.89 | 82.80 | 77.78 | 72.63/79.34 | **77.09** |
The benchmarking datasets are as follows:
* **SC:** **[Sentiment Classification](https://aclanthology.org/2021.findings-emnlp.278)**
* **NER:** **[Named Entity Recognition](https://multiconer.github.io/competition)**
* **NLI:** **[Natural Language Inference](https://github.com/csebuetnlp/banglabert/#datasets)**
* **QA:** **[Question Answering](https://github.com/csebuetnlp/banglabert/#datasets)**
## Citation
If you use this model, please cite the following paper:
```
@inproceedings{bhattacharjee-etal-2022-banglabert,
title = "{B}angla{BERT}: Language Model Pretraining and Benchmarks for Low-Resource Language Understanding Evaluation in {B}angla",
author = "Bhattacharjee, Abhik and
Hasan, Tahmid and
Ahmad, Wasi and
Mubasshir, Kazi Samin and
Islam, Md Saiful and
Iqbal, Anindya and
Rahman, M. Sohel and
Shahriyar, Rifat",
booktitle = "Findings of the Association for Computational Linguistics: NAACL 2022",
month = jul,
year = "2022",
address = "Seattle, United States",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2022.findings-naacl.98",
pages = "1318--1327",
abstract = "In this work, we introduce BanglaBERT, a BERT-based Natural Language Understanding (NLU) model pretrained in Bangla, a widely spoken yet low-resource language in the NLP literature. To pretrain BanglaBERT, we collect 27.5 GB of Bangla pretraining data (dubbed {`}Bangla2B+{'}) by crawling 110 popular Bangla sites. We introduce two downstream task datasets on natural language inference and question answering and benchmark on four diverse NLU tasks covering text classification, sequence labeling, and span prediction. In the process, we bring them under the first-ever Bangla Language Understanding Benchmark (BLUB). BanglaBERT achieves state-of-the-art results outperforming multilingual and monolingual models. We are making the models, datasets, and a leaderboard publicly available at \url{https://github.com/csebuetnlp/banglabert} to advance Bangla NLP.",
}
```
If you use the normalization module, please cite the following paper:
```
@inproceedings{hasan-etal-2020-low,
title = "Not Low-Resource Anymore: Aligner Ensembling, Batch Filtering, and New Datasets for {B}engali-{E}nglish Machine Translation",
author = "Hasan, Tahmid and
Bhattacharjee, Abhik and
Samin, Kazi and
Hasan, Masum and
Basak, Madhusudan and
Rahman, M. Sohel and
Shahriyar, Rifat",
booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)",
month = nov,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://www.aclweb.org/anthology/2020.emnlp-main.207",
doi = "10.18653/v1/2020.emnlp-main.207",
pages = "2612--2623",
abstract = "Despite being the seventh most widely spoken language in the world, Bengali has received much less attention in machine translation literature due to being low in resources. Most publicly available parallel corpora for Bengali are not large enough; and have rather poor quality, mostly because of incorrect sentence alignments resulting from erroneous sentence segmentation, and also because of a high volume of noise present in them. In this work, we build a customized sentence segmenter for Bengali and propose two novel methods for parallel corpus creation on low-resource setups: aligner ensembling and batch filtering. With the segmenter and the two methods combined, we compile a high-quality Bengali-English parallel corpus comprising of 2.75 million sentence pairs, more than 2 million of which were not available before. Training on neural models, we achieve an improvement of more than 9 BLEU score over previous approaches to Bengali-English machine translation. We also evaluate on a new test set of 1000 pairs made with extensive quality control. We release the segmenter, parallel corpus, and the evaluation set, thus elevating Bengali from its low-resource status. To the best of our knowledge, this is the first ever large scale study on Bengali-English machine translation. We believe our study will pave the way for future research on Bengali-English machine translation as well as other low-resource languages. Our data and code are available at https://github.com/csebuetnlp/banglanmt.",
}
```
|
{"language": ["bn"], "licenses": ["cc-by-nc-sa-4.0"]}
| null |
csebuetnlp/banglabert
|
[
"transformers",
"pytorch",
"electra",
"pretraining",
"bn",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"bn"
] |
TAGS
#transformers #pytorch #electra #pretraining #bn #endpoints_compatible #has_space #region-us
|
BanglaBERT
==========
This repository contains the pretrained discriminator checkpoint of the model BanglaBERT. This is an ELECTRA discriminator model pretrained with the Replaced Token Detection (RTD) objective. Finetuned models using this checkpoint achieve state-of-the-art results on many of the NLP tasks in bengali.
For finetuning on different downstream tasks such as 'Sentiment classification', 'Named Entity Recognition', 'Natural Language Inference' etc., refer to the scripts in the official GitHub repository.
Note: This model was pretrained using a specific normalization pipeline available here. All finetuning scripts in the official GitHub repository uses this normalization by default. If you need to adapt the pretrained model for a different task make sure the text units are normalized using this pipeline before tokenizing to get best results. A basic example is given below:
Using this model as a discriminator in 'transformers' (tested on 4.11.0.dev0)
-----------------------------------------------------------------------------
Benchmarks
----------
* Zero-shot cross-lingual transfer-learning
* Supervised fine-tuning
The benchmarking datasets are as follows:
* SC: Sentiment Classification
* NER: Named Entity Recognition
* NLI: Natural Language Inference
* QA: Question Answering
If you use this model, please cite the following paper:
If you use the normalization module, please cite the following paper:
|
[] |
[
"TAGS\n#transformers #pytorch #electra #pretraining #bn #endpoints_compatible #has_space #region-us \n"
] |
[
33
] |
[
"passage: TAGS\n#transformers #pytorch #electra #pretraining #bn #endpoints_compatible #has_space #region-us \n"
] |
[
-0.0004112394235562533,
0.041354142129421234,
-0.008528155274689198,
-0.03102412074804306,
0.07681901752948761,
0.02007954567670822,
0.012598119676113129,
0.0646592229604721,
0.04147854819893837,
0.043214455246925354,
0.18332809209823608,
0.09383107721805573,
-0.06503050029277802,
0.05114077031612396,
-0.01736832968890667,
-0.2509968876838684,
0.08372611552476883,
0.06038285046815872,
-0.08393190056085587,
0.10378177464008331,
0.047619886696338654,
-0.14532780647277832,
0.03572488948702812,
-0.017941409721970558,
-0.09317171573638916,
0.04069788381457329,
-0.04251347482204437,
-0.0805816501379013,
0.1405310183763504,
0.018291758373379707,
0.19978167116641998,
0.035918258130550385,
-0.08932095021009445,
-0.15195386111736298,
0.04345838353037834,
0.005574648734182119,
-0.08992258459329605,
0.0498429574072361,
-0.03951637074351311,
-0.06969453394412994,
0.11775260418653488,
0.0358562096953392,
0.003754012519493699,
0.007837334647774696,
-0.20779648423194885,
-0.14565764367580414,
-0.03541754558682442,
0.028747469186782837,
-0.006246470846235752,
0.05760728940367699,
-0.020497648045420647,
0.16387692093849182,
-0.1811470091342926,
0.033113595098257065,
0.16855975985527039,
-0.3143838346004486,
-0.008360587060451508,
0.10539181530475616,
0.14951273798942566,
0.08691398799419403,
-0.017013538628816605,
0.07504642754793167,
0.0594821572303772,
0.011466129682958126,
0.020147845149040222,
-0.05267513915896416,
-0.014599094167351723,
0.10996408760547638,
-0.14509563148021698,
-0.11043968796730042,
0.254686176776886,
-0.029262498021125793,
0.07830287516117096,
0.013772690668702126,
-0.07884658128023148,
-0.14590929448604584,
0.04520614072680473,
0.0026330274995416403,
-0.0049156006425619125,
0.038008399307727814,
0.047033220529556274,
-0.019544513896107674,
-0.15124498307704926,
0.06107987463474274,
-0.18533262610435486,
0.2643808126449585,
0.0003308326122350991,
0.0615878626704216,
-0.1735832691192627,
0.07805110514163971,
0.0036724291276186705,
-0.0839289128780365,
0.03248788043856621,
-0.10621437430381775,
0.02779177948832512,
0.02349206991493702,
-0.1280091106891632,
0.07184570282697678,
0.05408177524805069,
0.14981554448604584,
0.015863966196775436,
0.009631853550672531,
0.07638216763734818,
0.15075933933258057,
0.07397035509347916,
0.05969224125146866,
-0.05560861900448799,
0.00012562645133584738,
-0.03007574938237667,
-0.060922712087631226,
0.028188431635499,
-0.03684408962726593,
-0.11114306002855301,
-0.08288317918777466,
0.005051752086728811,
0.05996648594737053,
0.03790896385908127,
0.01578747294843197,
-0.08183763921260834,
0.05077175796031952,
-0.0016833044355735183,
-0.01786896027624607,
-0.026684697717428207,
-0.020236963406205177,
0.005471234209835529,
0.06323514878749847,
0.020098969340324402,
0.0027798283845186234,
0.036347974091768265,
0.049077894538640976,
-0.1082748994231224,
-0.014954476617276669,
-0.0567358173429966,
-0.06285220384597778,
0.05307726189494133,
-0.14453305304050446,
0.09401827305555344,
-0.16278260946273804,
0.006573964841663837,
0.009190497919917107,
0.07161904871463776,
-0.0009785270085558295,
0.002985185943543911,
0.04520949721336365,
-0.05650826916098595,
0.006915666162967682,
-0.07078620791435242,
-0.02853657677769661,
-0.07965071499347687,
0.08913742750883102,
-0.0038500959053635597,
0.10972831398248672,
-0.11476588994264603,
0.05240604653954506,
-0.0772668719291687,
0.04463693127036095,
-0.10606101900339127,
-0.017576292157173157,
-0.0167388953268528,
0.11604676395654678,
0.03217073529958725,
-0.059753261506557465,
-0.15440422296524048,
0.07252475619316101,
-0.010512527078390121,
0.13824570178985596,
-0.07674701511859894,
-0.06294940412044525,
0.15393252670764923,
-0.07869051396846771,
-0.13528120517730713,
0.034883566200733185,
-0.005420091561973095,
0.008535950444638729,
-0.012329958379268646,
0.18301606178283691,
-0.037178490310907364,
-0.07195069640874863,
-0.024386776611208916,
0.06593277305364609,
-0.10258170962333679,
-0.07009898126125336,
0.05975327640771866,
0.04804649204015732,
-0.012106447480618954,
-0.02111349068582058,
0.05216153711080551,
0.08940333127975464,
-0.08064303547143936,
-0.03747595474123955,
-0.016401177272200584,
-0.03268832713365555,
0.10689026117324829,
0.06618049740791321,
0.053169745951890945,
-0.09390737861394882,
-0.0634048730134964,
0.06981851905584335,
0.029447780922055244,
0.06398533284664154,
0.04899004474282265,
-0.049374427646398544,
0.09855900704860687,
-0.12223415076732635,
-0.04745597392320633,
-0.19901502132415771,
-0.09100515395402908,
-0.08051510900259018,
0.07951728999614716,
0.023484904319047928,
0.31558695435523987,
0.06812369078397751,
-0.10139766335487366,
-0.00983357708901167,
-0.04641420766711235,
0.039105188101530075,
0.0354425348341465,
-0.027270440012216568,
-0.06384865194559097,
-0.01288951002061367,
-0.09591913223266602,
-0.03103673830628395,
-0.08334320038557053,
0.04777867719531059,
0.08772335946559906,
0.1395852416753769,
-0.005023861303925514,
0.049047261476516724,
0.02013910561800003,
0.048047978430986404,
-0.05950259044766426,
0.01673496514558792,
0.058578941971063614,
-0.015111393295228481,
-0.08375831693410873,
0.08678558468818665,
-0.13623882830142975,
0.33996742963790894,
0.205155611038208,
-0.2672557234764099,
0.002807808807119727,
0.053086791187524796,
-0.052053097635507584,
0.03480076789855957,
0.08317737281322479,
-0.012364029884338379,
0.026694465428590775,
-0.013916105963289738,
0.08156011253595352,
-0.024807492271065712,
-0.054136257618665695,
-0.013335944153368473,
-0.05617549642920494,
-0.10000460594892502,
0.09495817869901657,
0.05787186324596405,
-0.10804971307516098,
0.18190988898277283,
0.3140229880809784,
-0.02174604870378971,
0.16751913726329803,
0.013977342285215855,
-0.012260744348168373,
0.0006993251736275852,
-0.03958621248602867,
-0.08525049686431885,
0.10000360757112503,
-0.26294824481010437,
-0.0648106038570404,
0.06170731037855148,
-0.012238415889441967,
0.0876331701874733,
-0.14857515692710876,
-0.0970139354467392,
0.033501897007226944,
0.06423191726207733,
-0.05850604921579361,
0.11584218591451645,
0.005187953822314739,
0.08774769306182861,
0.020191524177789688,
-0.06272605806589127,
0.05924614146351814,
0.0031453159172087908,
-0.015201346017420292,
0.12217328697443008,
-0.10419262945652008,
-0.20651744306087494,
-0.04602757841348648,
-0.023447461426258087,
0.054314497858285904,
0.011447029188275337,
0.07058338075876236,
-0.06666956841945648,
-0.0005737446481361985,
0.04733463004231453,
0.03820301592350006,
-0.17506633698940277,
0.04615093395113945,
-0.025863971561193466,
0.040316272526979446,
-0.09765812009572983,
-0.09337668120861053,
-0.0578727200627327,
-0.06581471115350723,
-0.024746540933847427,
0.11771070957183838,
-0.02752642333507538,
0.07919833809137344,
0.1304364949464798,
0.010744374245405197,
0.034686166793107986,
-0.0037712978664785624,
0.22229883074760437,
-0.11996106803417206,
-0.03056221641600132,
0.14757944643497467,
-0.00862937979400158,
0.048175305128097534,
0.15896180272102356,
0.040623728185892105,
-0.05521335452795029,
-0.006256227381527424,
-0.06589237600564957,
-0.10465794056653976,
-0.1532374918460846,
-0.11455032229423523,
-0.12561888992786407,
-0.025711558759212494,
-0.025202125310897827,
0.05679804086685181,
0.055055927485227585,
0.07254693657159805,
0.0174062829464674,
-0.12727132439613342,
-0.08102244138717651,
0.03782955929636955,
0.18572333455085754,
-0.06862401217222214,
0.09615868330001831,
-0.06657487899065018,
-0.09396278858184814,
0.047296226024627686,
0.06638147681951523,
0.1123565211892128,
0.11809718608856201,
-0.04697784036397934,
0.08952423185110092,
0.13554565608501434,
0.14420604705810547,
0.08188251405954361,
0.0027293458115309477,
-0.057943664491176605,
-0.026515068486332893,
0.0139937624335289,
-0.05774539336562157,
0.08537470549345016,
0.1401548534631729,
-0.10509054362773895,
-0.03348548337817192,
-0.25013354420661926,
0.06178712844848633,
0.005267721600830555,
0.09275928139686584,
-0.16262611746788025,
-0.017163557931780815,
0.10569140315055847,
-0.023304540663957596,
-0.06608039885759354,
0.06885390728712082,
0.08767291903495789,
-0.07681910693645477,
0.03659747913479805,
0.030028099194169044,
0.07109272480010986,
0.04431986063718796,
0.09253036230802536,
-0.10267242044210434,
-0.17551718652248383,
0.020433368161320686,
0.04239816218614578,
-0.2604841887950897,
0.26600635051727295,
-0.023440929129719734,
-0.11570149660110474,
-0.025261640548706055,
-0.05132715031504631,
-0.011573508381843567,
0.1544002890586853,
0.09223347902297974,
0.021953308954834938,
-0.12646235525608063,
-0.07862329483032227,
0.10081269592046738,
0.016526537016034126,
0.09337138384580612,
-0.05808178335428238,
0.0037325306329876184,
0.02074800617992878,
0.03850021958351135,
-0.0031783725135028362,
0.12884917855262756,
0.0013324243482202291,
-0.1212742030620575,
0.050439853221178055,
0.05078117176890373,
0.04463677108287811,
-0.006078772246837616,
-0.03601044416427612,
-0.09443075209856033,
0.06372044235467911,
-0.11305639892816544,
-0.04032048583030701,
-0.058239858597517014,
-0.10084503889083862,
0.1518031656742096,
-0.04940143600106239,
0.08152024447917938,
-0.03635422885417938,
-0.03719133138656616,
-0.10007737576961517,
-0.09648727625608444,
0.14003127813339233,
-0.10627901554107666,
-0.003970146644860506,
-0.04681471735239029,
0.12069917470216751,
-0.016997195780277252,
0.06503621488809586,
-0.03367353603243828,
0.08610087633132935,
-0.17201074957847595,
-0.061618171632289886,
0.04618167504668236,
-0.07060904055833817,
0.10353291034698486,
0.04016603156924248,
0.004439698066562414,
0.05417109653353691,
0.08670372515916824,
-0.007406284566968679,
0.19429141283035278,
0.2469898909330368,
-0.07810001820325851,
0.08597534149885178,
0.11952795833349228,
0.005204856861382723,
-0.2476789653301239,
-0.04672439396381378,
-0.16207727789878845,
0.002510861726477742,
0.020795919001102448,
-0.1389521062374115,
0.0340413935482502,
0.05091669037938118,
-0.06674852967262268,
0.10389768332242966,
-0.2622458338737488,
-0.04165757820010185,
0.14714911580085754,
-0.06934729218482971,
0.45425960421562195,
-0.13867256045341492,
-0.003335164161399007,
0.0226359311491251,
-0.15998339653015137,
0.11179657280445099,
-0.028145764023065567,
0.07109943777322769,
-0.021546967327594757,
0.04505598172545433,
0.02337975800037384,
-0.06924790143966675,
0.17911212146282196,
-0.01044986117631197,
0.027927592396736145,
-0.1006297692656517,
-0.15987257659435272,
0.08414573967456818,
-0.035950928926467896,
-0.04594564810395241,
0.014850537292659283,
-0.02157275192439556,
-0.2374611347913742,
0.015326544642448425,
-0.14154188334941864,
0.0829271599650383,
0.017750253900885582,
-0.05502743273973465,
-0.06905731558799744,
0.010341349989175797,
-0.016178030520677567,
0.035247739404439926,
0.29359692335128784,
-0.06546547263860703,
0.19652362167835236,
0.0784636065363884,
0.10050506889820099,
-0.15145273506641388,
-0.0012021649163216352,
0.015278860926628113,
-0.02782091312110424,
0.10278802365064621,
-0.12286192923784256,
-0.0009501700988039374,
0.12574073672294617,
-0.030363017693161964,
0.011328809894621372,
0.1143098920583725,
-0.011135652661323547,
0.03182440251111984,
0.1584305316209793,
-0.20599408447742462,
-0.09467098116874695,
-0.021394137293100357,
0.022021712735295296,
0.100815050303936,
0.07741749286651611,
0.08930154889822006,
0.006564377807080746,
-0.019130578264594078,
-0.0024126730859279633,
-0.05760333314538002,
-0.10254362225532532,
0.005163176916539669,
0.10288561135530472,
0.06162402406334877,
-0.07596814632415771,
0.029851263388991356,
0.029713930562138557,
-0.22709839046001434,
-0.01061203796416521,
0.09122344851493835,
-0.060750652104616165,
-0.15511764585971832,
-0.1457691341638565,
0.0071221888065338135,
-0.1545080691576004,
-0.021792348474264145,
0.005229126662015915,
-0.09271065145730972,
0.06839679926633835,
0.2846852242946625,
0.10493139922618866,
0.053527623414993286,
0.01305050402879715,
0.019522447139024734,
0.05965500324964523,
-0.05649113655090332,
-0.020886793732643127,
0.014745530672371387,
-0.09378314018249512,
0.04321848601102829,
-0.019735414534807205,
0.14989104866981506,
-0.09426362812519073,
-0.06038996949791908,
-0.14122411608695984,
0.06486719101667404,
-0.08878141641616821,
-0.13833411037921906,
-0.09733043611049652,
-0.08664953708648682,
0.007047866005450487,
-0.12046467512845993,
-0.06667592376470566,
-0.05373881012201309,
-0.13978376984596252,
0.07399661093950272,
0.027466297149658203,
0.03951488062739372,
-0.054180826991796494,
-0.017092091962695122,
0.1238366961479187,
-0.035174816846847534,
0.10916559398174286,
0.10287265479564667,
-0.04405292123556137,
0.06832891702651978,
0.0008871106547303498,
-0.10612627863883972,
0.09389644116163254,
-0.011457288637757301,
0.04233083128929138,
0.06217039003968239,
-0.02167568914592266,
-0.005965042859315872,
0.030982255935668945,
0.04551161825656891,
-0.04783423990011215,
-0.0701371431350708,
0.039051853120326996,
0.04166662320494652,
-0.11476851999759674,
-0.018522121012210846,
-0.08909609913825989,
0.1428602784872055,
0.008310544304549694,
0.0758599042892456,
0.047473419457674026,
0.02336987294256687,
-0.06503959000110626,
0.04340346157550812,
-0.021439045667648315,
-0.16291072964668274,
0.053254906088113785,
-0.04498935118317604,
0.010616892017424107,
-0.023154599592089653,
0.2873225510120392,
-0.02796005830168724,
-0.06096808984875679,
0.04890015348792076,
0.07178494334220886,
-0.05299663171172142,
0.041642915457487106,
0.1934858113527298,
0.11185499280691147,
-0.05809216946363449,
-0.10931473970413208,
0.0647563636302948,
0.012289133854210377,
-0.06262043863534927,
0.10341612994670868,
0.1374136507511139,
0.09263501316308975,
0.10938360542058945,
-0.037829913198947906,
0.013174560852348804,
-0.07008972018957138,
-0.12207391113042831,
0.03411508724093437,
0.011765895411372185,
-0.04041648283600807,
0.08072693645954132,
0.21589288115501404,
-0.02846800908446312,
0.03526124730706215,
-0.042620472609996796,
0.04816947504878044,
-0.11881574243307114,
-0.09438898414373398,
-0.009078496135771275,
-0.08662469685077667,
0.0022848232183605433,
-0.07049301266670227,
0.016432665288448334,
0.24730068445205688,
0.03331257775425911,
-0.012623870745301247,
0.1030958890914917,
0.10206084698438644,
-0.04290137439966202,
0.003899678122252226,
0.014852939173579216,
0.06601212173700333,
-0.05442531779408455,
-0.010322327725589275,
-0.09317927807569504,
-0.06140200421214104,
-0.06263742595911026,
0.02630556933581829,
-0.13982219994068146,
-0.043467819690704346,
-0.1103200912475586,
-0.09290221333503723,
-0.0724717527627945,
0.07025430351495743,
0.0006160566699691117,
0.08965422958135605,
-0.03241516649723053,
0.010654966346919537,
-0.0060256412252783775,
0.2343865931034088,
-0.08350787311792374,
-0.009714961051940918,
0.0026803791988641024,
0.07741864025592804,
0.02488582581281662,
0.12121956795454025,
-0.045554663985967636,
0.00017290985852014273,
-0.11001236736774445,
0.20291759073734283,
0.2568155527114868,
-0.08930061012506485,
0.08132679760456085,
0.06461839377880096,
0.03932706266641617,
0.06092759221792221,
0.01150329690426588,
0.11975438892841339,
0.20511725544929504,
-0.15357105433940887,
-0.034683819860219955,
-0.06375471502542496,
0.026345783844590187,
-0.051612306386232376,
0.030848387628793716,
0.03294576704502106,
-0.07466058433055878,
-0.06982763111591339,
0.044428497552871704,
-0.1759749948978424,
0.05744984373450279,
0.043016642332077026,
-0.25491923093795776,
-0.0645657330751419,
-0.010884488932788372,
0.1871432065963745,
-0.0246000736951828,
0.11718714237213135,
-0.05348975583910942,
-0.16744530200958252,
-0.00975964404642582,
0.02840207703411579,
-0.19237692654132843,
-0.07878107577562332,
0.14237932860851288,
0.07446548342704773,
0.012301168404519558,
-0.06267943978309631,
-0.02141178771853447,
0.09839288890361786,
0.04352439567446709,
-0.037776343524456024,
0.021074019372463226,
0.07845572382211685,
-0.03732490539550781,
-0.13300707936286926,
0.017899388447403908,
0.0317610539495945,
-0.040884777903556824,
0.09369798749685287,
-0.2153923064470291,
0.05646141618490219,
-0.014853804372251034,
-0.014459695667028427,
0.009441089816391468,
0.03283824771642685,
-0.013573242351412773,
0.07440705597400665,
0.021787064149975777,
-0.019397106021642685,
-0.04749598354101181,
-0.02434224635362625,
-0.040392909198999405,
0.06951574236154556,
0.04201528802514076,
-0.14545252919197083,
-0.013198891654610634,
-0.057603951543569565,
0.0432538278400898,
-0.013354936614632607,
-0.07987949997186661,
-0.022308066487312317,
-0.023925410583615303,
0.06286203861236572,
-0.05050002411007881,
0.026798952370882034,
0.06403107941150665,
0.027687007561326027,
-0.019890379160642624,
-0.01794293522834778,
0.04868467524647713,
0.05412912368774414,
-0.14384233951568604,
-0.055877987295389175
] |
null | null |
transformers
|
# mT5-m2o-english-CrossSum
This repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the [CrossSum](https://huggingface.co/datasets/csebuetnlp/CrossSum) dataset, where the target summary was in **english**, i.e. this model tries to **summarize text written in any language in English.** For finetuning details and scripts, see the [paper](https://arxiv.org/abs/2112.08804) and the [official repository](https://github.com/csebuetnlp/CrossSum).
## Using this model in `transformers` (tested on 4.11.0.dev0)
```python
import re
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
WHITESPACE_HANDLER = lambda k: re.sub('\s+', ' ', re.sub('\n+', ' ', k.strip()))
article_text = """Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs "spill over into misinformation about vaccines in general". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. "We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO," the post said, referring to the World Health Organization."""
model_name = "csebuetnlp/mT5_m2o_english_crossSum"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSeq2SeqLM.from_pretrained(model_name)
input_ids = tokenizer(
[WHITESPACE_HANDLER(article_text)],
return_tensors="pt",
padding="max_length",
truncation=True,
max_length=512
)["input_ids"]
output_ids = model.generate(
input_ids=input_ids,
max_length=84,
no_repeat_ngram_size=2,
num_beams=4
)[0]
summary = tokenizer.decode(
output_ids,
skip_special_tokens=True,
clean_up_tokenization_spaces=False
)
print(summary)
```
## Citation
If you use this model, please cite the following paper:
```
@article{hasan2021crosssum,
author = {Tahmid Hasan and Abhik Bhattacharjee and Wasi Uddin Ahmad and Yuan-Fang Li and Yong-bin Kang and Rifat Shahriyar},
title = {CrossSum: Beyond English-Centric Cross-Lingual Abstractive Text Summarization for 1500+ Language Pairs},
journal = {CoRR},
volume = {abs/2112.08804},
year = {2021},
url = {https://arxiv.org/abs/2112.08804},
eprinttype = {arXiv},
eprint = {2112.08804}
}
```
|
{"language": ["am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo"], "tags": ["summarization", "mT5"], "licenses": ["cc-by-nc-sa-4.0"], "widget": [{"text": "Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs \"spill over into misinformation about vaccines in general\". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. \"We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO,\" the post said, referring to the World Health Organization."}]}
|
summarization
|
csebuetnlp/mT5_m2o_english_crossSum
|
[
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"summarization",
"mT5",
"am",
"ar",
"az",
"bn",
"my",
"zh",
"en",
"fr",
"gu",
"ha",
"hi",
"ig",
"id",
"ja",
"rn",
"ko",
"ky",
"mr",
"ne",
"om",
"ps",
"fa",
"pcm",
"pt",
"pa",
"ru",
"gd",
"sr",
"si",
"so",
"es",
"sw",
"ta",
"te",
"th",
"ti",
"tr",
"uk",
"ur",
"uz",
"vi",
"cy",
"yo",
"arxiv:2112.08804",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2112.08804"
] |
[
"am",
"ar",
"az",
"bn",
"my",
"zh",
"en",
"fr",
"gu",
"ha",
"hi",
"ig",
"id",
"ja",
"rn",
"ko",
"ky",
"mr",
"ne",
"om",
"ps",
"fa",
"pcm",
"pt",
"pa",
"ru",
"gd",
"sr",
"si",
"so",
"es",
"sw",
"ta",
"te",
"th",
"ti",
"tr",
"uk",
"ur",
"uz",
"vi",
"cy",
"yo"
] |
TAGS
#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #arxiv-2112.08804 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# mT5-m2o-english-CrossSum
This repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.
## Using this model in 'transformers' (tested on 4.11.0.dev0)
If you use this model, please cite the following paper:
|
[
"# mT5-m2o-english-CrossSum\n\nThis repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.",
"## Using this model in 'transformers' (tested on 4.11.0.dev0)\n\n\n\n\n\n\nIf you use this model, please cite the following paper:"
] |
[
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #arxiv-2112.08804 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# mT5-m2o-english-CrossSum\n\nThis repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.",
"## Using this model in 'transformers' (tested on 4.11.0.dev0)\n\n\n\n\n\n\nIf you use this model, please cite the following paper:"
] |
[
157,
104,
31
] |
[
"passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #arxiv-2112.08804 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# mT5-m2o-english-CrossSum\n\nThis repository contains the many-to-one (m2o) mT5 checkpoint finetuned on all cross-lingual pairs of the CrossSum dataset, where the target summary was in english, i.e. this model tries to summarize text written in any language in English. For finetuning details and scripts, see the paper and the official repository.## Using this model in 'transformers' (tested on 4.11.0.dev0)\n\n\n\n\n\n\nIf you use this model, please cite the following paper:"
] |
[
-0.09608187526464462,
-0.03249843046069145,
-0.0047204685397446156,
-0.01687534712255001,
0.06920012831687927,
0.006863716058433056,
0.10838908702135086,
0.10223676264286041,
0.020791489630937576,
0.050613172352313995,
0.06287726014852524,
0.11210094392299652,
0.06030525639653206,
0.05553257837891579,
-0.00011314707808196545,
-0.24768857657909393,
0.025020962581038475,
-0.0036083338782191277,
-0.05488017573952675,
0.14682047069072723,
0.1400417536497116,
-0.07035364210605621,
0.09357466548681259,
-0.0022418475709855556,
-0.027669604867696762,
0.07798011600971222,
-0.006785335950553417,
-0.047171901911497116,
0.13866941630840302,
0.11156458407640457,
0.0483054481446743,
0.06813783198595047,
0.055489782243967056,
-0.22453677654266357,
0.03028535656630993,
0.019703790545463562,
-0.028052760288119316,
0.035020142793655396,
0.07017799466848373,
-0.03019106201827526,
0.17540346086025238,
-0.08139634132385254,
-0.05520252883434296,
0.08105015009641647,
-0.09385492652654648,
-0.13911405205726624,
-0.05919887498021126,
0.0802159532904625,
0.0880703553557396,
0.10516814142465591,
-0.07043712586164474,
0.11682868748903275,
-0.032962020486593246,
0.07597774267196655,
0.2404051572084427,
-0.27106037735939026,
-0.03808446228504181,
0.14333556592464447,
0.08687426894903183,
0.07947811484336853,
-0.037566523998975754,
0.05621614679694176,
0.04284936562180519,
0.012089194729924202,
-0.010167432017624378,
-0.07211910933256149,
0.09086395800113678,
0.048144228756427765,
-0.16866950690746307,
-0.04668673127889633,
0.23690523207187653,
0.04394909366965294,
-0.040766216814517975,
-0.017200199887156487,
-0.077814020216465,
0.03711554780602455,
0.0010816180147230625,
-0.0724213719367981,
0.030734390020370483,
-0.028897827491164207,
0.04810824990272522,
-0.04536290466785431,
-0.1073082685470581,
-0.05375918745994568,
-0.08834025263786316,
0.03744835779070854,
-0.0024967321660369635,
0.022974226623773575,
-0.044697415083646774,
0.07605062425136566,
-0.2249334305524826,
-0.08256937563419342,
-0.07205382734537125,
-0.05913691967725754,
-0.07020602375268936,
-0.051276691257953644,
-0.014143708162009716,
-0.15310733020305634,
0.052139706909656525,
0.16254988312721252,
0.041910022497177124,
0.019965626299381256,
0.030309252440929413,
0.03223820775747299,
0.10016662627458572,
0.10463674366474152,
-0.0996498391032219,
-0.08614253997802734,
0.00412668427452445,
0.005526178982108831,
0.003925191704183817,
0.017036376520991325,
-0.04538211598992348,
-0.030004402622580528,
0.04249507188796997,
0.06136051565408707,
0.01573854871094227,
0.04617364704608917,
-0.02365618757903576,
-0.02772068977355957,
0.07869653403759003,
-0.14465086162090302,
-0.03225550428032875,
0.009502166882157326,
-0.030133754014968872,
0.05407314747571945,
-0.022606993094086647,
0.025267889723181725,
-0.09484191238880157,
0.05275505781173706,
-0.09309876710176468,
0.015341418795287609,
-0.03994688764214516,
-0.13512912392616272,
0.039457790553569794,
-0.0036687457468360662,
-0.01406197901815176,
-0.12786255776882172,
-0.1884189248085022,
-0.06911540031433105,
-0.018809707835316658,
-0.02908303402364254,
0.0009287158027291298,
-0.04284804314374924,
-0.03230789676308632,
-0.0009852354414761066,
-0.00907447375357151,
0.07453096657991409,
-0.07364937663078308,
0.03672296553850174,
0.01885894313454628,
0.03538491949439049,
0.048528797924518585,
0.02300771325826645,
-0.08855453878641129,
0.007859001867473125,
-0.19586919248104095,
0.07917758822441101,
-0.09202078729867935,
-0.028630372136831284,
-0.1681988388299942,
-0.07167384773492813,
0.004165246617048979,
0.036687977612018585,
0.011195043101906776,
0.19063664972782135,
-0.15422306954860687,
-0.07232622057199478,
0.17202290892601013,
-0.1366683542728424,
-0.07409042119979858,
0.09872192144393921,
-0.0024950657971203327,
0.05295003205537796,
0.06852338463068008,
0.09493208676576614,
0.014046325348317623,
-0.10236319154500961,
-0.1042974665760994,
0.038755785673856735,
-0.04669509455561638,
0.03383014723658562,
0.14717692136764526,
-0.04979848861694336,
-0.007033463567495346,
-0.03301052376627922,
-0.07229963690042496,
-0.061372313648462296,
-0.06917070597410202,
-0.014514966867864132,
0.014452483505010605,
0.024546636268496513,
-0.017484430223703384,
-0.011045522056519985,
0.02824225276708603,
-0.02245512418448925,
-0.08469744771718979,
0.10468625277280807,
0.08053454011678696,
-0.013624956831336021,
0.02989398129284382,
-0.06481290608644485,
0.07545405626296997,
-0.14655271172523499,
0.011904168874025345,
-0.17814652621746063,
0.003918274771422148,
-0.035195495933294296,
-0.01831330917775631,
0.08493191003799438,
0.11567158997058868,
0.03843705728650093,
-0.0003502104082144797,
-0.05371128395199776,
-0.017226506024599075,
-0.020814118906855583,
0.012539055198431015,
-0.04840230941772461,
-0.14780540764331818,
-0.0603923462331295,
-0.047830671072006226,
0.038078270852565765,
-0.1947949230670929,
0.048610471189022064,
0.13488981127738953,
0.09240087866783142,
-0.02021961659193039,
-0.01544780284166336,
0.037636157125234604,
0.010542750358581543,
-0.0391099713742733,
-0.006070731673389673,
0.03888712450861931,
-0.030429869890213013,
-0.09433295577764511,
0.12121254205703735,
-0.06951451301574707,
0.04550592228770256,
0.09678824245929718,
-0.09512206166982651,
-0.04221123456954956,
-0.0015357627999037504,
-0.04806359484791756,
-0.018145881593227386,
-0.06481865793466568,
-0.10257669538259506,
0.08736004680395126,
0.03818660229444504,
0.08947106450796127,
-0.12211883813142776,
-0.05619807541370392,
-0.011669391766190529,
-0.04097792133688927,
-0.038024209439754486,
0.14730796217918396,
0.13614684343338013,
-0.22474130988121033,
0.09459280967712402,
0.09634040296077728,
0.056217752397060394,
0.11632884293794632,
0.03079274669289589,
-0.09302976727485657,
-0.03665049374103546,
-0.037349577993154526,
-0.011092516593635082,
0.022355396300554276,
-0.13813695311546326,
0.018614528700709343,
0.03626350685954094,
0.06274086982011795,
0.048110686242580414,
-0.08430922031402588,
0.023102769628167152,
0.034098707139492035,
-0.054320819675922394,
-0.06708180159330368,
0.05114087834954262,
-0.01752246357500553,
0.11275601387023926,
-0.013112070970237255,
0.054923027753829956,
-0.02163437381386757,
-0.03853237256407738,
-0.1621776521205902,
0.24190585315227509,
-0.07230739295482635,
-0.1990939825773239,
-0.14092570543289185,
-0.03290741890668869,
-0.004290411714464426,
0.006632637698203325,
0.04710689187049866,
-0.06208227202296257,
-0.057030946016311646,
-0.05827367678284645,
0.15698522329330444,
-0.05020911246538162,
-0.025685248896479607,
-0.026260916143655777,
0.013282785192131996,
-0.034448087215423584,
-0.11714126169681549,
-0.03257855027914047,
0.000629653048235923,
-0.03999554365873337,
0.05400373041629791,
-0.09403865784406662,
0.08558535575866699,
0.14661449193954468,
0.0055513703264296055,
0.013002083636820316,
-0.0360204242169857,
0.2374764382839203,
-0.09815853834152222,
0.08083433657884598,
0.126445934176445,
0.02106372080743313,
0.04833296686410904,
0.136552095413208,
0.03837266191840172,
-0.04840388149023056,
0.0078066689893603325,
0.02648116834461689,
-0.028180690482258797,
-0.2976570725440979,
-0.03613896667957306,
-0.040955014526844025,
0.006232014391571283,
-0.023737939074635506,
0.03058096021413803,
0.0006151704001240432,
0.03875765949487686,
0.00014052286860533059,
-0.06797593086957932,
0.07905148714780807,
0.08120037615299225,
0.17933610081672668,
-0.03187602758407593,
0.1018315926194191,
-0.05022238567471504,
0.014464686624705791,
0.0988161489367485,
-0.0317094512283802,
0.07845721393823624,
0.03673075512051582,
0.10731338709592819,
0.0854557529091835,
-0.06932787597179413,
0.060933079570531845,
0.09500563889741898,
0.03541182726621628,
-0.01575794629752636,
-0.05141366273164749,
-0.07150494307279587,
-0.013204546645283699,
0.07754538208246231,
-0.0006494509871117771,
0.030169246718287468,
-0.05510684847831726,
0.00025268536410294473,
0.0796356201171875,
0.12233056128025055,
0.11905698478221893,
-0.18897558748722076,
-0.0825224369764328,
0.03925305977463722,
-0.08570335805416107,
-0.028067778795957565,
0.013382963836193085,
0.050236549228429794,
-0.12715069949626923,
0.08691269159317017,
-0.003926199395209551,
0.0817563533782959,
-0.04231928661465645,
0.013595031574368477,
-0.05688280612230301,
0.04757945239543915,
-0.06635307520627975,
0.10871340334415436,
-0.21348366141319275,
0.24860087037086487,
0.028948677703738213,
0.003447059541940689,
-0.032121725380420685,
0.002351206261664629,
0.027420859783887863,
0.12465158104896545,
0.16898320615291595,
0.038206346333026886,
-0.11525604128837585,
-0.06141266971826553,
-0.057370707392692566,
0.019095610827207565,
0.055444035679101944,
-0.055309657007455826,
0.10468124598264694,
-0.01123357005417347,
-0.0286799855530262,
-0.05152622610330582,
0.08912047743797302,
-0.08574827760457993,
-0.059975869953632355,
0.055604953318834305,
-0.04861712083220482,
0.0570690780878067,
-0.039388760924339294,
-0.07933654636144638,
-0.008551347069442272,
0.20244550704956055,
-0.09379859268665314,
-0.08874478191137314,
-0.1119011789560318,
0.10880919545888901,
0.1179039403796196,
-0.10153573006391525,
-0.01801026239991188,
-0.00041161212720908225,
0.07553088665008545,
-0.06341224163770676,
-0.027070054784417152,
0.05060676857829094,
-0.09900392591953278,
-0.1290867030620575,
-0.006477785762399435,
0.17706312239170074,
0.028443120419979095,
0.07298622280359268,
-0.01846177503466606,
0.04566960036754608,
0.005181921645998955,
-0.09748733788728714,
0.031605977565050125,
0.07200680673122406,
0.03631093353033066,
0.10945894569158554,
-0.014866901561617851,
-0.1495436280965805,
-0.08084730058908463,
-0.08226050436496735,
0.13337749242782593,
0.2415807992219925,
-0.022731944918632507,
0.0059094661846756935,
0.14988526701927185,
-0.12702897191047668,
-0.17947930097579956,
-0.025609849020838737,
-0.017975356429815292,
0.03621434420347214,
-0.03056873381137848,
-0.0697714239358902,
0.09296654909849167,
0.11502140760421753,
0.01461400929838419,
0.012918460182845592,
-0.2774164080619812,
-0.14455026388168335,
0.05357075855135918,
0.04011279344558716,
0.06414824724197388,
-0.12101832777261734,
-0.08832836896181107,
-0.017887132242321968,
-0.15037941932678223,
-0.046370457857847214,
0.024574926123023033,
0.07341130077838898,
-0.0012781245168298483,
0.003961221314966679,
-0.0020799252670258284,
-0.062235649675130844,
0.1423482745885849,
0.07378239184617996,
-0.021500710397958755,
-0.049922551959753036,
-0.0834885984659195,
0.046477679163217545,
-0.01795426569879055,
0.0944494754076004,
0.028150035068392754,
0.030609557405114174,
-0.1693129688501358,
-0.06600363552570343,
-0.09165352582931519,
0.04092852771282196,
-0.04155939444899559,
-0.05823942646384239,
-0.05174296349287033,
0.05033516883850098,
-0.015236235223710537,
-0.02891722321510315,
0.0866067111492157,
-0.18337032198905945,
0.08640597015619278,
0.12675884366035461,
0.1816687285900116,
-0.051380548626184464,
-0.08014682680368423,
0.0003993920690845698,
-0.028780536726117134,
0.020264262333512306,
-0.14086386561393738,
0.01714113913476467,
0.1891166865825653,
0.03878457844257355,
0.1366960108280182,
0.03663665056228638,
-0.12230265885591507,
-0.0070900991559028625,
0.05075749382376671,
-0.10571947693824768,
-0.13921815156936646,
-0.061642084270715714,
-0.06681475043296814,
-0.06401066482067108,
0.011244140565395355,
0.18681524693965912,
-0.05103781446814537,
-0.03454102575778961,
0.013388270512223244,
0.04745827987790108,
-0.034712500870227814,
0.14446942508220673,
0.043610721826553345,
0.10284263640642166,
-0.04293866083025932,
0.06577582657337189,
0.04928402230143547,
-0.14590518176555634,
-0.0013801095774397254,
0.13444030284881592,
-0.10592292994260788,
-0.10968507081270218,
-0.08309195190668106,
0.04468484967947006,
-0.13626843690872192,
-0.040553636848926544,
-0.05217250436544418,
-0.09318459033966064,
0.11313868314027786,
0.26966196298599243,
0.07630538195371628,
0.005832205060869455,
0.002672710223123431,
-0.04942847415804863,
-0.03453722968697548,
0.07400050014257431,
0.06338871270418167,
0.0057105072773993015,
-0.05523018166422844,
0.04828348383307457,
0.005404562223702669,
0.06822384148836136,
-0.03664863854646683,
-0.009829435497522354,
-0.11073896288871765,
-0.00211991835385561,
-0.12346631288528442,
0.02706395648419857,
-0.11411421746015549,
-0.0448688268661499,
-0.0232267864048481,
-0.0666583776473999,
-0.06368216872215271,
0.001143145840615034,
-0.06493087112903595,
0.00845089741051197,
-0.03337518870830536,
0.07265631854534149,
-0.11740630120038986,
-0.01084820181131363,
0.0683927983045578,
-0.04303618520498276,
0.08845889568328857,
0.05756441876292229,
-0.025138691067695618,
0.04297187179327011,
-0.03983324021100998,
0.0012445427710190415,
0.03553915396332741,
0.027929389849305153,
0.010498234070837498,
-0.16496944427490234,
0.03163580223917961,
0.014075500890612602,
0.010047223418951035,
0.02202426828444004,
-0.020190052688121796,
-0.06885429471731186,
0.08063716441392899,
-0.015062405727803707,
-0.009842400439083576,
-0.0858205035328865,
0.03400293365120888,
0.06472918391227722,
0.011953581124544144,
0.09104674309492111,
-0.08441045135259628,
0.08935285359621048,
-0.19628871977329254,
0.06105251982808113,
-0.02586461789906025,
-0.06568820029497147,
0.007893064059317112,
-0.04067995399236679,
0.05260491371154785,
-0.07871285825967789,
0.13803057372570038,
-0.03530099615454674,
0.09094929695129395,
0.0954563170671463,
-0.01816081814467907,
-0.0009340798133052886,
0.01109381951391697,
0.08991771936416626,
0.035627368837594986,
-0.017645999789237976,
-0.006560726091265678,
0.05828696861863136,
0.01686730422079563,
-0.026306962594389915,
0.18915627896785736,
0.15456297993659973,
0.06830284744501114,
0.11857038736343384,
-0.020357679575681686,
-0.053671762347221375,
-0.06374875456094742,
-0.060787394642829895,
0.006244149059057236,
0.028175652027130127,
-0.011636503040790558,
0.13011476397514343,
0.24591515958309174,
-0.1355893313884735,
0.055136773735284805,
-0.0183431226760149,
-0.058465663343667984,
-0.15457986295223236,
-0.15160225331783295,
-0.09146776795387268,
-0.052706457674503326,
0.01212689932435751,
-0.14993134140968323,
0.031088154762983322,
0.0800442025065422,
0.08113407343626022,
0.03167780116200447,
0.10042259097099304,
-0.04176720231771469,
-0.10184504091739655,
0.054506558924913406,
-0.03298872336745262,
0.01669522561132908,
-0.01713918149471283,
0.05466693639755249,
0.0629722997546196,
-0.031763188540935516,
0.004489050712436438,
0.044671256095170975,
0.013491661287844181,
0.0033857894595712423,
-0.056724876165390015,
-0.0799194872379303,
-0.037689439952373505,
0.08722206205129623,
0.019582660868763924,
0.040729399770498276,
0.043964482843875885,
-0.03370688483119011,
0.02483776956796646,
0.1747635006904602,
-0.04714681953191757,
-0.14979234337806702,
-0.10011957585811615,
0.27194100618362427,
0.017094040289521217,
0.060714516788721085,
0.011510791257023811,
-0.08254165202379227,
0.04262172058224678,
0.17997372150421143,
0.25481027364730835,
0.022738616913557053,
0.002361099235713482,
0.040480490773916245,
0.02064366266131401,
0.08538824319839478,
-0.04714485630393028,
0.08051390945911407,
0.20523370802402496,
-0.10947464406490326,
0.06159772723913193,
-0.05394895002245903,
0.011531834490597248,
0.05711811035871506,
0.0914478749036789,
0.05253611505031586,
-0.02448766864836216,
0.01770862191915512,
0.11310328543186188,
-0.0520821250975132,
-0.14464548230171204,
-0.046666719019412994,
-0.07635517418384552,
-0.07667813450098038,
-0.017557887360453606,
-0.08716671168804169,
0.04671185836195946,
0.04439183697104454,
-0.005941525101661682,
-0.06776446849107742,
0.09240076690912247,
0.05372226610779762,
-0.15006721019744873,
-0.08835455030202866,
0.11770972609519958,
-0.051096998155117035,
0.04880569130182266,
-0.0333821177482605,
0.09975002706050873,
0.12282567471265793,
-0.02882368303835392,
-0.07001163065433502,
0.09133550524711609,
0.04952778294682503,
-0.008223874494433403,
0.02949834056198597,
0.06134914234280586,
0.035606276243925095,
0.13139943778514862,
0.03864523023366928,
-0.17284081876277924,
0.0519588403403759,
0.04460575431585312,
-0.04622754454612732,
-0.06262288242578506,
0.10816153138875961,
-0.1159312054514885,
0.11276450753211975,
0.1413266509771347,
-0.0021738146897405386,
0.002547777956351638,
-0.05706258490681648,
0.05900374799966812,
0.02661932073533535,
0.08303573727607727,
-0.0012838876573368907,
-0.13310682773590088,
0.047784414142370224,
-0.09077069908380508,
0.01665882207453251,
-0.25229036808013916,
-0.06172454357147217,
0.010941039770841599,
0.002646885346621275,
-0.0419132262468338,
0.09119468182325363,
0.07950460910797119,
0.045082613825798035,
-0.02342906780540943,
-0.13387075066566467,
-0.0014576348476111889,
0.1547166109085083,
-0.15400967001914978,
-0.05059441924095154
] |
null | null |
transformers
|
# mT5-multilingual-XLSum
This repository contains the mT5 checkpoint finetuned on the 45 languages of [XL-Sum](https://huggingface.co/datasets/csebuetnlp/xlsum) dataset. For finetuning details and scripts,
see the [paper](https://aclanthology.org/2021.findings-acl.413/) and the [official repository](https://github.com/csebuetnlp/xl-sum).
## Using this model in `transformers` (tested on 4.11.0.dev0)
```python
import re
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
WHITESPACE_HANDLER = lambda k: re.sub('\s+', ' ', re.sub('\n+', ' ', k.strip()))
article_text = """Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs "spill over into misinformation about vaccines in general". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. "We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO," the post said, referring to the World Health Organization."""
model_name = "csebuetnlp/mT5_multilingual_XLSum"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForSeq2SeqLM.from_pretrained(model_name)
input_ids = tokenizer(
[WHITESPACE_HANDLER(article_text)],
return_tensors="pt",
padding="max_length",
truncation=True,
max_length=512
)["input_ids"]
output_ids = model.generate(
input_ids=input_ids,
max_length=84,
no_repeat_ngram_size=2,
num_beams=4
)[0]
summary = tokenizer.decode(
output_ids,
skip_special_tokens=True,
clean_up_tokenization_spaces=False
)
print(summary)
```
## Benchmarks
Scores on the XL-Sum test sets are as follows:
Language | ROUGE-1 / ROUGE-2 / ROUGE-L
---------|----------------------------
Amharic | 20.0485 / 7.4111 / 18.0753
Arabic | 34.9107 / 14.7937 / 29.1623
Azerbaijani | 21.4227 / 9.5214 / 19.3331
Bengali | 29.5653 / 12.1095 / 25.1315
Burmese | 15.9626 / 5.1477 / 14.1819
Chinese (Simplified) | 39.4071 / 17.7913 / 33.406
Chinese (Traditional) | 37.1866 / 17.1432 / 31.6184
English | 37.601 / 15.1536 / 29.8817
French | 35.3398 / 16.1739 / 28.2041
Gujarati | 21.9619 / 7.7417 / 19.86
Hausa | 39.4375 / 17.6786 / 31.6667
Hindi | 38.5882 / 16.8802 / 32.0132
Igbo | 31.6148 / 10.1605 / 24.5309
Indonesian | 37.0049 / 17.0181 / 30.7561
Japanese | 48.1544 / 23.8482 / 37.3636
Kirundi | 31.9907 / 14.3685 / 25.8305
Korean | 23.6745 / 11.4478 / 22.3619
Kyrgyz | 18.3751 / 7.9608 / 16.5033
Marathi | 22.0141 / 9.5439 / 19.9208
Nepali | 26.6547 / 10.2479 / 24.2847
Oromo | 18.7025 / 6.1694 / 16.1862
Pashto | 38.4743 / 15.5475 / 31.9065
Persian | 36.9425 / 16.1934 / 30.0701
Pidgin | 37.9574 / 15.1234 / 29.872
Portuguese | 37.1676 / 15.9022 / 28.5586
Punjabi | 30.6973 / 12.2058 / 25.515
Russian | 32.2164 / 13.6386 / 26.1689
Scottish Gaelic | 29.0231 / 10.9893 / 22.8814
Serbian (Cyrillic) | 23.7841 / 7.9816 / 20.1379
Serbian (Latin) | 21.6443 / 6.6573 / 18.2336
Sinhala | 27.2901 / 13.3815 / 23.4699
Somali | 31.5563 / 11.5818 / 24.2232
Spanish | 31.5071 / 11.8767 / 24.0746
Swahili | 37.6673 / 17.8534 / 30.9146
Tamil | 24.3326 / 11.0553 / 22.0741
Telugu | 19.8571 / 7.0337 / 17.6101
Thai | 37.3951 / 17.275 / 28.8796
Tigrinya | 25.321 / 8.0157 / 21.1729
Turkish | 32.9304 / 15.5709 / 29.2622
Ukrainian | 23.9908 / 10.1431 / 20.9199
Urdu | 39.5579 / 18.3733 / 32.8442
Uzbek | 16.8281 / 6.3406 / 15.4055
Vietnamese | 32.8826 / 16.2247 / 26.0844
Welsh | 32.6599 / 11.596 / 26.1164
Yoruba | 31.6595 / 11.6599 / 25.0898
## Citation
If you use this model, please cite the following paper:
```
@inproceedings{hasan-etal-2021-xl,
title = "{XL}-Sum: Large-Scale Multilingual Abstractive Summarization for 44 Languages",
author = "Hasan, Tahmid and
Bhattacharjee, Abhik and
Islam, Md. Saiful and
Mubasshir, Kazi and
Li, Yuan-Fang and
Kang, Yong-Bin and
Rahman, M. Sohel and
Shahriyar, Rifat",
booktitle = "Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021",
month = aug,
year = "2021",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2021.findings-acl.413",
pages = "4693--4703",
}
```
|
{"language": ["am", "ar", "az", "bn", "my", "zh", "en", "fr", "gu", "ha", "hi", "ig", "id", "ja", "rn", "ko", "ky", "mr", "ne", "om", "ps", "fa", "pcm", "pt", "pa", "ru", "gd", "sr", "si", "so", "es", "sw", "ta", "te", "th", "ti", "tr", "uk", "ur", "uz", "vi", "cy", "yo"], "tags": ["summarization", "mT5"], "datasets": ["csebuetnlp/xlsum"], "licenses": ["cc-by-nc-sa-4.0"], "widget": [{"text": "Videos that say approved vaccines are dangerous and cause autism, cancer or infertility are among those that will be taken down, the company said. The policy includes the termination of accounts of anti-vaccine influencers. Tech giants have been criticised for not doing more to counter false health information on their sites. In July, US President Joe Biden said social media platforms were largely responsible for people's scepticism in getting vaccinated by spreading misinformation, and appealed for them to address the issue. YouTube, which is owned by Google, said 130,000 videos were removed from its platform since last year, when it implemented a ban on content spreading misinformation about Covid vaccines. In a blog post, the company said it had seen false claims about Covid jabs \"spill over into misinformation about vaccines in general\". The new policy covers long-approved vaccines, such as those against measles or hepatitis B. \"We're expanding our medical misinformation policies on YouTube with new guidelines on currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO,\" the post said, referring to the World Health Organization."}], "model-index": [{"name": "csebuetnlp/mT5_multilingual_XLSum", "results": [{"task": {"type": "summarization", "name": "Summarization"}, "dataset": {"name": "xsum", "type": "xsum", "config": "default", "split": "test"}, "metrics": [{"type": "rouge", "value": 36.5002, "name": "ROUGE-1", "verified": true}, {"type": "rouge", "value": 13.934, "name": "ROUGE-2", "verified": true}, {"type": "rouge", "value": 28.9876, "name": "ROUGE-L", "verified": true}, {"type": "rouge", "value": 28.9958, "name": "ROUGE-LSUM", "verified": true}, {"type": "loss", "value": 2.0674800872802734, "name": "loss", "verified": true}, {"type": "gen_len", "value": 26.9733, "name": "gen_len", "verified": true}]}]}]}
|
summarization
|
csebuetnlp/mT5_multilingual_XLSum
|
[
"transformers",
"pytorch",
"mt5",
"text2text-generation",
"summarization",
"mT5",
"am",
"ar",
"az",
"bn",
"my",
"zh",
"en",
"fr",
"gu",
"ha",
"hi",
"ig",
"id",
"ja",
"rn",
"ko",
"ky",
"mr",
"ne",
"om",
"ps",
"fa",
"pcm",
"pt",
"pa",
"ru",
"gd",
"sr",
"si",
"so",
"es",
"sw",
"ta",
"te",
"th",
"ti",
"tr",
"uk",
"ur",
"uz",
"vi",
"cy",
"yo",
"dataset:csebuetnlp/xlsum",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"am",
"ar",
"az",
"bn",
"my",
"zh",
"en",
"fr",
"gu",
"ha",
"hi",
"ig",
"id",
"ja",
"rn",
"ko",
"ky",
"mr",
"ne",
"om",
"ps",
"fa",
"pcm",
"pt",
"pa",
"ru",
"gd",
"sr",
"si",
"so",
"es",
"sw",
"ta",
"te",
"th",
"ti",
"tr",
"uk",
"ur",
"uz",
"vi",
"cy",
"yo"
] |
TAGS
#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #dataset-csebuetnlp/xlsum #model-index #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us
|
mT5-multilingual-XLSum
======================
This repository contains the mT5 checkpoint finetuned on the 45 languages of XL-Sum dataset. For finetuning details and scripts,
see the paper and the official repository.
Using this model in 'transformers' (tested on 4.11.0.dev0)
----------------------------------------------------------
Benchmarks
----------
Scores on the XL-Sum test sets are as follows:
If you use this model, please cite the following paper:
|
[] |
[
"TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #dataset-csebuetnlp/xlsum #model-index #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n"
] |
[
168
] |
[
"passage: TAGS\n#transformers #pytorch #mt5 #text2text-generation #summarization #mT5 #am #ar #az #bn #my #zh #en #fr #gu #ha #hi #ig #id #ja #rn #ko #ky #mr #ne #om #ps #fa #pcm #pt #pa #ru #gd #sr #si #so #es #sw #ta #te #th #ti #tr #uk #ur #uz #vi #cy #yo #dataset-csebuetnlp/xlsum #model-index #autotrain_compatible #endpoints_compatible #has_space #text-generation-inference #region-us \n"
] |
[
-0.08646338433027267,
-0.02580511011183262,
-0.006402633152902126,
0.014841850847005844,
0.11805697530508041,
0.03955939784646034,
0.07664599269628525,
0.1215146854519844,
0.04115590453147888,
0.06827494502067566,
0.1432306319475174,
0.13022325932979584,
0.066834457218647,
0.09859861433506012,
-0.020829707384109497,
-0.2796865999698639,
0.042838387191295624,
0.01805652119219303,
-0.06560946255922318,
0.14316698908805847,
0.07420980930328369,
-0.07293959707021713,
0.12529557943344116,
-0.059878136962652206,
-0.03137265890836716,
-0.009355505928397179,
-0.03475627303123474,
-0.03452621027827263,
0.10409579426050186,
0.08351270854473114,
0.0011220359010621905,
0.10217955708503723,
-0.009927472099661827,
-0.19220183789730072,
0.05530533567070961,
0.010962228290736675,
-0.06121176853775978,
0.06295283883810043,
0.06624240428209305,
-0.05702592432498932,
0.20684295892715454,
0.010274659842252731,
-0.10116490721702576,
0.07062696665525436,
-0.13991835713386536,
-0.10073749721050262,
-0.06109984964132309,
0.09174918383359909,
0.022892916575074196,
0.0860346108675003,
-0.041013505309820175,
0.12542006373405457,
-0.08066155761480331,
0.07819471508264542,
0.25837138295173645,
-0.2753373384475708,
-0.013708444312214851,
0.10153517872095108,
0.06425516307353973,
0.11544177681207657,
-0.0648888573050499,
0.07164723426103592,
0.03990350291132927,
0.011139418929815292,
0.014654522761702538,
-0.05404828488826752,
0.09234413504600525,
0.06270287185907364,
-0.12023982405662537,
-0.02208303101360798,
0.2882609963417053,
0.00890812836587429,
0.043102510273456573,
0.054322902113199234,
-0.10071641206741333,
-0.11903024464845657,
-0.03480162099003792,
0.0030002507846802473,
-0.032737456262111664,
0.035928044468164444,
-0.016664890572428703,
-0.007063908036798239,
-0.11757451295852661,
-0.012095363810658455,
-0.12029002606868744,
0.08956138789653778,
0.020348967984318733,
0.009425998665392399,
-0.05437120050191879,
0.08388996869325638,
-0.0407414436340332,
-0.1163187250494957,
0.01163626927882433,
-0.05937284231185913,
0.03361080586910248,
0.0175373163074255,
0.0023424476385116577,
-0.02653774805366993,
0.08220241963863373,
0.06497935205698013,
-0.09780006855726242,
0.045716144144535065,
0.05472186952829361,
0.09946230053901672,
0.04137740656733513,
0.03110693208873272,
-0.13020409643650055,
-0.12313950061798096,
-0.06357461214065552,
-0.004433367867022753,
-0.023852603510022163,
-0.04243079200387001,
-0.10859720408916473,
-0.02107400819659233,
0.0749984160065651,
0.02045246958732605,
0.03208538889884949,
0.11141199618577957,
-0.03804052993655205,
-0.015583659522235394,
0.059466972947120667,
-0.03752642124891281,
-0.01465922873467207,
0.054823342710733414,
-0.015965940430760384,
0.13291332125663757,
-0.01358629111200571,
0.005811266601085663,
-0.07705327868461609,
0.0746295377612114,
-0.09481575340032578,
0.011774844489991665,
-0.02097959630191326,
-0.07440358400344849,
0.044464945793151855,
-0.08328255265951157,
0.002461118856444955,
-0.15154998004436493,
-0.13312439620494843,
-0.011392655782401562,
-0.016019552946090698,
-0.0510411262512207,
0.006991948001086712,
-0.04892702028155327,
-0.08705591410398483,
0.04527730867266655,
-0.04525529593229294,
0.06288579851388931,
-0.09293904155492783,
0.11157049238681793,
-0.037971239537000656,
0.09139513224363327,
-0.06133711710572243,
0.05996314808726311,
-0.06494206935167313,
0.015841085463762283,
-0.04956357553601265,
0.02019020915031433,
-0.03552056849002838,
0.016650451347231865,
-0.08050727099180222,
-0.07340698689222336,
-0.10303705930709839,
0.042004916816949844,
-0.010168171487748623,
0.20999889075756073,
-0.12815506756305695,
-0.1229950413107872,
0.1728687435388565,
-0.07732796669006348,
-0.11980258673429489,
0.13947376608848572,
0.04672984033823013,
0.012982234358787537,
0.04674148187041283,
0.12775863707065582,
0.017619166523218155,
-0.056185781955718994,
-0.08064185827970505,
0.08863220363855362,
-0.0008219906012527645,
-0.021924324333667755,
0.12521758675575256,
0.04606008902192116,
-0.029642032459378242,
0.04374675825238228,
0.025907909497618675,
0.0524730421602726,
-0.09625004976987839,
-0.0479864627122879,
0.02365446835756302,
-0.025242431089282036,
0.06894630938768387,
0.05423135682940483,
0.04871244728565216,
-0.04982810840010643,
-0.05945693701505661,
-0.06991522014141083,
0.0721476674079895,
0.01388605684041977,
0.035613901913166046,
-0.05702692270278931,
0.1623762547969818,
-0.048149287700653076,
0.03330444544553757,
-0.16488759219646454,
0.0654362216591835,
-0.05291084200143814,
0.08847109973430634,
0.04705169051885605,
0.1513070911169052,
0.08368470519781113,
0.020035747438669205,
-0.07014743238687515,
-0.056721754372119904,
0.07458966970443726,
-0.0036694095470011234,
-0.07594125717878342,
-0.2001580148935318,
0.03095705807209015,
-0.05246477946639061,
0.0541391484439373,
-0.12220260500907898,
0.043900344520807266,
0.1203746572136879,
0.13357628881931305,
-0.013496517203748226,
0.028763482347130775,
-0.008525585755705833,
0.0537622794508934,
-0.07923253625631332,
0.030123241245746613,
0.09258973598480225,
-0.01786765456199646,
-0.09028387814760208,
0.1356632113456726,
-0.13933894038200378,
0.19381265342235565,
0.18729551136493683,
-0.16480375826358795,
-0.0106788519769907,
0.00305515737272799,
-0.04154086485505104,
-0.005570478271692991,
0.05160908401012421,
-0.058879874646663666,
0.021410977467894554,
0.012488342821598053,
0.13865934312343597,
-0.07460424304008484,
-0.04306861758232117,
0.026865415275096893,
-0.05621049180626869,
-0.06938046962022781,
0.17617401480674744,
0.13984617590904236,
-0.184221088886261,
0.2193295657634735,
0.22216098010540009,
0.07638496905565262,
0.20972466468811035,
0.02146581932902336,
-0.07523148506879807,
-0.0067572640255093575,
-0.03842756524682045,
-0.03014184907078743,
0.05528867989778519,
-0.17208194732666016,
-0.004855314735323191,
0.05302481725811958,
0.028850145637989044,
0.07292649149894714,
-0.09483453631401062,
-0.03986866772174835,
-0.020175183191895485,
-0.012613067403435707,
-0.005751879420131445,
0.09126702696084976,
-0.020969988778233528,
0.14618010818958282,
0.015278870239853859,
-0.01508206408470869,
-0.021914981305599213,
0.012596981599926949,
-0.11008677631616592,
0.2006336897611618,
-0.1434328556060791,
-0.2647164463996887,
-0.06791103631258011,
-0.10699635744094849,
-0.03864356875419617,
0.010203001089394093,
0.06351496279239655,
-0.1494152545928955,
-0.005632020533084869,
-0.0401131734251976,
0.1266208440065384,
-0.12256360799074173,
0.02756439335644245,
-0.08154728263616562,
0.04155242443084717,
-0.06654123961925507,
-0.05503842979669571,
-0.044777773320674896,
-0.0032507332507520914,
-0.03223313391208649,
0.15086406469345093,
-0.14564742147922516,
0.10209407657384872,
0.14454345405101776,
0.05491498485207558,
0.04454679787158966,
-0.01745566725730896,
0.18424974381923676,
-0.11019963026046753,
0.0546313039958477,
0.12980739772319794,
0.03108617290854454,
0.08790569007396698,
0.11257853358983994,
0.012022046372294426,
-0.03635305166244507,
-0.0031973563600331545,
-0.025873683393001556,
-0.050823986530303955,
-0.19269588589668274,
-0.12297598272562027,
-0.09652827680110931,
0.14376114308834076,
-0.013132900930941105,
0.07093209773302078,
0.06360015273094177,
0.06547761708498001,
-0.0629582554101944,
-0.061056386679410934,
-0.012075182981789112,
0.03522107005119324,
0.07370967417955399,
-0.0680118203163147,
0.10450724512338638,
-0.04872577264904976,
-0.04218374565243721,
0.10744914412498474,
0.024929512292146683,
0.031761534512043,
0.03159615397453308,
0.06261876225471497,
0.05901128798723221,
0.09176083654165268,
0.08170662075281143,
0.0828763023018837,
0.022844430059194565,
-0.03887558355927467,
-0.04769158363342285,
-0.05327599495649338,
-0.038153111934661865,
0.04842197522521019,
0.07148966938257217,
-0.019605575129389763,
-0.05496722459793091,
-0.018697328865528107,
0.09307091683149338,
0.02155941165983677,
0.09615350514650345,
-0.20223158597946167,
-0.04868096485733986,
0.049459151923656464,
0.024035051465034485,
-0.05029739439487457,
0.03762015700340271,
0.07877764850854874,
-0.10010737180709839,
0.048669926822185516,
0.019626609981060028,
0.1006847396492958,
-0.046271178871393204,
0.07306604832410812,
-0.011013571172952652,
0.0003125910006929189,
-0.027551395818591118,
0.08456190675497055,
-0.2643141448497772,
0.2700752913951874,
0.021132059395313263,
-0.04820789396762848,
-0.03734063729643822,
-0.05281893163919449,
0.03586646169424057,
0.17501327395439148,
0.11537172645330429,
0.054651711136102676,
-0.07836266607046127,
-0.14567236602306366,
-0.03072880581021309,
0.011262441985309124,
0.12249168008565903,
-0.0760888084769249,
0.06281032413244247,
-0.026217155158519745,
0.0003677707863971591,
-0.019980330020189285,
0.0870484709739685,
-0.0485142283141613,
-0.10698268562555313,
0.10248531401157379,
-0.004980690777301788,
0.039586085826158524,
-0.022746069356799126,
-0.07157531380653381,
-0.1603529155254364,
0.08944152295589447,
-0.10459475964307785,
-0.051959335803985596,
-0.10517973452806473,
0.01071182545274496,
0.06875112652778625,
-0.12003149837255478,
-0.09473168849945068,
-0.030588233843445778,
0.0022939175833016634,
-0.05684424936771393,
-0.09701033681631088,
0.08882134407758713,
-0.08890765905380249,
-0.16795499622821808,
-0.029955841600894928,
0.1478097140789032,
0.016809793189167976,
0.09889959543943405,
-0.06433264166116714,
0.032100699841976166,
-0.07950994372367859,
-0.09678886085748672,
0.08184576779603958,
0.03300520405173302,
0.05671609193086624,
0.05158074572682381,
-0.05103420838713646,
0.0024466700851917267,
-0.04485040903091431,
-0.09979300945997238,
0.1561969667673111,
0.2654974162578583,
-0.03831382840871811,
0.08190181851387024,
0.1898847222328186,
-0.06698823720216751,
-0.32883501052856445,
-0.1020924523472786,
-0.12290880084037781,
0.006095423363149166,
-0.054667167365550995,
-0.1428288221359253,
-0.007802795618772507,
0.020302485674619675,
0.008326495997607708,
0.061280831694602966,
-0.28666168451309204,
-0.0804954245686531,
0.11632534861564636,
0.026943285018205643,
0.20262247323989868,
-0.1612684428691864,
-0.04419039189815521,
-0.031638771295547485,
-0.11420679092407227,
-0.0013124970719218254,
-0.08047515153884888,
0.1119115874171257,
-0.05642223358154297,
0.03384925052523613,
0.0031363742891699076,
-0.0412670336663723,
0.14393746852874756,
-0.005251957569271326,
-0.01978449523448944,
-0.09357107430696487,
-0.06671282649040222,
0.05618290603160858,
0.020172258839011192,
-0.008968192152678967,
-0.13067586719989777,
0.002749864710494876,
-0.10936305671930313,
0.001712290570139885,
-0.10428362339735031,
0.042806562036275864,
-0.025254538282752037,
-0.07440433651208878,
-0.09491828083992004,
0.06143561005592346,
0.015914084389805794,
-0.015384151600301266,
0.15805964171886444,
-0.09795594960451126,
0.17690154910087585,
0.14932550489902496,
0.12163546681404114,
-0.10477377474308014,
-0.0029790252447128296,
-0.056381579488515854,
-0.04294315725564957,
0.03388737142086029,
-0.15141144394874573,
-0.006584265735000372,
0.15575098991394043,
-0.0026862001977860928,
0.08538544178009033,
0.0842437893152237,
-0.08821997046470642,
-0.0017864579567685723,
0.12189710140228271,
-0.18256902694702148,
-0.1587209552526474,
-0.07199584692716599,
-0.04493510350584984,
0.025237757712602615,
0.032388217747211456,
0.13786470890045166,
-0.06252606213092804,
0.007686957251280546,
-0.0044171069748699665,
0.03348146751523018,
-0.0261247456073761,
0.14706666767597198,
0.054604947566986084,
0.04910893738269806,
-0.10807984322309494,
0.0745367780327797,
-0.018365714699029922,
-0.09974934160709381,
-0.019174844026565552,
0.14350084960460663,
-0.13603082299232483,
-0.1390450894832611,
-0.04728773981332779,
0.07022230327129364,
-0.05800842121243477,
-0.04889875277876854,
-0.06902402639389038,
-0.11054613441228867,
0.10943786799907684,
0.18772326409816742,
0.06845293194055557,
0.03541647642850876,
-0.006301398854702711,
-0.005879333708435297,
-0.0032616930548101664,
0.07122199237346649,
0.047307852655649185,
0.008813084103167057,
-0.11481264978647232,
0.06905263662338257,
-0.01946692354977131,
0.14809025824069977,
-0.06259697675704956,
-0.038330283015966415,
-0.13862833380699158,
0.013722794130444527,
-0.11192870885133743,
-0.07710595428943634,
-0.12475071102380753,
-0.08276855945587158,
-0.034661807119846344,
-0.10541734099388123,
-0.060688551515340805,
-0.05759627744555473,
-0.08924674242734909,
-0.0077257161028683186,
-0.015023157000541687,
0.07556517422199249,
-0.05175592750310898,
0.011267132125794888,
0.103173166513443,
-0.05046331137418747,
0.08374372124671936,
0.10673636198043823,
-0.0431801974773407,
0.09389308094978333,
-0.13563120365142822,
-0.01780153624713421,
0.040358442813158035,
0.043687961995601654,
0.0316753163933754,
0.02403990365564823,
-0.01372868474572897,
-0.020762581378221512,
0.027986908331513405,
0.06853371858596802,
0.04971487075090408,
-0.05410198122262955,
0.15225590765476227,
-0.09667443484067917,
-0.056180987507104874,
-0.07200424373149872,
0.0443526990711689,
0.06957680732011795,
0.04593314230442047,
0.08024001121520996,
-0.0779152438044548,
0.05859170854091644,
-0.13844361901283264,
0.06522088497877121,
0.008075354620814323,
-0.15024586021900177,
-0.008301224559545517,
-0.061293747276067734,
0.05687815696001053,
-0.053940434008836746,
0.13230113685131073,
-0.0344119518995285,
-0.07035045325756073,
0.041370514780282974,
0.031971968710422516,
-0.026774905622005463,
-0.007100245915353298,
0.10602365434169769,
0.08839527517557144,
-0.04227568954229355,
-0.0873025432229042,
0.09140364825725555,
0.02627820521593094,
0.019871899858117104,
0.15128456056118011,
0.09062705188989639,
0.0976504534482956,
0.08410143107175827,
-0.046059153974056244,
-0.03251180425286293,
-0.0656374990940094,
-0.052535828202962875,
-0.03929354250431061,
-0.0025122733786702156,
-0.024486053735017776,
0.20401577651500702,
0.27737337350845337,
-0.0894511267542839,
0.028562651947140694,
-0.04784608259797096,
-0.03032585047185421,
-0.14754462242126465,
-0.14468172192573547,
-0.07479477673768997,
-0.11383165419101715,
-0.004045257344841957,
-0.11564740538597107,
-0.011477400548756123,
0.07924777269363403,
0.07451639324426651,
-0.004670888185501099,
0.06655564159154892,
0.05333368107676506,
-0.08459985256195068,
0.04941840097308159,
-0.010902351699769497,
-0.004488688427954912,
-0.0811287984251976,
0.011388746090233326,
0.010560373775660992,
-0.07143492996692657,
0.0036565193440765142,
0.04611961171030998,
-0.08140251040458679,
0.007694265339523554,
-0.09253548830747604,
-0.12355495989322662,
-0.048570416867733,
0.04097408428788185,
-0.01347232423722744,
0.10258661210536957,
0.0002143421588698402,
-0.015863237902522087,
0.017246780917048454,
0.18903088569641113,
-0.06997811794281006,
-0.028809864073991776,
-0.029772967100143433,
0.15632179379463196,
0.015792034566402435,
0.06093810871243477,
0.007155471947044134,
-0.033846162259578705,
-0.04435848817229271,
0.28692737221717834,
0.3263632357120514,
-0.09259019047021866,
0.040054187178611755,
0.04673274606466293,
0.04663088545203209,
0.05650675669312477,
0.019446365535259247,
0.11920265853404999,
0.20709876716136932,
-0.10886888951063156,
0.04138452932238579,
-0.09229446947574615,
-0.0092853344976902,
-0.01768067106604576,
0.05967481806874275,
0.07244032621383667,
-0.02558945305645466,
-0.07466726005077362,
0.07927776873111725,
-0.19803692400455475,
-0.04248494654893875,
-0.054057929664850235,
-0.16209806501865387,
-0.05589992552995682,
-0.010377940721809864,
0.05967050418257713,
0.08711781352758408,
0.076105497777462,
-0.019122811034321785,
-0.0351237989962101,
-0.03232823312282562,
0.05595700815320015,
-0.17337502539157867,
0.03204787150025368,
0.07457692921161652,
-0.04495998099446297,
0.03231082484126091,
-0.0657830536365509,
-0.015107393264770508,
0.12348698824644089,
0.004296677187085152,
-0.03410175442695618,
0.06970909982919693,
0.06582240015268326,
-0.03800734132528305,
-0.033503949642181396,
0.05947599187493324,
0.05240248143672943,
0.005586260464042425,
0.11067622154951096,
-0.09385061264038086,
0.05889048054814339,
0.02572837471961975,
-0.05974463373422623,
0.003003775840625167,
0.0928146168589592,
-0.034844622015953064,
0.11481953412294388,
0.09349633008241653,
-0.002403327263891697,
-0.030757077038288116,
-0.06704383343458176,
-0.028497202321887016,
-0.013408532366156578,
-0.0070684547536075115,
-0.008555013686418533,
-0.12491217255592346,
-0.06465089321136475,
-0.00555462297052145,
0.048126619309186935,
-0.18978245556354523,
-0.02049051970243454,
-0.04568195715546608,
0.023115165531635284,
-0.09293416142463684,
0.10014216601848602,
0.07051248103380203,
0.0012292563915252686,
-0.02616814337670803,
-0.10490471869707108,
0.029085248708724976,
0.11885254830121994,
-0.13522009551525116,
-0.05303233116865158
] |
null | null |
transformers
|
# FrALBERT Base Cased
Pretrained model on French language using a masked language modeling (MLM) objective. It was introduced in
[this paper](https://arxiv.org/abs/1909.11942) and first released in
[this repository](https://github.com/google-research/albert).
This model, unlike other ALBERT models, is cased: it does make a difference between french and French.
## Model description
FrALBERT is a transformers model pretrained on 16Go of French Wikipedia in a self-supervised fashion. This means it
was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of
publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it
was pretrained with two objectives:
- Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run
the entire masked sentence through the model and has to predict the masked words. This is different from traditional
recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like
GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the
sentence.
- Sentence Ordering Prediction (SOP): FrALBERT uses a pretraining loss based on predicting the ordering of two consecutive segments of text.
This way, the model learns an inner representation of the English language that can then be used to extract features
useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard
classifier using the features produced by the FrALBERT model as inputs.
FrALBERT is particular in that it shares its layers across its Transformer. Therefore, all layers have the same weights. Using repeating layers results in a small memory footprint, however, the computational cost remains similar to a BERT-like architecture with the same number of hidden layers as it has to iterate through the same number of (repeating) layers.
This is the second version of the base model.
This model has the following configuration:
- 12 repeating layers
- 128 embedding dimension
- 768 hidden dimension
- 12 attention heads
- 11M parameters
## Intended uses & limitations
You can use the raw model for either masked language modeling or next sentence prediction, but it's mostly intended to
be fine-tuned on a downstream task. See the [model hub](https://huggingface.co/models?filter=fralbert-base-cased) to look for
fine-tuned versions on a task that interests you.
Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)
to make decisions, such as sequence classification, token classification or question answering. For tasks such as text
generation you should look at model like GPT2.
### How to use
You can use this model directly with a pipeline for masked language modeling:
```python
>>> from transformers import pipeline
>>> unmasker = pipeline('fill-mask', model='cservan/fralbert-base-cased')
>>> unmasker("Paris est la capitale de la [MASK] .")
[
{
"sequence": "paris est la capitale de la france.",
"score": 0.6231236457824707,
"token": 3043,
"token_str": "france"
},
{
"sequence": "paris est la capitale de la region.",
"score": 0.2993471622467041,
"token": 10531,
"token_str": "region"
},
{
"sequence": "paris est la capitale de la societe.",
"score": 0.02028230018913746,
"token": 24622,
"token_str": "societe"
},
{
"sequence": "paris est la capitale de la bretagne.",
"score": 0.012089950032532215,
"token": 24987,
"token_str": "bretagne"
},
{
"sequence": "paris est la capitale de la chine.",
"score": 0.010002839379012585,
"token": 14860,
"token_str": "chine"
}
]
```
Here is how to use this model to get the features of a given text in PyTorch:
```python
from transformers import AlbertTokenizer, AlbertModel
tokenizer = AlbertTokenizer.from_pretrained('cservan/fralbert-base-cased')
model = AlbertModel.from_pretrained("cservan/fralbert-base-cased")
text = "Remplacez-moi par le texte en français que vous souhaitez."
encoded_input = tokenizer(text, return_tensors='pt')
output = model(**encoded_input)
```
and in TensorFlow:
```python
from transformers import AlbertTokenizer, TFAlbertModel
tokenizer = AlbertTokenizer.from_pretrained('cservan/fralbert-base-cased')
model = TFAlbertModel.from_pretrained("cservan/fralbert-base-cased")
text = "Remplacez-moi par le texte en français que vous souhaitez."
encoded_input = tokenizer(text, return_tensors='tf')
output = model(encoded_input)
```
## Training data
The FrALBERT model was pretrained on 4go of [French Wikipedia](https://fr.wikipedia.org/wiki/French_Wikipedia) (excluding lists, tables and
headers).
## Training procedure
### Preprocessing
The texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are
then of the form:
```
[CLS] Sentence A [SEP] Sentence B [SEP]
```
### Training
The FrALBERT procedure follows the BERT setup.
The details of the masking procedure for each sentence are the following:
- 15% of the tokens are masked.
- In 80% of the cases, the masked tokens are replaced by `[MASK]`.
- In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.
- In the 10% remaining cases, the masked tokens are left as is.
## Evaluation results
When fine-tuned on downstream tasks, the ALBERT models achieve the following results:
Slot-filling:
| | FrALBERT-base | FrALBERT-base-cased
|----------------|---------------|--------------------
| MEDIA | 81.76 (0.59) | 85.09 (0.14)
|
### BibTeX entry and citation info
```bibtex
@inproceedings{cattan2021fralbert,
author = {Oralie Cattan and
Christophe Servan and
Sophie Rosset},
booktitle = {Recent Advances in Natural Language Processing, RANLP 2021},
title = {{On the Usability of Transformers-based models for a French Question-Answering task}},
year = {2021},
address = {Online},
month = sep,
}
```
Link to the paper: [PDF](https://hal.archives-ouvertes.fr/hal-03336060)
|
{"language": "fr", "license": "apache-2.0", "datasets": ["wikipedia"]}
|
fill-mask
|
cservan/fralbert-base-cased
|
[
"transformers",
"pytorch",
"albert",
"fill-mask",
"fr",
"dataset:wikipedia",
"arxiv:1909.11942",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"1909.11942"
] |
[
"fr"
] |
TAGS
#transformers #pytorch #albert #fill-mask #fr #dataset-wikipedia #arxiv-1909.11942 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
FrALBERT Base Cased
===================
Pretrained model on French language using a masked language modeling (MLM) objective. It was introduced in
this paper and first released in
this repository.
This model, unlike other ALBERT models, is cased: it does make a difference between french and French.
Model description
-----------------
FrALBERT is a transformers model pretrained on 16Go of French Wikipedia in a self-supervised fashion. This means it
was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of
publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it
was pretrained with two objectives:
* Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run
the entire masked sentence through the model and has to predict the masked words. This is different from traditional
recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like
GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the
sentence.
* Sentence Ordering Prediction (SOP): FrALBERT uses a pretraining loss based on predicting the ordering of two consecutive segments of text.
This way, the model learns an inner representation of the English language that can then be used to extract features
useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard
classifier using the features produced by the FrALBERT model as inputs.
FrALBERT is particular in that it shares its layers across its Transformer. Therefore, all layers have the same weights. Using repeating layers results in a small memory footprint, however, the computational cost remains similar to a BERT-like architecture with the same number of hidden layers as it has to iterate through the same number of (repeating) layers.
This is the second version of the base model.
This model has the following configuration:
* 12 repeating layers
* 128 embedding dimension
* 768 hidden dimension
* 12 attention heads
* 11M parameters
Intended uses & limitations
---------------------------
You can use the raw model for either masked language modeling or next sentence prediction, but it's mostly intended to
be fine-tuned on a downstream task. See the model hub to look for
fine-tuned versions on a task that interests you.
Note that this model is primarily aimed at being fine-tuned on tasks that use the whole sentence (potentially masked)
to make decisions, such as sequence classification, token classification or question answering. For tasks such as text
generation you should look at model like GPT2.
### How to use
You can use this model directly with a pipeline for masked language modeling:
Here is how to use this model to get the features of a given text in PyTorch:
and in TensorFlow:
Training data
-------------
The FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and
headers).
Training procedure
------------------
### Preprocessing
The texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are
then of the form:
### Training
The FrALBERT procedure follows the BERT setup.
The details of the masking procedure for each sentence are the following:
* 15% of the tokens are masked.
* In 80% of the cases, the masked tokens are replaced by '[MASK]'.
* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.
* In the 10% remaining cases, the masked tokens are left as is.
Evaluation results
------------------
When fine-tuned on downstream tasks, the ALBERT models achieve the following results:
Slot-filling:
FrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)
FrALBERT-base: , FrALBERT-base-cased:
### BibTeX entry and citation info
Link to the paper: PDF
|
[
"### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:\n\n\nTraining data\n-------------\n\n\nThe FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and\nheaders).\n\n\nTraining procedure\n------------------",
"### Preprocessing\n\n\nThe texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are\nthen of the form:",
"### Training\n\n\nThe FrALBERT procedure follows the BERT setup.\n\n\nThe details of the masking procedure for each sentence are the following:\n\n\n* 15% of the tokens are masked.\n* In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n* In the 10% remaining cases, the masked tokens are left as is.\n\n\nEvaluation results\n------------------\n\n\nWhen fine-tuned on downstream tasks, the ALBERT models achieve the following results:\n\n\nSlot-filling:\n\n\nFrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)\nFrALBERT-base: , FrALBERT-base-cased:",
"### BibTeX entry and citation info\n\n\nLink to the paper: PDF"
] |
[
"TAGS\n#transformers #pytorch #albert #fill-mask #fr #dataset-wikipedia #arxiv-1909.11942 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:\n\n\nTraining data\n-------------\n\n\nThe FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and\nheaders).\n\n\nTraining procedure\n------------------",
"### Preprocessing\n\n\nThe texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are\nthen of the form:",
"### Training\n\n\nThe FrALBERT procedure follows the BERT setup.\n\n\nThe details of the masking procedure for each sentence are the following:\n\n\n* 15% of the tokens are masked.\n* In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n* In the 10% remaining cases, the masked tokens are left as is.\n\n\nEvaluation results\n------------------\n\n\nWhen fine-tuned on downstream tasks, the ALBERT models achieve the following results:\n\n\nSlot-filling:\n\n\nFrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)\nFrALBERT-base: , FrALBERT-base-cased:",
"### BibTeX entry and citation info\n\n\nLink to the paper: PDF"
] |
[
60,
86,
42,
185,
17
] |
[
"passage: TAGS\n#transformers #pytorch #albert #fill-mask #fr #dataset-wikipedia #arxiv-1909.11942 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nHere is how to use this model to get the features of a given text in PyTorch:\n\n\nand in TensorFlow:\n\n\nTraining data\n-------------\n\n\nThe FrALBERT model was pretrained on 4go of French Wikipedia (excluding lists, tables and\nheaders).\n\n\nTraining procedure\n------------------### Preprocessing\n\n\nThe texts are lowercased and tokenized using SentencePiece and a vocabulary size of 32,000. The inputs of the model are\nthen of the form:### Training\n\n\nThe FrALBERT procedure follows the BERT setup.\n\n\nThe details of the masking procedure for each sentence are the following:\n\n\n* 15% of the tokens are masked.\n* In 80% of the cases, the masked tokens are replaced by '[MASK]'.\n* In 10% of the cases, the masked tokens are replaced by a random token (different) from the one they replace.\n* In the 10% remaining cases, the masked tokens are left as is.\n\n\nEvaluation results\n------------------\n\n\nWhen fine-tuned on downstream tasks, the ALBERT models achieve the following results:\n\n\nSlot-filling:\n\n\nFrALBERT-base: MEDIA, FrALBERT-base-cased: 81.76 (0.59)\nFrALBERT-base: , FrALBERT-base-cased:### BibTeX entry and citation info\n\n\nLink to the paper: PDF"
] |
[
-0.03126373142004013,
0.015831228345632553,
-0.005797833204269409,
0.09988299757242203,
0.02111339382827282,
0.0010972425807267427,
0.0966000109910965,
0.0784265398979187,
-0.07355821877717972,
0.09975865483283997,
0.033671341836452484,
0.05030170455574989,
0.0803571566939354,
0.14475573599338531,
-0.0023422869853675365,
-0.25827556848526,
0.08825349807739258,
-0.007973266765475273,
0.05110989138484001,
0.07126928865909576,
0.08964456617832184,
-0.07734489440917969,
-0.014853608794510365,
0.03814234212040901,
-0.009291918016970158,
-0.052537646144628525,
-0.00396701181307435,
-0.05471326783299446,
0.08350078761577606,
0.03464047610759735,
0.0712599903345108,
0.04687013104557991,
0.002947910688817501,
-0.1370858997106552,
0.029356304556131363,
0.06817523390054703,
-0.036716002970933914,
0.07374241203069687,
0.08421848714351654,
-0.040386080741882324,
0.09577756375074387,
-0.01874254085123539,
0.056160181760787964,
0.04625200480222702,
-0.13657629489898682,
-0.24023714661598206,
-0.12885290384292603,
0.13538125157356262,
0.053403254598379135,
0.0030310465954244137,
-0.01399570144712925,
0.1315125674009323,
-0.021797573193907738,
0.1024475023150444,
0.2739783227443695,
-0.25322529673576355,
-0.010178772732615471,
-0.05497116968035698,
0.025915829464793205,
-0.05491338297724724,
-0.07798278331756592,
-0.03055965155363083,
0.0011877523502334952,
0.03656814619898796,
0.10937156528234482,
-0.027761679142713547,
-0.011256376281380653,
-0.06585794687271118,
-0.16706448793411255,
-0.05667265132069588,
0.009395605884492397,
0.006422834470868111,
-0.11433397978544235,
-0.13469940423965454,
-0.0777079313993454,
-0.028535936027765274,
-0.07338167726993561,
0.0164161529392004,
0.008116135373711586,
0.0033872511703521013,
0.07742544263601303,
-0.06011240556836128,
-0.05353892967104912,
-0.005297193303704262,
-0.05240635573863983,
0.17509743571281433,
0.0547267384827137,
0.010174150578677654,
-0.05038491263985634,
0.05312125012278557,
-0.16328538954257965,
-0.09796810150146484,
-0.05222005397081375,
-0.07414095103740692,
-0.07239633798599243,
-0.020173851400613785,
-0.026419268921017647,
-0.2143239974975586,
-0.03875771909952164,
0.21598637104034424,
-0.10344083607196808,
0.06714069098234177,
-0.1527629792690277,
0.0386815071105957,
0.01422026939690113,
0.10706017166376114,
-0.09526859223842621,
0.09090418368577957,
0.026210976764559746,
0.02120141126215458,
0.0038376757875084877,
-0.04634750261902809,
0.009663492441177368,
0.08582809567451477,
0.023847628384828568,
0.03734150528907776,
0.045383092015981674,
0.08064950257539749,
-0.06275366246700287,
-0.042710427194833755,
0.16599518060684204,
-0.15146586298942566,
0.019146475940942764,
0.01870676316320896,
-0.024260155856609344,
-0.0371098667383194,
0.07111024856567383,
-0.046655360609292984,
-0.11241098493337631,
0.10352036356925964,
-0.04689731076359749,
-0.019825445488095284,
-0.03816835954785347,
-0.17033204436302185,
0.008827226236462593,
-0.06556632369756699,
-0.09246677905321121,
-0.011509446427226067,
-0.08031734079122543,
-0.009881095960736275,
0.07152464985847473,
-0.005408904515206814,
0.05139679089188576,
-0.011237211525440216,
-0.05628899112343788,
0.025723814964294434,
0.04505803436040878,
-0.012099724262952805,
0.0007605130667798221,
0.025670524686574936,
-0.010379797779023647,
0.08783318847417831,
0.027804555371403694,
0.014454703778028488,
-0.051468752324581146,
0.05378050357103348,
-0.28227928280830383,
0.06671958416700363,
-0.10687236487865448,
-0.09950369596481323,
-0.0776059478521347,
-0.06238848716020584,
-0.025997288525104523,
-0.008059604093432426,
0.03582329675555229,
0.14051593840122223,
-0.1596592664718628,
-0.05518638715147972,
0.2829727232456207,
-0.1303163766860962,
0.028972167521715164,
0.13350529968738556,
-0.053812846541404724,
-0.040512632578611374,
0.02787865325808525,
0.15400102734565735,
-0.023335132747888565,
-0.17560267448425293,
-0.01134188286960125,
-0.059657201170921326,
-0.015845229849219322,
0.1756686568260193,
0.06165381893515587,
-0.05888708308339119,
0.06034056842327118,
0.0048727234825491905,
-0.04295628145337105,
-0.08135946094989777,
-0.024344725534319878,
-0.03407374396920204,
0.00859373714774847,
0.00977044552564621,
0.07994694262742996,
-0.0541708804666996,
-0.013235509395599365,
-0.05524790287017822,
-0.09228874742984772,
-0.06386640667915344,
0.04496641084551811,
-0.07454599440097809,
0.08350690454244614,
-0.07815075665712357,
0.1015094667673111,
-0.012317189015448093,
0.023776959627866745,
-0.21155209839344025,
-0.0191126000136137,
0.08792262524366379,
-0.11591916531324387,
0.037296511232852936,
-0.028142565861344337,
0.04210653156042099,
0.0815669372677803,
-0.052031632512807846,
-0.001386136980727315,
-0.03363979235291481,
-0.04485993832349777,
-0.0627434030175209,
-0.16748927533626556,
-0.06270036101341248,
-0.06299711763858795,
0.010145383886992931,
-0.046862512826919556,
0.03603888303041458,
0.02627391368150711,
0.08711986243724823,
0.03514697030186653,
-0.051629625260829926,
0.026504218578338623,
-0.00046325798030011356,
-0.027337193489074707,
-0.031706538051366806,
-0.013317739591002464,
0.02287021093070507,
-0.002746132668107748,
0.13962770998477936,
-0.15864698588848114,
-0.10874614864587784,
0.06298030912876129,
0.06440801918506622,
-0.08542723953723907,
0.029274536296725273,
-0.060880549252033234,
0.01563912257552147,
-0.049893055111169815,
0.0069137816317379475,
0.17443732917308807,
0.0478915199637413,
0.14362718164920807,
-0.06681738048791885,
-0.01114121824502945,
0.0418156273663044,
-0.004176169168204069,
-0.09554029256105423,
0.055709727108478546,
-0.024143161252141,
-0.10109300166368484,
0.050064072012901306,
-0.04925931245088577,
0.06828182190656662,
0.09725053608417511,
-0.008003286086022854,
-0.0964212417602539,
-0.06349979341030121,
0.05415915325284004,
0.07235915958881378,
0.08316905051469803,
0.022997582331299782,
0.010099820792675018,
0.08177686482667923,
0.04069485515356064,
0.0271091777831316,
-0.07987174391746521,
0.059414952993392944,
0.06271633505821228,
-0.027571868151426315,
-0.08267057687044144,
-0.018607409670948982,
0.027455154806375504,
0.099863700568676,
0.021624170243740082,
0.026943758130073547,
-0.03885194659233093,
-0.027828309684991837,
-0.08013005554676056,
0.1597585529088974,
-0.12858393788337708,
-0.23613305389881134,
-0.15396441519260406,
-0.06609634310007095,
-0.043687913566827774,
0.04208178445696831,
0.06474273651838303,
-0.01542885135859251,
-0.08142628520727158,
-0.1012234315276146,
0.03915518894791603,
-0.004087105393409729,
-0.0032715261913836002,
0.0385480672121048,
-0.06947419047355652,
0.03645993024110794,
-0.14376838505268097,
-0.025903897359967232,
-0.009863142855465412,
0.039599575102329254,
0.04916936531662941,
0.021375827491283417,
0.10668128728866577,
0.08709923923015594,
-0.00806447397917509,
0.00024632542044855654,
-0.0361468531191349,
0.18356619775295258,
-0.060598164796829224,
0.07967719435691833,
0.016597798094153404,
-0.07831062376499176,
0.12429367750883102,
0.06585169583559036,
0.0024800309911370277,
-0.06359836459159851,
-0.004548979923129082,
0.06313393265008926,
-0.035094697028398514,
-0.2279394418001175,
-0.015937382355332375,
-0.04802478849887848,
0.03575640544295311,
0.09041113406419754,
0.03848110884428024,
0.030289916321635246,
-0.03732447326183319,
-0.09795568138360977,
-0.006218831054866314,
0.06720910966396332,
0.1031751036643982,
-0.04279840737581253,
-0.013690667226910591,
0.08086207509040833,
-0.04891398921608925,
-0.015445621684193611,
0.08085103332996368,
-0.05766119807958603,
0.13784247636795044,
-0.0480380542576313,
0.19922897219657898,
0.1077343225479126,
0.040160927921533585,
0.02431841753423214,
0.15070828795433044,
-0.0660848319530487,
0.03725374490022659,
-0.029555590823292732,
-0.07264284044504166,
0.012659692205488682,
0.007637558039277792,
0.007564081810414791,
-0.007597476709634066,
-0.09747279435396194,
-0.024163594469428062,
0.06660313159227371,
0.2880840301513672,
0.034353725612163544,
-0.1150619313120842,
-0.08102244883775711,
-0.005385955795645714,
-0.03608572855591774,
-0.061926908791065216,
-0.0487096793949604,
0.03818075731396675,
-0.10015642642974854,
0.09023429453372955,
-0.06536151468753815,
0.0875980332493782,
-0.02604423649609089,
0.012455393560230732,
-0.07726864516735077,
0.026901274919509888,
-0.03213633596897125,
0.03691813349723816,
-0.18734893202781677,
0.19763745367527008,
0.0597897469997406,
0.040348559617996216,
-0.0948561429977417,
0.044490646570920944,
-0.005151356104761362,
0.0005061870906502008,
0.18142080307006836,
0.00040908955270424485,
-0.07156725972890854,
-0.063368059694767,
-0.05056830123066902,
-0.037200357764959335,
0.158786803483963,
-0.007862322963774204,
0.11508572101593018,
0.03774850815534592,
-0.025114746764302254,
0.011703078635036945,
0.10541470348834991,
-0.07836329936981201,
-0.1547905057668686,
0.09183057397603989,
-0.10034766793251038,
-0.1299218386411667,
-0.06824725866317749,
-0.07030791789293289,
-0.13805794715881348,
0.18633458018302917,
-0.074905164539814,
-0.009385260753333569,
-0.14310462772846222,
-0.027071423828601837,
0.09443897753953934,
-0.10559144616127014,
0.03905096650123596,
-0.06568122655153275,
0.15881872177124023,
-0.13868418335914612,
-0.0992833822965622,
0.08819664269685745,
-0.06285540014505386,
-0.1249719187617302,
-0.059734851121902466,
0.11206851899623871,
0.12209577858448029,
0.04852120205760002,
-0.001197535777464509,
0.05442117899656296,
0.07410480082035065,
-0.08006606996059418,
0.017028870061039925,
0.017288556322455406,
0.10815107822418213,
0.02366064116358757,
-0.09814003109931946,
-0.007213374134153128,
-0.0882011204957962,
0.06391803175210953,
0.08310677856206894,
0.3353902995586395,
-0.07252423465251923,
0.16271494328975677,
0.14252664148807526,
-0.07668792456388474,
-0.20947521924972534,
-0.021347226575016975,
0.026913313195109367,
0.02796291373670101,
0.04092103987932205,
-0.17783817648887634,
-0.06355960667133331,
0.0013000473845750093,
-0.026130441576242447,
-0.0126236816868186,
-0.1777590960264206,
-0.10271882265806198,
0.08665873855352402,
0.043620526790618896,
0.0065283640287816525,
-0.09256807714700699,
-0.046878669410943985,
-0.00457672169432044,
-0.10361000895500183,
0.056024886667728424,
-0.013085621409118176,
0.12116934359073639,
0.004780254792422056,
-0.10930193215608597,
0.02956816554069519,
-0.07464984059333801,
0.1123441830277443,
-0.01977194845676422,
0.08985209465026855,
-0.05692250281572342,
0.013511055149137974,
0.14900988340377808,
-0.054388049989938736,
0.14121153950691223,
0.044394392520189285,
0.04109340161085129,
-0.048373330384492874,
-0.05279785767197609,
-0.06245533004403114,
0.05107474699616432,
-0.07461639493703842,
-0.027469297870993614,
-0.05310722440481186,
0.04848920553922653,
0.12685120105743408,
-0.019933098927140236,
0.08295408636331558,
0.010572533123195171,
0.10640636831521988,
0.22146284580230713,
-0.01502812746912241,
0.03494073823094368,
-0.06962869316339493,
0.008946234360337257,
-0.006378259044140577,
0.036226239055395126,
-0.08927065879106522,
0.06682566553354263,
0.08172203600406647,
0.042319245636463165,
0.16358520090579987,
0.028316184878349304,
-0.17607273161411285,
-0.0473959855735302,
0.02111664041876793,
-0.14700847864151,
-0.18069210648536682,
0.020382406190037727,
-0.04149399697780609,
-0.0992322564125061,
0.0017833077581599355,
0.08241505175828934,
-0.0011721700429916382,
-0.046027421951293945,
-0.012518768198788166,
0.046917106956243515,
0.01707732118666172,
0.12415369600057602,
0.016802750527858734,
0.06322890520095825,
-0.07572177797555923,
0.08828515559434891,
0.10153067857027054,
-0.13533169031143188,
0.05686407908797264,
0.05569439381361008,
-0.07820814102888107,
-0.030651092529296875,
0.08646082878112793,
0.1007210984826088,
0.12700505554676056,
0.026836631819605827,
-0.10670231282711029,
-0.051309455186128616,
0.10241896659135818,
0.1838139146566391,
0.022932013496756554,
0.01699824072420597,
-0.0009611759451217949,
-0.005334113724529743,
-0.07186871767044067,
0.05726674571633339,
0.0448339618742466,
0.03151725232601166,
0.060971491038799286,
0.12230922281742096,
-0.019359102472662926,
0.05889289826154709,
-0.012284096330404282,
-0.04598497971892357,
-0.0842597559094429,
-0.02883070707321167,
-0.04671144485473633,
0.004115395713597536,
-0.08828680962324142,
-0.028980402275919914,
-0.02805166132748127,
0.03278094902634621,
0.0392017662525177,
0.05443492904305458,
-0.08304021507501602,
-0.036961816251277924,
-0.041627123951911926,
0.0011705616489052773,
-0.15885372459888458,
0.008719800040125847,
0.04749401658773422,
-0.07611341774463654,
0.0813336968421936,
0.039800092577934265,
-0.03396344557404518,
0.046178970485925674,
-0.06241088733077049,
-0.01585695706307888,
-0.02105962485074997,
-0.010883974842727184,
0.017934100702404976,
-0.1186569407582283,
0.016981730237603188,
-0.05976860225200653,
-0.05016003176569939,
-0.012023200280964375,
0.07823647558689117,
-0.10828373581171036,
-0.0005995439132675529,
0.038201238960027695,
-0.050452299416065216,
-0.028795145452022552,
0.11744943261146545,
0.04125773534178734,
-0.01021589059382677,
0.1024302989244461,
-0.04549725353717804,
0.07335717231035233,
-0.10074403136968613,
-0.02576710470020771,
0.001072047045454383,
-0.05448366701602936,
0.11990317702293396,
-0.04699785262346268,
0.03629770874977112,
-0.036475714296102524,
0.10681664198637009,
-0.05085902661085129,
-0.03252386301755905,
0.013116220943629742,
-0.047490306198596954,
-0.08789572864770889,
0.027507638558745384,
0.022560378536581993,
-0.05539262294769287,
-0.06943154335021973,
0.03815314173698425,
0.036775846034288406,
0.017720356583595276,
0.15225407481193542,
0.18536700308322906,
0.12625649571418762,
0.13579820096492767,
0.022001221776008606,
0.02848350629210472,
-0.05781369283795357,
-0.09623599052429199,
-0.043217457830905914,
0.02232157625257969,
0.08444730192422867,
-0.00278739002533257,
-0.030395152047276497,
0.09834554046392441,
-0.13139191269874573,
0.19244931638240814,
0.03308645635843277,
-0.06233701854944229,
-0.08219818770885468,
-0.2261313945055008,
-0.020105749368667603,
0.0671517625451088,
-0.004894794896245003,
-0.10086239129304886,
0.05543721839785576,
-0.03630228713154793,
0.04245966672897339,
0.015983201563358307,
0.12757064402103424,
-0.12569718062877655,
-0.060717325657606125,
0.08646707981824875,
-0.002724743215367198,
-0.06538991630077362,
0.09519529342651367,
-0.007444037590175867,
0.037833407521247864,
0.011956132017076015,
0.06479208916425705,
0.051378607749938965,
0.11107685416936874,
0.03836570680141449,
-0.06732919067144394,
-0.09800158441066742,
0.043532006442546844,
-0.022174913436174393,
0.0569736547768116,
0.22106319665908813,
0.02599242515861988,
-0.040562357753515244,
-0.01581711508333683,
0.10791146010160446,
-0.0187229011207819,
-0.07740899175405502,
-0.15278717875480652,
0.18583831191062927,
0.0731990709900856,
0.00641772523522377,
0.0076959701254963875,
-0.10091305524110794,
-0.022714365273714066,
0.19675381481647491,
0.17682857811450958,
0.07894260436296463,
0.06011941283941269,
0.035742055624723434,
0.007595642935484648,
0.08625811338424683,
0.02564554661512375,
0.039802368730306625,
0.07414482533931732,
-0.00877761747688055,
0.041036609560251236,
-0.05290122702717781,
-0.03725820034742355,
-0.04803069680929184,
0.17605628073215485,
0.007108672987669706,
-0.013499867171049118,
-0.05459071695804596,
0.012450768612325191,
0.003111298428848386,
-0.2592938244342804,
-0.024055704474449158,
-0.09063160419464111,
-0.11535383015871048,
-0.0035498221404850483,
-0.0695076733827591,
0.038260165601968765,
0.08520223200321198,
0.06321877241134644,
-0.012289637699723244,
0.20652809739112854,
0.03389843553304672,
-0.012417060323059559,
-0.07257089763879776,
0.03928513452410698,
-0.15838338434696198,
0.1955021768808365,
0.07029462605714798,
0.0062413448467850685,
0.11987299472093582,
0.005583424586802721,
-0.06181890144944191,
0.045237645506858826,
0.027626272290945053,
0.018370099365711212,
-0.02937506139278412,
0.19683299958705902,
-0.012648436240851879,
-0.03161856532096863,
-0.003585538361221552,
-0.1006641834974289,
0.06310900300741196,
-0.11803949624300003,
-0.052048247307538986,
-0.060140449553728104,
0.08627867698669434,
-0.07969646155834198,
0.08128221333026886,
0.20717179775238037,
0.019680267199873924,
0.005136874038726091,
-0.0767279788851738,
-0.014154604636132717,
0.007332661189138889,
0.04697529599070549,
-0.03806417062878609,
-0.19388775527477264,
0.04737045243382454,
-0.04374242573976517,
0.06101393699645996,
-0.24061523377895355,
-0.07307584583759308,
0.056562915444374084,
-0.07962829619646072,
0.009198267944157124,
0.07023456692695618,
0.0343620739877224,
0.024302758276462555,
-0.04297609254717827,
-0.05238189175724983,
0.012070182710886002,
0.08260146528482437,
-0.07417739927768707,
0.017056994140148163
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-large-xls-r-1b-bemba-fds
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-1b](https://huggingface.co/facebook/wav2vec2-xls-r-1b) on the [BembaSpeech](https://github.com/csikasote/BembaSpeech) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2898
- Wer: 0.3435
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 4
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 8
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 15
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 1.7986 | 0.34 | 500 | 0.4549 | 0.7292 |
| 0.5358 | 0.67 | 1000 | 0.3325 | 0.4491 |
| 0.4559 | 1.01 | 1500 | 0.3090 | 0.3954 |
| 0.3983 | 1.35 | 2000 | 0.3067 | 0.4105 |
| 0.4067 | 1.68 | 2500 | 0.2838 | 0.3678 |
| 0.3722 | 2.02 | 3000 | 0.2824 | 0.3762 |
| 0.3286 | 2.36 | 3500 | 0.2810 | 0.3670 |
| 0.3239 | 2.69 | 4000 | 0.2643 | 0.3501 |
| 0.3187 | 3.03 | 4500 | 0.2838 | 0.3754 |
| 0.2801 | 3.36 | 5000 | 0.2815 | 0.3507 |
| 0.2806 | 3.7 | 5500 | 0.2725 | 0.3486 |
| 0.2714 | 4.04 | 6000 | 0.2898 | 0.3435 |
### Framework versions
- Transformers 4.16.2
- Pytorch 1.10.0+cu111
- Datasets 1.18.3
- Tokenizers 0.11.0
|
{"license": "apache-2.0", "tags": ["generated_from_trainer", "bem", "robust-speech-event"], "model-index": [{"name": "wav2vec2-large-xls-r-1b-bemba-fds", "results": []}]}
|
automatic-speech-recognition
|
csikasote/wav2vec2-large-xls-r-1b-bemba-fds
|
[
"transformers",
"pytorch",
"tensorboard",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"bem",
"robust-speech-event",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us
|
wav2vec2-large-xls-r-1b-bemba-fds
=================================
This model is a fine-tuned version of facebook/wav2vec2-xls-r-1b on the BembaSpeech dataset.
It achieves the following results on the evaluation set:
* Loss: 0.2898
* Wer: 0.3435
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-05
* train\_batch\_size: 4
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 8
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* num\_epochs: 15
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.16.2
* Pytorch 1.10.0+cu111
* Datasets 1.18.3
* Tokenizers 0.11.0
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0"
] |
[
66,
159,
4,
35
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 8\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 15\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.16.2\n* Pytorch 1.10.0+cu111\n* Datasets 1.18.3\n* Tokenizers 0.11.0"
] |
[
-0.11465804278850555,
0.0729159489274025,
-0.003411562880501151,
0.05645596608519554,
0.1167273223400116,
0.006166545208543539,
0.08970609307289124,
0.14165210723876953,
-0.07306689023971558,
0.08457065373659134,
0.10702508687973022,
0.09272751957178116,
0.05682758986949921,
0.12470467388629913,
-0.02086191624403,
-0.31695669889450073,
0.008644022047519684,
0.013749250210821629,
-0.1080833449959755,
0.1188637763261795,
0.09670364111661911,
-0.1104830801486969,
0.01604842022061348,
0.017682787030935287,
-0.1203666478395462,
0.010089853778481483,
-0.028642775490880013,
-0.07182403653860092,
0.11872098594903946,
0.03271980211138725,
0.0941983237862587,
0.01801241934299469,
0.08845213055610657,
-0.26146501302719116,
0.017468051984906197,
0.060490284115076065,
0.04079675301909447,
0.07159990072250366,
0.11077789217233658,
-0.01215266901999712,
0.11997175216674805,
-0.07529519498348236,
0.07582072168588638,
0.03586238995194435,
-0.10904458165168762,
-0.30871808528900146,
-0.079629085958004,
0.040390171110630035,
0.12419858574867249,
0.0947251170873642,
-0.030139686539769173,
0.07717125117778778,
-0.08391585201025009,
0.08233310282230377,
0.2338806837797165,
-0.2553621232509613,
-0.07997356355190277,
-0.042304519563913345,
0.04617169499397278,
0.0484105683863163,
-0.12355860322713852,
-0.03512994945049286,
0.02445291168987751,
0.03886450082063675,
0.11416621506214142,
0.004651391878724098,
-0.03317520394921303,
0.018279917538166046,
-0.1564047634601593,
-0.04874109476804733,
0.09482134133577347,
0.07333564013242722,
-0.021545208990573883,
-0.09844724088907242,
-0.019445078447461128,
-0.18673738837242126,
-0.055682770907878876,
0.005414553452283144,
0.03029453381896019,
-0.04023255035281181,
-0.10070443898439407,
0.013943028636276722,
-0.08755800127983093,
-0.08377258479595184,
0.014507779851555824,
0.13889020681381226,
0.049072638154029846,
-0.025331629440188408,
-0.0004927952541038394,
0.10312613099813461,
0.039224572479724884,
-0.1331842839717865,
0.003712362376973033,
0.041197482496500015,
-0.10960282385349274,
-0.01897326484322548,
-0.03130091354250908,
-0.02863425575196743,
-0.0017270103562623262,
0.11698354780673981,
-0.030381467193365097,
0.08707468956708908,
0.024512337520718575,
0.03189198672771454,
-0.08301780372858047,
0.1762590855360031,
-0.06963635981082916,
-0.028689783066511154,
-0.057863641530275345,
0.09910757094621658,
-0.022618943825364113,
-0.008321649394929409,
-0.06389568746089935,
0.032693710178136826,
0.10189776122570038,
0.04133057966828346,
-0.01868143118917942,
0.023220181465148926,
-0.06539179384708405,
-0.02705351449549198,
-0.035207122564315796,
-0.10217826068401337,
0.042340755462646484,
0.03196457028388977,
-0.08098205178976059,
0.005847196560353041,
0.008040736429393291,
0.017546167597174644,
-0.03202008083462715,
0.1363699734210968,
-0.060650669038295746,
0.02123217098414898,
-0.08559124171733856,
-0.10204990208148956,
0.03434909135103226,
-0.048630062490701675,
-0.001579316915012896,
-0.07215074449777603,
-0.10872471332550049,
-0.046992696821689606,
0.06687621027231216,
-0.052045729011297226,
-0.07044627517461777,
-0.06542174518108368,
-0.06382743269205093,
0.048201221972703934,
-0.03310520574450493,
0.17923149466514587,
-0.06275186687707901,
0.10930201411247253,
0.009898796677589417,
0.04393261671066284,
0.045221272855997086,
0.07575681805610657,
-0.03808322921395302,
0.035179175436496735,
-0.13919302821159363,
0.0759599581360817,
-0.0879511684179306,
0.05317996069788933,
-0.15111590921878815,
-0.12070140987634659,
-0.0273148063570261,
0.007198154926300049,
0.09825922548770905,
0.0946950688958168,
-0.1697043627500534,
-0.09949410706758499,
0.1888808012008667,
-0.0636448785662651,
-0.08538898825645447,
0.13942573964595795,
-0.03193500265479088,
-0.0033397749066352844,
0.0357743538916111,
0.17291919887065887,
0.07795273512601852,
-0.09831254929304123,
0.028188295662403107,
-0.04748594015836716,
0.10749498009681702,
0.01837144047021866,
0.09269227832555771,
-0.045231323689222336,
0.040062423795461655,
-0.006286522839218378,
-0.03188760578632355,
0.07762469351291656,
-0.08819938451051712,
-0.08365539461374283,
-0.014384817332029343,
-0.07356658577919006,
0.02287701517343521,
0.057874131947755814,
0.04168354347348213,
-0.08838514983654022,
-0.1314988136291504,
0.03825211897492409,
0.09806649386882782,
-0.09878745675086975,
0.034033410251140594,
-0.07617934793233871,
0.0376780666410923,
-0.03052746318280697,
-0.00965869054198265,
-0.17085058987140656,
0.0009372279746457934,
0.02661212719976902,
-0.04143603518605232,
0.030777109786868095,
-0.0032901172526180744,
0.08567831665277481,
0.02942909114062786,
-0.04472601041197777,
-0.07264291495084763,
-0.06535214930772781,
-0.006914857774972916,
-0.08942613750696182,
-0.2281254678964615,
-0.07109202444553375,
-0.03306642919778824,
0.15681514143943787,
-0.22735461592674255,
0.006822066381573677,
0.014779395423829556,
0.10274270921945572,
0.03748010843992233,
-0.04477597400546074,
-0.010423821397125721,
0.08078636974096298,
-0.00904101599007845,
-0.06199902668595314,
0.045869309455156326,
-0.002089110901579261,
-0.11663604527711868,
0.03044520691037178,
-0.11194563657045364,
0.0929076299071312,
0.10881432145833969,
-0.024921398609876633,
-0.08674421161413193,
-0.06010991334915161,
-0.0667123794555664,
-0.06847027689218521,
-0.03197035565972328,
0.0055082389153540134,
0.22282421588897705,
0.039874766021966934,
0.12354154884815216,
-0.07152245938777924,
-0.057848576456308365,
0.029067683964967728,
0.004046874586492777,
-0.01600673981010914,
0.14721687138080597,
0.07891273498535156,
-0.03861532732844353,
0.09765312820672989,
0.10653368383646011,
-0.08523266017436981,
0.14924311637878418,
-0.0682525485754013,
-0.12872377038002014,
-0.01713109388947487,
0.030145512893795967,
0.02617504447698593,
0.11897778511047363,
-0.15225085616111755,
-0.01231789868324995,
0.012990618124604225,
0.026569994166493416,
0.029713785275816917,
-0.21502336859703064,
-0.011365440674126148,
0.04339442774653435,
-0.06994078308343887,
-0.02636057883501053,
-0.021857235580682755,
0.007816189900040627,
0.09266335517168045,
0.009805914014577866,
-0.07301336526870728,
-0.008978160098195076,
-0.02535031922161579,
-0.08181121200323105,
0.18968629837036133,
-0.09315365552902222,
-0.14080514013767242,
-0.11690033227205276,
-0.038552042096853256,
0.012376666069030762,
-0.02406039834022522,
0.04863140732049942,
-0.11400895565748215,
-0.0323059968650341,
-0.06206660345196724,
0.04691091552376747,
-0.0588332824409008,
0.030682072043418884,
-0.0022099833004176617,
0.005909142550081015,
0.07713494449853897,
-0.09788522124290466,
0.02251456491649151,
-0.03687961772084236,
-0.03416372835636139,
0.03535033017396927,
0.037076640874147415,
0.088966004550457,
0.17458751797676086,
0.037667158991098404,
0.022147314622998238,
-0.04370063543319702,
0.1360781341791153,
-0.10045673698186874,
-0.03533150255680084,
0.10981252044439316,
-0.012900268658995628,
0.04497615247964859,
0.11990101635456085,
0.05808635801076889,
-0.07652527093887329,
0.017612114548683167,
0.049648355692625046,
-0.012775846756994724,
-0.24490858614444733,
-0.033738065510988235,
-0.06024482473731041,
-0.013661086559295654,
0.12493278086185455,
0.0319879949092865,
-0.030148280784487724,
0.024877365678548813,
0.00038008380215615034,
-0.010481288656592369,
-0.0018099669832736254,
0.059515345841646194,
0.06250002235174179,
0.034609466791152954,
0.11636824160814285,
-0.015246170572936535,
-0.05010418966412544,
0.025503262877464294,
0.004829812794923782,
0.25913628935813904,
-0.00002113589653163217,
0.16799314320087433,
0.05143214017152786,
0.1585153490304947,
0.013110645115375519,
0.0614105723798275,
0.008158217184245586,
-0.027765445411205292,
0.013531479984521866,
-0.04886045306921005,
-0.015871815383434296,
0.0477352999150753,
0.10790809243917465,
0.030619872733950615,
-0.11673513799905777,
-0.03547490015625954,
0.025883955880999565,
0.35830268263816833,
0.06122482195496559,
-0.28662410378456116,
-0.08033651113510132,
0.0027683842927217484,
-0.09504377096891403,
-0.038235217332839966,
0.03147856518626213,
0.11149973422288895,
-0.09588140994310379,
0.04095664620399475,
-0.08359957486391068,
0.10061703622341156,
-0.05957798287272453,
0.005650090053677559,
0.09092158079147339,
0.07738037407398224,
0.0030934421811252832,
0.05793919786810875,
-0.2666623294353485,
0.30875128507614136,
-0.01928752474486828,
0.07001787424087524,
-0.04718998819589615,
0.03159574419260025,
0.03347394987940788,
-0.0680202841758728,
0.0763864740729332,
-0.01799360290169716,
-0.0964474007487297,
-0.18794690072536469,
-0.08248554915189743,
0.022296791896224022,
0.12869687378406525,
-0.04282720014452934,
0.12166975438594818,
-0.021652724593877792,
-0.009814586490392685,
0.06052037328481674,
-0.07758308947086334,
-0.09337455779314041,
-0.10886475443840027,
0.025225773453712463,
0.0161589328199625,
0.05695087090134621,
-0.10088509321212769,
-0.11603646725416183,
-0.08493371307849884,
0.1581026017665863,
-0.09952891618013382,
-0.010108173824846745,
-0.12803015112876892,
0.09493040293455124,
0.15948888659477234,
-0.06869146227836609,
0.052216824144124985,
0.028950160369277,
0.10829783231019974,
0.0258466973900795,
-0.013217565603554249,
0.12339247763156891,
-0.08515296876430511,
-0.1805020421743393,
-0.057040926069021225,
0.1582033485174179,
0.04713145270943642,
0.06842582672834396,
-0.024611515924334526,
0.021077850833535194,
-0.02122957818210125,
-0.07866830378770828,
0.06269138306379318,
0.0070550814270973206,
0.011770983226597309,
0.05388942360877991,
-0.03803795203566551,
-0.012445102445781231,
-0.08088935911655426,
-0.06406118720769882,
0.1680067926645279,
0.2763645052909851,
-0.08723480999469757,
0.039395254105329514,
0.03939918056130409,
-0.04854800924658775,
-0.15163478255271912,
0.025840414687991142,
0.14123649895191193,
0.03279050439596176,
0.01629791408777237,
-0.2261388748884201,
0.05704843997955322,
0.0900326743721962,
-0.02135155349969864,
0.07912681251764297,
-0.34417444467544556,
-0.13178205490112305,
0.12466740608215332,
0.10615707188844681,
-0.02286899834871292,
-0.15823805332183838,
-0.05385438725352287,
0.007116110064089298,
-0.08417138457298279,
0.0668724998831749,
-0.041439007967710495,
0.12175412476062775,
-0.0004911621217615902,
0.06833121180534363,
0.018702805042266846,
-0.05559452250599861,
0.14005762338638306,
-0.015934228897094727,
0.054680489003658295,
-0.006386719178408384,
0.043917857110500336,
-0.012733899988234043,
-0.04636942222714424,
0.007204148918390274,
-0.06961990892887115,
0.00434862170368433,
-0.13828659057617188,
-0.028931856155395508,
-0.08625278621912003,
0.023634405806660652,
-0.037344470620155334,
-0.042113956063985825,
0.004308102186769247,
0.041049446910619736,
0.06180603429675102,
0.0057421294040977955,
0.11786676198244095,
-0.05680074915289879,
0.1428832709789276,
0.07794640213251114,
0.09179182350635529,
-0.0038341328036040068,
-0.11457566916942596,
-0.021404245868325233,
-0.00882038939744234,
0.05385178327560425,
-0.11863362789154053,
0.02721124328672886,
0.1454533040523529,
0.04411330074071884,
0.153669536113739,
0.06116523593664169,
-0.07936936616897583,
0.02028798684477806,
0.063605397939682,
-0.08400379121303558,
-0.11589139699935913,
-0.0022862947080284357,
0.06656406074762344,
-0.13528025150299072,
-0.001793139846995473,
0.10693079233169556,
-0.04954594001173973,
-0.012932416051626205,
0.013385909609496593,
0.01976807788014412,
-0.05273105949163437,
0.23116420209407806,
0.026781344786286354,
0.07347119599580765,
-0.09077958762645721,
0.079373799264431,
0.06021811440587044,
-0.18612957000732422,
0.027225831523537636,
0.0896759033203125,
-0.03062223084270954,
-0.021516527980566025,
0.030065637081861496,
0.07806956022977829,
0.020655840635299683,
-0.054434556514024734,
-0.11046270281076431,
-0.14972352981567383,
0.09341984987258911,
0.10901831090450287,
0.03405376523733139,
0.02205805294215679,
-0.045820947736501694,
0.04814862087368965,
-0.10193610191345215,
0.08900731801986694,
0.10672987252473831,
0.06962910294532776,
-0.12947487831115723,
0.15112201869487762,
0.012357527390122414,
0.013198789209127426,
0.0013188612647354603,
-0.004249102436006069,
-0.10612910240888596,
0.03373146057128906,
-0.10775202512741089,
-0.0212356299161911,
-0.04826769232749939,
-0.0034540712367743254,
0.010564284399151802,
-0.05062839388847351,
-0.047148656100034714,
0.015634896233677864,
-0.1240227222442627,
-0.039637576788663864,
-0.006906122900545597,
0.06637855619192123,
-0.10561075806617737,
-0.028990864753723145,
0.0438205748796463,
-0.10511544346809387,
0.10308638960123062,
0.06234126538038254,
0.020992159843444824,
0.03801307454705238,
-0.1357487142086029,
0.00665896013379097,
0.04002389684319496,
-0.0027673854492604733,
0.0239156112074852,
-0.15673336386680603,
-0.012788317166268826,
-0.02935931645333767,
0.038234539330005646,
-0.005211932584643364,
0.028872564435005188,
-0.13593384623527527,
-0.0394141711294651,
-0.03293919563293457,
-0.07213085889816284,
-0.05868282541632652,
0.04416143521666527,
0.0432087779045105,
0.04572809487581253,
0.15897540748119354,
-0.09119901806116104,
0.05427839607000351,
-0.21388612687587738,
0.015857785940170288,
-0.03424825519323349,
-0.06355158239603043,
-0.05466584116220474,
-0.030466757714748383,
0.08793677389621735,
-0.0689416229724884,
0.0870811715722084,
-0.038802195340394974,
0.03561116382479668,
0.03726715222001076,
-0.121780626475811,
0.02228458970785141,
0.04998272657394409,
0.21091605722904205,
0.04472191631793976,
-0.02356491982936859,
0.055778998881578445,
0.012624809518456459,
0.05729154869914055,
0.15730193257331848,
0.1587958037853241,
0.19344058632850647,
0.061144519597291946,
0.09419524669647217,
0.06006782129406929,
-0.11463050544261932,
-0.11116614192724228,
0.1278000921010971,
-0.03475121408700943,
0.12978215515613556,
-0.019207848235964775,
0.2534855306148529,
0.09744298458099365,
-0.19280439615249634,
0.047736089676618576,
-0.0359317809343338,
-0.0852498859167099,
-0.09182455390691757,
-0.04217104613780975,
-0.06380519270896912,
-0.17535658180713654,
0.010174648836255074,
-0.10154163092374802,
0.057794034481048584,
0.0592498853802681,
0.046883925795555115,
0.022670958191156387,
0.12295301258563995,
0.0526496097445488,
-0.006831868086010218,
0.1301276683807373,
0.016083097085356712,
-0.009887893684208393,
-0.055895786732435226,
-0.08547303080558777,
0.030557259917259216,
-0.03102712333202362,
0.044698312878608704,
-0.05051182582974434,
-0.1089639887213707,
0.0627068504691124,
0.00767051987349987,
-0.10593350976705551,
0.01965685747563839,
-0.013170073740184307,
0.07301969826221466,
0.0961654856801033,
0.03366152569651604,
-0.007710068020969629,
-0.02927478775382042,
0.2534811496734619,
-0.10421669483184814,
-0.053058747202157974,
-0.12722285091876984,
0.2409396916627884,
0.010610807687044144,
-0.015738368034362793,
0.010018600150942802,
-0.06611084938049316,
-0.0013128221035003662,
0.16565552353858948,
0.13441741466522217,
-0.0418873094022274,
-0.011236431077122688,
0.01822277531027794,
-0.009532976895570755,
-0.05019773915410042,
0.08721230179071426,
0.13398852944374084,
0.07951795309782028,
-0.08063901960849762,
-0.04520833492279053,
-0.04388820752501488,
-0.04458744451403618,
-0.03970989212393761,
0.05586434528231621,
0.03372360020875931,
-0.00788616668432951,
-0.030714042484760284,
0.10977508127689362,
-0.06908083707094193,
-0.11668902635574341,
0.011216970160603523,
-0.18445949256420135,
-0.19061565399169922,
-0.0358768068253994,
0.07944762706756592,
0.03496045991778374,
0.051622726023197174,
-0.008183971047401428,
-0.023791372776031494,
0.08853431046009064,
0.00301368348300457,
-0.042341940104961395,
-0.10826966166496277,
0.08794477581977844,
-0.0967019572854042,
0.1787789762020111,
-0.04385281354188919,
0.032416652888059616,
0.11710511893033981,
0.07866038382053375,
-0.06696678698062897,
0.046546898782253265,
0.06976951658725739,
-0.1333034634590149,
0.04785196855664253,
0.18845361471176147,
-0.03645295277237892,
0.13342595100402832,
0.043881434947252274,
-0.1124177873134613,
0.027876203879714012,
-0.10619135200977325,
-0.059690993279218674,
-0.05666695907711983,
-0.00679200142621994,
-0.03924848884344101,
0.13430576026439667,
0.22071141004562378,
-0.06213665008544922,
-0.01892172545194626,
-0.06094592809677124,
0.0014529037289321423,
0.03220866620540619,
0.12829981744289398,
-0.05180111527442932,
-0.26491057872772217,
0.017862115055322647,
-0.004963963758200407,
0.019309109076857567,
-0.24263381958007812,
-0.09641294926404953,
0.040871307253837585,
-0.05532550811767578,
-0.05992060527205467,
0.11141109466552734,
0.07717228680849075,
0.05313652381300926,
-0.053489673882722855,
-0.0904872864484787,
-0.0340319499373436,
0.18386074900627136,
-0.1734144389629364,
-0.054290201514959335
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# wav2vec2-large-xls-r-300m-bemba-fds
This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the [BembaSpeech](https://github.com/csikasote/BembaSpeech) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3594
- Wer: 0.3838
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0003
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 16
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 500
- num_epochs: 30
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 2.9961 | 0.67 | 500 | 0.5157 | 0.7133 |
| 0.5903 | 1.34 | 1000 | 0.3663 | 0.4989 |
| 0.4804 | 2.02 | 1500 | 0.3547 | 0.4653 |
| 0.4146 | 2.69 | 2000 | 0.3274 | 0.4345 |
| 0.3792 | 3.36 | 2500 | 0.3586 | 0.4640 |
| 0.3509 | 4.03 | 3000 | 0.3360 | 0.4316 |
| 0.3114 | 4.7 | 3500 | 0.3382 | 0.4303 |
| 0.2935 | 5.38 | 4000 | 0.3263 | 0.4091 |
| 0.2723 | 6.05 | 4500 | 0.3348 | 0.4175 |
| 0.2502 | 6.72 | 5000 | 0.3317 | 0.4147 |
| 0.2334 | 7.39 | 5500 | 0.3542 | 0.4030 |
| 0.2287 | 8.06 | 6000 | 0.3594 | 0.4067 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.10.0+cu111
- Datasets 1.13.3
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer", "bem", "robust-speech-event"], "model-index": [{"name": "wav2vec2-large-xls-r-300m-bemba-fds", "results": []}]}
|
automatic-speech-recognition
|
csikasote/wav2vec2-large-xls-r-300m-bemba-fds
|
[
"transformers",
"pytorch",
"tensorboard",
"wav2vec2",
"automatic-speech-recognition",
"generated_from_trainer",
"bem",
"robust-speech-event",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us
|
wav2vec2-large-xls-r-300m-bemba-fds
===================================
This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the BembaSpeech dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3594
* Wer: 0.3838
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0003
* train\_batch\_size: 8
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 16
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* lr\_scheduler\_warmup\_steps: 500
* num\_epochs: 30
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.10.0+cu111
* Datasets 1.13.3
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3"
] |
[
66,
158,
4,
33
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #wav2vec2 #automatic-speech-recognition #generated_from_trainer #bem #robust-speech-event #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0003\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 16\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* num\\_epochs: 30\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.10.0+cu111\n* Datasets 1.13.3\n* Tokenizers 0.10.3"
] |
[
-0.11642204970121384,
0.07105162739753723,
-0.0032327650114893913,
0.05935138836503029,
0.10804925858974457,
0.008824776858091354,
0.09769383817911148,
0.1461581140756607,
-0.08202971518039703,
0.0781208798289299,
0.10415594279766083,
0.09022080153226852,
0.057156842201948166,
0.11077716201543808,
-0.01727232336997986,
-0.32496529817581177,
0.00855289027094841,
0.018603701144456863,
-0.10595814883708954,
0.11788330227136612,
0.0965743213891983,
-0.11080584675073624,
0.01853482984006405,
0.02176075428724289,
-0.10881074517965317,
0.008794841356575489,
-0.03086256980895996,
-0.07456693798303604,
0.12404712289571762,
0.02672526054084301,
0.09635687619447708,
0.01481279544532299,
0.08993694186210632,
-0.2617224156856537,
0.01805497705936432,
0.0601360909640789,
0.0452447384595871,
0.06894588470458984,
0.11001991480588913,
-0.01918674446642399,
0.1439713090658188,
-0.06951609253883362,
0.07789056748151779,
0.0381220206618309,
-0.11263908445835114,
-0.3121357560157776,
-0.07556230574846268,
0.039194945245981216,
0.11500761657953262,
0.09932972490787506,
-0.02624616026878357,
0.07412296533584595,
-0.08596387505531311,
0.08044642955064774,
0.23574751615524292,
-0.24767117202281952,
-0.08367420732975006,
-0.03843418508768082,
0.044787317514419556,
0.0508071668446064,
-0.12937705218791962,
-0.03697773441672325,
0.022230105474591255,
0.039575207978487015,
0.11633675545454025,
0.003035113448277116,
-0.025166595354676247,
0.020055586472153664,
-0.14748865365982056,
-0.04856153205037117,
0.09924551099538803,
0.06699060648679733,
-0.024013152346014977,
-0.1031462624669075,
-0.016670098528265953,
-0.19324225187301636,
-0.05854259058833122,
0.012616897001862526,
0.027831731364130974,
-0.0390564501285553,
-0.1074926033616066,
0.013527637347579002,
-0.08634815365076065,
-0.09000428766012192,
0.02399410307407379,
0.13273650407791138,
0.047846537083387375,
-0.027514217421412468,
-0.0033466094173491,
0.10992901027202606,
0.03789927065372467,
-0.13441012799739838,
-0.0015383580466732383,
0.03848718851804733,
-0.10261216014623642,
-0.016479594632983208,
-0.029522741213440895,
-0.022602232173085213,
-0.005432269535958767,
0.11180628836154938,
-0.03467527776956558,
0.08259224146604538,
0.022198209539055824,
0.031548719853162766,
-0.08750500530004501,
0.18048010766506195,
-0.06367941945791245,
-0.035749245434999466,
-0.0621141642332077,
0.10002798587083817,
-0.022160494700074196,
-0.009863988496363163,
-0.06187991425395012,
0.028657488524913788,
0.10093221068382263,
0.04218939319252968,
-0.025469662621617317,
0.02639133669435978,
-0.06383515894412994,
-0.026714803650975227,
-0.042377445846796036,
-0.0998840406537056,
0.04405368119478226,
0.03523106500506401,
-0.08237608522176743,
0.007194072008132935,
0.007452397607266903,
0.016093892976641655,
-0.031514331698417664,
0.1380993276834488,
-0.060353729873895645,
0.020023714751005173,
-0.09330327063798904,
-0.09817984700202942,
0.03174858167767525,
-0.05332774668931961,
0.0014473290648311377,
-0.0659838318824768,
-0.11715225875377655,
-0.037345681339502335,
0.07214713841676712,
-0.053171996027231216,
-0.06172272935509682,
-0.06468234956264496,
-0.0643346980214119,
0.0473281666636467,
-0.02786576934158802,
0.17190037667751312,
-0.0649404525756836,
0.11146839708089828,
0.010632720775902271,
0.042202942073345184,
0.041975025087594986,
0.07583526521921158,
-0.0471305213868618,
0.03618790954351425,
-0.1367097645998001,
0.07246523350477219,
-0.08542721718549728,
0.05528459697961807,
-0.14423692226409912,
-0.12663748860359192,
-0.035410746932029724,
0.005450117401778698,
0.09583024680614471,
0.08968889713287354,
-0.17589305341243744,
-0.09818606823682785,
0.18748538196086884,
-0.0658225268125534,
-0.09061046689748764,
0.13710302114486694,
-0.03514309599995613,
0.001434636884368956,
0.03586574271321297,
0.16498295962810516,
0.08187981694936752,
-0.09542520344257355,
0.030670829117298126,
-0.04751589521765709,
0.1085071936249733,
0.019410010427236557,
0.09359801560640335,
-0.04591483622789383,
0.040867824107408524,
-0.008653232827782631,
-0.024437393993139267,
0.06962781399488449,
-0.08650228381156921,
-0.0918339267373085,
-0.016018547117710114,
-0.07248746603727341,
0.01806606538593769,
0.05936158075928688,
0.03623482957482338,
-0.09356402605772018,
-0.13542614877223969,
0.020977597683668137,
0.09724149107933044,
-0.0999893769621849,
0.03318372368812561,
-0.07781395316123962,
0.03291673958301544,
-0.0234936885535717,
-0.006928673014044762,
-0.16472488641738892,
0.006022842135280371,
0.027724526822566986,
-0.03251330181956291,
0.02942669577896595,
0.007198340259492397,
0.09034429490566254,
0.03366789594292641,
-0.04564465954899788,
-0.07492038607597351,
-0.07198088616132736,
-0.009423617273569107,
-0.08505536615848541,
-0.2303803712129593,
-0.0728880986571312,
-0.03313658386468887,
0.14923390746116638,
-0.22927798330783844,
0.007641419768333435,
0.018822025507688522,
0.09565579146146774,
0.03894369304180145,
-0.045168861746788025,
-0.013378158211708069,
0.08079294860363007,
-0.014222204685211182,
-0.06478890776634216,
0.045129623264074326,
0.0003432764206081629,
-0.11647335439920425,
0.03198867291212082,
-0.11463712155818939,
0.09648185223340988,
0.10946719348430634,
-0.029050694778561592,
-0.08319709450006485,
-0.058812499046325684,
-0.07038350403308868,
-0.06764727830886841,
-0.03204961493611336,
0.007347938138991594,
0.22847974300384521,
0.039383381605148315,
0.12236560881137848,
-0.07488413900136948,
-0.057170454412698746,
0.027011068537831306,
0.0036384763661772013,
-0.015733839944005013,
0.14567887783050537,
0.06663595139980316,
-0.02865545265376568,
0.09468844532966614,
0.10481377691030502,
-0.08937918394804001,
0.16136722266674042,
-0.06713955849409103,
-0.13388969004154205,
-0.014283857308328152,
0.026248082518577576,
0.0318816676735878,
0.13149094581604004,
-0.16994845867156982,
-0.011627382598817348,
0.012051718309521675,
0.02743375673890114,
0.028924981132149696,
-0.21799778938293457,
-0.014090554788708687,
0.048563748598098755,
-0.07135152816772461,
-0.030067671090364456,
-0.023785635828971863,
0.009864375926554203,
0.0921010747551918,
0.003547406056895852,
-0.07470991462469101,
-0.009074850007891655,
-0.026401594281196594,
-0.07883517444133759,
0.19178971648216248,
-0.085688516497612,
-0.1335299015045166,
-0.12261681258678436,
-0.041579440236091614,
0.010261958464980125,
-0.01946091651916504,
0.04814016446471214,
-0.11200560629367828,
-0.0299681406468153,
-0.0651434138417244,
0.053100213408470154,
-0.06704670935869217,
0.03087404929101467,
-0.004574017599225044,
0.009628809057176113,
0.07936310768127441,
-0.10203389078378677,
0.01904400624334812,
-0.03219730034470558,
-0.04056921973824501,
0.038246069103479385,
0.03636636957526207,
0.09174990653991699,
0.1730836182832718,
0.03168663755059242,
0.02193032205104828,
-0.04221764951944351,
0.13141675293445587,
-0.09798125177621841,
-0.03518134728074074,
0.11046680063009262,
-0.007410194259136915,
0.04529428854584694,
0.11518555879592896,
0.06299767643213272,
-0.07526032626628876,
0.01587788760662079,
0.052214063704013824,
-0.017150668427348137,
-0.23942255973815918,
-0.02804531157016754,
-0.062119197100400925,
-0.009484624490141869,
0.12589356303215027,
0.030341779813170433,
-0.019592707976698875,
0.028553687036037445,
-0.0007235237862914801,
-0.0033322246745228767,
-0.0012167568784207106,
0.0631004050374031,
0.05352884158492088,
0.03585305064916611,
0.11341435462236404,
-0.015800822526216507,
-0.054018791764974594,
0.02248472347855568,
0.006871303077787161,
0.2613057494163513,
-0.002978159347549081,
0.17276810109615326,
0.052130475640296936,
0.15718844532966614,
0.014976056292653084,
0.0653093233704567,
0.003019133349880576,
-0.02921583503484726,
0.01114977989345789,
-0.055232226848602295,
-0.02408219687640667,
0.04849093407392502,
0.09847261756658554,
0.04298056289553642,
-0.11722464859485626,
-0.03057108260691166,
0.02557172067463398,
0.35804688930511475,
0.05881059169769287,
-0.2943343222141266,
-0.07927178591489792,
0.004134088754653931,
-0.08846427500247955,
-0.04031017795205116,
0.032678406685590744,
0.1088128313422203,
-0.09653764218091965,
0.034130487591028214,
-0.08836890757083893,
0.1008419618010521,
-0.06662992388010025,
0.007587817031890154,
0.09294367581605911,
0.0750386118888855,
-0.00037260568933561444,
0.06314448267221451,
-0.2689807713031769,
0.3096538186073303,
-0.02033594809472561,
0.07992976903915405,
-0.05180973559617996,
0.03139610216021538,
0.03785758465528488,
-0.06360232084989548,
0.07726127654314041,
-0.0199615266174078,
-0.09144127368927002,
-0.19417549669742584,
-0.07940112799406052,
0.021342435851693153,
0.1294575333595276,
-0.043080635368824005,
0.12498253583908081,
-0.0238997470587492,
-0.005502240266650915,
0.06190606206655502,
-0.07609972357749939,
-0.0963478535413742,
-0.11330961436033249,
0.025220856070518494,
0.018669266253709793,
0.05978403985500336,
-0.10657833516597748,
-0.11162808537483215,
-0.0737803727388382,
0.1587395966053009,
-0.0933208018541336,
-0.0075712488032877445,
-0.12869760394096375,
0.08382958918809891,
0.1607457399368286,
-0.06697754561901093,
0.046112995594739914,
0.027778232470154762,
0.11211419105529785,
0.02475745789706707,
-0.01023709774017334,
0.11652958393096924,
-0.08930914103984833,
-0.17883701622486115,
-0.059259310364723206,
0.16129136085510254,
0.04708833247423172,
0.0722474604845047,
-0.022078804671764374,
0.017122970893979073,
-0.02367626689374447,
-0.07888272404670715,
0.059478726238012314,
0.0019942910876125097,
0.013575609773397446,
0.04687574878334999,
-0.038936953991651535,
0.004059562459588051,
-0.08018998801708221,
-0.05716080963611603,
0.16430659592151642,
0.2663310766220093,
-0.0837044045329094,
0.033973969519138336,
0.038778144866228104,
-0.048922717571258545,
-0.1503930687904358,
0.02971569448709488,
0.1277240365743637,
0.030548470094799995,
0.005215685814619064,
-0.23457874357700348,
0.05807854235172272,
0.08690203726291656,
-0.01924259215593338,
0.08075001090765,
-0.34997645020484924,
-0.12842272222042084,
0.12790314853191376,
0.11160246282815933,
-0.028205951675772667,
-0.1571047008037567,
-0.05232829973101616,
0.006147623527795076,
-0.08941793441772461,
0.060492176562547684,
-0.04468868672847748,
0.12465638667345047,
-0.008387075737118721,
0.07196033746004105,
0.018478073179721832,
-0.059567514806985855,
0.14030463993549347,
-0.018645944073796272,
0.06053952872753143,
-0.008291688747704029,
0.053259074687957764,
0.0009981045732274652,
-0.04881560429930687,
0.012655270285904408,
-0.07762975990772247,
0.003297524293884635,
-0.1392897516489029,
-0.03173503279685974,
-0.08175375312566757,
0.02543814666569233,
-0.03829570114612579,
-0.05272878333926201,
0.006541144102811813,
0.04404396936297417,
0.060537129640579224,
0.00399318290874362,
0.11765255779027939,
-0.059166572988033295,
0.1423650085926056,
0.06166262924671173,
0.10472243279218674,
-0.013692797161638737,
-0.11587861180305481,
-0.018775708973407745,
-0.0031067815143615007,
0.058151811361312866,
-0.11414626985788345,
0.026379374787211418,
0.14308024942874908,
0.04463285207748413,
0.15381598472595215,
0.06504663079977036,
-0.07863503694534302,
0.019573718309402466,
0.06480123847723007,
-0.08416147530078888,
-0.11178526282310486,
-0.006011219695210457,
0.0755656287074089,
-0.1414736807346344,
0.0045037842355668545,
0.10113336890935898,
-0.05399082601070404,
-0.007004304323345423,
0.013653994537889957,
0.01372851338237524,
-0.05382975563406944,
0.22513410449028015,
0.023262914270162582,
0.07283252477645874,
-0.08913706988096237,
0.0782613679766655,
0.05121917650103569,
-0.18275173008441925,
0.024669094011187553,
0.0900406688451767,
-0.02616008184850216,
-0.020464003086090088,
0.029390821233391762,
0.08891602605581284,
0.018866000697016716,
-0.055180590599775314,
-0.11484213173389435,
-0.15117982029914856,
0.09277691692113876,
0.10148010402917862,
0.03244103118777275,
0.02657465636730194,
-0.059381451457738876,
0.04984970763325691,
-0.10537686944007874,
0.08840467035770416,
0.10146014392375946,
0.07159218192100525,
-0.12910374999046326,
0.14832879602909088,
0.009839209727942944,
0.006337940227240324,
0.004112366586923599,
-0.008717585355043411,
-0.10116711258888245,
0.037157054990530014,
-0.11039368808269501,
-0.02392484061419964,
-0.054976850748062134,
-0.0016911604907363653,
0.006276225205510855,
-0.04820113256573677,
-0.04745686799287796,
0.014724994078278542,
-0.11911686509847641,
-0.04186311364173889,
-0.0065686083398759365,
0.06895825266838074,
-0.106974296271801,
-0.02320483885705471,
0.046767715364694595,
-0.10935318470001221,
0.09944076836109161,
0.05987754464149475,
0.0168056171387434,
0.03753725066781044,
-0.13740654289722443,
0.01219850778579712,
0.035885993391275406,
-0.003142037196084857,
0.02874649316072464,
-0.15354806184768677,
-0.012834065593779087,
-0.026474183425307274,
0.038067519664764404,
-0.004230188671499491,
0.020253852009773254,
-0.1395770162343979,
-0.03791797161102295,
-0.025256024673581123,
-0.0799713060259819,
-0.05594110116362572,
0.05011704936623573,
0.053903527557849884,
0.03896036744117737,
0.1659637987613678,
-0.09072128683328629,
0.05486588925123215,
-0.21362453699111938,
0.01662796549499035,
-0.02860872820019722,
-0.06160346046090126,
-0.05748927965760231,
-0.03425170108675957,
0.08527806401252747,
-0.0718810111284256,
0.09161219000816345,
-0.033815860748291016,
0.03035951964557171,
0.035811588168144226,
-0.12240120768547058,
0.0337725505232811,
0.04660840705037117,
0.22958451509475708,
0.04829967021942139,
-0.02561136707663536,
0.06149125099182129,
0.009579875506460667,
0.057599637657403946,
0.16682027280330658,
0.1645125299692154,
0.19963935017585754,
0.07329437881708145,
0.09309066832065582,
0.06433558464050293,
-0.11104727536439896,
-0.11860987544059753,
0.12292242795228958,
-0.02486320771276951,
0.1247721016407013,
-0.01674807071685791,
0.2606149911880493,
0.10573980957269669,
-0.19518589973449707,
0.04914055019617081,
-0.0347851924598217,
-0.08653022348880768,
-0.08988930284976959,
-0.0518781803548336,
-0.06154841557145119,
-0.17124047875404358,
0.011974550783634186,
-0.10315363854169846,
0.05867791920900345,
0.06041288375854492,
0.04719462990760803,
0.018299778923392296,
0.13236099481582642,
0.05613773688673973,
-0.00013302125444170088,
0.1280936896800995,
0.01984352618455887,
-0.0070845335721969604,
-0.05445149168372154,
-0.08836081624031067,
0.026036176830530167,
-0.03955768421292305,
0.043172646313905716,
-0.053678177297115326,
-0.10786096751689911,
0.05654522404074669,
0.00924211461097002,
-0.1051776111125946,
0.02060002088546753,
-0.021230138838291168,
0.07467207312583923,
0.09668188542127609,
0.033540233969688416,
-0.008884420618414879,
-0.028330523520708084,
0.2552233636379242,
-0.1063055619597435,
-0.05909974128007889,
-0.12636949121952057,
0.24111811816692352,
0.010326492600142956,
-0.01855415850877762,
0.010307423770427704,
-0.06437242776155472,
0.007066073827445507,
0.167831152677536,
0.14516139030456543,
-0.04295959696173668,
-0.012839007191359997,
0.015348630025982857,
-0.01036688219755888,
-0.05425727367401123,
0.08574160188436508,
0.13324017822742462,
0.07465913891792297,
-0.07852044701576233,
-0.04272184148430824,
-0.04796802997589111,
-0.045292407274246216,
-0.042928364127874374,
0.06502865999937057,
0.038791585713624954,
-0.008285760879516602,
-0.03336536884307861,
0.10913986712694168,
-0.07068584859371185,
-0.10683175176382065,
0.014461176469922066,
-0.17723605036735535,
-0.18719615042209625,
-0.03738768398761749,
0.08035410195589066,
0.030862746760249138,
0.05375700443983078,
-0.0074777379631996155,
-0.02240670472383499,
0.08420631289482117,
0.005653626751154661,
-0.044513002038002014,
-0.10858403891324997,
0.08933018893003464,
-0.08807437866926193,
0.19239313900470734,
-0.04636577144265175,
0.02029094286262989,
0.12016241997480392,
0.0779624730348587,
-0.066758893430233,
0.05166080966591835,
0.07169485837221146,
-0.133012592792511,
0.04257615655660629,
0.19021637737751007,
-0.03350203111767769,
0.14098648726940155,
0.044737521559000015,
-0.11258760094642639,
0.030386250466108322,
-0.10244165360927582,
-0.05987321585416794,
-0.05715717375278473,
-0.010790384374558926,
-0.03547893464565277,
0.13327933847904205,
0.22671107947826385,
-0.0602031834423542,
-0.021650169044733047,
-0.0603790245950222,
0.002122508594766259,
0.03706914186477661,
0.1202511340379715,
-0.05504283308982849,
-0.26333460211753845,
0.010316315107047558,
0.000727058679331094,
0.015040190890431404,
-0.25290733575820923,
-0.10330981761217117,
0.04995787516236305,
-0.05252068117260933,
-0.06006643921136856,
0.1072792336344719,
0.07924424856901169,
0.0580766536295414,
-0.05177401751279831,
-0.08006090670824051,
-0.0351671501994133,
0.18152499198913574,
-0.1717243492603302,
-0.05104629695415497
] |
null | null |
transformers
|
# Wav2Vec2-Large-XLSR-53-Bemba
Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Bemba language of Zambia using the [BembaSpeech](https://csikasote.github.io/BembaSpeech). When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
```python
import torch
import torchaudio
from datasets import load_dataset
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
test_dataset = load_dataset("csv", data_files={"test": "/content/test.csv"}, delimiter="\t")["test"] # Adapt the path to test.csv
processor = Wav2Vec2Processor.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba")
model = Wav2Vec2ForCTC.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba")
#BembaSpeech is sample at 16kHz so we you do not need to resample
#resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = speech_array.squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits
predicted_ids = torch.argmax(logits, dim=-1)
print("Prediction:", processor.batch_decode(predicted_ids))
print("Reference:", test_dataset["sentence"][:2])
```
## Evaluation
The model can be evaluated as follows on the Bemba test data of BembaSpeech.
```python
import torch
import torchaudio
from datasets import load_dataset, load_metric
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
import re
test_dataset = load_dataset("csv", data_files={"test": "/content/test.csv"}, delimiter="\\t")["test"]
wer = load_metric("wer")
processor = Wav2Vec2Processor.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba")
model = Wav2Vec2ForCTC.from_pretrained("csikasote/wav2vec2-large-xlsr-bemba")
model.to("cuda")
chars_to_ignore_regex = '[\,\_\?\.\!\;\:\"\“]'
#resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower()
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = speech_array.squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def evaluate(batch):
inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits
pred_ids = torch.argmax(logits, dim=-1)
batch["pred_strings"] = processor.batch_decode(pred_ids)
return batch
result = test_dataset.map(evaluate, batched=True, batch_size=8)
print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"], references=result["sentence"])))
```
**Test Result**: 42.17 %
## Training
The BembaSpeech `train`, `dev` and `test` datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found [here](https://colab.research.google.com/drive/1aplFHfaXE68HGDwBYV2KqUWPasrk7bXv?usp=sharing).
|
{"language": "bem", "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["BembaSpeech"], "metrics": ["wer"], "model-index": [{"name": "XLSR Wav2Vec2 Bemba by Claytone Sikasote", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "BembaSpeech bem", "type": "bembaspeech", "args": "bem"}, "metrics": [{"type": "wer", "value": 42.17, "name": "Test WER"}]}]}]}
|
automatic-speech-recognition
|
csikasote/wav2vec2-large-xlsr-bemba
|
[
"transformers",
"pytorch",
"jax",
"wav2vec2",
"automatic-speech-recognition",
"audio",
"speech",
"xlsr-fine-tuning-week",
"bem",
"dataset:BembaSpeech",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"bem"
] |
TAGS
#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #bem #dataset-BembaSpeech #license-apache-2.0 #model-index #endpoints_compatible #region-us
|
# Wav2Vec2-Large-XLSR-53-Bemba
Fine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
## Evaluation
The model can be evaluated as follows on the Bemba test data of BembaSpeech.
Test Result: 42.17 %
## Training
The BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here.
|
[
"# Wav2Vec2-Large-XLSR-53-Bemba\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Bemba test data of BembaSpeech. \n\n\n\n\nTest Result: 42.17 %",
"## Training\n\nThe BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here."
] |
[
"TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #bem #dataset-BembaSpeech #license-apache-2.0 #model-index #endpoints_compatible #region-us \n",
"# Wav2Vec2-Large-XLSR-53-Bemba\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Bemba test data of BembaSpeech. \n\n\n\n\nTest Result: 42.17 %",
"## Training\n\nThe BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here."
] |
[
80,
70,
20,
32,
52
] |
[
"passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #bem #dataset-BembaSpeech #license-apache-2.0 #model-index #endpoints_compatible #region-us \n# Wav2Vec2-Large-XLSR-53-Bemba\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Bemba language of Zambia using the BembaSpeech. When using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:## Evaluation\n\nThe model can be evaluated as follows on the Bemba test data of BembaSpeech. \n\n\n\n\nTest Result: 42.17 %## Training\n\nThe BembaSpeech 'train', 'dev' and 'test' datasets were used for training, development and evaluation respectively. The script used for evaluating the model on the test dataset can be found here."
] |
[
-0.11402676999568939,
0.05119815841317177,
-0.002567385323345661,
0.032641369849443436,
0.02792268805205822,
-0.007559408899396658,
0.1468786597251892,
0.09148366749286652,
0.04825704172253609,
0.0037539098411798477,
0.041110049933195114,
-0.01816781610250473,
0.019245078787207603,
0.13581177592277527,
-0.031562089920043945,
-0.15989825129508972,
0.01000458188354969,
0.04719199240207672,
0.011200756765902042,
0.13568225502967834,
0.09662226587533951,
-0.05515754595398903,
0.011758833192288876,
0.053429264575242996,
-0.12224258482456207,
0.0031478942837566137,
0.009304547682404518,
-0.10648516565561295,
0.050544317811727524,
0.06362245976924896,
0.06863436847925186,
0.032085105776786804,
0.05212550610303879,
-0.1724461019039154,
0.018365079537034035,
0.038605883717536926,
0.04521987587213516,
-0.010347997769713402,
0.09459565579891205,
0.049519505351781845,
0.10866860300302505,
0.07134600728750229,
-0.024740854278206825,
0.06101509928703308,
-0.05834802985191345,
-0.17510481178760529,
-0.048470787703990936,
-0.02722611464560032,
0.09817183017730713,
0.1095299944281578,
-0.047259457409381866,
0.21910113096237183,
-0.1298525184392929,
0.03394460678100586,
0.057418908923864365,
-0.12852613627910614,
-0.0030671358108520508,
0.032356955111026764,
-0.005085684824734926,
0.10380692034959793,
-0.03927881270647049,
-0.02348422445356846,
0.05534601956605911,
-0.00011405583063606173,
0.019994694739580154,
-0.040368854999542236,
-0.07918297499418259,
-0.07066065818071365,
-0.17165187001228333,
-0.0385994054377079,
0.2746437191963196,
-0.03183497488498688,
-0.08278901875019073,
-0.168972909450531,
-0.029163489118218422,
0.12276619672775269,
0.0026826246175915003,
-0.03797613084316254,
-0.01996498927474022,
0.034884482622146606,
-0.00927745457738638,
-0.022918803617358208,
-0.08447576314210892,
-0.09975846856832504,
0.01561116985976696,
0.06140368431806564,
0.0043815672397613525,
0.02420131303369999,
-0.15158821642398834,
0.09410829097032547,
-0.036046840250492096,
-0.11366578191518784,
-0.09445445984601974,
-0.050803110003471375,
-0.0589204765856266,
-0.04888574406504631,
-0.041607268154621124,
-0.1221025213599205,
0.013835631310939789,
0.08382418751716614,
0.0627203956246376,
0.019495561718940735,
-0.05641818046569824,
-0.022904550656676292,
0.07554906606674194,
0.15911853313446045,
-0.009862028062343597,
-0.12195843458175659,
-0.004000307992100716,
-0.014971519820392132,
-0.04096590727567673,
-0.010266456753015518,
-0.01742146909236908,
-0.13591083884239197,
0.029558679088950157,
0.03982339799404144,
0.014316733926534653,
-0.007207429502159357,
-0.09192502498626709,
-0.09773623943328857,
0.044380296021699905,
-0.10779235512018204,
0.01029337476938963,
0.0954606831073761,
0.026097120717167854,
0.061006635427474976,
0.06529638171195984,
0.033245258033275604,
-0.07873187214136124,
0.00894546415656805,
-0.009607864543795586,
0.043795835226774216,
-0.06110897660255432,
0.0011063746642321348,
0.03484431654214859,
0.03041890449821949,
-0.009663070552051067,
-0.05551423132419586,
-0.1877417415380478,
-0.021451566368341446,
0.01582864671945572,
-0.026279568672180176,
0.04641132801771164,
-0.01708274334669113,
0.011126321740448475,
-0.01919720135629177,
-0.07026433944702148,
0.048882585018873215,
-0.05165340378880501,
0.04389283433556557,
0.10050851851701736,
0.017603067681193352,
0.06058492511510849,
0.0639244094491005,
-0.051123231649398804,
-0.017634257674217224,
-0.0659731924533844,
0.16305950284004211,
-0.047074273228645325,
-0.08798190951347351,
-0.08302280306816101,
-0.0628179982304573,
-0.09615723788738251,
0.012734967283904552,
0.0009433082304894924,
0.14242075383663177,
-0.19441474974155426,
-0.10037834197282791,
0.29011282324790955,
-0.1533249467611313,
-0.03706808760762215,
0.2002486139535904,
-0.06116309016942978,
0.12458428740501404,
0.0732731968164444,
0.09326325356960297,
0.25242769718170166,
-0.24607202410697937,
0.0595317967236042,
-0.0028976909816265106,
-0.015646815299987793,
-0.028511714190244675,
0.10774178057909012,
-0.07178083807229996,
0.034416262060403824,
0.022609593346714973,
-0.11185327917337418,
0.04741741716861725,
0.012579974718391895,
-0.04390370100736618,
-0.06353918462991714,
-0.043154027312994,
0.0898401290178299,
-0.007273537106812,
-0.032990243285894394,
-0.017568524926900864,
-0.032375603914260864,
0.11288773268461227,
0.11997359991073608,
-0.09233096987009048,
0.02224144898355007,
-0.12038425356149673,
0.08138405531644821,
-0.12840905785560608,
-0.04613185301423073,
-0.1431664526462555,
0.1678144484758377,
0.04481362923979759,
0.05265092849731445,
0.06366736441850662,
0.10593932867050171,
0.01963040977716446,
-0.02659965120255947,
0.00006569059769390151,
-0.00092760578263551,
-0.0036655620206147432,
-0.047775913029909134,
-0.023103656247258186,
-0.049400102347135544,
-0.033814121037721634,
-0.043489232659339905,
0.1380770355463028,
-0.16648966073989868,
-0.024239934980869293,
0.06385377049446106,
0.043203435838222504,
0.03233013302087784,
-0.02036481909453869,
0.07500220835208893,
0.05005182325839996,
-0.0022660617250949144,
-0.0036004353314638138,
0.011058742180466652,
0.023394852876663208,
0.007206321228295565,
0.08963929861783981,
-0.07012096047401428,
-0.07655948400497437,
0.10755623131990433,
0.05983589217066765,
0.03671487793326378,
0.03412112221121788,
0.006359199061989784,
-0.0194699764251709,
-0.09747394174337387,
-0.03364308923482895,
0.2849104404449463,
-0.0002595496771391481,
0.12454771995544434,
-0.13422122597694397,
-0.11305107921361923,
0.004002481698989868,
-0.09278848767280579,
0.0023345891386270523,
-0.007501774933189154,
-0.025169292464852333,
0.06136335805058479,
0.03525887802243233,
0.09008271247148514,
-0.05469901114702225,
0.37327003479003906,
-0.07062440365552902,
-0.15040352940559387,
-0.03721587359905243,
-0.01442366000264883,
-0.03502550721168518,
0.06296265870332718,
-0.21201345324516296,
0.0013183228438720107,
0.022777849808335304,
0.08300813287496567,
0.06733010709285736,
-0.10727293789386749,
0.05026540532708168,
0.07194896787405014,
-0.12010014057159424,
-0.08505628257989883,
0.06597176194190979,
-0.004756307229399681,
0.027366047725081444,
-0.10573527961969376,
0.03062012977898121,
0.008399078622460365,
-0.060784853994846344,
-0.18085569143295288,
0.12640699744224548,
-0.07644478976726532,
-0.12343776971101761,
-0.17758555710315704,
0.02291776053607464,
0.009832089766860008,
0.01341453567147255,
0.04111046716570854,
-0.10137458890676498,
-0.01079022977501154,
-0.07074816524982452,
0.0749952495098114,
0.003562260651960969,
0.0011466802097856998,
0.08747672289609909,
0.04891251400113106,
0.07298069447278976,
-0.14551319181919098,
0.030319849029183388,
-0.025761524215340614,
0.0289551243185997,
0.08175504207611084,
-0.06016882136464119,
-0.023810306563973427,
0.1665678322315216,
0.0023630065843462944,
0.015042845159769058,
0.029394501820206642,
0.1878678798675537,
-0.08391479402780533,
-0.009103856980800629,
0.24412864446640015,
-0.08055933564901352,
-0.024373240768909454,
0.030973365530371666,
-0.0415944866836071,
-0.07421433925628662,
0.018266255035996437,
-0.027148820459842682,
-0.0476236455142498,
-0.2726791501045227,
-0.10725116729736328,
-0.07684226334095001,
-0.0397978313267231,
-0.020245671272277832,
-0.0624493770301342,
-0.017107617110013962,
0.02362840808928013,
-0.0230970848351717,
-0.09212023764848709,
0.03189076855778694,
0.010160288773477077,
0.06616007536649704,
0.019883714616298676,
0.07480582594871521,
-0.052167050540447235,
0.0016543969977647066,
0.06360036879777908,
0.04046715423464775,
0.15992900729179382,
0.06493605673313141,
0.1402808129787445,
0.044693246483802795,
0.05988204479217529,
0.1088060736656189,
0.050519488751888275,
-0.07994088530540466,
0.013482474721968174,
0.012778524309396744,
-0.06213885918259621,
-0.10586826503276825,
0.04130791872739792,
0.09067783504724503,
-0.048366330564022064,
0.004199671093374491,
-0.014635931700468063,
0.03456130996346474,
0.1383078545331955,
0.09571094810962677,
-0.1588689237833023,
-0.0734986737370491,
0.021597711369395256,
-0.07000913470983505,
-0.03145158663392067,
0.01377718336880207,
0.14760246872901917,
-0.08939633518457413,
0.002753143198788166,
-0.05190659686923027,
0.07609601318836212,
-0.004292440600693226,
-0.022738557308912277,
-0.02656693384051323,
-0.0002134158567059785,
-0.0016307021724060178,
0.03233605623245239,
-0.21354453265666962,
0.11329182982444763,
0.005600000731647015,
0.1148214042186737,
-0.0709080770611763,
0.05218496173620224,
0.003848955500870943,
0.014175275340676308,
0.08534117043018341,
0.02938135899603367,
-0.055840786546468735,
-0.06571288406848907,
-0.044679850339889526,
0.03460681810975075,
0.04495524615049362,
0.11052785813808441,
0.13031268119812012,
-0.03636139631271362,
0.03810178115963936,
0.026198802515864372,
-0.014020930044353008,
-0.14199389517307281,
-0.06736090779304504,
0.05907151475548744,
0.08022863417863846,
0.07071982324123383,
-0.03623070567846298,
-0.03135858476161957,
-0.030399955809116364,
-0.029320264235138893,
-0.30610525608062744,
-0.09821724146604538,
-0.09336109459400177,
-0.08102483302354813,
0.16308315098285675,
-0.056585896760225296,
0.0067940219305455685,
0.04960436746478081,
0.18473686277866364,
-0.021877571940422058,
-0.053759459406137466,
-0.01611659675836563,
-0.11844930052757263,
-0.16067031025886536,
-0.06691883504390717,
0.14397098124027252,
0.12278534471988678,
0.05514903739094734,
0.13213564455509186,
-0.007941613905131817,
0.016118640080094337,
-0.0777367427945137,
-0.013303716666996479,
-0.008064281195402145,
-0.00722178490832448,
-0.013772809877991676,
-0.020284265279769897,
-0.16410236060619354,
-0.11972325295209885,
-0.0932646244764328,
0.09428597241640091,
0.0773499608039856,
-0.036065611988306046,
0.13466379046440125,
0.131798654794693,
-0.12319918721914291,
-0.18839648365974426,
-0.005505163222551346,
0.15768221020698547,
0.09725981950759888,
-0.03528701141476631,
-0.2258608192205429,
0.07612390071153641,
-0.004147874657064676,
-0.03167828917503357,
0.0010643138084560633,
-0.3411238193511963,
-0.1571797877550125,
0.15210069715976715,
-0.05897950381040573,
0.13372956216335297,
-0.031187715008854866,
-0.014116554521024227,
0.008616571314632893,
-0.06502488255500793,
0.06197519972920418,
-0.17603731155395508,
0.06579435616731644,
0.006767685525119305,
0.14619959890842438,
0.023661985993385315,
-0.031049374490976334,
0.10495965927839279,
0.11981704086065292,
0.010064659640192986,
0.011250914074480534,
0.1281462013721466,
0.002638953970745206,
-0.013070867396891117,
0.12075597792863846,
0.004360131919384003,
0.04252878576517105,
-0.19425877928733826,
-0.1115245670080185,
-0.05914901942014694,
0.06670618057250977,
-0.005736277438700199,
-0.07461367547512054,
0.08949266374111176,
-0.001211735769174993,
0.015025469474494457,
-0.02263505570590496,
-0.09886185079813004,
-0.1328015774488449,
-0.10228448361158371,
0.1783398985862732,
0.19953781366348267,
-0.04518771171569824,
-0.05807019770145416,
-0.022539330646395683,
0.02340712770819664,
0.13001057505607605,
-0.06613221764564514,
0.09600427746772766,
-0.005525417160242796,
0.05094051733613014,
0.08875790983438492,
-0.008935141377151012,
-0.1000806912779808,
0.061726756393909454,
0.012642735615372658,
-0.03448742628097534,
0.007508476264774799,
0.03973696753382683,
-0.023532560095191002,
-0.07782895117998123,
-0.029690144583582878,
0.13595251739025116,
-0.03479207307100296,
-0.029690463095903397,
-0.015380775555968285,
-0.04193135350942612,
-0.09177077561616898,
0.21457697451114655,
-0.008167346939444542,
0.052927032113075256,
-0.06374526768922806,
0.04346054047346115,
-0.034701451659202576,
0.04031415656208992,
0.03317999094724655,
-0.06606732308864594,
-0.09464219957590103,
-0.031064288690686226,
-0.04751535877585411,
0.06856422871351242,
0.02863999456167221,
-0.11312317848205566,
-0.01259578112512827,
-0.059983618557453156,
0.020391426980495453,
0.17140142619609833,
0.08305946737527847,
0.048882149159908295,
-0.08487977832555771,
-0.05058855935931206,
-0.10209020227193832,
0.0842592790722847,
0.12565277516841888,
-0.004672727547585964,
-0.0964738130569458,
0.10078300535678864,
0.0010211686603724957,
-0.013930203393101692,
-0.08237078785896301,
-0.049396514892578125,
-0.008221421390771866,
0.05862531065940857,
-0.15691453218460083,
0.01930248737335205,
-0.034537363797426224,
-0.0016348998760804534,
-0.015143400058150291,
-0.09842921793460846,
-0.0047645606100559235,
0.08801037818193436,
-0.04367654398083687,
0.05770228058099747,
-0.03978825733065605,
0.10553249716758728,
-0.06316021084785461,
0.006277130916714668,
0.12002372741699219,
-0.10096951574087143,
0.09244203567504883,
0.12858596444129944,
-0.07513650506734848,
0.07524171471595764,
-0.18421579897403717,
-0.05984693765640259,
-0.0063031925819814205,
0.07542303204536438,
-0.03346122056245804,
-0.17882612347602844,
0.05773866921663284,
0.06747439503669739,
0.08202527463436127,
0.0021239868365228176,
0.16336041688919067,
-0.07043690234422684,
0.00866671558469534,
-0.037976332008838654,
-0.06292599439620972,
0.009330593049526215,
0.03590119630098343,
0.03899691626429558,
0.08188971132040024,
0.20004884898662567,
-0.10805771499872208,
0.062675341963768,
-0.15348194539546967,
0.001283116522245109,
0.010913359932601452,
0.023781538009643555,
-0.08169926702976227,
-0.07565008848905563,
0.03164950758218765,
-0.01775483600795269,
0.1698802262544632,
0.018968505784869194,
-0.019118165597319603,
0.011408350430428982,
-0.11767680197954178,
0.013037394732236862,
-0.01622205413877964,
0.32799655199050903,
0.07933727651834488,
0.0526682510972023,
0.019118532538414,
-0.03185566887259483,
0.031174685806035995,
0.021244777366518974,
0.027955137193202972,
0.23655539751052856,
-0.03584771975874901,
0.055941395461559296,
0.026248235255479813,
-0.03956519812345505,
-0.04291485622525215,
-0.06063557416200638,
-0.14225977659225464,
0.0072090644389390945,
-0.005212206393480301,
0.13297975063323975,
0.1491418182849884,
-0.1566811352968216,
0.038653019815683365,
0.033466484397649765,
-0.09304343163967133,
-0.14850623905658722,
-0.04349571466445923,
-0.0751299187541008,
-0.12249922752380371,
0.05198162421584129,
-0.11120176315307617,
0.04862847924232483,
0.013661680743098259,
0.054477509111166,
-0.03944515064358711,
0.1572827249765396,
0.034164804965257645,
-0.15979093313217163,
0.08589629083871841,
-0.04941673204302788,
-0.002540275687351823,
0.022945189848542213,
0.018409019336104393,
0.09687808156013489,
0.019978446885943413,
0.059022605419158936,
0.027956482023000717,
-0.06983789801597595,
-0.018371352925896645,
-0.021409625187516212,
-0.08530139923095703,
-0.05306343361735344,
-0.007301513105630875,
0.030104996636509895,
0.15259894728660583,
0.08321331441402435,
-0.032554350793361664,
-0.004840237554162741,
0.025256795808672905,
-0.02610851265490055,
-0.1591571867465973,
-0.17115171253681183,
0.06047922745347023,
-0.020567791536450386,
0.03853590041399002,
-0.009661458432674408,
-0.04909299314022064,
0.008531340397894382,
0.2331211119890213,
0.0982486680150032,
-0.011599550023674965,
0.01166444830596447,
-0.015591875649988651,
0.010415596887469292,
-0.022088050842285156,
0.11301323771476746,
0.03368637338280678,
0.22746892273426056,
-0.000146206803037785,
0.021272625774145126,
-0.05836465582251549,
-0.07232623547315598,
-0.03929532319307327,
0.009007580578327179,
-0.11145032942295074,
-0.08763757348060608,
0.005727680400013924,
0.08912855386734009,
-0.05028896406292915,
-0.10550379008054733,
-0.041181594133377075,
-0.008882356807589531,
-0.08305074274539948,
0.014646690338850021,
-0.0023573474027216434,
0.02787480689585209,
0.007349254097789526,
-0.04862155765295029,
-0.02827600948512554,
0.20125311613082886,
-0.02425425872206688,
-0.033055298030376434,
0.018468398600816727,
0.03013998456299305,
-0.07238476723432541,
0.06435370445251465,
0.0243894774466753,
0.21409352123737335,
0.018085859715938568,
0.05105030536651611,
-0.06675748527050018,
0.11822754144668579,
0.031270138919353485,
0.017909174785017967,
-0.003152647754177451,
0.09750186651945114,
-0.023254379630088806,
0.15412300825119019,
0.05374908447265625,
-0.055436231195926666,
0.018474822863936424,
-0.056974444538354874,
-0.07902318984270096,
-0.14260268211364746,
0.06208081543445587,
-0.04710603877902031,
0.10597787797451019,
0.10424135625362396,
-0.102015420794487,
-0.06680101156234741,
-0.06005467101931572,
0.09355487674474716,
0.013736818917095661,
0.05787495896220207,
-0.0724249929189682,
-0.20198598504066467,
-0.007800431922078133,
0.017532341182231903,
0.0069350916892290115,
-0.21665997803211212,
0.026965105906128883,
0.042214978486299515,
-0.062201905995607376,
0.04779614135622978,
0.07420890033245087,
0.03364920616149902,
0.049475278705358505,
0.024295317009091377,
-0.006542949005961418,
0.035829219967126846,
0.06735040247440338,
-0.14909765124320984,
-0.06513954699039459
] |
null | null |
transformers
|
### marianmt-th-zh_cn
* source languages: th
* target languages: zh_cn
* dataset:
* model: transformer-align
* pre-processing: normalization + SentencePiece
* test set translations:
* test set scores:
## Training
Training scripts from [LalitaDeelert/NLP-ZH_TH-Project](https://github.com/LalitaDeelert/NLP-ZH_TH-Project). Experiments tracked at [cstorm125/marianmt-th-zh_cn](https://wandb.ai/cstorm125/marianmt-th-zh_cn).
```
export WANDB_PROJECT=marianmt-th-zh_cn
python train_model.py --input_fname ../data/v1/Train.csv \
--output_dir ../models/marianmt-th-zh_cn \
--source_lang th --target_lang zh \
--metric_tokenize zh --fp16
```
## Usage
```
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("cstorm125/marianmt-zh_cn-th")
model = AutoModelForSeq2SeqLM.from_pretrained("cstorm125/marianmt-zh_cn-th").cpu()
src_text = [
'ฉันรักคุณ',
'ฉันอยากกินข้าว',
]
translated = model.generate(**tokenizer(src_text, return_tensors="pt", padding=True))
print([tokenizer.decode(t, skip_special_tokens=True) for t in translated])
> ['我爱你', '我想吃饭。']
```
## Requirements
```
transformers==4.6.0
torch==1.8.0
```
|
{"tags": ["translation", "torch==1.8.0"], "widget": [{"text": "Inference Unavailable"}]}
|
translation
|
cstorm125/marianmt-th-zh_cn
|
[
"transformers",
"pytorch",
"marian",
"text2text-generation",
"translation",
"torch==1.8.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us
|
### marianmt-th-zh_cn
* source languages: th
* target languages: zh_cn
* dataset:
* model: transformer-align
* pre-processing: normalization + SentencePiece
* test set translations:
* test set scores:
## Training
Training scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.
## Usage
## Requirements
|
[
"### marianmt-th-zh_cn\n* source languages: th\n* target languages: zh_cn\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:",
"## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.",
"## Usage",
"## Requirements"
] |
[
"TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### marianmt-th-zh_cn\n* source languages: th\n* target languages: zh_cn\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:",
"## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.",
"## Usage",
"## Requirements"
] |
[
49,
63,
41,
3,
5
] |
[
"passage: TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n### marianmt-th-zh_cn\n* source languages: th\n* target languages: zh_cn\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-th-zh_cn.## Usage## Requirements"
] |
[
-0.17309942841529846,
-0.02402368374168873,
-0.0025235656648874283,
0.04286745563149452,
0.08233149349689484,
-0.004156056325882673,
0.0526319295167923,
0.08169356733560562,
-0.023285694420337677,
0.006210209336131811,
0.12547250092029572,
0.08747147768735886,
0.005561168305575848,
0.030811857432127,
-0.053131479769945145,
-0.17585749924182892,
0.025161365047097206,
0.040346842259168625,
-0.1667453944683075,
0.13419905304908752,
0.11687688529491425,
-0.049648385494947433,
0.1047242134809494,
0.03446459770202637,
-0.04025081545114517,
0.02850279025733471,
-0.022890605032444,
-0.11238297075033188,
0.13362814486026764,
0.060239315032958984,
0.0926518589258194,
0.10185325890779495,
0.07365555316209793,
-0.1769273728132248,
0.01952873356640339,
-0.0172830019146204,
-0.0032912064343690872,
0.02604632079601288,
0.09201551228761673,
-0.05003215745091438,
0.2323170006275177,
-0.09471257776021957,
-0.026848401874303818,
0.07056120783090591,
-0.05314893275499344,
-0.0943462997674942,
-0.06700225174427032,
-0.03450635075569153,
0.024889755994081497,
0.1406448781490326,
-0.012420358136296272,
0.19792751967906952,
-0.169630765914917,
0.09684810042381287,
0.19849856197834015,
-0.2584282159805298,
-0.04796166718006134,
0.07071937620639801,
0.11320475488901138,
-0.012481819838285446,
0.013054423034191132,
0.09186478704214096,
0.09372807294130325,
0.0055234432220458984,
-0.09208069741725922,
-0.09758717566728592,
-0.020047591999173164,
0.04602647200226784,
-0.15966439247131348,
-0.007107562851160765,
0.2540832757949829,
-0.014960519969463348,
-0.04581663757562637,
-0.00582644110545516,
-0.03126008063554764,
-0.018019165843725204,
-0.010455912910401821,
-0.007116223219782114,
-0.04998723790049553,
-0.0371963307261467,
-0.04042372107505798,
-0.026547111570835114,
-0.10129772126674652,
-0.09563121199607849,
-0.04754158854484558,
0.16776001453399658,
0.0699511170387268,
-0.005491908639669418,
-0.0990767851471901,
0.11782071739435196,
0.14426645636558533,
-0.12656310200691223,
-0.05872180312871933,
-0.06933476030826569,
-0.03659417852759361,
-0.006343928165733814,
-0.04225388169288635,
-0.11000965535640717,
0.11715041100978851,
0.08201926201581955,
-0.021041065454483032,
0.10778217017650604,
0.001935038948431611,
0.019849035888910294,
-0.05578485131263733,
0.14707127213478088,
-0.05746309086680412,
-0.06542734056711197,
0.023519758135080338,
-0.028671463951468468,
-0.04796788841485977,
-0.01634110137820244,
-0.09203075617551804,
-0.12278519570827484,
0.046650223433971405,
0.11805443465709686,
-0.04782428219914436,
0.06854087859392166,
0.022504117339849472,
-0.029098495841026306,
-0.0323047861456871,
-0.10502062737941742,
-0.0392073318362236,
0.06343823671340942,
-0.05906146019697189,
0.08269798010587692,
0.04426981508731842,
0.0035111564211547375,
-0.0659901425242424,
0.015080384910106659,
-0.03539261221885681,
0.0008680910686962306,
-0.06180749088525772,
-0.06929893791675568,
0.0389714241027832,
0.04387116804718971,
-0.008799808099865913,
-0.15968622267246246,
-0.12808287143707275,
-0.030135702341794968,
0.030380362644791603,
-0.02957158349454403,
0.0541713647544384,
-0.10342281311750412,
-0.039654407650232315,
0.024808695539832115,
-0.0017189773498103023,
0.0968913584947586,
-0.07448779046535492,
0.07540237158536911,
0.05002184212207794,
0.018965231254696846,
0.009443058632314205,
-0.005015185568481684,
-0.07232912629842758,
0.03970874100923538,
0.008217696100473404,
0.09840381890535355,
-0.0866256132721901,
0.07900069653987885,
-0.12384587526321411,
-0.15830077230930328,
-0.021173574030399323,
0.0738418698310852,
0.04902176558971405,
0.2069670855998993,
-0.19652488827705383,
-0.0033183456398546696,
0.19358505308628082,
-0.03489898517727852,
-0.11817865073680878,
0.10936225950717926,
-0.06359244883060455,
0.0815722793340683,
0.04769326001405716,
0.11103356629610062,
0.056711990386247635,
-0.11193760484457016,
0.12159236520528793,
-0.0028071526903659105,
0.021005742251873016,
-0.006069888826459646,
0.07641284167766571,
-0.028389275074005127,
-0.16928140819072723,
0.04283297806978226,
-0.12158514559268951,
0.05412532016634941,
-0.06922708451747894,
-0.08070934563875198,
-0.012272477149963379,
-0.06023050472140312,
0.08354542404413223,
0.0158791933208704,
0.11044079810380936,
-0.10404433310031891,
-0.06327320635318756,
-0.07115773111581802,
0.11394862085580826,
-0.07092136144638062,
0.037652768194675446,
-0.10532156378030777,
0.07905411720275879,
0.05001009255647659,
0.012283749878406525,
-0.11332421004772186,
0.047407716512680054,
0.0023041635286062956,
0.04077019542455673,
0.10494048148393631,
0.06898488849401474,
0.04754549637436867,
0.03771291673183441,
-0.02898671291768551,
-0.007534065283834934,
0.03452224284410477,
0.0017252754187211394,
-0.026350637897849083,
-0.158772811293602,
0.02678435668349266,
-0.005190092138946056,
0.13702809810638428,
-0.26260218024253845,
-0.0030683435034006834,
0.09647921472787857,
0.0755094662308693,
-0.06580416113138199,
0.06516500562429428,
0.002857421524822712,
0.04417876899242401,
-0.05408836528658867,
-0.009079898707568645,
0.04844564571976662,
-0.03333953395485878,
-0.13650773465633392,
0.11324939131736755,
-0.04970283806324005,
-0.005108973011374474,
0.0745224729180336,
-0.14704445004463196,
-0.029950158670544624,
-0.03673172742128372,
0.0034908857196569443,
-0.008188499137759209,
0.021215111017227173,
-0.00013446417870000005,
0.15426190197467804,
-0.019916292279958725,
0.140868678689003,
-0.11148179322481155,
-0.004799454007297754,
-0.008153132162988186,
-0.08285219967365265,
0.014448937959969044,
0.18609996140003204,
0.0275820754468441,
-0.2075110375881195,
0.022409066557884216,
0.07184971868991852,
-0.09625409543514252,
0.2416023313999176,
-0.03930353373289108,
-0.04577493295073509,
0.011831541545689106,
0.04939308762550354,
0.004171349573880434,
0.0750691294670105,
-0.1275360882282257,
-0.008584984578192234,
0.047494884580373764,
0.04751748591661453,
0.0732058435678482,
-0.1384473741054535,
-0.030592158436775208,
0.02623257413506508,
-0.042806774377822876,
-0.0424383170902729,
0.11348497122526169,
0.009912990033626556,
0.07502260059118271,
-0.04508271440863609,
-0.09459389001131058,
-0.007362199015915394,
-0.030892860144376755,
-0.15273548662662506,
0.25472819805145264,
-0.09272509813308716,
-0.1703944206237793,
-0.1568310558795929,
0.005125884432345629,
-0.08741629868745804,
-0.02388095296919346,
0.06397917121648788,
-0.12564383447170258,
-0.012833958491683006,
-0.02189537137746811,
0.0726134181022644,
-0.06364124268293381,
-0.03750508278608322,
-0.05572205036878586,
0.05365900322794914,
-0.044352851808071136,
-0.13338373601436615,
-0.013231364078819752,
-0.029469404369592667,
-0.07872219383716583,
0.06789351254701614,
-0.14816582202911377,
0.11905292421579361,
0.14961297810077667,
-0.0026049634907394648,
0.06268330663442612,
-0.03279966861009598,
0.12500081956386566,
-0.1050175130367279,
-0.02488104999065399,
0.1680292785167694,
0.023929297924041748,
-0.007530268281698227,
0.04846853390336037,
0.007486885413527489,
-0.053075261414051056,
0.022217897698283195,
-0.047446079552173615,
-0.07162372022867203,
-0.30478140711784363,
-0.14570069313049316,
-0.09227485209703445,
-0.01726452261209488,
0.014626799151301384,
0.024795731529593468,
0.12410324066877365,
0.07270807027816772,
0.002623711945489049,
-0.05200747027993202,
0.001263757236301899,
0.052593428641557693,
0.1368359476327896,
0.05086134746670723,
0.08163908123970032,
-0.06845174729824066,
-0.07385435700416565,
0.028603944927453995,
-0.029064487665891647,
0.22408849000930786,
0.00978158414363861,
0.12217734009027481,
0.07615545392036438,
0.17717459797859192,
0.05047805979847908,
0.09153976291418076,
0.05800618603825569,
-0.015570526011288166,
0.016323670744895935,
-0.07876899093389511,
-0.008104098029434681,
0.04277665540575981,
-0.011060495860874653,
-0.052835095673799515,
-0.09553717076778412,
0.0205510463565588,
0.06569597870111465,
0.12301753461360931,
0.018621299415826797,
-0.18077747523784637,
-0.016367556527256966,
-0.007439970970153809,
0.03519948199391365,
-0.04336400702595711,
0.08164379745721817,
0.03435178101062775,
-0.15441283583641052,
0.09560991823673248,
-0.047310929745435715,
0.11614561080932617,
0.021589204668998718,
0.021296149119734764,
-0.06581444293260574,
0.028698353096842766,
0.004840277601033449,
0.17152561247348785,
-0.3451928198337555,
0.2990512251853943,
0.008542178198695183,
0.06149239093065262,
-0.10734815895557404,
-0.050877850502729416,
0.03178733214735985,
0.10086182504892349,
0.10845470428466797,
0.00913910660892725,
-0.17183426022529602,
-0.08004745095968246,
-0.013421379029750824,
0.03317137807607651,
0.10091470181941986,
0.0491049699485302,
0.06769727170467377,
0.0012100661406293511,
-0.02534019574522972,
-0.01058289222419262,
0.04236100614070892,
-0.20138375461101532,
-0.07912564277648926,
0.035814303904771805,
-0.00665730657055974,
-0.07957152277231216,
-0.09094364196062088,
-0.08208860456943512,
-0.06409646570682526,
0.05529634281992912,
-0.046510253101587296,
-0.01754147745668888,
-0.1009678989648819,
-0.044222839176654816,
0.1489141434431076,
-0.11145270615816116,
0.024399876594543457,
-0.0005426584393717349,
-0.011931357905268669,
-0.019802285358309746,
-0.06847517192363739,
0.08033864945173264,
-0.09157039225101471,
-0.13536757230758667,
-0.023677386343479156,
0.1740117073059082,
0.04028617590665817,
0.06963475048542023,
0.05438428744673729,
-0.026667211204767227,
-0.024019284173846245,
-0.11095430701971054,
-0.10125245153903961,
-0.05135587975382805,
-0.024294529110193253,
0.03024164028465748,
-0.08634275197982788,
-0.028709005564451218,
-0.09596384316682816,
-0.13025537133216858,
0.21809548139572144,
0.17642022669315338,
-0.06547719240188599,
0.022165028378367424,
0.07140405476093292,
-0.07467879354953766,
-0.2617526650428772,
0.03567102551460266,
0.009737713262438774,
0.08301573991775513,
-0.05593804642558098,
-0.14389941096305847,
0.03337109461426735,
-0.005115840118378401,
-0.002010377123951912,
0.091753751039505,
-0.30239415168762207,
-0.15985319018363953,
0.17575116455554962,
0.04408952593803406,
0.16842348873615265,
-0.07745727896690369,
-0.04626963660120964,
-0.08862372487783432,
-0.1613837033510208,
0.01745631732046604,
-0.13710619509220123,
0.08443039655685425,
-0.004860365763306618,
0.044515855610370636,
0.03236960619688034,
-0.04765687137842178,
0.16282692551612854,
0.032018501311540604,
0.0022555955220013857,
-0.05253055691719055,
-0.0284102950245142,
0.015865053981542587,
0.010205673985183239,
0.06991399824619293,
-0.09129000455141068,
0.05740198493003845,
-0.11072210967540741,
-0.03575708717107773,
-0.07650195062160492,
0.09432979673147202,
-0.02949357032775879,
-0.023772917687892914,
-0.0936819463968277,
0.015516435727477074,
0.042089544236660004,
-0.005614600144326687,
0.09954240173101425,
-0.09891607612371445,
-0.009508956223726273,
0.06970719248056412,
0.21415680646896362,
-0.022291401401162148,
0.11474169045686722,
0.006859428249299526,
0.002399260411038995,
0.07840695977210999,
-0.10240977257490158,
0.029813023284077644,
0.1600416600704193,
0.030940214172005653,
0.048370134085416794,
0.03478633612394333,
-0.0847863256931305,
-0.01618010364472866,
0.09740892052650452,
-0.05667397007346153,
-0.06620978564023972,
-0.11278241127729416,
0.01553039439022541,
0.04860310256481171,
0.05132000148296356,
0.14488773047924042,
-0.04196063429117203,
-0.00999775342643261,
-0.028079308569431305,
-0.057266440242528915,
-0.08481968939304352,
0.22570106387138367,
0.08142334967851639,
0.06723113358020782,
-0.09747568517923355,
0.046900518238544464,
-0.004282205365598202,
0.052145328372716904,
0.015221801586449146,
0.1500464677810669,
-0.11685601621866226,
-0.1065080314874649,
0.05865577235817909,
0.20680458843708038,
-0.16253674030303955,
-0.1027313619852066,
-0.13419878482818604,
-0.12574055790901184,
0.00001887303005787544,
0.18706728518009186,
0.09727148711681366,
-0.031144948676228523,
-0.04324769973754883,
-0.07612881809473038,
-0.018254395574331284,
0.03545002266764641,
0.14522315561771393,
0.03071916662156582,
-0.06846729665994644,
0.13008908927440643,
-0.017566977068781853,
0.13429653644561768,
-0.05956333503127098,
-0.01742156408727169,
-0.08573991060256958,
0.09530481696128845,
-0.2410985827445984,
-0.0036721033975481987,
-0.04739873856306076,
-0.024494251236319542,
-0.01440828014165163,
-0.06755571067333221,
-0.04911162704229355,
0.03924141451716423,
-0.1115669384598732,
0.032847125083208084,
-0.06913909316062927,
0.07981332391500473,
-0.00066593405790627,
-0.009565822780132294,
0.0608842559158802,
-0.07456718385219574,
0.02720438875257969,
0.07621970772743225,
-0.06784428656101227,
0.1712464541196823,
-0.16704878211021423,
0.00987914763391018,
-0.005955812521278858,
0.06324853003025055,
0.03447266295552254,
-0.053403157740831375,
0.01752512902021408,
0.05431919917464256,
0.10847380012273788,
0.019280223175883293,
-0.04035598039627075,
-0.0533885657787323,
-0.061516981571912766,
-0.0662013441324234,
-0.08295472711324692,
-0.01662747748196125,
0.0699634850025177,
0.04369480535387993,
0.06998354941606522,
0.09057410806417465,
-0.09036428481340408,
0.05520870164036751,
-0.06796623766422272,
-0.014291825704276562,
-0.0010600771056488156,
-0.07726658880710602,
-0.06337477266788483,
-0.09290598332881927,
0.0908130407333374,
-0.01759311743080616,
0.13196249306201935,
-0.0513492077589035,
0.022470448166131973,
-0.022075414657592773,
0.04379909858107567,
0.08818678557872772,
0.026418952271342278,
0.2282741814851761,
0.017376871779561043,
0.0482858307659626,
-0.05074381083250046,
0.01502111367881298,
0.016096075996756554,
0.1301531344652176,
0.14554902911186218,
0.15665075182914734,
0.01771457865834236,
0.13063283264636993,
-0.004922846332192421,
0.003859135787934065,
0.026925865560770035,
-0.07870876789093018,
0.010444995015859604,
0.021940559148788452,
-0.025761500000953674,
0.12158038467168808,
0.1719643622636795,
-0.13841122388839722,
0.03378140553832054,
-0.05607050284743309,
-0.10122702270746231,
-0.13785967230796814,
-0.1561506986618042,
-0.10358734428882599,
-0.10271783173084259,
0.01642220839858055,
-0.07444076985120773,
-0.019282791763544083,
0.017019499093294144,
0.10242869704961777,
-0.04732147976756096,
0.2053244411945343,
0.03686041384935379,
-0.11324216425418854,
0.09450255334377289,
-0.04309156909584999,
0.046816423535346985,
0.055454373359680176,
-0.012533345259726048,
-0.003562218975275755,
-0.08385278284549713,
0.01518509816378355,
-0.002535533858463168,
-0.09232174605131149,
0.01724018156528473,
-0.03499394282698631,
-0.06580178439617157,
-0.017819073051214218,
0.02497720718383789,
0.09400846809148788,
0.18169663846492767,
0.04489925876259804,
-0.019087864086031914,
-0.002948748879134655,
0.10735724121332169,
-0.008198583498597145,
-0.1732204556465149,
-0.11524084955453873,
0.18152636289596558,
0.06977560371160507,
0.039565376937389374,
0.015861190855503082,
-0.008571803569793701,
0.03156546875834465,
0.3101840913295746,
0.24211104214191437,
-0.06820940971374512,
-0.01604117639362812,
0.00863416027277708,
0.02128016948699951,
0.007528241258114576,
0.09413257986307144,
0.040861744433641434,
0.1572205275297165,
-0.06085661053657532,
-0.08381308615207672,
-0.06388851255178452,
-0.024455612525343895,
-0.09744902700185776,
0.11584045737981796,
0.046661462634801865,
-0.09468964487314224,
-0.0020476195495575666,
0.13607625663280487,
-0.10675752907991409,
-0.006493146065622568,
-0.039259810000658035,
-0.08940247446298599,
-0.11121157556772232,
-0.049552373588085175,
0.017957204952836037,
0.038400933146476746,
0.040942948311567307,
-0.07127345353364944,
-0.03116253763437271,
-0.0031364497262984514,
0.024196932092308998,
-0.10156998038291931,
-0.06509953737258911,
0.12446589767932892,
-0.06011210009455681,
0.13325874507427216,
-0.013444619253277779,
0.15759721398353577,
0.11073623597621918,
0.058475129306316376,
-0.009872660040855408,
0.10305201262235641,
0.07947872579097748,
-0.04193814471364021,
0.06109168380498886,
0.06091545149683952,
-0.02382410503923893,
0.000053315288823796436,
0.03365492448210716,
-0.11576233804225922,
0.09029227495193481,
0.07355352491140366,
0.01118296105414629,
-0.09939298033714294,
0.08047472685575485,
-0.10547266900539398,
0.11455691605806351,
0.13963451981544495,
-0.031615231186151505,
0.012306365184485912,
-0.08323150873184204,
0.05889449268579483,
0.013147806748747826,
-0.007723946124315262,
-0.044504135847091675,
-0.19014431536197662,
-0.048822421580553055,
-0.0698038712143898,
-0.0065048690885305405,
-0.1704726219177246,
-0.025378098711371422,
-0.07050401717424393,
-0.047361403703689575,
-0.024475332349538803,
0.14638826251029968,
0.0927625298500061,
0.019667061045765877,
0.02854015678167343,
-0.12102469056844711,
-0.0026769954711198807,
0.0945979505777359,
-0.17445798218250275,
-0.07904449850320816
] |
null | null |
transformers
|
### marianmt-zh_cn-th
* source languages: zh_cn
* target languages: th
* dataset:
* model: transformer-align
* pre-processing: normalization + SentencePiece
* test set translations:
* test set scores:
## Training
Training scripts from [LalitaDeelert/NLP-ZH_TH-Project](https://github.com/LalitaDeelert/NLP-ZH_TH-Project). Experiments tracked at [cstorm125/marianmt-zh_cn-th](https://wandb.ai/cstorm125/marianmt-zh_cn-th).
```
export WANDB_PROJECT=marianmt-zh_cn-th
python train_model.py --input_fname ../data/v1/Train.csv \
\\t--output_dir ../models/marianmt-zh_cn-th \
\\t--source_lang zh --target_lang th \
\\t--metric_tokenize th_syllable --fp16
```
## Usage
```
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
tokenizer = AutoTokenizer.from_pretrained("cstorm125/marianmt-zh_cn-th")
model = AutoModelForSeq2SeqLM.from_pretrained("cstorm125/marianmt-zh_cn-th").cpu()
src_text = [
'我爱你',
'我想吃米饭',
]
translated = model.generate(**tokenizer(src_text, return_tensors="pt", padding=True))
print([tokenizer.decode(t, skip_special_tokens=True) for t in translated])
> ['ผมรักคุณนะ', 'ฉันอยากกินข้าว']
```
## Requirements
```
transformers==4.6.0
torch==1.8.0
```
|
{"tags": ["translation", "torch==1.8.0"], "widget": [{"text": "Inference Unavailable"}]}
|
translation
|
cstorm125/marianmt-zh_cn-th
|
[
"transformers",
"pytorch",
"marian",
"text2text-generation",
"translation",
"torch==1.8.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us
|
### marianmt-zh_cn-th
* source languages: zh_cn
* target languages: th
* dataset:
* model: transformer-align
* pre-processing: normalization + SentencePiece
* test set translations:
* test set scores:
## Training
Training scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.
## Usage
## Requirements
|
[
"### marianmt-zh_cn-th\n* source languages: zh_cn\n* target languages: th\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:",
"## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.",
"## Usage",
"## Requirements"
] |
[
"TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### marianmt-zh_cn-th\n* source languages: zh_cn\n* target languages: th\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:",
"## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.",
"## Usage",
"## Requirements"
] |
[
49,
63,
41,
3,
5
] |
[
"passage: TAGS\n#transformers #pytorch #marian #text2text-generation #translation #torch==1.8.0 #autotrain_compatible #endpoints_compatible #region-us \n### marianmt-zh_cn-th\n* source languages: zh_cn\n* target languages: th\n* dataset: \n* model: transformer-align\n* pre-processing: normalization + SentencePiece\n* test set translations: \n* test set scores:## Training\n\nTraining scripts from LalitaDeelert/NLP-ZH_TH-Project. Experiments tracked at cstorm125/marianmt-zh_cn-th.## Usage## Requirements"
] |
[
-0.16600242257118225,
-0.027812577784061432,
-0.002706715138629079,
0.04715290293097496,
0.0873231440782547,
-0.005283234175294638,
0.050016067922115326,
0.07865055650472641,
-0.025380514562129974,
0.008193920366466045,
0.12634031474590302,
0.09176836162805557,
0.004395538941025734,
0.04229692742228508,
-0.05487890541553497,
-0.18166756629943848,
0.02715873345732689,
0.03612595424056053,
-0.15910914540290833,
0.13827678561210632,
0.11968296766281128,
-0.0454656258225441,
0.10597463697195053,
0.02751600183546543,
-0.04340993985533714,
0.02637927047908306,
-0.01568298041820526,
-0.11511150747537613,
0.13452479243278503,
0.06041799113154411,
0.09637763351202011,
0.09812972694635391,
0.06990458816289902,
-0.1823704093694687,
0.019417934119701385,
-0.015727100893855095,
-0.0027074534446001053,
0.026586124673485756,
0.09203238785266876,
-0.05236917361617088,
0.23085398972034454,
-0.0931372120976448,
-0.02185753732919693,
0.06848341971635818,
-0.06224217265844345,
-0.08235563337802887,
-0.06336579471826553,
-0.033810101449489594,
0.023781949654221535,
0.1438513845205307,
-0.012827564030885696,
0.20232518017292023,
-0.16352961957454681,
0.10443394631147385,
0.19104807078838348,
-0.2612217962741852,
-0.05128660053014755,
0.0704120546579361,
0.11383768171072006,
-0.008452331647276878,
0.011386653408408165,
0.09152668714523315,
0.09218355268239975,
0.010467529296875,
-0.08692561089992523,
-0.09838569909334183,
-0.026912959292531013,
0.037766166031360626,
-0.16374622285366058,
-0.012138135731220245,
0.25750306248664856,
-0.016383042559027672,
-0.048943303525447845,
-0.0032462929375469685,
-0.03288828581571579,
-0.017772000283002853,
-0.012550790794193745,
-0.014735860750079155,
-0.05290371552109718,
-0.032794516533613205,
-0.04475047066807747,
-0.028066884726285934,
-0.10192220658063889,
-0.09086021780967712,
-0.05372299998998642,
0.1610981971025467,
0.06459642946720123,
-0.0057967123575508595,
-0.10393962264060974,
0.1128072589635849,
0.12517982721328735,
-0.12105956673622131,
-0.05493324249982834,
-0.07226864248514175,
-0.03490560129284859,
-0.012327487580478191,
-0.038940753787755966,
-0.11001955717802048,
0.11101428419351578,
0.09088649600744247,
-0.02764444425702095,
0.10269812494516373,
-0.008176573552191257,
0.020563943311572075,
-0.05590398609638214,
0.1417982578277588,
-0.050612710416316986,
-0.0661921501159668,
0.020641490817070007,
-0.035388071089982986,
-0.04853130131959915,
-0.017048489302396774,
-0.09110485017299652,
-0.1212567463517189,
0.04320840910077095,
0.11824239790439606,
-0.03739992901682854,
0.07652276009321213,
0.01705155335366726,
-0.029689468443393707,
-0.027813391759991646,
-0.1035325676202774,
-0.03750718757510185,
0.06006040424108505,
-0.05108531937003136,
0.08169142156839371,
0.044794365763664246,
0.002572472672909498,
-0.06938183307647705,
0.020949555560946465,
-0.03465719893574715,
-0.00320239644497633,
-0.06106717139482498,
-0.07386945188045502,
0.042201295495033264,
0.03693694621324539,
-0.007780165411531925,
-0.16202138364315033,
-0.12623170018196106,
-0.030258784070611,
0.030196428298950195,
-0.030354760587215424,
0.047904904931783676,
-0.10270994156599045,
-0.04460665583610535,
0.030744319781661034,
-0.0032439175993204117,
0.09237226098775864,
-0.07521375268697739,
0.07761859893798828,
0.051780588924884796,
0.02371702715754509,
0.007212596479803324,
-0.005828367546200752,
-0.07800965011119843,
0.03814557194709778,
0.0036834687925875187,
0.09365436434745789,
-0.08777495473623276,
0.0759875625371933,
-0.12740223109722137,
-0.15419414639472961,
-0.027834242209792137,
0.07481652498245239,
0.048086751252412796,
0.20854295790195465,
-0.1960277110338211,
-0.0016304492019116879,
0.2061179280281067,
-0.039377741515636444,
-0.11675256490707397,
0.11152754724025726,
-0.05863349139690399,
0.08647800981998444,
0.050932589918375015,
0.1141979917883873,
0.05660087242722511,
-0.11092456430196762,
0.11810749024152756,
0.00032358794123865664,
0.019335387274622917,
-0.003408460645005107,
0.07471564412117004,
-0.026273809373378754,
-0.17271658778190613,
0.04261317849159241,
-0.12311361730098724,
0.054921168833971024,
-0.06896557658910751,
-0.08009664714336395,
-0.012438623234629631,
-0.052473004907369614,
0.08658628165721893,
0.012518600560724735,
0.10329394042491913,
-0.10541462898254395,
-0.06359580159187317,
-0.07361330091953278,
0.10716305673122406,
-0.06200804188847542,
0.03794100880622864,
-0.10600699484348297,
0.07538964599370956,
0.04501007869839668,
0.016266493126749992,
-0.1196778267621994,
0.040778595954179764,
0.0030583448242396116,
0.050615690648555756,
0.10592199862003326,
0.06795231252908707,
0.05045321583747864,
0.031600471585989,
-0.03011801466345787,
-0.005340940784662962,
0.03659328818321228,
-0.0014882220420986414,
-0.027680950239300728,
-0.16059237718582153,
0.03289502114057541,
-0.010936220176517963,
0.1258634328842163,
-0.2624475061893463,
-0.0027807841543108225,
0.09944988787174225,
0.0718110203742981,
-0.05926883965730667,
0.06324192136526108,
0.001985258189961314,
0.04677413031458855,
-0.05641396343708038,
-0.008025256916880608,
0.05014386028051376,
-0.031540222465991974,
-0.12830029428005219,
0.12023009359836578,
-0.0548446848988533,
-0.006136483978480101,
0.07933022081851959,
-0.1498950868844986,
-0.03918980434536934,
-0.03444627299904823,
0.0024539725854992867,
-0.006251748651266098,
0.020438551902770996,
0.0008608216303400695,
0.14227432012557983,
-0.015983257442712784,
0.13813892006874084,
-0.10802624374628067,
-0.0019361082231625915,
-0.006177693605422974,
-0.08398593217134476,
0.009107859805226326,
0.18368418514728546,
0.03625183552503586,
-0.203501358628273,
0.030389169231057167,
0.07886873185634613,
-0.08925731480121613,
0.23722721636295319,
-0.03560017794370651,
-0.0458141528069973,
0.014596851542592049,
0.051900047808885574,
0.008299252949655056,
0.06714285910129547,
-0.12686368823051453,
-0.006035390309989452,
0.0486767403781414,
0.05088723450899124,
0.06917206197977066,
-0.1350783109664917,
-0.031377945095300674,
0.027296368032693863,
-0.043602585792541504,
-0.04272513836622238,
0.11083832383155823,
0.008693534880876541,
0.07852812111377716,
-0.0383252389729023,
-0.09249567240476608,
-0.004940792918205261,
-0.02683286927640438,
-0.15330855548381805,
0.25412553548812866,
-0.09317144751548767,
-0.18210142850875854,
-0.1621117740869522,
-0.0001317799324169755,
-0.08391600102186203,
-0.01865096390247345,
0.06586698442697525,
-0.12591174244880676,
-0.016631541773676872,
-0.02159576676785946,
0.07674067467451096,
-0.06989327818155289,
-0.03806963562965393,
-0.057541631162166595,
0.05623767897486687,
-0.04329857975244522,
-0.12918917834758759,
-0.013490560464560986,
-0.02207852713763714,
-0.07758182287216187,
0.07212129980325699,
-0.14578720927238464,
0.12565648555755615,
0.1452740728855133,
-0.0006020978908054531,
0.05618072673678398,
-0.0343228280544281,
0.13339866697788239,
-0.10851246863603592,
-0.02129307948052883,
0.1627371609210968,
0.019851956516504288,
-0.004910178482532501,
0.057189635932445526,
0.0037521633785218,
-0.05638308823108673,
0.026800164952874184,
-0.044402457773685455,
-0.07179545611143112,
-0.3084999620914459,
-0.1458289474248886,
-0.0931360200047493,
-0.005453138146549463,
0.016010740771889687,
0.02685544081032276,
0.12059278786182404,
0.07292500883340836,
-0.0010048558469861746,
-0.04861126095056534,
0.0034431512467563152,
0.05761023238301277,
0.142024427652359,
0.04527931660413742,
0.08428727090358734,
-0.06972901523113251,
-0.07721636444330215,
0.03432466834783554,
-0.02352486364543438,
0.2227739542722702,
0.018063466995954514,
0.125370055437088,
0.07492278516292572,
0.17070885002613068,
0.05151119828224182,
0.09465314447879791,
0.05807521194219589,
-0.013927099294960499,
0.013535981066524982,
-0.07713786512613297,
-0.00875845830887556,
0.04069213569164276,
-0.016141366213560104,
-0.05759594589471817,
-0.09132090210914612,
0.020606085658073425,
0.06786098331212997,
0.12438427656888962,
0.016406018286943436,
-0.18401233851909637,
-0.021796321496367455,
-0.004772582091391087,
0.03269033879041672,
-0.04830925166606903,
0.08209218084812164,
0.0335112139582634,
-0.15679356455802917,
0.09498900175094604,
-0.04613291844725609,
0.11839796602725983,
0.0187680646777153,
0.022331731393933296,
-0.06551235169172287,
0.02165280655026436,
0.004399498458951712,
0.1713276505470276,
-0.3509644567966461,
0.3038150668144226,
0.00868775974959135,
0.05817576125264168,
-0.10433991998434067,
-0.048698414117097855,
0.022091854363679886,
0.10782203078269958,
0.11770376563072205,
0.009901098906993866,
-0.1770709604024887,
-0.08797946572303772,
-0.012727437540888786,
0.02802317962050438,
0.10863403975963593,
0.04824315756559372,
0.06945089995861053,
-0.002446170197799802,
-0.024690330028533936,
-0.009284021332859993,
0.034253936260938644,
-0.1990301012992859,
-0.0837135910987854,
0.03520051762461662,
-0.00629215594381094,
-0.07812930643558502,
-0.08408542722463608,
-0.07853744179010391,
-0.057220056653022766,
0.068489208817482,
-0.054186683148145676,
-0.014424082823097706,
-0.10529156774282455,
-0.0462985597550869,
0.14372289180755615,
-0.11302772164344788,
0.02424662746489048,
-0.005315183196216822,
-0.009964118711650372,
-0.021679624915122986,
-0.06904258579015732,
0.08661238104104996,
-0.09566308557987213,
-0.1328759491443634,
-0.02206914685666561,
0.16943557560443878,
0.03566956892609596,
0.07135098427534103,
0.053163591772317886,
-0.019357791170477867,
-0.02102123573422432,
-0.1105794757604599,
-0.09787429869174957,
-0.04967039078474045,
-0.029093975201249123,
0.026381075382232666,
-0.09089897572994232,
-0.03279408439993858,
-0.09718990325927734,
-0.13266201317310333,
0.2263089120388031,
0.17884261906147003,
-0.06458063423633575,
0.024409234523773193,
0.07130778580904007,
-0.07880304008722305,
-0.26867708563804626,
0.037747375667095184,
0.007822367362678051,
0.07625961303710938,
-0.051242806017398834,
-0.14312531054019928,
0.029443323612213135,
-0.012245007790625095,
-0.0007672050851397216,
0.09235537797212601,
-0.29802948236465454,
-0.15575692057609558,
0.17060205340385437,
0.043730489909648895,
0.1686781495809555,
-0.08069398254156113,
-0.05032600089907646,
-0.08706816285848618,
-0.158794566988945,
0.025325428694486618,
-0.13051554560661316,
0.0806502103805542,
0.0006698024226352572,
0.049134500324726105,
0.03194265067577362,
-0.045712992548942566,
0.1579827517271042,
0.030016692355275154,
-0.0005587919731624424,
-0.05735376477241516,
-0.026291532441973686,
0.01596114970743656,
0.013952843844890594,
0.07258599996566772,
-0.08012593537569046,
0.05515005439519882,
-0.110883928835392,
-0.03982483595609665,
-0.07660875469446182,
0.09744945168495178,
-0.03101804107427597,
-0.026918288320302963,
-0.08824873715639114,
0.014109198935329914,
0.03811926394701004,
-0.00908446591347456,
0.0962422788143158,
-0.09888793528079987,
-0.0015149010578170419,
0.0730455294251442,
0.2172602266073227,
-0.008591579273343086,
0.11385484039783478,
0.010044820606708527,
-0.0023165473248809576,
0.07641846686601639,
-0.10259316861629486,
0.028711779043078423,
0.15646803379058838,
0.026797190308570862,
0.04815082997083664,
0.03338273987174034,
-0.08310088515281677,
-0.018814269453287125,
0.1001296266913414,
-0.06542666256427765,
-0.0711185559630394,
-0.10911905020475388,
0.010754607617855072,
0.04661419242620468,
0.04682145640254021,
0.1488005816936493,
-0.04327710345387459,
-0.008818818256258965,
-0.028316006064414978,
-0.05072356015443802,
-0.0828469768166542,
0.22147905826568604,
0.07815995812416077,
0.06286395341157913,
-0.09437797963619232,
0.04521525651216507,
0.0011203160975128412,
0.03209218382835388,
0.020366793498396873,
0.15299633145332336,
-0.11939382553100586,
-0.10868803411722183,
0.06397537887096405,
0.1939624845981598,
-0.16425278782844543,
-0.09930428117513657,
-0.12582385540008545,
-0.11866503953933716,
0.004289319273084402,
0.18659508228302002,
0.09276051074266434,
-0.02953184023499489,
-0.04292576014995575,
-0.07236737012863159,
-0.015460293740034103,
0.035987626761198044,
0.14336061477661133,
0.03393687680363655,
-0.06965192407369614,
0.12923552095890045,
-0.01902848854660988,
0.13263966143131256,
-0.0588928647339344,
-0.01582559011876583,
-0.08991747349500656,
0.09431697428226471,
-0.22629013657569885,
-0.0028061505872756243,
-0.04602424055337906,
-0.02380671724677086,
-0.01618875004351139,
-0.06845083832740784,
-0.049387045204639435,
0.032949719578027725,
-0.11157773435115814,
0.03023529052734375,
-0.0697985291481018,
0.07955125719308853,
-0.006573433522135019,
-0.006925638299435377,
0.0663638561964035,
-0.07090011984109879,
0.026212746277451515,
0.07666855305433273,
-0.06712788343429565,
0.17251069843769073,
-0.16754990816116333,
0.013837786391377449,
-0.002336397534236312,
0.062187910079956055,
0.03356492891907692,
-0.04491019994020462,
0.017494244500994682,
0.05624759942293167,
0.11172628402709961,
0.02285376563668251,
-0.032507799565792084,
-0.06222148612141609,
-0.058056384325027466,
-0.061434514820575714,
-0.0922132134437561,
-0.018379483371973038,
0.06736551225185394,
0.046705763787031174,
0.06289549171924591,
0.09685657173395157,
-0.09362667798995972,
0.05784488841891289,
-0.06787732988595963,
-0.010172479785978794,
-0.0012354189530014992,
-0.07692597061395645,
-0.06505681574344635,
-0.09087013453245163,
0.09123918414115906,
-0.020442213863134384,
0.12641768157482147,
-0.04637434706091881,
0.031214188784360886,
-0.017408521845936775,
0.03366268053650856,
0.07865840196609497,
0.023569956421852112,
0.2205260545015335,
0.02470816858112812,
0.04164375364780426,
-0.05750175192952156,
0.015269921161234379,
0.0156038086861372,
0.12529154121875763,
0.14662835001945496,
0.16087184846401215,
0.015528787858784199,
0.12855572998523712,
-0.0013865503715351224,
0.007953155785799026,
0.019481051713228226,
-0.08353973180055618,
0.007047333754599094,
0.02232562191784382,
-0.02714814990758896,
0.12401817739009857,
0.17382900416851044,
-0.13781563937664032,
0.03483574092388153,
-0.061411626636981964,
-0.10050879418849945,
-0.141856849193573,
-0.1615498661994934,
-0.106770358979702,
-0.09695596992969513,
0.019070366397500038,
-0.08032909035682678,
-0.016655895859003067,
0.015577757731080055,
0.10164356231689453,
-0.04961265251040459,
0.206912562251091,
0.032367367297410965,
-0.11533848196268082,
0.0952874943614006,
-0.043212004005908966,
0.050738900899887085,
0.056018680334091187,
-0.010894078761339188,
-0.007717194966971874,
-0.07637061178684235,
0.016146177425980568,
0.0034932508133351803,
-0.09210208803415298,
0.015182727947831154,
-0.03978579491376877,
-0.06738632917404175,
-0.02116398513317108,
0.029746050015091896,
0.08788377046585083,
0.1830219030380249,
0.04262876510620117,
-0.01955346204340458,
0.0006461511948145926,
0.11416105926036835,
-0.012963689863681793,
-0.17403925955295563,
-0.11950952559709549,
0.1871212273836136,
0.06487806141376495,
0.04618682339787483,
0.008408553898334503,
-0.009127442725002766,
0.02706526592373848,
0.3117702901363373,
0.2465764880180359,
-0.06424082070589066,
-0.014098159037530422,
0.011757227592170238,
0.02521596848964691,
0.01346118375658989,
0.09998629242181778,
0.042985327541828156,
0.16384638845920563,
-0.06529894471168518,
-0.08434753119945526,
-0.06327579915523529,
-0.020851388573646545,
-0.09378155320882797,
0.11766384541988373,
0.04687928408384323,
-0.08843761682510376,
-0.005371768027544022,
0.14156338572502136,
-0.10860739648342133,
-0.006172310560941696,
-0.049296800047159195,
-0.09114862978458405,
-0.10960453003644943,
-0.04970785602927208,
0.016255877912044525,
0.03765437752008438,
0.042112771421670914,
-0.06716272979974747,
-0.03251567855477333,
-0.004731037188321352,
0.025191999971866608,
-0.10520124435424805,
-0.06153574958443642,
0.12089362740516663,
-0.047358401119709015,
0.1266084611415863,
-0.011368351057171822,
0.15603423118591309,
0.1102476492524147,
0.05746765434741974,
-0.011550636030733585,
0.1020030602812767,
0.0783248320221901,
-0.03264625743031502,
0.062273088842630386,
0.06223654747009277,
-0.02739381231367588,
0.001606744364835322,
0.037881962954998016,
-0.11702331900596619,
0.09138418734073639,
0.07537859678268433,
0.012805039063096046,
-0.10182074457406998,
0.08361755311489105,
-0.11058178544044495,
0.11374874413013458,
0.1463797241449356,
-0.028194086626172066,
0.010148935951292515,
-0.08713217824697495,
0.05788881331682205,
0.010874228551983833,
-0.010617641732096672,
-0.046551238745450974,
-0.19032250344753265,
-0.05414820834994316,
-0.06366172432899475,
0.0024869453627616167,
-0.18242672085762024,
-0.02467728778719902,
-0.06765307486057281,
-0.03724600747227669,
-0.03213454410433769,
0.14669546484947205,
0.09171243011951447,
0.016895977780222893,
0.02629663608968258,
-0.11891276389360428,
0.0018964578630402684,
0.0965452566742897,
-0.17318075895309448,
-0.07812336832284927
] |
null | null |
transformers
|
# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa
Finetuning `airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`.
Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py).
Run with:
```
export MODEL_NAME=wangchan-deberta_v1-base-wiki-20210520-news-spm
CUDA_LAUNCH_BLOCKING=1 python train_question_answering_lm_finetuning.py \
--model_name $MODEL_NAME \
--dataset_name chimera_qa \
--revision mlm@ckp-41100 \
--output_dir $MODEL_NAME-finetune-chimera_qa-model \
--log_dir $MODEL_NAME-finetune-chimera_qa-log \
--model_max_length 400 \
--pad_on_right \
--fp16 \
--use_auth_token
```
|
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
|
question-answering
|
cstorm125/wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa
|
[
"transformers",
"pytorch",
"deberta",
"question-answering",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #deberta #question-answering #endpoints_compatible #region-us
|
# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa
Finetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.
Trained with thai2transformers.
Run with:
|
[
"# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
"TAGS\n#transformers #pytorch #deberta #question-answering #endpoints_compatible #region-us \n",
"# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
31,
166
] |
[
"passage: TAGS\n#transformers #pytorch #deberta #question-answering #endpoints_compatible #region-us \n# wangchan-deberta_v1-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearch/wangchan-deberta_v1-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
-0.03964553028345108,
-0.13523806631565094,
-0.0036429017782211304,
0.0469239205121994,
0.041265372186899185,
-0.001220006844960153,
0.09441833198070526,
0.08874376863241196,
-0.010469491593539715,
0.013886729255318642,
0.06614906340837479,
0.037891749292612076,
0.050557609647512436,
0.05073443427681923,
-0.07806675136089325,
-0.2249918282032013,
0.09185003489255905,
0.11173123121261597,
-0.04190700128674507,
0.14337725937366486,
0.13323982059955597,
-0.06521935015916824,
0.07074622064828873,
0.06181171163916588,
-0.09304570406675339,
0.005544273182749748,
0.02328803390264511,
-0.09098009020090103,
0.14152809977531433,
0.005325653590261936,
0.14552246034145355,
0.0884888768196106,
-0.03917224705219269,
-0.19798417389392853,
0.04782523587346077,
-0.004278290085494518,
0.037588175386190414,
-0.005781333427876234,
-0.026305461302399635,
0.05793054774403572,
-0.027301188558340073,
-0.027283474802970886,
0.036811623722314835,
0.00043573984294198453,
-0.0988268181681633,
-0.07832182198762894,
-0.04809620976448059,
0.01812036707997322,
0.14755919575691223,
0.09371072053909302,
-0.03486338257789612,
0.18911856412887573,
-0.19289329648017883,
0.04097588360309601,
0.11081963777542114,
-0.39115962386131287,
-0.013333627954125404,
0.04986076056957245,
0.024028578773140907,
0.047075774520635605,
-0.03275908902287483,
0.015083156526088715,
0.05320301651954651,
0.0334831178188324,
-0.05875849351286888,
-0.06755463778972626,
-0.027838924899697304,
-0.02737538330256939,
-0.09628593921661377,
0.016434159129858017,
0.16484282910823822,
0.05211958289146423,
-0.06287345290184021,
-0.10004466027021408,
-0.04829955846071243,
0.033349934965372086,
-0.003967587370425463,
-0.12086115032434464,
-0.029394039884209633,
0.017244262620806694,
0.00012411287752911448,
0.015140073373913765,
-0.10678932815790176,
-0.07268576323986053,
-0.1480409801006317,
0.07979077845811844,
0.07017774134874344,
0.04277492314577103,
-0.1441257745027542,
0.02068425342440605,
-0.020301802083849907,
-0.13658525049686432,
-0.07542064785957336,
-0.05524318665266037,
0.0034836085978895426,
0.0201125331223011,
-0.08767865598201752,
-0.0466904379427433,
0.08556964993476868,
0.08771400153636932,
-0.15118153393268585,
0.04142029955983162,
-0.08731068670749664,
0.051801420748233795,
-0.027307147160172462,
0.15556924045085907,
-0.0225241519510746,
0.06899938732385635,
0.035446371883153915,
0.016630591824650764,
0.0004519057983998209,
0.012227696366608143,
-0.000016570536899962462,
-0.08900296688079834,
0.03326911851763725,
0.028358573094010353,
-0.08285384625196457,
0.08981844037771225,
-0.05114540830254555,
-0.027130307629704475,
0.12821324169635773,
-0.11152497678995132,
-0.09489023685455322,
0.011423435993492603,
-0.028599195182323456,
0.08693736791610718,
0.0006118241581134498,
0.09738261997699738,
-0.021307772025465965,
0.09444485604763031,
-0.020560843870043755,
-0.013734548352658749,
0.03463394194841385,
-0.04074323549866676,
0.021270081400871277,
-0.08508365601301193,
0.0050997124053537846,
-0.15369150042533875,
-0.12586164474487305,
-0.00020341953495517373,
-0.027092838659882545,
0.027918575331568718,
-0.01926829293370247,
-0.03554418310523033,
0.009732340462505817,
-0.06910140812397003,
-0.011426771059632301,
-0.02034059911966324,
-0.03312075883150101,
0.10256940871477127,
0.07174728810787201,
0.12190507352352142,
-0.10256006568670273,
0.04119809344410896,
-0.14950823783874512,
0.05795414000749588,
-0.10341285914182663,
0.0869172215461731,
-0.045411527156829834,
-0.013585643842816353,
-0.06106827035546303,
-0.023377010598778725,
-0.09593096375465393,
0.025166692212224007,
0.0063321953639388084,
0.10588479787111282,
-0.11816687881946564,
-0.0661119893193245,
0.20574630796909332,
-0.12208522856235504,
-0.14152272045612335,
0.17930607497692108,
-0.057388875633478165,
0.0964154452085495,
0.10660958290100098,
0.2176109254360199,
0.031135106459259987,
-0.13595151901245117,
0.06215289980173111,
-0.035483431071043015,
-0.025775430724024773,
0.06995284557342529,
0.06545154750347137,
0.023516010493040085,
-0.03247624635696411,
0.07479282468557358,
-0.1184515431523323,
0.03395149111747742,
-0.049229834228754044,
-0.07727234065532684,
-0.032693151384592056,
-0.08685877174139023,
0.09819796681404114,
-0.038094040006399155,
0.11750875413417816,
-0.0475754514336586,
-0.05192602053284645,
0.0036638176534324884,
0.06261371821165085,
-0.002463730750605464,
0.014301792718470097,
-0.15599124133586884,
0.051133282482624054,
-0.042097847908735275,
-0.0025282977148890495,
-0.09984544664621353,
0.0026534132193773985,
0.04040621221065521,
0.001586543396115303,
0.03158341720700264,
0.10859646648168564,
-0.009829345159232616,
-0.056467317044734955,
-0.05927613377571106,
0.04705396667122841,
-0.03321235254406929,
-0.001267920946702361,
-0.03152746334671974,
-0.00716171320527792,
0.09088297933340073,
-0.024986356496810913,
0.10676365345716476,
-0.0850147157907486,
0.03000103309750557,
0.035515494644641876,
0.11061518639326096,
0.007873614318668842,
0.06687530130147934,
0.034992579370737076,
0.0825955793261528,
0.03209371492266655,
0.033357515931129456,
0.03680216893553734,
0.008967868983745575,
-0.13732998073101044,
0.10755743086338043,
-0.05534199997782707,
0.14929375052452087,
0.13464659452438354,
-0.11067931354045868,
0.015888415277004242,
0.015085071325302124,
-0.0430341437458992,
-0.029752029106020927,
-0.09956617653369904,
0.0459517277777195,
0.1586785465478897,
-0.015766171738505363,
0.12873856723308563,
-0.1174500435590744,
-0.025319507345557213,
-0.018858464434742928,
-0.0122576504945755,
-0.011127769947052002,
0.03594670072197914,
0.017452795058488846,
-0.17317788302898407,
0.059294480830430984,
0.1913202553987503,
-0.047849081456661224,
0.13561101257801056,
-0.06919945776462555,
-0.0709945484995842,
0.005010632798075676,
0.07735864818096161,
-0.06159668043255806,
0.028194820508360863,
-0.11914651095867157,
0.04332182928919792,
0.054001010954380035,
0.06397905945777893,
0.058258961886167526,
-0.1263829618692398,
-0.0708836242556572,
0.009215966798365116,
-0.0748254582285881,
-0.1285163015127182,
0.10426085442304611,
0.05070926249027252,
0.07624394446611404,
0.04690077155828476,
0.03602917492389679,
0.04494396969676018,
0.009085949510335922,
-0.13419803977012634,
0.15050143003463745,
-0.034831639379262924,
-0.34912797808647156,
-0.03409019485116005,
0.09434914588928223,
-0.04539947211742401,
-0.04525769129395485,
0.061148907989263535,
-0.1666322946548462,
0.01588713563978672,
-0.00037349952617660165,
0.038898006081581116,
-0.0007825231878086925,
0.015726864337921143,
0.0822867676615715,
0.027400799095630646,
0.03714154660701752,
-0.08496090769767761,
-0.039934828877449036,
-0.08539316803216934,
-0.06339694559574127,
0.08496460318565369,
-0.12961672246456146,
0.038331545889377594,
0.02689221315085888,
-0.06154846027493477,
0.020044395700097084,
-0.021918315440416336,
0.25879985094070435,
-0.08735295385122299,
0.010247420519590378,
0.14928843080997467,
-0.12924548983573914,
0.02571197785437107,
0.1469881683588028,
-0.021672742441296577,
-0.09273982793092728,
0.0680340975522995,
0.056351084262132645,
-0.01574140600860119,
-0.24390535056591034,
-0.01983964629471302,
-0.017649643123149872,
-0.00639861635863781,
-0.0904071107506752,
0.030973218381404877,
0.06429192423820496,
0.08317839354276657,
-0.006789624225348234,
-0.058485325425863266,
-0.030097227543592453,
0.01833752728998661,
0.2222054898738861,
0.03759608417749405,
0.12540358304977417,
-0.04600447043776512,
-0.0639946237206459,
0.024275964125990868,
0.09815912693738937,
0.08099458366632462,
0.04964635893702507,
-0.01685021072626114,
0.10912510752677917,
0.19492806494235992,
0.17979542911052704,
0.026352746412158012,
-0.06892112642526627,
-0.049460191279649734,
-0.01391538791358471,
-0.014122571796178818,
-0.06931345909833908,
0.03594903647899628,
-0.022734595462679863,
-0.0301434975117445,
-0.012731348164379597,
0.06825247406959534,
0.06559862196445465,
0.22153283655643463,
0.04618977755308151,
-0.11729921400547028,
-0.10630511492490768,
0.019386667758226395,
-0.092210553586483,
0.007764512673020363,
0.06473853439092636,
0.0319364033639431,
-0.1373714953660965,
0.07070440798997879,
0.029348159208893776,
0.12758395075798035,
-0.07663482427597046,
0.07175026834011078,
-0.05240427702665329,
-0.18168215453624725,
0.01806592382490635,
0.03802090883255005,
-0.36218082904815674,
0.1855057328939438,
0.010046476498246193,
0.026293180882930756,
-0.04538657143712044,
-0.04299401864409447,
-0.03909990191459656,
0.05689476430416107,
0.07100257277488708,
-0.020124539732933044,
0.007341781165450811,
-0.09641318768262863,
-0.03052595630288124,
0.09976986795663834,
0.05611196160316467,
0.038535211235284805,
0.035100311040878296,
0.01787007600069046,
0.048928212374448776,
0.018775269389152527,
0.05187604948878288,
-0.2168058305978775,
-0.024653209373354912,
0.0328778401017189,
0.015289828181266785,
0.03332621604204178,
-0.04894150421023369,
-0.05026279762387276,
-0.1221374049782753,
0.11329550296068192,
-0.11129748076200485,
-0.11473153531551361,
-0.06498581916093826,
0.06818735599517822,
0.01573241502046585,
-0.07742141932249069,
-0.005606579128652811,
-0.019919319078326225,
0.022299811244010925,
-0.04385669156908989,
-0.05367192625999451,
0.07615046948194504,
-0.02322353795170784,
-0.1251642256975174,
0.023103075101971626,
0.09773983806371689,
0.026317337527871132,
0.05572840943932533,
0.08485281467437744,
-0.026214003562927246,
0.04837996885180473,
-0.0709821954369545,
-0.03537744656205177,
0.001126654096879065,
-0.06717360019683838,
-0.036802563816308975,
-0.09041166305541992,
0.008905279450118542,
-0.13697311282157898,
-0.07513025403022766,
0.16341543197631836,
0.15817369520664215,
-0.051550570875406265,
0.09390424937009811,
0.17113140225410461,
0.013964718207716942,
-0.12953655421733856,
-0.008122354745864868,
0.011574974283576012,
0.039610009640455246,
-0.05266452953219414,
-0.05424536392092705,
0.12159707397222519,
0.11533783376216888,
-0.02428716979920864,
0.003065894590690732,
-0.16135235130786896,
-0.11518917977809906,
0.1041150838136673,
-0.017578257247805595,
0.29201704263687134,
-0.12603570520877838,
-0.04217913746833801,
0.00486409105360508,
-0.2568013072013855,
0.07150240242481232,
-0.06523620337247849,
0.058289334177970886,
-0.03482051566243172,
0.07154475152492523,
-0.008053096011281013,
-0.001326818484812975,
0.13984255492687225,
0.044788505882024765,
0.007591368164867163,
-0.060701511800289154,
-0.012231211178004742,
0.06510938704013824,
0.0378466472029686,
0.10012119263410568,
0.003920756746083498,
0.054012712091207504,
-0.17143192887306213,
-0.06457504630088806,
-0.08572402596473694,
-0.03594553470611572,
-0.0452277772128582,
-0.06023741513490677,
-0.03854816406965256,
0.034546274691820145,
0.014912388287484646,
0.011260184459388256,
0.019980430603027344,
-0.08583595603704453,
0.05478881672024727,
0.003247666172683239,
0.08813011646270752,
-0.07277432829141617,
0.04613110050559044,
-0.041859615594148636,
-0.05187537521123886,
0.12257750332355499,
-0.178974911570549,
0.04606940597295761,
0.10309114307165146,
-0.014022739604115486,
0.08661479502916336,
0.03414091095328331,
-0.03467842936515808,
0.09165123105049133,
0.037856586277484894,
-0.106887087225914,
-0.10682990401983261,
0.037379730492830276,
-0.13799163699150085,
0.023923518136143684,
0.01002503465861082,
0.117428719997406,
-0.022164050489664078,
-0.041465479880571365,
0.009220607578754425,
-0.013071568682789803,
-0.0800926461815834,
0.0819300189614296,
0.05743185058236122,
0.03321155160665512,
-0.08499567955732346,
0.09456615895032883,
0.04912583902478218,
-0.15450482070446014,
0.0023081479594111443,
0.027613256126642227,
-0.17443880438804626,
-0.042807597666978836,
-0.11280424147844315,
0.05440044403076172,
-0.02909133769571781,
-0.0850832387804985,
-0.10940232872962952,
-0.1427372843027115,
0.04591602459549904,
0.1431533694267273,
0.07215593010187149,
0.03838910534977913,
0.03570856153964996,
0.005145614966750145,
0.0005491864285431802,
0.030433781445026398,
0.07339853048324585,
0.05918349698185921,
-0.09474094957113266,
-0.04255056008696556,
0.02775123529136181,
0.17242343723773956,
-0.05170850083231926,
-0.022351011633872986,
-0.08758708089590073,
0.07108794897794724,
-0.24875445663928986,
0.060475368052721024,
-0.028482867404818535,
-0.03546518459916115,
-0.03008725307881832,
-0.15927094221115112,
-0.05026839300990105,
0.01764511875808239,
-0.03457817807793617,
0.031150933355093002,
-0.03245268017053604,
-0.021383823826909065,
-0.07610439509153366,
-0.018650807440280914,
0.11815597116947174,
-0.06765275448560715,
0.1063288152217865,
0.06456310302019119,
-0.07578178495168686,
0.10065814107656479,
-0.08372348546981812,
-0.04496735706925392,
0.059810999780893326,
0.025451544672250748,
0.027658823877573013,
0.003491915762424469,
0.045853979885578156,
0.05110572278499603,
0.05981012433767319,
0.05928643047809601,
0.2085205316543579,
-0.11118753254413605,
-0.04339536651968956,
-0.03814433515071869,
-0.0679766833782196,
-0.04707787558436394,
-0.0881853923201561,
0.16501273214817047,
0.062335219234228134,
0.12267941236495972,
-0.07359664887189865,
0.05564722418785095,
-0.10055836290121078,
-0.005725170020014048,
0.013124492950737476,
-0.07603725045919418,
0.01857525110244751,
-0.04866505786776543,
0.02705702930688858,
-0.0535009391605854,
0.2014784961938858,
-0.11974319815635681,
-0.004596707411110401,
0.06004063040018082,
-0.04909021034836769,
-0.022361215204000473,
0.029932817444205284,
0.2606382966041565,
0.11327296495437622,
-0.010474716313183308,
-0.08197222650051117,
-0.0530007928609848,
-0.020716916769742966,
-0.017303647473454475,
0.06908117234706879,
0.2986431121826172,
-0.02387140318751335,
0.0765424445271492,
0.07696174085140228,
0.1154627725481987,
-0.05433383211493492,
-0.03465309366583824,
-0.11879149824380875,
-0.03074863739311695,
0.032842569053173065,
-0.014248042367398739,
0.26189202070236206,
-0.08219341933727264,
0.02795613743364811,
-0.04316955804824829,
-0.05906490236520767,
-0.13355639576911926,
-0.012706642970442772,
-0.13744962215423584,
-0.10553054511547089,
0.05745919421315193,
-0.07193827629089355,
-0.06217426434159279,
0.1376897245645523,
0.07039018720388412,
-0.044483792036771774,
0.21120195090770721,
0.08641792088747025,
-0.0694768950343132,
0.08345131576061249,
-0.04902343451976776,
-0.007738600485026836,
0.05599581450223923,
0.011617873795330524,
0.04914214462041855,
0.056667134165763855,
0.004958674777299166,
-0.003949407488107681,
-0.05323909968137741,
-0.046926774084568024,
-0.10858152061700821,
-0.0458916611969471,
-0.008828933350741863,
0.038102224469184875,
-0.0017246523639187217,
0.07838600128889084,
0.055669937282800674,
0.036806803196668625,
-0.002607713220641017,
0.09152407944202423,
0.03463519364595413,
-0.1870381236076355,
-0.18481120467185974,
0.037760209292173386,
0.028327811509370804,
0.06573953479528427,
0.040074799209833145,
-0.056129105389118195,
-0.03259839490056038,
0.31550073623657227,
0.19431228935718536,
-0.05566665157675743,
0.04109997674822807,
0.05412213131785393,
0.042672500014305115,
-0.004334264900535345,
0.04501243680715561,
0.11548613756895065,
0.14313171803951263,
-0.06790722906589508,
-0.09975873678922653,
-0.06202331930398941,
-0.08381101489067078,
-0.08324384689331055,
0.04168587177991867,
0.06578278541564941,
-0.007668302860110998,
-0.08741346746683121,
0.08800508826971054,
-0.04951449856162071,
0.00251093041151762,
-0.05109037831425667,
-0.14309506118297577,
-0.11874251812696457,
-0.03325094282627106,
-0.07695502042770386,
0.0003530853136908263,
0.014396986924111843,
-0.0074133556336164474,
0.009965546429157257,
0.04610896110534668,
0.01752399280667305,
-0.024230988696217537,
0.002997289877384901,
0.1286107897758484,
0.08856961876153946,
-0.03553637117147446,
0.04571190103888512,
0.16436034440994263,
0.0471607968211174,
0.09231231361627579,
0.025476500391960144,
0.15932036936283112,
0.06029670685529709,
0.06884951889514923,
-0.07559294998645782,
0.058610476553440094,
0.03401279076933861,
0.027434593066573143,
0.12680977582931519,
-0.16601425409317017,
0.05144305154681206,
-0.03349217772483826,
-0.06154589354991913,
-0.1470811516046524,
0.11567503213882446,
-0.04427654296159744,
0.08556554466485977,
0.1527124047279358,
-0.03687634691596031,
0.053801365196704865,
-0.040937405079603195,
0.046814195811748505,
0.022740036249160767,
-0.075853630900383,
-0.06855916976928711,
-0.1666407734155655,
0.027590613812208176,
0.019091598689556122,
0.0025507472455501556,
-0.2284461408853531,
-0.03560779243707657,
-0.02970220521092415,
0.003730073105543852,
0.003377749351784587,
0.06043329834938049,
0.06482624262571335,
0.034312035888433456,
-0.0032150449696928263,
-0.2682606279850006,
0.02524673007428646,
0.06748870760202408,
-0.13957887887954712,
-0.09159059822559357
] |
null | null |
transformers
|
# airesearch/wangchanberta-base-att-spm-uncased
Finetuning `airesearch/wangchanberta-base-att-spm-uncased` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`.
Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py).
Run with:
```
export MODEL_NAME=airesearch/wangchanberta-base-att-spm-uncased
python train_question_answering_lm_finetuning.py \
--model_name $MODEL_NAME \
--dataset_name chimera_qa \
--output_dir $MODEL_NAME-finetune-chimera_qa-model \
--log_dir $MODEL_NAME-finetune-chimera_qa-log \
--lowercase \
--pad_on_right \
--fp16
```
|
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
|
question-answering
|
cstorm125/wangchanberta-base-att-spm-uncased-finetune-qa
|
[
"transformers",
"pytorch",
"camembert",
"question-answering",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us
|
# airesearch/wangchanberta-base-att-spm-uncased
Finetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.
Trained with thai2transformers.
Run with:
|
[
"# airesearch/wangchanberta-base-att-spm-uncased\n\nFinetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
"TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n",
"# airesearch/wangchanberta-base-att-spm-uncased\n\nFinetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
31,
154
] |
[
"passage: TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n# airesearch/wangchanberta-base-att-spm-uncased\n\nFinetuning 'airesearch/wangchanberta-base-att-spm-uncased' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
-0.06169361621141434,
-0.10002221912145615,
-0.004205552861094475,
0.07189788669347763,
0.053869858384132385,
0.00043351310887373984,
0.07556740939617157,
0.09258310496807098,
0.0015330344904214144,
0.018894976004958153,
0.056050099432468414,
0.03379638120532036,
0.06882385164499283,
0.02088298462331295,
-0.07881523668766022,
-0.14780306816101074,
0.06325415521860123,
0.07154237478971481,
-0.039426371455192566,
0.1499026119709015,
0.08045346289873123,
-0.05171479284763336,
0.08370348066091537,
0.06119164824485779,
-0.10609216243028641,
-0.015914801508188248,
0.013856706209480762,
-0.10142084211111069,
0.1307935267686844,
-0.029467513784766197,
0.18172860145568848,
0.05570966750383377,
-0.07346896827220917,
-0.17744092643260956,
0.05200200900435448,
0.0028096248861402273,
0.06438665091991425,
0.004760744050145149,
-0.013892127200961113,
0.07254758477210999,
-0.027095070108771324,
0.055308498442173004,
0.04892750829458237,
0.015893742442131042,
-0.10955753177404404,
-0.022485611960291862,
-0.06075170263648033,
-0.014564082957804203,
0.12420249730348587,
0.05757579207420349,
-0.019274257123470306,
0.2233375608921051,
-0.2180650681257248,
0.04848141968250275,
0.03335811570286751,
-0.37762749195098877,
-0.01927879825234413,
0.03250092640519142,
-0.00508640892803669,
0.03554008528590202,
-0.07425438612699509,
-0.018375463783740997,
0.02229929529130459,
0.03983553498983383,
-0.04115912690758705,
-0.08321043103933334,
-0.13996732234954834,
-0.02841419354081154,
-0.07779299467802048,
0.01802234724164009,
0.15968483686447144,
0.062126852571964264,
-0.08258774131536484,
-0.07678709924221039,
-0.02983308769762516,
0.06028047204017639,
0.0077493153512477875,
-0.08891028165817261,
-0.03643720969557762,
0.007560596335679293,
-0.015414334833621979,
0.020807800814509392,
-0.10042005777359009,
-0.06494557112455368,
-0.11268854141235352,
0.03826236352324486,
0.06577445566654205,
0.041410256177186966,
-0.09102389216423035,
0.009106609970331192,
-0.030024448409676552,
-0.11758098006248474,
-0.0799388661980629,
-0.074864961206913,
-0.042097192257642746,
0.0007490995340049267,
-0.07643626630306244,
-0.046362247318029404,
0.12150228768587112,
0.0990038588643074,
-0.06157189980149269,
0.04946226254105568,
-0.08012381196022034,
0.05913054570555687,
-0.05930085480213165,
0.1384335309267044,
0.013986639678478241,
0.10091322660446167,
0.026065785437822342,
0.010957028716802597,
-0.039922911673784256,
0.004855569452047348,
0.030675673857331276,
-0.047453414648771286,
0.03447260335087776,
0.025958500802516937,
-0.09219276905059814,
0.08182504028081894,
-0.07938089966773987,
-0.037510983645915985,
0.15957383811473846,
-0.09615491330623627,
-0.07862246781587601,
0.018472490832209587,
-0.007903626188635826,
0.06077247112989426,
0.021369000896811485,
0.10068856924772263,
-0.024465274065732956,
0.04754210263490677,
-0.008798479102551937,
0.010557956993579865,
0.04452326521277428,
-0.03458027169108391,
0.012442853301763535,
-0.11534678936004639,
0.0009347318555228412,
-0.14984340965747833,
-0.1437915414571762,
-0.0070157768204808235,
0.006285584066063166,
0.0446736179292202,
-0.006384925451129675,
-0.028707843273878098,
0.016606729477643967,
-0.08181065320968628,
-0.008981768041849136,
-0.02999919466674328,
-0.03934856876730919,
0.08228738605976105,
0.047177936881780624,
0.12361925095319748,
-0.125436931848526,
0.04347721114754677,
-0.18220451474189758,
0.045577261596918106,
-0.1444200724363327,
0.054821815341711044,
-0.05909958481788635,
0.02893843501806259,
-0.04693695902824402,
-0.015292626805603504,
-0.0862935408949852,
-0.0027207338716834784,
0.015572995878756046,
0.12878727912902832,
-0.1123671680688858,
-0.06375480443239212,
0.1725902259349823,
-0.1311982125043869,
-0.18728207051753998,
0.19393785297870636,
-0.04058343544602394,
0.07396263629198074,
0.06410014629364014,
0.2733392119407654,
0.02717333287000656,
-0.08730173110961914,
0.046496644616127014,
-0.04424462467432022,
-0.025206154212355614,
0.0005662143230438232,
0.04915955290198326,
0.00440028915181756,
-0.04320681467652321,
0.08445252478122711,
-0.08673129230737686,
0.05889536440372467,
-0.046149078756570816,
-0.09840618818998337,
-0.0703902542591095,
-0.11425910145044327,
0.06548812240362167,
-0.027742184698581696,
0.12231721729040146,
-0.05326024815440178,
0.006483196746557951,
-0.0592370368540287,
0.04442058131098747,
0.012288780882954597,
0.0022907035890966654,
-0.11051249504089355,
0.08227857202291489,
-0.04858686402440071,
-0.023635245859622955,
-0.11662784963846207,
-0.020959611982107162,
0.0523233562707901,
0.009824077598750591,
0.04187772423028946,
0.037129007279872894,
0.004784129094332457,
-0.01726248860359192,
-0.041229620575904846,
0.02251034416258335,
0.0057913619093596935,
-0.005497370380908251,
-0.031967002898454666,
-0.02054222859442234,
0.06335511803627014,
-0.03324764594435692,
0.08697859197854996,
-0.10404743999242783,
0.027222832664847374,
-0.015048091299831867,
0.12252667546272278,
0.02222803421318531,
0.011955438181757927,
0.08058122545480728,
0.10088168829679489,
0.030368493869900703,
0.023635977879166603,
0.061980314552783966,
0.0252189002931118,
-0.13756632804870605,
0.06765339523553848,
0.0053394390270113945,
0.14501088857650757,
0.1414213329553604,
-0.12246610224246979,
-0.01764488033950329,
-0.017051201313734055,
-0.04647308960556984,
-0.04518350586295128,
-0.08538533002138138,
0.10162870585918427,
0.200114905834198,
0.013411130756139755,
0.1528967171907425,
-0.10508503764867783,
-0.004123337101191282,
-0.011907669715583324,
-0.018922701478004456,
0.014597728848457336,
0.049043554812669754,
0.05241959169507027,
-0.14901019632816315,
0.06038863956928253,
0.22179122269153595,
-0.05515649542212486,
0.141051784157753,
-0.08156585693359375,
-0.057337693870067596,
0.02244514413177967,
0.12324333935976028,
-0.04422465339303017,
0.0336410328745842,
-0.14592836797237396,
0.014853546395897865,
0.06479912996292114,
0.0414838008582592,
0.022096576169133186,
-0.13086377084255219,
-0.057797327637672424,
-0.0037631848827004433,
-0.052512235939502716,
-0.1335785835981369,
0.08514571934938431,
0.05557866394519806,
0.07111447304487228,
0.005498175974935293,
-0.0034974655136466026,
0.02184721827507019,
0.007310779765248299,
-0.10709719359874725,
0.14943477511405945,
-0.027205105870962143,
-0.2909665107727051,
-0.031847432255744934,
0.13552425801753998,
-0.09623327851295471,
-0.049343179911375046,
0.04693479835987091,
-0.14313484728336334,
0.03745276480913162,
0.004610451404005289,
-0.017236188054084778,
0.04192980378866196,
0.01729697361588478,
0.017933903262019157,
0.011490057222545147,
0.041749339550733566,
-0.06797100603580475,
-0.03891343995928764,
-0.07998894155025482,
-0.07915297895669937,
0.10557576268911362,
-0.09593117982149124,
0.043032750487327576,
0.017215635627508163,
-0.07243600487709045,
0.021132053807377815,
-0.004550217650830746,
0.24405474960803986,
-0.05479152128100395,
0.012433988973498344,
0.10716351866722107,
-0.1253141164779663,
0.02880443073809147,
0.10679998993873596,
-0.014378624968230724,
-0.09109317511320114,
0.04782459884881973,
0.03865228220820427,
-0.03956577926874161,
-0.23330718278884888,
-0.021192627027630806,
-0.03961390256881714,
0.0038678369019180536,
-0.05782794579863548,
0.021200846880674362,
0.024960583075881004,
0.10769889503717422,
0.02499094232916832,
-0.08388269692659378,
-0.09722240269184113,
0.00567483389750123,
0.11865472793579102,
0.047723475843667984,
0.1299983710050583,
-0.04046669229865074,
-0.04323042184114456,
0.02878458984196186,
0.12913280725479126,
0.1296088546514511,
0.03773890808224678,
-0.029365822672843933,
0.10636596381664276,
0.2556578814983368,
0.15492616593837738,
0.052548978477716446,
-0.07707565277814865,
-0.06477530300617218,
-0.006206491030752659,
-0.022761518135666847,
-0.10691992193460464,
0.0416632741689682,
0.012434408068656921,
-0.005094287917017937,
-0.03379865735769272,
0.04600908234715462,
0.07021338492631912,
0.25410622358322144,
0.02924523502588272,
-0.09427227824926376,
-0.08052118122577667,
-0.008222493343055248,
-0.09003116190433502,
-0.008498786017298698,
0.06659353524446487,
0.037630561739206314,
-0.15649546682834625,
0.04546568542718887,
0.009362280368804932,
0.12366414815187454,
-0.07129386812448502,
0.05941012129187584,
-0.0680471882224083,
-0.15080858767032623,
0.01797897182404995,
0.01667250692844391,
-0.3056921660900116,
0.19709573686122894,
0.008104043081402779,
0.02410631813108921,
-0.025889290496706963,
-0.04305456206202507,
0.007793188560754061,
0.022794902324676514,
0.09057972580194473,
-0.006996447220444679,
-0.018983198329806328,
-0.11567658931016922,
-0.031826868653297424,
0.12249711155891418,
0.06111735850572586,
0.0763811469078064,
0.017325384542346,
0.03975439816713333,
0.059196632355451584,
0.0039195953868329525,
0.05778883025050163,
-0.20031100511550903,
-0.009421936236321926,
0.011063692159950733,
0.029392186552286148,
-0.008061188273131847,
-0.032133717089891434,
-0.031081847846508026,
-0.17016977071762085,
0.17616112530231476,
-0.06621050089597702,
-0.10067577660083771,
-0.07693468034267426,
0.003513741074129939,
0.03810020908713341,
-0.06551644206047058,
0.029722420498728752,
-0.06026227027177811,
-0.04726657271385193,
-0.02681538462638855,
-0.048768751323223114,
0.07808596640825272,
-0.03581555560231209,
-0.07505828887224197,
0.005627943202853203,
0.06264296174049377,
-0.004832218401134014,
0.05743733048439026,
0.07679804414510727,
-0.011898175813257694,
0.017700601369142532,
-0.08890075981616974,
-0.014457624405622482,
-0.0444558821618557,
-0.052028998732566833,
0.040820639580488205,
-0.10606168210506439,
0.04327971488237381,
-0.11109162122011185,
-0.08143185079097748,
0.19476056098937988,
0.16286590695381165,
-0.03830110281705856,
0.11106874793767929,
0.17610420286655426,
-0.0002615197154227644,
-0.11108866333961487,
-0.004705268424004316,
0.01022527739405632,
0.052097249776124954,
-0.059899818152189255,
-0.06701533496379852,
0.16358929872512817,
0.11553844809532166,
-0.008818134665489197,
-0.013109995983541012,
-0.08157234638929367,
-0.10417088866233826,
0.1317957043647766,
0.03267379477620125,
0.2772156596183777,
-0.11232812702655792,
-0.05089246481657028,
0.038667865097522736,
-0.1873830407857895,
0.03267952799797058,
-0.06122628599405289,
0.08475341647863388,
-0.05026750639081001,
0.04026272892951965,
-0.00038236932596191764,
-0.02417699806392193,
0.13777956366539001,
0.07060838490724564,
0.03893033042550087,
-0.033857300877571106,
0.030088260769844055,
0.01789572462439537,
0.04270599037408829,
0.11460387706756592,
0.0022367697674781084,
0.06534150242805481,
-0.1819716989994049,
-0.01358972117304802,
-0.10792557895183563,
-0.00410406943410635,
-0.005451282951980829,
-0.05070118606090546,
-0.027830207720398903,
0.04011179506778717,
0.005807036999613047,
0.032246701419353485,
0.0017646305495873094,
-0.06203322485089302,
0.1301763951778412,
0.08305943757295609,
0.08800799399614334,
-0.11015021055936813,
0.04797990992665291,
-0.05137540400028229,
-0.04442062973976135,
0.1126914769411087,
-0.14321351051330566,
0.03404460474848747,
0.12610577046871185,
0.017370332032442093,
0.07634872198104858,
0.036785539239645004,
0.009356587193906307,
0.08525122702121735,
0.047443728893995285,
-0.10432633757591248,
-0.059030722826719284,
0.032023631036281586,
-0.14456824958324432,
-0.007398600224405527,
0.016361868008971214,
0.1019783616065979,
-0.03853105008602142,
-0.027260305359959602,
-0.00385312270373106,
-0.03982248529791832,
-0.08822116255760193,
0.09679339081048965,
0.07738488167524338,
0.03591972589492798,
-0.0967877134680748,
0.08390846848487854,
0.02563735470175743,
-0.12706442177295685,
-0.0032031163573265076,
-0.0039563109166920185,
-0.1714022159576416,
-0.05908457562327385,
-0.07526203989982605,
0.10358908027410507,
-0.018426012247800827,
-0.05482950061559677,
-0.10769602656364441,
-0.12631656229496002,
0.04333760589361191,
0.12553802132606506,
0.08480799943208694,
0.044392045587301254,
0.019968733191490173,
-0.025111857801675797,
0.0007954351603984833,
0.002413736889138818,
0.047946564853191376,
0.06741069257259369,
-0.11232953518629074,
-0.027234936133027077,
0.016353042796254158,
0.14100387692451477,
-0.03659394383430481,
-0.04088450223207474,
-0.11104282736778259,
0.06190106272697449,
-0.18850429356098175,
0.04973035305738449,
-0.05154803395271301,
-0.009293662384152412,
0.0048852404579520226,
-0.15151241421699524,
-0.04334920644760132,
0.01382733415812254,
-0.062490709125995636,
0.024343056604266167,
-0.016252733767032623,
0.02488892711699009,
-0.0672866627573967,
-0.025007013231515884,
0.12215842306613922,
-0.058295249938964844,
0.08162006735801697,
0.05741802975535393,
-0.09396599233150482,
0.09168445318937302,
-0.10742981731891632,
-0.05108426511287689,
0.05277048423886299,
0.03596753627061844,
0.03542184457182884,
-0.0027388064190745354,
0.0464644655585289,
0.049456678330898285,
0.06271960586309433,
0.0708705484867096,
0.2423771470785141,
-0.11448174715042114,
-0.10107527673244476,
-0.020830774679780006,
-0.06300143897533417,
-0.06520339846611023,
-0.07181468605995178,
0.18103745579719543,
0.057584863156080246,
0.1323997527360916,
-0.056769996881484985,
0.06504740566015244,
-0.13220000267028809,
-0.018026063218712807,
0.009753167629241943,
-0.10372498631477356,
0.027559585869312286,
-0.05973218008875847,
0.017645899206399918,
-0.03413645550608635,
0.18920549750328064,
-0.1295851171016693,
0.025772910565137863,
0.031217779964208603,
0.0029166999738663435,
-0.03343433141708374,
0.029039662331342697,
0.2753395736217499,
0.11997071653604507,
-0.03170079365372658,
-0.028389951214194298,
-0.03135001286864281,
-0.035072509199380875,
0.007990779355168343,
0.036549393087625504,
0.29954794049263,
-0.06481378525495529,
0.0648762509226799,
0.08465846627950668,
0.12617439031600952,
-0.019791804254055023,
-0.03305554389953613,
-0.07437150180339813,
0.02939327247440815,
0.03638555109500885,
-0.05411858111619949,
0.23659482598304749,
-0.10902774333953857,
0.0519835501909256,
-0.043244630098342896,
-0.07163432985544205,
-0.09900432825088501,
0.00798000954091549,
-0.10727941989898682,
-0.08851034194231033,
0.06724265962839127,
-0.0633554682135582,
-0.06808631122112274,
0.16642814874649048,
0.06810217350721359,
-0.007940524257719517,
0.19051098823547363,
0.05629001557826996,
-0.040219634771347046,
0.10916294902563095,
-0.06081356108188629,
-0.02990959770977497,
0.019412606954574585,
-0.008496788330376148,
0.032619066536426544,
0.04330091550946236,
-0.00928207952529192,
-0.0000866249029058963,
-0.04226485639810562,
-0.039829641580581665,
-0.12233851850032806,
-0.05587826669216156,
-0.012137824669480324,
-0.002866057213395834,
-0.003344410564750433,
0.07090970873832703,
0.03338225558400154,
0.034170836210250854,
0.00014552322681993246,
0.09995167702436447,
0.005156655330210924,
-0.14912573993206024,
-0.23349186778068542,
0.05662763863801956,
-0.0038171224296092987,
0.014516431838274002,
0.02510402910411358,
-0.029946263879537582,
0.00011264016211498529,
0.3191787600517273,
0.17379260063171387,
-0.06751825660467148,
0.04835344851016998,
0.0514242947101593,
0.02553579956293106,
-0.03010503575205803,
0.0723547711968422,
0.10658762603998184,
0.04303156957030296,
-0.07991938292980194,
-0.10427272319793701,
-0.08924830704927444,
-0.11073092371225357,
-0.09611870348453522,
0.04778620973229408,
0.08633558452129364,
-0.009221549145877361,
-0.105065256357193,
0.07006194442510605,
-0.05468941852450371,
0.06834042817354202,
-0.05919628217816353,
-0.14584235846996307,
-0.14787869155406952,
-0.07504476606845856,
-0.0999111607670784,
0.0444989874958992,
0.0006293095648288727,
-0.022529518231749535,
0.006499325856566429,
0.026564303785562515,
0.006221753545105457,
-0.027467550709843636,
-0.04456228390336037,
0.11044169217348099,
0.04708419740200043,
-0.04304299131035805,
0.06578782200813293,
0.13966000080108643,
0.04367126524448395,
0.09135967493057251,
0.023108504712581635,
0.1198631152510643,
0.04138518497347832,
0.07996600866317749,
-0.06453801691532135,
0.07830676436424255,
0.04432335123419762,
0.057898759841918945,
0.07512499392032623,
-0.12583927810192108,
0.05939020961523056,
-0.017162257805466652,
-0.04777595400810242,
-0.11666560173034668,
0.11338791996240616,
-0.027847182005643845,
0.1042686253786087,
0.15346252918243408,
-0.02051912620663643,
0.05718081444501877,
-0.04712582752108574,
0.05831754952669144,
0.023700280115008354,
-0.1251724660396576,
-0.1089579239487648,
-0.15888993442058563,
0.02765781618654728,
0.0028320623096078634,
-0.02728802151978016,
-0.21526964008808136,
-0.06561625003814697,
-0.02503202296793461,
-0.016123821958899498,
-0.014764483086764812,
0.07415689527988434,
0.0719742402434349,
0.04055998846888542,
0.0011646741768345237,
-0.22894616425037384,
0.04626140743494034,
0.0538916252553463,
-0.0918298065662384,
-0.09028176218271255
] |
null | null |
transformers
|
# wangchanberta-base-wiki-20210520-news-spm-finetune-qa
Finetuning `airesearchth/wangchanberta-base-wiki-20210520-news-spm` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`.
Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py).
Run with:
```
export MODEL_NAME=airesearchth/wangchanberta-base-wiki-20210520-news-spm
CUDA_LAUNCH_BLOCKING=1 python train_question_answering_lm_finetuning.py \
--model_name $MODEL_NAME \
--dataset_name chimera_qa \
--output_dir $MODEL_NAME-finetune-chimera_qa-model \
--log_dir $MODEL_NAME-finetune-chimera_qa-log \
--model_max_length 400 \
--pad_on_right \
--fp16
```
|
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
|
question-answering
|
cstorm125/wangchanberta-base-wiki-20210520-news-spm-finetune-qa
|
[
"transformers",
"pytorch",
"camembert",
"question-answering",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us
|
# wangchanberta-base-wiki-20210520-news-spm-finetune-qa
Finetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.
Trained with thai2transformers.
Run with:
|
[
"# wangchanberta-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
"TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n",
"# wangchanberta-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
31,
159
] |
[
"passage: TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n# wangchanberta-base-wiki-20210520-news-spm-finetune-qa\n\nFinetuning 'airesearchth/wangchanberta-base-wiki-20210520-news-spm' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
-0.05998677387833595,
-0.11627022922039032,
-0.003607646794989705,
0.047817789018154144,
0.026026777923107147,
0.0034189617726951838,
0.07305090874433517,
0.08659139275550842,
-0.0019146837294101715,
0.0176298338919878,
0.06823267787694931,
0.005805567372590303,
0.0698540061712265,
0.02400720678269863,
-0.08120398968458176,
-0.19051747024059296,
0.073115274310112,
0.09388849884271622,
-0.06575071066617966,
0.14523082971572876,
0.10974657535552979,
-0.06184523552656174,
0.080986388027668,
0.08090835064649582,
-0.1015891507267952,
-0.013785232789814472,
0.026258286088705063,
-0.09847381711006165,
0.14216487109661102,
0.0077484166249632835,
0.15683990716934204,
0.06991502642631531,
-0.05883506312966347,
-0.2115025818347931,
0.05869748443365097,
-0.010367019101977348,
0.04850764945149422,
0.005671354476362467,
-0.003966958727687597,
0.06872488558292389,
-0.005680215544998646,
0.006985425483435392,
0.024920504540205002,
0.013415164314210415,
-0.10346293449401855,
-0.015129061415791512,
-0.05383995547890663,
0.02034040167927742,
0.14747627079486847,
0.08754530549049377,
-0.021264584735035896,
0.20202967524528503,
-0.20548678934574127,
0.04154345393180847,
0.10880858451128006,
-0.4099293053150177,
-0.03323443606495857,
0.02706247940659523,
0.0024889043997973204,
0.0353328213095665,
-0.05015769600868225,
0.014098210260272026,
0.028940493240952492,
0.0359792523086071,
-0.08083879202604294,
-0.0603475384414196,
-0.07095978409051895,
-0.023585917428135872,
-0.09508562833070755,
0.010909431613981724,
0.17128241062164307,
0.051367223262786865,
-0.05666772276163101,
-0.09266793727874756,
-0.026610128581523895,
0.03722511976957321,
-0.0026004831306636333,
-0.08817540854215622,
-0.06186773255467415,
0.010040314868092537,
-0.005798691883683205,
0.008502845652401447,
-0.11521129310131073,
-0.07405633479356766,
-0.12667782604694366,
0.08711286634206772,
0.07361230254173279,
0.05218959227204323,
-0.12901048362255096,
0.013267606496810913,
-0.006514499895274639,
-0.13073286414146423,
-0.07734176516532898,
-0.06050126627087593,
-0.007644480559974909,
0.013665328733623028,
-0.08424083143472672,
-0.03474488481879234,
0.09256439656019211,
0.036221083253622055,
-0.13586024940013885,
0.04196052625775337,
-0.0655650720000267,
0.058089353144168854,
-0.04812217131257057,
0.14354924857616425,
0.02502429112792015,
0.09696486592292786,
0.018876347690820694,
0.02874775230884552,
-0.03350912779569626,
0.015201115980744362,
0.018791284412145615,
-0.06562302261590958,
0.0007265793974511325,
0.025031480938196182,
-0.0861712172627449,
0.07147690653800964,
-0.05354004353284836,
-0.021903200075030327,
0.12694492936134338,
-0.10511471331119537,
-0.09168524295091629,
0.01941707730293274,
-0.0173184871673584,
0.07617122679948807,
0.011001328006386757,
0.10739396512508392,
-0.03328050673007965,
0.06726278364658356,
-0.013346186839044094,
-0.0019135072361677885,
0.04832175374031067,
0.0002830341982189566,
0.018117375671863556,
-0.08613725751638412,
-0.003364939009770751,
-0.14503854513168335,
-0.11391928046941757,
0.013430907391011715,
-0.0041270204819738865,
0.041052620857954025,
-0.030090106651186943,
-0.025343716144561768,
0.009870430454611778,
-0.06396173685789108,
0.008846846409142017,
0.00934659969061613,
-0.029680361971259117,
0.099256232380867,
0.048019591718912125,
0.11215152591466904,
-0.09519494324922562,
0.03242276981472969,
-0.1421627700328827,
0.05968758836388588,
-0.09636510163545609,
0.06828965991735458,
-0.050707198679447174,
-0.013686919584870338,
-0.05853891745209694,
-0.0249848123639822,
-0.10386598855257034,
0.008634260855615139,
-0.008142286911606789,
0.13131672143936157,
-0.11295925080776215,
-0.07453663647174835,
0.1861482411623001,
-0.12715598940849304,
-0.15154238045215607,
0.188584566116333,
-0.0492490790784359,
0.07511419802904129,
0.08338732272386551,
0.2285599410533905,
0.02193031646311283,
-0.11407876759767532,
0.036245010793209076,
-0.034521620720624924,
-0.004931426141411066,
0.04894110932946205,
0.0696987584233284,
0.027594206854701042,
-0.05499277263879776,
0.07585965842008591,
-0.09218164533376694,
0.04294548183679581,
-0.05192537233233452,
-0.1046672984957695,
-0.05558261275291443,
-0.1113847866654396,
0.08209778368473053,
-0.043060820549726486,
0.12381788343191147,
-0.045256972312927246,
-0.03992076590657234,
-0.07498989254236221,
0.06477459520101547,
0.012280493974685669,
0.005949456710368395,
-0.15539875626564026,
0.06921307742595673,
-0.042045336216688156,
-0.01527697779238224,
-0.09236979484558105,
-0.0014133088989183307,
0.03924965485930443,
0.0031116134487092495,
0.03959512710571289,
0.07103856652975082,
-0.001514190691523254,
-0.03565216436982155,
-0.0558171309530735,
0.051967959851026535,
-0.02307497337460518,
-0.01788361929357052,
-0.04781492426991463,
-0.019967708736658096,
0.0824524313211441,
-0.023761626332998276,
0.11796523630619049,
-0.12608154118061066,
0.026503846049308777,
0.030350370332598686,
0.11845937371253967,
0.018997617065906525,
0.03520263358950615,
0.05705440044403076,
0.0787416324019432,
0.03369918838143349,
0.04560023546218872,
0.04032342880964279,
0.012306369841098785,
-0.14395660161972046,
0.09543802589178085,
-0.008855076506733894,
0.13177305459976196,
0.13109134137630463,
-0.09498357027769089,
0.03510291129350662,
-0.019703427329659462,
-0.0535796582698822,
-0.03002641163766384,
-0.08879055827856064,
0.0902794823050499,
0.18155179917812347,
0.00586409866809845,
0.13442453742027283,
-0.11845973879098892,
-0.026207629591226578,
-0.010776662267744541,
-0.008617518469691277,
-0.009033379144966602,
0.05705016851425171,
0.04754915460944176,
-0.17212574183940887,
0.036987822502851486,
0.21095480024814606,
-0.03795456513762474,
0.143288254737854,
-0.07381082326173782,
-0.06091918796300888,
0.022797975689172745,
0.09733492136001587,
-0.04849855229258537,
0.037956010550260544,
-0.12058445066213608,
0.045068662613630295,
0.06597510725259781,
0.039790280163288116,
0.051910217851400375,
-0.1406411975622177,
-0.0841236338019371,
0.008663168177008629,
-0.06085124611854553,
-0.156227245926857,
0.09724122285842896,
0.05246598273515701,
0.07250657677650452,
0.0327121801674366,
0.03767922893166542,
0.02638568915426731,
0.0034181135706603527,
-0.12206479907035828,
0.14171043038368225,
-0.01979479193687439,
-0.31973567605018616,
-0.05301983281970024,
0.09529411792755127,
-0.055073801428079605,
-0.06265915930271149,
0.04594616964459419,
-0.18548797070980072,
0.03484101966023445,
0.0005574728129431605,
0.014795361086726189,
0.017753805965185165,
0.02234623022377491,
0.07616691291332245,
0.01950986683368683,
0.02790115214884281,
-0.06627893447875977,
-0.03109339252114296,
-0.09040703624486923,
-0.07687472552061081,
0.08244894444942474,
-0.13455946743488312,
0.037059977650642395,
0.026631463319063187,
-0.052343063056468964,
0.036270588636398315,
-0.009029584005475044,
0.27541863918304443,
-0.08937236666679382,
0.012951698154211044,
0.1425587236881256,
-0.1382226049900055,
0.031114917248487473,
0.11329995840787888,
-0.007124569732695818,
-0.08432348072528839,
0.04594920203089714,
0.041338007897138596,
-0.023339400067925453,
-0.25878116488456726,
-0.025960346683859825,
-0.018996620550751686,
-0.01347443275153637,
-0.0801951140165329,
0.03662737086415291,
0.027114450931549072,
0.10499394685029984,
-0.0019254089565947652,
-0.0792451798915863,
-0.0718548595905304,
0.00654886057600379,
0.12957876920700073,
0.056117717176675797,
0.11434309184551239,
-0.047412533313035965,
-0.04137048497796059,
0.02861369587481022,
0.1084737703204155,
0.10065539181232452,
0.03980393707752228,
-0.023310648277401924,
0.10510362684726715,
0.25631898641586304,
0.19542300701141357,
0.03714650869369507,
-0.08448204398155212,
-0.07295016944408417,
0.0008509985636919737,
-0.01901201903820038,
-0.08075105398893356,
0.05210105702280998,
-0.024804340675473213,
0.002590315882116556,
-0.012421180494129658,
0.09413672983646393,
0.051052920520305634,
0.2727530896663666,
0.03070620633661747,
-0.07661200314760208,
-0.10693885385990143,
0.011560567654669285,
-0.11977285146713257,
0.001187590998597443,
0.07540040463209152,
0.06606664508581161,
-0.15002067387104034,
0.04686130955815315,
0.017017148435115814,
0.13418740034103394,
-0.05813944712281227,
0.06168919429183006,
-0.06995542347431183,
-0.18959668278694153,
0.013998341746628284,
0.0410933792591095,
-0.2963142395019531,
0.20991797745227814,
0.013450887985527515,
0.03601989150047302,
-0.05275307968258858,
-0.0480181984603405,
-0.016603443771600723,
0.06089313328266144,
0.07905038446187973,
-0.014426850713789463,
0.0006508055957965553,
-0.10448193550109863,
-0.047985486686229706,
0.11609014868736267,
0.07263720035552979,
0.0307297483086586,
0.033748771995306015,
0.024013133719563484,
0.05345909297466278,
0.005401741713285446,
0.04360348358750343,
-0.19607169926166534,
-0.040530361235141754,
0.033217694610357285,
0.03578021004796028,
0.02007213793694973,
-0.05675423517823219,
-0.04586167261004448,
-0.17265909910202026,
0.1194014698266983,
-0.12125591188669205,
-0.10105548053979874,
-0.06607739627361298,
0.03425133600831032,
0.02642533741891384,
-0.07550635933876038,
0.006112344097346067,
-0.02472393587231636,
0.021129924803972244,
-0.048270780593156815,
-0.043363768607378006,
0.05512123927474022,
-0.04243355244398117,
-0.12590958178043365,
0.02178768813610077,
0.11736694723367691,
0.027989903464913368,
0.07399840652942657,
0.08684127777814865,
-0.01891312561929226,
0.01863466016948223,
-0.07731688022613525,
-0.021218545734882355,
-0.03483981266617775,
-0.059023622423410416,
-0.005724424961954355,
-0.0896216332912445,
0.028614968061447144,
-0.14648400247097015,
-0.09555945545434952,
0.1862419992685318,
0.14917904138565063,
-0.05205561965703964,
0.09610815346240997,
0.14812906086444855,
0.023759106174111366,
-0.10670413076877594,
-0.014553889632225037,
0.028176939114928246,
0.045083943754434586,
-0.06282171607017517,
-0.060084328055381775,
0.15254093706607819,
0.11028483510017395,
-0.029182342812418938,
-0.008228513412177563,
-0.13771402835845947,
-0.10536394268274307,
0.08013562113046646,
-0.0032225081231445074,
0.2698207199573517,
-0.12010712921619415,
-0.04342171549797058,
0.031208621338009834,
-0.24128150939941406,
0.0580989308655262,
-0.08066021651029587,
0.05707031860947609,
-0.04218451678752899,
0.08004917204380035,
-0.007990825921297073,
-0.01268966868519783,
0.14632229506969452,
0.03721940889954567,
0.0003658049099612981,
-0.04197666421532631,
-0.016157690435647964,
0.007578840013593435,
0.04460960999131203,
0.07370810955762863,
-0.022231237962841988,
0.06233539432287216,
-0.19562667608261108,
-0.045763835310935974,
-0.10244742035865784,
-0.026644781231880188,
-0.040239498019218445,
-0.04016419127583504,
-0.028338788077235222,
0.03314116969704628,
0.008546906523406506,
0.03357600420713425,
0.024830717593431473,
-0.09351760894060135,
0.10233718156814575,
0.02491975575685501,
0.11228886991739273,
-0.07992538064718246,
0.040910426527261734,
-0.060340188443660736,
-0.050025053322315216,
0.11687002331018448,
-0.1875673532485962,
0.03317240998148918,
0.09684830158948898,
0.006125830579549074,
0.06255091726779938,
0.02666187472641468,
-0.009551660157740116,
0.09539449214935303,
0.051177944988012314,
-0.11091668903827667,
-0.12445065379142761,
0.014568577520549297,
-0.054898012429475784,
-0.00044072410673834383,
0.006758151575922966,
0.10075923800468445,
-0.01588866487145424,
-0.03663083538413048,
-0.0010708653135225177,
-0.031225528568029404,
-0.0762796625494957,
0.07573699206113815,
0.07252006232738495,
0.04070081561803818,
-0.10375235974788666,
0.07122351229190826,
0.0398109145462513,
-0.13759852945804596,
-0.0005435931961983442,
0.04584469646215439,
-0.17321044206619263,
-0.05296416953206062,
-0.10671980679035187,
0.06103084981441498,
-0.018618693575263023,
-0.07593514025211334,
-0.08974909037351608,
-0.14864614605903625,
0.05389104411005974,
0.16787207126617432,
0.05826585739850998,
0.016978327184915543,
0.023017289116978645,
-0.01562504656612873,
0.005454711616039276,
0.02216821350157261,
0.05903239548206329,
0.05132446065545082,
-0.10151340812444687,
-0.03191627562046051,
0.011674672365188599,
0.18397732079029083,
-0.04176994785666466,
-0.0238204225897789,
-0.06769641488790512,
0.07774517685174942,
-0.22090879082679749,
0.06979905813932419,
-0.03590122610330582,
-0.016994567587971687,
-0.0289923083037138,
-0.17664021253585815,
-0.059354059398174286,
0.010113739408552647,
-0.05746536701917648,
0.04008884355425835,
-0.03620194271206856,
-0.0111400056630373,
-0.047727398574352264,
-0.0007748480420559645,
0.14051760733127594,
-0.06595069169998169,
0.09950956702232361,
0.05964333564043045,
-0.08085672557353973,
0.12623770534992218,
-0.07133127003908157,
-0.05040488392114639,
0.03311267867684364,
0.03304349631071091,
0.022721657529473305,
0.024826617911458015,
0.04753779247403145,
0.04059547185897827,
0.05642355978488922,
0.06854487210512161,
0.21480756998062134,
-0.11187073588371277,
-0.05771850794553757,
-0.0036319270730018616,
-0.06279648095369339,
-0.04307873547077179,
-0.08058926463127136,
0.17455732822418213,
0.03269685059785843,
0.1346166878938675,
-0.05678367614746094,
0.060748208314180374,
-0.11418306827545166,
0.004021932370960712,
0.018021471798419952,
-0.08916378021240234,
0.004342753440141678,
-0.057323090732097626,
0.02286280132830143,
-0.05184945836663246,
0.17615842819213867,
-0.15037038922309875,
-0.010270455852150917,
0.04696404188871384,
-0.02051754854619503,
-0.03902424871921539,
0.020506327971816063,
0.27192240953445435,
0.15053458511829376,
-0.011173578910529613,
-0.05573581904172897,
-0.07319046556949615,
-0.033770062029361725,
0.0059178913943469524,
0.06528232991695404,
0.29661568999290466,
-0.009653138928115368,
0.0831252932548523,
0.07925058156251907,
0.11808599531650543,
-0.009784885682165623,
-0.031774770468473434,
-0.1380470246076584,
-0.018467923626303673,
0.030947979539632797,
-0.027171071618795395,
0.2718485891819,
-0.09488452970981598,
0.0218422319740057,
-0.04000731557607651,
-0.05533602461218834,
-0.12487273663282394,
-0.012654115445911884,
-0.12676464021205902,
-0.08280695974826813,
0.07238540053367615,
-0.06295689195394516,
-0.0512939877808094,
0.15865522623062134,
0.07207449525594711,
-0.021424241364002228,
0.21425561606884003,
0.08011376857757568,
-0.05023388937115669,
0.10213205963373184,
-0.05106305703520775,
-0.002883884822949767,
0.024087127298116684,
0.010742886923253536,
0.05424150079488754,
0.021989846602082253,
0.00632523512467742,
-0.0009779547108337283,
-0.056659918278455734,
-0.04546131566166878,
-0.11286120116710663,
-0.055469296872615814,
-0.008094334974884987,
0.012495160102844238,
0.02212664484977722,
0.051780443638563156,
0.03127003088593483,
0.06127306446433067,
-0.013768339529633522,
0.09228281676769257,
0.02290290966629982,
-0.17281222343444824,
-0.20026624202728271,
0.004288277123123407,
0.0005470117903314531,
0.0381028912961483,
0.028483837842941284,
-0.037359971553087234,
-0.017819635570049286,
0.362479031085968,
0.17092789709568024,
-0.06823226809501648,
0.04575801640748978,
0.05250988155603409,
0.045429181307554245,
-0.03461220860481262,
0.03886113315820694,
0.10650050640106201,
0.10244264453649521,
-0.08905567973852158,
-0.11330749094486237,
-0.06687948107719421,
-0.08176212757825851,
-0.0903545543551445,
0.04693090170621872,
0.07576346397399902,
-0.014754929579794407,
-0.0940021425485611,
0.07614202797412872,
-0.051483504474163055,
0.05495968461036682,
-0.07858002185821533,
-0.1499720960855484,
-0.1367143988609314,
-0.05900723487138748,
-0.07792842388153076,
0.02581060864031315,
0.006975416559726,
-0.00787192489951849,
-0.009087282232940197,
0.010646896436810493,
0.015641668811440468,
-0.023387962952256203,
-0.022182593122124672,
0.15200182795524597,
0.06244107708334923,
-0.03503631427884102,
0.06585492938756943,
0.15055938065052032,
0.038191474974155426,
0.07604879885911942,
0.03273149952292442,
0.12362753599882126,
0.06322918832302094,
0.06080253794789314,
-0.0908081904053688,
0.08971697837114334,
0.04993118345737457,
0.04043011739850044,
0.11748308688402176,
-0.1207563653588295,
0.06113775074481964,
-0.037966467440128326,
-0.04998144507408142,
-0.11828859895467758,
0.12855030596256256,
-0.031691063195466995,
0.10061809420585632,
0.16853441298007965,
-0.03329012170433998,
0.05353747680783272,
-0.03274111822247505,
0.07262856513261795,
0.024150948971509933,
-0.07946788519620895,
-0.07369479537010193,
-0.17624782025814056,
0.022387376055121422,
-0.02043765038251877,
-0.014464063569903374,
-0.23627908527851105,
-0.04915015771985054,
-0.03033972531557083,
0.0019466548692435026,
-0.0006079393206164241,
0.054665446281433105,
0.11854720115661621,
0.03636569529771805,
0.006296330597251654,
-0.23919245600700378,
0.043719738721847534,
0.04323435202240944,
-0.11095013469457626,
-0.07466676831245422
] |
null | null |
transformers
|
# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa
Finetuning `airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask` with the training set of `iapp_wiki_qa_squad`, `thaiqa_squad`, and `nsc_qa` (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 `newmm` words). Benchmarks shared on [wandb](https://wandb.ai/cstorm125/wangchanberta-qa) using validation and test sets of `iapp_wiki_qa_squad`.
Trained with [thai2transformers](https://github.com/vistec-AI/thai2transformers/blob/dev/scripts/downstream/train_question_answering_lm_finetuning.py).
Run with:
```
export MODEL_NAME=airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask
CUDA_LAUNCH_BLOCKING=1 python train_question_answering_lm_finetuning.py \
--model_name $MODEL_NAME \
--dataset_name chimera_qa \
--output_dir $MODEL_NAME-finetune-chimera_qa-model \
--log_dir $MODEL_NAME-finetune-chimera_qa-log \
--model_max_length 400 \
--pad_on_right \
--fp16 \
--use_auth_token
```
|
{"widget": [{"text": "\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2d\u0e30\u0e44\u0e23", "context": "\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e2a\u0e27\u0e19\u0e01\u0e38\u0e2b\u0e25\u0e32\u0e1a\u0e27\u0e34\u0e17\u0e22\u0e32\u0e25\u0e31\u0e22 (Suankularb Wittayalai School) (\u0e2d\u0e31\u0e01\u0e29\u0e23\u0e22\u0e48\u0e2d : \u0e2a.\u0e01. / S.K.) \u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e0a\u0e32\u0e22\u0e25\u0e49\u0e27\u0e19 \u0e23\u0e30\u0e14\u0e31\u0e1a\u0e0a\u0e31\u0e49\u0e19\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e19\u0e32\u0e14\u0e43\u0e2b\u0e0d\u0e48\u0e1e\u0e34\u0e40\u0e28\u0e29 \u0e2a\u0e31\u0e07\u0e01\u0e31\u0e14\u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e40\u0e02\u0e15\u0e1e\u0e37\u0e49\u0e19\u0e17\u0e35\u0e48\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e21\u0e31\u0e18\u0e22\u0e21\u0e28\u0e36\u0e01\u0e29\u0e32\u0e40\u0e02\u0e15 1 \u0e2a\u0e33\u0e19\u0e31\u0e01\u0e07\u0e32\u0e19\u0e04\u0e13\u0e30\u0e01\u0e23\u0e23\u0e21\u0e01\u0e32\u0e23\u0e01\u0e32\u0e23\u0e28\u0e36\u0e01\u0e29\u0e32\u0e02\u0e31\u0e49\u0e19\u0e1e\u0e37\u0e49\u0e19\u0e10\u0e32\u0e19 (\u0e0a\u0e37\u0e48\u0e2d\u0e40\u0e14\u0e34\u0e21: \u0e01\u0e23\u0e21\u0e2a\u0e32\u0e21\u0e31\u0e0d\u0e28\u0e36\u0e01\u0e29\u0e32) \u0e01\u0e23\u0e30\u0e17\u0e23\u0e27\u0e07\u0e28\u0e36\u0e01\u0e29\u0e32\u0e18\u0e34\u0e01\u0e32\u0e23 \u0e01\u0e48\u0e2d\u0e15\u0e31\u0e49\u0e07\u0e42\u0e14\u0e22 \u0e1e\u0e23\u0e30\u0e1a\u0e32\u0e17\u0e2a\u0e21\u0e40\u0e14\u0e47\u0e08\u0e1e\u0e23\u0e30\u0e08\u0e38\u0e25\u0e08\u0e2d\u0e21\u0e40\u0e01\u0e25\u0e49\u0e32\u0e40\u0e08\u0e49\u0e32\u0e2d\u0e22\u0e39\u0e48\u0e2b\u0e31\u0e27 \u0e44\u0e14\u0e49\u0e23\u0e31\u0e1a\u0e01\u0e32\u0e23\u0e2a\u0e16\u0e32\u0e1b\u0e19\u0e32\u0e02\u0e36\u0e49\u0e19\u0e43\u0e19\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 8 \u0e21\u0e35\u0e19\u0e32\u0e04\u0e21 \u0e1e.\u0e28. 2424 (\u0e02\u0e13\u0e30\u0e19\u0e31\u0e49\u0e19\u0e19\u0e31\u0e1a\u0e27\u0e31\u0e19\u0e17\u0e35\u0e48 1 \u0e40\u0e21\u0e29\u0e32\u0e22\u0e19 \u0e40\u0e1b\u0e47\u0e19\u0e27\u0e31\u0e19\u0e02\u0e36\u0e49\u0e19\u0e1b\u0e35\u0e43\u0e2b\u0e21\u0e48 \u0e40\u0e21\u0e37\u0e48\u0e2d\u0e19\u0e31\u0e1a\u0e2d\u0e22\u0e48\u0e32\u0e07\u0e2a\u0e32\u0e01\u0e25\u0e16\u0e37\u0e2d\u0e40\u0e1b\u0e47\u0e19 \u0e1e.\u0e28. 2425) \u0e42\u0e14\u0e22\u0e40\u0e1b\u0e47\u0e19\u0e42\u0e23\u0e07\u0e40\u0e23\u0e35\u0e22\u0e19\u0e23\u0e31\u0e10\u0e1a\u0e32\u0e25\u0e41\u0e2b\u0e48\u0e07\u0e41\u0e23\u0e01\u0e02\u0e2d\u0e07\u0e1b\u0e23\u0e30\u0e40\u0e17\u0e28\u0e44\u0e17\u0e22"}]}
|
question-answering
|
cstorm125/wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa
|
[
"transformers",
"pytorch",
"camembert",
"question-answering",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us
|
# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa
Finetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.
Trained with thai2transformers.
Run with:
|
[
"# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa\n\nFinetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
"TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n",
"# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa\n\nFinetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
31,
168
] |
[
"passage: TAGS\n#transformers #pytorch #camembert #question-answering #endpoints_compatible #region-us \n# wangchanberta-base-wiki-20210520-news-spm_span-mask-finetune-qa\n\nFinetuning 'airesearch/wangchanberta-base-wiki-20210520-news-spm_span-mask' with the training set of 'iapp_wiki_qa_squad', 'thaiqa_squad', and 'nsc_qa' (removed examples which have cosine similarity with validation and test examples over 0.8; contexts of the latter two are trimmed to be around 300 'newmm' words). Benchmarks shared on wandb using validation and test sets of 'iapp_wiki_qa_squad'.\nTrained with thai2transformers.\n\nRun with:"
] |
[
-0.05645165219902992,
-0.15578137338161469,
-0.003975315950810909,
0.06371725350618362,
0.028402362018823624,
-0.00334802339784801,
0.10726091265678406,
0.09047532081604004,
-0.009096259251236916,
0.02531369775533676,
0.06127292662858963,
-0.0026147738099098206,
0.07015769183635712,
0.09560193121433258,
-0.048190489411354065,
-0.24415038526058197,
0.08125502616167068,
0.0861651599407196,
-0.05217559263110161,
0.1618712991476059,
0.11556685715913773,
-0.07393321394920349,
0.06784527748823166,
0.0834026113152504,
-0.10574711859226227,
-0.01679716818034649,
0.03422338888049126,
-0.10367725044488907,
0.1333068460226059,
-0.004372229799628258,
0.17342953383922577,
0.0700506865978241,
-0.03057202883064747,
-0.17775164544582367,
0.05375754460692406,
-0.004029496107250452,
0.04726012051105499,
0.015807438641786575,
-0.006195432506501675,
0.06686002761125565,
-0.030359333381056786,
-0.024774329736828804,
0.02383538708090782,
0.015091247856616974,
-0.09852325916290283,
-0.066645547747612,
-0.04286876320838928,
0.025934411212801933,
0.11792894452810287,
0.07881458103656769,
-0.03761140629649162,
0.18603329360485077,
-0.19750891625881195,
0.05044373497366905,
0.13119077682495117,
-0.4157165288925171,
-0.02768372744321823,
0.02105094864964485,
0.033281054347753525,
0.028933078050613403,
-0.04602588713169098,
0.01290867943316698,
0.03625987470149994,
0.02726447768509388,
-0.026248546317219734,
-0.06289022415876389,
-0.06225086376070976,
-0.048625219613313675,
-0.1089450940489769,
0.02139902673661709,
0.11461251229047775,
0.03860693424940109,
-0.06399373710155487,
-0.10389754921197891,
-0.06502378731966019,
0.036381665617227554,
-0.009629866108298302,
-0.07338462024927139,
-0.0364442877471447,
0.020629502832889557,
-0.02692311443388462,
0.008887811563909054,
-0.10264462232589722,
-0.06737029552459717,
-0.1558917611837387,
0.11004480719566345,
0.0732748806476593,
0.05009687691926956,
-0.13868533074855804,
-0.00897529348731041,
-0.03203938528895378,
-0.14262373745441437,
-0.06516434252262115,
-0.059015028178691864,
-0.012103626504540443,
0.04680560901761055,
-0.07336738705635071,
-0.08792723715305328,
0.091684490442276,
0.0738968625664711,
-0.16564004123210907,
0.053737737238407135,
-0.09410662949085236,
0.07177819311618805,
-0.055066462606191635,
0.14851242303848267,
-0.02236366458237171,
0.10206802189350128,
0.03132730722427368,
-0.0076262252405285835,
0.021591030061244965,
-0.0034246535506099463,
0.00953826867043972,
-0.08524952083826065,
-0.004357611760497093,
0.01668895035982132,
-0.06778482347726822,
0.09788382798433304,
-0.05657027289271355,
-0.020758964121341705,
0.08207695186138153,
-0.12052488327026367,
-0.08154972642660141,
0.004460074007511139,
-0.004799318965524435,
0.03615960851311684,
0.024416524916887283,
0.08840680867433548,
-0.019984766840934753,
0.10322345048189163,
-0.021340323612093925,
0.007727453950792551,
0.03279855102300644,
-0.052327536046504974,
0.016964823007583618,
-0.09368434548377991,
-0.02776949480175972,
-0.13772724568843842,
-0.09753260761499405,
0.02911374159157276,
-0.015969393774867058,
0.04605834186077118,
-0.01214260421693325,
-0.014105312526226044,
0.0061781806871294975,
-0.06043658405542374,
-0.0018248173873871565,
0.021213388070464134,
-0.03478660807013512,
0.10106168687343597,
0.06477238237857819,
0.14710521697998047,
-0.08211812376976013,
0.02419007197022438,
-0.15100987255573273,
0.06369543820619583,
-0.13181306421756744,
0.07371138781309128,
-0.04689563810825348,
-0.017399175092577934,
-0.05628116801381111,
-0.04840604588389397,
-0.11448999494314194,
0.02137947641313076,
0.019602999091148376,
0.10521119832992554,
-0.11747425049543381,
-0.0731862410902977,
0.2813548147678375,
-0.13584426045417786,
-0.1443224549293518,
0.1942521631717682,
-0.05075673758983612,
0.06908208876848221,
0.06882669776678085,
0.23642200231552124,
0.032922569662332535,
-0.1572342813014984,
0.05905246362090111,
-0.024904360994696617,
-0.03180669620633125,
0.0466734804213047,
0.0809616893529892,
0.010784859769046307,
-0.03827410936355591,
0.07427258789539337,
-0.10888773202896118,
0.05550534278154373,
-0.05131752789020538,
-0.0932859405875206,
-0.052332282066345215,
-0.09054175764322281,
0.1292944997549057,
-0.041898056864738464,
0.12166837602853775,
-0.05191865563392639,
-0.05994275584816933,
-0.06824152171611786,
0.03887249901890755,
-0.005055443849414587,
0.03493788093328476,
-0.17229850590229034,
0.07678613066673279,
-0.03695080801844597,
-0.009906494058668613,
-0.1222047507762909,
0.010653662495315075,
0.03956134244799614,
0.019036397337913513,
0.04009195789694786,
0.03871827945113182,
0.004358310718089342,
-0.03245937451720238,
-0.06898299604654312,
0.03870321437716484,
-0.01142857689410448,
-0.02049609273672104,
-0.042482778429985046,
-0.016320331022143364,
0.06578545272350311,
-0.030722135677933693,
0.10358788073062897,
-0.054744310677051544,
0.02723276987671852,
0.013491441495716572,
0.09330595284700394,
0.007189548574388027,
0.04844487085938454,
0.054848477244377136,
0.07434631884098053,
0.03050103969871998,
0.030394168570637703,
0.048753827810287476,
0.015145153738558292,
-0.17928318679332733,
0.10594724118709564,
-0.05684894695878029,
0.12418027967214584,
0.12928356230258942,
-0.13985568284988403,
0.025113634765148163,
0.013372362591326237,
-0.06022104248404503,
-0.02567988820374012,
-0.11626686900854111,
0.07160074263811111,
0.15131038427352905,
-0.002860398031771183,
0.12099212408065796,
-0.11867613345384598,
0.0030787510331720114,
-0.007004202343523502,
-0.02378193289041519,
-0.004700901452451944,
0.03332814946770668,
0.012058869004249573,
-0.1314307153224945,
0.04624590277671814,
0.20292189717292786,
-0.038550712168216705,
0.13664980232715607,
-0.06490408629179001,
-0.054673902690410614,
0.007971933111548424,
0.09531109780073166,
-0.044260118156671524,
0.017422642558813095,
-0.10744903981685638,
0.031825121492147446,
0.05897724628448486,
0.04162351414561272,
0.04420775547623634,
-0.14425188302993774,
-0.06665658950805664,
0.0011216654675081372,
-0.0744219645857811,
-0.15799081325531006,
0.09589588642120361,
0.051240473985672,
0.07317370921373367,
0.0456082709133625,
0.0340593196451664,
0.03998574614524841,
0.0068251946941018105,
-0.12711025774478912,
0.15868569910526276,
-0.040327273309230804,
-0.3563890755176544,
-0.046893712133169174,
0.061563100665807724,
-0.045807842165231705,
-0.04441053047776222,
0.04723352566361427,
-0.16923876106739044,
0.023387938737869263,
-0.0022869673557579517,
0.025988757610321045,
0.01587662659585476,
0.04229474812746048,
0.060286957770586014,
0.01624622941017151,
0.059984732419252396,
-0.07098153978586197,
-0.03893434256315231,
-0.0876036286354065,
-0.0726843774318695,
0.10448531061410904,
-0.12638235092163086,
0.0618516206741333,
0.035028740763664246,
-0.059566840529441833,
0.025313934311270714,
-0.033398669213056564,
0.23763255774974823,
-0.0788927972316742,
0.004982548765838146,
0.1569685935974121,
-0.1251358538866043,
0.04411948844790459,
0.1573708951473236,
-0.01271434873342514,
-0.0709698274731636,
0.06387582421302795,
0.04476165398955345,
-0.04996141418814659,
-0.2243911623954773,
-0.02413898892700672,
-0.03477882966399193,
-0.011948734521865845,
-0.07173380255699158,
0.02517900988459587,
0.027421507984399796,
0.09205605834722519,
0.0001933203311637044,
-0.07960685342550278,
-0.032453037798404694,
0.019291533157229424,
0.1323135495185852,
0.0453004315495491,
0.13742691278457642,
-0.03815677389502525,
-0.05465081334114075,
0.02154124341905117,
0.06903959065675735,
0.1262403279542923,
0.030950428918004036,
-0.02074783481657505,
0.1303718090057373,
0.2521170675754547,
0.1931368112564087,
0.04449344053864479,
-0.06666159629821777,
-0.045453865081071854,
-0.002646553795784712,
-0.03264681249856949,
-0.07596889138221741,
0.025024618953466415,
-0.009739906527101994,
0.005826374515891075,
-0.02140367403626442,
0.0904291495680809,
0.0687747374176979,
0.23931416869163513,
0.04921963810920715,
-0.10510202497243881,
-0.10872305184602737,
0.010597066022455692,
-0.07664304971694946,
0.013129840604960918,
0.06450960040092468,
0.053273558616638184,
-0.1424940824508667,
0.09238754212856293,
-0.006918950472027063,
0.12409234791994095,
-0.018174726516008377,
0.0694262683391571,
-0.08389951288700104,
-0.14666932821273804,
0.013733753003180027,
0.03447822108864784,
-0.33849450945854187,
0.25234460830688477,
0.00713298050686717,
0.034632839262485504,
-0.050640009343624115,
-0.03494016081094742,
0.0028525409288704395,
0.036094311624765396,
0.10685926675796509,
-0.01118082832545042,
-0.013135116547346115,
-0.09911953657865524,
-0.03177577629685402,
0.10172165185213089,
0.08495207130908966,
0.0704447478055954,
0.028829388320446014,
0.02387024648487568,
0.03336026892066002,
0.02035253494977951,
0.08619460463523865,
-0.21086567640304565,
-0.014499422162771225,
0.03288273140788078,
0.033718809485435486,
-0.010357825085520744,
-0.06234145909547806,
-0.058893557637929916,
-0.15615105628967285,
0.14148657023906708,
-0.11031268537044525,
-0.11262644827365875,
-0.07787282764911652,
0.0885225236415863,
0.03802981972694397,
-0.09562625735998154,
0.00989309512078762,
-0.03763686865568161,
0.04117213562130928,
-0.04490004852414131,
-0.052452750504016876,
0.0821690782904625,
-0.02067621424794197,
-0.1188993901014328,
0.030677499249577522,
0.09560912847518921,
0.016302907839417458,
0.07696668803691864,
0.08186209946870804,
-0.003333958564326167,
0.03153949975967407,
-0.08143407851457596,
-0.02790510654449463,
-0.04256097972393036,
-0.06286454200744629,
0.011274293065071106,
-0.09618093073368073,
-0.026925988495349884,
-0.12939485907554626,
-0.058172211050987244,
0.19188842177391052,
0.1516365110874176,
-0.06966100633144379,
0.11370330303907394,
0.1601092368364334,
0.012129274196922779,
-0.16243842244148254,
-0.017171340063214302,
0.015334893018007278,
0.05995607003569603,
-0.030482754111289978,
-0.050392504781484604,
0.12976886332035065,
0.08510090410709381,
-0.029459398239850998,
-0.04841260612010956,
-0.14722906053066254,
-0.1233307421207428,
0.09791356325149536,
0.00761268800124526,
0.26061221957206726,
-0.11959756910800934,
-0.05337437987327576,
0.0010068112751469016,
-0.21429729461669922,
0.04421420767903328,
-0.08569993823766708,
0.08155646175146103,
-0.03101450577378273,
0.05889633670449257,
-0.00683694938197732,
-0.020256081596016884,
0.12315419316291809,
0.012659359723329544,
-0.007534236181527376,
-0.05974123254418373,
-0.00598837761208415,
0.07186516374349594,
0.03798792138695717,
0.07320533692836761,
0.003622655989602208,
0.042057473212480545,
-0.14047503471374512,
-0.03874216228723526,
-0.09903033077716827,
-0.022127274423837662,
-0.03819279372692108,
-0.05495528504252434,
-0.038980767130851746,
0.04265791177749634,
0.04265185818076134,
0.03039495274424553,
0.0528106726706028,
-0.08112549781799316,
0.10376928001642227,
-0.0020388534758239985,
0.09085554629564285,
-0.09089762717485428,
0.03489893302321434,
-0.04310543090105057,
-0.03939225152134895,
0.10782921314239502,
-0.1849430650472641,
0.04541826993227005,
0.10224716365337372,
0.0053063202649354935,
0.07715350389480591,
0.05674015358090401,
-0.02221371978521347,
0.10788600891828537,
0.06146606057882309,
-0.09638608992099762,
-0.11751127988100052,
0.04755084216594696,
-0.13107463717460632,
-0.02532198093831539,
0.000405985425459221,
0.10311058163642883,
-0.013571041636168957,
-0.03571176156401634,
-0.00037755814264528453,
0.000621594546828419,
-0.09030323475599289,
0.08523041009902954,
0.07044848799705505,
0.03537596017122269,
-0.0855194553732872,
0.07624754309654236,
0.04448126628994942,
-0.11791151016950607,
-0.01402331329882145,
0.044152386486530304,
-0.17679978907108307,
-0.040157996118068695,
-0.07682318985462189,
0.025651317089796066,
0.004485803190618753,
-0.07336962968111038,
-0.10682182013988495,
-0.15395870804786682,
0.03279250115156174,
0.15345053374767303,
0.06887253373861313,
0.009461878798902035,
0.019313979893922806,
0.0020494575146585703,
0.008692781440913677,
0.02367996796965599,
0.07320357859134674,
0.05315642058849335,
-0.07568085193634033,
0.029136957600712776,
0.03456750512123108,
0.1862347275018692,
-0.05804112181067467,
-0.020328717306256294,
-0.09454434365034103,
0.0799933448433876,
-0.20637862384319305,
0.06910024583339691,
-0.030562210828065872,
-0.04060971364378929,
-0.045557886362075806,
-0.15620961785316467,
-0.054748643189668655,
0.025947153568267822,
-0.038972653448581696,
0.028549320995807648,
-0.034839265048503876,
-0.02635052800178528,
-0.0800986960530281,
-0.014757287688553333,
0.12921614944934845,
-0.05508222430944443,
0.09601178765296936,
0.058791376650333405,
-0.09166315197944641,
0.09740674495697021,
-0.13584481179714203,
-0.0753355473279953,
0.055561210960149765,
0.021988825872540474,
0.018057290464639664,
0.009738794527947903,
0.04446424916386604,
0.04967714846134186,
0.05413934960961342,
0.06503071635961533,
0.2236274778842926,
-0.1047842726111412,
-0.029805220663547516,
-0.04008534923195839,
-0.05438491702079773,
-0.04365277290344238,
-0.06317665427923203,
0.14987410604953766,
0.0645315945148468,
0.11730875074863434,
-0.08226016908884048,
0.06199562922120094,
-0.11232834309339523,
-0.007497240789234638,
0.0049714744091033936,
-0.08312064409255981,
0.006797037087380886,
-0.05524333566427231,
0.03332368656992912,
-0.0449766144156456,
0.17435061931610107,
-0.10551266372203827,
0.008485675789415836,
0.04954545199871063,
0.005280391313135624,
-0.028978366404771805,
0.013038692064583302,
0.2741159200668335,
0.09899080544710159,
-0.020456206053495407,
0.0067034270614385605,
-0.05657985433936119,
-0.03087596781551838,
0.018583107739686966,
0.10116090625524521,
0.2798546552658081,
0.012569633312523365,
0.07931564003229141,
0.07070605456829071,
0.14550256729125977,
-0.05268070474267006,
-0.06165361776947975,
-0.11771334707736969,
-0.017518222332000732,
0.0337199866771698,
-0.03332908824086189,
0.2294331192970276,
-0.08726039528846741,
0.04942779988050461,
-0.03250725567340851,
-0.07401353865861893,
-0.14127188920974731,
-0.020210707560181618,
-0.13061438500881195,
-0.05716592073440552,
0.06348954141139984,
-0.06945988535881042,
-0.06099657714366913,
0.15722978115081787,
0.06486516445875168,
-0.018841980025172234,
0.23413832485675812,
0.03509972244501114,
-0.05381190404295921,
0.09144502133131027,
-0.04897899925708771,
0.007790836971253157,
0.06953942030668259,
0.003064451040700078,
0.04981398954987526,
0.06989816576242447,
0.014387210831046104,
-0.00792248360812664,
-0.04984894022345543,
-0.027987414970993996,
-0.12140931934118271,
-0.059778448194265366,
-0.016094928607344627,
0.01943817548453808,
0.027810780331492424,
0.06886626034975052,
0.04837511479854584,
0.03828977420926094,
-0.02126351185142994,
0.10680917650461197,
0.015447737649083138,
-0.18841877579689026,
-0.188200443983078,
0.029639899730682373,
0.0005057085654698312,
0.025602994486689568,
0.023490356281399727,
-0.04816146939992905,
-0.029371798038482666,
0.34897422790527344,
0.2282310277223587,
-0.06503435969352722,
0.04994513466954231,
0.05080592259764671,
0.03726101666688919,
-0.0010308146011084318,
0.06878799945116043,
0.10701823979616165,
0.1320493370294571,
-0.0748148187994957,
-0.07790536433458328,
-0.06437265127897263,
-0.09954023361206055,
-0.1299128234386444,
0.015754355117678642,
0.05160360410809517,
0.001727293711155653,
-0.0874411016702652,
0.08896085619926453,
-0.05334225669503212,
0.017634548246860504,
-0.046170689165592194,
-0.15938560664653778,
-0.12217520922422409,
-0.035359252244234085,
-0.06232749670743942,
0.03437512740492821,
0.0079569760710001,
-0.01440426055341959,
0.03553403913974762,
0.004970451816916466,
0.03173965588212013,
-0.008316252380609512,
-0.003199238795787096,
0.11449633538722992,
0.03580276295542717,
0.010280794464051723,
0.06237341836094856,
0.13421103358268738,
0.0513298325240612,
0.08147658407688141,
0.026739811524748802,
0.14336702227592468,
0.05537400022149086,
0.06243152171373367,
-0.07312624156475067,
0.07120147347450256,
0.02347569167613983,
-0.006148068234324455,
0.1134309321641922,
-0.14451684057712555,
0.06678295880556107,
-0.04926222562789917,
-0.041110560297966,
-0.12792548537254333,
0.11433670669794083,
-0.05923744663596153,
0.08394601941108704,
0.16661304235458374,
-0.03404637426137924,
0.031223097816109657,
-0.040925946086645126,
0.02898281067609787,
0.008738464675843716,
-0.06407328695058823,
-0.085347980260849,
-0.17400729656219482,
0.01590942218899727,
0.020673677325248718,
-0.008004342205822468,
-0.2563948631286621,
-0.056728608906269073,
-0.04525475949048996,
0.0005499205435626209,
-0.014929153956472874,
0.05711062252521515,
0.09244736284017563,
0.02335730753839016,
0.0036072642542421818,
-0.2897733449935913,
0.045558493584394455,
0.06405538320541382,
-0.13738404214382172,
-0.0778832659125328
] |
null | null |
k2
|
# Introduction
This repo contains pre-trained model using
<https://github.com/k2-fsa/icefall/pull/219>.
It is trained on [AIShell](https://www.openslr.org/33/) dataset
using modified transducer from [optimized_transducer](https://github.com/csukuangfj/optimized_transducer).
Also, it uses [aidatatang_200zh](http://www.openslr.org/62/) as extra training data.
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01
cd icefall-aishell-transducer-stateless-modified-2-2022-03-01
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `TODO`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout TODO
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/TODO/egs/aishell/ASR/transducer_stateless_modified-2/train.py#L232>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the AIShell dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```bash
cd egs/aishell/ASR
./prepare.sh --stop-stage 6
./prepare_aidatatang_200zh.sh
export CUDA_VISIBLE_DEVICES="0,1,2"
./transducer_stateless_modified-2/train.py \
--world-size 3 \
--num-epochs 90 \
--start-epoch 0 \
--exp-dir transducer_stateless_modified-2/exp-2 \
--max-duration 250 \
--lr-factor 2.0 \
--context-size 2 \
--modified-transducer-prob 0.25 \
--datatang-prob 0.2
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/oG72ZlWaSGua6fXkcGRRjA/>
The commands for decoding are
```bash
# greedy search
for epoch in 89; do
for avg in 38; do
./transducer_stateless_modified-2/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_modified-2/exp-2 \
--max-duration 100 \
--context-size 2 \
--decoding-method greedy_search \
--max-sym-per-frame 1
done
done
# modified beam search
for epoch in 89; do
for avg in 38; do
./transducer_stateless_modified-2/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_modified-2/exp-2 \
--max-duration 100 \
--context-size 2 \
--decoding-method modified_beam_search \
--beam-size 4
done
done
```
You can find the decoding log for the above command in this
repo (in the folder [log][log]).
The WER for the test dataset is
| | test |comment |
|------------------------|------|----------------------------------------------------------------|
| greedy search | 4.94 |--epoch 89, --avg 38, --max-duration 100, --max-sym-per-frame 1 |
| modified beam search | 4.68 |--epoch 89, --avg 38, --max-duration 100 --beam-size 4 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```bash
epoch=89
avg=38
./transducer_stateless_modified-2/export.py \
--exp-dir ./transducer_stateless_modified-2/exp-2 \
--lang-dir ./data/lang_char \
--epoch $epoch \
--avg $avg
```
**HINT**: To use `pretrained.pt` to compute the WER for the `test` dataset,
just do the following:
```bash
cp icefall-aishell-transducer-stateless-modified-2-2022-03-01/exp/pretrained.pt \
/path/to/icefall/egs/aishell/ASR/transducer_stateless_modified-2/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless_modified-2/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/aishell/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{"language": "en", "license": "apache-2.0", "tags": ["icefall", "k2", "transducer", "aishell", "ASR", "stateless transducer", "PyTorch"], "datasets": ["aishell", "aidatatang_200zh"], "metrics": ["WER"]}
| null |
csukuangfj/icefall-aishell-transducer-stateless-modified-2-2022-03-01
|
[
"k2",
"icefall",
"transducer",
"aishell",
"ASR",
"stateless transducer",
"PyTorch",
"en",
"dataset:aishell",
"dataset:aidatatang_200zh",
"license:apache-2.0",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #dataset-aidatatang_200zh #license-apache-2.0 #region-us
|
Introduction
============
This repo contains pre-trained model using
<URL
It is trained on AIShell dataset
using modified transducer from optimized\_transducer.
Also, it uses aidatatang\_200zh as extra training data.
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit 'TODO'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the AIShell dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The commands for decoding are
You can find the decoding log for the above command in this
repo (in the folder [log](URL)).
The WER for the test dataset is
test: greedy search, comment: 4.94
test: modified beam search, comment: 4.68
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for the 'test' dataset,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_modified-2/URL'.
|
[] |
[
"TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #dataset-aidatatang_200zh #license-apache-2.0 #region-us \n"
] |
[
59
] |
[
"passage: TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #dataset-aidatatang_200zh #license-apache-2.0 #region-us \n"
] |
[
-0.1148015484213829,
0.10091722756624222,
-0.005711867939680815,
0.07724861800670624,
0.011987430043518543,
-0.029239220544695854,
0.20202097296714783,
0.05115976929664612,
0.18732497096061707,
-0.08075223863124847,
0.19758202135562897,
0.11648067086935043,
0.03819235414266586,
0.15277323126792908,
0.013063925318419933,
-0.23367811739444733,
0.029090605676174164,
-0.05850335583090782,
-0.06745374202728271,
0.08112197369337082,
0.07438448816537857,
-0.02390199527144432,
0.022162649780511856,
0.0009555229917168617,
0.05282209813594818,
-0.000254948710789904,
0.02319534309208393,
-0.0818755030632019,
0.106153205037117,
-0.051216769963502884,
0.029070869088172913,
0.05592772364616394,
0.026166610419750214,
-0.10673113912343979,
0.023400070145726204,
-0.0324745699763298,
-0.050855886191129684,
0.06551963835954666,
0.018584974110126495,
0.025359569117426872,
-0.034253790974617004,
-0.07581587135791779,
-0.05255710333585739,
0.03154519572854042,
-0.053218912333250046,
-0.1781885325908661,
-0.08833780884742737,
0.0696774423122406,
0.029151100665330887,
0.06046759709715843,
0.044264234602451324,
0.1484770029783249,
-0.10634864866733551,
0.08207681775093079,
0.09348854422569275,
-0.38248297572135925,
0.03998624160885811,
-0.0035798142198473215,
-0.08166542649269104,
0.07526382803916931,
0.024904796853661537,
0.018727976828813553,
0.005931975319981575,
-0.02555565908551216,
0.017922701314091682,
-0.004578035324811935,
-0.13343076407909393,
0.06466837227344513,
-0.09468796104192734,
-0.020967308431863785,
0.24798037111759186,
0.001363372546620667,
0.03105809912085533,
0.07612738758325577,
-0.05686473473906517,
-0.06683357805013657,
0.06773005425930023,
0.06384042650461197,
0.007082329131662846,
0.08480021357536316,
0.024228207767009735,
-0.0003360282862558961,
-0.1182275041937828,
-0.01695646531879902,
-0.16147975623607635,
0.11133717000484467,
-0.005225215572863817,
0.12750962376594543,
-0.15582223236560822,
0.023647334426641464,
0.027938229963183403,
-0.1396714746952057,
0.04334425553679466,
-0.06869920343160629,
0.03529244288802147,
0.10652870684862137,
-0.017306316643953323,
0.023390330374240875,
0.12033690512180328,
0.17717312276363373,
0.07021744549274445,
-0.0026351886335760355,
0.03152914345264435,
0.1294643133878708,
0.07635761797428131,
0.031011853367090225,
0.027817318215966225,
-0.025541765615344048,
0.033667512238025665,
-0.016377722844481468,
0.10481766611337662,
-0.06630479544401169,
-0.048740487545728683,
-0.027710415422916412,
0.007423692848533392,
0.08481381088495255,
0.10501531511545181,
-0.06651268899440765,
-0.09562108665704727,
0.014560188166797161,
0.15716271102428436,
-0.06698399037122726,
-0.0008345293463207781,
0.026797259226441383,
-0.010839767754077911,
0.06782073527574539,
-0.013921457342803478,
0.07382947951555252,
0.03167569637298584,
0.04981767758727074,
-0.07938603311777115,
-0.014168706722557545,
0.030012192204594612,
0.03652556613087654,
0.08975958079099655,
-0.14743782579898834,
0.04161372035741806,
-0.13468100130558014,
-0.18266251683235168,
0.0567227303981781,
0.04342533275485039,
0.0025394605472683907,
-0.06508456915616989,
0.04062739387154579,
-0.01978158950805664,
0.03579029440879822,
-0.07235369831323624,
-0.10035927593708038,
-0.05030660703778267,
0.008216163143515587,
-0.12174719572067261,
0.06130810081958771,
-0.13179577887058258,
-0.007826377637684345,
-0.16415561735630035,
0.00982065312564373,
0.07549134641885757,
-0.012398339807987213,
-0.1588166356086731,
0.09383723139762878,
-0.0494166761636734,
-0.0022445477079600096,
-0.05397187918424606,
-0.06588722765445709,
0.05580703541636467,
0.1483178287744522,
-0.2645792067050934,
-0.06368101388216019,
0.07377907633781433,
-0.13664370775222778,
-0.09225593507289886,
0.09673742204904556,
0.05494162440299988,
-0.06880464404821396,
0.038225430995225906,
0.18308226764202118,
-0.04957277327775955,
0.00995603110641241,
-0.09152165055274963,
0.13488750159740448,
-0.14360177516937256,
-0.17740345001220703,
0.12049251794815063,
-0.028780797496438026,
0.027948463335633278,
0.023580100387334824,
0.0673627257347107,
0.11822225898504257,
-0.023646125569939613,
-0.15073691308498383,
-0.057993389666080475,
-0.07341615110635757,
0.052298784255981445,
-0.012518909759819508,
0.017202099785208702,
-0.012901557609438896,
0.005458363331854343,
-0.06565352529287338,
0.03329962491989136,
0.09046793729066849,
0.047037381678819656,
-0.051521461457014084,
-0.0030687961261719465,
0.01338664535433054,
-0.0037942028138786554,
-0.06672512739896774,
-0.066722072660923,
-0.017620760947465897,
-0.04774579405784607,
-0.017878739163279533,
0.06663636863231659,
0.07280325144529343,
-0.029358793050050735,
0.012992985546588898,
-0.00006046418275218457,
0.027517156675457954,
0.06472375243902206,
-0.04667041078209877,
-0.20747056603431702,
0.030041834339499474,
-0.03561898693442345,
0.04022201523184776,
0.021193675696849823,
0.02493033930659294,
0.13669008016586304,
0.05899077653884888,
-0.014099531807005405,
0.051306117326021194,
-0.03099757991731167,
-0.021341243758797646,
0.018016474321484566,
0.0062652211636304855,
0.09037521481513977,
0.015094535425305367,
-0.16625846922397614,
0.1532743275165558,
0.010977949015796185,
0.08053315430879593,
0.11544673889875412,
0.002508155070245266,
0.13293395936489105,
0.013532248325645924,
-0.06271301209926605,
-0.08457291126251221,
0.07565964013338089,
0.059033505618572235,
0.027567381039261818,
0.058439452201128006,
0.02992931753396988,
0.032459720969200134,
-0.008283737115561962,
-0.008900943212211132,
-0.07277532666921616,
-0.04095272719860077,
0.06842021644115448,
0.031034037470817566,
-0.18014082312583923,
0.10507648438215256,
0.3310094177722931,
-0.009461190551519394,
0.14952638745307922,
-0.09993211925029755,
-0.06344986706972122,
-0.04271498695015907,
-0.09525483846664429,
-0.044503483921289444,
0.16427797079086304,
0.023340653628110886,
0.039982568472623825,
0.08674288541078568,
0.009699800983071327,
-0.0018702357774600387,
-0.095360167324543,
-0.06608012318611145,
-0.021184531971812248,
0.03721543028950691,
-0.20433136820793152,
0.06325513124465942,
-0.09995322674512863,
0.03270009532570839,
-0.021619463339447975,
-0.119318388402462,
0.1257971227169037,
-0.011602917686104774,
-0.026071697473526,
0.12296675145626068,
-0.1650463044643402,
-0.13475733995437622,
-0.06287362426519394,
0.12879055738449097,
-0.020791806280612946,
-0.08198481053113937,
0.056459713727235794,
-0.12323685735464096,
-0.02755378745496273,
0.009753264486789703,
-0.06083650887012482,
0.029620200395584106,
0.02392806112766266,
0.08798269182443619,
-0.02565748430788517,
-0.01103246957063675,
-0.10330172628164291,
-0.03757186233997345,
-0.028709353879094124,
-0.002745653036981821,
0.1473575234413147,
0.02740565501153469,
0.02039276249706745,
0.08389554172754288,
0.07063369452953339,
0.00495782308280468,
0.011808724142611027,
0.09281165152788162,
-0.02293284423649311,
0.03077532723546028,
0.13363592326641083,
-0.03932245820760727,
0.08806796371936798,
0.07557173073291779,
0.0631294772028923,
0.010427074506878853,
-0.025052839890122414,
-0.02001882530748844,
-0.10672782361507416,
-0.22966593503952026,
-0.07061851769685745,
-0.05051672086119652,
0.18572653830051422,
0.00008583541784901172,
0.06081751361489296,
0.009941480122506618,
0.09022683650255203,
0.07332286983728409,
-0.08076893538236618,
-0.06572859734296799,
-0.009571438655257225,
0.08542108535766602,
-0.0028580473735928535,
0.08933636546134949,
-0.067864328622818,
0.036803800612688065,
0.13075418770313263,
0.10097254067659378,
0.17850039899349213,
0.07061534374952316,
0.054705314338207245,
0.0720491111278534,
0.23296688497066498,
0.044114693999290466,
0.11057999730110168,
0.06678697466850281,
-0.013932537287473679,
-0.0076210591942071915,
-0.02170269377529621,
-0.057605963200330734,
0.05974460020661354,
-0.05721434950828552,
-0.13072127103805542,
0.09600239992141724,
0.13247254490852356,
0.082707479596138,
0.2932901680469513,
0.0664379894733429,
-0.08956611156463623,
-0.03453537076711655,
0.04506199061870575,
0.01009838841855526,
0.034305404871702194,
0.09125459939241409,
-0.022582322359085083,
-0.06447640061378479,
0.07140086591243744,
-0.06434786319732666,
0.02617541514337063,
0.009931505657732487,
0.03295358642935753,
-0.0072774216532707214,
-0.08536293357610703,
0.05464601516723633,
0.07279948145151138,
-0.21003960072994232,
0.22071564197540283,
-0.012300503440201283,
-0.02240607887506485,
-0.06691130995750427,
-0.03293239325284958,
0.07363315671682358,
0.08371198922395706,
0.0707201361656189,
0.09919799119234085,
-0.0989479348063469,
-0.0015371923800557852,
-0.20447106659412384,
0.08307737112045288,
-0.0029684058390557766,
0.012418256141245365,
-0.10903327167034149,
-0.02186156064271927,
0.04366733878850937,
0.041408970952034,
-0.029706353321671486,
-0.18491800129413605,
-0.04468965902924538,
0.08639267832040787,
0.14922116696834564,
0.03894747048616409,
-0.05360452085733414,
-0.13112182915210724,
-0.0036891307681798935,
0.041440997272729874,
-0.12044662982225418,
-0.09576204419136047,
-0.07608950883150101,
-0.05073711648583412,
0.12482964247465134,
-0.08441577106714249,
0.02391030266880989,
-0.00690739369019866,
-0.06632465124130249,
-0.022457744926214218,
-0.16547898948192596,
0.09471742808818817,
-0.1201910600066185,
-0.14242209494113922,
0.02904740534722805,
0.03980318084359169,
0.07152464985847473,
0.04241269826889038,
-0.05331417918205261,
0.08339832723140717,
-0.09783415496349335,
-0.0906013697385788,
0.10238237679004669,
-0.018963336944580078,
-0.013795220293104649,
0.03320249170064926,
0.055919039994478226,
-0.05139300227165222,
0.008331102319061756,
-0.08144025504589081,
0.19805340468883514,
0.2969817519187927,
-0.09824652969837189,
0.2299693524837494,
0.24374297261238098,
-0.011523820459842682,
-0.23339492082595825,
-0.1753263920545578,
-0.1291065663099289,
-0.05536515265703201,
0.12574587762355804,
-0.1965564638376236,
0.11102968454360962,
0.1354658454656601,
-0.1131785586476326,
0.09919629991054535,
-0.23892073333263397,
-0.06148752570152283,
0.19973769783973694,
-0.09727280586957932,
0.3090699017047882,
-0.10259290784597397,
-0.04503165930509567,
-0.03563323989510536,
-0.09421144425868988,
0.12061432749032974,
-0.12152884900569916,
0.029258938506245613,
-0.05064014345407486,
0.038011323660612106,
-0.022474752739071846,
-0.0666390061378479,
0.12864261865615845,
0.034910403192043304,
-0.015314378775656223,
-0.04856010898947716,
-0.049115560948848724,
0.15301190316677094,
-0.017435722053050995,
0.04057178646326065,
-0.15213564038276672,
0.07323353737592697,
-0.04989440739154816,
0.04412394016981125,
-0.10939613729715347,
0.08143379539251328,
-0.00045677594607695937,
-0.08974894136190414,
0.017379555851221085,
0.015520214103162289,
0.0069947452284395695,
-0.017514511942863464,
0.21325655281543732,
0.10157203674316406,
-0.09806722402572632,
0.0690135657787323,
0.036232803016901016,
-0.09059890359640121,
0.007226115558296442,
-0.12738728523254395,
-0.054685965180397034,
0.0643787607550621,
-0.10978461056947708,
0.0510978065431118,
0.11754311621189117,
0.021787559613585472,
-0.04200707748532295,
0.0677449181675911,
-0.047744929790496826,
0.007186018396168947,
0.10690263658761978,
-0.16860921680927277,
-0.030130213126540184,
0.024418221786618233,
0.14354895055294037,
0.07778661698102951,
0.16595713794231415,
0.16886548697948456,
-0.0014274085406213999,
-0.01944117806851864,
-0.00520746735855937,
0.11769451946020126,
-0.0671636238694191,
0.007200663909316063,
0.10181494802236557,
0.006079449318349361,
-0.10545367002487183,
0.16792921721935272,
-0.05657822638750076,
-0.007519190199673176,
0.05405334383249283,
0.05909876525402069,
-0.09745743125677109,
-0.10224730521440506,
-0.06640386581420898,
-0.052639931440353394,
-0.1781580001115799,
-0.10955797135829926,
-0.04319985955953598,
-0.042737022042274475,
-0.005648382939398289,
0.0017959390534088016,
0.05505835637450218,
0.07234556972980499,
-0.010789436288177967,
-0.06746252626180649,
0.06552114337682724,
0.048114702105522156,
-0.07925665378570557,
-0.021262967959046364,
-0.129445418715477,
-0.0810498297214508,
-0.029748426750302315,
0.10815152525901794,
-0.0420716293156147,
-0.01422110479325056,
-0.03993930295109749,
0.012782135978341103,
-0.0544874481856823,
-0.08381707966327667,
-0.09103242307901382,
-0.008020336739718914,
0.007274485658854246,
-0.08861040323972702,
-0.02808341756463051,
0.06212720274925232,
-0.09209234267473221,
0.008680742233991623,
-0.00958373211324215,
0.03371232748031616,
-0.1680988371372223,
-0.06492803990840912,
0.08960868418216705,
0.025081193074584007,
0.13699910044670105,
0.1099754124879837,
-0.006110748741775751,
0.07390075922012329,
-0.024433569982647896,
0.010918797925114632,
0.042463190853595734,
0.03341212123632431,
0.032477930188179016,
0.04447389394044876,
-0.08023177087306976,
0.03975401073694229,
0.0018375942017883062,
0.05786708742380142,
0.12794379889965057,
-0.07527396827936172,
-0.019702745601534843,
0.0350460410118103,
-0.10218365490436554,
-0.07671384513378143,
-0.10002622753381729,
0.07336261868476868,
0.06702952831983566,
0.1783405840396881,
-0.009416835382580757,
0.011883482336997986,
-0.01511877216398716,
0.026678331196308136,
0.036386363208293915,
-0.07881330698728561,
-0.043482810258865356,
0.01830647885799408,
-0.023587306961417198,
-0.062283869832754135,
0.16358621418476105,
-0.0335439071059227,
-0.1494579017162323,
0.0484217070043087,
-0.006114899646490812,
-0.06856246292591095,
0.01217444147914648,
0.15546225011348724,
0.01707691140472889,
-0.0010288683697581291,
-0.09997974336147308,
0.026580221951007843,
-0.06798923015594482,
-0.08481443673372269,
0.13246016204357147,
0.0024109340738505125,
0.07813962548971176,
0.039905447512865067,
0.06216132268309593,
-0.09243014454841614,
-0.13563430309295654,
-0.12757523357868195,
-0.0933513194322586,
0.07063347846269608,
0.06336767226457596,
0.1578318178653717,
0.1248278096318245,
0.018958615139126778,
-0.019824815914034843,
-0.06728646159172058,
-0.04965100809931755,
-0.12956605851650238,
-0.14456047117710114,
0.00004444519800017588,
-0.09188754856586456,
0.03445559740066528,
-0.021311776712536812,
0.036771468818187714,
0.1971183717250824,
0.03747646510601044,
-0.010775414295494556,
-0.016659341752529144,
-0.012138805352151394,
0.05873779207468033,
0.007611739914864302,
-0.020882828161120415,
-0.07346170395612717,
-0.03684861212968826,
-0.02391767129302025,
-0.04635409638285637,
0.012399713508784771,
-0.0715688094496727,
-0.03235851600766182,
-0.07524725049734116,
0.021394232288002968,
-0.0269028190523386,
-0.07859896123409271,
-0.06683474034070969,
-0.006011278834193945,
0.020925350487232208,
0.07303860038518906,
0.06478817015886307,
0.07325267791748047,
0.04698638617992401,
0.12860339879989624,
-0.023450205102562904,
-0.10800128430128098,
-0.052480410784482956,
0.044411856681108475,
-0.0652337446808815,
0.03148108348250389,
-0.018023915588855743,
-0.05711107328534126,
-0.04742569848895073,
0.029931195080280304,
0.19975987076759338,
-0.06587405502796173,
0.016222313046455383,
-0.06352584809064865,
0.00852664653211832,
-0.04027804732322693,
0.06103311851620674,
0.08890693634748459,
0.1376587599515915,
-0.12213917821645737,
-0.006943514570593834,
-0.07608677446842194,
-0.020005587488412857,
-0.13620126247406006,
0.0064850496128201485,
-0.05119314789772034,
-0.05175105482339859,
-0.0734827071428299,
0.11291849613189697,
-0.0721370056271553,
0.07872288674116135,
-0.043865978717803955,
-0.09845808148384094,
-0.03448650613427162,
-0.024293003603816032,
0.1709374338388443,
0.035199593752622604,
0.08750098198652267,
-0.059315990656614304,
-0.010157641023397446,
0.01288573071360588,
0.04547644406557083,
-0.182626411318779,
0.020508253946900368,
0.07102642208337784,
-0.10771642625331879,
0.1531551629304886,
-0.016674213111400604,
0.14644810557365417,
0.10287383943796158,
0.03800848498940468,
-0.1295185387134552,
0.023571547120809555,
0.02512074075639248,
-0.019764738157391548,
-0.11407361179590225,
-0.07629604637622833,
0.028142137452960014,
-0.08464132249355316,
0.06089257448911667,
0.05766482651233673,
-0.005139557644724846,
0.15443135797977448,
0.03270568698644638,
0.00476139597594738,
0.07068683952093124,
-0.022975077852606773,
0.07021715492010117,
-0.027968736365437508,
-0.04925323650240898,
-0.06437640637159348,
-0.04676100239157677,
-0.03053719736635685,
0.01083003357052803,
-0.19969512522220612,
-0.057057905942201614,
0.08547097444534302,
0.022587521001696587,
0.05868929997086525,
0.0220847949385643,
0.05767853930592537,
-0.058477647602558136,
-0.167682945728302,
-0.022224415093660355,
-0.09989747405052185,
-0.010763054713606834,
0.023328151553869247,
-0.02353227697312832,
0.028539419174194336,
-0.02381320856511593,
0.018801046535372734,
-0.026001261547207832,
-0.13209518790245056,
-0.07569220662117004
] |
null | null |
k2
|
# Introduction
This repo contains pre-trained model using
<https://github.com/k2-fsa/icefall/pull/219>.
It is trained on [AIShell](https://www.openslr.org/33/) dataset
using modified transducer from [optimized_transducer](https://github.com/csukuangfj/optimized_transducer).
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01
cd icefall-aishell-transducer-stateless-modified-2022-03-01
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `TODO`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout TODO
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/TODO/egs/aishell/ASR/transducer_stateless_modified/train.py#L232>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the AIShell dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```bash
cd egs/aishell/ASR
./prepare.sh --stop-stage 6
export CUDA_VISIBLE_DEVICES="0,1,2"
./transducer_stateless_modified/train.py \
--world-size 3 \
--num-epochs 90 \
--start-epoch 0 \
--exp-dir transducer_stateless_modified/exp-4 \
--max-duration 250 \
--lr-factor 2.0 \
--context-size 2 \
--modified-transducer-prob 0.25
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/C27M8YxRQCa1t2XglTqlWg>
The commands for decoding are
```bash
# greedy search
for epoch in 64; do
for avg in 33; do
./transducer_stateless_modified-2/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_modified/exp-4 \
--max-duration 100 \
--context-size 2 \
--decoding-method greedy_search \
--max-sym-per-frame 1
done
done
# modified beam search
for epoch in 64; do
for avg in 33; do
./transducer_stateless_modified/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_modified/exp-4 \
--max-duration 100 \
--context-size 2 \
--decoding-method modified_beam_search \
--beam-size 4
done
done
```
You can find the decoding log for the above command in this
repo (in the folder [log][log]).
The WER for the test dataset is
| | test |comment |
|------------------------|------|----------------------------------------------------------------|
| greedy search | 5.22 |--epoch 64, --avg 33, --max-duration 100, --max-sym-per-frame 1 |
| modified beam search | 5.02 |--epoch 64, --avg 33, --max-duration 100 --beam-size 4 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```bash
epoch=64
avg=33
./transducer_stateless_modified/export.py \
--exp-dir ./transducer_stateless_modified/exp-4 \
--lang-dir ./data/lang_char \
--epoch $epoch \
--avg $avg
```
**HINT**: To use `pretrained.pt` to compute the WER for the `test` dataset,
just do the following:
```bash
cp icefall-aishell-transducer-stateless-modified-2022-03-01/exp/pretrained.pt \
/path/to/icefall/egs/aishell/ASR/transducer_stateless_modified/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless_modified/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/aishell/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{"language": "en", "license": "apache-2.0", "tags": ["icefall", "k2", "transducer", "aishell", "ASR", "stateless transducer", "PyTorch"], "datasets": ["aishell"], "metrics": ["WER"]}
| null |
csukuangfj/icefall-aishell-transducer-stateless-modified-2022-03-01
|
[
"k2",
"icefall",
"transducer",
"aishell",
"ASR",
"stateless transducer",
"PyTorch",
"en",
"dataset:aishell",
"license:apache-2.0",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #license-apache-2.0 #region-us
|
Introduction
============
This repo contains pre-trained model using
<URL
It is trained on AIShell dataset
using modified transducer from optimized\_transducer.
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit 'TODO'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 512-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the AIShell dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The commands for decoding are
You can find the decoding log for the above command in this
repo (in the folder [log](URL)).
The WER for the test dataset is
test: greedy search, comment: 5.22
test: modified beam search, comment: 5.02
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for the 'test' dataset,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_modified/URL'.
|
[] |
[
"TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #license-apache-2.0 #region-us \n"
] |
[
49
] |
[
"passage: TAGS\n#k2 #icefall #transducer #aishell #ASR #stateless transducer #PyTorch #en #dataset-aishell #license-apache-2.0 #region-us \n"
] |
[
-0.0912378653883934,
0.07988960295915604,
-0.006669369526207447,
0.07851304858922958,
0.008801380172371864,
-0.03460881486535072,
0.1909128725528717,
0.05409783869981766,
0.18148064613342285,
-0.0828222706913948,
0.2081434726715088,
0.14270251989364624,
0.02130713500082493,
0.1469823122024536,
0.0018595680594444275,
-0.2536882758140564,
0.03303275257349014,
-0.07665423303842545,
-0.06312529742717743,
0.08486305177211761,
0.06858454644680023,
-0.021544598042964935,
0.013000999577343464,
0.0025744843296706676,
0.05133645981550217,
-0.0026905182749032974,
0.029876604676246643,
-0.08128637820482254,
0.10192607343196869,
-0.053133197128772736,
0.039574362337589264,
0.06003212556242943,
0.012832049280405045,
-0.12373247742652893,
0.030157431960105896,
-0.03602173551917076,
-0.05816726014018059,
0.06906068325042725,
0.020647751167416573,
0.0068633113987743855,
-0.05441071838140488,
-0.06818012148141861,
-0.03350450471043587,
0.03624728322029114,
-0.06317850947380066,
-0.1613914966583252,
-0.07557716220617294,
0.06197865679860115,
0.020470058545470238,
0.046767041087150574,
0.046882305294275284,
0.16259972751140594,
-0.12858189642429352,
0.09164070338010788,
0.09944629669189453,
-0.3794322907924652,
0.04365764185786247,
-0.00095846236217767,
-0.07714761793613434,
0.07382962852716446,
0.0221920907497406,
0.02707475796341896,
-0.0040580881759524345,
-0.030639678239822388,
0.017736317589879036,
-0.002135912189260125,
-0.14537127315998077,
0.06697852164506912,
-0.09807232767343521,
-0.023945223540067673,
0.22957146167755127,
-0.00728967972099781,
0.034552887082099915,
0.07471691817045212,
-0.07000577449798584,
-0.060767434537410736,
0.07472936064004898,
0.05742533877491951,
0.015246530994772911,
0.08809284120798111,
0.030423996970057487,
-0.005438430234789848,
-0.11032744497060776,
-0.017874307930469513,
-0.16556604206562042,
0.12381387501955032,
-0.012242615222930908,
0.13673777878284454,
-0.15366560220718384,
0.02018345147371292,
0.031106892973184586,
-0.1360904425382614,
0.04168323799967766,
-0.08854102343320847,
0.042259540408849716,
0.10368205606937408,
-0.021453138440847397,
0.0228680819272995,
0.12416563183069229,
0.17513138055801392,
0.05956288427114487,
-0.01219430286437273,
0.03027266636490822,
0.13897959887981415,
0.08803866803646088,
0.02809186279773712,
0.05078687146306038,
0.0002833935141097754,
0.04131162539124489,
-0.02649529092013836,
0.09184984117746353,
-0.07380345463752747,
-0.06052558869123459,
-0.011226381175220013,
0.004484857432544231,
0.09055197238922119,
0.10788839310407639,
-0.0635041743516922,
-0.11025644838809967,
0.010831205174326897,
0.1636449545621872,
-0.07003968954086304,
0.00603801105171442,
0.03614818677306175,
-0.007903944700956345,
0.07127092033624649,
-0.009302046149969101,
0.07987377047538757,
0.03484390676021576,
0.04085765779018402,
-0.07472047954797745,
-0.00532520329579711,
0.018744785338640213,
0.032718874514102936,
0.0898662656545639,
-0.15068235993385315,
0.03934498876333237,
-0.13456763327121735,
-0.1865530014038086,
0.04674740880727768,
0.0331353098154068,
0.006312964018434286,
-0.07606945931911469,
0.029336875304579735,
-0.01591426320374012,
0.035490475594997406,
-0.07179666310548782,
-0.1183822974562645,
-0.0454021655023098,
0.012591739185154438,
-0.13687865436077118,
0.06004229560494423,
-0.13905766606330872,
-0.010852500796318054,
-0.1709751933813095,
0.0038335579447448254,
0.06900954991579056,
0.009904221631586552,
-0.15955688059329987,
0.09217049181461334,
-0.04582381993532181,
-0.0018194542499259114,
-0.06471173465251923,
-0.06852979958057404,
0.0274965837597847,
0.16167975962162018,
-0.2606445848941803,
-0.06605077534914017,
0.08944328874349594,
-0.1412372887134552,
-0.07851171493530273,
0.09584653377532959,
0.04632703214883804,
-0.08110708743333817,
0.03424641862511635,
0.1972573846578598,
-0.062363509088754654,
0.026740623638033867,
-0.06983506679534912,
0.1493120789527893,
-0.16207154095172882,
-0.15197616815567017,
0.12212496250867844,
-0.03475996106863022,
0.022455710917711258,
0.01562957465648651,
0.06247011199593544,
0.10661417245864868,
-0.02652014046907425,
-0.14012092351913452,
-0.05510948225855827,
-0.06013490632176399,
0.07225550711154938,
-0.004663246683776379,
0.026414921507239342,
-0.013939259573817253,
0.0076739112846553326,
-0.019916612654924393,
0.012247229926288128,
0.09415968507528305,
0.05433037504553795,
-0.048170190304517746,
-0.0018609627149999142,
0.019639911130070686,
-0.012817144393920898,
-0.07991112768650055,
-0.1020396426320076,
-0.01813538931310177,
-0.0333963967859745,
-0.0067701926454901695,
0.08474551141262054,
0.08417068421840668,
-0.028222449123859406,
0.021036790683865547,
0.000030475124731310643,
0.02479027770459652,
0.06133440136909485,
-0.044664401561021805,
-0.21472348272800446,
0.030629323795437813,
-0.043714508414268494,
0.0173184797167778,
0.031560298055410385,
0.03290265426039696,
0.14808349311351776,
0.05829736962914467,
-0.016148971393704414,
0.042777638882398605,
-0.03166559338569641,
-0.01655816286802292,
0.021473772823810577,
0.007769769057631493,
0.10401847213506699,
0.008981847204267979,
-0.18405045568943024,
0.1666388362646103,
-0.009725176729261875,
0.09301548451185226,
0.11492563039064407,
-0.03621193766593933,
0.13533613085746765,
0.00034020503517240286,
-0.07012379169464111,
-0.07224441319704056,
0.0763489380478859,
0.06403760612010956,
0.0668460801243782,
0.05717892199754715,
0.03476601094007492,
0.03281814977526665,
-0.024287691339850426,
-0.014253119938075542,
-0.07533284276723862,
-0.03367980942130089,
0.06819384545087814,
0.01745135895907879,
-0.17125874757766724,
0.10776758939027786,
0.3296218812465668,
-0.004476088099181652,
0.160936176776886,
-0.0984550341963768,
-0.05269044265151024,
-0.03202863410115242,
-0.09679681062698364,
-0.04375945404171944,
0.12936779856681824,
0.013030963018536568,
0.0383952371776104,
0.08960234373807907,
0.008655755780637264,
-0.007258477155119181,
-0.09566189348697662,
-0.07011207938194275,
-0.012708052061498165,
0.043217629194259644,
-0.22362090647220612,
0.06542042642831802,
-0.08990563452243805,
0.02482079155743122,
-0.017326360568404198,
-0.11404162645339966,
0.12900184094905853,
-0.011971475556492805,
-0.02978554368019104,
0.1255762130022049,
-0.16270938515663147,
-0.13144958019256592,
-0.07185633480548859,
0.13855727016925812,
-0.015364228747785091,
-0.07895947992801666,
0.06630139797925949,
-0.11734949052333832,
-0.024616481736302376,
0.011017793789505959,
-0.05074223875999451,
0.02330404706299305,
0.018224254250526428,
0.07981763035058975,
-0.03156455606222153,
-0.02406851015985012,
-0.0973866656422615,
-0.040477506816387177,
-0.034757256507873535,
-0.02620820328593254,
0.15048518776893616,
0.023970117792487144,
0.016635555773973465,
0.09501165896654129,
0.06400292366743088,
0.001875722547993064,
0.0042271665297448635,
0.09301481395959854,
-0.024582967162132263,
0.021343937143683434,
0.1320800632238388,
-0.06217304244637489,
0.09606315195560455,
0.06780920177698135,
0.0670260339975357,
0.016589229926466942,
-0.024558909237384796,
-0.019342387095093727,
-0.11219567060470581,
-0.2339407056570053,
-0.0689534991979599,
-0.04289085045456886,
0.18920962512493134,
-0.012022891081869602,
0.061017945408821106,
0.022438189014792442,
0.0818771943449974,
0.08090834319591522,
-0.08446662127971649,
-0.04491174593567848,
-0.005343109834939241,
0.0920477882027626,
-0.006565549876540899,
0.09325553476810455,
-0.059098981320858,
0.031006043776869774,
0.13661040365695953,
0.09189024567604065,
0.15743738412857056,
0.07265490293502808,
0.06164751574397087,
0.08071920275688171,
0.2106257677078247,
0.06398405879735947,
0.11643870919942856,
0.06173894926905632,
0.0003481822204776108,
-0.010527604259550571,
-0.019198084250092506,
-0.0627213716506958,
0.05561863258481026,
-0.03361527994275093,
-0.1359178125858307,
0.09359019994735718,
0.122604601085186,
0.09443344175815582,
0.30849936604499817,
0.052944984287023544,
-0.09825537353754044,
-0.030246753245592117,
0.05059781298041344,
0.004838303197175264,
0.03711705282330513,
0.09356563538312912,
-0.032572999596595764,
-0.0751810222864151,
0.07253054529428482,
-0.06378952413797379,
0.02651267871260643,
0.014406802132725716,
0.035934653133153915,
-0.00044957027421332896,
-0.10247434675693512,
0.0653790608048439,
0.07156770676374435,
-0.19068898260593414,
0.21089856326580048,
-0.017426539212465286,
-0.03855733573436737,
-0.051503077149391174,
-0.03231717273592949,
0.07364355027675629,
0.09814643114805222,
0.07624343782663345,
0.1003316193819046,
-0.10205621272325516,
-0.0030949125066399574,
-0.2048426866531372,
0.07306124269962311,
0.01624487154185772,
-0.007220481988042593,
-0.11370912939310074,
-0.028865238651633263,
0.048266924917697906,
0.040572185069322586,
-0.04346488416194916,
-0.1654147356748581,
-0.04027123749256134,
0.08227952569723129,
0.11921751499176025,
0.051582250744104385,
-0.04684380441904068,
-0.1345936506986618,
-0.02395833283662796,
0.04973499849438667,
-0.14804254472255707,
-0.09184578061103821,
-0.07757463306188583,
-0.05341194570064545,
0.11481121182441711,
-0.07870765775442123,
0.02731703780591488,
-0.017524149268865585,
-0.06935573369264603,
-0.02935604751110077,
-0.17036864161491394,
0.10326726734638214,
-0.1116199865937233,
-0.1531403809785843,
0.032947368919849396,
0.04079924523830414,
0.05331959202885628,
0.04850704222917557,
-0.052280180156230927,
0.09681210666894913,
-0.10598265379667282,
-0.0931326299905777,
0.12043843418359756,
-0.0038707368075847626,
-0.009804202243685722,
0.03848946467041969,
0.04774193465709686,
-0.031873807311058044,
0.01947372779250145,
-0.05870721489191055,
0.22439776360988617,
0.3064878284931183,
-0.10551347583532333,
0.2530943751335144,
0.232868954539299,
-0.021601304411888123,
-0.2518948018550873,
-0.17849886417388916,
-0.13449695706367493,
-0.06718666851520538,
0.14296524226665497,
-0.1871662437915802,
0.09939563274383545,
0.1399552822113037,
-0.10915409028530121,
0.08808479458093643,
-0.24154254794120789,
-0.06261148303747177,
0.18601901829242706,
-0.10382705181837082,
0.3112789988517761,
-0.11591673642396927,
-0.049192845821380615,
-0.029166093096137047,
-0.0726439580321312,
0.11274299770593643,
-0.129652738571167,
0.033185284584760666,
-0.04787520319223404,
0.0231630839407444,
-0.01982324756681919,
-0.06958181411027908,
0.12531019747257233,
0.023997360840439796,
-0.01691919006407261,
-0.05182144418358803,
-0.053042761981487274,
0.1500619351863861,
-0.022016003727912903,
0.025026241317391396,
-0.1263105571269989,
0.07398365437984467,
-0.044837817549705505,
0.04168490320444107,
-0.10527857393026352,
0.08941194415092468,
0.005248651374131441,
-0.08044013381004333,
0.026511162519454956,
0.0012417531106621027,
0.015155581757426262,
-0.00895716529339552,
0.23401859402656555,
0.07795356959104538,
-0.08085978031158447,
0.09375496953725815,
0.05962701514363289,
-0.06345726549625397,
-0.011563708074390888,
-0.12432615458965302,
-0.05835427716374397,
0.05946287512779236,
-0.10460516810417175,
0.060526810586452484,
0.11194875836372375,
0.021192869171500206,
-0.04262573644518852,
0.08463691920042038,
-0.030982602387666702,
0.018708202987909317,
0.09947454929351807,
-0.1750207394361496,
-0.0526166632771492,
0.027329187840223312,
0.12590566277503967,
0.08218211680650711,
0.16406187415122986,
0.16672024130821228,
0.008738663978874683,
-0.023008527234196663,
-0.004100600723177195,
0.11766185611486435,
-0.06703564524650574,
-0.016708016395568848,
0.09397455304861069,
0.012226168066263199,
-0.10942722111940384,
0.15835663676261902,
-0.06255161017179489,
-0.021798506379127502,
0.050581805408000946,
0.06369170546531677,
-0.09594488888978958,
-0.11283961683511734,
-0.049256909638643265,
-0.05734429880976677,
-0.1745336949825287,
-0.09350460022687912,
-0.042272135615348816,
-0.05557214096188545,
-0.0027147396467626095,
0.04348263144493103,
0.05519729107618332,
0.08399542421102524,
-0.0030440788250416517,
-0.04849390685558319,
0.07509474456310272,
0.0517427921295166,
-0.101190946996212,
-0.02502637729048729,
-0.12566472589969635,
-0.07464843988418579,
-0.035303350538015366,
0.1101202443242073,
-0.05189479514956474,
-0.02026687189936638,
-0.04881826043128967,
0.00905164610594511,
-0.017723189666867256,
-0.07829699665307999,
-0.07852557301521301,
-0.002520667389035225,
0.014191958121955395,
-0.07859302312135696,
-0.016515860334038734,
0.06147865206003189,
-0.09526724368333817,
0.01253076083958149,
-0.013051867485046387,
0.024914147332310677,
-0.17246019840240479,
-0.06659027934074402,
0.0976024940609932,
0.033472683280706406,
0.14777353405952454,
0.11196228861808777,
-0.021210787817835808,
0.0676933079957962,
-0.04208976402878761,
0.013453011400997639,
0.054835572838783264,
0.026889638975262642,
0.0271427221596241,
0.06818908452987671,
-0.08139735460281372,
0.030318889766931534,
0.000246904935920611,
0.0626215785741806,
0.12740010023117065,
-0.07643812149763107,
-0.00841483660042286,
0.0471816323697567,
-0.12746913731098175,
-0.07372988760471344,
-0.11302640289068222,
0.05542151257395744,
0.06523116677999496,
0.18119804561138153,
-0.009731446392834187,
0.026415813714265823,
-0.011357779614627361,
0.02826249785721302,
0.0509478896856308,
-0.0767626166343689,
-0.05829155072569847,
0.01954500935971737,
-0.023083651438355446,
-0.05946911498904228,
0.18500490486621857,
-0.007898562587797642,
-0.14597561955451965,
0.055813390761613846,
-0.002432363573461771,
-0.04977600648999214,
0.022805001586675644,
0.14182807505130768,
0.014843160286545753,
-0.008572817780077457,
-0.10036437958478928,
0.029652049764990807,
-0.07150762528181076,
-0.11461387574672699,
0.14716260135173798,
0.004678407218307257,
0.06345683336257935,
0.034838367253541946,
0.06739576160907745,
-0.06207776442170143,
-0.13529038429260254,
-0.14786063134670258,
-0.07946500182151794,
0.06876452267169952,
0.05247573181986809,
0.13741017878055573,
0.1240936666727066,
0.02286904864013195,
-0.008667477406561375,
-0.0590241476893425,
-0.04896598309278488,
-0.16247017681598663,
-0.13699871301651,
-0.006556198466569185,
-0.09643697738647461,
0.03727190941572189,
-0.024525433778762817,
0.04021124914288521,
0.1975860744714737,
0.04075571149587631,
0.007448310498148203,
-0.02112867310643196,
-0.0286518856883049,
0.07285694032907486,
0.006339365616440773,
-0.032252609729766846,
-0.06877052038908005,
-0.04114345461130142,
-0.023358801379799843,
-0.04452336207032204,
0.0012511782115325332,
-0.07133675366640091,
-0.031438764184713364,
-0.07171868532896042,
0.010047467425465584,
-0.050128474831581116,
-0.08779004216194153,
-0.06320007890462875,
0.008802184835076332,
-0.003513575065881014,
0.07676175236701965,
0.0700116977095604,
0.0721297338604927,
0.04681841656565666,
0.1285131722688675,
-0.022378871217370033,
-0.09976514428853989,
-0.052428457885980606,
0.052817825227975845,
-0.06167568266391754,
0.047914065420627594,
-0.017426693812012672,
-0.0557255744934082,
-0.047955453395843506,
0.02605794556438923,
0.21408070623874664,
-0.06597311049699783,
0.022219963371753693,
-0.05037020146846771,
0.01047042477875948,
-0.014693941920995712,
0.06612900644540787,
0.08440433442592621,
0.1376197189092636,
-0.13136446475982666,
-0.01656896434724331,
-0.06262663006782532,
-0.02060304582118988,
-0.13870801031589508,
-0.010879376903176308,
-0.053852055221796036,
-0.045270830392837524,
-0.07633572816848755,
0.11109621822834015,
-0.10148157924413681,
0.09727495908737183,
-0.02727482281625271,
-0.10534919798374176,
-0.022639401257038116,
-0.03090241178870201,
0.1807427853345871,
0.02348906733095646,
0.09999065846204758,
-0.05316198244690895,
-0.04720750078558922,
0.020008506253361702,
0.04680195078253746,
-0.20023256540298462,
0.04067599028348923,
0.06809109449386597,
-0.12123260647058487,
0.14734168350696564,
-0.0183585025370121,
0.15419432520866394,
0.09288114309310913,
0.031404681503772736,
-0.11442625522613525,
0.013878867030143738,
0.020791780203580856,
-0.029323097318410873,
-0.11869718134403229,
-0.05416688323020935,
0.029434235766530037,
-0.08485594391822815,
0.05812680348753929,
0.03892120346426964,
0.0016652296762913465,
0.13654734194278717,
0.004301830660551786,
0.010969337075948715,
0.06596948206424713,
-0.0204599741846323,
0.05306898429989815,
-0.0233455803245306,
-0.043262235820293427,
-0.054493606090545654,
-0.05146151781082153,
-0.04442114755511284,
0.0075681558810174465,
-0.19835273921489716,
-0.07324900478124619,
0.09942108392715454,
0.022010236978530884,
0.06941249966621399,
0.025664355605840683,
0.05531547963619232,
-0.05885900929570198,
-0.16657498478889465,
-0.011496917344629765,
-0.11001061648130417,
-0.012739116325974464,
0.022808190435171127,
-0.02144463174045086,
0.03070256859064102,
-0.05084160715341568,
0.04164474830031395,
-0.026119444519281387,
-0.1252989023923874,
-0.06487127393484116
] |
null | null | null |
# Introduction
This repo contains pre-trained model using
<https://github.com/k2-fsa/icefall/pull/213>.
It is trained on train-clean-100 subset of the LibriSpeech dataset.
Also, it uses the `S` subset from GigaSpeech as extra training data.
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21
cd icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `2332ba312d7ce72f08c7bac1e3312f7e3dd722dc`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout 2332ba312d7ce72f08c7bac1e3312f7e3dd722dc
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/2332ba312d7ce72f08c7bac1e3312f7e3dd722dc/egs/librispeech/ASR/transducer_stateless_multi_datasets/train.py#L198>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
./prepare_giga_speech.sh
export CUDA_VISIBLE_DEVICES="0,1"
./transducer_stateless_multi_datasets/train.py \
--world-size 2 \
--num-epochs 60 \
--start-epoch 0 \
--exp-dir transducer_stateless_multi_datasets/exp-100-2 \
--full-libri 0 \
--max-duration 300 \
--lr-factor 1 \
--bpe-model data/lang_bpe_500/bpe.model \
--modified-transducer-prob 0.25
--giga-prob 0.2
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/qUEKzMnrTZmOz1EXPda9RA/>
The command for decoding is:
```
epoch=57
avg=17
## greedy search
for epoch in 57; do
for avg in 17; do
for sym in 1 2 3; do
./transducer_stateless_multi_datasets/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_multi_datasets/exp-100-2 \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--context-size 2 \
--max-sym-per-frame $sym
done
done
done
## modified beam search
epoch=57
avg=17
./transducer_stateless_multi_datasets/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_multi_datasets/exp-100-2 \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--context-size 2 \
--decoding-method modified_beam_search \
--beam-size 4
```
You can find the decoding log for the above command in this
repo (in the folder `log`).
The WERs for the test datasets are
| | test-clean | test-other | comment |
|-------------------------------------|------------|------------|------------------------------------------|
| greedy search (max sym per frame 1) | 6.34 | 16.7 | --epoch 57, --avg 17, --max-duration 100 |
| greedy search (max sym per frame 2) | 6.34 | 16.7 | --epoch 57, --avg 17, --max-duration 100 |
| greedy search (max sym per frame 3) | 6.34 | 16.7 | --epoch 57, --avg 17, --max-duration 100 |
| modified beam search (beam size 4) | 6.31 | 16.3 | --epoch 57, --avg 17, --max-duration 100 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```bash
./transducer_stateless_multi_datasets/export.py \
--epoch 57 \
--avg 17 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer_stateless_multi_datasets/exp-full
```
**HINT**: To use `pretrained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer_stateless_multi_datasets/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless_multi_datasets/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-100h-transducer-stateless-multi-datasets-bpe-500-2022-02-21
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
This repo contains pre-trained model using
<URL
It is trained on train-clean-100 subset of the LibriSpeech dataset.
Also, it uses the 'S' subset from GigaSpeech as extra training data.
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit '2332ba312d7ce72f08c7bac1e3312f7e3dd722dc'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The command for decoding is:
You can find the decoding log for the above command in this
repo (in the folder 'log').
The WERs for the test datasets are
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_multi\_datasets/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09
cd icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
-----
## Description
This repo provides pre-trained conformer CTC model for the librispeech dataset
using [icefall][icefall].
The commands for training are:
```
cd egs/librispeech/ASR/conformer_ctc
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./conformer_ctc/train.py \
--exp-dir conformer_ctc/exp_500_att0.8 \
--lang-dir data/lang_bpe_500 \
--att-rate 0.8 \
--full-libri 1 \
--max-duration 200 \
--concatenate-cuts 0 \
--world-size 4 \
--bucketing-sampler 1 \
--start-epoch 0 \
--num-epochs 90
```
The command for decoding is:
```
./conformer_ctc/decode.py \
--exp-dir conformer_ctc/exp_500_att0.8 \
--lang-dir data/lang_bpe_500 \
--max-duration 30 \
--concatenate-cuts 0 \
--bucketing-sampler 1 \
--num-paths 1000 \
--epoch 77 \
--avg 55 \
--method attention-decoder \
--nbest-scale 0.5
```
You can find the decoding log for the above command in this
repo: [log/log-decode-2021-11-09-17-38-28](log/log-decode-2021-11-09-17-38-28).
The best WER for the librispeech test dataset is:
| | test-clean | test-other |
|-----|------------|------------|
| WER | 2.42 | 5.73 |
Scale values used in n-gram LM rescoring and attention rescoring for the best WERs are:
| ngram_lm_scale | attention_scale |
|----------------|-----------------|
| 2.0 | 2.0 |
# File description
- [log][log], this directory contains the decoding log
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
Note: For the `data/lm` directory, we provide only `G_4_gram.pt`. If you need other files
in this directory, please run [prepare.sh][prepare].
- [exp][exp], this directory contains two files: `preprained.pt` and `cpu_jit.pt`.
`exp/pretrained.pt` is generated by the following command:
```
./conformer_ctc/export.py \
--epoch 77 \
--avg 55 \
--jit 0 \
--lang-dir data/lang_bpe_500 \
--exp-dir conformer_ctc/exp_500_att0.8
```
**HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/conformer_ctc/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `conformer_ctc/decode.py`.
`exp/cpu_jit.pt` is generated by the following command:
```
./conformer_ctc/export.py \
--epoch 77 \
--avg 55 \
--jit 1 \
--lang-dir data/lang_bpe_500 \
--exp-dir conformer_ctc/exp_500_att0.8
```
# Deploy your model in C++ using k2
To deploy your model in C++ using k2 without depending on Python, do the following:
```
# Note: It requires torch >= 1.8.0
git clone https://github.com/k2-fsa/k2
cd k2
git checkout v2.0-pre
mkdir build_release
cd build_release
cmake -DCMAKE_BUILD_TYPE=Release ..
make -j ctc_decode hlg_decode ngram_lm_rescore attention_rescore
```
## CTC decoding
```
cd k2/build_release
./bin/ctc_decode \
--use_gpu true \
--nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \
--bpe_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/bpe.model \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav
```
## HLG decoding
```
./bin/hlg_decode \
--use_gpu true \
--nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \
--hlg ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/HLG.pt \
--word_table ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/words.txt \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav
```
## HLG decoding + n-gram LM rescoring
**NOTE**: V100 GPU with 16 GB RAM is known NOT to work because of OOM.
V100 GPU with 32 GB RAM is known to work.
```
./bin/ngram_lm_rescore \
--use_gpu true \
--nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \
--hlg ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/HLG.pt \
--g ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lm/G_4_gram.pt \
--ngram_lm_scale 1.0 \
--word_table ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/words.txt \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav
```
## HLG decoding + n-gram LM rescoring + attention decoder rescoring
**NOTE**: V100 GPU with 16 GB RAM is known NOT to work because of OOM.
V100 GPU with 32 GB RAM is known to work.
```
./bin/attention_rescore \
--use_gpu true \
--nn_model ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/exp/cpu_jit.pt \
--hlg ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/HLG.pt \
--g ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lm/G_4_gram.pt \
--ngram_lm_scale 2.0 \
--attention_scale 2.0 \
--num_paths 100 \
--nbest_scale 0.5 \
--word_table ./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/data/lang_bpe_500/words.txt \
--sos_id 1 \
--eos_id 1 \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1089-134686-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0001.wav \
./icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/test_wavs/1221-135766-0002.wav
```
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-conformer-ctc-jit-bpe-500-2021-11-09
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
---
Description
-----------
This repo provides pre-trained conformer CTC model for the librispeech dataset
using [icefall](URL).
The commands for training are:
The command for decoding is:
You can find the decoding log for the above command in this
repo: log/log-decode-2021-11-09-17-38-28.
The best WER for the librispeech test dataset is:
test-clean: WER, test-other: 2.42
Scale values used in n-gram LM rescoring and attention rescoring for the best WERs are:
File description
================
* [log](URL), this directory contains the decoding log
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
Note: For the 'data/lm' directory, we provide only 'G\_4\_gram.pt'. If you need other files
in this directory, please run <URL>.
* [exp](URL), this directory contains two files: 'URL' and 'cpu\_jit.pt'.
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'conformer\_ctc/URL'.
'exp/cpu\_jit.pt' is generated by the following command:
Deploy your model in C++ using k2
=================================
To deploy your model in C++ using k2 without depending on Python, do the following:
CTC decoding
------------
HLG decoding
------------
HLG decoding + n-gram LM rescoring
----------------------------------
NOTE: V100 GPU with 16 GB RAM is known NOT to work because of OOM.
V100 GPU with 32 GB RAM is known to work.
HLG decoding + n-gram LM rescoring + attention decoder rescoring
----------------------------------------------------------------
NOTE: V100 GPU with 16 GB RAM is known NOT to work because of OOM.
V100 GPU with 32 GB RAM is known to work.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17
cd icefall-asr-librispeech-transducer-bpe-500-2021-12-17
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `cb04c8a7509425ab45fae888b0ca71bbbd23f0de`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout cb04c8a7509425ab45fae888b0ca71bbbd23f0de
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/cb04c8a7509425ab45fae888b0ca71bbbd23f0de/egs/librispeech/ASR/transducer/train.py#L196>
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer, plus a 4-layer LSTM with hidden size 512.
-----
## Description
This repo provides pre-trained RNN-T Conformer model for the librispeech dataset
using [icefall][icefall].
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer/train.py \
--world-size 4 \
--num-epochs 30 \
--start-epoch 0 \
--exp-dir transducer/exp-lr-2.5-full \
--full-libri 1 \
--max-duration 250 \
--lr-factor 2.5
```
The command for decoding is:
```
epoch=26
avg=12
./transducer/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer/exp-lr-2.5-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100
```
You can find the decoding log for the above command in this
repo: [log/log-decode-epoch-26-avg-12-2021-12-17-09-33-04](log/log-decode-epoch-26-avg-12-2021-12-17-09-33-04).
The best WER using greedy search is:
| | test-clean | test-other |
|-----|------------|------------|
| WER | 3.16 | 7.71 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```
./transducer/export.py \
--epoch 26 \
--avg 12 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer/exp-lr-2.5-full
```
**HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-bpe-500-2021-12-17/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-17
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit 'cb04c8a7509425ab45fae888b0ca71bbbd23f0de'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer, plus a 4-layer LSTM with hidden size 512.
---
Description
-----------
This repo provides pre-trained RNN-T Conformer model for the librispeech dataset
using [icefall](URL).
The commands for training are:
The command for decoding is:
You can find the decoding log for the above command in this
repo: log/log-decode-epoch-26-avg-12-2021-12-17-09-33-04.
The best WER using greedy search is:
test-clean: WER, test-other: 3.16
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23
cd icefall-asr-librispeech-transducer-bpe-500-2021-12-23
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `5b6699a8354b70b23b252b371c612a35ed186ec2`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout 5b6699a8354b70b23b252b371c612a35ed186ec2
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/5b6699a8354b70b23b252b371c612a35ed186ec2/egs/librispeech/ASR/transducer/train.py#L191>
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer, plus a 2-layer LSTM with hidden size 512.
-----
## Description
This repo provides pre-trained RNN-T Conformer model for the librispeech dataset
using [icefall][icefall].
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer/train.py \
--world-size 4 \
--num-epochs 35 \
--start-epoch 0 \
--exp-dir transducer/exp-lr-2.5-full \
--full-libri 1 \
--max-duration 180 \
--lr-factor 2.5
```
The command for decoding is:
```
epoch=34
avg=11
./transducer/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer/exp-lr-2.5-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100
```
You can find the decoding log for the above command in the `log` folder
of this repo.
The best WER using greedy search is:
| | test-clean | test-other |
|-----|------------|------------|
| WER | 3.07 | 7.51 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```
./transducer/export.py \
--epoch 34 \
--avg 11 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer/exp-lr-2.5-full
```
**HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-bpe-500-2021-12-23/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-transducer-bpe-500-2021-12-23
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit '5b6699a8354b70b23b252b371c612a35ed186ec2'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer, plus a 2-layer LSTM with hidden size 512.
---
Description
-----------
This repo provides pre-trained RNN-T Conformer model for the librispeech dataset
using [icefall](URL).
The commands for training are:
The command for decoding is:
You can find the decoding log for the above command in the 'log' folder
of this repo.
The best WER using greedy search is:
test-clean: WER, test-other: 3.07
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22
cd icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32/egs/librispeech/ASR/transducer_stateless/train.py#L195>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer_stateless/train.py \
--world-size 4 \
--num-epochs 30 \
--start-epoch 0 \
--exp-dir transducer_stateless/exp-full \
--full-libri 1 \
--max-duration 250 \
--lr-factor 3
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/PsJ3LgkEQfOmzedAlYfVeg/#scalars&_smoothingWeight=0>
The command for decoding is:
```
epoch=20
avg=10
## greedy search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100
## beam search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--decoding-method beam_search \
--beam-size 4
```
You can find the decoding log for the above command in this
repo (in the folder `log`).
The WERs for the test datasets are
| | test-clean | test-other | comment |
|---------------------------|------------|------------|------------------------------------------|
| greedy search | 2.99 | 7.52 | --epoch 20, --avg 10, --max-duration 100 |
| beam search (beam size 2) | 2.95 | 7.43 | |
| beam search (beam size 3) | 2.94 | 7.37 | |
| beam search (beam size 4) | 2.92 | 7.37 | |
| beam search (beam size 5) | 2.93 | 7.38 | |
| beam search (beam size 8) | 2.92 | 7.38 | |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```
./transducer_stateless/export.py \
--epoch 20 \
--avg 10 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer_stateless/exp-full
```
**HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-22
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit 'fb6a57e9e01dd8aae2af2a6b4568daad8bc8ab32'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The command for decoding is:
You can find the decoding log for the above command in this
repo (in the folder 'log').
The WERs for the test datasets are
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27
cd icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `14c93add507982306f5a478cd144e0e32e0f970d`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout 14c93add507982306f5a478cd144e0e32e0f970d
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/14c93add507982306f5a478cd144e0e32e0f970d/egs/librispeech/ASR/transducer_stateless/train.py#L198>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer_stateless/train.py \
--world-size 4 \
--num-epochs 30 \
--start-epoch 0 \
--exp-dir transducer_stateless/exp-full \
--full-libri 1 \
--max-duration 250 \
--lr-factor 3
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/Mjx7MeTgR3Oyr1yBCwjozw/>
The command for decoding is:
```
epoch=29
avg=13
## greedy search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100
## beam search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--decoding-method beam_search \
--beam-size 4
```
You can find the decoding log for the above command in this
repo (in the folder `log`).
The WERs for the test datasets are
| | test-clean | test-other | comment |
|---------------------------|------------|------------|------------------------------------------|
| greedy search | 2.85 | 7.30 | --epoch 29, --avg 13, --max-duration 100 |
| beam search (beam size 4) | 2.83 | 7.19 | |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```
./transducer_stateless/export.py \
--epoch 29 \
--avg 13 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer_stateless/exp-full
```
**HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2021-12-27
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit '14c93add507982306f5a478cd144e0e32e0f970d'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The command for decoding is:
You can find the decoding log for the above command in this
repo (in the folder 'log').
The WERs for the test datasets are
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10
cd icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `4c1b3665ee6efb935f4dd93a80ff0e154b13efb6`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout 4c1b3665ee6efb935f4dd93a80ff0e154b13efb6
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/273e5fb2f3ac2620bafdffe2689b8b3ee10173d3/egs/librispeech/ASR/transducer_stateless/train.py#L198>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer_stateless/train.py \
--world-size 4 \
--num-epochs 76 \
--start-epoch 0 \
--exp-dir transducer_stateless/exp-full \
--full-libri 1 \
--max-duration 250 \
--lr-factor 3
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/qGdqzHnxS0WJ695OXfZDzA/>
The command for decoding is:
```
epoch=71
avg=15
## greedy search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100
## beam search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--decoding-method beam_search \
--beam-size 4
```
You can find the decoding log for the above command in this
repo (in the folder `log`).
The WERs for the test datasets are
| | test-clean | test-other | comment |
|---------------------------|------------|------------|------------------------------------------|
| greedy search | 2.69 | 6.81 | --epoch 71, --avg 15, --max-duration 100 |
| beam search (beam size 4) | 2.68 | 6.72 | --epoch 71, --avg 15, --max-duration 100 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```
./transducer_stateless/export.py \
--epoch 71 \
--avg 15 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer_stateless/exp-full
```
**HINT**: To use `pre-trained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-01-10
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit '4c1b3665ee6efb935f4dd93a80ff0e154b13efb6'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The command for decoding is:
You can find the decoding log for the above command in this
repo (in the folder 'log').
The WERs for the test datasets are
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07
cd icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `a8150021e01d34ecbd6198fe03a57eacf47a16f2`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout a8150021e01d34ecbd6198fe03a57eacf47a16f2
```
to download `icefall`.
You can find the model information by visiting <https://github.com/k2-fsa/icefall/blob/a8150021e01d34ecbd6198fe03a57eacf47a16f2/egs/librispeech/ASR/transducer_stateless/train.py#L198>.
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer_stateless/train.py \
--world-size 4 \
--num-epochs 76 \
--start-epoch 0 \
--exp-dir transducer_stateless/exp-full \
--full-libri 1 \
--max-duration 300 \
--lr-factor 5 \
--bpe-model data/lang_bpe_500/bpe.model \
--modified-transducer-prob 0.25
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/qgvWkbF2R46FYA6ZMNmOjA/>
The command for decoding is:
```
epoch=63
avg=19
## greedy search
for sym in 1 2 3; do
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--max-sym-per-frame $sym
done
## modified beam search
./transducer_stateless/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless/exp-full \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--context-size 2 \
--decoding-method modified_beam_search \
--beam-size 4
```
You can find the decoding log for the above command in this
repo (in the folder `log`).
The WERs for the test datasets are
| | test-clean | test-other | comment |
|-------------------------------------|------------|------------|------------------------------------------|
| greedy search (max sym per frame 1) | 2.67 | 6.67 | --epoch 63, --avg 19, --max-duration 100 |
| greedy search (max sym per frame 2) | 2.67 | 6.67 | --epoch 63, --avg 19, --max-duration 100 |
| greedy search (max sym per frame 3) | 2.67 | 6.67 | --epoch 63, --avg 19, --max-duration 100 |
| modified beam search (beam size 4) | 2.67 | 6.57 | --epoch 63, --avg 19, --max-duration 100 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```
./transducer_stateless/export.py \
--epoch 63 \
--avg 19 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer_stateless/exp-full
```
**HINT**: To use `pretrained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer_stateless/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{}
| null |
csukuangfj/icefall-asr-librispeech-transducer-stateless-bpe-500-2022-02-07
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
Introduction
============
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit 'a8150021e01d34ecbd6198fe03a57eacf47a16f2'.
You can use
to download 'icefall'.
You can find the model information by visiting <URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The command for decoding is:
You can find the decoding log for the above command in this
repo (in the folder 'log').
The WERs for the test datasets are
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless/URL'.
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null |
k2
|
# Introduction
This repo contains pre-trained model using
<https://github.com/k2-fsa/icefall/pull/213>.
It is trained on full LibriSpeech dataset.
Also, it uses the `L` subset from [GigaSpeech](https://github.com/SpeechColab/GigaSpeech)
as extra training data.
## How to clone this repo
```
sudo apt-get install git-lfs
git clone https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01
cd icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01
git lfs pull
```
**Catuion**: You have to run `git lfs pull`. Otherwise, you will be SAD later.
The model in this repo is trained using the commit `2332ba312d7ce72f08c7bac1e3312f7e3dd722dc`.
You can use
```
git clone https://github.com/k2-fsa/icefall
cd icefall
git checkout 2332ba312d7ce72f08c7bac1e3312f7e3dd722dc
```
to download `icefall`.
You can find the model information by visiting
<https://github.com/k2-fsa/icefall/blob/2332ba312d7ce72f08c7bac1e3312f7e3dd722dc/egs/librispeech/ASR/transducer_stateless_multi_datasets/train.py#L218>
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
[Rnn-Transducer with Stateless Prediction Network](https://ieeexplore.ieee.org/document/9054419).
A Conv1d layer is placed right after the input embedding layer.
-----
## Description
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall][icefall]. There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
```
cd egs/librispeech/ASR/
./prepare.sh
./prepare_giga_speech.sh
export CUDA_VISIBLE_DEVICES="0,1,2,3"
./transducer_stateless_multi_datasets/train.py \
--world-size 4 \
--num-epochs 40 \
--start-epoch 0 \
--exp-dir transducer_stateless_multi_datasets/exp-full-2 \
--full-libri 1 \
--max-duration 300 \
--lr-factor 5 \
--bpe-model data/lang_bpe_500/bpe.model \
--modified-transducer-prob 0.25 \
--giga-prob 0.2
```
The tensorboard training log can be found at
<https://tensorboard.dev/experiment/xmo5oCgrRVelH9dCeOkYBg/>
The command for decoding is:
```bash
epoch=39
avg=15
sym=1
# greedy search
./transducer_stateless_multi_datasets/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_multi_datasets/exp-full-2 \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--context-size 2 \
--max-sym-per-frame $sym
# modified beam search
./transducer_stateless_multi_datasets/decode.py \
--epoch $epoch \
--avg $avg \
--exp-dir transducer_stateless_multi_datasets/exp-full-2 \
--bpe-model ./data/lang_bpe_500/bpe.model \
--max-duration 100 \
--context-size 2 \
--decoding-method modified_beam_search \
--beam-size 4
```
You can find the decoding log for the above command in this
repo (in the folder `log`).
The WERs for the test datasets are
| | test-clean | test-other | comment |
|-------------------------------------|------------|------------|------------------------------------------|
| greedy search (max sym per frame 1) | 2.64 | 6.55 | --epoch 39, --avg 15, --max-duration 100 |
| modified beam search (beam size 4) | 2.61 | 6.46 | --epoch 39, --avg 15, --max-duration 100 |
# File description
- [log][log], this directory contains the decoding log and decoding results
- [test_wavs][test_wavs], this directory contains wave files for testing the pre-trained model
- [data][data], this directory contains files generated by [prepare.sh][prepare]
- [exp][exp], this directory contains only one file: `preprained.pt`
`exp/pretrained.pt` is generated by the following command:
```bash
./transducer_stateless_multi_datasets/export.py \
--epoch 39 \
--avg 15 \
--bpe-model data/lang_bpe_500/bpe.model \
--exp-dir transducer_stateless_multi_datasets/exp-full-2
```
**HINT**: To use `pretrained.pt` to compute the WER for test-clean and test-other,
just do the following:
```
cp icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/exp/pretrained.pt \
/path/to/icefall/egs/librispeech/ASR/transducer_stateless_multi_datasets/exp/epoch-999.pt
```
and pass `--epoch 999 --avg 1` to `transducer_stateless_multi_datasets/decode.py`.
[icefall]: https://github.com/k2-fsa/icefall
[prepare]: https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/prepare.sh
[exp]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/exp
[data]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/data
[test_wavs]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/test_wavs
[log]: https://huggingface.co/csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01/tree/main/log
[icefall]: https://github.com/k2-fsa/icefall
|
{"language": "en", "license": "apache-2.0", "tags": ["icefall", "k2", "transducer", "librispeech", "ASR", "stateless transducer", "PyTorch", "RNN-T", "speech recognition"], "datasets": ["librispeech"], "metrics": ["WER"]}
| null |
csukuangfj/icefall-asr-librispeech-transducer-stateless-multi-datasets-bpe-500-2022-03-01
|
[
"k2",
"icefall",
"transducer",
"librispeech",
"ASR",
"stateless transducer",
"PyTorch",
"RNN-T",
"speech recognition",
"en",
"dataset:librispeech",
"license:apache-2.0",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#k2 #icefall #transducer #librispeech #ASR #stateless transducer #PyTorch #RNN-T #speech recognition #en #dataset-librispeech #license-apache-2.0 #region-us
|
Introduction
============
This repo contains pre-trained model using
<URL
It is trained on full LibriSpeech dataset.
Also, it uses the 'L' subset from GigaSpeech
as extra training data.
How to clone this repo
----------------------
Catuion: You have to run 'git lfs pull'. Otherwise, you will be SAD later.
The model in this repo is trained using the commit '2332ba312d7ce72f08c7bac1e3312f7e3dd722dc'.
You can use
to download 'icefall'.
You can find the model information by visiting
<URL
In short, the encoder is a Conformer model with 8 heads, 12 encoder layers, 512-dim attention, 2048-dim feedforward;
the decoder contains a 1024-dim embedding layer and a Conv1d with kernel size 2.
The decoder architecture is modified from
Rnn-Transducer with Stateless Prediction Network.
A Conv1d layer is placed right after the input embedding layer.
---
Description
-----------
This repo provides pre-trained transducer Conformer model for the LibriSpeech dataset
using [icefall](URL). There are no RNNs in the decoder. The decoder is stateless
and contains only an embedding layer and a Conv1d.
The commands for training are:
The tensorboard training log can be found at
<URL
The command for decoding is:
You can find the decoding log for the above command in this
repo (in the folder 'log').
The WERs for the test datasets are
File description
================
* [log](URL), this directory contains the decoding log and decoding results
* [test\_wavs](URL), this directory contains wave files for testing the pre-trained model
* [data](URL), this directory contains files generated by <URL>
* [exp](URL), this directory contains only one file: 'URL'
'exp/URL' is generated by the following command:
HINT: To use 'URL' to compute the WER for test-clean and test-other,
just do the following:
and pass '--epoch 999 --avg 1' to 'transducer\_stateless\_multi\_datasets/URL'.
|
[] |
[
"TAGS\n#k2 #icefall #transducer #librispeech #ASR #stateless transducer #PyTorch #RNN-T #speech recognition #en #dataset-librispeech #license-apache-2.0 #region-us \n"
] |
[
60
] |
[
"passage: TAGS\n#k2 #icefall #transducer #librispeech #ASR #stateless transducer #PyTorch #RNN-T #speech recognition #en #dataset-librispeech #license-apache-2.0 #region-us \n"
] |
[
-0.06383673846721649,
0.10928380489349365,
-0.007234589196741581,
0.05235107243061066,
-0.013617466203868389,
-0.04787064343690872,
0.14397504925727844,
0.08755352348089218,
0.12339134514331818,
-0.07475242018699646,
0.18447791039943695,
0.16649699211120605,
0.015006504952907562,
0.08972150087356567,
-0.005499931052327156,
-0.27207478880882263,
0.03771843761205673,
-0.06523624062538147,
0.02768099308013916,
0.09897862374782562,
0.07782378047704697,
0.005415977910161018,
-0.006837237626314163,
0.019284622743725777,
0.026077914983034134,
0.027298033237457275,
0.04625777155160904,
-0.11693038046360016,
0.12307829409837723,
-0.016338223591446877,
-0.01370417233556509,
0.0958828404545784,
0.019184835255146027,
-0.18825173377990723,
0.021731054410338402,
-0.03515755757689476,
-0.05755608528852463,
0.032191820442676544,
-0.012311664409935474,
-0.02449170872569084,
0.022781435400247574,
-0.06276869773864746,
-0.015734972432255745,
0.05784756317734718,
-0.030432358384132385,
-0.23722290992736816,
-0.028044672682881355,
0.10369934141635895,
-0.03447456657886505,
0.022027235478162766,
0.003631148487329483,
0.14839264750480652,
-0.1168379932641983,
0.10449028015136719,
0.06641534715890884,
-0.3441197872161865,
0.05054883286356926,
-0.009389100596308708,
-0.015874184668064117,
0.08305826783180237,
0.0012050634250044823,
0.049367960542440414,
0.004133995622396469,
-0.030407389625906944,
-0.03979415446519852,
-0.026273034512996674,
-0.1335282325744629,
0.06239426136016846,
-0.11267382651567459,
0.02115130051970482,
0.18995831906795502,
-0.022819358855485916,
0.03494962304830551,
0.03362118825316429,
-0.04124321788549423,
-0.02464749664068222,
0.06788698583841324,
0.005917070899158716,
-0.0009863493032753468,
0.07599521428346634,
-0.03279581665992737,
-0.07071486860513687,
-0.08110107481479645,
-0.03992684558033943,
-0.14506396651268005,
0.13959719240665436,
0.014829132705926895,
0.10182486474514008,
-0.12479282915592194,
-0.008670518174767494,
-0.010906199924647808,
-0.11360354721546173,
0.011912094429135323,
-0.07094273716211319,
0.04019530490040779,
0.09798784554004669,
-0.03292367234826088,
0.03114907443523407,
0.1590505838394165,
0.17247051000595093,
0.07883824408054352,
0.012803126126527786,
-0.04353862255811691,
0.12587924301624298,
0.02626929059624672,
0.04395899176597595,
0.03509999066591263,
-0.009383091703057289,
0.03678654134273529,
-0.05695854127407074,
0.08210653066635132,
-0.06491635739803314,
-0.08518802374601364,
-0.029652327299118042,
0.0007303226739168167,
0.12614524364471436,
0.07568629086017609,
-0.00868998747318983,
-0.02399742417037487,
0.025724301114678383,
0.06868362426757812,
-0.08440397679805756,
0.007368502207100391,
0.04523409157991409,
0.005521790124475956,
0.08182868361473083,
0.00943367276340723,
0.0699942484498024,
-0.010708951391279697,
0.013633018359541893,
-0.023496506735682487,
0.0014902260154485703,
0.056734491139650345,
-0.0013342555612325668,
0.07336751371622086,
-0.16461902856826782,
0.027974482625722885,
-0.1673319935798645,
-0.14570698142051697,
0.03862711042165756,
-0.015249224379658699,
-0.0033039345871657133,
-0.05420326441526413,
-0.0502631776034832,
-0.0346265584230423,
0.03338634967803955,
-0.09465475380420685,
-0.11548414081335068,
-0.057220470160245895,
0.03745858743786812,
-0.10069489479064941,
0.10636470466852188,
-0.174094557762146,
0.006645421963185072,
-0.1335250437259674,
-0.02767690271139145,
0.0828544870018959,
0.05642374977469444,
-0.15360940992832184,
0.06354877352714539,
-0.05998864024877548,
-0.025411663576960564,
-0.07998380810022354,
-0.036626193672418594,
-0.00012705964036285877,
0.13209623098373413,
-0.2973765730857849,
-0.09720461070537567,
0.15061113238334656,
-0.14727330207824707,
-0.013910443522036076,
0.13978064060211182,
0.03583008423447609,
-0.05290021002292633,
0.0809539258480072,
0.32908040285110474,
-0.02669154480099678,
-0.03706354647874832,
-0.035011496394872665,
0.12676000595092773,
-0.1539485901594162,
-0.1114703118801117,
0.08442046493291855,
-0.04972956329584122,
0.011421691626310349,
0.01475114282220602,
0.09025336056947708,
0.10513971745967865,
-0.016642607748508453,
-0.11816781759262085,
-0.07015377283096313,
-0.06325346976518631,
0.06908552348613739,
-0.03085336647927761,
0.00017471704632043839,
-0.07647177577018738,
-0.002812117338180542,
0.029956970363855362,
0.020288199186325073,
0.03733455017209053,
0.07814943045377731,
-0.08468569070100784,
0.007884971797466278,
0.07875169813632965,
0.004136864561587572,
-0.07424233853816986,
0.052879251539707184,
-0.03117980621755123,
-0.017686888575553894,
0.10106021165847778,
0.0586705282330513,
0.0763435885310173,
-0.053537845611572266,
-0.02206852100789547,
0.010453728958964348,
0.08113359659910202,
0.07039126008749008,
0.010098753497004509,
-0.2281980961561203,
0.07841155678033829,
-0.04415570944547653,
-0.012529093772172928,
0.005883242469280958,
0.01653343439102173,
0.1373090147972107,
0.06356740742921829,
-0.02558819204568863,
0.04956354200839996,
-0.033013422042131424,
-0.006804578006267548,
0.01447339728474617,
0.013948456384241581,
0.08944689482450485,
0.007511379662901163,
-0.1434481143951416,
0.23762771487236023,
-0.02063487097620964,
0.05827820301055908,
0.16489127278327942,
-0.05455051362514496,
0.1187986508011818,
0.061117447912693024,
-0.031928837299346924,
-0.056505683809518814,
0.10163718461990356,
0.031415026634931564,
0.12988026440143585,
-0.0021414789371192455,
0.06356602907180786,
-0.019012806937098503,
-0.03869553655385971,
-0.007663022726774216,
-0.06751442700624466,
-0.03944137692451477,
0.11198440194129944,
-0.007330853492021561,
-0.16084474325180054,
0.0852813795208931,
0.21986934542655945,
-0.07926451414823532,
0.20620208978652954,
-0.07047122716903687,
-0.03434964269399643,
-0.018637273460626602,
-0.062436170876026154,
-0.025717560201883316,
0.09926514327526093,
-0.07218924909830093,
-0.0042004757560789585,
0.07041069120168686,
-0.00838548131287098,
0.033561401069164276,
-0.10604636371135712,
-0.054019127041101456,
-0.033738017082214355,
-0.023134242743253708,
-0.1980949193239212,
0.0655890703201294,
-0.07914240658283234,
0.053586144000291824,
-0.048078425228595734,
-0.1796693205833435,
0.09582934528589249,
-0.04113204777240753,
-0.04829468950629234,
0.09672339260578156,
-0.18099553883075714,
-0.20031633973121643,
-0.08144374191761017,
0.051472458988428116,
-0.0250119436532259,
-0.0438435859978199,
0.10688114166259766,
-0.154110848903656,
0.003085094504058361,
-0.002262434922158718,
0.018510492518544197,
-0.03562604635953903,
0.03346918895840645,
-0.011853564530611038,
-0.0015507154166698456,
-0.023377999663352966,
-0.12879124283790588,
-0.01612759195268154,
-0.05924277380108833,
0.021308278664946556,
0.1307649463415146,
-0.008685196749866009,
0.011266396380960941,
0.1347532421350479,
0.0255938358604908,
0.012470548041164875,
-0.025856995955109596,
0.07712766528129578,
-0.0063236732967197895,
0.005406167823821306,
0.09370675683021545,
-0.08147824555635452,
0.06762120872735977,
0.09689605236053467,
0.03078939951956272,
-0.012301148846745491,
-0.0036492827348411083,
-0.0249185748398304,
-0.11072329431772232,
-0.25078582763671875,
-0.08511092513799667,
-0.0353110209107399,
0.16134227812290192,
-0.015982134267687798,
0.06407089531421661,
0.07735896110534668,
0.005045383237302303,
0.07892639935016632,
-0.11263807862997055,
0.01914583146572113,
-0.0008859268855303526,
0.2221902310848236,
-0.04734035208821297,
0.11230625957250595,
-0.06481903791427612,
-0.02045087330043316,
0.1211727187037468,
0.09284618496894836,
0.13620537519454956,
0.1136825829744339,
0.11566207557916641,
0.0721355676651001,
0.19752515852451324,
0.09913990646600723,
0.07909317314624786,
0.05828459560871124,
0.03441697731614113,
0.012093550525605679,
-0.05020944029092789,
-0.015524584800004959,
0.06317158788442612,
0.10492084920406342,
-0.1031675636768341,
0.10075171291828156,
0.0346064418554306,
0.08077337592840195,
0.22613804042339325,
0.061994485557079315,
-0.07376567274332047,
0.014972859062254429,
0.03424295783042908,
-0.03065166249871254,
0.028094742447137833,
0.14542467892169952,
0.028966596350073814,
-0.012762444093823433,
0.07594116777181625,
-0.02191336452960968,
0.022266194224357605,
-0.011959102004766464,
0.030802441760897636,
-0.04243350774049759,
-0.11124362051486969,
0.050925612449645996,
0.05514153838157654,
-0.22515079379081726,
0.20457223057746887,
-0.015622676350176334,
-0.016076190397143364,
-0.03274664282798767,
-0.031808532774448395,
0.05455966293811798,
0.07981669902801514,
0.13152150809764862,
0.06594642996788025,
-0.09834402799606323,
0.011795088648796082,
-0.18484456837177277,
0.0789112001657486,
0.09462329000234604,
0.05038155987858772,
-0.10649733245372772,
-0.02351195365190506,
0.025850627571344376,
0.05798051878809929,
-0.0828981101512909,
-0.21324555575847626,
-0.06599409878253937,
0.051851026713848114,
0.19234022498130798,
0.016267433762550354,
-0.020522311329841614,
-0.1450522243976593,
-0.07479250431060791,
0.09856642782688141,
-0.13939440250396729,
-0.08292295783758163,
-0.06182308495044708,
-0.12789562344551086,
0.13508833944797516,
-0.06088065356016159,
0.028409719467163086,
0.0019550109282135963,
-0.066381074488163,
-0.051733843982219696,
-0.17846468091011047,
0.11098723113536835,
-0.07848584651947021,
-0.12493202835321426,
0.023249901831150055,
0.12574993073940277,
0.05630917102098465,
0.07449319958686829,
0.0006226514233276248,
0.09745419025421143,
-0.0933104157447815,
-0.08482041954994202,
0.10120019316673279,
-0.02326878532767296,
-0.07510068267583847,
0.0026262281462550163,
0.017197731882333755,
-0.1086815893650055,
-0.029801541939377785,
-0.029741330072283745,
0.26513946056365967,
0.2669987380504608,
-0.0909537523984909,
0.2705761194229126,
0.26193711161613464,
-0.043350957334041595,
-0.27131420373916626,
-0.15351715683937073,
-0.1365477591753006,
-0.07271701097488403,
0.01663065329194069,
-0.1932041049003601,
0.06609666347503662,
0.0940566286444664,
-0.08955689519643784,
0.103122279047966,
-0.23574478924274445,
-0.10695315897464752,
0.23966282606124878,
-0.11636923253536224,
0.2810322940349579,
-0.11508533358573914,
-0.09393299371004105,
-0.06432218849658966,
-0.012674026191234589,
0.16043274104595184,
-0.14283804595470428,
0.06503663957118988,
-0.02338833548128605,
0.01232925709336996,
0.004681346472352743,
-0.03640098497271538,
0.10726048052310944,
0.0787605568766594,
-0.045639291405677795,
-0.05339784175157547,
-0.0328964926302433,
0.0667746514081955,
-0.0005917632952332497,
0.05452211946249008,
-0.14563657343387604,
0.06152821704745293,
-0.030606089159846306,
-0.01654770039021969,
-0.1172460988163948,
0.11825813353061676,
0.03410654515028,
-0.09356779605150223,
-0.0021695420145988464,
-0.033174045383930206,
0.02261686511337757,
-0.004465071484446526,
0.19595251977443695,
-0.016001712530851364,
-0.05549061298370361,
0.13084469735622406,
0.09578170627355576,
-0.047929778695106506,
0.03782368451356888,
-0.07049311697483063,
-0.09529798477888107,
0.06894277036190033,
-0.08143354952335358,
0.043276987969875336,
0.09312831610441208,
0.05156025290489197,
-0.012010463513433933,
0.07226061075925827,
-0.055778197944164276,
-0.00023127463646233082,
0.08877293765544891,
-0.15551337599754333,
-0.052338287234306335,
0.01868550106883049,
0.032458771020174026,
0.1370583176612854,
0.16822269558906555,
0.1889338493347168,
-0.02341107465326786,
-0.015031300485134125,
-0.01472497545182705,
0.07509530335664749,
-0.09525543451309204,
0.044571653008461,
0.15734584629535675,
-0.00654529919847846,
-0.135562002658844,
0.13359549641609192,
-0.06466805189847946,
-0.016468582674860954,
0.04379289597272873,
0.05565175786614418,
-0.10707290470600128,
-0.1104918122291565,
-0.09491173177957535,
-0.05776533484458923,
-0.12917909026145935,
-0.11286789178848267,
-0.007838161662220955,
-0.0813290923833847,
0.026621365919709206,
0.1753275841474533,
0.04951454699039459,
0.07509913295507431,
-0.04569308087229729,
-0.06147531792521477,
0.06482944637537003,
0.058677203953266144,
-0.1234220415353775,
-0.021951917558908463,
-0.08187119662761688,
-0.020893801003694534,
-0.01740695908665657,
0.08591844886541367,
-0.06704383343458176,
-0.030348528176546097,
-0.09554983675479889,
0.034303516149520874,
-0.047013744711875916,
-0.032984066754579544,
-0.04857245087623596,
0.022611942142248154,
0.031791508197784424,
-0.08699704706668854,
-0.014085481874644756,
0.0653146356344223,
-0.09223891794681549,
-0.003986412193626165,
-0.0048046233132481575,
0.04389054328203201,
-0.1976051926612854,
-0.0456167533993721,
0.058134954422712326,
0.021636705845594406,
0.16555781662464142,
0.1256776750087738,
-0.07660824805498123,
0.11742395162582397,
-0.09869498759508133,
-0.02244754135608673,
0.1117674931883812,
0.036437567323446274,
0.002997276373207569,
0.044210243970155716,
-0.0440310463309288,
0.05848920717835426,
0.03338479623198509,
0.07383851706981659,
0.13101662695407867,
-0.07768823951482773,
0.030128369107842445,
-0.042837031185626984,
-0.12950456142425537,
-0.032028310000896454,
-0.10896497964859009,
0.05543706566095352,
0.10185205191373825,
0.16718822717666626,
-0.03868614509701729,
-0.016639648005366325,
0.016077576205134392,
0.02386396750807762,
0.02848334051668644,
-0.087058424949646,
-0.044038332998752594,
0.03010496310889721,
-0.018386367708444595,
-0.035880256444215775,
0.21487922966480255,
-0.026319758966565132,
-0.17064960300922394,
0.044613003730773926,
-0.020065510645508766,
-0.12587395310401917,
0.03674069419503212,
0.2626139521598816,
0.038719795644283295,
-0.012884745374321938,
-0.08493516594171524,
-0.01828238181769848,
-0.05413510650396347,
-0.09558413922786713,
0.06751565635204315,
0.06801538169384003,
0.09223853051662445,
0.06254774332046509,
0.12975148856639862,
-0.025546874850988388,
-0.05470895767211914,
-0.05177507549524307,
-0.05686228722333908,
0.06639227271080017,
0.027699220925569534,
0.209966778755188,
0.12033633142709732,
0.028490228578448296,
-0.006040135398507118,
-0.023606663569808006,
-0.060566723346710205,
-0.18969568610191345,
-0.12327016890048981,
-0.03732560947537422,
-0.07416074723005295,
0.0384359285235405,
-0.030362924560904503,
0.05626482516527176,
0.12329961359500885,
0.07700580358505249,
-0.0009011514484882355,
-0.06464418023824692,
-0.05121458321809769,
-0.0001124790869653225,
0.0348195806145668,
-0.057012204080820084,
-0.0755947157740593,
-0.10105768591165543,
-0.04394053667783737,
0.015768691897392273,
0.03133382648229599,
-0.07752838730812073,
-0.008578682318329811,
-0.07873491197824478,
-0.017232496291399002,
-0.10239437222480774,
-0.050892915576696396,
-0.03472457453608513,
0.024232419207692146,
0.043623678386211395,
0.09894315898418427,
0.10073515772819519,
-0.03718595206737518,
0.08316752314567566,
0.11457402259111404,
-0.031032219529151917,
-0.18016408383846283,
-0.012126639485359192,
0.07565637677907944,
-0.05338020995259285,
0.10025064647197723,
-0.03842099756002426,
-0.0558636412024498,
-0.05840787664055824,
0.07573410868644714,
0.2607494592666626,
-0.0174150038510561,
0.04491695761680603,
-0.04660075157880783,
0.024686288088560104,
0.002875139005482197,
0.018702974542975426,
0.13356642425060272,
0.12520143389701843,
-0.06304773688316345,
-0.05177099257707596,
-0.05929160863161087,
-0.019041750580072403,
-0.12145159393548965,
0.011404255405068398,
-0.06015898287296295,
-0.0897805467247963,
-0.07275954633951187,
0.12121274322271347,
-0.1598210483789444,
0.058650076389312744,
-0.055437929928302765,
-0.1029277965426445,
-0.03295855224132538,
-0.009737720713019371,
0.1659109741449356,
0.033398956060409546,
0.04384378343820572,
-0.03790423646569252,
-0.08001556992530823,
-0.011370447464287281,
0.037753403186798096,
-0.25237637758255005,
0.03721905127167702,
0.03265601396560669,
-0.07988566160202026,
0.09082521498203278,
-0.016132451593875885,
0.16533517837524414,
0.07192424684762955,
0.06218798831105232,
-0.05425003916025162,
0.08918782323598862,
0.016535332426428795,
-0.05450943857431412,
-0.08191263675689697,
-0.004588240757584572,
0.032828230410814285,
-0.005491178948432207,
0.0703296959400177,
0.06606639176607132,
0.02495018020272255,
0.0801570937037468,
-0.02230537123978138,
-0.05728645995259285,
0.05666109174489975,
-0.05386669188737869,
0.03400818258523941,
-0.03569469600915909,
-0.03584764152765274,
-0.04073994234204292,
-0.055036261677742004,
-0.05986405164003372,
0.009979198686778545,
-0.20742474496364594,
-0.052534062415361404,
0.040035542100667953,
-0.022007115185260773,
0.009698151610791683,
0.04001221805810928,
-0.057599470019340515,
-0.04128093644976616,
-0.16891920566558838,
0.008959856815636158,
-0.0872000902891159,
0.00420183502137661,
0.02597952075302601,
-0.00715241115540266,
0.04049311578273773,
-0.1003064215183258,
0.0926964282989502,
-0.0023711728863418102,
-0.15293952822685242,
-0.06006823107600212
] |
null | null | null |
## Pre-trained TDNN models for the yesno dataset with icefall.
Refer to <https://github.com/k2-fsa/icefall/tree/master/egs/yesno/ASR>
for more information about this pre-trained model.
You can find usage instructions there.
## Sound files for testing the pre-trained model
The folder `test_waves` contains test sound files. They
are downloaded from <https://www.openslr.org/1/>.
There are 60 files in the dataset, 30 are used for training.
The remaining 30 files, contained in `test_waves` are kept for testing.
The code for splitting the dataset can be found at
<https://github.com/lhotse-speech/lhotse/blob/master/lhotse/recipes/yesno.py#L138>
```python
wave_files = list(corpus_dir.glob("*.wav"))
assert len(wave_files) == 60
wave_files.sort()
train_set = wave_files[::2]
test_set = wave_files[1::2]
assert len(train_set) == 30
assert len(test_set) == 30
```
|
{}
| null |
csukuangfj/icefall_asr_yesno_tdnn
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
## Pre-trained TDNN models for the yesno dataset with icefall.
Refer to <URL
for more information about this pre-trained model.
You can find usage instructions there.
## Sound files for testing the pre-trained model
The folder 'test_waves' contains test sound files. They
are downloaded from <URL
There are 60 files in the dataset, 30 are used for training.
The remaining 30 files, contained in 'test_waves' are kept for testing.
The code for splitting the dataset can be found at
<URL
|
[
"## Pre-trained TDNN models for the yesno dataset with icefall.\n\nRefer to <URL\nfor more information about this pre-trained model.\n\nYou can find usage instructions there.",
"## Sound files for testing the pre-trained model\n\nThe folder 'test_waves' contains test sound files. They\nare downloaded from <URL\n\nThere are 60 files in the dataset, 30 are used for training.\nThe remaining 30 files, contained in 'test_waves' are kept for testing.\n\nThe code for splitting the dataset can be found at\n<URL"
] |
[
"TAGS\n#region-us \n",
"## Pre-trained TDNN models for the yesno dataset with icefall.\n\nRefer to <URL\nfor more information about this pre-trained model.\n\nYou can find usage instructions there.",
"## Sound files for testing the pre-trained model\n\nThe folder 'test_waves' contains test sound files. They\nare downloaded from <URL\n\nThere are 60 files in the dataset, 30 are used for training.\nThe remaining 30 files, contained in 'test_waves' are kept for testing.\n\nThe code for splitting the dataset can be found at\n<URL"
] |
[
6,
41,
81
] |
[
"passage: TAGS\n#region-us \n## Pre-trained TDNN models for the yesno dataset with icefall.\n\nRefer to <URL\nfor more information about this pre-trained model.\n\nYou can find usage instructions there.## Sound files for testing the pre-trained model\n\nThe folder 'test_waves' contains test sound files. They\nare downloaded from <URL\n\nThere are 60 files in the dataset, 30 are used for training.\nThe remaining 30 files, contained in 'test_waves' are kept for testing.\n\nThe code for splitting the dataset can be found at\n<URL"
] |
[
-0.1880803406238556,
0.03516754135489464,
0.0037605923134833574,
0.08916329592466354,
0.09411059319972992,
-0.0985790342092514,
0.13371308147907257,
-0.038229405879974365,
0.045969851315021515,
0.05418291687965393,
0.04140839725732803,
-0.01079460047185421,
-0.0759991854429245,
0.10129813104867935,
0.051073409616947174,
-0.2129470705986023,
0.025276580825448036,
-0.06405456364154816,
-0.12554752826690674,
0.09399820864200592,
0.016470279544591904,
-0.05874006077647209,
-0.08588943630456924,
0.01319935917854309,
-0.03758620098233223,
-0.004969341214746237,
-0.1254567950963974,
-0.00523633835837245,
0.045211952179670334,
-0.14443345367908478,
0.10328345000743866,
0.1271306425333023,
0.06523473560810089,
-0.13005033135414124,
0.049243975430727005,
0.031092120334506035,
0.030909419059753418,
0.02750231884419918,
0.09619278460741043,
0.017562050372362137,
-0.06475088745355606,
0.09874220192432404,
0.02966049499809742,
0.006916831247508526,
-0.04515824466943741,
-0.07976207882165909,
0.019393568858504295,
-0.19507046043872833,
0.10158983618021011,
0.16218900680541992,
-0.08040091395378113,
0.21530915796756744,
-0.17462323606014252,
0.11686546355485916,
0.10638821870088577,
-0.053615741431713104,
-0.03134378045797348,
0.06319116801023483,
0.06173056364059448,
0.09945931285619736,
0.022796690464019775,
0.07296309620141983,
0.1229342669248581,
-0.0019071371061727405,
-0.19702036678791046,
0.06646496057510376,
0.03056703694164753,
-0.011435262858867645,
-0.17553530633449554,
0.05971093848347664,
0.28510934114456177,
0.04556708037853241,
-0.07942427694797516,
0.04922850802540779,
-0.006346989888697863,
-0.11432261765003204,
0.0782657116651535,
0.009105861186981201,
0.027147209271788597,
-0.0238703154027462,
-0.14547686278820038,
-0.016239160671830177,
-0.12616929411888123,
-0.07100939750671387,
0.04357782378792763,
0.14047232270240784,
-0.0725814551115036,
0.030603397637605667,
-0.12209223955869675,
0.04829442873597145,
-0.05621516704559326,
-0.05323147401213646,
-0.037545859813690186,
-0.056509748101234436,
-0.06067828834056854,
-0.030005771666765213,
-0.1612614244222641,
-0.15554024279117584,
0.12327475100755692,
0.17358329892158508,
0.10815785825252533,
0.024381788447499275,
-0.05290433391928673,
0.08120643347501755,
0.09849200397729874,
-0.04396292567253113,
0.08815259486436844,
-0.11450499296188354,
-0.045963354408741,
-0.06195472180843353,
-0.01931038498878479,
-0.10564753413200378,
-0.1013089194893837,
-0.0666094496846199,
0.06501788645982742,
0.023261724039912224,
0.11363302916288376,
-0.1631072610616684,
0.004912975709885359,
-0.06561914086341858,
0.03409762308001518,
0.013484282419085503,
-0.0028537895996123552,
0.036633506417274475,
0.02478090114891529,
-0.05373018980026245,
-0.003472196636721492,
0.1531943380832672,
0.06860754638910294,
0.011965225450694561,
-0.021414926275610924,
0.05217456817626953,
-0.053679611533880234,
-0.07200586795806885,
0.041585225611925125,
-0.0997256487607956,
0.02167610079050064,
-0.08497106283903122,
-0.13500086963176727,
-0.12684567272663116,
0.060020171105861664,
0.05123390629887581,
-0.01448065135627985,
-0.09673312306404114,
-0.03199588879942894,
-0.06993341445922852,
-0.016200626268982887,
0.07634968310594559,
-0.06432810425758362,
-0.004888932686299086,
-0.0854637399315834,
0.08704666048288345,
-0.043007053434848785,
-0.025861017405986786,
-0.08532154560089111,
-0.015161080285906792,
0.078467957675457,
0.09866147488355637,
-0.11619049310684204,
-0.052058856934309006,
-0.018256470561027527,
-0.09885408729314804,
-0.17820176482200623,
0.035835590213537216,
0.16481859982013702,
0.20083700120449066,
-0.36362218856811523,
-0.007617845665663481,
-0.06635487824678421,
-0.1795448213815689,
0.07191967219114304,
0.10408735275268555,
-0.05125323310494423,
0.05859067291021347,
0.04766402393579483,
0.19982051849365234,
0.01621687412261963,
-0.07900933176279068,
0.10755137354135513,
-0.10098730772733688,
0.025282809510827065,
-0.17718638479709625,
0.07351180166006088,
-0.054028116166591644,
-0.20390821993350983,
0.001011505490168929,
0.023509813472628593,
-0.008746406994760036,
-0.10207290202379227,
-0.05277483910322189,
0.0026974030770361423,
-0.06470954418182373,
-0.12710054218769073,
0.03374272212386131,
-0.019253689795732498,
-0.10906028002500534,
-0.01989438384771347,
-0.009530546143651009,
-0.006218917667865753,
-0.02464672364294529,
0.06190510839223862,
0.07517065852880478,
0.15349486470222473,
-0.20044094324111938,
-0.1382375806570053,
-0.13257543742656708,
-0.052933841943740845,
-0.027448363602161407,
0.10647639632225037,
0.07040469348430634,
0.1413593739271164,
0.09919121116399765,
0.07920893281698227,
-0.0024474537931382656,
0.028995700180530548,
-0.007044538855552673,
0.054056670516729355,
-0.00548539636656642,
-0.15622587502002716,
0.061291586607694626,
-0.07371541857719421,
0.10046681016683578,
-0.1724107414484024,
-0.053808823227882385,
0.029040802270174026,
0.0698317363858223,
-0.046113040298223495,
-0.09723630547523499,
0.09477979689836502,
0.034482475370168686,
-0.0635899007320404,
-0.004778503440320492,
0.006865119561553001,
0.034035440534353256,
-0.09058579057455063,
0.026914848014712334,
0.06114351004362106,
-0.13840840756893158,
0.057356785982847214,
0.059259843081235886,
-0.05729283019900322,
-0.06367228180170059,
-0.11714624613523483,
-0.03644217923283577,
-0.042505886405706406,
0.18368352949619293,
-0.03979511931538582,
-0.023941513150930405,
0.01548993494361639,
-0.03459645062685013,
0.05124691128730774,
0.06690623611211777,
-0.0793534442782402,
-0.028981804847717285,
0.11116449534893036,
0.08603693544864655,
-0.07380107790231705,
-0.0060881213285028934,
-0.01457975897938013,
-0.17614060640335083,
0.23533286154270172,
-0.022432640194892883,
-0.16989192366600037,
-0.043786972761154175,
0.042748309671878815,
0.008515702560544014,
0.16809335350990295,
0.04763534665107727,
0.186717689037323,
0.011754332110285759,
0.04757480323314667,
0.12989182770252228,
-0.11772117763757706,
-0.0543564073741436,
0.06520412117242813,
0.08714562654495239,
-0.2310389131307602,
0.016454454511404037,
-0.1555488407611847,
-0.022673429921269417,
-0.08689798414707184,
-0.10128414630889893,
0.023664414882659912,
0.010486166924238205,
-0.050046712160110474,
0.1622432917356491,
-0.09719298779964447,
-0.091519296169281,
-0.13970860838890076,
0.16134384274482727,
-0.018023362383246422,
-0.03349437192082405,
0.13703522086143494,
-0.24961186945438385,
-0.03376070410013199,
-0.06807844340801239,
-0.03244219347834587,
-0.07880979776382446,
0.025089535862207413,
-0.007128856610506773,
-0.022533245384693146,
-0.05777993053197861,
-0.05706455931067467,
0.07653402537107468,
-0.08401590585708618,
0.08929071575403214,
0.054533835500478745,
-0.07880259305238724,
0.07826261222362518,
0.12880447506904602,
0.024671407416462898,
0.022014161571860313,
0.0667610689997673,
0.20953930914402008,
-0.10072354972362518,
0.06827656179666519,
0.008020239882171154,
-0.05179929733276367,
-0.0038393139839172363,
-0.014092104509472847,
-0.0017098176758736372,
-0.0670737773180008,
0.017380382865667343,
-0.021422309800982475,
-0.15908962488174438,
-0.2667270600795746,
-0.0895099937915802,
-0.015691068023443222,
-0.04498283937573433,
-0.024022771045565605,
0.07066603749990463,
0.09906835854053497,
0.08853377401828766,
0.11569865792989731,
-0.25246796011924744,
-0.0473039485514164,
0.049667470157146454,
-0.07171810418367386,
0.035767633467912674,
0.10974135994911194,
-0.034038037061691284,
0.04686891660094261,
0.044859398156404495,
0.00836525484919548,
0.18976840376853943,
0.05605759099125862,
0.030783560127019882,
0.07689478248357773,
-0.020121131092309952,
0.10279682278633118,
0.16104456782341003,
0.06945963203907013,
0.017502587288618088,
0.030026348307728767,
-0.0638400986790657,
0.06764572113752365,
0.0670454204082489,
0.16118542850017548,
-0.1364947408437729,
0.08045457303524017,
0.08626628667116165,
-0.12928782403469086,
0.09292671084403992,
0.08071666955947876,
-0.21918855607509613,
-0.07121524959802628,
-0.055910952389240265,
0.07025707513093948,
-0.04964647814631462,
0.14251098036766052,
0.11859282851219177,
0.050527773797512054,
0.013012445531785488,
-0.013436323963105679,
0.12219846248626709,
0.11537163704633713,
-0.05677303671836853,
0.03342222794890404,
0.0616806223988533,
-0.05917765572667122,
-0.003161888336762786,
-0.040345776826143265,
0.08576057106256485,
-0.05140453949570656,
0.020687052980065346,
0.023112133145332336,
-0.04934481531381607,
-0.08199036866426468,
0.10058603435754776,
0.14929983019828796,
0.026758620515465736,
-0.11579558253288269,
0.0778113454580307,
-0.18934635818004608,
-0.016071220859885216,
0.07397694885730743,
-0.019250618293881416,
-0.012560996226966381,
0.09078111499547958,
0.019789038226008415,
0.04325196519494057,
-0.08889564126729965,
-0.24859343469142914,
-0.008485395461320877,
-0.01316903904080391,
0.08529143780469894,
0.015116185881197453,
-0.0010972145246341825,
-0.2105419933795929,
-0.08042071014642715,
0.12412706762552261,
-0.04023183137178421,
0.04051656275987625,
-0.08928649127483368,
-0.022309886291623116,
0.26249682903289795,
0.0955180898308754,
0.05500561371445656,
0.016650522127747536,
0.007031741086393595,
-0.04914754629135132,
0.037465110421180725,
0.04465643689036369,
-0.13681164383888245,
-0.07299251854419708,
-0.06772947311401367,
0.09234361350536346,
0.10216247290372849,
0.02191145345568657,
-0.03767137601971626,
0.1474452167749405,
-0.11260522902011871,
-0.009633632376790047,
0.01889931410551071,
-0.06072483956813812,
-0.03360285237431526,
0.07160796225070953,
0.03605387732386589,
0.10475824773311615,
-0.048862241208553314,
-0.061500951647758484,
0.09422484040260315,
0.23554688692092896,
-0.03771080821752548,
0.06876280158758163,
0.1897028088569641,
-0.019002219662070274,
-0.13525304198265076,
0.191705122590065,
0.08961044996976852,
0.07430620491504669,
-0.020162593573331833,
-0.14702066779136658,
0.17199602723121643,
0.04296516999602318,
0.004227552097290754,
0.22108587622642517,
0.07901644706726074,
-0.13058985769748688,
0.14348483085632324,
-0.004469249863177538,
0.2310977280139923,
-0.01216872502118349,
-0.014167473651468754,
-0.018301136791706085,
0.1286603957414627,
0.11279963701963425,
-0.11537078022956848,
0.09106860309839249,
0.07014113664627075,
-0.0375753678381443,
0.0597432516515255,
-0.11299272626638412,
0.09872599691152573,
0.046176739037036896,
-0.03847417235374451,
-0.008420834317803383,
0.13115547597408295,
0.05301862210035324,
-0.027423234656453133,
0.04892246797680855,
0.1281399428844452,
0.08195583522319794,
0.10591867566108704,
-0.1381068229675293,
0.005175342317670584,
0.10645557194948196,
-0.04958011209964752,
-0.10977187007665634,
0.004706472624093294,
0.05425455793738365,
0.0029812308494001627,
-0.01708311401307583,
0.11712510883808136,
-0.06945076584815979,
0.05598149821162224,
0.14037470519542694,
0.20128558576107025,
0.19222602248191833,
-0.03411080315709114,
0.058024872094392776,
-0.11026100069284439,
0.07553450763225555,
-0.07116461545228958,
-0.05156226083636284,
0.07755260169506073,
0.11373381316661835,
-0.10926572233438492,
0.038273558020591736,
-0.11769787967205048,
-0.07355311512947083,
0.06193762645125389,
-0.09158773720264435,
-0.15664330124855042,
-0.0717846155166626,
-0.18077850341796875,
0.12368230521678925,
0.10038097947835922,
0.0938948392868042,
-0.10368628054857254,
0.05702895671129227,
-0.012828623875975609,
-0.0023148648906499147,
-0.0397268570959568,
0.1345454752445221,
0.1529543101787567,
0.0269547738134861,
-0.0898466631770134,
0.013664951547980309,
-0.06783914566040039,
-0.010830731131136417,
0.09350378066301346,
-0.016349591314792633,
-0.046742185950279236,
-0.07364950329065323,
0.05130678787827492,
0.14223171770572662,
-0.017420535907149315,
-0.12675057351589203,
-0.12660536170005798,
0.1294516623020172,
0.028934458270668983,
0.0719735324382782,
0.02500183694064617,
-0.06526338309049606,
-0.04762294143438339,
0.02275496907532215,
-0.11612140387296677,
0.05885294824838638,
0.012937248684465885,
0.08760103583335876,
-0.05749513581395149,
-0.07170621305704117,
-0.061339084059000015,
0.003962191753089428,
-0.06954706460237503,
-0.09226793050765991,
0.0073660556226968765,
0.0627317950129509,
-0.06898543983697891,
-0.028346531093120575,
0.03314107283949852,
0.016234194859862328,
0.021220538765192032,
-0.09065773338079453,
-0.07539457827806473,
0.09454125910997391,
-0.08111590147018433,
0.09823060780763626,
-0.04252798855304718,
-0.0863637924194336,
-0.05018015578389168,
0.03791600093245506,
-0.011547812260687351,
-0.053234945982694626,
0.0000036145959256828064,
0.0360046848654747,
-0.024569328874349594,
0.037071771919727325,
-0.05516587197780609,
0.0641762763261795,
0.0377277173101902,
-0.008358045481145382,
-0.00594691326841712,
-0.010121333412826061,
-0.04464638978242874,
0.03455571457743645,
0.09581062197685242,
-0.04050808027386665,
0.115634486079216,
-0.023880261927843094,
-0.123137928545475,
-0.003963941242545843,
-0.12060756236314774,
-0.05525849014520645,
-0.04438696429133415,
0.001304642646573484,
0.23423050343990326,
0.17397119104862213,
0.01649363338947296,
0.1338963806629181,
0.0014619898283854127,
-0.07445178925991058,
0.04637031629681587,
0.011198535561561584,
-0.06354699283838272,
0.02441956289112568,
0.07738718390464783,
-0.06741291284561157,
0.1058025062084198,
-0.05187036842107773,
0.0721098855137825,
-0.08073976635932922,
-0.012541322037577629,
0.007985328324139118,
0.011652573011815548,
0.2823854386806488,
0.06548558175563812,
0.05258315056562424,
-0.14060966670513153,
0.03982189670205116,
0.008712248876690865,
-0.0011668190127238631,
0.044497568160295486,
0.11069953441619873,
-0.02619631215929985,
0.03823019936680794,
0.05197237432003021,
-0.04645945131778717,
-0.05290082097053528,
0.04686750844120979,
-0.04828691855072975,
0.04973731189966202,
0.0652984157204628,
0.03169749304652214,
0.05563609302043915,
-0.025718389078974724,
-0.0008987401961348951,
0.0228491872549057,
-0.05863450467586517,
-0.1766359657049179,
-0.08641547709703445,
-0.08577493578195572,
-0.05892852321267128,
0.11976355314254761,
-0.05838410183787346,
-0.015348201617598534,
-0.018019163981080055,
0.11874111741781235,
0.034453075379133224,
0.20500612258911133,
0.10804431885480881,
-0.017237462103366852,
-0.023363828659057617,
-0.06590338796377182,
-0.027652887627482414,
-0.027114208787679672,
-0.12261044979095459,
0.09802374988794327,
-0.12005537003278732,
-0.037191472947597504,
-0.04824597388505936,
0.008904218673706055,
0.040346112102270126,
0.05049528554081917,
-0.06990770995616913,
0.015580754727125168,
0.0403723306953907,
0.04036719724535942,
0.03236125037074089,
0.056287918239831924,
-0.051328882575035095,
0.06142207607626915,
0.01002642884850502,
0.04683253541588783,
-0.15586970746517181,
-0.08769048750400543,
0.1915932446718216,
-0.204417884349823,
0.0751795694231987,
-0.041238024830818176,
-0.09567756205797195,
-0.01619727723300457,
0.023160217329859734,
0.13313907384872437,
0.00657283142209053,
-0.06472329795360565,
0.11280492693185806,
-0.023473823443055153,
-0.12871691584587097,
0.04598352685570717,
-0.03098123148083687,
0.09579265862703323,
-0.020874805748462677,
-0.0641927421092987,
-0.1415039300918579,
-0.07586441934108734,
-0.005609135143458843,
0.02936459332704544,
0.03622617945075035,
-0.11949607729911804,
-0.1379450559616089,
0.2502917945384979,
-0.10771580785512924,
-0.06893083453178406,
0.04288943484425545,
-0.050828780978918076,
-0.1614222675561905,
-0.10655736923217773,
0.05694383382797241,
-0.09830833971500397,
0.0017199902795255184,
-0.09569007903337479,
0.005520959384739399,
-0.1378660500049591,
-0.027344709262251854,
-0.002576518803834915,
0.04759492725133896,
0.006811853963881731,
0.0006127515225671232,
-0.046690210700035095,
0.0036663522478193045,
0.19717122614383698,
0.054321348667144775,
0.07485773414373398,
-0.01776333712041378,
0.10638520121574402,
0.0383731983602047,
-0.10454320162534714,
-0.05723842605948448,
0.10190889239311218,
-0.05262109637260437,
0.02690570428967476,
0.0315016470849514,
-0.13123682141304016,
0.027081554755568504,
0.040012557059526443,
-0.014203151687979698,
-0.10656747967004776,
0.11120550334453583,
-0.07360580563545227,
0.089621901512146,
-0.04234406352043152,
-0.012779094278812408,
-0.0547390878200531,
-0.09301595389842987,
0.11870621144771576,
0.021647432819008827,
-0.06330536305904388,
0.030716950073838234,
-0.10112572461366653,
-0.05946144834160805,
0.040822140872478485,
-0.01488802582025528,
-0.1946020871400833,
-0.020478660240769386,
-0.01247361395508051,
-0.08905453979969025,
0.08287765085697174,
-0.044822197407484055,
0.08992630243301392,
-0.006076700985431671,
0.008891244418919086,
0.14491501450538635,
0.05614021420478821,
0.021667158231139183,
-0.1251002699136734,
-0.08228767663240433
] |
null | null | null |
See
https://colab.research.google.com/drive/14MozS-9jWD3XQ0o-dZ-meqnblgHs70P2?usp=sharing
|
{}
| null |
csukuangfj/test-data-for-optimized-transducer
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
See
URL
|
[] |
[
"TAGS\n#region-us \n"
] |
[
6
] |
[
"passage: TAGS\n#region-us \n"
] |
[
0.024608636274933815,
-0.026205500587821007,
-0.009666500613093376,
-0.10395516455173492,
0.08638657629489899,
0.059816278517246246,
0.01882290467619896,
0.020661840215325356,
0.23975107073783875,
-0.005599027033895254,
0.1219947561621666,
0.0015615287702530622,
-0.037353623658418655,
0.03733762726187706,
-0.0035912662278860807,
-0.17583473026752472,
0.03876631706953049,
-0.018274923786520958,
0.01843859627842903,
0.026470553129911423,
-0.07776834815740585,
-0.07564429938793182,
0.015296397730708122,
-0.10247814655303955,
-0.083692267537117,
0.11002834886312485,
0.031466204673051834,
-0.019670886918902397,
0.10779199749231339,
-0.04243955761194229,
0.18699054419994354,
-0.011512263678014278,
-0.11213519424200058,
-0.2536850869655609,
0.021806683391332626,
-0.01765260472893715,
-0.08747660368680954,
0.01506110467016697,
0.0665089413523674,
-0.09014441072940826,
-0.0588928684592247,
0.0795099288225174,
-0.01132340170443058,
0.04246443510055542,
-0.27593839168548584,
-0.12684126198291779,
-0.05297930911183357,
-0.1421966552734375,
0.08651168644428253,
0.04035491496324539,
0.008764253929257393,
0.15506891906261444,
-0.20897391438484192,
0.004104613792151213,
0.08255259692668915,
-0.2538507878780365,
0.05591634660959244,
0.17671173810958862,
0.03623908758163452,
0.18037272989749908,
0.0060391901060938835,
0.11029672622680664,
0.0716743916273117,
-0.024263937026262283,
-0.17590197920799255,
-0.08127854019403458,
-0.04696211963891983,
0.16642488539218903,
-0.06727185100317001,
-0.14248386025428772,
0.34701237082481384,
0.00015008423360995948,
0.009657775051891804,
0.16921205818653107,
-0.059524230659008026,
-0.09972117841243744,
0.07259953022003174,
0.016484731808304787,
0.018492350354790688,
0.1471305936574936,
0.16307872533798218,
-0.0458691343665123,
-0.13837823271751404,
-0.018630273640155792,
-0.22798998653888702,
0.17510560154914856,
-0.03248048573732376,
0.13137903809547424,
-0.27447956800460815,
0.01684025302529335,
-0.2570667266845703,
0.0032130838371813297,
0.04178816080093384,
-0.06004921346902847,
-0.0226522795855999,
-0.013265985064208508,
-0.08018817007541656,
0.004899587947875261,
0.06192673370242119,
0.1266920566558838,
-0.06128726154565811,
0.06128238886594772,
-0.09319206327199936,
0.141696035861969,
0.07166698575019836,
0.07868369668722153,
0.13037432730197906,
0.041205424815416336,
-0.07187089323997498,
-0.21872246265411377,
-0.0026476888451725245,
-0.06275863200426102,
-0.09502086788415909,
-0.0020165652967989445,
-0.11606067419052124,
0.17244569957256317,
-0.030802514404058456,
-0.09825427830219269,
-0.11208184063434601,
0.09148659557104111,
-0.032992321997880936,
-0.03437839448451996,
-0.03552987426519394,
-0.020977836102247238,
0.019381176680326462,
0.04704452306032181,
-0.1548958420753479,
-0.005131472367793322,
0.07039852440357208,
0.11502562463283539,
-0.1346137970685959,
-0.003783059772104025,
-0.07908964157104492,
0.03039063885807991,
0.07654735445976257,
-0.16510222852230072,
0.03158547356724739,
-0.1124754324555397,
-0.07531405985355377,
0.002912673633545637,
-0.015710093080997467,
-0.016202643513679504,
0.166526660323143,
-0.0020451415330171585,
0.0714716836810112,
-0.026345307007431984,
-0.05890209600329399,
-0.11243434250354767,
-0.08489254862070084,
0.05390460044145584,
0.03670717030763626,
0.03266148269176483,
-0.2193479984998703,
0.014805203303694725,
-0.12762966752052307,
0.1360815018415451,
-0.10566820204257965,
-0.04705966264009476,
-0.022842247039079666,
0.20562705397605896,
0.037286072969436646,
0.08762791007757187,
-0.22171171009540558,
0.039756543934345245,
-0.05404696613550186,
0.18480908870697021,
-0.1502426266670227,
-0.0799463614821434,
0.20813211798667908,
-0.07964949309825897,
-0.10115210711956024,
0.021235812455415726,
0.020391687750816345,
0.026287272572517395,
0.0766737088561058,
0.4564172327518463,
-0.09766800701618195,
-0.09146861732006073,
0.10178250074386597,
0.17055274546146393,
-0.12427149713039398,
-0.1827561855316162,
0.06446871906518936,
-0.16666454076766968,
-0.1973118633031845,
0.0018917324487119913,
0.09222044050693512,
0.038269978016614914,
-0.07875611633062363,
-0.020746968686580658,
0.06325206160545349,
-0.0007678253459744155,
0.09095914661884308,
0.03755716234445572,
0.09034032374620438,
-0.08716782182455063,
0.11115926504135132,
-0.05017651244997978,
0.004037132486701012,
0.1343354731798172,
0.027325427159667015,
-0.03223329409956932,
0.08694463223218918,
-0.0485352948307991,
0.05295134335756302,
-0.1662379503250122,
-0.15068690478801727,
0.03398871049284935,
0.06283251196146011,
0.03186952322721481,
0.1280253529548645,
0.08141885697841644,
-0.10732853412628174,
0.022690722718834877,
-0.004228927195072174,
0.058398615568876266,
0.03891623765230179,
0.006107209715992212,
0.008764320984482765,
0.0961301177740097,
-0.10607069730758667,
-0.13589619100093842,
-0.07336436957120895,
-0.014715781435370445,
0.14371353387832642,
-0.0302802175283432,
0.07690227776765823,
-0.004240254405885935,
0.00013200697139836848,
0.06930823624134064,
0.08137880265712738,
0.016412746161222458,
0.08971183747053146,
-0.05237193778157234,
-0.05160155147314072,
0.10863113403320312,
-0.13533565402030945,
0.17837053537368774,
0.14053137600421906,
-0.20532016456127167,
0.029453208670020103,
-0.06838275492191315,
0.03670361638069153,
-0.008162540383636951,
0.0975119024515152,
-0.08272241055965424,
-0.02106042578816414,
0.013134466484189034,
0.0052274600602686405,
-0.013007243163883686,
0.017682146281003952,
-0.07295988500118256,
-0.07787393033504486,
-0.10233919322490692,
0.08436838537454605,
0.11562882363796234,
-0.10282530635595322,
0.14214380085468292,
0.4384984076023102,
0.11495281755924225,
0.21582984924316406,
-0.09581480920314789,
-0.0412987545132637,
0.007486371789127588,
0.0001535322517156601,
-0.04476691037416458,
0.08031861484050751,
-0.15973517298698425,
-0.038901735097169876,
0.027348900213837624,
0.07128690183162689,
0.11475157737731934,
-0.14959022402763367,
-0.09639324247837067,
-0.00793045200407505,
0.0022841424215584993,
-0.1249532699584961,
0.023905446752905846,
-0.03974650055170059,
0.04015624523162842,
0.07232289016246796,
-0.021535737439990044,
0.13939237594604492,
-0.04166141897439957,
-0.0639561116695404,
0.07585346698760986,
-0.2017085999250412,
-0.23179671168327332,
-0.12309670448303223,
-0.14680525660514832,
0.04366797208786011,
0.05154111236333847,
0.01726446859538555,
-0.17635835707187653,
-0.015074856579303741,
0.07706750929355621,
0.07820965349674225,
-0.20886357128620148,
-0.022814949974417686,
-0.004290030337870121,
0.0895976573228836,
-0.10227091610431671,
-0.0017130117630586028,
-0.04419664293527603,
-0.10150232166051865,
0.0017003051470965147,
0.07279510796070099,
-0.137485533952713,
0.13807645440101624,
0.21589438617229462,
0.07225540280342102,
0.07359948754310608,
-0.019093448296189308,
0.09936179965734482,
-0.10856141895055771,
-0.16549113392829895,
0.08348225057125092,
-0.06234746053814888,
0.047262318432331085,
0.17534415423870087,
0.03307317942380905,
-0.13904969394207,
-0.015682822093367577,
-0.0402069091796875,
-0.15603256225585938,
-0.238995760679245,
-0.09178274869918823,
-0.1182505264878273,
0.16442428529262543,
0.0009358620154671371,
0.06651917099952698,
0.08258313685655594,
-0.022042419761419296,
0.16447891294956207,
-0.07379321753978729,
-0.07578866183757782,
-0.006978808436542749,
0.12375060468912125,
-0.056660156697034836,
-0.03080669604241848,
-0.10566964000463486,
-0.008295975625514984,
0.1151021271944046,
0.15304014086723328,
0.12214863300323486,
0.2957419455051422,
0.08268889784812927,
0.026645636186003685,
0.08958091586828232,
0.17622539401054382,
0.09495089203119278,
0.07838419824838638,
-0.045413073152303696,
-0.014814783819019794,
0.014317171648144722,
-0.04022889584302902,
0.010141594335436821,
0.14683100581169128,
-0.2679629921913147,
-0.006678564939647913,
-0.2710230350494385,
0.0965198427438736,
-0.10913380235433578,
0.11837165057659149,
-0.01015760749578476,
0.10194015502929688,
0.11082887649536133,
0.03233652561903,
-0.03858073800802231,
0.16613617539405823,
0.08450309932231903,
-0.11277695000171661,
0.001758623169735074,
0.03737903758883476,
0.09715615212917328,
-0.02818971499800682,
0.12721189856529236,
-0.11048974841833115,
-0.1464834064245224,
0.013753619976341724,
0.07152791321277618,
-0.15373679995536804,
0.3138748109340668,
0.012069208547472954,
-0.13481520116329193,
-0.01481647603213787,
-0.09957809001207352,
-0.006440147757530212,
0.1254177987575531,
0.09333524852991104,
0.07935678958892822,
-0.2185502052307129,
-0.13339371979236603,
0.05872276425361633,
-0.00575496768578887,
0.22408108413219452,
-0.034034017473459244,
-0.11356475204229355,
-0.027013886719942093,
0.04241163283586502,
-0.06043251231312752,
0.08524788916110992,
0.023536119610071182,
-0.08113526552915573,
-0.032957352697849274,
0.05323701351881027,
0.012368366122245789,
0.00524376705288887,
0.09360801428556442,
0.020107939839363098,
-0.0009265501867048442,
0.01785753294825554,
0.047885000705718994,
-0.0675911232829094,
-0.1984109878540039,
0.09357594698667526,
-0.05215044692158699,
0.0015536568826064467,
-0.08013670891523361,
-0.15122665464878082,
-0.08837161958217621,
-0.16009655594825745,
0.12540200352668762,
-0.034406669437885284,
0.12700119614601135,
-0.06619787961244583,
0.17341409623622894,
-0.07871770113706589,
0.04481020197272301,
-0.047349292784929276,
0.050332702696323395,
-0.007268077693879604,
-0.07756082713603973,
0.16585899889469147,
-0.15564003586769104,
0.01809087023139,
0.19572502374649048,
-0.018915493041276932,
0.07177707552909851,
0.021322092041373253,
-0.0636206790804863,
0.23147478699684143,
0.3014698624610901,
0.008138049393892288,
0.1665448248386383,
0.3018903136253357,
-0.07466315478086472,
-0.2642788887023926,
-0.05505012720823288,
-0.2841376066207886,
-0.05371501296758652,
0.10716094076633453,
-0.22523896396160126,
0.06986407935619354,
0.14383509755134583,
-0.06471995264291763,
0.30228954553604126,
-0.21825523674488068,
0.012589273042976856,
0.15434536337852478,
-0.08868814259767532,
0.5515313148498535,
-0.1133413165807724,
-0.17677772045135498,
-0.008122089318931103,
-0.08741296827793121,
0.10602109134197235,
-0.0340677872300148,
0.06877441704273224,
0.013465235009789467,
0.04797380417585373,
0.048932258039712906,
-0.03111894056200981,
0.22701001167297363,
0.008710170164704323,
0.09015397727489471,
-0.07378865778446198,
-0.18624304234981537,
0.11639340221881866,
-0.04359482601284981,
-0.08891059458255768,
0.0849778801202774,
-0.05942516401410103,
-0.11078983545303345,
0.04663389176130295,
-0.07950539886951447,
-0.024862350896000862,
0.08423490077257156,
-0.04678233340382576,
-0.042606171220541,
-0.008054176345467567,
-0.1618063747882843,
-0.0002289071271661669,
0.31360217928886414,
-0.07096036523580551,
0.16695955395698547,
0.03677211329340935,
0.00038613268407061696,
-0.11027684062719345,
0.030288029462099075,
-0.05203165486454964,
-0.021576624363660812,
0.09578979015350342,
-0.11096979677677155,
0.03204701095819473,
0.14160704612731934,
-0.04864364117383957,
0.05846960097551346,
0.09256096184253693,
-0.0849417969584465,
0.007583672646433115,
0.17753590643405914,
-0.17537221312522888,
-0.1273445188999176,
-0.006135711446404457,
-0.09862716495990753,
0.14055661857128143,
0.04394126310944557,
0.05191568285226822,
0.16669964790344238,
0.03967129811644554,
-0.029474308714270592,
-0.02817419543862343,
-0.1153380498290062,
-0.0201893113553524,
0.040153320878744125,
0.00045633706031367183,
-0.08791285753250122,
0.2262638509273529,
0.06409153342247009,
-0.1328488290309906,
-0.051157206296920776,
0.2161225974559784,
-0.06805316358804703,
-0.04911920800805092,
-0.223562553524971,
0.10752306133508682,
-0.07112517952919006,
-0.0965060144662857,
0.05453834682703018,
-0.02270081453025341,
0.005106312222778797,
0.181985542178154,
0.03941008821129799,
0.11070270836353302,
0.03738937899470329,
-0.02448922023177147,
0.15798696875572205,
-0.142850860953331,
-0.14191335439682007,
-0.025354057550430298,
-0.08757315576076508,
-0.13844476640224457,
-0.026804137974977493,
0.1617041826248169,
-0.09177309274673462,
-0.14772607386112213,
-0.2621181011199951,
0.10968475043773651,
-0.16432365775108337,
-0.10192688554525375,
-0.03469514101743698,
-0.08968492597341537,
0.0696166530251503,
0.030301768332719803,
-0.03093348816037178,
-0.06706760823726654,
-0.18593791127204895,
0.0816768929362297,
0.06349513679742813,
0.045533183962106705,
-0.017847947776317596,
0.0067379772663116455,
0.1720137596130371,
0.025955144315958023,
0.10040043294429779,
0.16762186586856842,
0.011397695168852806,
0.2246655523777008,
-0.1671202927827835,
-0.11496317386627197,
0.1336962729692459,
-0.026543032377958298,
0.06762003898620605,
0.16792191565036774,
-0.0772583931684494,
0.015526676550507545,
-0.028136352077126503,
0.07066910713911057,
-0.11003983020782471,
-0.105624258518219,
0.007937257178127766,
0.02567129209637642,
-0.2755882740020752,
-0.005599735304713249,
-0.19717298448085785,
0.14788752794265747,
0.02579621411859989,
0.03297143429517746,
0.10257530212402344,
0.10404334217309952,
0.08312062919139862,
-0.0017710148822516203,
0.03226327523589134,
-0.1176818460226059,
0.02753005363047123,
-0.059239376336336136,
-0.020663779228925705,
0.017624232918024063,
0.36952024698257446,
-0.03603357449173927,
-0.046802736818790436,
0.003710439894348383,
0.1307835876941681,
-0.02139742486178875,
0.017395347356796265,
0.13209912180900574,
0.12607666850090027,
-0.08595693111419678,
-0.1504845917224884,
0.04888554662466049,
-0.04565655067563057,
-0.02836887165904045,
0.1464131623506546,
0.05905961990356445,
0.1050296202301979,
0.0908031314611435,
-0.014463032595813274,
-0.00318976235575974,
0.012856799177825451,
-0.15486004948616028,
0.06223496049642563,
-0.010558074340224266,
0.012565906159579754,
0.017934376373887062,
0.15238402783870697,
-0.005540105979889631,
0.07739730179309845,
-0.09889880567789078,
0.004208535887300968,
-0.13498884439468384,
-0.07913459837436676,
0.03617347031831741,
-0.13393273949623108,
0.04141177982091904,
-0.01871878281235695,
0.029611799865961075,
0.30386561155319214,
0.02558239921927452,
-0.020639164373278618,
0.12512871623039246,
-0.1214587539434433,
-0.12050267308950424,
-0.001594188273884356,
-0.029960084706544876,
0.0791488066315651,
-0.02633434161543846,
-0.0997740775346756,
-0.1001306027173996,
-0.15166029334068298,
-0.09759195148944855,
0.05182836204767227,
-0.04993441700935364,
-0.059362251311540604,
-0.17634081840515137,
-0.05707859992980957,
-0.05147340148687363,
0.14025864005088806,
-0.12263951450586319,
0.15159130096435547,
-0.014490418136119843,
0.004084470681846142,
0.04405883327126503,
0.1950942426919937,
-0.03644494712352753,
0.08714226633310318,
0.0154351145029068,
0.1522706001996994,
-0.05119588226079941,
0.14720745384693146,
-0.10931728035211563,
-0.04014137014746666,
-0.06710435450077057,
0.21513493359088898,
0.25630924105644226,
-0.06136954948306084,
-0.008937356993556023,
-0.012760217301547527,
0.058654606342315674,
0.1073930487036705,
0.16049085557460785,
0.002326392102986574,
0.2802925705909729,
-0.03133585304021835,
0.04815128445625305,
0.02901598811149597,
0.013607407920062542,
-0.06336209923028946,
0.03397751972079277,
0.07539387792348862,
-0.035039983689785004,
-0.1412304788827896,
0.15837742388248444,
-0.21980468928813934,
0.18157227337360382,
0.11640069633722305,
-0.19996967911720276,
-0.013728445395827293,
-0.04882071167230606,
0.1689416468143463,
-0.0856364443898201,
0.1637246012687683,
-0.0903693437576294,
-0.2108195722103119,
-0.2056000679731369,
0.03867346793413162,
-0.34623071551322937,
-0.254462867975235,
0.10422009229660034,
0.1488201916217804,
0.04015883058309555,
-0.018507536500692368,
-0.019967829808592796,
-0.018367022275924683,
0.04877542704343796,
-0.0067357709631323814,
0.06014643982052803,
0.031397558748722076,
-0.02988368645310402,
-0.24127542972564697,
-0.029804671183228493,
0.023964406922459602,
-0.07093082368373871,
0.07464958727359772,
-0.06874357163906097,
-0.022495782002806664,
0.08059766888618469,
-0.03066304884850979,
0.03298592567443848,
-0.035373736172914505,
-0.16326889395713806,
0.027529051527380943,
0.03900543600320816,
0.036012712866067886,
0.00634160777553916,
0.0008072225609794259,
-0.03455270454287529,
0.0644603744149208,
-0.16716794669628143,
-0.16015739738941193,
0.14140215516090393,
-0.06745140254497528,
0.2779497504234314,
-0.05812826007604599,
-0.0809100940823555,
0.04766704887151718,
-0.03426874056458473,
0.1807648241519928,
-0.07756473124027252,
0.047254521399736404,
0.12766779959201813,
0.011127962730824947,
0.03121316432952881,
-0.3092964291572571,
0.11082969605922699,
-0.000795336440205574,
-0.006093299947679043,
-0.07581598311662674
] |
null | null | null |
# Introduction
This repo contains the benchmark results for <https://github.com/csukuangfj/transducer-loss-benchmarking>
## Usage
First, install `git-lfs`.
Second, use the following command to clone this repo:
```bash
git lfs install
git clone https://huggingface.co/csukuangfj/transducer-loss-benchmarking
```
**Caution**: You have to run `git lfs install` first. Otherwise, you will be **SAD** later.
Third,
```
pip install torch-tb-profiler
cd transducer-loss-benchmarking
tensorboard --logdir ./log/torchaudio-30 --port 6006
tensorboard --logdir ./log/optimized_transducer-30 --port 6007
```
Fourth, open your browser and go to
- <http://localhost:6006/#pytorch_profiler>
- <http://localhost:6006/#pytorch_profiler>
You will see the following images:


|
{}
| null |
csukuangfj/transducer-loss-benchmarking
|
[
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#region-us
|
# Introduction
This repo contains the benchmark results for <URL
## Usage
First, install 'git-lfs'.
Second, use the following command to clone this repo:
Caution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.
Third,
Fourth, open your browser and go to
- <http://localhost:6006/#pytorch_profiler>
- <http://localhost:6006/#pytorch_profiler>
You will see the following images:


|
[
"# Introduction\n\nThis repo contains the benchmark results for <URL",
"## Usage\n\nFirst, install 'git-lfs'.\n\nSecond, use the following command to clone this repo:\n\n\n\nCaution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.\n\nThird,\n\n\nFourth, open your browser and go to\n\n- <http://localhost:6006/#pytorch_profiler>\n- <http://localhost:6006/#pytorch_profiler>\n\nYou will see the following images:\n\n\n\n"
] |
[
"TAGS\n#region-us \n",
"# Introduction\n\nThis repo contains the benchmark results for <URL",
"## Usage\n\nFirst, install 'git-lfs'.\n\nSecond, use the following command to clone this repo:\n\n\n\nCaution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.\n\nThird,\n\n\nFourth, open your browser and go to\n\n- <http://localhost:6006/#pytorch_profiler>\n- <http://localhost:6006/#pytorch_profiler>\n\nYou will see the following images:\n\n\n\n"
] |
[
6,
13,
126
] |
[
"passage: TAGS\n#region-us \n# Introduction\n\nThis repo contains the benchmark results for <URL## Usage\n\nFirst, install 'git-lfs'.\n\nSecond, use the following command to clone this repo:\n\n\n\nCaution: You have to run 'git lfs install' first. Otherwise, you will be SAD later.\n\nThird,\n\n\nFourth, open your browser and go to\n\n- <http://localhost:6006/#pytorch_profiler>\n- <http://localhost:6006/#pytorch_profiler>\n\nYou will see the following images:\n\n\n\n"
] |
[
-0.02542124129831791,
0.03949194401502609,
-0.00569039024412632,
0.03029646910727024,
0.15798062086105347,
0.06563792377710342,
0.005004215054214001,
0.11876966059207916,
0.1389773041009903,
-0.03474846109747887,
0.10570523887872696,
0.07168398797512054,
0.03597390651702881,
0.006856535095721483,
0.09946094453334808,
-0.14118137955665588,
0.03565117344260216,
0.02677096612751484,
-0.027888702228665352,
0.06124399974942207,
0.027276938781142235,
-0.005685562267899513,
0.10487893223762512,
-0.011493954807519913,
-0.09881938248872757,
0.12230752408504486,
0.0668758675456047,
0.09102863818407059,
0.07771401852369308,
-0.0777004212141037,
0.08737878501415253,
0.006320204585790634,
0.02546062506735325,
-0.1212213933467865,
0.031736019998788834,
0.10375207662582397,
0.0017107665771618485,
0.013533188961446285,
0.04734112694859505,
-0.012345441617071629,
0.08432330936193466,
0.12501157820224762,
-0.10546160489320755,
0.01811571605503559,
-0.10976442694664001,
-0.15065604448318481,
-0.10962342470884323,
-0.04082049801945686,
0.042896244674921036,
0.06779275834560394,
0.04510906711220741,
0.17093846201896667,
-0.13583184778690338,
0.10052292793989182,
0.3424539864063263,
-0.13635200262069702,
-0.006249954923987389,
0.14926259219646454,
-0.008803877048194408,
0.030699344351887703,
0.006012419238686562,
0.049080658704042435,
0.04878765717148781,
0.09442135691642761,
-0.13933853805065155,
-0.051625192165374756,
-0.03469425067305565,
-0.011297542601823807,
-0.08394971489906311,
-0.16070103645324707,
0.16098980605602264,
-0.003186493646353483,
-0.02377847209572792,
0.20606623589992523,
-0.13232196867465973,
-0.07832968980073929,
0.013392572291195393,
0.03551657497882843,
0.046079520136117935,
0.09721159934997559,
-0.06436315178871155,
-0.17887741327285767,
-0.06205330044031143,
-0.14596512913703918,
-0.05639423429965973,
0.154863640666008,
0.04226602613925934,
0.11560418456792831,
-0.029083918780088425,
0.11863987892866135,
-0.04809886962175369,
-0.04772686958312988,
-0.03866185247898102,
-0.08499163389205933,
0.0646316185593605,
0.09442473202943802,
0.03356286510825157,
-0.04980652406811714,
0.18203236162662506,
0.11727700382471085,
-0.06721208989620209,
0.010920548811554909,
-0.07301649451255798,
0.09293748438358307,
-0.06821933388710022,
0.0818755254149437,
-0.047964975237846375,
0.05040851980447769,
0.09353213757276535,
0.007092399522662163,
0.023929258808493614,
0.030300544574856758,
-0.0362628810107708,
-0.02428545244038105,
-0.009271373972296715,
0.1343303620815277,
0.11644051969051361,
0.04203709214925766,
-0.05755114182829857,
0.05009700357913971,
0.06833435595035553,
-0.058048151433467865,
-0.012461488135159016,
-0.0652666762471199,
-0.06976303458213806,
-0.03816479444503784,
0.1003500297665596,
-0.04443192854523659,
-0.12630051374435425,
-0.0492045022547245,
-0.020710689947009087,
0.022462937980890274,
-0.11961879581212997,
-0.01935717463493347,
0.02533537894487381,
-0.02111201547086239,
0.0163379218429327,
-0.12913769483566284,
-0.1367192417383194,
-0.01144877914339304,
0.04071750491857529,
0.036016952246427536,
0.057693589478731155,
0.05928829684853554,
-0.022316664457321167,
-0.1053871437907219,
-0.003975801169872284,
-0.02867055870592594,
-0.10038106143474579,
0.08455590903759003,
0.016460126265883446,
0.07234293967485428,
-0.13635049760341644,
0.010356729850172997,
-0.08054128289222717,
0.09785942733287811,
-0.2537221312522888,
-0.027702229097485542,
-0.14250284433364868,
0.12126395851373672,
-0.09616537392139435,
-0.027097193524241447,
-0.10397369414567947,
0.047596171498298645,
0.11450565606355667,
0.16379325091838837,
-0.07210060954093933,
0.004325130954384804,
0.19067321717739105,
-0.1544007658958435,
-0.0273793563246727,
0.1433413177728653,
0.05464264005422592,
-0.022896286100149155,
0.011656633578240871,
0.2133939415216446,
0.06101327762007713,
-0.1118059828877449,
0.06810576468706131,
0.1304563581943512,
-0.22718754410743713,
-0.09922246634960175,
-0.04743439704179764,
-0.06668106466531754,
-0.1886359453201294,
0.04180271178483963,
-0.09688772261142731,
0.10838035494089127,
-0.05269677937030792,
0.006683373358100653,
-0.04941016063094139,
-0.04865690693259239,
-0.10150662809610367,
0.006199758965522051,
-0.025073217228055,
-0.0342007651925087,
0.009314772672951221,
-0.16244356334209442,
0.07771121710538864,
-0.02629026398062706,
0.059290092438459396,
-0.02098623290657997,
0.16409635543823242,
0.005605682730674744,
0.050041817128658295,
-0.0493931844830513,
-0.05657093599438667,
0.036209139972925186,
0.09424713999032974,
0.02772318571805954,
-0.004245390184223652,
0.05432363972067833,
-0.08407299965620041,
0.013679321855306625,
-0.06848926842212677,
0.08281739801168442,
-0.0822228342294693,
0.023099351674318314,
-0.005058069247752428,
0.025957981124520302,
0.012298902496695518,
-0.24074780941009521,
0.011696476489305496,
-0.010991102084517479,
0.021719928830862045,
0.08908716589212418,
-0.025280296802520752,
-0.03079243376851082,
0.03371774032711983,
-0.053919170051813126,
0.028808867558836937,
-0.04542066901922226,
0.03527407348155975,
0.0018162665655836463,
0.011580448597669601,
0.04559032991528511,
0.12669463455677032,
0.06402084976434708,
0.12846730649471283,
-0.20070764422416687,
0.02070499025285244,
0.11335578560829163,
-0.05010831728577614,
0.015790803357958794,
0.024064065888524055,
-0.10441389679908752,
0.08917324244976044,
-0.015320184640586376,
0.01748458296060562,
-0.05758540332317352,
0.04939229413866997,
-0.034285131841897964,
-0.0639674961566925,
-0.030582791194319725,
0.06302329152822495,
0.1877654492855072,
-0.07878882437944412,
0.01285009179264307,
0.03127181902527809,
-0.08418890088796616,
-0.04296692833304405,
-0.05096244066953659,
-0.03788591921329498,
0.05100804194808006,
0.13397862017154694,
-0.018319357186555862,
0.11145909875631332,
-0.06191515550017357,
0.020633338019251823,
0.00599531177431345,
0.007930319756269455,
0.103995680809021,
-0.15217216312885284,
-0.06628813594579697,
-0.07537953555583954,
-0.11746140569448471,
-0.08299671113491058,
0.01398754958063364,
-0.010264592245221138,
0.07317224144935608,
-0.020194437354803085,
-0.06206116825342178,
0.014323538169264793,
0.005855511408299208,
-0.010992439463734627,
0.1060434877872467,
-0.12403790652751923,
-0.18316121399402618,
-0.1390041559934616,
0.019690493121743202,
-0.03793065622448921,
0.050613731145858765,
0.04511871188879013,
-0.15145306289196014,
-0.05745471268892288,
0.019190100952982903,
0.13513445854187012,
-0.08389925956726074,
-0.019729487597942352,
-0.018310867249965668,
0.031031517311930656,
0.09043405205011368,
-0.14924955368041992,
-0.02845916710793972,
-0.009680014103651047,
-0.10305856913328171,
0.05945766344666481,
-0.10307402163743973,
0.12349431216716766,
-0.043512631207704544,
0.03703862428665161,
0.04649031162261963,
0.0707184448838234,
0.23496678471565247,
-0.10600655525922775,
-0.02342139557003975,
0.14187608659267426,
0.02663690596818924,
0.0909716933965683,
0.05843571946024895,
0.01143388357013464,
-0.09643906354904175,
-0.003895375644788146,
0.07066582888364792,
-0.10154858231544495,
-0.17055056989192963,
0.048838984221220016,
0.008785346522927284,
0.0028001766186207533,
0.01722424291074276,
0.09219128638505936,
-0.10908050835132599,
0.06434693187475204,
-0.04552764073014259,
-0.0257840808480978,
-0.08687549829483032,
0.01708759367465973,
-0.07308731228113174,
-0.06929425895214081,
-0.025904634967446327,
-0.052040185779333115,
-0.0627511516213417,
0.10546381026506424,
0.03838774189352989,
0.12907133996486664,
-0.06870044022798538,
0.03937621787190437,
0.09246775507926941,
0.16665613651275635,
-0.027879171073436737,
0.16676828265190125,
-0.07850228995084763,
0.025294233113527298,
0.008347397670149803,
0.020197713747620583,
-0.05026829242706299,
-0.02444005385041237,
0.04624190554022789,
-0.0724167600274086,
0.022597631439566612,
-0.024366676807403564,
0.09367228299379349,
0.1825949102640152,
-0.003202797146514058,
-0.16608192026615143,
0.05449492111802101,
-0.028407806530594826,
-0.026563020423054695,
-0.05067245662212372,
0.12175020575523376,
0.04590843990445137,
0.01255951076745987,
0.023369435220956802,
0.04429003223776817,
0.07742928713560104,
-0.07653791457414627,
0.040263235569000244,
-0.00012902090384159237,
0.07953004539012909,
-0.013195126317441463,
0.02182517573237419,
-0.15790539979934692,
0.17718306183815002,
0.006093142554163933,
-0.04378070682287216,
-0.007371751591563225,
-0.014076082035899162,
0.10092872381210327,
0.06663253903388977,
0.10329952090978622,
0.06418765336275101,
-0.05497659742832184,
-0.0995246097445488,
-0.09043996036052704,
0.04676083102822304,
0.07275275141000748,
-0.016400029882788658,
0.00774424010887742,
-0.010933969169855118,
-0.017893381416797638,
-0.008081141859292984,
-0.07498174160718918,
-0.044873256236314774,
0.05502277985215187,
-0.0029622141737490892,
-0.014801861718297005,
0.009707305580377579,
-0.06246155872941017,
0.06477110832929611,
0.06163123622536659,
0.1384488344192505,
0.05322999134659767,
-0.08552147448062897,
-0.02475615032017231,
0.040224313735961914,
0.13964632153511047,
-0.025910774245858192,
-0.043936796486377716,
-0.0361718125641346,
-0.09593261778354645,
-0.02148343063890934,
-0.1377158761024475,
0.020851461216807365,
-0.05424780026078224,
-0.032466646283864975,
0.031087379902601242,
0.1637202948331833,
-0.02732613869011402,
-0.013392877764999866,
0.014504863880574703,
-0.08425328880548477,
-0.02490132860839367,
-0.11571141332387924,
0.11729013174772263,
-0.09075281023979187,
-0.008734052069485188,
0.15122510492801666,
-0.05045941844582558,
0.13829204440116882,
-0.051059551537036896,
0.0660296082496643,
0.04357083886861801,
0.21692471206188202,
-0.011889193207025528,
-0.062451910227537155,
0.06513264030218124,
-0.05709492415189743,
-0.25870105624198914,
-0.05476486310362816,
-0.11111680418252945,
-0.10268969088792801,
-0.03556250408291817,
-0.10326461493968964,
0.12030358612537384,
0.039257727563381195,
-0.02068895660340786,
0.23436596989631653,
-0.14804691076278687,
-0.0014080788241699338,
0.08973236382007599,
0.07447222620248795,
0.2616164982318878,
-0.1581805944442749,
-0.050412945449352264,
0.005102965515106916,
-0.07376537472009659,
-0.03169664740562439,
-0.02270009182393551,
0.08170104026794434,
-0.002129398984834552,
0.10771606117486954,
0.054193757474422455,
-0.056345779448747635,
0.11557665467262268,
-0.07692348212003708,
0.05273488536477089,
-0.09282301366329193,
-0.07705720514059067,
-0.07215652614831924,
-0.031595103442668915,
0.12359874695539474,
0.06974758952856064,
0.018404332920908928,
0.001562151126563549,
0.035649482160806656,
-0.11972612887620926,
0.10880950838327408,
0.01058280747383833,
0.006376942619681358,
-0.0979970246553421,
-0.04816737771034241,
-0.03946343809366226,
0.07461481541395187,
0.0641208365559578,
0.009905919432640076,
-0.04582098871469498,
0.06339234113693237,
-0.05396907776594162,
-0.08225112408399582,
-0.05743599683046341,
0.07830045372247696,
-0.0006671942537650466,
0.0718834400177002,
-0.06105370819568634,
0.07732722163200378,
0.054296448826789856,
-0.039660047739744186,
-0.06558429449796677,
0.037915050983428955,
-0.13030517101287842,
-0.037855297327041626,
0.13231632113456726,
-0.13066191971302032,
-0.03985043987631798,
-0.06814771890640259,
-0.13105224072933197,
0.05321972444653511,
0.040799401700496674,
0.1112852618098259,
0.03261219337582588,
0.018094979226589203,
0.016707615926861763,
0.04251882806420326,
-0.11192989349365234,
0.019360121339559555,
0.08724123239517212,
0.0014117771061137319,
-0.12154471129179001,
0.08931677043437958,
-0.024227309972047806,
-0.015097249299287796,
-0.050071194767951965,
0.11527813971042633,
-0.05157534033060074,
-0.09323737770318985,
-0.015011731535196304,
0.03207055851817131,
0.004348193760961294,
-0.04648907855153084,
-0.14422254264354706,
0.047194164246320724,
-0.007313756737858057,
0.0016347793862223625,
0.07437305152416229,
0.02208271250128746,
0.12717397511005402,
0.05304620414972305,
-0.0221925787627697,
-0.09205697476863861,
-0.002779782982543111,
0.045276954770088196,
-0.13327103853225708,
0.0196037869900465,
0.010616387240588665,
0.1352834552526474,
-0.04175747185945511,
0.049042508006095886,
-0.19654394686222076,
0.060524843633174896,
-0.11683634668588638,
-0.01408008299767971,
-0.016162514686584473,
0.004029357805848122,
0.059144701808691025,
-0.048111442476511,
-0.07747997343540192,
0.05023004859685898,
-0.1441849321126938,
-0.03909369558095932,
0.03764098882675171,
-0.04529472067952156,
-0.030489670112729073,
-0.03578140586614609,
0.1138034462928772,
-0.016257010400295258,
0.053262677043676376,
0.10995437949895859,
-0.03349513188004494,
0.05718691647052765,
-0.038012657314538956,
-0.11353650689125061,
0.08379587531089783,
0.015273964032530785,
-0.04656492546200752,
0.158168762922287,
0.05180281028151512,
-0.0004638626123778522,
0.003399703186005354,
-0.029686832800507545,
0.11520019173622131,
-0.12500733137130737,
0.02171526476740837,
0.03619151934981346,
-0.06631509959697723,
0.01066004578024149,
-0.03168908879160881,
0.15887561440467834,
0.14102640748023987,
0.03675331547856331,
-0.024793870747089386,
0.013340595178306103,
-0.08672191947698593,
0.02346678264439106,
0.032988760620355606,
-0.1191280335187912,
0.1178278997540474,
-0.002072911011055112,
0.03122624009847641,
-0.016920052468776703,
0.18331430852413177,
0.07332565635442734,
0.09227284789085388,
-0.04892707243561745,
0.09799786657094955,
0.07272951304912567,
-0.029249699786305428,
-0.03791983053088188,
-0.05414079129695892,
0.0008669908856973052,
-0.06691677868366241,
-0.010216495022177696,
0.10702603310346603,
-0.13073912262916565,
0.11059661954641342,
0.0711246132850647,
0.11420761048793793,
0.046675220131874084,
0.007682275027036667,
-0.09417540580034256,
0.003226321889087558,
-0.27735987305641174,
0.014420630410313606,
0.04232138395309448,
-0.006395766045898199,
0.00926254689693451,
0.06013351306319237,
0.023231761530041695,
0.06301369518041611,
0.021115003153681755,
-0.028958044946193695,
-0.10635848343372345,
-0.0892127975821495,
-0.0034427575301378965,
-0.23070846498012543,
0.06751485168933868,
-0.06532055884599686,
-0.03626978397369385,
0.06626082956790924,
-0.028629109263420105,
-0.07502519339323044,
0.1422542929649353,
0.0005086823366582394,
-0.09618565440177917,
0.011628407984972,
-0.018845418468117714,
-0.05173828452825546,
0.1255066692829132,
-0.07261651754379272,
0.01841672509908676,
0.009116314351558685,
0.018805984407663345,
0.04357634112238884,
-0.016129035502672195,
0.030680246651172638,
-0.09700105339288712,
-0.0023370555136352777,
-0.04841763153672218,
0.04236746206879616,
-0.04436787590384483,
0.12097995728254318,
0.020209673792123795,
-0.001249745604582131,
0.02422333136200905,
0.158956378698349,
-0.0386795848608017,
-0.004008327145129442,
-0.0431671068072319,
0.14417417347431183,
-0.04673672839999199,
0.013094000518321991,
-0.08734310418367386,
-0.09823784977197647,
-0.08022740483283997,
0.3548494577407837,
0.1809517741203308,
-0.06328412145376205,
-0.00647373404353857,
-0.03895387798547745,
0.029175905510783195,
0.03215290233492851,
0.1221129447221756,
0.1435927450656891,
0.10263554751873016,
-0.057384192943573,
-0.07824314385652542,
-0.011000593192875385,
0.023262236267328262,
-0.156717911362648,
-0.035802580416202545,
0.08758579194545746,
-0.026693524792790413,
-0.09704694896936417,
0.1399378478527069,
-0.17539042234420776,
-0.04096825420856476,
-0.059287603944540024,
-0.05228449031710625,
0.018210915848612785,
-0.01370219886302948,
0.009020060300827026,
-0.004191912245005369,
0.05249554291367531,
-0.08659780025482178,
0.04662751033902168,
0.022425418719649315,
-0.0356622114777565,
-0.2487495094537735,
-0.15451474487781525,
0.04531734809279442,
-0.07342660427093506,
0.22479906678199768,
0.028345463797450066,
-0.04042429104447365,
0.009024086408317089,
-0.06344927102327347,
-0.08631565421819687,
0.04993360862135887,
-0.007006560917943716,
-0.20056839287281036,
-0.06861616671085358,
0.025904580950737,
-0.03771137073636055,
-0.1133868545293808,
-0.033543702214956284,
0.013024828396737576,
-0.03384968265891075,
0.1348607987165451,
0.029166748747229576,
-0.09195594489574432,
-0.03288853168487549,
-0.22248713672161102,
0.05359429493546486,
0.11111146211624146,
0.01153834443539381,
0.06490431725978851,
-0.09412461519241333,
0.05362457409501076,
0.06048983708024025,
-0.0555117167532444,
-0.04397399723529816,
-0.07565689086914062,
-0.07373347878456116,
0.08524724096059799,
-0.061941955238580704,
-0.020994149148464203,
0.020474478602409363,
-0.017823511734604836,
0.005102504976093769,
-0.05827386677265167,
0.05740026384592056,
-0.0038290382362902164,
0.04940532147884369,
0.04116668924689293,
-0.11312902718782425,
0.02849142625927925,
0.06447269022464752,
-0.10685338079929352,
-0.0942755863070488
] |
null | null |
transformers
|
# Wav2Vec2-Large-XLSR-53-Cantonese
Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on Cantonese using the [Common Voice](https://huggingface.co/datasets/common_voice).
When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
```python
import torch
import torchaudio
from datasets import load_dataset
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
test_dataset = load_dataset("common_voice", "zh-HK", split="test[:2%]")
processor = Wav2Vec2Processor.from_pretrained("ctl/wav2vec2-large-xlsr-cantonese")
model = Wav2Vec2ForCTC.from_pretrained("ctl/wav2vec2-large-xlsr-cantonese")
resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = resampler(speech_array).squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
inputs = processor(test_dataset["speech"][:2], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values, attention_mask=inputs.attention_mask).logits
predicted_ids = torch.argmax(logits, dim=-1)
print("Prediction:", processor.batch_decode(predicted_ids))
print("Reference:", test_dataset["sentence"][:2])
```
## Evaluation
The model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice.
```python
!pip install jiwer
import torch
import torchaudio
from datasets import load_dataset, load_metric
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor
import re
import argparse
lang_id = "zh-HK"
model_id = "ctl/wav2vec2-large-xlsr-cantonese"
chars_to_ignore_regex = '[\,\?\.\!\-\;\:"\“\%\‘\”\�\.\⋯\!\-\:\–\。\》\,\)\,\?\;\~\~\…\︰\,\(\」\‧\《\﹔\、\—\/\,\「\﹖\·\']'
test_dataset = load_dataset("common_voice", f"{lang_id}", split="test")
cer = load_metric("cer")
processor = Wav2Vec2Processor.from_pretrained(f"{model_id}")
model = Wav2Vec2ForCTC.from_pretrained(f"{model_id}")
model.to("cuda")
resampler = torchaudio.transforms.Resample(48_000, 16_000)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def speech_file_to_array_fn(batch):
batch["sentence"] = re.sub(chars_to_ignore_regex, '', batch["sentence"]).lower()
speech_array, sampling_rate = torchaudio.load(batch["path"])
batch["speech"] = resampler(speech_array).squeeze().numpy()
return batch
test_dataset = test_dataset.map(speech_file_to_array_fn)
# Preprocessing the datasets.
# We need to read the aduio files as arrays
def evaluate(batch):
inputs = processor(batch["speech"], sampling_rate=16_000, return_tensors="pt", padding=True)
with torch.no_grad():
logits = model(inputs.input_values.to("cuda"), attention_mask=inputs.attention_mask.to("cuda")).logits
pred_ids = torch.argmax(logits, dim=-1)
batch["pred_strings"] = processor.batch_decode(pred_ids)
return batch
result = test_dataset.map(evaluate, batched=True, batch_size=16)
print("CER: {:2f}".format(100 * cer.compute(predictions=result["pred_strings"], references=result["sentence"])))
```
**Test Result**: 15.51 %
## Training
The Common Voice `train`, `validation` were used for training.
The script used for training will be posted [here](https://github.com/chutaklee/CantoASR)
|
{"language": ["yue"], "license": "apache-2.0", "tags": ["audio", "automatic-speech-recognition", "speech", "xlsr-fine-tuning-week"], "datasets": ["common_voice"], "metrics": ["cer"], "language_bcp47": ["zh-HK"], "model-index": [{"name": "wav2vec2-large-xlsr-cantonese", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Speech Recognition"}, "dataset": {"name": "Common Voice zh-HK", "type": "common_voice", "args": "zh-HK"}, "metrics": [{"type": "cer", "value": 15.36, "name": "Test CER"}]}]}]}
|
automatic-speech-recognition
|
ctl/wav2vec2-large-xlsr-cantonese
|
[
"transformers",
"pytorch",
"jax",
"wav2vec2",
"automatic-speech-recognition",
"audio",
"speech",
"xlsr-fine-tuning-week",
"yue",
"dataset:common_voice",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"yue"
] |
TAGS
#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #yue #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us
|
# Wav2Vec2-Large-XLSR-53-Cantonese
Fine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.
When using this model, make sure that your speech input is sampled at 16kHz.
## Usage
The model can be used directly (without a language model) as follows:
## Evaluation
The model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice.
Test Result: 15.51 %
## Training
The Common Voice 'train', 'validation' were used for training.
The script used for training will be posted here
|
[
"# Wav2Vec2-Large-XLSR-53-Cantonese\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.\nWhen using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. \n\n\n\n\n\nTest Result: 15.51 %",
"## Training\n\nThe Common Voice 'train', 'validation' were used for training.\n\nThe script used for training will be posted here"
] |
[
"TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #yue #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n",
"# Wav2Vec2-Large-XLSR-53-Cantonese\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.\nWhen using this model, make sure that your speech input is sampled at 16kHz.",
"## Usage\n\nThe model can be used directly (without a language model) as follows:",
"## Evaluation\n\nThe model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. \n\n\n\n\n\nTest Result: 15.51 %",
"## Training\n\nThe Common Voice 'train', 'validation' were used for training.\n\nThe script used for training will be posted here"
] |
[
85,
65,
20,
32,
29
] |
[
"passage: TAGS\n#transformers #pytorch #jax #wav2vec2 #automatic-speech-recognition #audio #speech #xlsr-fine-tuning-week #yue #dataset-common_voice #license-apache-2.0 #model-index #endpoints_compatible #has_space #region-us \n# Wav2Vec2-Large-XLSR-53-Cantonese\n\nFine-tuned facebook/wav2vec2-large-xlsr-53 on Cantonese using the Common Voice.\nWhen using this model, make sure that your speech input is sampled at 16kHz.## Usage\n\nThe model can be used directly (without a language model) as follows:## Evaluation\n\nThe model can be evaluated as follows on the Chinese (Hong Kong) test data of Common Voice. \n\n\n\n\n\nTest Result: 15.51 %## Training\n\nThe Common Voice 'train', 'validation' were used for training.\n\nThe script used for training will be posted here"
] |
[
-0.12746071815490723,
0.009867233224213123,
-0.002203010953962803,
-0.047473274171352386,
0.10282795876264572,
-0.08455982804298401,
0.18695911765098572,
0.11480991542339325,
-0.03922693431377411,
0.005476950667798519,
0.0020250030793249607,
-0.03411377966403961,
0.08279486000537872,
0.08098917454481125,
0.017794588580727577,
-0.2215646654367447,
0.011786747723817825,
0.057323966175317764,
0.0935540571808815,
0.12098769843578339,
0.09703639149665833,
-0.07620083540678024,
0.018378769978880882,
0.1069604828953743,
-0.133029967546463,
0.03383960947394371,
0.01843075454235077,
-0.1140306293964386,
0.14766642451286316,
0.03350585699081421,
0.08046090602874756,
0.049755215644836426,
0.07678398489952087,
-0.21656860411167145,
0.024134280160069466,
0.0342620313167572,
0.045025311410427094,
0.02592499554157257,
0.06772521138191223,
0.051542650908231735,
0.08500636368989944,
0.11455529928207397,
-0.0242848452180624,
0.10411297529935837,
-0.05459537357091904,
-0.20345556735992432,
-0.020226141437888145,
0.06053198501467705,
0.13161225616931915,
0.14668717980384827,
-0.07816218584775925,
0.13324101269245148,
-0.15846413373947144,
0.08707088977098465,
0.09278307110071182,
-0.22948230803012848,
0.014009037055075169,
0.1314043551683426,
0.11137472093105316,
0.05054888874292374,
-0.07986295223236084,
-0.005406043492257595,
0.06375308334827423,
0.010662487708032131,
-0.026049675419926643,
-0.05175243690609932,
-0.19789010286331177,
-0.02729373797774315,
-0.12710073590278625,
-0.016212107613682747,
0.2468602955341339,
-0.002693296642974019,
-0.0909520611166954,
-0.10784990340471268,
0.00911711435765028,
-0.028875650838017464,
0.0048630028031766415,
-0.07551291584968567,
0.007265950553119183,
0.051234353333711624,
-0.00296372571028769,
-0.057379573583602905,
-0.11268537491559982,
-0.15047170221805573,
0.0138632170855999,
0.012829956598579884,
0.026925547048449516,
-0.01096296589821577,
-0.17365294694900513,
0.09387179464101791,
-0.06827438622713089,
-0.06753332912921906,
-0.020620150491595268,
0.04527261480689049,
-0.04437718912959099,
-0.008502179756760597,
-0.03577717766165733,
-0.12254542112350464,
0.008878174237906933,
0.0026123393326997757,
0.02584523893892765,
0.04784281551837921,
-0.08518083393573761,
0.08703405410051346,
0.009676074609160423,
0.15231627225875854,
-0.06883752346038818,
0.03619677200913429,
0.04351776838302612,
0.03136517480015755,
-0.054156284779310226,
-0.004376332275569439,
-0.08205366879701614,
-0.039245475083589554,
0.07161279767751694,
0.06867021322250366,
-0.03798737749457359,
0.014138611033558846,
-0.06206245720386505,
-0.020369920879602432,
0.07333002984523773,
-0.08603891730308533,
-0.03334351256489754,
0.07356683909893036,
-0.03370217978954315,
0.10351880639791489,
0.06503278762102127,
0.07085491716861725,
-0.0784715786576271,
-0.01036395039409399,
0.005909696687012911,
0.07970980554819107,
-0.0038927667774260044,
-0.06014368683099747,
0.020841244608163834,
-0.02413717471063137,
-0.024596357718110085,
-0.08324981480836868,
-0.10750987380743027,
-0.08183608949184418,
-0.05313943699002266,
0.05087575316429138,
-0.035424355417490005,
-0.09075375646352768,
-0.04543381556868553,
-0.05045556277036667,
-0.0575714036822319,
0.06257046014070511,
-0.03522275760769844,
0.07507907599210739,
0.07949066907167435,
0.03786410763859749,
0.04982072114944458,
0.07395675033330917,
-0.09618727117776871,
-0.02697264961898327,
-0.025785574689507484,
0.1175551488995552,
-0.003912885207682848,
-0.04333268851041794,
-0.08576548844575882,
-0.0701189860701561,
-0.06509426236152649,
0.06849490851163864,
0.034311484545469284,
0.12080971896648407,
-0.2853754162788391,
-0.08932636678218842,
0.21093016862869263,
-0.12967392802238464,
-0.04018797352910042,
0.21343524754047394,
0.002847545314580202,
0.10052747279405594,
0.10664302110671997,
0.22728334367275238,
0.05461852625012398,
-0.2379201501607895,
0.030547009781003,
-0.0018178577302023768,
-0.02961411513388157,
-0.046562351286411285,
0.087440624833107,
-0.054620686918497086,
-0.012970051728188992,
0.0179065503180027,
-0.10403173416852951,
0.048116397112607956,
-0.04537526145577431,
-0.05577057972550392,
-0.02621554583311081,
-0.05539092421531677,
0.019584624096751213,
0.0378132201731205,
0.03240156173706055,
-0.027508437633514404,
-0.06322427839040756,
0.050579335540533066,
0.107469342648983,
-0.1117129772901535,
0.06763087958097458,
-0.12454096227884293,
0.08614616841077805,
-0.019853565841913223,
-0.009633881971240044,
-0.12647905945777893,
0.199935644865036,
0.004600026644766331,
0.02549174800515175,
0.0753818154335022,
0.1535877287387848,
-0.013215024024248123,
0.006604690570384264,
-0.0558585450053215,
-0.03344951942563057,
0.024145256727933884,
-0.03805268555879593,
-0.03212185204029083,
-0.06324603408575058,
-0.01641944795846939,
-0.04573288559913635,
0.0983731746673584,
-0.1989627480506897,
0.00013275582750793546,
0.04746904969215393,
0.020886577665805817,
-0.0148461377248168,
-0.0029626640025526285,
0.10408952087163925,
0.10187181085348129,
-0.022141823545098305,
0.007614923641085625,
0.033729977905750275,
0.030351536348462105,
-0.10072933882474899,
0.17000658810138702,
-0.1651647686958313,
-0.04271123558282852,
0.13307815790176392,
-0.07313984632492065,
-0.015889279544353485,
0.022977618500590324,
-0.01689937338232994,
-0.03198962286114693,
-0.09532851725816727,
0.014107550494372845,
0.3200870454311371,
-0.033486805856227875,
0.1535319685935974,
-0.11870317161083221,
0.014568887650966644,
0.0016873261192813516,
-0.09502275288105011,
0.028309771791100502,
0.05727161094546318,
-0.003691375255584717,
0.032553885132074356,
0.027320193126797676,
-0.08252330124378204,
-0.09057573974132538,
0.2272271066904068,
-0.01402521412819624,
-0.06526478379964828,
0.01195299532264471,
0.040403492748737335,
-0.006270790006965399,
0.044589996337890625,
-0.18494360148906708,
-0.06278755515813828,
0.010949689894914627,
0.05781404674053192,
0.06868959963321686,
-0.19209139049053192,
0.0032034781761467457,
-0.0052596330642700195,
-0.14285191893577576,
-0.1501619964838028,
0.056228917092084885,
-0.04227873682975769,
0.04470990225672722,
-0.09621012955904007,
-0.025757398456335068,
-0.002098778961226344,
-0.030033929273486137,
-0.1821015179157257,
0.10610940307378769,
-0.07413607090711594,
-0.24974612891674042,
-0.12689542770385742,
0.04715600237250328,
0.055005814880132675,
0.024939686059951782,
0.08809228986501694,
-0.1501728892326355,
0.0017184352036565542,
-0.0342099629342556,
0.041425447911024094,
-0.020099099725484848,
-0.043646275997161865,
-0.03725479170680046,
0.04521883651614189,
0.07739274948835373,
-0.12787780165672302,
0.013006925582885742,
-0.04673732444643974,
-0.08362239599227905,
0.018116870895028114,
-0.03236944228410721,
-0.0069108689203858376,
0.18354792892932892,
0.03315519914031029,
-0.009091177955269814,
-0.04193872958421707,
0.11837702244520187,
-0.10336536914110184,
-0.023418037220835686,
0.22328630089759827,
0.004746129736304283,
-0.017646508291363716,
0.08591236174106598,
-0.0016837334260344505,
-0.06344553083181381,
0.033705636858940125,
-0.021080341190099716,
-0.07697522640228271,
-0.256820410490036,
-0.07864347845315933,
-0.06851265579462051,
-0.03455473110079765,
-0.04573523998260498,
0.018304912373423576,
0.10292859375476837,
-0.010854392312467098,
-0.044484943151474,
-0.05706975609064102,
0.05710681527853012,
0.017024973407387733,
0.050850410014390945,
0.011524503119289875,
0.09380704164505005,
-0.03528820723295212,
-0.02042754366993904,
0.0001450782729079947,
0.06042521446943283,
0.08854573220014572,
0.07316013425588608,
0.04868924245238304,
0.083837129175663,
0.15402555465698242,
0.15268167853355408,
0.01655757613480091,
-0.05722891911864281,
-0.02320002019405365,
0.0047270930372178555,
-0.06038770079612732,
-0.03343654423952103,
0.058195676654577255,
0.18408919870853424,
-0.05388990417122841,
-0.04410609230399132,
0.002596795093268156,
0.004911500494927168,
0.16827134788036346,
0.08785233646631241,
-0.15941573679447174,
-0.0660964772105217,
-0.01817644201219082,
-0.05643367022275925,
0.014151654206216335,
0.08550605177879333,
0.18008309602737427,
-0.16060906648635864,
0.024582669138908386,
0.03305364027619362,
0.10292154550552368,
-0.008660621009767056,
0.03115776553750038,
-0.0775982066988945,
-0.0008094153017736971,
0.011672403663396835,
0.06926818937063217,
-0.3304884433746338,
0.2175893932580948,
-0.005982426926493645,
0.10373298078775406,
-0.040512897074222565,
0.008105899207293987,
0.009279299527406693,
0.049219392240047455,
0.1270233392715454,
0.005462083965539932,
0.02793983928859234,
-0.11949758976697922,
-0.06705310195684433,
0.034362051635980606,
0.009118110872805119,
0.06823813170194626,
0.024716632440686226,
0.029448464512825012,
-0.00814851839095354,
-0.006785155739635229,
-0.06368333846330643,
-0.15404067933559418,
-0.01151900552213192,
0.0031182188540697098,
0.16679343581199646,
0.11825362592935562,
-0.02656880021095276,
-0.07185079902410507,
-0.10549537092447281,
0.05445234850049019,
-0.12449735403060913,
-0.0209230687469244,
-0.03143016993999481,
-0.051969800144433975,
0.08596684783697128,
-0.0627909004688263,
-0.017607970163226128,
0.07119470834732056,
0.09905868023633957,
-0.02391345612704754,
0.005028614308685064,
0.0716281309723854,
-0.10113935172557831,
-0.12782467901706696,
-0.022054333239793777,
0.2077244073152542,
0.06735368818044662,
0.08872399479150772,
0.0499352365732193,
-0.014257288537919521,
0.002135803224518895,
-0.04525597020983696,
-0.014855324290692806,
0.05319685861468315,
-0.05739425867795944,
0.027721164748072624,
-0.02690090797841549,
-0.1545579582452774,
-0.16038204729557037,
-0.07931819558143616,
0.11372138559818268,
0.11778703331947327,
-0.034518368542194366,
0.12329907715320587,
0.21702252328395844,
-0.08513771742582321,
-0.17905274033546448,
-0.02716694585978985,
0.0777006521821022,
0.1042976975440979,
-0.03257915377616882,
-0.17814777791500092,
0.057216569781303406,
-0.0052196611650288105,
-0.03496802970767021,
-0.07379245012998581,
-0.26735740900039673,
-0.17376427352428436,
0.1024850532412529,
-0.05227157101035118,
0.08078571408987045,
-0.038833312690258026,
-0.04972926527261734,
0.030033553019165993,
-0.005815607029944658,
-0.031093837693333626,
-0.0659475177526474,
0.11305796355009079,
0.029769890010356903,
0.04989367723464966,
0.030369199812412262,
-0.041287798434495926,
0.07955539971590042,
0.08887680619955063,
-0.011397137306630611,
-0.008275296539068222,
0.016407595947384834,
0.02121841162443161,
0.022397538647055626,
0.16842417418956757,
-0.08634020388126373,
0.032464154064655304,
-0.12703773379325867,
-0.07772824168205261,
-0.06558067351579666,
0.05627346783876419,
0.008740750141441822,
-0.01631668396294117,
0.003221290884539485,
-0.001129458425566554,
-0.027679594233632088,
0.027916057035326958,
-0.008672425523400307,
-0.12800224125385284,
0.06270068138837814,
0.17255635559558868,
0.19659437239170074,
-0.02678162418305874,
-0.0439860038459301,
-0.024133065715432167,
-0.02274073287844658,
0.12328124046325684,
-0.07640804350376129,
0.023198066279292107,
0.05577969178557396,
0.04460792616009712,
0.14795280992984772,
0.021314434707164764,
-0.08704790472984314,
0.11448829621076584,
0.014455028809607029,
-0.04245374724268913,
-0.12124349921941757,
-0.03367811068892479,
-0.007023499347269535,
0.006981853861361742,
0.020841538906097412,
0.07725001871585846,
-0.09789305925369263,
-0.02729937434196472,
-0.014023000374436378,
0.015256284736096859,
-0.12005103379487991,
0.18404355645179749,
0.007539636921137571,
0.08857695013284683,
-0.09953693300485611,
0.03544991463422775,
0.050045549869537354,
-0.01615661382675171,
0.04971859231591225,
-0.008397727273404598,
-0.08819599449634552,
-0.049784231930971146,
-0.04895342141389847,
0.10730215907096863,
0.085471972823143,
-0.1263781487941742,
-0.061428252607584,
-0.09565296024084091,
-0.003688790602609515,
0.06599976122379303,
0.048039376735687256,
0.025602631270885468,
-0.08654402196407318,
-0.05611300840973854,
-0.11708489060401917,
0.02293941006064415,
0.08725240081548691,
-0.006867141928523779,
-0.12026005238294601,
0.23479272425174713,
0.11925530433654785,
0.06096901744604111,
-0.030576562508940697,
-0.08459799736738205,
-0.05571766197681427,
0.10060151666402817,
-0.05657067149877548,
0.012178526259958744,
-0.04793380945920944,
-0.009933914057910442,
-0.014166651293635368,
-0.09182919561862946,
-0.011205730028450489,
0.08994780480861664,
-0.07490510493516922,
0.055585574358701706,
-0.001447310671210289,
0.07351596653461456,
-0.07738351076841354,
0.027432898059487343,
0.058151621371507645,
-0.0632934644818306,
0.07063289731740952,
0.12072435021400452,
-0.11354897171258926,
0.0987531915307045,
-0.18992553651332855,
-0.06308989226818085,
0.08635488152503967,
0.05663678050041199,
-0.04662567377090454,
-0.08948996663093567,
0.03359157592058182,
0.09689554572105408,
0.08925791084766388,
-0.02309250831604004,
0.09379082918167114,
-0.0626104474067688,
-0.05736691877245903,
-0.08205379545688629,
-0.02995513379573822,
-0.040531400591135025,
0.05724991112947464,
0.08725869655609131,
0.13094235956668854,
0.14133882522583008,
-0.12254258245229721,
0.09597701579332352,
-0.13797084987163544,
0.006835282780230045,
-0.05512333661317825,
-0.014728052541613579,
-0.13265083730220795,
-0.08165706694126129,
0.06903526932001114,
-0.04631805419921875,
0.10135392844676971,
-0.023171626031398773,
0.03865688294172287,
-0.009822872467339039,
-0.07537106424570084,
-0.016814973205327988,
-0.007295588031411171,
0.2293815165758133,
0.06276324391365051,
0.022304024547338486,
0.020021209493279457,
-0.003430603537708521,
0.005009505897760391,
0.09622325748205185,
0.02951815165579319,
0.14114506542682648,
-0.02908160164952278,
0.09871802479028702,
0.0838107094168663,
-0.05426323413848877,
-0.07832134515047073,
-0.03948558494448662,
-0.17205794155597687,
0.020115463063120842,
-0.0670996829867363,
0.18658848106861115,
0.1605735421180725,
-0.09515485912561417,
0.09224820882081985,
0.02246711403131485,
-0.1054266020655632,
-0.15686452388763428,
-0.1205635517835617,
-0.06116029620170593,
-0.1492573618888855,
0.03841530531644821,
-0.05499941110610962,
0.026528645306825638,
0.014884881675243378,
0.05028776079416275,
-0.05184149369597435,
0.1605154126882553,
0.022837763652205467,
-0.10220008343458176,
0.07632068544626236,
-0.08707983791828156,
0.02067508175969124,
-0.08190339058637619,
0.028214596211910248,
0.16948163509368896,
-0.021335385739803314,
0.06795485317707062,
0.019066868349909782,
-0.04153605177998543,
0.056004222482442856,
-0.07578045129776001,
-0.062077343463897705,
0.007794241886585951,
-0.040271881967782974,
0.0984102115035057,
0.128941610455513,
0.11234597861766815,
-0.05695551633834839,
0.02183709107339382,
0.12915821373462677,
-0.032857514917850494,
-0.12034349143505096,
-0.14496932923793793,
0.18736408650875092,
0.027602363377809525,
-0.009473422542214394,
-0.02008584700524807,
-0.030391588807106018,
0.0014382685767486691,
0.26397886872291565,
0.18849323689937592,
0.03509325534105301,
0.020540202036499977,
-0.035794299095869064,
-0.009614736773073673,
-0.042181696742773056,
0.026994889602065086,
0.11537710577249527,
0.17195577919483185,
0.0033777104690670967,
0.02011292800307274,
-0.0733630433678627,
-0.0878719612956047,
-0.016280626878142357,
-0.0031844843178987503,
-0.07025963813066483,
-0.09079641103744507,
-0.013028711080551147,
0.1116478443145752,
-0.08200601488351822,
-0.08149518072605133,
-0.09943146258592606,
-0.08783300966024399,
-0.05778159946203232,
0.017467377707362175,
0.030130615457892418,
0.11828924715518951,
-0.0013920519268140197,
-0.05944781005382538,
0.0413462333381176,
0.11170587688684464,
0.002438825787976384,
-0.02644984796643257,
-0.06965066492557526,
0.03735581040382385,
-0.08819371461868286,
0.02834842912852764,
0.002463347977027297,
0.14860846102237701,
0.008365350775420666,
0.10223882645368576,
0.0016303478041663766,
0.16919763386249542,
-0.02807922288775444,
-0.045084066689014435,
0.013764540664851665,
0.14234791696071625,
-0.014960071071982384,
0.12251440435647964,
0.009218120947480202,
-0.09781023859977722,
0.06277918815612793,
-0.13623754680156708,
0.014386257156729698,
-0.13658595085144043,
0.07753930240869522,
-0.04842418432235718,
0.1017984002828598,
0.08620229363441467,
-0.06963185220956802,
-0.030822616070508957,
-0.04120481386780739,
0.048898790031671524,
0.000311294017592445,
-0.05832398310303688,
-0.04970679432153702,
-0.23466132581233978,
0.027778606861829758,
-0.11267463117837906,
-0.008928108960390091,
-0.19655638933181763,
-0.017889106646180153,
-0.009715316817164421,
-0.09296458214521408,
-0.014206902123987675,
0.02159084565937519,
0.09425710141658783,
0.03407401964068413,
-0.009493691846728325,
-0.08804582804441452,
0.05373304709792137,
0.13355903327465057,
-0.17060115933418274,
-0.11093630641698837
] |
null | null |
transformers
|
# My Awesome Model
|
{"tags": ["conversational"]}
|
text-generation
|
cumtowndiscord/DialoGPT-small-joshua
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# My Awesome Model
|
[
"# My Awesome Model"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# My Awesome Model"
] |
[
51,
4
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# My Awesome Model"
] |
[
-0.05259015038609505,
0.05521034821867943,
-0.005910294596105814,
0.017722278833389282,
0.15250112116336823,
0.02286236733198166,
0.07657632976770401,
0.09513414651155472,
-0.025391526520252228,
-0.047348517924547195,
0.15119488537311554,
0.19781284034252167,
-0.020334534347057343,
0.101333387196064,
-0.04688440263271332,
-0.3143521845340729,
0.06439975649118423,
0.05463787540793419,
-0.015605635941028595,
0.12023304402828217,
0.09468326717615128,
-0.0530015267431736,
0.08742043375968933,
-0.012155864387750626,
-0.1293085366487503,
-0.0027921805158257484,
-0.002384399762377143,
-0.10180269181728363,
0.11194873601198196,
0.033712033182382584,
0.05166437849402428,
0.0182647667825222,
-0.05843055993318558,
-0.139859139919281,
0.03845210000872612,
-0.015005595050752163,
-0.05602653697133064,
0.05648263916373253,
0.059830192476511,
-0.07164353132247925,
0.1669619083404541,
0.13275989890098572,
-0.04237370565533638,
0.056127581745386124,
-0.17620700597763062,
0.017941240221261978,
0.01800798624753952,
0.019184142351150513,
0.05306641012430191,
0.10830496996641159,
-0.03932326287031174,
0.09217294305562973,
-0.11410652846097946,
0.08313368260860443,
0.07800983637571335,
-0.29151955246925354,
-0.025312699377536774,
0.10440942645072937,
0.06437138468027115,
0.048375632613897324,
-0.013386772945523262,
0.0621674507856369,
0.02149512618780136,
0.008602659218013287,
0.02225899137556553,
-0.06727100163698196,
-0.05789240449666977,
0.032748885452747345,
-0.0967593789100647,
-0.03634428232908249,
0.19753605127334595,
-0.024647634476423264,
0.053590498864650726,
-0.06265407055616379,
-0.11300963163375854,
-0.039751436561346054,
-0.050429005175828934,
-0.029761891812086105,
-0.05090925097465515,
0.09489558637142181,
0.004352911841124296,
-0.09534718841314316,
-0.13405443727970123,
-0.01370926946401596,
-0.1618979275226593,
0.15892250835895538,
0.012579603120684624,
0.046201955527067184,
-0.19210097193717957,
0.11465331166982651,
-0.03857925534248352,
-0.08259090781211853,
0.030513519421219826,
-0.12010065466165543,
0.03160654753446579,
-0.008132083341479301,
-0.019599268212914467,
-0.049325279891490936,
0.061037879437208176,
0.08101806789636612,
0.018783701583743095,
0.005755073390901089,
0.018167443573474884,
0.05343452841043472,
0.05891622602939606,
0.10033947974443436,
-0.02891627699136734,
-0.0625043511390686,
0.0025436533614993095,
-0.12051084637641907,
-0.01122665498405695,
-0.05357983708381653,
-0.18095199763774872,
0.002246231772005558,
0.02455340512096882,
0.05192234739661217,
0.011778532527387142,
0.09955989569425583,
-0.028496338054537773,
-0.026898741722106934,
0.06898727267980576,
0.002862759632989764,
-0.015707949176430702,
-0.005368964280933142,
-0.010934269987046719,
0.11485416442155838,
-0.023099146783351898,
0.04774846136569977,
-0.12022071331739426,
0.020393015816807747,
-0.07851235568523407,
-0.0019349842332303524,
-0.06214260309934616,
-0.04864754155278206,
-0.0019346009939908981,
-0.06985589861869812,
0.021118074655532837,
-0.14833110570907593,
-0.17990200221538544,
-0.005064866971224546,
0.021302316337823868,
-0.052403319627046585,
-0.09162671118974686,
-0.0982397273182869,
-0.02586611732840538,
0.03574685752391815,
-0.05873546749353409,
0.013170980848371983,
-0.06884536147117615,
0.06542801111936569,
0.0029820678755640984,
0.05682007595896721,
-0.14085575938224792,
0.08719147741794586,
-0.12582023441791534,
-0.023288866505026817,
-0.061977192759513855,
0.1109607070684433,
0.024780582636594772,
0.1267160177230835,
0.004311583004891872,
-0.0033308975398540497,
-0.08729329705238342,
0.08271238207817078,
-0.04243258014321327,
0.22770646214485168,
-0.10479787737131119,
-0.08809807151556015,
0.2632525563240051,
-0.05423165112733841,
-0.16432519257068634,
0.10179096460342407,
-0.014350244775414467,
0.12198644131422043,
0.13850919902324677,
0.16080057621002197,
0.007628654129803181,
0.03313867375254631,
0.10115300863981247,
0.08631709218025208,
-0.08573295921087265,
-0.0611947737634182,
0.023627014830708504,
-0.011463395319879055,
-0.10670105367898941,
0.046802595257759094,
0.04794782027602196,
0.08188598603010178,
-0.04982871189713478,
-0.028600862249732018,
-0.01972118206322193,
-0.044152840971946716,
0.05264130234718323,
0.007675500120967627,
0.13217447698116302,
-0.03674980252981186,
-0.03692879155278206,
-0.023745311424136162,
0.01699630729854107,
-0.03115241602063179,
0.007061392068862915,
-0.05687357112765312,
0.11091547459363937,
-0.03406180441379547,
0.051789235323667526,
-0.16953988373279572,
-0.04873261600732803,
-0.02087729424238205,
0.1402055323123932,
0.04973345249891281,
0.1329866498708725,
0.06287940591573715,
-0.010758201591670513,
0.00859389640390873,
0.007998145185410976,
0.13181665539741516,
0.007865442894399166,
-0.07660657912492752,
-0.047718439251184464,
0.09176599979400635,
-0.05973208695650101,
0.06147782504558563,
-0.098741315305233,
-0.004747362341731787,
-0.01433002483099699,
0.08674649894237518,
0.006352655589580536,
0.029382232576608658,
-0.006192679051309824,
0.003654100699350238,
-0.06161240115761757,
0.017873648554086685,
0.12492607533931732,
-0.01421504095196724,
-0.07439801841974258,
0.22084392607212067,
-0.15798072516918182,
0.18006981909275055,
0.18165533244609833,
-0.3081994652748108,
0.024602634832262993,
-0.08860466629266739,
-0.036338552832603455,
0.03426366671919823,
0.0491504967212677,
-0.034147560596466064,
0.16587987542152405,
-0.016766328364610672,
0.201018825173378,
-0.03547777235507965,
-0.01287798210978508,
-0.010399105958640575,
-0.03656993433833122,
-0.010632630437612534,
0.09065473079681396,
0.15122920274734497,
-0.1677125245332718,
0.18270380795001984,
0.1660280078649521,
0.06873020529747009,
0.17776396870613098,
0.034313347190618515,
-0.006856906693428755,
0.07112615555524826,
-0.022670727223157883,
-0.07675548642873764,
-0.049287427216768265,
-0.26302891969680786,
-0.027947327122092247,
0.06471601128578186,
0.04510856419801712,
0.11924877762794495,
-0.10971947014331818,
-0.037208184599876404,
0.010892451740801334,
-0.013165894895792007,
0.02132410928606987,
0.09682225435972214,
0.01171150617301464,
0.11804302036762238,
-0.021027036011219025,
-0.05209195241332054,
0.0898953229188919,
0.02727191150188446,
-0.0787680521607399,
0.19168277084827423,
-0.10074768215417862,
-0.3233809769153595,
-0.11354339867830276,
-0.18166927993297577,
-0.017843691632151604,
0.05878754332661629,
0.08049646019935608,
-0.09228580445051193,
-0.02625267766416073,
-0.01639235019683838,
0.0758359357714653,
-0.09145816415548325,
-0.015880629420280457,
-0.09367848187685013,
0.034986745566129684,
-0.10827737301588058,
-0.07011983543634415,
-0.05141967162489891,
-0.03368452936410904,
-0.04457031562924385,
0.13157756626605988,
-0.12242637574672699,
0.06396433711051941,
0.2076517641544342,
0.06227295100688934,
0.05622440204024315,
-0.0229496993124485,
0.23288212716579437,
-0.10842552781105042,
0.02383521944284439,
0.1717897206544876,
-0.03566030040383339,
0.0727933868765831,
0.13435456156730652,
0.006721907295286655,
-0.08144525438547134,
0.03465581312775612,
-0.04592517390847206,
-0.08630958944559097,
-0.20441576838493347,
-0.14156180620193481,
-0.12814727425575256,
0.07913564145565033,
0.03285396471619606,
0.05478321388363838,
0.15024253726005554,
0.11386489123106003,
0.007987297140061855,
0.00976672861725092,
-0.006888182368129492,
0.05438044294714928,
0.17482298612594604,
-0.05838097631931305,
0.10041683167219162,
-0.037591226398944855,
-0.1924494504928589,
0.08022978901863098,
0.04309763014316559,
0.08280511945486069,
0.07474655658006668,
0.0856199786067009,
0.013537914492189884,
0.03723837807774544,
0.10897084325551987,
0.1165735274553299,
0.031679023057222366,
-0.038079675287008286,
-0.04882059991359711,
-0.026300756260752678,
-0.03285675123333931,
0.05745977535843849,
0.07790146768093109,
-0.1608346849679947,
-0.06348084658384323,
-0.06350091099739075,
0.07662643492221832,
0.09017108380794525,
0.11811108142137527,
-0.21219493448734283,
0.01579318381845951,
0.092556893825531,
-0.0494147390127182,
-0.1304239183664322,
0.07402537018060684,
-0.00466050673276186,
-0.1397053301334381,
0.037663187831640244,
-0.014095795340836048,
0.1359514445066452,
-0.0778401643037796,
0.10336452722549438,
-0.08307972550392151,
-0.06147889420390129,
0.03632286190986633,
0.1355396956205368,
-0.30774354934692383,
0.2137020230293274,
-0.022472934797406197,
-0.05296783149242401,
-0.10508129745721817,
-0.011727629229426384,
0.020913105458021164,
0.09079049527645111,
0.10090240091085434,
-0.0025442070327699184,
0.0061299679800868034,
-0.0345483273267746,
-0.053218815475702286,
0.024456629529595375,
0.07957815378904343,
-0.08542889356613159,
0.0017540202243253589,
-0.02361489273607731,
-0.004407065454870462,
-0.032844748347997665,
-0.01189463958144188,
-0.011617658659815788,
-0.16786961257457733,
0.06556065380573273,
-0.002625665394589305,
0.11129079759120941,
0.03491498529911041,
0.0024013579823076725,
-0.1009332686662674,
0.19977013766765594,
0.01796281896531582,
-0.08052749931812286,
-0.08830537647008896,
-0.03254766762256622,
0.03660419583320618,
-0.06121435388922691,
0.027481911703944206,
-0.06916457414627075,
0.033381566405296326,
-0.06441576033830643,
-0.18325145542621613,
0.1268530637025833,
-0.10945470631122589,
-0.03609596937894821,
-0.04321056231856346,
0.18323224782943726,
-0.00929707009345293,
-0.0011623724130913615,
0.05866571143269539,
0.0032208464108407497,
-0.1347510665655136,
-0.10740556567907333,
0.020214511081576347,
-0.015275230631232262,
0.009142245166003704,
0.05559912323951721,
-0.009665844030678272,
0.00045268211397342384,
-0.039558928459882736,
-0.023234419524669647,
0.32348164916038513,
0.10732097923755646,
-0.04944206401705742,
0.17007054388523102,
0.13087597489356995,
-0.0827672928571701,
-0.30699312686920166,
-0.10971353948116302,
-0.10529600828886032,
-0.026918673887848854,
-0.037983208894729614,
-0.19617970287799835,
0.09504909813404083,
-0.03528566658496857,
-0.022136637941002846,
0.11253651231527328,
-0.2759084105491638,
-0.0770430713891983,
0.1826775223016739,
0.003314757253974676,
0.3998824954032898,
-0.10265109688043594,
-0.08777514100074768,
-0.06741699576377869,
-0.1120782196521759,
0.2033512443304062,
-0.05560711398720741,
0.08663415163755417,
-0.00517998356372118,
0.15513743460178375,
0.055607251822948456,
-0.02176513522863388,
0.08932057023048401,
-0.005811662413179874,
-0.0546204075217247,
-0.1219351515173912,
-0.03444604203104973,
-0.009159418754279613,
0.007239421829581261,
0.03589896112680435,
-0.04242607578635216,
0.01279151439666748,
-0.1399589478969574,
-0.045490626245737076,
-0.0764620453119278,
0.024699507281184196,
0.021008269861340523,
-0.0652410089969635,
-0.01643640361726284,
-0.03945036977529526,
-0.012804778292775154,
0.03164318576455116,
0.15236099064350128,
-0.06478006392717361,
0.1476556956768036,
0.04904455319046974,
0.15412139892578125,
-0.14745712280273438,
-0.02258288487792015,
-0.06896031647920609,
-0.05498642474412918,
0.04900865629315376,
-0.10053684562444687,
0.050061121582984924,
0.1202658861875534,
-0.0742902010679245,
0.0987328365445137,
0.0922594666481018,
-0.01938629150390625,
0.0012483424507081509,
0.1226617842912674,
-0.2489612102508545,
-0.07742628455162048,
-0.10509459674358368,
0.013337249867618084,
0.10138551890850067,
0.06995654851198196,
0.17304721474647522,
-0.0037713919300585985,
-0.036284226924180984,
-0.0064643872901797295,
0.025414984673261642,
-0.03540204465389252,
0.05724727362394333,
-0.002706433180719614,
0.016663886606693268,
-0.15213344991207123,
0.060368724167346954,
-0.00024176653823815286,
-0.1438901126384735,
-0.013603870756924152,
0.16073721647262573,
-0.11208858340978622,
-0.15145981311798096,
-0.007263668347150087,
0.13685113191604614,
-0.13171035051345825,
-0.03302847594022751,
-0.03708777576684952,
-0.170182466506958,
0.07439173012971878,
0.1024777740240097,
0.08549231290817261,
0.08025266975164413,
-0.06620611250400543,
-0.00807863101363182,
-0.011656313203275204,
-0.026087598875164986,
0.031810320913791656,
-0.023377234116196632,
-0.09044221043586731,
0.03872343525290489,
-0.026654237881302834,
0.13591371476650238,
-0.09607382118701935,
-0.09331836551427841,
-0.135749951004982,
0.039314381778240204,
-0.12405620515346527,
-0.08138058334589005,
-0.12200927734375,
-0.0591500885784626,
0.00224387738853693,
-0.0001289021165575832,
-0.035674065351486206,
-0.06687422841787338,
-0.13582271337509155,
0.04366770386695862,
-0.04484611004590988,
0.0013091047294437885,
-0.040241483598947525,
0.04561002552509308,
0.06766383349895477,
-0.03493715822696686,
0.13722217082977295,
0.11722734570503235,
-0.07864081114530563,
0.08946478366851807,
-0.16657429933547974,
-0.0683990865945816,
0.08854512125253677,
0.008173754438757896,
0.06165994703769684,
0.06743349134922028,
0.033807408064603806,
0.06109451875090599,
0.04151686280965805,
0.03488299250602722,
0.01739438995718956,
-0.09271225333213806,
0.015541021712124348,
0.022296719253063202,
-0.1294609159231186,
-0.04801803454756737,
-0.029226921498775482,
0.00939185917377472,
0.008117396384477615,
0.11003357172012329,
-0.0426274873316288,
0.09439733624458313,
-0.05888751894235611,
0.036728594452142715,
0.016222506761550903,
-0.16461637616157532,
-0.020102784037590027,
-0.11915475130081177,
0.028684545308351517,
-0.0033096212428063154,
0.25625869631767273,
0.06346847862005234,
0.020517030730843544,
0.01250078622251749,
0.08567021042108536,
0.07241600006818771,
0.02562166005373001,
0.1956365555524826,
0.10854171961545944,
-0.05020022392272949,
-0.12334850430488586,
0.09686340391635895,
0.034720368683338165,
0.06432123482227325,
0.13385434448719025,
-0.026959087699651718,
0.002498799469321966,
0.11019360274076462,
0.011678861454129219,
0.04961980879306793,
-0.09859088063240051,
-0.16400282084941864,
-0.00994415208697319,
0.061864156275987625,
-0.04559077322483063,
0.12240655720233917,
0.11382720619440079,
-0.020697353407740593,
0.03180128335952759,
-0.010503606870770454,
-0.05694027617573738,
-0.16998925805091858,
-0.1630837321281433,
-0.08357038348913193,
-0.11794789135456085,
-0.0027763545513153076,
-0.11386270076036453,
0.013879159465432167,
0.06452289968729019,
0.0604364387691021,
-0.09019444137811661,
0.08891061693429947,
0.0687386617064476,
-0.11843101680278778,
0.08828350901603699,
-0.033263903111219406,
0.07249268144369125,
0.0015160300536081195,
0.003872724948450923,
-0.13800905644893646,
0.032393742352724075,
-0.008493867702782154,
0.04159298539161682,
-0.09244006127119064,
0.022458361461758614,
-0.11297028511762619,
-0.07659684121608734,
-0.07971972227096558,
0.05093973129987717,
-0.03541257977485657,
0.1390930563211441,
0.001295371213927865,
-0.035233911126852036,
0.024190181866288185,
0.22729112207889557,
-0.06350252777338028,
-0.030667411163449287,
-0.0618741400539875,
0.21414142847061157,
0.024466563016176224,
0.10703565180301666,
-0.016775688156485558,
0.019240234047174454,
-0.0764411985874176,
0.3689337372779846,
0.344390869140625,
-0.1225387305021286,
-0.0015968306688591838,
0.031062176451086998,
0.036916591227054596,
0.11621878296136856,
0.12602226436138153,
0.057955991476774216,
0.2995031177997589,
-0.08396036922931671,
-0.002026971662417054,
-0.02688612788915634,
-0.03624163940548897,
-0.04409930482506752,
0.10547586530447006,
0.06835740804672241,
-0.03330419585108757,
-0.027012333273887634,
0.1376710683107376,
-0.2966996431350708,
0.12323499470949173,
-0.15714547038078308,
-0.1487535685300827,
-0.06873904913663864,
-0.005042468197643757,
0.08589684963226318,
0.04748665541410446,
0.1069009080529213,
-0.019124338403344154,
-0.08203735202550888,
0.05766449123620987,
0.0320524163544178,
-0.22844897210597992,
0.011852608993649483,
0.08361081779003143,
-0.06153005734086037,
0.011767351068556309,
-0.017906347289681435,
0.038472190499305725,
0.07790610194206238,
0.025976579636335373,
-0.032770540565252304,
0.06325861811637878,
-0.005814229138195515,
-0.05033424496650696,
0.04302205145359039,
0.05059972032904625,
0.017107632011175156,
-0.1511564701795578,
0.07320158183574677,
-0.1762860119342804,
0.0566408596932888,
-0.005331212189048529,
-0.04948166385293007,
0.000018263708625454456,
0.01998119056224823,
-0.06808236241340637,
0.05880929157137871,
0.0952666699886322,
-0.012173139490187168,
-0.002317852806299925,
-0.056667573750019073,
0.007662574760615826,
-0.0679154172539711,
-0.0747012197971344,
-0.10497893393039703,
-0.1338900774717331,
-0.11392296850681305,
0.10846775025129318,
-0.011928223073482513,
-0.19833622872829437,
0.02906924858689308,
-0.11258108913898468,
0.04933213070034981,
-0.13360801339149475,
0.08599711954593658,
0.1282832771539688,
0.021543797105550766,
-0.01265349704772234,
0.04020093381404877,
0.01591683179140091,
0.08550478518009186,
-0.09200563281774521,
-0.10515180230140686
] |
null | null |
transformers
|
Fine tuning LayoutLMv2 model on Vietnamese bill dataset
```python
from transformers import LayoutLMv2ForTokenClassification
model = LayoutLMv2ForTokenClassification.from_pretrained('cuongngm/layoutlm-bill', num_labels=len(labels))
```
labels = ['price',
'storename',
'total_cost',
'phone',
'address',
'unitprice',
'item',
'subitem',
'other',
'time',
'unit',
'total refunds',
'total_qty',
'seller',
'total_received']
|
{}
|
token-classification
|
cuongngm/layoutlm-bill
|
[
"transformers",
"pytorch",
"layoutlmv2",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #layoutlmv2 #token-classification #autotrain_compatible #endpoints_compatible #region-us
|
Fine tuning LayoutLMv2 model on Vietnamese bill dataset
labels = ['price',
'storename',
'total_cost',
'phone',
'address',
'unitprice',
'item',
'subitem',
'other',
'time',
'unit',
'total refunds',
'total_qty',
'seller',
'total_received']
|
[] |
[
"TAGS\n#transformers #pytorch #layoutlmv2 #token-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
41
] |
[
"passage: TAGS\n#transformers #pytorch #layoutlmv2 #token-classification #autotrain_compatible #endpoints_compatible #region-us \n"
] |
[
-0.05778699368238449,
0.03974124416708946,
-0.008774658665060997,
0.03134288638830185,
0.16732493042945862,
0.034163497388362885,
0.07008584588766098,
0.09666585177183151,
0.01093202829360962,
-0.03180955722928047,
0.11920414119958878,
0.2536323070526123,
-0.02541540376842022,
0.12405043840408325,
-0.08940679579973221,
-0.2751603424549103,
0.05289534479379654,
0.07812897115945816,
-0.0509544312953949,
0.10437314957380295,
0.09426096081733704,
-0.096477210521698,
0.08267372101545334,
-0.017689958214759827,
-0.15998992323875427,
0.031330011785030365,
0.031133079901337624,
-0.11463280022144318,
0.09397059679031372,
0.0392887182533741,
0.17064033448696136,
0.03290946036577225,
-0.023194191977381706,
-0.1294083297252655,
0.026020493358373642,
0.026482179760932922,
-0.06576782464981079,
0.06989692151546478,
0.09297885000705719,
-0.0724194347858429,
0.0004459601186681539,
0.023980289697647095,
0.02180878445506096,
0.04127785190939903,
-0.1154012456536293,
-0.11993054300546646,
-0.028065886348485947,
0.06803034991025925,
0.06359178572893143,
0.028339453041553497,
0.04528860002756119,
0.21416284143924713,
-0.12959837913513184,
0.1112191379070282,
0.11915595829486847,
-0.2852734923362732,
-0.005756948608905077,
0.14852507412433624,
-0.0030514465179294348,
-0.03204598277807236,
-0.02021097019314766,
0.04491802304983139,
0.021199705079197884,
0.018341125920414925,
0.027534527704119682,
-0.06974580883979797,
-0.10134192556142807,
0.02801455929875374,
-0.09212496131658554,
-0.03259341046214104,
0.19970397651195526,
-0.03201518580317497,
0.06338144093751907,
-0.0024487206246703863,
-0.10643959790468216,
-0.02835744433104992,
-0.020721422508358955,
0.0026460534427314997,
-0.027708260342478752,
0.0392601303756237,
0.006277269683778286,
0.015453265979886055,
-0.10328508913516998,
0.008988311514258385,
-0.2241227626800537,
0.22883152961730957,
0.021964799612760544,
0.07942602038383484,
-0.16387812793254852,
0.07133140414953232,
0.00432487390935421,
-0.09671732783317566,
0.0443706251680851,
-0.10392092913389206,
0.00154713389929384,
-0.050665080547332764,
-0.02832726575434208,
0.016673464328050613,
0.07684864103794098,
0.11847390979528427,
0.09207979589700699,
0.04180806502699852,
0.0005521529237739742,
0.09081310778856277,
0.031043238937854767,
0.10453872382640839,
-0.008934730663895607,
-0.026695383712649345,
0.05356279015541077,
-0.11922793090343475,
0.013831875286996365,
-0.04324895516037941,
-0.1353645622730255,
-0.052963029593229294,
0.07332616299390793,
0.10282445698976517,
0.025591719895601273,
0.06446130573749542,
-0.06997820734977722,
-0.05468494072556496,
0.10225611925125122,
-0.07712759077548981,
0.02380007691681385,
0.0021684167440980673,
0.010550598613917828,
0.11851963400840759,
-0.007207420188933611,
0.0061905160546302795,
-0.04666688293218613,
0.11484462022781372,
-0.05645641312003136,
0.006140651181340218,
-0.045840680599212646,
-0.06910344213247299,
0.046140436083078384,
-0.14090818166732788,
0.044356945902109146,
-0.16669055819511414,
-0.11685968935489655,
0.035181231796741486,
0.041286733001470566,
0.0077555058524012566,
-0.04102035239338875,
0.020996762439608574,
0.0021928229834884405,
0.0021646725945174694,
-0.06196162849664688,
-0.014183941297233105,
-0.04863809049129486,
0.06290029734373093,
-0.011364633217453957,
0.06171398237347603,
-0.08672089129686356,
0.05835375934839249,
-0.09694076329469681,
0.02440287359058857,
-0.10379631072282791,
-0.0006411506328731775,
-0.052533239126205444,
0.16130277514457703,
-0.03289404883980751,
-0.06531120091676712,
-0.040451355278491974,
0.00815553031861782,
-0.042518362402915955,
0.11008085310459137,
-0.08943292498588562,
-0.1067400798201561,
0.1347787082195282,
-0.11807852983474731,
-0.1288880556821823,
0.06096221134066582,
0.006109313108026981,
-0.010931028053164482,
0.06116866692900658,
0.09868606925010681,
0.12345214188098907,
-0.034919194877147675,
0.05594572052359581,
0.12384092807769775,
-0.15172681212425232,
-0.15482929348945618,
0.01971510797739029,
0.012406782247126102,
-0.12012148648500443,
0.05494942516088486,
0.06585130095481873,
0.07786384969949722,
-0.07147765904664993,
-0.04589385911822319,
-0.04318201541900635,
-0.015889273956418037,
0.10509875416755676,
0.054638952016830444,
0.09694869816303253,
-0.0420546755194664,
0.02332139015197754,
0.04700779914855957,
0.03256576508283615,
0.034252576529979706,
0.015309659764170647,
-0.08613849431276321,
0.11357367783784866,
-0.07669250667095184,
0.009631022810935974,
-0.19131550192832947,
-0.1182670146226883,
0.005608411505818367,
0.059924058616161346,
-0.033211253583431244,
0.13797996938228607,
0.07247842103242874,
-0.03458826616406441,
-0.0023120781406760216,
-0.027080101892352104,
0.15149423480033875,
0.03204483911395073,
-0.06698699295520782,
-0.10485642403364182,
-0.0035018303897231817,
-0.06614898145198822,
-0.03379574418067932,
-0.03812050446867943,
0.013367078267037868,
0.07384876906871796,
0.16941611468791962,
0.01876668632030487,
0.07587146013975143,
-0.015248815529048443,
0.05340377986431122,
-0.06478475034236908,
-0.005989662371575832,
0.11225032806396484,
-0.008588879369199276,
-0.05581550672650337,
0.12216658890247345,
-0.11563436686992645,
0.3249482214450836,
0.185505673289299,
-0.2773588299751282,
0.017563272267580032,
-0.04416828602552414,
-0.023483986034989357,
-0.0004700309655163437,
0.023501502349972725,
0.05075979232788086,
0.02893635630607605,
0.013826107606291771,
0.16484588384628296,
-0.017082910984754562,
-0.05047663673758507,
0.015549123287200928,
-0.06421256810426712,
-0.04786786064505577,
0.06996762007474899,
0.101493239402771,
-0.19506941735744476,
0.1725781261920929,
0.21628913283348083,
0.0018555746646597981,
0.09177366644144058,
-0.013344105333089828,
0.023524761199951172,
0.031723566353321075,
-0.03686145320534706,
-0.027301767840981483,
-0.008122175000607967,
-0.1615874469280243,
-0.02990984171628952,
0.08935689181089401,
0.029235607013106346,
0.05835922434926033,
-0.1354506015777588,
-0.01868385449051857,
0.016246246173977852,
0.039081692695617676,
-0.015885232016444206,
0.09966475516557693,
0.051478609442710876,
0.07956003397703171,
-0.00868115946650505,
-0.07854697853326797,
0.10971519351005554,
0.004792613442987204,
-0.06752726435661316,
0.15722130239009857,
-0.12969624996185303,
-0.2947699725627899,
-0.14390388131141663,
-0.19127389788627625,
-0.05046501010656357,
0.04131399095058441,
0.048620935529470444,
-0.09950754046440125,
-0.059498172253370285,
0.06847665458917618,
0.001623176853172481,
-0.06055719032883644,
0.07062265276908875,
-0.03540179878473282,
0.08509432524442673,
-0.018519936129450798,
-0.06954783946275711,
-0.055927615612745285,
-0.0414201058447361,
-0.02613709680736065,
0.13159692287445068,
-0.07070932537317276,
0.061695270240306854,
0.16283351182937622,
-0.013026673346757889,
0.06530477106571198,
-0.008078161627054214,
0.14188720285892487,
-0.04140574112534523,
-0.013613970950245857,
0.22200967371463776,
-0.05633757263422012,
0.08353669196367264,
0.1540241837501526,
0.041764017194509506,
-0.0403645895421505,
-0.00221289717592299,
-0.044953569769859314,
-0.11235775053501129,
-0.1929023265838623,
-0.14358146488666534,
-0.11976822465658188,
0.022163430228829384,
0.0670013353228569,
0.06063150241971016,
0.10678660124540329,
0.10436863452196121,
0.02886897139251232,
0.03517468273639679,
-0.05179718881845474,
0.07515324652194977,
0.2329932153224945,
0.0010029739933088422,
0.13521945476531982,
-0.06918307393789291,
-0.12653949856758118,
0.0786515474319458,
0.06753218173980713,
0.16002078354358673,
0.08846481889486313,
-0.03066476806998253,
0.0255486648529768,
0.13737152516841888,
0.16751514375209808,
0.1297721564769745,
0.019602876156568527,
-0.03937277942895889,
0.005858137737959623,
0.00009884304745355621,
-0.050046294927597046,
0.01805928908288479,
0.11923974007368088,
-0.11102869361639023,
-0.049453530460596085,
-0.08008626103401184,
0.0741138905286789,
0.10086892545223236,
0.05682622641324997,
-0.22243084013462067,
0.023685850203037262,
0.07992563396692276,
0.00009739531378727406,
-0.07107511907815933,
0.041660066694021225,
-0.05711185559630394,
-0.13931119441986084,
0.09855468571186066,
-0.04457836225628853,
0.1122964546084404,
-0.08670590817928314,
0.04690837860107422,
-0.012499743141233921,
-0.044510915875434875,
0.04440256580710411,
0.11083681136369705,
-0.23993900418281555,
0.23940879106521606,
0.013277639634907246,
-0.06741482764482498,
-0.08327994495630264,
0.0008145290776155889,
0.038868650794029236,
0.2019420564174652,
0.054901283234357834,
0.018034491688013077,
-0.1020551547408104,
-0.18951259553432465,
-0.04604224115610123,
0.011967078782618046,
0.08163987845182419,
-0.0021220948547124863,
-0.02166498824954033,
-0.03929366171360016,
-0.026156701147556305,
-0.018793383613228798,
-0.033498529344797134,
-0.0002000758977374062,
-0.11958488076925278,
0.06002609431743622,
0.04885181412100792,
0.01973840221762657,
0.0027429915498942137,
-0.06634359061717987,
-0.11346053332090378,
0.23760053515434265,
-0.08375486731529236,
-0.07633186131715775,
-0.12529438734054565,
-0.08247952908277512,
0.07333134859800339,
-0.08693837374448776,
0.07178293913602829,
-0.09247510880231857,
0.046413157135248184,
-0.03039172664284706,
-0.20090904831886292,
0.1285516619682312,
-0.12863905727863312,
-0.041191618889570236,
-0.06193387880921364,
0.14642776548862457,
-0.09815216064453125,
0.018241681158542633,
0.01747041381895542,
0.017957033589482307,
-0.07901791483163834,
-0.08617274463176727,
0.016419190913438797,
0.032105859369039536,
0.03878277912735939,
0.050481166690588,
-0.057005301117897034,
-0.03258875384926796,
0.018971407786011696,
0.02980881556868553,
0.24965596199035645,
0.18610970675945282,
-0.07782752066850662,
0.12343915551900864,
0.11683381348848343,
-0.044916391372680664,
-0.3175329566001892,
-0.07109321653842926,
-0.09933280944824219,
-0.040871161967515945,
-0.03415766730904579,
-0.13326579332351685,
0.1367558240890503,
0.022940225899219513,
-0.04264963045716286,
0.07325160503387451,
-0.16267172992229462,
-0.08590833842754364,
0.1943003535270691,
-0.0010218152310699224,
0.3578220307826996,
-0.06566820293664932,
-0.0798448771238327,
-0.04600640386343002,
-0.15766489505767822,
0.09459023177623749,
0.002373378723859787,
0.06756804883480072,
-0.04350960999727249,
0.041615575551986694,
0.03665178641676903,
-0.06291329115629196,
0.10352667421102524,
0.03191421926021576,
0.04651443660259247,
-0.09885291755199432,
-0.0890415608882904,
0.046507179737091064,
-0.036033984273672104,
0.009482534602284431,
0.049101680517196655,
0.03696424141526222,
-0.1246163547039032,
-0.014286420308053493,
-0.06409062445163727,
0.08560869097709656,
0.04267823323607445,
-0.057284217327833176,
0.007375302724540234,
-0.029944051057100296,
-0.005789119750261307,
-0.00421120272949338,
0.2532790005207062,
0.02436981536448002,
0.11846763640642166,
0.14298470318317413,
0.09473973512649536,
-0.1688377559185028,
-0.027834711596369743,
-0.07952225208282471,
-0.060129642486572266,
0.07923906296491623,
-0.04621187970042229,
0.08262167870998383,
0.1278066486120224,
-0.029770877212285995,
0.03190390393137932,
0.11209224164485931,
0.05402187630534172,
-0.043240662664175034,
0.153057262301445,
-0.1937270313501358,
0.0442107692360878,
-0.027363670989871025,
0.0004532164894044399,
0.048240721225738525,
0.12736783921718597,
0.1153838038444519,
0.041781287640333176,
-0.023382162675261497,
0.006082751788198948,
-0.011888482607901096,
-0.053514204919338226,
0.0843736082315445,
0.07893380522727966,
0.048407651484012604,
-0.1457613855600357,
0.05378333106637001,
0.027309643104672432,
-0.09572086483240128,
-0.05806012824177742,
0.06238840892910957,
-0.1553935408592224,
-0.11065811663866043,
0.011939547024667263,
0.09843193739652634,
-0.13321523368358612,
-0.06422793120145798,
-0.053000614047050476,
-0.1297502964735031,
0.07072881609201431,
0.1531151980161667,
0.13062144815921783,
0.09846726804971695,
-0.04803524166345596,
-0.04502645507454872,
-0.02932196669280529,
-0.018670959398150444,
0.010611818172037601,
0.060962654650211334,
-0.18999697268009186,
0.038105450570583344,
-0.00807112455368042,
0.16742603480815887,
-0.09692341834306717,
-0.06846752017736435,
-0.1336631029844284,
0.050985608249902725,
-0.08091622591018677,
-0.06551729142665863,
-0.0959804505109787,
-0.016431493684649467,
0.024287445470690727,
-0.06683551520109177,
-0.04708725959062576,
-0.014623597264289856,
-0.11040210723876953,
0.04908880591392517,
0.003515340853482485,
0.03214099630713463,
-0.053075410425662994,
-0.046283118426799774,
0.0833362266421318,
-0.03921474888920784,
0.08558426797389984,
0.08461696654558182,
-0.05467032268643379,
0.06570801883935928,
-0.07886151969432831,
-0.13805243372917175,
0.14046891033649445,
0.04486068710684776,
0.09652195870876312,
0.017176570370793343,
0.038593705743551254,
0.073255255818367,
0.011806592345237732,
0.050993017852306366,
0.07536526024341583,
-0.12348739057779312,
0.032782990485429764,
-0.06000329554080963,
-0.1613915115594864,
-0.02635570615530014,
-0.05736926570534706,
0.09324377775192261,
0.017503922805190086,
0.15200082957744598,
-0.004327565431594849,
0.07368715852499008,
-0.06430529057979584,
0.002900370629504323,
-0.031497515738010406,
-0.18479296565055847,
-0.03858397901058197,
-0.04583295062184334,
0.012515569105744362,
-0.006835192907601595,
0.2555631995201111,
0.04336102306842804,
0.01708845980465412,
0.04016496613621712,
0.09534046053886414,
-0.011532729491591454,
0.017750654369592667,
0.15677224099636078,
0.09642043709754944,
-0.01634051837027073,
-0.03656552731990814,
0.08647118508815765,
0.016823289915919304,
-0.0778827890753746,
0.11005629599094391,
0.03772580996155739,
-0.04265972599387169,
0.05467662960290909,
0.032995134592056274,
0.01356444600969553,
-0.16403166949748993,
-0.1561090648174286,
-0.059449054300785065,
0.07617752999067307,
0.023158937692642212,
0.02678116224706173,
0.11130081862211227,
-0.028325339779257774,
0.02580459602177143,
-0.015071915462613106,
-0.02922098897397518,
-0.18761996924877167,
-0.12038294225931168,
-0.09932973980903625,
-0.12261207401752472,
0.011479394510388374,
-0.03784283250570297,
-0.028312478214502335,
0.11179319769144058,
0.049811072647571564,
-0.024361172690987587,
0.05118126422166824,
0.0034409374929964542,
-0.011275487020611763,
0.008979369886219501,
-0.01128071453422308,
0.010722801089286804,
-0.01220316719263792,
-0.014854183420538902,
-0.13901731371879578,
-0.02910418063402176,
-0.04148200526833534,
0.005580550990998745,
-0.04760715737938881,
0.022339073941111565,
-0.09673412889242172,
-0.11421101540327072,
-0.04191211983561516,
0.023931344971060753,
-0.05582542344927788,
0.08719328790903091,
-0.005910171195864677,
0.02947339415550232,
0.018003767356276512,
0.14212849736213684,
-0.07178422808647156,
-0.08833976089954376,
-0.047104042023420334,
0.24008573591709137,
0.05422874167561531,
0.09437024593353271,
0.007490250747650862,
0.011500077322125435,
-0.08013186603784561,
0.277520090341568,
0.2691807746887207,
-0.03454088047146797,
0.05437979847192764,
0.020099295303225517,
0.014366094022989273,
0.08129112422466278,
0.12801529467105865,
0.0743185356259346,
0.20950961112976074,
-0.08197618275880814,
-0.052265483886003494,
-0.051328178495168686,
-0.016310935840010643,
-0.11299603432416916,
0.0442340113222599,
0.028652755543589592,
-0.05005773529410362,
-0.06117720156908035,
0.07672857493162155,
-0.1584158092737198,
0.17652533948421478,
0.05530504137277603,
-0.17637746036052704,
-0.0781833678483963,
-0.02176763489842415,
0.1454465687274933,
-0.003720478154718876,
0.06764408946037292,
-0.04851401224732399,
-0.0823097974061966,
0.05232630670070648,
0.010331254452466965,
-0.22684168815612793,
-0.05646951496601105,
0.08493104577064514,
0.016064852476119995,
0.019817477092146873,
-0.032334569841623306,
0.07010794430971146,
0.07780664414167404,
0.05638008192181587,
-0.058445267379283905,
0.020888328552246094,
-0.00633151875808835,
-0.08150964975357056,
-0.00883753877133131,
-0.0008433987968601286,
0.0026467349380254745,
-0.06789134442806244,
0.03673473745584488,
-0.13965782523155212,
0.017946675419807434,
-0.07979290932416916,
-0.01338474452495575,
-0.015799034386873245,
0.02234521508216858,
-0.017149455845355988,
0.05813609063625336,
0.06571108847856522,
0.0014315495500341058,
-0.04241492599248886,
-0.06686784327030182,
-0.009505338035523891,
0.03682727366685867,
-0.11386525630950928,
-0.1286592334508896,
-0.0847516804933548,
-0.04587195813655853,
0.06421384960412979,
-0.0035427496768534184,
-0.05032660812139511,
-0.049098946154117584,
-0.09952948987483978,
0.005699475761502981,
-0.14046946167945862,
0.06746573001146317,
0.05839042738080025,
0.03217732906341553,
-0.007235072087496519,
-0.03762112557888031,
0.02165811136364937,
0.03594008460640907,
-0.13543765246868134,
-0.10064350068569183
] |
null | null |
transformers
|
# Harry Potter DialoGPT Model
|
{"tags": ["conversational"]}
|
text-generation
|
cutiebunny639/DialoGPT-small-harry
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Harry Potter DialoGPT Model
|
[
"# Harry Potter DialoGPT Model"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Harry Potter DialoGPT Model"
] |
[
51,
8
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Harry Potter DialoGPT Model"
] |
[
-0.0009023238671943545,
0.07815738022327423,
-0.006546166725456715,
0.07792752981185913,
0.10655936598777771,
0.048972971737384796,
0.17639793455600739,
0.12185695022344589,
0.016568755730986595,
-0.04774167761206627,
0.11647630482912064,
0.2130284160375595,
-0.002118367003276944,
0.024608047679066658,
-0.05022026598453522,
-0.3065771162509918,
0.0474756620824337,
0.014356585219502449,
-0.07174845039844513,
0.11724270135164261,
0.09064973145723343,
-0.046179238706827164,
0.08330509811639786,
-0.009135239757597446,
-0.13198648393154144,
-0.039482954889535904,
0.019292812794446945,
-0.11745545268058777,
0.1662212759256363,
0.05298272892832756,
0.02469746209681034,
-0.008447164669632912,
-0.06598151475191116,
-0.15036040544509888,
0.037190426141023636,
-0.027472136542201042,
-0.01080626156181097,
0.05462246760725975,
0.023526115342974663,
-0.07521048933267593,
0.170567125082016,
0.17678891122341156,
0.0833497866988182,
0.0349111407995224,
-0.14917024970054626,
-0.045548245310783386,
0.008950977586209774,
0.05421316996216774,
-0.017893504351377487,
0.09349167346954346,
-0.019903047010302544,
0.11801653355360031,
-0.04491448402404785,
0.09210366010665894,
0.15255063772201538,
-0.4016275703907013,
-0.027563704177737236,
0.08920855820178986,
0.05989706888794899,
0.12076901644468307,
-0.10560955852270126,
0.03972794860601425,
-0.0039703017100691795,
0.01236654631793499,
-0.014540530741214752,
-0.08304883539676666,
-0.07308239489793777,
0.032504837960004807,
-0.1272556483745575,
0.008525865152478218,
0.23756256699562073,
-0.10643257945775986,
0.037069112062454224,
-0.09791990369558334,
-0.07414398342370987,
0.048336777836084366,
-0.053761593997478485,
-0.081727035343647,
-0.054839808493852615,
0.06347949057817459,
0.004366500303149223,
-0.06301609426736832,
-0.08326146006584167,
-0.0006536149303428829,
-0.12781435251235962,
0.17595994472503662,
0.061243366450071335,
0.041611745953559875,
-0.21322020888328552,
0.08940251916646957,
0.04477722570300102,
-0.04711297154426575,
0.007116159424185753,
-0.11796226352453232,
0.04023287072777748,
0.005483259446918964,
-0.03256071358919144,
-0.021854614838957787,
0.0393419973552227,
0.13909944891929626,
-0.01777748204767704,
0.03252175822854042,
0.006831915583461523,
0.05811219662427902,
0.08162496984004974,
0.02222144603729248,
0.019291909411549568,
-0.0818009302020073,
0.019385190680623055,
-0.08128736168146133,
-0.0030400939285755157,
-0.048940129578113556,
-0.17071883380413055,
-0.07477642595767975,
0.052610911428928375,
0.020047198981046677,
0.03746970370411873,
0.08054786175489426,
-0.0017944995779544115,
-0.05560554191470146,
0.03284840285778046,
0.01671096310019493,
-0.020622212439775467,
-0.010361049324274063,
-0.02412462793290615,
0.19123271107673645,
0.019619356840848923,
0.014111656695604324,
-0.12379156798124313,
0.10023640841245651,
-0.08179095387458801,
0.0037731381598860025,
0.02743307314813137,
-0.04204464703798294,
-0.004716555587947369,
0.02917117439210415,
0.023101668804883957,
-0.1252521574497223,
-0.1099385917186737,
-0.0030569476075470448,
-0.012054097838699818,
-0.036421261727809906,
-0.10490952432155609,
-0.08483029156923294,
-0.012153145857155323,
0.0449371263384819,
-0.013397793285548687,
0.007936403155326843,
-0.05143149942159653,
0.0985720232129097,
-0.0514979362487793,
0.09873400628566742,
-0.08342572301626205,
0.06359215080738068,
-0.09124887734651566,
-0.061886150389909744,
-0.11452563107013702,
0.05216052383184433,
0.012905281968414783,
0.066250741481781,
0.016998225823044777,
-0.044836658984422684,
-0.014836243353784084,
0.05253177136182785,
-0.07656687498092651,
0.1940697431564331,
-0.041674621403217316,
-0.12459053844213486,
0.24146439135074615,
-0.09138800948858261,
-0.1802034229040146,
0.12973085045814514,
-0.022254703566432,
0.08523941785097122,
0.12802475690841675,
0.20380465686321259,
-0.00019822151807602495,
-0.01302915159612894,
0.07281201332807541,
0.07031642645597458,
-0.09803894907236099,
0.06239739805459976,
0.029653839766979218,
-0.008071083575487137,
-0.08906278014183044,
0.05762826278805733,
0.046033453196287155,
-0.010650773532688618,
-0.035073768347501755,
-0.001896020956337452,
-0.012895751744508743,
-0.022185025736689568,
0.14126582443714142,
-0.02006692811846733,
0.1300428807735443,
-0.06926563382148743,
-0.03515486419200897,
-0.009500149637460709,
0.03533667325973511,
-0.04091939330101013,
0.08151165395975113,
-0.0436173714697361,
0.10586477071046829,
0.09034156054258347,
0.053724925965070724,
-0.13120363652706146,
0.00466286763548851,
-0.015246815048158169,
0.17014820873737335,
0.08964069187641144,
0.05222717300057411,
0.06265474855899811,
-0.0020888058934360743,
-0.06708643585443497,
0.045407816767692566,
0.13778303563594818,
-0.037020038813352585,
-0.12218865007162094,
-0.1755627691745758,
0.051157694309949875,
-0.045444171875715256,
0.10855234414339066,
-0.10010123997926712,
0.022670533508062363,
-0.055906031280756,
0.07772238552570343,
-0.024998966604471207,
0.020512236282229424,
-0.0013405600329861045,
-0.021700702607631683,
-0.08356887847185135,
-0.002377772703766823,
0.08597290515899658,
-0.02048647589981556,
-0.06707409024238586,
0.16556480526924133,
-0.16400809586048126,
0.1631954461336136,
0.2116095870733261,
-0.28542569279670715,
-0.005696662236005068,
-0.15163889527320862,
-0.0208092350512743,
0.019645055755972862,
0.07834604382514954,
0.026225795969367027,
0.2044338881969452,
-0.012928472831845284,
0.16565458476543427,
-0.05699567869305611,
-0.07730039209127426,
-0.06881127506494522,
-0.048101142048835754,
0.013522743247449398,
0.09095205366611481,
0.04542696103453636,
-0.11962861567735672,
0.13119758665561676,
0.1054433062672615,
0.06484298408031464,
0.12711186707019806,
0.1030748188495636,
-0.008113685995340347,
0.07252490520477295,
-0.03624548763036728,
-0.03462279960513115,
-0.09254947304725647,
-0.30446043610572815,
-0.04840317741036415,
0.0939924493432045,
0.007963384501636028,
0.09285714477300644,
-0.0919896736741066,
-0.03311870992183685,
0.006042704917490482,
0.009473444893956184,
0.028337622061371803,
0.09653715789318085,
0.013490920886397362,
0.15320514142513275,
-0.008011690340936184,
-0.03430786728858948,
0.05891305208206177,
0.017982570454478264,
-0.09147711098194122,
0.17280617356300354,
-0.17050009965896606,
-0.27190929651260376,
-0.06990014761686325,
-0.21745692193508148,
-0.013139115646481514,
0.05258983001112938,
0.0786920040845871,
-0.11818131804466248,
-0.018352627754211426,
-0.006239492911845446,
0.05685517191886902,
-0.2425733357667923,
0.0004911290016025305,
-0.1354890614748001,
0.0501418262720108,
-0.1974833607673645,
-0.09718500077724457,
-0.02271542325615883,
-0.013450481928884983,
-0.0464281290769577,
0.13365240395069122,
-0.1448695808649063,
-0.011572926305234432,
0.2329535037279129,
0.032479673624038696,
0.027794739231467247,
-0.05020907148718834,
0.19788463413715363,
-0.0958966314792633,
-0.023973820731043816,
0.11024576425552368,
-0.05038975924253464,
0.04834126681089401,
0.06649978458881378,
-0.012981836684048176,
-0.08557141572237015,
0.023789849132299423,
-0.068336620926857,
-0.03150583803653717,
-0.27926525473594666,
-0.0930178239941597,
-0.09319330751895905,
0.11305391043424606,
0.04079577326774597,
0.06421639025211334,
0.16545771062374115,
0.05191578343510628,
-0.024325082078576088,
-0.03006586618721485,
0.11609793454408646,
0.12905290722846985,
0.2277202159166336,
-0.06067761778831482,
0.10221996158361435,
0.009445492178201675,
-0.08203992247581482,
0.06062209978699684,
0.056782789528369904,
0.06324724853038788,
0.02584579586982727,
0.03694582358002663,
-0.030939655378460884,
0.1121687963604927,
0.12571842968463898,
0.05258069559931755,
0.0481170229613781,
0.0002127334737451747,
-0.0561506561934948,
-0.008168719708919525,
-0.05726633965969086,
0.06774696707725525,
0.061340972781181335,
-0.12918008863925934,
-0.08061543852090836,
0.0011613310780376196,
0.06660808622837067,
-0.016230419278144836,
0.06823775917291641,
-0.13560809195041656,
-0.03582429885864258,
0.0790911465883255,
-0.07693151384592056,
-0.14156894385814667,
0.11972879618406296,
-0.026570770889520645,
-0.19904157519340515,
0.05265914276242256,
0.007704653777182102,
0.0908159390091896,
-0.06360849738121033,
0.05343840271234512,
-0.13023801147937775,
-0.12935101985931396,
-0.018437571823596954,
0.07945099472999573,
-0.3450873792171478,
0.13536721467971802,
-0.013286802917718887,
-0.02876877970993519,
-0.06474969536066055,
-0.02640824392437935,
0.013905409723520279,
0.12719078361988068,
0.08667250722646713,
0.0008821099763736129,
0.0991629809141159,
0.03823768347501755,
0.04188435152173042,
-0.002011700300499797,
0.10950417071580887,
0.0050011589191854,
0.004797275178134441,
-0.04982118681073189,
0.007274609990417957,
-0.05164213851094246,
-0.07472953200340271,
0.08393982797861099,
-0.20678792893886566,
0.09087453782558441,
-0.03378438204526901,
0.08427679538726807,
0.04304937273263931,
-0.018965769559144974,
-0.1001204177737236,
0.19745583832263947,
-0.012206900864839554,
-0.11405988782644272,
-0.07517550885677338,
-0.02810264565050602,
0.09103139489889145,
-0.013817726634442806,
0.012886416167020798,
-0.045470476150512695,
0.032183047384023666,
-0.1263762265443802,
-0.1597503274679184,
0.08734500408172607,
-0.04441224783658981,
-0.10894393920898438,
-0.025462759658694267,
0.20382575690746307,
-0.007266622502356768,
0.08242089301347733,
0.01605331338942051,
0.010653935372829437,
-0.18066231906414032,
-0.04018142446875572,
0.02645772136747837,
-0.0016437612939625978,
0.005979063920676708,
0.047698814421892166,
0.019091911613941193,
0.06207629665732384,
-0.1069745197892189,
-0.013920160941779613,
0.3158324360847473,
0.15978319942951202,
-0.00912671908736229,
0.14943915605545044,
0.1093616932630539,
-0.08669080585241318,
-0.17238758504390717,
-0.1171615794301033,
-0.1210922971367836,
-0.08425768464803696,
-0.10681738704442978,
-0.1525043100118637,
0.09535340964794159,
-0.03392014652490616,
0.03498011827468872,
0.14615866541862488,
-0.280263751745224,
-0.10949636250734329,
0.13820378482341766,
0.010744688101112843,
0.3510635495185852,
-0.12303631007671356,
-0.044944874942302704,
-0.06214528530836105,
-0.16933435201644897,
0.08021392673254013,
-0.031203703954815865,
0.11581093072891235,
-0.0744495838880539,
0.19395925104618073,
0.01719796098768711,
0.014287159778177738,
0.0916559100151062,
0.05038322135806084,
-0.05808406323194504,
-0.07368700206279755,
-0.10248131304979324,
0.010812131687998772,
0.03546109423041344,
0.010252019390463829,
-0.008802837692201138,
0.0211968794465065,
-0.11341743916273117,
-0.050869911909103394,
-0.06302189081907272,
0.0072614275850355625,
-0.01001308299601078,
-0.042155615985393524,
-0.05533592775464058,
-0.022557416930794716,
-0.020093943923711777,
0.02266426384449005,
0.14185629785060883,
-0.07527699321508408,
0.18586260080337524,
0.02357078716158867,
0.1586609035730362,
-0.11956068128347397,
-0.06724818795919418,
-0.029193658381700516,
-0.05280323326587677,
0.06468886137008667,
-0.08884575963020325,
-0.027708567678928375,
0.1332162618637085,
-0.01903904788196087,
0.04655366763472557,
0.12936700880527496,
0.02046884410083294,
0.015383756719529629,
0.034968774765729904,
-0.2578005790710449,
-0.07463036477565765,
-0.03505445644259453,
-0.012416874058544636,
0.05272092670202255,
0.05525677278637886,
0.19735674560070038,
-0.03551921248435974,
-0.08521962910890579,
0.020131373777985573,
0.02735883742570877,
-0.02776256389915943,
0.10749414563179016,
0.019579345360398293,
-0.004837906453758478,
-0.16151933372020721,
0.08257976174354553,
-0.005964108742773533,
-0.08297000825405121,
0.028665626421570778,
0.2024049311876297,
-0.12141239643096924,
-0.10309756547212601,
-0.06804922968149185,
0.07315051555633545,
-0.09220825880765915,
0.016043387353420258,
-0.005091092549264431,
-0.1521538347005844,
0.06916408240795135,
0.07598215341567993,
0.04075418785214424,
0.06513199955224991,
-0.11743064224720001,
-0.015730571001768112,
-0.04170290008187294,
-0.002195435343310237,
0.03521120920777321,
0.01863143965601921,
-0.057492829859256744,
0.15846455097198486,
-0.0676199421286583,
0.08538917452096939,
-0.0744810476899147,
-0.1058846190571785,
-0.1395980566740036,
0.04660497233271599,
-0.08038312196731567,
-0.07247276604175568,
-0.12832807004451752,
-0.052204377949237823,
-0.0067099276930093765,
-0.03388519585132599,
0.006552806124091148,
-0.06627799570560455,
-0.10922821611166,
0.01822470687329769,
-0.00743203004822135,
-0.009385870769619942,
-0.06096754968166351,
0.026706209406256676,
0.06246216222643852,
-0.039788868278265,
0.15730851888656616,
0.22509248554706573,
-0.13591648638248444,
0.11564400047063828,
-0.09797432273626328,
-0.105463907122612,
0.046008042991161346,
0.009427277371287346,
0.03594303876161575,
0.0503489226102829,
-0.03594081476330757,
0.0044484552927315235,
0.03905477747321129,
0.08074651658535004,
0.08456914126873016,
-0.06776505708694458,
0.020801106467843056,
-0.05122765153646469,
-0.14904099702835083,
-0.016655439510941505,
-0.0464773029088974,
0.06876829266548157,
-0.006725262850522995,
0.11020535975694656,
-0.0515950471162796,
0.07739507406949997,
-0.07558431476354599,
0.050614211708307266,
0.021146971732378006,
-0.14688286185264587,
-0.006612539757043123,
-0.07093682140111923,
0.042144812643527985,
-0.008834975771605968,
0.20241086184978485,
-0.03228091076016426,
0.010342049412429333,
0.033811055123806,
0.06203942745923996,
-0.01957780309021473,
0.009357001632452011,
0.2014283686876297,
0.12640917301177979,
-0.08496357500553131,
-0.02679651789367199,
0.06793134659528732,
0.07248228788375854,
0.07093550264835358,
0.10807815194129944,
-0.015352966263890266,
0.028434239327907562,
0.07829629629850388,
-0.060215238481760025,
0.07576877623796463,
-0.08603982627391815,
-0.11668483167886734,
0.05793621391057968,
0.012955795042216778,
-0.055695828050374985,
0.20305177569389343,
0.19142870604991913,
-0.026278704404830933,
0.018410727381706238,
-0.0029499190859496593,
-0.10117456316947937,
-0.15619947016239166,
-0.05423750728368759,
-0.07170962542295456,
-0.1319410353899002,
-0.004549739416688681,
-0.16646917164325714,
0.022016216069459915,
-0.01132756657898426,
0.09506805986166,
-0.06855440139770508,
-0.01345991250127554,
0.1364889293909073,
-0.1055467277765274,
0.0847758799791336,
-0.024517204612493515,
0.07877567410469055,
-0.03746940940618515,
-0.018209461122751236,
-0.10342709720134735,
0.007514837197959423,
0.01131442841142416,
0.06840907037258148,
-0.10897937417030334,
0.02432350255548954,
-0.12208317965269089,
-0.08617185056209564,
-0.026142612099647522,
0.09279687702655792,
-0.0403008833527565,
0.15116846561431885,
0.02645145356655121,
-0.06710928678512573,
-0.004313822835683823,
0.2646709978580475,
-0.08046227693557739,
-0.08319197595119476,
-0.030799202620983124,
0.2152107208967209,
0.04053696244955063,
0.06396269053220749,
0.019140036776661873,
0.038027774542570114,
-0.07184682041406631,
0.2957373559474945,
0.34401440620422363,
-0.1318037211894989,
-0.007773484103381634,
0.04225075617432594,
0.04406323283910751,
0.14687567949295044,
0.07998795062303543,
0.11360671371221542,
0.2849363386631012,
-0.09197647124528885,
0.016657205298542976,
-0.04230864346027374,
-0.01424806285649538,
-0.06908884644508362,
0.045314885675907135,
0.08216670155525208,
-0.09241747111082077,
-0.022950593382120132,
0.08125471323728561,
-0.29741767048835754,
0.10791494697332382,
-0.15600289404392242,
-0.14948409795761108,
-0.05027429759502411,
-0.008771711029112339,
0.014683255925774574,
0.019041186198592186,
0.09663030505180359,
0.025651484727859497,
-0.07275258749723434,
0.07816889137029648,
0.024486342445015907,
-0.23020237684249878,
-0.01345184724777937,
0.1456068754196167,
-0.06789913028478622,
-0.025938833132386208,
-0.021313713863492012,
0.051610056310892105,
0.05763651058077812,
0.09027529507875443,
-0.03809558227658272,
-0.0746568813920021,
-0.007141788024455309,
-0.022818787023425102,
0.01914946548640728,
0.0597183033823967,
0.06841408461332321,
-0.0920223817229271,
0.1167774423956871,
-0.07350476831197739,
0.0650370642542839,
0.037623800337314606,
-0.022277191281318665,
0.0018526542698964477,
0.013183658011257648,
-0.06512464582920074,
0.05533479526638985,
0.1295643299818039,
-0.025459708645939827,
-0.002524374984204769,
-0.028180841356515884,
-0.0767761766910553,
-0.024015206843614578,
-0.04643676429986954,
-0.09101243317127228,
-0.18130090832710266,
-0.12738600373268127,
0.041754670441150665,
-0.03240608796477318,
-0.2046082615852356,
0.0060346988029778,
-0.1128578633069992,
0.03700976446270943,
-0.14154092967510223,
0.10004086047410965,
0.07216610759496689,
0.004716616589576006,
0.006774604320526123,
0.0675399899482727,
0.045677728950977325,
0.14796748757362366,
-0.16543124616146088,
-0.04919974133372307
] |
null | null |
transformers
|
**Disclaimer**: *This model is still under testing and may change in the future, we will try to keep backwards compatibility. For any questions reach us at [email protected]*
# MediaWatch News Topics (Greek)
Fine-tuned model for multi-label text-classification (SequenceClassification), based on [roberta-el-news](https://huggingface.co/cvcio/roberta-el-news), using [Hugging Face's](https://huggingface.co/) [Transformers](https://github.com/huggingface/transformers) library. This model is to classify news in real-time on upto 33 topics including: *AFFAIRS*, *AGRICULTURE*, *ARTS_AND_CULTURE*, *BREAKING_NEWS*, *BUSINESS*, *COVID*, *ECONOMY*, *EDUCATION*, *ELECTIONS*, *ENTERTAINMENT*, *ENVIRONMENT*, *FOOD*, *HEALTH*, *INTERNATIONAL*, *LAW_AND_ORDER*, *MILITARY*, *NON_PAPER*, *OPINION*, *POLITICS*, *REFUGEE*, *REGIONAL*, *RELIGION*, *SCIENCE*, *SOCIAL_MEDIA*, *SOCIETY*, *SPORTS*, *TECH*, *TOURISM*, *TRANSPORT*, *TRAVEL*, *WEATHER*, *CRIME*, *JUSTICE*.
## How to use
You can use this model directly with a pipeline for text-classification:
```python
from transformers import pipeline
pipe = pipeline(
task="text-classification",
model="cvcio/mediawatch-el-topics",
tokenizer="cvcio/roberta-el-news" # or cvcio/mediawatch-el-topics
)
topics = pipe(
"Η βιασύνη αρκετών χωρών να άρουν τους περιορισμούς κατά του κορονοϊού, "+
"αν όχι να κηρύξουν το τέλος της πανδημίας, με το σκεπτικό ότι έφτασε "+
"πλέον η ώρα να συμβιώσουμε με την Covid-19, έχει κάνει μερικούς πιο "+
"επιφυλακτικούς επιστήμονες να προειδοποιούν ότι πρόκειται μάλλον "+
"για «ενδημική αυταπάτη» και ότι είναι πρόωρη τέτοια υπερβολική "+
"χαλάρωση. Καθώς τα κρούσματα της Covid-19, μετά το αιφνιδιαστικό "+
"μαζικό κύμα της παραλλαγής Όμικρον, εμφανίζουν τάση υποχώρησης σε "+
"Ευρώπη και Βόρεια Αμερική, όπου περισσεύει η κόπωση μεταξύ των "+
"πολιτών μετά από δύο χρόνια πανδημίας, ειδικοί και μη αδημονούν να "+
"«ξεμπερδέψουν» με τον κορονοϊό.",
padding=True,
truncation=True,
max_length=512,
return_all_scores=True
)
print(topics)
# outputs
[
[
{'label': 'AFFAIRS', 'score': 0.0018806682201102376},
{'label': 'AGRICULTURE', 'score': 0.00014653144171461463},
{'label': 'ARTS_AND_CULTURE', 'score': 0.0012948638759553432},
{'label': 'BREAKING_NEWS', 'score': 0.0001729220530251041},
{'label': 'BUSINESS', 'score': 0.0028276608791202307},
{'label': 'COVID', 'score': 0.4407998025417328},
{'label': 'ECONOMY', 'score': 0.039826102554798126},
{'label': 'EDUCATION', 'score': 0.0019098613411188126},
{'label': 'ELECTIONS', 'score': 0.0003333651984576136},
{'label': 'ENTERTAINMENT', 'score': 0.004249618388712406},
{'label': 'ENVIRONMENT', 'score': 0.0015828514005988836},
{'label': 'FOOD', 'score': 0.0018390495097264647},
{'label': 'HEALTH', 'score': 0.1204477995634079},
{'label': 'INTERNATIONAL', 'score': 0.25892165303230286},
{'label': 'LAW_AND_ORDER', 'score': 0.07646272331476212},
{'label': 'MILITARY', 'score': 0.00033025629818439484},
{'label': 'NON_PAPER', 'score': 0.011991199105978012},
{'label': 'OPINION', 'score': 0.16166265308856964},
{'label': 'POLITICS', 'score': 0.0008890336030162871},
{'label': 'REFUGEE', 'score': 0.0011504743015393615},
{'label': 'REGIONAL', 'score': 0.0008734092116355896},
{'label': 'RELIGION', 'score': 0.0009001944563351572},
{'label': 'SCIENCE', 'score': 0.05075162276625633},
{'label': 'SOCIAL_MEDIA', 'score': 0.00039615994319319725},
{'label': 'SOCIETY', 'score': 0.0043518817983567715},
{'label': 'SPORTS', 'score': 0.002416545059531927},
{'label': 'TECH', 'score': 0.0007818648009561002},
{'label': 'TOURISM', 'score': 0.011870541609823704},
{'label': 'TRANSPORT', 'score': 0.0009422845905646682},
{'label': 'TRAVEL', 'score': 0.03004464879631996},
{'label': 'WEATHER', 'score': 0.00040286066359840333},
{'label': 'CRIME', 'score': 0.0005416403291746974},
{'label': 'JUSTICE', 'score': 0.000990519649349153}
]
]
```
## Labels
All labels, except *NON_PAPER*, retrieved by source articles during the data collection step, without any preprocessing, assuming that journalists and newsrooms assign correct tags to the articles. We disregarded all articles with more than 6 tags to reduce bias and tag manipulation.
| label | roc_auc | samples |
|-------:|--------:|--------:|
| AFFAIRS | 0.9872 | 6,314 |
| AGRICULTURE | 0.9799 | 1,254 |
| ARTS_AND_CULTURE | 0.9838 | 15,968 |
| BREAKING_NEWS | 0.9675 | 827 |
| BUSINESS | 0.9811 | 6,507 |
| COVID | 0.9620 | 50,000 |
| CRIME | 0.9885 | 34,421 |
| ECONOMY | 0.9765 | 45,474 |
| EDUCATION | 0.9865 | 10,111 |
| ELECTIONS | 0.9940 | 7,571 |
| ENTERTAINMENT | 0.9925 | 23,323 |
| ENVIRONMENT | 0.9847 | 23,060 |
| FOOD | 0.9934 | 3,712 |
| HEALTH | 0.9723 | 16,852 |
| INTERNATIONAL | 0.9624 | 50,000 |
| JUSTICE | 0.9862 | 4,860 |
| LAW_AND_ORDER | 0.9177 | 50,000 |
| MILITARY | 0.9838 | 6,536 |
| NON_PAPER | 0.9595 | 4,589 |
| OPINION | 0.9624 | 6,296 |
| POLITICS | 0.9773 | 50,000 |
| REFUGEE | 0.9949 | 4,536 |
| REGIONAL | 0.9520 | 50,000 |
| RELIGION | 0.9922 | 11,533 |
| SCIENCE | 0.9837 | 1,998 |
| SOCIAL_MEDIA | 0.991 | 6,212 |
| SOCIETY | 0.9439 | 50,000 |
| SPORTS | 0.9939 | 31,396 |
| TECH | 0.9923 | 8,225 |
| TOURISM | 0.9900 | 8,081 |
| TRANSPORT | 0.9879 | 3,211 |
| TRAVEL | 0.9832 | 4,638 |
| WEATHER | 0.9950 | 19,931 |
| loss | 0.0533 | - |
| roc_auc | 0.9855 | - |
## Pretraining
The model was pretrained using an NVIDIA A10 GPU for 15 epochs (~ approx 59K steps, 8 hours training) with a batch size of 128. The optimizer used is Adam with a learning rate of 1e-5, and weight decay 0.01. We used roc_auc_micro to evaluate the results.
### Framework versions
- Transformers 4.13.0
- Pytorch 1.9.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
## Authors
Dimitris Papaevagelou - [@andefined](https://github.com/andefined)
## About Us
[Civic Information Office](https://cvcio.org/) is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
|
{"language": "el", "license": "gpl-3.0", "tags": ["roberta", "Greek", "news", "transformers", "text-classification"], "pipeline_tag": "text-classification", "widget": [{"text": "\u03a0\u03b1\u03c1\u2019 \u03bf\u03bb\u03af\u03b3\u03bf\u03bd \u00ab\u03b8\u03b5\u03c1\u03bc\u03cc\u00bb \u03b5\u03c0\u03b5\u03b9\u03c3\u03cc\u03b4\u03b9\u03bf \u03c4\u03bf\u03c5\u03c1\u03ba\u03b9\u03ba\u03bf\u03cd \u03c0\u03bf\u03bb\u03b5\u03bc\u03b9\u03ba\u03bf\u03cd \u03c0\u03bb\u03bf\u03af\u03bf\u03c5 \u03bc\u03b5 \u03b5\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03cc \u03c9\u03ba\u03b5\u03b1\u03bd\u03bf\u03b3\u03c1\u03b1\u03c6\u03b9\u03ba\u03cc \u03c3\u03c4\u03b7\u03bd \u03c0\u03b5\u03c1\u03b9\u03bf\u03c7\u03ae \u03bc\u03b5\u03c4\u03b1\u03be\u03cd \u03a1\u03cc\u03b4\u03bf\u03c5 \u03ba\u03b1\u03b9 \u039a\u03b1\u03c3\u03c4\u03b5\u03bb\u03cc\u03c1\u03b9\u03b6\u03bf\u03c5, \u03c3\u03c4\u03bf \u03b4\u03b9\u03ac\u03c3\u03c4\u03b7\u03bc\u03b1 20-23 \u03a3\u03b5\u03c0\u03c4\u03b5\u03bc\u03b2\u03c1\u03af\u03bf\u03c5, \u03b1\u03c0\u03bf\u03ba\u03ac\u03bb\u03c5\u03c8\u03b5 \u03c4\u03bf \u039f\u03a1\u0395\u039d. \u03a3\u03cd\u03bc\u03c6\u03c9\u03bd\u03b1 \u03bc\u03b5 \u03c0\u03bb\u03b7\u03c1\u03bf\u03c6\u03bf\u03c1\u03af\u03b5\u03c2 \u03c0\u03bf\u03c5 \u03bc\u03b5\u03c4\u03ad\u03b4\u03c9\u03c3\u03b5 \u03c4\u03bf \u03ba\u03b5\u03bd\u03c4\u03c1\u03b9\u03ba\u03cc \u03b4\u03b5\u03bb\u03c4\u03af\u03bf \u03b5\u03b9\u03b4\u03ae\u03c3\u03b5\u03c9\u03bd, \u03cc\u03c4\u03b1\u03bd \u03c4\u03bf \u03b5\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03cc \u03b5\u03c1\u03b5\u03c5\u03bd\u03b7\u03c4\u03b9\u03ba\u03cc \u00ab \u0391\u0399\u0393\u0391\u0399\u039f \u00bb \u03c0\u03bf\u03c5 \u03b1\u03bd\u03ae\u03ba\u03b5\u03b9 \u03c3\u03c4\u03bf \u0395\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03cc \u039a\u03ad\u03bd\u03c4\u03c1\u03bf \u0398\u03b1\u03bb\u03b1\u03c3\u03c3\u03af\u03c9\u03bd \u0395\u03c1\u03b5\u03c5\u03bd\u03ce\u03bd \u03b2\u03b3\u03ae\u03ba\u03b5 \u03ad\u03be\u03c9 \u03b1\u03c0\u03cc \u03c4\u03b1 6 \u03bd.\u03bc, \u03c3\u03b5 \u03b4\u03b9\u03b5\u03b8\u03bd\u03ae \u03cd\u03b4\u03b1\u03c4\u03b1, \u03c4\u03bf \u03c0\u03c1\u03bf\u03c3\u03ad\u03b3\u03b3\u03b9\u03c3\u03b5 \u03c4\u03bf\u03c5\u03c1\u03ba\u03b9\u03ba\u03cc \u03c0\u03bf\u03bb\u03b5\u03bc\u03b9\u03ba\u03cc \u03c0\u03bb\u03bf\u03af\u03bf, \u03bf \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03ae\u03c4\u03b7\u03c2 \u03c4\u03bf\u03c5 \u03bf\u03c0\u03bf\u03af\u03bf\u03c5 \u03b6\u03ae\u03c4\u03b7\u03c3\u03b5 \u03b4\u03cd\u03bf \u03c6\u03bf\u03c1\u03ad\u03c2 \u03bc\u03ad\u03c3\u03c9 \u03b1\u03c3\u03c5\u03c1\u03bc\u03ac\u03c4\u03bf\u03c5 \u03bd\u03b1 \u03b5\u03bd\u03b7\u03bc\u03b5\u03c1\u03c9\u03b8\u03b5\u03af \u03b3\u03b9\u03b1 \u03c4\u03b1 \u03c3\u03c4\u03bf\u03b9\u03c7\u03b5\u03af\u03b1 \u03c4\u03bf\u03c5 \u03c0\u03bb\u03bf\u03af\u03bf\u03c5, \u03b1\u03bb\u03bb\u03ac \u03ba\u03b1\u03b9 \u03b3\u03b9\u03b1 \u03c4\u03b7\u03bd \u03b1\u03c0\u03bf\u03c3\u03c4\u03bf\u03bb\u03ae \u03c4\u03bf\u03c5. \u039f \u03c0\u03bb\u03bf\u03af\u03b1\u03c1\u03c7\u03bf\u03c2 \u03c4\u03bf\u03c5 \u03b5\u03bb\u03bb\u03b7\u03bd\u03b9\u03ba\u03bf\u03cd \u03b5\u03c1\u03b5\u03c5\u03bd\u03b7\u03c4\u03b9\u03ba\u03bf\u03cd \u03b4\u03b5\u03bd \u03b1\u03c0\u03ac\u03bd\u03c4\u03b7\u03c3\u03b5 \u03ba\u03b1\u03b9 \u03c4\u03b5\u03bb\u03b9\u03ba\u03ac \u03c4\u03bf \u03c4\u03bf\u03c5\u03c1\u03ba\u03b9\u03ba\u03cc \u03c0\u03bf\u03bb\u03b5\u03bc\u03b9\u03ba\u03cc \u03b1\u03c0\u03bf\u03bc\u03b1\u03ba\u03c1\u03cd\u03bd\u03b8\u03b7\u03ba\u03b5.", "example_title": "Topic AFFAIRS"}, {"text": "\u0397 \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03b7\u03c4\u03b9\u03ba\u03ae \u03b1\u03bd\u03b9\u03ba\u03b1\u03bd\u03cc\u03c4\u03b7\u03c4\u03b1 \u03bf\u03b4\u03b7\u03b3\u03b5\u03af \u03c4\u03b7\u03bd \u03c7\u03ce\u03c1\u03b1 \u03c3\u03c4\u03bf \u03c7\u03ac\u03bf\u03c2. \u0397 \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03b7\u03c3\u03b7 \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03b1\u03ba\u03b7 \u03b1\u03b4\u03c5\u03bd\u03b1\u03c4\u03b5\u03af \u03bd\u03b1 \u03b4\u03b9\u03b1\u03c7\u03b5\u03b9\u03c1\u03b9\u03c3\u03c4\u03b5\u03af \u03c4\u03b7\u03bd \u03c0\u03b1\u03bd\u03b4\u03b7\u03bc\u03af\u03b1. \u0394\u03b5\u03bd \u03bc\u03c0\u03bf\u03c1\u03b5\u03b9 \u03bf\u03cd\u03c4\u03b5 \u03bd\u03b1 \u03c0\u03b5\u03af\u03c3\u03b5\u03b9 \u03c4\u03bf\u03bd \u03ba\u03cc\u03c3\u03bc\u03bf \u03bd\u03b1 \u03b5\u03bc\u03b2\u03bf\u03bb\u03b9\u03b1\u03c3\u03c4\u03b5\u03af, \u03c0\u03bf\u03c5 \u03ae\u03c4\u03b1\u03bd \u03c4\u03bf \u03c0\u03b9\u03bf \u03b1\u03c0\u03bb\u03bf \u03c0\u03c1\u03ac\u03b3\u03bc\u03b1. \u03a3\u03b7\u03bc\u03b5\u03c1\u03b1 \u03bb\u03bf\u03b9\u03c0\u03cc\u03bd \u03c6\u03c4\u03ac\u03c3\u03b1\u03bc\u03b5 \u03c3\u03c4\u03bf \u03c3\u03b7\u03bc\u03b5\u03af\u03bf \u03bd\u03b1 \u03bc\u03b9\u03bb\u03ac\u03bc\u03b5 \u03b3\u03b9\u03b1 \u03b5\u03c0\u03b1\u03bd\u03b1\u03c6\u03bf\u03c1\u03ac \u03c4\u03b7\u03c2 \u03c7\u03c1\u03ae\u03c3\u03b7\u03c2 \u03bc\u03ac\u03c3\u03ba\u03b1\u03c2 \u03c3\u03b5 \u03b5\u03be\u03c9\u03c4\u03b5\u03c1\u03b9\u03ba\u03bf\u03cd\u03c2 \u03c7\u03ce\u03c1\u03bf\u03c5\u03c2 \u03b1\u03ba\u03cc\u03bc\u03b7 \u03ba\u03b1\u03b9 \u03cc\u03c0\u03bf\u03c5 \u03b4\u03b5\u03bd \u03c5\u03c0\u03ac\u03c1\u03c7\u03b5\u03b9 \u03c3\u03c5\u03b3\u03c7\u03c1\u03c9\u03c4\u03b9\u03c3\u03bc\u03cc\u03c2. \u03a3\u03c4\u03b9\u03c2 \u03c3\u03c5\u03b6\u03b7\u03c4\u03ae\u03c3\u03b5\u03b9\u03c2 \u03c4\u03c9\u03bd \u03b5\u03b9\u03b4\u03b9\u03ba\u03ce\u03bd \u03b8\u03b1 \u03b2\u03c1\u03b5\u03b8\u03b5\u03af \u03b5\u03c0\u03af\u03c3\u03b7\u03c2 \u03c4\u03bf \u03b5\u03bd\u03b4\u03b5\u03c7\u03cc\u03bc\u03b5\u03bd\u03bf \u03b3\u03b9\u03b1 \u03c4\u03bf\u03c0\u03b9\u03ba\u03ac lockdown \u03c3\u03b5 \u03c0\u03b5\u03c1\u03b9\u03bf\u03c7\u03ad\u03c2 \u03bc\u03b5 \u03b2\u03b1\u03c1\u03cd \u03b9\u03b9\u03ba\u03cc \u03c6\u03bf\u03c1\u03c4\u03af\u03bf \u03b3\u03b9\u03b1 \u03bd\u03b1 \u03bc\u03b7\u03bd \u03be\u03b5\u03c6\u03cd\u03b3\u03b5\u03b9 \u03b7 \u03ba\u03b1\u03c4\u03ac\u03c3\u03c4\u03b1\u03c3\u03b7, \u03b5\u03bd\u03ce \u03b8\u03b1 \u03c7\u03c1\u03b5\u03b9\u03ac\u03b6\u03b5\u03c4\u03b1\u03b9 \u03ba\u03ac\u03c0\u03bf\u03b9\u03bf\u03c2 \u03b3\u03b9\u03b1 \u03c4\u03b9\u03c2 \u03bc\u03b5\u03c4\u03b1\u03ba\u03b9\u03bd\u03ae\u03c3\u03b5\u03b9\u03c2 \u03c4\u03bf\u03c5 \u03b5\u03af\u03c4\u03b5 \u03c0\u03b9\u03c3\u03c4\u03bf\u03c0\u03bf\u03b9\u03b7\u03c4\u03b9\u03ba\u03cc \u03b5\u03bc\u03b2\u03bf\u03bb\u03b9\u03b1\u03c3\u03bc\u03bf\u03cd \u03ae \u03bd\u03cc\u03c3\u03b7\u03c3\u03b7\u03c2 \u03ba\u03b1\u03b9 \u03bf\u03b9 \u03b1\u03bd\u03b5\u03bc\u03b2\u03bf\u03bb\u03af\u03b1\u03c3\u03c4\u03bf\u03b9 rapid \u03ae \u03bc\u03bf\u03c1\u03b9\u03b1\u03ba\u03cc \u03c4\u03b5\u03c3\u03c4.", "example_title": "Topic COVID"}, {"text": "\u0397 \u00ab\u03c9\u03c1\u03b1\u03af\u03b1 \u0395\u03bb\u03ad\u03bd\u03b7\u00bb \u03b5\u03c0\u03ad\u03c3\u03c4\u03c1\u03b5\u03c8\u03b5 \u03c3\u03c4\u03b7\u03bd \u03c4\u03b7\u03bb\u03b5\u03cc\u03c1\u03b1\u03c3\u03b7, \u03bc\u03ad\u03c3\u03b1 \u03b1\u03c0\u03cc \u03c4\u03b7 \u03c3\u03c5\u03c7\u03bd\u03cc\u03c4\u03b7\u03c4\u03b1 \u03c4\u03bf\u03c5 MEGA \u03ba\u03b1\u03b9 \u03ac\u03c6\u03b7\u03c3\u03b5 \u03c4\u03b9\u03c2 \u03ba\u03b1\u03bb\u03cd\u03c4\u03b5\u03c1\u03b5\u03c2 \u03b5\u03bd\u03c4\u03c5\u03c0\u03ce\u03c3\u03b5\u03b9\u03c2. \u03a4\u03bf \u03c0\u03bb\u03b1\u03c4\u03cc \u03b1\u03c0\u03cc \u03c4\u03bf \u03bf\u03c0\u03bf\u03af\u03bf \u03b5\u03bc\u03c6\u03b1\u03bd\u03af\u03b6\u03b5\u03c4\u03b1\u03b9 \u03b7 \u0395\u03bb\u03ad\u03bd\u03b7 \u039c\u03b5\u03bd\u03b5\u03b3\u03ac\u03ba\u03b7 \u03ad\u03c7\u03b5\u03b9 \u03c6\u03c4\u03b9\u03b1\u03c7\u03c4\u03b5\u03af \u03b1\u03c0\u03cc \u03c4\u03b7\u03bd \u03b1\u03c1\u03c7\u03ae \u03b3\u03b9\u03b1 \u03c4\u03b7\u03bd \u03b5\u03ba\u03c0\u03bf\u03bc\u03c0\u03ae \u03c4\u03b7\u03c2. \u03a3\u03ae\u03bc\u03b5\u03c1\u03b1, \u03c3\u03c4\u03bf \u03ba\u03bb\u03b5\u03af\u03c3\u03b9\u03bc\u03bf \u03c4\u03b7\u03c2 \u03b5\u03ba\u03c0\u03bf\u03bc\u03c0\u03ae\u03c2 \u03b7 \u0395\u03bb\u03ad\u03bd\u03b7 \u03c0\u03ad\u03c1\u03b1\u03c3\u03b5 \u03b1\u03bd\u03ac\u03bc\u03b5\u03c3\u03b1 \u03b1\u03c0\u03cc \u03c4\u03b9\u03c2 \u03ba\u03ac\u03bc\u03b5\u03c1\u03b5\u03c2 \u03b3\u03b9\u03b1 \u03bd\u03b1 \u03bc\u03c0\u03b5\u03b9 \u03c3\u03c4\u03bf \u03ba\u03b1\u03bc\u03b1\u03c1\u03af\u03bd\u03b9 \u03c4\u03b7\u03c2 \u00ab\u039c\u03b7\u03bd \u03c4\u03c1\u03bf\u03bc\u03bf\u03ba\u03c1\u03b1\u03c4\u03b5\u03af\u03c3\u03c4\u03b5, \u03b5\u03af\u03bc\u03b1\u03b9 \u03b7 \u0395\u03bb\u03ad\u03bd\u03b7 \u039c\u03b5\u03bd\u03b5\u03b3\u03ac\u03ba\u03b7, \u03c4\u03b1 \u03ba\u03ac\u03bd\u03c9 \u03b1\u03c5\u03c4\u03ac. \u039c\u03b5 \u03c3\u03c5\u03b3\u03c7\u03c9\u03c1\u03b5\u03af\u03c4\u03b1\u03b9, \u03ad\u03c7\u03c9 \u03c8\u03c5\u03c7\u03bf\u03bb\u03bf\u03b3\u03b9\u03ba\u03ac \u03b1\u03bd \u03b4\u03b5\u03bd \u03b5\u03af\u03bc\u03b1\u03b9 \u03b5\u03bb\u03b5\u03cd\u03b8\u03b5\u03c1\u03b7\u00bb \u03b5\u03af\u03c0\u03b5 \u03b1\u03c1\u03c7\u03b9\u03ba\u03ac \u03b7 \u03c0\u03b1\u03c1\u03bf\u03c5\u03c3\u03b9\u03ac\u03c3\u03c4\u03c1\u03b9\u03b1 \u03c3\u03c4\u03bf\u03c5\u03c2 \u03c3\u03c5\u03bd\u03b5\u03c1\u03b3\u03ac\u03c4\u03b5\u03c2 \u03c4\u03b7\u03c2 \u03ba\u03b1\u03b9 \u03c0\u03c1\u03cc\u03c3\u03b8\u03b5\u03c3\u03b5 \u03c3\u03c4\u03b7 \u03c3\u03c5\u03bd\u03ad\u03c7\u03b5\u03b9\u03b1: \u00ab\u0397 \u0395\u03bb\u03ad\u03bd\u03b7 \u03bf\u03bb\u03bf\u03ba\u03bb\u03ae\u03c1\u03c9\u03c3\u03b5. \u039c\u03c0\u03bf\u03c1\u03b5\u03af\u03c4\u03b5 \u03bd\u03b1 \u03c3\u03c5\u03bd\u03b5\u03c7\u03af\u03c3\u03b5\u03c4\u03b5 \u03bc\u03b5 \u03c4\u03bf \u03c5\u03c0\u03cc\u03bb\u03bf\u03b9\u03c0\u03bf \u03c0\u03c1\u03cc\u03b3\u03c1\u03b1\u03bc\u03bc\u03b1 \u03c4\u03bf\u03c5 Mega. \u0395\u03b3\u03ce \u03b1\u03bd\u03bf\u03af\u03b3\u03c9 \u03c4\u03bf \u03ba\u03b1\u03bc\u03b1\u03c1\u03af\u03bd\u03b9, \u03b1\u03bd \u03bc\u03b5 \u03b1\u03c6\u03ae\u03c3\u03bf\u03c5\u03bd. \u039c\u03c0\u03b1\u03af\u03bd\u03c9 \u03ba\u03b1\u03bc\u03b1\u03c1\u03af\u03bd\u03b9\u00bb. \u0394\u03b5\u03af\u03c4\u03b5 \u03c4\u03bf \u03b1\u03c0\u03cc\u03c3\u03c0\u03b1\u03c3\u03bc\u03b1!", "example_title": "Topic ENTERTAINMENT"}, {"text": "\u0388\u03bd\u03b1 \u03b5\u03be\u03b1\u03b9\u03c1\u03b5\u03c4\u03b9\u03ba\u03ac \u03b5\u03bd\u03b4\u03b9\u03b1\u03c6\u03ad\u03c1\u03bf\u03bd \u00ab\u03ba\u03bf\u03c5\u03c4\u03c3\u03bf\u03bc\u03c0\u03bf\u03bb\u03b9\u03cc\u00bb \u03b5\u03bd\u03c4\u03cc\u03c0\u03b9\u03c3\u03b1\u03bd \u03bf\u03b9 \u03ba\u03b5\u03c1\u03b1\u03af\u03b5\u03c2 \u03c4\u03b7\u03c2 \u03c3\u03c4\u03ae\u03bb\u03b7\u03c2 \u03c0\u03ad\u03c1\u03b9\u03be \u03c4\u03bf\u03c5 \u039c\u03b5\u03b3\u03ac\u03c1\u03bf\u03c5 \u039c\u03b1\u03be\u03af\u03bc\u03bf\u03c5 : \u03c4\u03bf \u03ba\u03b1\u03c4\u03ac \u03c0\u03cc\u03c3\u03bf\u03bd, \u03b4\u03b7\u03bb\u03b1\u03b4\u03ae, \u03bf \u00ab\u03b5\u03be \u03b1\u03c0\u03bf\u03c1\u03c1\u03ae\u03c4\u03c9\u03bd\u00bb \u03c4\u03bf\u03c5 \u039a\u03c5\u03c1\u03b9\u03ac\u03ba\u03bf\u03c5 \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03ac\u03ba\u03b7 , \u0393\u03b9\u03ce\u03c1\u03b3\u03bf\u03c2 \u0393\u03b5\u03c1\u03b1\u03c0\u03b5\u03c4\u03c1\u03af\u03c4\u03b7\u03c2 \u03bc\u03b5\u03c4\u03ad\u03c7\u03b5\u03b9 \u03c3\u03c4\u03b7 \u03b4\u03b9\u03b1\u03c7\u03b5\u03af\u03c1\u03b9\u03c3\u03b7 \u03c4\u03b7\u03c2 \u03c0\u03b1\u03bd\u03b4\u03b7\u03bc\u03af\u03b1\u03c2 \u03ba\u03b1\u03b9 \u03c3\u03c4\u03b7\u03bd \u03b4\u03b9\u03b1\u03b4\u03b9\u03ba\u03b1\u03c3\u03af\u03b1 \u03bb\u03ae\u03c8\u03b7\u03c2 \u03b1\u03c0\u03bf\u03c6\u03ac\u03c3\u03b5\u03c9\u03bd. \u03a4\u03bf \u03b5\u03bd \u03bb\u03cc\u03b3\u03c9 \u00ab\u03ba\u03bf\u03c5\u03c4\u03c3\u03bf\u03bc\u03c0\u03bf\u03bb\u03b9\u03cc\u00bb \u03c0\u03c5\u03c1\u03bf\u03b4\u03cc\u03c4\u03b7\u03c3\u03b5 \u03c4\u03bf \u03b3\u03b5\u03b3\u03bf\u03bd\u03cc\u03c2 \u03cc\u03c4\u03b9 \u03c3\u03b5 \u03c3\u03b1\u03b2\u03b2\u03b1\u03c4\u03b9\u03ac\u03c4\u03b9\u03ba\u03b7 \u03b5\u03c6\u03b7\u03bc\u03b5\u03c1\u03af\u03b4\u03b1 \u03b4\u03b7\u03bc\u03bf\u03c3\u03b9\u03b5\u03cd\u03b8\u03b7\u03ba\u03b1\u03bd \u03c0\u03c1\u03bf\u03c7\u03b8\u03ad\u03c2 \u03b4\u03b7\u03bb\u03ce\u03c3\u03b5\u03b9\u03c2 \u03c4\u03bf\u03c5 \u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03bf\u03cd \u0395\u03c0\u03b9\u03ba\u03c1\u03b1\u03c4\u03b5\u03af\u03b1\u03c2 \u03bc\u03b5 \u03c4\u03b9\u03c2 \u03bf\u03c0\u03bf\u03af\u03b5\u03c2 \u03b1\u03c0\u03ad\u03ba\u03bb\u03b5\u03b9\u03b5 \u03ba\u03ac\u03b8\u03b5 \u03c3\u03b5\u03bd\u03ac\u03c1\u03b9\u03bf \u03bd\u03ad\u03c9\u03bd \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03c9\u03bd \u03bc\u03ad\u03c4\u03c1\u03c9\u03bd \u03ba\u03b1\u03b9 \u03c4\u03b7\u03bd \u03af\u03b4\u03b9\u03b1 \u03ce\u03c1\u03b1, \u03c4\u03bf \u039c\u03b1\u03be\u03af\u03bc\u03bf\u03c5 \u03b1\u03bd\u03ae\u03b3\u03b3\u03b5\u03bb\u03bb\u03b5\u2026 \u03ba\u03b1\u03c1\u03b1\u03bd\u03c4\u03af\u03bd\u03b1 \u03c3\u03c4\u03b7 \u039c\u03cd\u03ba\u03bf\u03bd\u03bf. \u00ab\u0395\u03af\u03bd\u03b1\u03b9 \u03b1\u03c5\u03c4\u03bf\u03bd\u03cc\u03b7\u03c4\u03bf \u03cc\u03c4\u03b9 \u03b7 \u03ba\u03bf\u03b9\u03bd\u03c9\u03bd\u03af\u03b1 \u03ba\u03b1\u03b9 \u03b7 \u03bf\u03b9\u03ba\u03bf\u03bd\u03bf\u03bc\u03af\u03b1 \u03b4\u03b5\u03bd \u03b1\u03bd\u03c4\u03ad\u03c7\u03bf\u03c5\u03bd \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03bf\u03c5\u03c2 \u03c0\u03b5\u03c1\u03b9\u03bf\u03c1\u03b9\u03c3\u03bc\u03bf\u03cd\u03c2\u00bb, \u03ad\u03bb\u03b5\u03b3\u03b5 \u03c7\u03b1\u03c1\u03b1\u03ba\u03c4\u03b7\u03c1\u03b9\u03c3\u03c4\u03b9\u03ba\u03ac \u03bf \u0393\u03b5\u03c1\u03b1\u03c0\u03b5\u03c4\u03c1\u03af\u03c4\u03b7\u03c2, \u03c4\u03b7\u03bd \u03ce\u03c1\u03b1 \u03c0\u03bf\u03c5 \u03b7 \u03ba\u03c5\u03b2\u03ad\u03c1\u03bd\u03b7\u03c3\u03b7 \u03b1\u03bd\u03b1\u03ba\u03bf\u03af\u03bd\u03c9\u03bd\u03b5\u2026 \u03b1\u03c5\u03c4\u03bf\u03cd\u03c2 \u03c4\u03bf\u03c5\u03c2 \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03bf\u03c5\u03c2 \u03c0\u03b5\u03c1\u03b9\u03bf\u03c1\u03b9\u03c3\u03bc\u03bf\u03cd\u03c2. \u03a9\u03c2 \u03b5\u03ba \u03c4\u03bf\u03cd\u03c4\u03c9\u03bd, \u03b4\u03cd\u03bf \u03c4\u03b9\u03bd\u03ac \u03bc\u03c0\u03bf\u03c1\u03b5\u03af \u03bd\u03b1 \u03c3\u03c5\u03bc\u03b2\u03b1\u03af\u03bd\u03bf\u03c5\u03bd: \u03b5\u03af\u03c4\u03b5 \u03bf \u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03cc\u03c2 \u0395\u03c0\u03b9\u03ba\u03c1\u03b1\u03c4\u03b5\u03af\u03b1\u03c2 \u03b4\u03b5\u03bd \u03bc\u03b5\u03c4\u03ad\u03c7\u03b5\u03b9 \u03c0\u03bb\u03ad\u03bf\u03bd \u03c3\u03c4\u03b7 \u03bb\u03ae\u03c8\u03b7 \u03c4\u03c9\u03bd \u03b1\u03c0\u03bf\u03c6\u03ac\u03c3\u03b5\u03c9\u03bd, \u03b5\u03af\u03c4\u03b5 \u03b7 \u03b1\u03c0\u03cc\u03c6\u03b1\u03c3\u03b7 \u03b3\u03b9\u03b1 \u03bf\u03c1\u03b9\u03b6\u03cc\u03bd\u03c4\u03b9\u03b1 \u03bc\u03ad\u03c4\u03c1\u03b1 \u03b5\u03bb\u03ae\u03c6\u03b8\u03b7 \u03c5\u03c0\u03cc \u03c4\u03bf \u03ba\u03c1\u03ac\u03c4\u03bf\u03c2 \u03c0\u03b1\u03bd\u03b9\u03ba\u03bf\u03cd \u03c4\u03bf \u03c0\u03c1\u03c9\u03af \u03c4\u03bf\u03c5 \u03a3\u03b1\u03b2\u03b2\u03ac\u03c4\u03bf\u03c5, \u03cc\u03c4\u03b1\u03bd \u03ad\u03c6\u03c4\u03b1\u03c3\u03b5 \u03c3\u03c4\u03bf \u039c\u03b1\u03be\u03af\u03bc\u03bf\u03c5 \u03b7 \u03c4\u03b5\u03bb\u03b5\u03c5\u03c4\u03b1\u03af\u03b1 \u00ab\u03c6\u03bf\u03c5\u03c1\u03bd\u03b9\u03ac\u00bb \u03c4\u03c9\u03bd \u03b5\u03c0\u03b9\u03b4\u03b7\u03bc\u03b9\u03bf\u03bb\u03bf\u03b3\u03b9\u03ba\u03ce\u03bd \u03b4\u03b5\u03b4\u03bf\u03bc\u03ad\u03bd\u03c9\u03bd \u03b3\u03b9\u03b1 \u03c4\u03bf \u03bd\u03b7\u03c3\u03af \u03c4\u03c9\u03bd \u03b1\u03bd\u03ad\u03bc\u03c9\u03bd\u2026", "example_title": "Topic NON_PAPER"}, {"text": "\u0395\u03af\u03bd\u03b1\u03b9 \u03be\u03b5\u03ba\u03ac\u03b8\u03b1\u03c1\u03bf \u03cc\u03c4\u03b9 \u03bc\u03b5\u03c4\u03ac \u03c4\u03bf \u03c0\u03bb\u03ae\u03b3\u03bc\u03b1 \u03c0\u03bf\u03c5 \u03b4\u03ad\u03c7\u03b8\u03b7\u03ba\u03b5 \u03b7 \u03ba\u03c5\u03b2\u03ad\u03c1\u03bd\u03b7\u03c3\u03ae \u03c4\u03bf\u03c5 \u03b1\u03c0\u03cc \u03c4\u03b9\u03c2 \u03b1\u03b4\u03c5\u03bd\u03b1\u03bc\u03af\u03b5\u03c2 \u03c3\u03c4\u03b7\u03bd \u03b1\u03bd\u03c4\u03b9\u03bc\u03b5\u03c4\u03ce\u03c0\u03b9\u03c3\u03b7 \u03c4\u03c9\u03bd \u03ba\u03b1\u03c4\u03b1\u03c3\u03c4\u03c1\u03bf\u03c6\u03b9\u03ba\u03ce\u03bd \u03c0\u03c5\u03c1\u03ba\u03b1\u03b3\u03b9\u03ce\u03bd \u03c4\u03bf \u03bc\u03b5\u03b3\u03ac\u03bb\u03bf \u03c3\u03c4\u03bf\u03af\u03c7\u03b7\u03bc\u03b1 \u03b3\u03b9\u03b1 \u03c4\u03bf\u03bd \u039a\u03c5\u03c1\u03b9\u03ac\u03ba\u03bf \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03ac\u03ba\u03b7 \u03b5\u03af\u03bd\u03b1\u03b9 \u03bd\u03b1 \u03c0\u03c1\u03bf\u03c7\u03c9\u03c1\u03ae\u03c3\u03b5\u03b9 \u03c3\u03c5\u03bd\u03c4\u03b5\u03c4\u03b1\u03b3\u03bc\u03ad\u03bd\u03b1 \u03ba\u03b1\u03b9 \u03c7\u03c9\u03c1\u03af\u03c2 \u03c0\u03b1\u03c1\u03b1\u03c4\u03c1\u03ac\u03b3\u03bf\u03c5\u03b4\u03b1 \u03bf \u03c3\u03c7\u03b5\u03b4\u03b9\u03b1\u03c3\u03bc\u03cc\u03c2 \u03b3\u03b9\u03b1 \u03c4\u03b7\u03bd \u03b1\u03c0\u03bf\u03ba\u03b1\u03c4\u03ac\u03c3\u03c4\u03b1\u03c3\u03b7 \u03c4\u03c9\u03bd \u03b6\u03b7\u03bc\u03b9\u03ce\u03bd. \u039f \u03a0\u03c1\u03c9\u03b8\u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03cc\u03c2 \u03ad\u03c7\u03b5\u03b9 \u03ae\u03b4\u03b7 \u03c6\u03c4\u03b9\u03ac\u03be\u03b5\u03b9 \u03bc\u03b9\u03b1 \u03bf\u03bc\u03ac\u03b4\u03b1 \u03ba\u03c1\u03bf\u03cd\u03c3\u03b7\u03c2 \u03c4\u03b7\u03bd \u03bf\u03c0\u03bf\u03af\u03b1 \u03b1\u03c0\u03bf\u03c4\u03b5\u03bb\u03bf\u03cd\u03bd 9 \u03c5\u03c0\u03bf\u03c5\u03c1\u03b3\u03bf\u03af. \u03a4\u03b1 \u03bc\u03ad\u03bb\u03b7 \u03c0\u03bf\u03c5 \u03b1\u03c0\u03b1\u03c1\u03c4\u03af\u03b6\u03bf\u03c5\u03bd \u03c4\u03b7\u03bd \u03bf\u03bc\u03ac\u03b4\u03b1 \u03ba\u03c1\u03bf\u03cd\u03c3\u03b7\u03c2 \u03ba\u03b1\u03b9 \u03c4\u03b1 \u03bf\u03c0\u03bf\u03af\u03b1 \u03b2\u03c1\u03af\u03c3\u03ba\u03bf\u03bd\u03c4\u03b1\u03b9 \u03c3\u03b5 \u03c3\u03c5\u03bd\u03b5\u03c7\u03ae, \u03ba\u03b1\u03b8\u03b7\u03bc\u03b5\u03c1\u03b9\u03bd\u03ae \u03b5\u03c0\u03b1\u03c6\u03ae \u03bc\u03b5 \u03c4\u03bf\u03bd \u039a\u03c5\u03c1\u03b9\u03ac\u03ba\u03bf \u039c\u03b7\u03c4\u03c3\u03bf\u03c4\u03ac\u03ba\u03b7 \u03b5\u03af\u03bd\u03b1\u03b9, \u03cc\u03c0\u03c9\u03c2 \u03bc\u03b1\u03c2 \u03c0\u03bb\u03b7\u03c1\u03bf\u03c6\u03bf\u03c1\u03b5\u03af \u03b7 \u03c3\u03c4\u03ae\u03bb\u03b7 \u00ab\u0398\u03b5\u03c9\u03c1\u03b5\u03af\u03bf\u00bb \u03c4\u03b7\u03c2 \u00ab\u039a\u03b1\u03b8\u03b7\u03bc\u03b5\u03c1\u03b9\u03bd\u03ae\u03c2\u00bb \u03b5\u03af\u03bd\u03b1\u03b9 \u03bf\u03b9: \u0393. \u0393\u03b5\u03c1\u03b1\u03c0\u03b5\u03c4\u03c1\u03af\u03c4\u03b7\u03c2, \u0391. \u03a3\u03ba\u03ad\u03c1\u03c4\u03c3\u03bf\u03c2, \u03a7\u03c1. \u03a4\u03c1\u03b9\u03b1\u03bd\u03c4\u03cc\u03c0\u03bf\u03c5\u03bb\u03bf\u03c2, \u039a. \u039a\u03b1\u03c1\u03b1\u03bc\u03b1\u03bd\u03bb\u03ae\u03c2, \u039a. \u03a3\u03ba\u03c1\u03ad\u03ba\u03b1\u03c2, \u03a3\u03c4. \u03a0\u03ad\u03c4\u03c3\u03b1\u03c2, \u03a3\u03c0. \u039b\u03b9\u03b2\u03b1\u03bd\u03cc\u03c2 \u03ba\u03b1\u03b9 \u03c6\u03c5\u03c3\u03b9\u03ba\u03ac \u03bf\u03b9 \u03a7\u03c1. \u03a3\u03c4\u03b1\u03b9\u03ba\u03bf\u03cd\u03c1\u03b1\u03c2 \u03ba\u03b1\u03b9 \u0398. \u03a3\u03ba\u03c5\u03bb\u03b1\u03ba\u03ac\u03ba\u03b7\u03c2.", "example_title": "Topic OPINION"}]}
|
text-classification
|
cvcio/mediawatch-el-topics
|
[
"transformers",
"pytorch",
"safetensors",
"roberta",
"text-classification",
"Greek",
"news",
"el",
"doi:10.57967/hf/0711",
"license:gpl-3.0",
"model-index",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"el"
] |
TAGS
#transformers #pytorch #safetensors #roberta #text-classification #Greek #news #el #doi-10.57967/hf/0711 #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
|
Disclaimer: *This model is still under testing and may change in the future, we will try to keep backwards compatibility. For any questions reach us at info@URL*
MediaWatch News Topics (Greek)
==============================
Fine-tuned model for multi-label text-classification (SequenceClassification), based on roberta-el-news, using Hugging Face's Transformers library. This model is to classify news in real-time on upto 33 topics including: *AFFAIRS*, *AGRICULTURE*, *ARTS\_AND\_CULTURE*, *BREAKING\_NEWS*, *BUSINESS*, *COVID*, *ECONOMY*, *EDUCATION*, *ELECTIONS*, *ENTERTAINMENT*, *ENVIRONMENT*, *FOOD*, *HEALTH*, *INTERNATIONAL*, *LAW\_AND\_ORDER*, *MILITARY*, *NON\_PAPER*, *OPINION*, *POLITICS*, *REFUGEE*, *REGIONAL*, *RELIGION*, *SCIENCE*, *SOCIAL\_MEDIA*, *SOCIETY*, *SPORTS*, *TECH*, *TOURISM*, *TRANSPORT*, *TRAVEL*, *WEATHER*, *CRIME*, *JUSTICE*.
How to use
----------
You can use this model directly with a pipeline for text-classification:
Labels
------
All labels, except *NON\_PAPER*, retrieved by source articles during the data collection step, without any preprocessing, assuming that journalists and newsrooms assign correct tags to the articles. We disregarded all articles with more than 6 tags to reduce bias and tag manipulation.
Pretraining
-----------
The model was pretrained using an NVIDIA A10 GPU for 15 epochs (~ approx 59K steps, 8 hours training) with a batch size of 128. The optimizer used is Adam with a learning rate of 1e-5, and weight decay 0.01. We used roc\_auc\_micro to evaluate the results.
### Framework versions
* Transformers 4.13.0
* Pytorch 1.9.0+cu111
* Datasets 1.16.1
* Tokenizers 0.10.3
Authors
-------
Dimitris Papaevagelou - @andefined
About Us
--------
Civic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
|
[
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
"TAGS\n#transformers #pytorch #safetensors #roberta #text-classification #Greek #news #el #doi-10.57967/hf/0711 #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
73,
80
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #roberta #text-classification #Greek #news #el #doi-10.57967/hf/0711 #license-gpl-3.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
-0.07905156910419464,
0.016066186130046844,
-0.0049019064754247665,
0.03880559280514717,
0.11747734248638153,
-0.025776969268918037,
0.20655949413776398,
0.023515531793236732,
0.09398451447486877,
-0.035853855311870575,
0.08108482509851456,
0.13080668449401855,
-0.01706824265420437,
0.13460232317447662,
-0.10768435150384903,
-0.2179214358329773,
0.04876385256648064,
0.059240736067295074,
0.038918256759643555,
0.10573556274175644,
0.11775710433721542,
-0.07309464365243912,
0.05233168601989746,
0.011716045439243317,
-0.029098110273480415,
-0.01694837585091591,
0.0437735989689827,
-0.14913766086101532,
0.13871905207633972,
0.01901128515601158,
0.06971719115972519,
0.030787194147706032,
-0.03959740325808525,
-0.020768949761986732,
0.02976277843117714,
-0.05001736432313919,
-0.07685453444719315,
0.03856084123253822,
0.0774054080247879,
-0.09264584630727768,
0.22410902380943298,
-0.04826197028160095,
-0.02507856860756874,
-0.026140892878174782,
-0.18929359316825867,
-0.17477203905582428,
-0.08207891881465912,
0.08641457557678223,
-0.0029855973552912474,
0.17250950634479523,
-0.0005893466295674443,
0.23581741750240326,
-0.1179843321442604,
0.05142107978463173,
0.09087765961885452,
-0.22526168823242188,
-0.04224886745214462,
-0.01954675279557705,
-0.011130494996905327,
0.07322602719068527,
0.051999080926179886,
0.10758133977651596,
-0.012330609373748302,
0.05367635563015938,
-0.010189964435994625,
-0.015050264075398445,
-0.020826661959290504,
-0.048403043299913406,
-0.10805384814739227,
-0.07228998839855194,
0.40614813566207886,
0.007922004908323288,
0.025292828679084778,
0.011750380508601665,
0.0372244156897068,
0.08725155144929886,
0.019502777606248856,
0.008639977313578129,
-0.07492609322071075,
-0.0020763948559761047,
-0.0047498526982963085,
-0.0020464530680328608,
-0.1405879557132721,
0.05894821882247925,
-0.2148965746164322,
0.30826687812805176,
0.004124253056943417,
0.08928117901086807,
-0.08919161558151245,
0.0682256817817688,
0.004526443313807249,
-0.07995238900184631,
0.03974155709147453,
-0.014473062008619308,
0.18287824094295502,
-0.01804495044052601,
0.03627626225352287,
-0.06715558469295502,
0.06998106092214584,
0.05266755819320679,
-0.12373145669698715,
-0.11770708858966827,
0.06013720855116844,
0.1580992341041565,
0.003073663916438818,
0.03428022563457489,
-0.02064763754606247,
0.11379121989011765,
0.007440127432346344,
-0.050111956894397736,
0.10481560975313187,
0.040511976927518845,
-0.07614447921514511,
0.04119361191987991,
0.008364083245396614,
0.06572917103767395,
0.08163964003324509,
0.10945655405521393,
-0.044597379863262177,
0.03301619738340378,
0.030676262453198433,
-0.004262801259756088,
-0.03644109517335892,
-0.09167786687612534,
-0.04038318246603012,
0.051797185093164444,
0.019621284678578377,
-0.0063141887076199055,
-0.10253237187862396,
0.18034130334854126,
-0.023524435237050056,
-0.0037325453013181686,
0.04880881309509277,
-0.008259469643235207,
0.08567515760660172,
-0.06980317085981369,
0.07870353013277054,
-0.21037639677524567,
-0.01590961590409279,
-0.018545446917414665,
0.033460211008787155,
-0.039072178304195404,
-0.06771659851074219,
0.04810429364442825,
-0.011538718827068806,
-0.07987897098064423,
-0.07828878611326218,
-0.10040206462144852,
-0.0833248645067215,
0.14627806842327118,
-0.025289874523878098,
-0.002541313413530588,
-0.15797701478004456,
0.01438634842634201,
-0.12299038469791412,
-0.025580275803804398,
0.01404849998652935,
-0.060020484030246735,
-0.06140865013003349,
0.04915881156921387,
-0.014268248341977596,
0.01477106474339962,
-0.005782968830317259,
0.02736896462738514,
0.018744798377156258,
0.2065264880657196,
0.1033283993601799,
-0.049460697919130325,
0.07388360053300858,
-0.10293258726596832,
-0.17130506038665771,
0.10728342831134796,
0.021540803834795952,
0.2002110630273819,
0.15043146908283234,
0.12083660066127777,
-0.04143926873803139,
-0.2440742701292038,
-0.124014712870121,
-0.03180209919810295,
-0.13156302273273468,
-0.16875553131103516,
0.0792856365442276,
-0.008437369018793106,
-0.048763908445835114,
0.06337876617908478,
-0.17701639235019684,
0.07775087654590607,
-0.1047544777393341,
-0.017046216875314713,
-0.006067186594009399,
-0.04492441192269325,
0.07228617370128632,
0.029269343242049217,
0.12957000732421875,
-0.10294214636087418,
-0.11609885096549988,
-0.007408870384097099,
-0.018585985526442528,
0.02860712632536888,
0.04171581193804741,
-0.1279069036245346,
0.2010825276374817,
0.0026780462358146906,
0.011963294818997383,
-0.14728105068206787,
-0.025612127035856247,
-0.038988590240478516,
0.017139218747615814,
0.13663256168365479,
0.03289603441953659,
0.06497713178396225,
-0.055197395384311676,
-0.07474981993436813,
0.026658669114112854,
-0.021680859848856926,
0.020950034260749817,
0.012049364857375622,
-0.10211198776960373,
0.1379554718732834,
-0.0030765950214117765,
0.06184045970439911,
-0.027830421924591064,
0.039787400513887405,
0.1227339580655098,
0.06321415305137634,
-0.01956787146627903,
0.07797710597515106,
-0.09574095904827118,
0.012861215509474277,
0.043793316930532455,
0.05039913207292557,
0.10464818775653839,
0.04198932275176048,
-0.04288307577371597,
0.05320959538221359,
-0.0715266615152359,
0.30254417657852173,
0.1649775356054306,
-0.15363357961177826,
-0.00864849891513586,
0.028866786509752274,
-0.06867839395999908,
0.008006897754967213,
-0.05050034075975418,
0.06152502819895744,
0.07091879099607468,
0.01131339929997921,
0.048810362815856934,
-0.04114878922700882,
-0.009644967503845692,
0.06370881199836731,
-0.0809469074010849,
-0.07761562615633011,
0.113695427775383,
0.08861785382032394,
-0.11550793796777725,
0.12867817282676697,
0.24427343904972076,
-0.025449300184845924,
0.16465333104133606,
-0.028325561434030533,
-0.006184936501085758,
-0.017112651839852333,
-0.038291990756988525,
0.030565768480300903,
0.1008598580956459,
-0.08202656358480453,
0.06379181891679764,
0.02702038176357746,
0.02969461865723133,
0.04719728231430054,
-0.14034394919872284,
-0.09229160845279694,
-0.031031835824251175,
0.0031919104512780905,
-0.041403379291296005,
0.10363079607486725,
-0.04811647906899452,
0.13456983864307404,
0.024305203929543495,
-0.12250987440347672,
0.08113261312246323,
0.025028610602021217,
-0.053999755531549454,
0.11433570832014084,
0.05179646983742714,
-0.2513314485549927,
-0.1288934051990509,
-0.04801210016012192,
-0.06301472336053848,
-0.005332339555025101,
0.10348119586706161,
0.0018156748265028,
-0.014680111780762672,
0.006388868670910597,
0.08230960369110107,
0.0011456909123808146,
-0.044404998421669006,
-0.02346412092447281,
0.07631855458021164,
-0.04890120401978493,
-0.04689262807369232,
-0.09067222476005554,
-0.14195440709590912,
-0.0469246543943882,
0.07898595184087753,
-0.11124217510223389,
0.13365332782268524,
0.01845884509384632,
-0.0207054540514946,
0.05476762726902962,
-0.02992958016693592,
0.1642698347568512,
-0.12289495766162872,
0.00785911176353693,
0.16223673522472382,
0.03571459278464317,
0.0487816259264946,
0.16499149799346924,
0.07327986508607864,
-0.015464264899492264,
-0.018215207383036613,
-0.06948459148406982,
-0.06237003207206726,
-0.23353180289268494,
-0.14150427281856537,
-0.06326974928379059,
0.08560456335544586,
-0.03722201660275459,
0.049858711659908295,
0.09736707806587219,
0.09369911253452301,
-0.0016173750627785921,
-0.04920295625925064,
0.043477632105350494,
0.058272477239370346,
0.32008683681488037,
0.045560624450445175,
0.05827372893691063,
-0.08194613456726074,
-0.090177521109581,
0.09432409703731537,
-0.03725990280508995,
0.1414913535118103,
0.09309114515781403,
0.010401099920272827,
0.0883086770772934,
0.04105585068464279,
0.054453618824481964,
0.07624442875385284,
0.1265261024236679,
-0.012536932714283466,
0.021291881799697876,
-0.04359199479222298,
0.012473474256694317,
-0.06126183271408081,
-0.061331845819950104,
-0.06539339572191238,
-0.0579284131526947,
-0.16420705616474152,
0.07231360673904419,
0.17198772728443146,
-0.04180096462368965,
-0.15339700877666473,
-0.11631231755018234,
-0.01774965599179268,
0.021945443004369736,
-0.05020976811647415,
-0.05482683330774307,
-0.08455388993024826,
-0.12936528027057648,
0.043442949652671814,
-0.06576204299926758,
0.08092467486858368,
-0.10213358700275421,
0.05048695206642151,
-0.03785102069377899,
-0.17559689283370972,
0.030457906424999237,
0.07661973685026169,
-0.11421297490596771,
0.40565258264541626,
0.003894887864589691,
0.09705089032649994,
-0.09226953238248825,
-0.05685539171099663,
-0.03080804832279682,
0.24695883691310883,
0.1527511030435562,
0.005106863100081682,
0.09046876430511475,
-0.06607294082641602,
-0.10175330936908722,
0.0376276895403862,
0.011601217091083527,
-0.10967247188091278,
-0.011045828461647034,
-0.021886296570301056,
0.004188053775578737,
-0.0045255753211677074,
-0.11314139515161514,
-0.06476585566997528,
-0.1146467924118042,
0.07927031069993973,
0.026423675939440727,
0.0890086367726326,
-0.005756240803748369,
-0.13909828662872314,
-0.1050180196762085,
0.07441014051437378,
-0.07633800059556961,
-0.11048093438148499,
-0.07214498519897461,
-0.06491955369710922,
-0.059838779270648956,
-0.10936790704727173,
0.01974494941532612,
-0.034727003425359726,
-0.13410361111164093,
-0.03592384606599808,
-0.09146445244550705,
0.08265800774097443,
-0.04895348846912384,
-0.07285014539957047,
0.011054795235395432,
0.12401725351810455,
0.04754355177283287,
-0.006722358986735344,
-0.047325197607278824,
0.061488308012485504,
-0.060731858015060425,
-0.08301005512475967,
0.007680921815335751,
-0.04606366157531738,
0.1697828471660614,
-0.017076054587960243,
-0.0022847021464258432,
-0.1729886680841446,
-0.06473414599895477,
-0.11108708381652832,
0.11489366739988327,
0.2472728043794632,
-0.03954218700528145,
-0.03150080889463425,
0.18619106709957123,
-0.035497523844242096,
-0.19222146272659302,
-0.11347322165966034,
-0.07648374885320663,
-0.018484655767679214,
0.04742422699928284,
-0.058283593505620956,
0.03614451363682747,
0.18936705589294434,
-0.12764230370521545,
-0.01026194915175438,
-0.182393878698349,
-0.06700186431407928,
0.13412779569625854,
-0.021112365648150444,
0.3150877058506012,
-0.09788988530635834,
-0.013313760980963707,
0.049299802631139755,
-0.1421067863702774,
0.194784015417099,
-0.0711418017745018,
0.048901788890361786,
-0.04514353349804878,
0.025388553738594055,
-0.014314047060906887,
-0.020132800564169884,
0.10155848413705826,
-0.08169084787368774,
0.009927231818437576,
-0.11294453591108322,
-0.218669131398201,
0.1525668501853943,
0.02312736213207245,
0.07664075493812561,
0.07905647158622742,
-0.014181428588926792,
-0.03973982855677605,
-0.05419604852795601,
-0.09464540332555771,
0.12931156158447266,
0.019267085939645767,
-0.051164042204618454,
-0.05409229174256325,
0.048249468207359314,
-0.09136934578418732,
-0.025251328945159912,
0.25070270895957947,
-0.06632523983716965,
0.14726491272449493,
0.09508731216192245,
0.16815604269504547,
-0.07712742686271667,
-0.029186977073550224,
-0.04886763542890549,
-0.07865157723426819,
0.029957298189401627,
0.007722865324467421,
-0.05791911482810974,
0.107343889772892,
-0.04054728150367737,
-0.016876760870218277,
0.045852094888687134,
-0.1120837852358818,
0.017131030559539795,
0.12331654131412506,
-0.19800019264221191,
-0.11464234441518784,
-0.06697020679712296,
0.050328608602285385,
0.03459818288683891,
0.21057164669036865,
0.10516494512557983,
-0.03607027605175972,
-0.018009625375270844,
-0.004584568552672863,
-0.03264394402503967,
-0.008898952975869179,
0.0018283737590536475,
0.036131780594587326,
-0.05163019523024559,
-0.06712224334478378,
0.05517550930380821,
0.14291729032993317,
-0.13375425338745117,
-0.10168039053678513,
0.08095203340053558,
-0.09714141488075256,
-0.07709605991840363,
-0.016606910154223442,
-0.017011696472764015,
-0.11627133935689926,
-0.09274189919233322,
-0.13169480860233307,
-0.12963362038135529,
0.04311990737915039,
0.18794186413288116,
0.09437926113605499,
0.018005549907684326,
-0.030760375782847404,
-0.017421698197722435,
-0.0790015235543251,
0.0357574000954628,
0.057184722274541855,
-0.06418772041797638,
-0.07586827874183655,
-0.06342267245054245,
0.06200478971004486,
0.09574098140001297,
-0.07904942333698273,
-0.05426356568932533,
-0.11647078394889832,
0.06089983135461807,
-0.00015918503049761057,
-0.030481796711683273,
-0.07619906216859818,
-0.02037152089178562,
-0.06675498187541962,
-0.09689316153526306,
-0.07947942614555359,
-0.009867089800536633,
-0.08119391649961472,
0.06663035601377487,
0.02238517440855503,
0.041794151067733765,
-0.08405487984418869,
-0.0015534409321844578,
0.06697762757539749,
0.007446419447660446,
0.11802634596824646,
0.07375215739011765,
-0.006978836841881275,
0.08092793822288513,
-0.13336312770843506,
0.037245094776153564,
0.05402139201760292,
-0.0427747517824173,
0.050406116992235184,
-0.025980481877923012,
-0.026337526738643646,
0.10011487454175949,
0.07630239427089691,
0.0948743000626564,
0.014126522466540337,
-0.1116320863366127,
0.12348209321498871,
0.09533653408288956,
-0.09367065876722336,
-0.06078478321433067,
0.0041611408814787865,
0.019288521260023117,
0.035025209188461304,
0.18027453124523163,
-0.07326167076826096,
0.015191925689578056,
-0.02430771477520466,
0.039028316736221313,
0.0009159499895758927,
-0.18432989716529846,
-0.08108390867710114,
-0.07052313536405563,
0.010983340442180634,
0.024525538086891174,
0.2038526087999344,
0.11190135031938553,
-0.0318143367767334,
0.059344783425331116,
0.11365635693073273,
-0.03965449333190918,
-0.04950934648513794,
0.025841284543275833,
-0.0341835618019104,
0.03185869753360748,
-0.041344307363033295,
0.0011842164676636457,
-0.0552426278591156,
-0.10125155001878738,
0.16131652891635895,
0.03482813760638237,
0.030197633430361748,
0.007759596221148968,
0.03625420853495598,
0.09041175991296768,
-0.058205123990774155,
-0.25125470757484436,
-0.02007182314991951,
-0.0028701634146273136,
-0.038413722068071365,
0.034852284938097,
0.14274650812149048,
-0.00892299972474575,
-0.008887450210750103,
-0.04873444139957428,
-0.012241778895258904,
-0.06304444372653961,
-0.20082087814807892,
-0.06638459861278534,
-0.08093824237585068,
0.010114840231835842,
-0.035679832100868225,
-0.020641272887587547,
0.1406691074371338,
-0.0034296601079404354,
-0.12716999650001526,
0.03229699283838272,
-0.01665939949452877,
-0.06792793422937393,
0.08303274214267731,
-0.038935381919145584,
0.040127936750650406,
-0.22674869000911713,
0.016170579940080643,
-0.1815677136182785,
0.050135087221860886,
0.0017431817250326276,
0.08342792093753815,
-0.10371152311563492,
-0.045436255633831024,
-0.1379595249891281,
-0.07426227629184723,
-0.01799686625599861,
0.0884542167186737,
0.031040219590067863,
0.09513135254383087,
-0.00196011527441442,
0.03625982999801636,
0.0985291451215744,
0.23088963329792023,
0.007066212594509125,
-0.12133582681417465,
-0.02953982539474964,
0.1274951547384262,
-0.022778771817684174,
0.04515240341424942,
0.02626165747642517,
-0.04615398123860359,
0.020020464435219765,
0.22301128506660461,
0.28949734568595886,
0.05030130222439766,
0.05545840412378311,
-0.026307767257094383,
0.03788668289780617,
0.06186242401599884,
0.0498608760535717,
0.008102988824248314,
0.26661738753318787,
-0.10350821167230606,
0.04527818411588669,
-0.020013712346553802,
0.0858309343457222,
-0.0981597974896431,
0.005000196397304535,
0.006694204639643431,
-0.08067399263381958,
-0.09468644112348557,
0.16954025626182556,
-0.11147204786539078,
0.0933748111128807,
-0.05141610652208328,
-0.1513429433107376,
-0.11686163395643234,
-0.005179329309612513,
0.024015169590711594,
-0.018160145729780197,
0.10851892828941345,
-0.03280145674943924,
-0.05001852661371231,
-0.04298075661063194,
0.07752435654401779,
-0.16285473108291626,
-0.043853290379047394,
0.13766008615493774,
0.06729619204998016,
0.16848713159561157,
0.011142611503601074,
0.11874738335609436,
0.05429711937904358,
-0.04777509719133377,
-0.06885569542646408,
0.13780082762241364,
0.03173977509140968,
-0.025108702480793,
0.03558271750807762,
-0.1447448432445526,
-0.0023040957748889923,
-0.034646037966012955,
0.057562634348869324,
-0.07454302161931992,
0.0972728356719017,
-0.05725298076868057,
-0.002785838209092617,
-0.042283229529857635,
0.06518951058387756,
-0.06917330622673035,
0.02568676881492138,
0.06168215349316597,
-0.0242376197129488,
-0.07041893899440765,
-0.027002492919564247,
0.04758426919579506,
0.1241825595498085,
-0.10297702252864838,
-0.018490131944417953,
-0.0838925763964653,
-0.04463762417435646,
0.03344569727778435,
0.06391146034002304,
-0.07080045342445374,
-0.0033302591182291508,
-0.0945274606347084,
0.00631204666569829,
-0.1274549812078476,
-0.0466446615755558,
0.05659324675798416,
-0.06101299077272415,
-0.0368962436914444,
-0.01062831748276949,
0.01462324894964695,
0.038160860538482666,
-0.1153833419084549,
-0.06712779402732849
] |
null | null |
transformers
|
# RoBERTa Greek base model
Pretrained model on Greek language with the Masked Language Modeling (MLM) objective using [Hugging Face's](https://huggingface.co/) [Transformers](https://github.com/huggingface/transformers) library. This model is *NOT* case-sensitive and all Greek diacritics retained.
### How to use
You can use this model directly with a pipeline for masked language modeling:
```python
# example url
# https://www.news247.gr/politiki/misologa-maximoy-gia-tin-ekthesi-tsiodra-lytra-gia-ti-thnitotita-ektos-meth.9462425.html
# not present in train/eval set
from transformers import pipeline
pipe = pipeline('fill-mask', model='cvcio/roberta-el-news')
pipe(
'Η κυβέρνηση μουδιασμένη από τη <mask> της έκθεσης Τσιόδρα-Λύτρα, '
'επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.'
)
# outputs
[
{
'sequence': 'Η κυβέρνηση μουδιασμένη από τη δημοσιοποίηση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.',
'score': 0.5881184339523315, 'token': 20235, 'token_str': ' δημοσιοποίηση'
},
{
'sequence': 'Η κυβέρνηση μουδιασμένη από τη δημοσίευση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.',
'score': 0.05952141433954239, 'token': 9696, 'token_str': ' δημοσίευση'
},
{
'sequence': 'Η κυβέρνηση μουδιασμένη από τη διαχείριση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.',
'score': 0.029887061566114426, 'token': 4315, 'token_str': ' διαχείριση'
},
{
'sequence': 'Η κυβέρνηση μουδιασμένη από τη διαρροή της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.',
'score': 0.022848669439554214, 'token': 24940, 'token_str': ' διαρροή'
},
{
'sequence': 'Η κυβέρνηση μουδιασμένη από τη ματαίωση της έκθεσης Τσιόδρα-Λύτρα, επιχειρεί χωρίς να απαντά ουσιαστικά να ρίξει ευθύνες στον ΣΥΡΙΖΑ, που κυβερνούσε πριν... 2 χρόνια.',
'score': 0.01729060709476471, 'token': 46913, 'token_str': ' ματαίωση'
}
]
```
## Training data
The model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with [MediaWatch](https://mediawatch.io/), from October 2016 upto December 2021.
## Preprocessing
The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. `&` => `&`).
## Pretraining
The model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.
### Training results
| epochs | steps | train/train_loss | train/loss | eval/loss |
|-------:|--------:|-----------------:|------------:|----------:|
| 3 | 765,414 | 0.3960 | 1.2356 | 0.9028 |
### Evaluation results
The model fine-tuned on ner task using the [elNER](https://github.com/nmpartzio/elner) dataset and achieved the following results:
| task | epochs | lr | batch | dataset | precision | recall | f1 | accuracy |
|-----:|-------:|-----:|------:|--------:|----------:|-------:|-------:|---------:|
| ner | 5 | 1e-5 | 16/16 | elNER4 | 0.8954 | 0.9280 | 0.9114 | 0.9872 |
| ner | 5 | 1e-4 | 16/16 | elNER18 | 0.9069 | 0.9268 | 0.9168 | 0.9823 |
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-5
- train_batch_size: 14
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 2
- total_train_batch_size: 28
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3.0
### Framework versions
- Transformers 4.13.0
- Pytorch 1.9.0+cu111
- Datasets 1.16.1
- Tokenizers 0.10.3
## Authors
Dimitris Papaevagelou - [@andefined](https://github.com/andefined)
## About Us
[Civic Information Office](https://cvcio.org/) is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
|
{"language": "el", "license": "gpl-3.0", "tags": ["generated_from_trainer", "roberta", "Greek", "news", "transformers"], "widget": [{"text": "\u0397 \u03ba\u03c5\u03b2\u03ad\u03c1\u03bd\u03b7\u03c3\u03b7 \u03bc\u03bf\u03c5\u03b4\u03b9\u03b1\u03c3\u03bc\u03ad\u03bd\u03b7 \u03b1\u03c0\u03cc \u03c4\u03b7 <mask> \u03c4\u03b7\u03c2 \u03ad\u03ba\u03b8\u03b5\u03c3\u03b7\u03c2 \u03a4\u03c3\u03b9\u03cc\u03b4\u03c1\u03b1-\u039b\u03cd\u03c4\u03c1\u03b1, \u03b5\u03c0\u03b9\u03c7\u03b5\u03b9\u03c1\u03b5\u03af \u03c7\u03c9\u03c1\u03af\u03c2 \u03bd\u03b1 \u03b1\u03c0\u03b1\u03bd\u03c4\u03ac \u03bf\u03c5\u03c3\u03b9\u03b1\u03c3\u03c4\u03b9\u03ba\u03ac \u03bd\u03b1 \u03c1\u03af\u03be\u03b5\u03b9 \u03b5\u03c5\u03b8\u03cd\u03bd\u03b5\u03c2 \u03c3\u03c4\u03bf\u03bd \u03a3\u03a5\u03a1\u0399\u0396\u0391, \u03c0\u03bf\u03c5 \u03ba\u03c5\u03b2\u03b5\u03c1\u03bd\u03bf\u03cd\u03c3\u03b5 \u03c0\u03c1\u03b9\u03bd... 2 \u03c7\u03c1\u03cc\u03bd\u03b9\u03b1."}], "model-index": [{"name": "roberta-el-news", "results": []}]}
|
fill-mask
|
cvcio/roberta-el-news
|
[
"transformers",
"pytorch",
"safetensors",
"roberta",
"fill-mask",
"generated_from_trainer",
"Greek",
"news",
"el",
"doi:10.57967/hf/0712",
"license:gpl-3.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"el"
] |
TAGS
#transformers #pytorch #safetensors #roberta #fill-mask #generated_from_trainer #Greek #news #el #doi-10.57967/hf/0712 #license-gpl-3.0 #autotrain_compatible #endpoints_compatible #region-us
|
RoBERTa Greek base model
========================
Pretrained model on Greek language with the Masked Language Modeling (MLM) objective using Hugging Face's Transformers library. This model is *NOT* case-sensitive and all Greek diacritics retained.
### How to use
You can use this model directly with a pipeline for masked language modeling:
Training data
-------------
The model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.
Preprocessing
-------------
The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').
Pretraining
-----------
The model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.
### Training results
### Evaluation results
The model fine-tuned on ner task using the elNER dataset and achieved the following results:
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 5e-5
* train\_batch\_size: 14
* eval\_batch\_size: 8
* seed: 42
* gradient\_accumulation\_steps: 2
* total\_train\_batch\_size: 28
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 3.0
### Framework versions
* Transformers 4.13.0
* Pytorch 1.9.0+cu111
* Datasets 1.16.1
* Tokenizers 0.10.3
Authors
-------
Dimitris Papaevagelou - @andefined
About Us
--------
Civic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
|
[
"### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nTraining data\n-------------\n\n\nThe model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.\n\n\nPreprocessing\n-------------\n\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').\n\n\nPretraining\n-----------\n\n\nThe model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.",
"### Training results",
"### Evaluation results\n\n\nThe model fine-tuned on ner task using the elNER dataset and achieved the following results:",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-5\n* train\\_batch\\_size: 14\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 28\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
"TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #generated_from_trainer #Greek #news #el #doi-10.57967/hf/0712 #license-gpl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nTraining data\n-------------\n\n\nThe model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.\n\n\nPreprocessing\n-------------\n\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').\n\n\nPretraining\n-----------\n\n\nThe model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.",
"### Training results",
"### Evaluation results\n\n\nThe model fine-tuned on ner task using the elNER dataset and achieved the following results:",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-5\n* train\\_batch\\_size: 14\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 28\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0",
"### Framework versions\n\n\n* Transformers 4.13.0\n* Pytorch 1.9.0+cu111\n* Datasets 1.16.1\n* Tokenizers 0.10.3\n\n\nAuthors\n-------\n\n\nDimitris Papaevagelou - @andefined\n\n\nAbout Us\n--------\n\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
76,
226,
4,
27,
126,
80
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #generated_from_trainer #Greek #news #el #doi-10.57967/hf/0712 #license-gpl-3.0 #autotrain_compatible #endpoints_compatible #region-us \n### How to use\n\n\nYou can use this model directly with a pipeline for masked language modeling:\n\n\nTraining data\n-------------\n\n\nThe model was pretrained on 8 millon unique news articles (~ approx 160M sentences, 33GB of text), collected with MediaWatch, from October 2016 upto December 2021.\n\n\nPreprocessing\n-------------\n\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50,265. During the preprocessing we only unescaped html text to the correspoing Unicode characters (ex. '&' => '&').\n\n\nPretraining\n-----------\n\n\nThe model was pretrained using an NVIDIA A10 GPU for 3 epochs (~ approx 760K steps, 182 hours) with a batch size of 14 (x2 gradient accumulation steps = 28) and a sequence length of 512 tokens. The optimizer used is Adam with a learning rate of 5e-5, and linear decay of the learning rate.### Training results### Evaluation results\n\n\nThe model fine-tuned on ner task using the elNER dataset and achieved the following results:### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-5\n* train\\_batch\\_size: 14\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 28\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 3.0"
] |
[
-0.07185406982898712,
0.11020053178071976,
-0.0020101454574614763,
0.07151404768228531,
0.09931200742721558,
0.03591727465391159,
0.08194401860237122,
0.094209223985672,
-0.05608896166086197,
0.1316375434398651,
0.12507982552051544,
-0.02822117507457733,
0.10092224925756454,
0.12001820653676987,
0.03529959172010422,
-0.2515507936477661,
0.052757762372493744,
-0.04426714777946472,
0.009512585587799549,
0.12931206822395325,
0.08196597546339035,
-0.12235694378614426,
0.07491415739059448,
0.01327538676559925,
-0.07076792418956757,
-0.05833911895751953,
-0.03820396959781647,
-0.03132692351937294,
0.09277646988630295,
0.04503678157925606,
0.05884741619229317,
0.0020982606802135706,
0.09628027677536011,
-0.22324785590171814,
0.011861863546073437,
0.09503134340047836,
-0.010271716862916946,
0.053635887801647186,
0.08407444506883621,
0.03167155757546425,
0.15461379289627075,
-0.10139526426792145,
0.06784902513027191,
0.015596142038702965,
-0.1412007063627243,
-0.1642000675201416,
-0.09539168328046799,
-0.025481585413217545,
0.14440006017684937,
0.1097228154540062,
-0.050172947347164154,
0.07752158492803574,
-0.06950666755437851,
0.03488064929842949,
0.09671923518180847,
-0.30072346329689026,
-0.06944169849157333,
-0.039398279041051865,
0.01907891407608986,
0.048086389899253845,
-0.10234393924474716,
-0.0004109617730136961,
0.023874029517173767,
0.012304898351430893,
0.11129923164844513,
0.022033169865608215,
0.16414964199066162,
-0.02672925963997841,
-0.15574049949645996,
-0.11093519628047943,
0.04375205561518669,
0.05959800258278847,
-0.02653542347252369,
-0.15382297337055206,
-0.0668570026755333,
-0.19411307573318481,
-0.022222958505153656,
0.007135681342333555,
0.01255261804908514,
-0.029796579852700233,
-0.014489947818219662,
-0.046348124742507935,
-0.09493106603622437,
-0.07088738679885864,
0.005201473366469145,
0.2106115221977234,
0.07870911806821823,
0.018948256969451904,
0.02405151166021824,
0.12197966128587723,
-0.02305910363793373,
-0.12074458599090576,
-0.04494404047727585,
0.012937341816723347,
-0.13571526110172272,
0.013535771518945694,
-0.036050695925951004,
-0.0577268972992897,
-0.04473615810275078,
0.18846024572849274,
-0.03200173005461693,
0.06144675612449646,
0.12016166001558304,
-0.014138953760266304,
-0.02456015720963478,
0.149918332695961,
-0.08196529000997543,
-0.08572033047676086,
-0.01733807474374771,
0.132459819316864,
0.048143960535526276,
-0.05116267874836922,
-0.03264826536178589,
0.043841518461704254,
0.06975394487380981,
0.011679037474095821,
0.01255602017045021,
0.01977311074733734,
-0.05533013120293617,
-0.038596030324697495,
0.12867365777492523,
-0.10663089901208878,
0.02737193927168846,
0.02880108915269375,
-0.058407012373209,
-0.056902751326560974,
-0.01650478132069111,
-0.01833200454711914,
-0.0589773990213871,
0.1166563555598259,
-0.07182776927947998,
-0.04328848421573639,
-0.1309424489736557,
-0.05936560034751892,
-0.013762394897639751,
-0.051752299070358276,
-0.041275180876255035,
-0.07643824815750122,
-0.12946456670761108,
-0.0602986104786396,
0.0659528598189354,
-0.03755760192871094,
-0.06623371690511703,
-0.05247502401471138,
-0.03188144043087959,
0.04886458069086075,
-0.02712554857134819,
0.11720864474773407,
-0.04686165973544121,
0.07971690595149994,
0.03266407176852226,
0.031956952065229416,
0.05212508514523506,
0.030347708612680435,
-0.08331653475761414,
0.057328712195158005,
-0.02266489528119564,
0.09355362504720688,
0.0019262864952906966,
-0.0032297559082508087,
-0.07618514448404312,
-0.09286241978406906,
-0.09859472513198853,
-0.013445399701595306,
0.09084899723529816,
0.07298643887042999,
-0.13307073712348938,
-0.035141605883836746,
0.13234049081802368,
-0.09735114127397537,
-0.07283042371273041,
0.16803759336471558,
-0.04533328488469124,
-0.08233422040939331,
0.05385136976838112,
0.08622001856565475,
0.08196540176868439,
-0.09472561627626419,
-0.05510856956243515,
-0.07693305611610413,
-0.005619229283183813,
0.06793057918548584,
0.1037115752696991,
-0.006399104371666908,
0.029749583452939987,
0.0029883799143135548,
-0.09823591262102127,
0.021184438839554787,
-0.02934424579143524,
-0.08680412918329239,
-0.003142555942758918,
-0.04496948421001434,
0.014539177529513836,
0.06526894867420197,
-0.013183975592255592,
-0.07654117047786713,
-0.16912102699279785,
-0.10717600584030151,
0.0962219387292862,
-0.09612838178873062,
0.04031955823302269,
-0.08471278101205826,
0.06717902421951294,
-0.041435904800891876,
0.010772467590868473,
-0.13629208505153656,
-0.06657683849334717,
0.04721017926931381,
-0.11646462231874466,
0.019068481400609016,
-0.05355122685432434,
0.040732432156801224,
0.0753236934542656,
-0.04633935168385506,
-0.06475894898176193,
-0.09743712097406387,
-0.04126061871647835,
-0.06609248369932175,
-0.08483579754829407,
-0.049631938338279724,
-0.0271261278539896,
0.156561017036438,
-0.1096983551979065,
-0.011253024451434612,
0.02315649390220642,
0.08140744268894196,
0.000609612266998738,
-0.07652103900909424,
-0.023821450769901276,
0.0180530846118927,
0.0006471815868280828,
-0.05998864024877548,
0.029118487611413002,
0.008508373983204365,
-0.02961329184472561,
-0.007702960167080164,
-0.202151358127594,
-0.04679238796234131,
0.11255981028079987,
0.06739436089992523,
-0.05387398228049278,
-0.0016396676655858755,
-0.029719369485974312,
-0.04506360739469528,
-0.04032915458083153,
0.017191054299473763,
0.27443933486938477,
0.031346987932920456,
0.10944178700447083,
-0.08952943235635757,
-0.012640354223549366,
0.010811815038323402,
0.029385752975940704,
-0.04524534195661545,
0.0773715004324913,
0.06761052459478378,
-0.12247780710458755,
0.057287558913230896,
0.04664841666817665,
0.005071609281003475,
0.11644437909126282,
0.013497387990355492,
-0.11169003695249557,
-0.04063928499817848,
0.008566707372665405,
0.005566836334764957,
0.12797926366329193,
-0.03999416157603264,
0.0066584753803908825,
0.024178562685847282,
0.00605736905708909,
0.02599770948290825,
-0.17498908936977386,
0.03988979011774063,
0.04808010905981064,
-0.03842262923717499,
0.0519871823489666,
-0.04207285866141319,
-0.057573989033699036,
0.05236804857850075,
0.0950121283531189,
-0.018040457740426064,
0.0023588489275425673,
-0.05434110760688782,
-0.08546552062034607,
0.18474993109703064,
-0.061822883784770966,
-0.21794798970222473,
-0.11180149763822556,
0.0395452044904232,
-0.053364239633083344,
0.008728304877877235,
0.0022266958840191364,
-0.03169891983270645,
-0.087907575070858,
-0.0970178171992302,
0.03632175177335739,
0.004741096403449774,
-0.0012732025934383273,
-0.01164042018353939,
-0.03791671246290207,
0.009114168584346771,
-0.06810378283262253,
0.035368483513593674,
-0.044340234249830246,
-0.046273373067379,
0.01155314315110445,
0.07294822484254837,
0.13136333227157593,
0.12861214578151703,
0.01863664574921131,
0.016033664345741272,
-0.0015709275612607598,
0.20627471804618835,
-0.1566697359085083,
0.004104440566152334,
0.10678649693727493,
0.0004288779746275395,
0.04893965646624565,
0.11355797946453094,
0.016615407541394234,
-0.08407532423734665,
0.035623084753751755,
0.04617719352245331,
-0.0561102032661438,
-0.13496607542037964,
-0.059429462999105453,
-0.050410594791173935,
-0.07429693639278412,
0.16089887917041779,
0.04086637124419212,
-0.038491081446409225,
0.027763020247220993,
-0.06599388271570206,
-0.04462018236517906,
0.04957875236868858,
0.08230289071798325,
0.0543704517185688,
0.057318028062582016,
0.07495813816785812,
-0.030311236158013344,
-0.004535183776170015,
0.07087840139865875,
-0.011272366158664227,
0.2127874344587326,
-0.04961210489273071,
0.23132984340190887,
0.0422450490295887,
0.10803171247243881,
0.021112889051437378,
0.014942602254450321,
-0.01637766696512699,
0.00719436164945364,
0.0030950908549129963,
-0.019939148798584938,
-0.07945790141820908,
0.07111254334449768,
0.062002357095479965,
0.0206292811781168,
-0.06740671396255493,
0.022013336420059204,
0.05455024540424347,
0.28054219484329224,
0.07995034754276276,
-0.27806568145751953,
-0.05922145023941994,
0.010180174373090267,
-0.06256093084812164,
-0.07615336775779724,
0.04506111890077591,
0.09154027700424194,
-0.08829803764820099,
0.05521168187260628,
-0.09438443928956985,
0.06251194328069687,
-0.1134408712387085,
-0.017006507143378258,
0.13589709997177124,
0.13072898983955383,
0.00011486838775454089,
0.06885562092065811,
-0.23723819851875305,
0.17503713071346283,
-0.0058779241517186165,
0.116221122443676,
-0.03710933029651642,
0.007874177768826485,
-0.002497006906196475,
-0.012902414426207542,
0.10582059621810913,
-0.027827871963381767,
-0.00841281097382307,
-0.03151471167802811,
-0.14578858017921448,
0.026819318532943726,
0.11042775958776474,
-0.06589142978191376,
0.12510435283184052,
-0.023144938051700592,
-0.005120978225022554,
-0.010875081643462181,
-0.02734730765223503,
-0.03302321583032608,
-0.16355153918266296,
0.05511185899376869,
-0.07265496999025345,
0.01947750337421894,
-0.07540475577116013,
-0.06735359877347946,
-0.12881304323673248,
0.12852007150650024,
-0.07091439515352249,
-0.061805419623851776,
-0.11325322091579437,
0.03202259913086891,
0.08847687393426895,
-0.07462440431118011,
0.0601089671254158,
0.003155659418553114,
0.11886188387870789,
-0.02533089555799961,
-0.03274589776992798,
0.05769838020205498,
-0.013179282657802105,
-0.2088620364665985,
-0.0060546682216227055,
0.13127769529819489,
0.07089327275753021,
0.04799012839794159,
-0.026848623529076576,
0.0830584466457367,
-0.05059683322906494,
-0.09636177867650986,
0.026689067482948303,
-0.005584605038166046,
0.036915089935064316,
-0.006521482486277819,
0.08801400661468506,
0.00362120708450675,
-0.09257543832063675,
0.014782004058361053,
0.07829855382442474,
0.2352587729692459,
-0.07278896868228912,
0.05247014760971069,
0.11733114719390869,
-0.02284904569387436,
-0.20959973335266113,
-0.04864414036273956,
0.06195041537284851,
0.0628746747970581,
-0.05600442364811897,
-0.16103416681289673,
-0.037113506346940994,
0.06303604692220688,
-0.033449672162532806,
0.0680554136633873,
-0.32333579659461975,
-0.1378476619720459,
0.04512961953878403,
0.10613708198070526,
-0.02218526042997837,
-0.09761545062065125,
-0.032237302511930466,
0.002894036704674363,
-0.019831843674182892,
0.12437979876995087,
0.043804533779621124,
0.10352488607168198,
-0.004831692669540644,
-0.030692532658576965,
0.021999893710017204,
-0.07414553314447403,
0.17011956870555878,
-0.03736650198698044,
0.07712319493293762,
-0.0398935042321682,
0.06979871541261673,
-0.0017706938087940216,
-0.07099968940019608,
0.05675681680440903,
-0.047075286507606506,
0.01777949556708336,
-0.08542147278785706,
-0.06554347276687622,
-0.08292677253484726,
-0.010893544182181358,
-0.03974643722176552,
-0.04012518376111984,
-0.06877780705690384,
0.07618860900402069,
0.07335074245929718,
0.016311118379235268,
0.09559816122055054,
-0.0034471589606255293,
-0.05484217777848244,
0.08221183717250824,
0.10545933246612549,
0.009208356030285358,
-0.11820897459983826,
-0.03278358653187752,
0.02292104810476303,
0.06932924687862396,
-0.13164030015468597,
0.0020080709364265203,
0.10572721809148788,
0.016455451026558876,
0.12429090589284897,
0.004073291085660458,
-0.13831011950969696,
0.011572230607271194,
0.06334107369184494,
-0.14335380494594574,
-0.19313569366931915,
-0.010593285784125328,
-0.07028192281723022,
-0.11160703003406525,
-0.06711044907569885,
0.06471017003059387,
-0.05115637555718422,
-0.00023821981449145824,
0.0015217306790873408,
0.06182873249053955,
-0.0041116951033473015,
0.12039858847856522,
0.02416480891406536,
0.04302436485886574,
-0.10719870030879974,
0.12950082123279572,
0.09132536500692368,
-0.14904208481311798,
0.005179012659937143,
0.13954216241836548,
-0.08054418116807938,
-0.05602901056408882,
-0.06866953521966934,
0.03592974692583084,
0.0471491701900959,
-0.05981709808111191,
-0.0730459913611412,
-0.07262987643480301,
0.06225256994366646,
-0.022792832925915718,
0.05104200169444084,
0.01945771835744381,
-0.00900298822671175,
0.019178356975317,
-0.13554875552654266,
0.1141137182712555,
0.0098203020170331,
0.026542698964476585,
-0.06437524408102036,
0.08795344829559326,
0.04418971762061119,
0.010493339039385319,
-0.015240214765071869,
-0.006913907825946808,
-0.02565569244325161,
0.01572325825691223,
-0.00911996141076088,
-0.020428434014320374,
-0.05911092832684517,
0.020671894773840904,
-0.018609002232551575,
-0.01785651408135891,
-0.06191953271627426,
0.016813868656754494,
-0.06329427659511566,
-0.08189120888710022,
-0.012657606042921543,
-0.016648756340146065,
-0.08220615237951279,
-0.021921096369624138,
0.037334274500608444,
-0.09351279586553574,
0.10599227249622345,
-0.0092931455001235,
0.036491334438323975,
0.027427135035395622,
-0.0020149715710431337,
0.0450349897146225,
-0.010839872062206268,
0.02439805306494236,
-0.01656108908355236,
-0.11191315948963165,
0.02171798050403595,
-0.014594229869544506,
-0.056297991424798965,
-0.009111369960010052,
0.04781639575958252,
-0.12610794603824615,
0.06495877355337143,
-0.012408577837049961,
-0.014421703293919563,
-0.04386652261018753,
0.04201646149158478,
-0.012321169488132,
0.03638618066906929,
0.1691243201494217,
-0.05192379280924797,
0.06096911057829857,
-0.16536101698875427,
-0.014872882515192032,
0.02120707742869854,
-0.02551104500889778,
-0.076688751578331,
-0.03280771151185036,
0.08394236117601395,
-0.03126969933509827,
0.11082910746335983,
0.019033627584576607,
-0.0032831435091793537,
0.04644467309117317,
0.00909892376512289,
-0.025222910568118095,
0.05253493785858154,
0.08927629142999649,
0.022039640694856644,
-0.021908387541770935,
0.0785214975476265,
0.019688589498400688,
0.023374607786536217,
0.0719025582075119,
0.16097977757453918,
0.16848769783973694,
0.16991586983203888,
0.09371142834424973,
-0.01783278025686741,
-0.11835605651140213,
-0.13233010470867157,
0.1455545872449875,
-0.10853997617959976,
0.07041566073894501,
0.0058260271325707436,
0.08961547911167145,
0.10356831550598145,
-0.18717069923877716,
0.054933223873376846,
0.006534218788146973,
-0.08190929889678955,
-0.05040914937853813,
-0.12769974768161774,
-0.0627150684595108,
-0.06976951658725739,
0.027344688773155212,
-0.08574602007865906,
0.024551142007112503,
0.10028550773859024,
0.0111938351765275,
-0.04643639177083969,
0.15799428522586823,
-0.010947237722575665,
-0.016681576147675514,
0.08049069344997406,
0.04294682294130325,
0.03497375175356865,
0.04579649493098259,
-0.05680013447999954,
0.009166440926492214,
-0.04929622262716293,
0.10461067408323288,
-0.020668303593993187,
-0.014975650236010551,
0.08785098791122437,
0.05988873913884163,
-0.07958970963954926,
0.026606380939483643,
0.008396962657570839,
0.08661576360464096,
0.14947055280208588,
0.04090038686990738,
0.002188713289797306,
-0.03311668336391449,
0.14396515488624573,
-0.02390092797577381,
-0.007086784113198519,
-0.13713234663009644,
0.13126240670681,
0.01802097260951996,
-0.00045952745131216943,
0.03339806944131851,
-0.10666674375534058,
0.008592077530920506,
0.17924313247203827,
0.09499464929103851,
-0.09825854003429413,
-0.02391960285604,
-0.0006441191653721035,
-0.003384453011676669,
-0.0023597220424562693,
0.09916000068187714,
0.029252037405967712,
0.15118245780467987,
-0.06597130745649338,
-0.02311490848660469,
-0.052561551332473755,
-0.040618058294057846,
-0.029713500291109085,
0.0919921025633812,
-0.0001755131670506671,
0.032362766563892365,
-0.10686829686164856,
0.03251758962869644,
0.022777942940592766,
-0.11082077771425247,
0.09259284287691116,
-0.1309712529182434,
-0.14821919798851013,
0.03192056342959404,
0.047064848244190216,
-0.013387485407292843,
0.07577630877494812,
0.03366602957248688,
0.009281572885811329,
0.077886663377285,
0.018017303198575974,
-0.06747723370790482,
-0.07112441211938858,
0.07172827422618866,
-0.03465615212917328,
0.27842777967453003,
-0.014862722717225552,
0.04433949664235115,
0.09323861449956894,
0.012378190644085407,
-0.13580498099327087,
0.09562138468027115,
0.04148133471608162,
-0.06887788325548172,
0.048702891916036606,
0.151450514793396,
-0.002663266146555543,
0.0560603030025959,
0.07632419466972351,
-0.016926869750022888,
0.033881381154060364,
-0.09983348846435547,
-0.0305529423058033,
-0.07771182060241699,
0.049290671944618225,
-0.05329463630914688,
0.16638629138469696,
0.18977369368076324,
-0.020364413037896156,
-0.011340725235641003,
-0.01777900941669941,
0.03907085210084915,
-0.025303157046437263,
0.20080798864364624,
-0.00263361819088459,
-0.1861673891544342,
0.01832820102572441,
-0.05068984627723694,
-0.007943380624055862,
-0.220947727560997,
-0.0748271718621254,
0.020516974851489067,
-0.06341355293989182,
-0.019357847049832344,
0.11116600036621094,
0.017225580289959908,
0.013630336150527,
-0.06398793309926987,
-0.015186019241809845,
-0.028850525617599487,
0.09815862774848938,
-0.14240211248397827,
-0.09659236669540405
] |
null | null |
transformers
|
# Greek RoBERTa Uncased (v1)
Pretrained model on Greek language using a masked language modeling (MLM) objective using [Hugging Face's](https://huggingface.co/) [Transformers](https://github.com/huggingface/transformers) library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).
### Training data
This model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.
### Preprocessing
The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.
```python
import unicodedata
from transformers import pipeline
def normalize_tweet(tweet, do_lower = True, do_strip_accents = True, do_split_word_numbers = False, user_fill = '', url_fill = ''):
# your tweet pre-processing logic goes here
# example...
# remove extra spaces, escape HTML, replace non-standard punctuation
# replace any @user with blank
# replace any link with blank
# explode hashtags to strings (ex. #EU2019 ==> EU 2019)
# remove all emojis
# if do_split_word_numbers:
# splited strings containing any numbers
# standardize punctuation
# remove unicode symbols
if do_lower:
tweet = tweet.lower()
if do_strip_accents:
tweet = strip_accents(tweet)
return tweet.strip()
def strip_accents(s):
return ''.join(c for c in unicodedata.normalize('NFD', s)
if unicodedata.category(c) != 'Mn')
nlp = pipeline('fill-mask', model = 'cvcio/roberta-el-uncased-twitter-v1')
print(
nlp(
normalize_tweet(
'<mask>: Μεγάλη υποχώρηση του ιικού φορτίου σε Αττική και Θεσσαλονίκη'
)
)
)
```
### Pretraining
The model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.
### Authors
Dimitris Papaevagelou - [@andefined](https://github.com/andefined)
### About Us
[Civic Information Office](https://cvcio.org/) is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
|
{"language": "el", "tags": ["roberta", "twitter", "Greek"], "widget": [{"text": "<mask>: \u03bc\u03b5\u03b3\u03b1\u03bb\u03b7 \u03c5\u03c0\u03bf\u03c7\u03c9\u03c1\u03b7\u03c3\u03b7 \u03c4\u03bf\u03c5 \u03b9\u03b9\u03ba\u03bf\u03c5 \u03c6\u03bf\u03c1\u03c4\u03b9\u03bf\u03c5 \u03c3\u03b5 \u03b1\u03c4\u03c4\u03b9\u03ba\u03b7 \u03ba\u03b1\u03b9 \u03b8\u03b5\u03c3\u03c3\u03b1\u03bb\u03bf\u03bd\u03b9\u03ba\u03b7"}]}
|
fill-mask
|
cvcio/roberta-el-uncased-twitter-v1
|
[
"transformers",
"pytorch",
"safetensors",
"roberta",
"fill-mask",
"twitter",
"Greek",
"el",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"el"
] |
TAGS
#transformers #pytorch #safetensors #roberta #fill-mask #twitter #Greek #el #autotrain_compatible #endpoints_compatible #region-us
|
# Greek RoBERTa Uncased (v1)
Pretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).
### Training data
This model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.
### Preprocessing
The texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.
### Pretraining
The model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.
### Authors
Dimitris Papaevagelou - @andefined
### About Us
Civic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest.
|
[
"# Greek RoBERTa Uncased (v1)\n\nPretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).",
"### Training data\n\nThis model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.",
"### Preprocessing\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.",
"### Pretraining\n\nThe model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.",
"### Authors\n\nDimitris Papaevagelou - @andefined",
"### About Us\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
"TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #twitter #Greek #el #autotrain_compatible #endpoints_compatible #region-us \n",
"# Greek RoBERTa Uncased (v1)\n\nPretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).",
"### Training data\n\nThis model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.",
"### Preprocessing\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.",
"### Pretraining\n\nThe model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.",
"### Authors\n\nDimitris Papaevagelou - @andefined",
"### About Us\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
49,
66,
35,
77,
77,
15,
31
] |
[
"passage: TAGS\n#transformers #pytorch #safetensors #roberta #fill-mask #twitter #Greek #el #autotrain_compatible #endpoints_compatible #region-us \n# Greek RoBERTa Uncased (v1)\n\nPretrained model on Greek language using a masked language modeling (MLM) objective using Hugging Face's Transformers library. This model is case-sensitive and has no Greek diacritics (uncased, no-accents).### Training data\n\nThis model was pretrained on almost 18M unique tweets, all Greek, collected between 2008-2021, from almost 450K distinct users.### Preprocessing\n\nThe texts are tokenized using a byte version of Byte-Pair Encoding (BPE) and a vocabulary size of 50256. For the tokenizer we splited strings containing any numbers (ex. EU2019 ==> EU 2019). The tweet normalization logic described in the example listed bellow.### Pretraining\n\nThe model was pretrained on a T4 GPU for 1.2M steps with a batch size of 96 and a sequence length of 96. The optimizer used is Adam with a learning rate of 1e-5, gradient accumulation steps of 8, learning rate warmup for 50000 steps and linear decay of the learning rate after.### Authors\n\nDimitris Papaevagelou - @andefined### About Us\n\nCivic Information Office is a Non Profit Organization based in Athens, Greece focusing on creating technology and research products for the public interest."
] |
[
-0.13889215886592865,
0.030834985896945,
-0.0015109189553186297,
0.08547939360141754,
0.08499906212091446,
-0.016043299809098244,
0.06009538099169731,
0.04357997700572014,
0.011621802113950253,
0.11339075118303299,
0.11407652497291565,
-0.0004191815387457609,
0.053026653826236725,
0.18937434256076813,
0.009774249978363514,
-0.28796571493148804,
0.058847442269325256,
-0.07105852663516998,
0.009257730096578598,
0.0975346639752388,
0.08654413372278214,
-0.0694483295083046,
0.06383232772350311,
-0.027119213715195656,
-0.013890231028199196,
-0.021831821650266647,
-0.01621975377202034,
-0.0564420185983181,
0.10838375985622406,
0.0049655805341899395,
0.0001803670747904107,
0.016937028616666794,
0.0835934728384018,
-0.12039882689714432,
0.012864972464740276,
0.07720614224672318,
0.020155994221568108,
0.03367253392934799,
0.08931837230920792,
-0.03399285674095154,
0.18137940764427185,
-0.09555332362651825,
0.09859590977430344,
0.010343542322516441,
-0.10839520394802094,
-0.06632605195045471,
-0.0924772322177887,
0.050536300987005234,
0.131745845079422,
0.08190163224935532,
-0.04173910617828369,
0.050641998648643494,
-0.06242506206035614,
0.0668710470199585,
0.05042340233922005,
-0.22819258272647858,
-0.04150019586086273,
0.050753459334373474,
0.005399129819124937,
0.060707468539476395,
-0.010835074819624424,
0.026827286928892136,
0.02524695172905922,
0.010281451046466827,
0.03504930064082146,
0.021263262256979942,
0.16130752861499786,
-0.03491296246647835,
-0.15042772889137268,
-0.042819540947675705,
0.07277262955904007,
0.018377628177404404,
-0.04303625226020813,
-0.12056078761816025,
-0.03297596052289009,
-0.05710771307349205,
-0.013027737848460674,
-0.010352504439651966,
-0.02997669391334057,
0.020856505259871483,
0.024063678458333015,
-0.01080864854156971,
-0.07287675142288208,
0.015986425802111626,
-0.03026767633855343,
0.2834857404232025,
0.05434795841574669,
0.03548193722963333,
0.02100761979818344,
0.07527301460504532,
-0.02980560064315796,
-0.06328009814023972,
0.033617500215768814,
0.03359132260084152,
-0.11746644228696823,
-0.013973170891404152,
-0.03572867438197136,
-0.08954455703496933,
-0.03466826304793358,
0.12650208175182343,
-0.01594570279121399,
0.052815984934568405,
0.14070016145706177,
0.00918540544807911,
0.0020594701636582613,
0.09310995787382126,
-0.07356396317481995,
-0.04964419826865196,
0.03709018602967262,
0.10069368779659271,
0.007404905278235674,
-0.0627676323056221,
-0.05409747362136841,
0.025055183097720146,
0.10519139468669891,
0.05127861350774765,
0.004602259956300259,
0.05049071088433266,
-0.04753684625029564,
-0.03072742000222206,
0.09258073568344116,
-0.12902753055095673,
-0.025553859770298004,
-0.035591766238212585,
-0.04197143018245697,
0.004831272643059492,
-0.006055430043488741,
-0.03155745938420296,
-0.11586140841245651,
0.12312450259923935,
-0.05851628631353378,
-0.06189403682947159,
-0.04144364222884178,
-0.05945318937301636,
0.0027333530597388744,
-0.11997803300619125,
-0.04176472872495651,
-0.10706999152898788,
-0.15946917235851288,
-0.04144585505127907,
0.0837322250008583,
-0.02009168453514576,
-0.062369443476200104,
-0.06051623076200485,
-0.03822134807705879,
-0.009675348177552223,
0.00838912557810545,
0.06466296315193176,
-0.051081299781799316,
0.07884889841079712,
-0.09043184667825699,
0.05664859339594841,
-0.013861074112355709,
-0.00041311135282739997,
-0.13878974318504333,
-0.01608668826520443,
-0.09728749096393585,
0.1162363737821579,
-0.028988424688577652,
0.0036310849245637655,
-0.1029307171702385,
-0.04355468973517418,
-0.03494392707943916,
0.01365678384900093,
0.06540026515722275,
0.17536702752113342,
-0.1680482029914856,
-0.06050771847367287,
0.03588971868157387,
-0.11005803942680359,
0.0008418802754022181,
0.16519129276275635,
-0.02476557530462742,
0.021853096783161163,
0.09719745814800262,
0.1862206608057022,
-0.023417485877871513,
-0.11329659819602966,
-0.04620855301618576,
-0.05440265312790871,
-0.023827925324440002,
0.061634719371795654,
0.016633963212370872,
-0.048907600343227386,
-0.053946685045957565,
0.032640594989061356,
-0.0857607051730156,
0.040263157337903976,
-0.045312270522117615,
-0.030846424400806427,
0.016885191202163696,
-0.0744595155119896,
0.04293488711118698,
0.010941319167613983,
0.022692883387207985,
-0.07335828989744186,
-0.13566993176937103,
-0.060882359743118286,
0.05777646601200104,
-0.08503541350364685,
0.03406120091676712,
-0.08917705714702606,
0.11662416160106659,
-0.024414854124188423,
0.019507277756929398,
-0.1829083114862442,
-0.037066686898469925,
0.06082339212298393,
-0.09653367847204208,
0.08871281892061234,
-0.09694564342498779,
0.07442518323659897,
0.030173325911164284,
-0.06027163565158844,
-0.033112578094005585,
-0.021510349586606026,
-0.011259582825005054,
-0.11905843019485474,
-0.13567613065242767,
0.012154909782111645,
-0.04174782708287239,
0.007858308963477612,
-0.08387603610754013,
0.018306149169802666,
0.07018029689788818,
0.11774113029241562,
0.03007475845515728,
-0.10898873955011368,
-0.015244349837303162,
0.01696367934346199,
0.036803122609853745,
-0.08647317439317703,
0.006020023487508297,
0.00848044641315937,
-0.06306672096252441,
0.07039991766214371,
-0.23210154473781586,
-0.07518859207630157,
0.09138783812522888,
0.09664598852396011,
-0.11621090769767761,
-0.011662313714623451,
-0.035641446709632874,
-0.01941792480647564,
-0.0536186508834362,
0.05447180196642876,
0.27448293566703796,
-0.0005915474030189216,
0.0866001695394516,
-0.09551713615655899,
-0.03383766859769821,
0.05487916246056557,
0.010953330434858799,
-0.03260405734181404,
0.09229989349842072,
0.04865572974085808,
-0.12871767580509186,
-0.00874178484082222,
0.026395713910460472,
0.013699891045689583,
0.13973166048526764,
-0.020058520138263702,
-0.08617880940437317,
0.004346927162259817,
0.027522427961230278,
0.023857159540057182,
0.1628355234861374,
-0.0682796984910965,
0.0027571478858590126,
-0.0024120875168591738,
0.07058870047330856,
0.04642999917268753,
-0.114796943962574,
0.03486732393503189,
0.0037705181166529655,
-0.031301308423280716,
0.009647940285503864,
0.024880075827240944,
-0.051382940262556076,
0.0994889959692955,
0.022970642894506454,
-0.05435511842370033,
-0.023533614352345467,
0.004022557288408279,
-0.11988946795463562,
0.21260084211826324,
-0.053811147809028625,
-0.17164839804172516,
-0.06023256108164787,
0.06613385677337646,
0.0010106588015332818,
-0.03896765038371086,
-0.005513446405529976,
-0.08464598655700684,
-0.08009810745716095,
-0.11454927176237106,
0.08793626725673676,
-0.016452474519610405,
0.07449153065681458,
-0.02331405319273472,
-0.03443829342722893,
-0.038323499262332916,
-0.057199664413928986,
-0.03537757322192192,
-0.1045909970998764,
-0.03914112225174904,
0.004698712378740311,
-0.057384103536605835,
0.05192112177610397,
0.1942073553800583,
-0.0680900514125824,
0.03437689691781998,
-0.04542793333530426,
0.15743064880371094,
-0.10589908808469772,
0.10582911223173141,
0.09739401191473007,
0.007095348555594683,
0.03229719027876854,
0.10170156508684158,
-0.0017564500449225307,
-0.11614775657653809,
0.049617379903793335,
0.05505840480327606,
-0.09056475758552551,
-0.18526312708854675,
-0.12036076933145523,
-0.05746270716190338,
-0.011569581925868988,
0.1128380224108696,
0.04673232510685921,
-0.058146655559539795,
0.004124309867620468,
-0.014849056489765644,
-0.015985406935214996,
0.09379352629184723,
0.0848848968744278,
0.06191689148545265,
0.05820617079734802,
0.07360504567623138,
-0.022065257653594017,
0.01496030017733574,
0.11895903944969177,
-0.048157695680856705,
0.25741150975227356,
-0.0858839899301529,
0.029251694679260254,
0.014543390832841396,
0.035674989223480225,
-0.0003799172700382769,
0.05970175564289093,
0.01793246902525425,
0.019914662465453148,
-0.04041799530386925,
-0.03910832107067108,
-0.008001098409295082,
0.04749684035778046,
-0.006671961396932602,
-0.10548973828554153,
-0.10477648675441742,
0.06226939707994461,
0.07230488955974579,
0.25524598360061646,
0.07386273890733719,
-0.223506897687912,
-0.11473111063241959,
-0.02628651075065136,
-0.05994514748454094,
-0.04070160537958145,
0.05113445967435837,
0.0212734192609787,
-0.05210626870393753,
0.0669521912932396,
-0.04132915660738945,
0.06253974884748459,
-0.11208408325910568,
-0.028040161356329918,
0.000924514839425683,
0.10250742733478546,
-0.042886193841695786,
0.01669193245470524,
-0.17128998041152954,
0.19777309894561768,
0.014604667201638222,
0.15845274925231934,
-0.04020419716835022,
0.015461409464478493,
-0.003248240565881133,
0.0401575081050396,
0.12174463272094727,
0.04079097881913185,
-0.04133627936244011,
-0.006465288810431957,
-0.14903151988983154,
0.03773225098848343,
0.10414217412471771,
-0.05092526972293854,
0.029764046892523766,
-0.01029679924249649,
0.02596236579120159,
-0.027392111718654633,
-0.19671744108200073,
-0.08213359862565994,
-0.147467240691185,
0.05088776722550392,
-0.11649809032678604,
-0.07326240092515945,
-0.054815877228975296,
-0.10591621696949005,
0.009133022278547287,
0.1953299194574356,
0.03417939320206642,
-0.06766388565301895,
-0.11981935799121857,
-0.014713726006448269,
0.13230928778648376,
-0.07034007459878922,
0.054055724292993546,
0.0027560596354305744,
0.008802722208201885,
-0.08662381768226624,
-0.06806470453739166,
0.13252179324626923,
-0.06977096945047379,
-0.19745363295078278,
-0.03751804679632187,
0.12691713869571686,
0.09710227698087692,
0.003912351094186306,
-0.0029478538781404495,
0.056875716894865036,
0.006085332948714495,
-0.10549645125865936,
0.02242925576865673,
0.017879903316497803,
0.003912637941539288,
0.058802492916584015,
0.08836042135953903,
-0.14412078261375427,
-0.08851473033428192,
-0.011216102167963982,
-0.017044605687260628,
0.30608510971069336,
-0.04454493522644043,
0.05891108140349388,
0.2315647304058075,
-0.05275919288396835,
-0.18885394930839539,
0.011645983904600143,
0.015861550346016884,
0.03526277467608452,
-0.10113989561796188,
-0.09448014944791794,
0.01462764386087656,
0.07183791697025299,
-0.029614388942718506,
-0.02073974721133709,
-0.2644074559211731,
-0.16140323877334595,
0.07155248522758484,
0.060483936220407486,
0.21439853310585022,
-0.0458269864320755,
0.02689659409224987,
0.03901373967528343,
0.05086878687143326,
0.08379879593849182,
0.11475405097007751,
0.05636497214436531,
0.024090619757771492,
-0.020657572895288467,
0.035110149532556534,
-0.05366680771112442,
0.09693602472543716,
0.03696287423372269,
0.009229999966919422,
-0.033254001289606094,
-0.01400453969836235,
-0.025979695841670036,
-0.037248414009809494,
0.11161135137081146,
0.009551459923386574,
0.05877988785505295,
-0.09477690607309341,
-0.07864458858966827,
-0.0666072741150856,
0.05877253785729408,
-0.009792137891054153,
-0.08317770063877106,
-0.07168682664632797,
0.08146300166845322,
0.0488695353269577,
-0.008364320732653141,
-0.051629941910505295,
-0.05967927724123001,
-0.054841142147779465,
0.0699208676815033,
0.15894049406051636,
-0.08498968929052353,
-0.03450244292616844,
0.02885415218770504,
-0.05186077207326889,
0.05749012902379036,
-0.16626684367656708,
-0.045946668833494186,
0.10089036822319031,
0.011310816742479801,
0.07516943663358688,
0.014034229330718517,
-0.16668252646923065,
0.005119022913277149,
0.05670760199427605,
-0.13262589275836945,
-0.14759545028209686,
0.03166773170232773,
-0.020151961594820023,
-0.05694343522191048,
0.06014106422662735,
0.09671883285045624,
-0.10075151920318604,
-0.008959256112575531,
-0.010561985895037651,
0.07667050510644913,
0.0017134346999228,
0.1091192215681076,
0.027595899999141693,
0.010182269848883152,
-0.07375553250312805,
0.15510039031505585,
0.06968935579061508,
-0.1798023134469986,
0.03273286670446396,
0.14710085093975067,
-0.12567606568336487,
-0.026467997580766678,
0.028500445187091827,
-0.005493116565048695,
0.009767541661858559,
-0.006053413730114698,
-0.068646639585495,
-0.08446177840232849,
0.07818963378667831,
0.09299354255199432,
0.032967325299978256,
-0.015752481296658516,
-0.0793476402759552,
0.008755392394959927,
-0.10526282340288162,
0.04278825595974922,
0.1259368658065796,
0.012197923846542835,
-0.015460792928934097,
0.0864538699388504,
0.01906404085457325,
-0.022213350981473923,
-0.03301222249865532,
0.00034356117248535156,
-0.08030682802200317,
-0.0154968136921525,
0.06133369728922844,
-0.03495471924543381,
-0.07730121910572052,
0.016471313312649727,
-0.0360553152859211,
-0.05333331599831581,
-0.005540343467146158,
0.01870899647474289,
-0.04236839339137077,
-0.007842550054192543,
-0.0053185042925179005,
-0.027561809867620468,
-0.05814288929104805,
-0.004766130354255438,
0.03889811038970947,
-0.08628130704164505,
0.178212970495224,
-0.012286379933357239,
-0.02678641676902771,
0.09452126920223236,
-0.08431747555732727,
0.090990349650383,
0.03721470385789871,
0.0038274338003247976,
0.0028489951509982347,
-0.05611877143383026,
0.01853238046169281,
-0.014109376817941666,
0.02459586225450039,
0.04231778159737587,
0.10296720266342163,
-0.06613322347402573,
0.11179281026124954,
-0.009177609346807003,
-0.018171051517128944,
-0.09601890295743942,
0.07786044478416443,
-0.0006152652786113322,
0.0850493535399437,
0.05934595689177513,
-0.06858760863542557,
0.014521765522658825,
-0.12303241342306137,
-0.014619279652833939,
0.02826182171702385,
-0.05174340680241585,
-0.04559171944856644,
-0.0036679457407444715,
0.11322351545095444,
-0.008440245874226093,
0.1104946881532669,
0.02270655892789364,
-0.013155879452824593,
0.03725620359182358,
0.024561448022723198,
0.0031236037611961365,
0.04467571899294853,
0.03185216709971428,
-0.04174024239182472,
-0.0046828798949718475,
-0.049553826451301575,
-0.001386473304592073,
-0.00031340442365035415,
0.047034502029418945,
0.18116480112075806,
0.06473365426063538,
0.0742889866232872,
0.08664427697658539,
-0.0454479455947876,
-0.04651385918259621,
-0.14235077798366547,
0.04538944363594055,
-0.08137072622776031,
0.09477076679468155,
0.012328378856182098,
0.11615276336669922,
0.1516081988811493,
-0.1500978171825409,
0.044408414512872696,
0.02679440565407276,
-0.10936439782381058,
-0.08180243521928787,
-0.17902618646621704,
-0.019815364852547646,
-0.02534199319779873,
0.0027534663677215576,
-0.08128737658262253,
-0.0008985170279629529,
0.0721656009554863,
0.03387413173913956,
0.008194572292268276,
0.0727866068482399,
-0.025558069348335266,
-0.006133462768048048,
0.1373906433582306,
0.008048200979828835,
0.02414938434958458,
0.10265649110078812,
-0.08629734069108963,
-0.04807807877659798,
-0.017796291038393974,
0.060495927929878235,
0.02473585493862629,
-0.059596385806798935,
0.06130366772413254,
0.01274432148784399,
-0.0410507395863533,
0.01775008998811245,
-0.012676114216446877,
0.036547254770994186,
0.1664261668920517,
0.06863336265087128,
-0.054542578756809235,
-0.0016500625060871243,
0.22080670297145844,
-0.00030185465584509075,
-0.07510529458522797,
-0.13354940712451935,
0.1500498354434967,
-0.013743002898991108,
0.03673967346549034,
0.050718553364276886,
-0.08562523871660233,
0.0014097228413447738,
0.14164437353610992,
0.12958692014217377,
-0.016614500433206558,
-0.006028094794601202,
-0.02674633450806141,
-0.019631410017609596,
0.017613288015127182,
0.12533612549304962,
-0.009127453900873661,
0.22829194366931915,
-0.06395868957042694,
0.05709519609808922,
-0.060017745941877365,
-0.060041166841983795,
-0.09386958181858063,
0.10072840750217438,
0.02160189114511013,
-0.02783544920384884,
-0.08564281463623047,
0.12098096311092377,
-0.023614676669239998,
-0.15995189547538757,
-0.03217460960149765,
-0.12081661820411682,
-0.12870751321315765,
-0.0165437962859869,
-0.02873748354613781,
-0.02570880576968193,
0.11883123219013214,
0.007501387502998114,
0.02042268216609955,
-0.053202223032712936,
0.04284951463341713,
-0.06990204006433487,
-0.08126174658536911,
0.04896772652864456,
0.0948154404759407,
0.19075095653533936,
0.03095564804971218,
0.09717079997062683,
0.06333848834037781,
0.02926539070904255,
-0.1010890081524849,
0.05723383650183678,
0.001958828419446945,
0.07219676673412323,
0.041773322969675064,
0.0718727633357048,
-0.034371595829725266,
-0.004037057980895042,
0.07638771831989288,
-0.011987896636128426,
0.056716252118349075,
-0.15766558051109314,
-0.0388057604432106,
-0.06856923550367355,
0.06822583079338074,
-0.05377673730254173,
0.10045503824949265,
0.1340218186378479,
-0.013896612450480461,
0.01700476184487343,
-0.04576883092522621,
0.004195115063339472,
-0.0006409371271729469,
0.03142477571964264,
-0.010731512680649757,
-0.13830289244651794,
-0.020136887207627296,
-0.029735183343291283,
0.03414566069841385,
-0.09466246515512466,
-0.01722305826842785,
-0.0966418907046318,
-0.01781177893280983,
-0.05387331172823906,
0.1676945984363556,
0.02598477154970169,
-0.04100999981164932,
-0.032887816429138184,
0.025507282465696335,
0.039965614676475525,
0.06651737540960312,
-0.11982947587966919,
-0.061949241906404495
] |
null | null |
transformers
|
## Hello World
|
{}
|
token-classification
|
cwtpc/wangchanberta-ner-8989
|
[
"transformers",
"pytorch",
"camembert",
"token-classification",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #camembert #token-classification #autotrain_compatible #endpoints_compatible #region-us
|
## Hello World
|
[
"## Hello World"
] |
[
"TAGS\n#transformers #pytorch #camembert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n",
"## Hello World"
] |
[
39,
3
] |
[
"passage: TAGS\n#transformers #pytorch #camembert #token-classification #autotrain_compatible #endpoints_compatible #region-us \n## Hello World"
] |
[
-0.020713642239570618,
0.10025706142187119,
-0.008806237950921059,
0.03443806990981102,
0.1759691834449768,
0.013023915700614452,
0.051259372383356094,
0.08922743797302246,
0.06123132258653641,
-0.004442824982106686,
0.1286446899175644,
0.22500000894069672,
-0.02998882532119751,
0.11595501750707626,
-0.05177784338593483,
-0.3157079517841339,
0.0666830763220787,
0.054703060537576675,
-0.07739651948213577,
0.09963671863079071,
0.08838541805744171,
-0.09702940285205841,
0.09185191988945007,
-0.03266112133860588,
-0.11767882108688354,
0.05003638193011284,
0.010805368423461914,
-0.09966360032558441,
0.10540498048067093,
0.02655402198433876,
0.15496188402175903,
0.039344191551208496,
-0.04194609075784683,
-0.13709315657615662,
0.03231203183531761,
-0.010647383518517017,
-0.043654270470142365,
0.04409871995449066,
0.04905543103814125,
-0.09302467107772827,
0.004680538084357977,
0.046962328255176544,
0.022645652294158936,
0.06551717966794968,
-0.14305689930915833,
-0.1424890011548996,
-0.03294418007135391,
0.020055016502738,
0.028606481850147247,
0.049244627356529236,
0.02853645198047161,
0.22696399688720703,
-0.11606154590845108,
0.10424420982599258,
0.10246043652296066,
-0.26055219769477844,
-0.005851516034454107,
0.10068424046039581,
0.017376504838466644,
0.003740768413990736,
-0.031285203993320465,
0.03467675298452377,
0.032232724130153656,
0.007534687872976065,
0.005132865626364946,
-0.06762788444757462,
-0.0619046725332737,
0.00778287835419178,
-0.09605979919433594,
-0.01625429093837738,
0.1957307904958725,
-0.03221413493156433,
0.03792569786310196,
0.02621643990278244,
-0.09021186083555222,
-0.09153009206056595,
-0.02880677953362465,
-0.02577345445752144,
-0.015620185993611813,
0.0492507703602314,
0.028436673805117607,
0.009643921628594398,
-0.10939407348632812,
0.04698970913887024,
-0.21852096915245056,
0.2455911636352539,
0.01844179816544056,
0.04956270754337311,
-0.18072767555713654,
0.03295471891760826,
-0.004912005737423897,
-0.07361463457345963,
0.017371149733662605,
-0.10063320398330688,
0.023588255047798157,
-0.043418917804956436,
0.003337407950311899,
0.03444160148501396,
0.09316720068454742,
0.14404872059822083,
0.04844031110405922,
0.07339771836996078,
0.00003969831232097931,
0.07872821390628815,
0.06872697174549103,
0.06915637105703354,
0.017412113025784492,
-0.09729529172182083,
0.04192199558019638,
-0.1475607454776764,
-0.017041297629475594,
-0.05506838485598564,
-0.11554975062608719,
-0.0159319955855608,
0.049245428293943405,
0.08820835500955582,
0.04297875985503197,
0.08941751718521118,
-0.03105776198208332,
-0.029473816975951195,
0.12078247964382172,
-0.06752034276723862,
0.022869793698191643,
0.014581876806914806,
-0.020684242248535156,
0.10731153935194016,
-0.05739589408040047,
-0.004364282824099064,
-0.04104932025074959,
0.09061844646930695,
-0.07272090762853622,
-0.012453820556402206,
-0.037840619683265686,
-0.032529961317777634,
0.04809851199388504,
-0.13318097591400146,
0.04288773611187935,
-0.1690087914466858,
-0.04472148045897484,
0.024060124531388283,
0.024774761870503426,
0.016080940142273903,
-0.03183286264538765,
-0.012304326519370079,
0.008822249248623848,
0.009222088381648064,
-0.04431065171957016,
-0.06483852863311768,
-0.07180867344141006,
0.08584655076265335,
-0.0018455618992447853,
0.07158781588077545,
-0.1629096418619156,
0.06694493442773819,
-0.11375857144594193,
0.021222036331892014,
-0.11069060117006302,
-0.01607503369450569,
-0.08891256153583527,
0.15100371837615967,
-0.0035545174032449722,
-0.04584035649895668,
-0.033399004489183426,
0.034961897879838943,
-0.04265034571290016,
0.1513187438249588,
-0.073786661028862,
-0.10726507753133774,
0.17530547082424164,
-0.11254021525382996,
-0.14282234013080597,
0.07108346372842789,
0.005023092031478882,
-0.002250480931252241,
0.04734962806105614,
0.10155566036701202,
0.05017751827836037,
-0.0009348875028081238,
0.041054435074329376,
0.07190616428852081,
-0.14960771799087524,
-0.11221160739660263,
-0.006965389009565115,
0.001447196933440864,
-0.12095314264297485,
0.065846748650074,
0.08743921667337418,
0.07194831222295761,
-0.08080489933490753,
-0.02914266474545002,
-0.014715813100337982,
-0.010685648769140244,
0.10104677081108093,
0.06878065317869186,
0.0948602706193924,
-0.0829242467880249,
0.029559001326560974,
-0.004609512630850077,
0.03443969041109085,
0.045053571462631226,
0.005128407385200262,
-0.0933147519826889,
0.14765727519989014,
0.008999976329505444,
0.004132147878408432,
-0.1566007137298584,
-0.06273510307073593,
0.010939753614366055,
0.10357926040887833,
0.012308657169342041,
0.13809436559677124,
0.08407579362392426,
-0.041781485080718994,
0.015817072242498398,
-0.04216063395142555,
0.1812426745891571,
0.01259149145334959,
-0.03554088994860649,
-0.08627676218748093,
0.07397095859050751,
-0.060726314783096313,
0.006247902754694223,
-0.05278143659234047,
0.008820130489766598,
0.11348483711481094,
0.15358762443065643,
0.021858956664800644,
0.06042046099901199,
-0.03262457251548767,
0.051080409437417984,
-0.09330898523330688,
0.008726900443434715,
0.1054287925362587,
-0.015805194154381752,
-0.07656118273735046,
0.13032017648220062,
-0.14830201864242554,
0.2823004126548767,
0.2150687277317047,
-0.3542865216732025,
-0.02235523797571659,
-0.004951581358909607,
-0.005400826223194599,
0.0117062171921134,
0.056819625198841095,
0.06605638563632965,
0.0244549922645092,
-0.005566845647990704,
0.1680983304977417,
0.005398008041083813,
-0.013982637785375118,
0.015258053317666054,
-0.07670781761407852,
-0.05635349079966545,
0.05855857953429222,
0.049835868179798126,
-0.1815086305141449,
0.18188084661960602,
0.22186391055583954,
0.008949787355959415,
0.1249474361538887,
0.0002269129327032715,
0.03897581994533539,
0.06337998062372208,
-0.0118575319647789,
-0.018605271354317665,
-0.014490635134279728,
-0.219120591878891,
-0.040474649518728256,
0.05525413900613785,
0.019941408187150955,
0.032248493283987045,
-0.08600815385580063,
-0.03597691282629967,
0.0068766712211072445,
0.05025280639529228,
0.03763992711901665,
0.09085102379322052,
0.04612063243985176,
0.09765227138996124,
0.020610833540558815,
-0.09944778680801392,
0.08449171483516693,
0.009522094391286373,
-0.04823427274823189,
0.17175476253032684,
-0.11070796847343445,
-0.3022056519985199,
-0.08747333288192749,
-0.2122466266155243,
-0.03772692009806633,
0.04600506275892258,
0.06692832708358765,
-0.08840576559305191,
-0.04835090786218643,
0.06905083358287811,
-0.005860600620508194,
-0.07049162685871124,
0.05978662893176079,
-0.0652894377708435,
0.07188611477613449,
-0.02904854714870453,
-0.03839289769530296,
-0.059211667627096176,
-0.031544603407382965,
-0.038829948753118515,
0.1504257619380951,
-0.07631048560142517,
0.09016036987304688,
0.15911734104156494,
-0.018633155152201653,
0.05312039703130722,
-0.013438583351671696,
0.1879112422466278,
-0.08537967503070831,
-0.026246367022395134,
0.13793042302131653,
-0.06505429744720459,
0.08559147268533707,
0.1441076099872589,
0.05202966555953026,
-0.0733344629406929,
0.023693496361374855,
-0.017835363745689392,
-0.11205238848924637,
-0.13722074031829834,
-0.1297808289527893,
-0.0921013131737709,
0.07737217098474503,
0.03852121904492378,
0.07958061993122101,
0.1442013680934906,
0.0929555669426918,
0.04273306950926781,
-0.0337039940059185,
-0.050543297082185745,
0.08779928088188171,
0.15276868641376495,
-0.023936448618769646,
0.13166190683841705,
-0.05969808250665665,
-0.15984736382961273,
0.07792395353317261,
0.07268506288528442,
0.06611090153455734,
0.12471627444028854,
-0.01089635118842125,
0.022145921364426613,
0.1313825249671936,
0.15682563185691833,
0.10584491491317749,
0.030644699931144714,
-0.023952431976795197,
-0.0037815854884684086,
-0.0023720005992799997,
-0.08931370079517365,
0.012191473506391048,
0.15358787775039673,
-0.10905009508132935,
-0.06942235678434372,
-0.12985356152057648,
0.05432172492146492,
0.11613300442695618,
0.0591675266623497,
-0.14988401532173157,
0.020110178738832474,
0.07833822071552277,
-0.009368429891765118,
-0.06330853700637817,
0.07762768119573593,
-0.07587604224681854,
-0.12451773881912231,
0.08275100588798523,
-0.01360382977873087,
0.10531105846166611,
-0.08659174293279648,
0.05745347589254379,
-0.04997837170958519,
-0.10137386620044708,
0.037645161151885986,
0.07064469158649445,
-0.24529077112674713,
0.2502822279930115,
-0.0028862268663942814,
-0.08947984874248505,
-0.08015529811382294,
-0.016405945643782616,
0.03678901121020317,
0.21500301361083984,
0.10434021800756454,
0.010352516546845436,
-0.11373793333768845,
-0.19400548934936523,
0.009370436891913414,
-0.02567845582962036,
0.11108183860778809,
-0.009282871149480343,
-0.008996783755719662,
-0.03829415887594223,
-0.0260236207395792,
-0.05468940734863281,
-0.026748383417725563,
0.03069904074072838,
-0.1169932633638382,
0.06183454394340515,
0.03738247975707054,
0.06252629309892654,
0.015991536900401115,
-0.02790035866200924,
-0.13001935184001923,
0.1628253310918808,
-0.07771766185760498,
-0.022129608318209648,
-0.10633914172649384,
-0.11818438023328781,
0.03016403503715992,
-0.09198429435491562,
0.06251797825098038,
-0.10365168750286102,
-0.01877828873693943,
-0.029879380017518997,
-0.14972320199012756,
0.1461184024810791,
-0.07163476198911667,
-0.05796762555837631,
-0.06619498878717422,
0.12738247215747833,
-0.050333499908447266,
0.0000485123491671402,
-0.0012934646802023053,
0.01606133207678795,
-0.06243335083127022,
-0.07223925739526749,
0.027741001918911934,
-0.0366252101957798,
0.027264641597867012,
0.05685688182711601,
-0.04516279324889183,
-0.047843363136053085,
-0.02252037078142166,
0.02736760303378105,
0.24212627112865448,
0.21672853827476501,
-0.06996624171733856,
0.10443873703479767,
0.16671624779701233,
-0.03726040944457054,
-0.3421591818332672,
-0.06910421699285507,
-0.13674597442150116,
-0.061615727841854095,
-0.06236092001199722,
-0.13815806806087494,
0.11003949493169785,
0.019402239471673965,
-0.058313872665166855,
0.0995226502418518,
-0.12131083756685257,
-0.059225600212812424,
0.17481385171413422,
0.018278468400239944,
0.3587537407875061,
-0.09558431059122086,
-0.05680680647492409,
-0.018896881490945816,
-0.14855614304542542,
0.1158064752817154,
0.06817933917045593,
0.09693427383899689,
-0.030711622908711433,
0.023491734638810158,
0.045115403831005096,
-0.05320701748132706,
0.08714737743139267,
-0.010712971910834312,
0.037921879440546036,
-0.12850987911224365,
-0.09939440339803696,
-0.0284144077450037,
-0.010073098354041576,
0.0007203428540378809,
0.034742049872875214,
0.014208880253136158,
-0.10686993598937988,
-0.014508910477161407,
-0.0921170637011528,
0.12706205248832703,
0.029085543006658554,
-0.07623688131570816,
-0.008392123505473137,
-0.023684682324528694,
-0.022706402465701103,
0.023025358095765114,
0.2847757339477539,
-0.04533744230866432,
0.14722940325737,
0.113134004175663,
0.10718613117933273,
-0.1832704097032547,
0.01908540539443493,
-0.06889280676841736,
-0.053723402321338654,
0.06515437364578247,
-0.06213353946805,
0.07258941978216171,
0.12684616446495056,
-0.04594774916768074,
0.025150850415229797,
0.10704398900270462,
0.018880033865571022,
-0.02282615937292576,
0.15382783114910126,
-0.21087245643138885,
-0.05144799128174782,
-0.01887974701821804,
0.03558744862675667,
0.09622368961572647,
0.11664950102567673,
0.1020505502820015,
0.04160699248313904,
-0.03918534889817238,
0.0061564575880765915,
-0.021960211917757988,
-0.035340264439582825,
0.04256027191877365,
0.06650233268737793,
0.029446953907608986,
-0.13197557628154755,
0.049299269914627075,
0.0318022258579731,
-0.20102903246879578,
-0.04306856915354729,
0.16430051624774933,
-0.11977240443229675,
-0.13310669362545013,
0.005645920522511005,
0.11983425915241241,
-0.14886312186717987,
-0.037850309163331985,
-0.012266790494322777,
-0.11563581973314285,
0.07378631085157394,
0.1960964798927307,
0.12024813890457153,
0.0926031842827797,
0.0034893148113042116,
-0.016391199082136154,
-0.0018585233483463526,
-0.038484424352645874,
-0.0383073128759861,
0.046073075383901596,
-0.1370113044977188,
-0.008825665339827538,
-0.004770656116306782,
0.16718018054962158,
-0.10605547577142715,
-0.0929407998919487,
-0.14693240821361542,
0.018168341368436813,
-0.0625934898853302,
-0.10274574160575867,
-0.11409033089876175,
-0.03313785418868065,
0.020585304126143456,
-0.10201369971036911,
-0.010639951564371586,
-0.0626838281750679,
-0.11526976525783539,
0.0367356613278389,
0.02645130269229412,
0.020174484699964523,
-0.04644716903567314,
-0.026963436976075172,
0.11505013704299927,
-0.025109151378273964,
0.12781652808189392,
0.10290892422199249,
-0.06445075571537018,
0.07864794880151749,
-0.08940636366605759,
-0.0846819058060646,
0.10623971372842789,
0.013580354861915112,
0.11034788936376572,
0.07662410289049149,
0.01721748523414135,
0.03700981289148331,
0.03688126057386398,
0.05881321057677269,
0.08363072574138641,
-0.11449918895959854,
0.03720887750387192,
-0.0013651682529598475,
-0.17211057245731354,
-0.010408339090645313,
-0.06377691775560379,
0.12145452201366425,
-0.008172819390892982,
0.12612970173358917,
-0.025000743567943573,
0.10061054676771164,
-0.019242456182837486,
0.01686469465494156,
-0.01795363985002041,
-0.21729062497615814,
-0.031327612698078156,
-0.05904213339090347,
0.009234683588147163,
-0.02017296478152275,
0.19593703746795654,
-0.010924240574240685,
0.03002255968749523,
0.057841986417770386,
0.04793616011738777,
-0.0015675697941333055,
0.04097912460565567,
0.15516862273216248,
0.08325263857841492,
-0.061982154846191406,
-0.05456039682030678,
0.08805672079324722,
0.0035377680324018,
-0.043313585221767426,
0.09956426918506622,
0.05261862277984619,
-0.022016452625393867,
0.04049248620867729,
0.01638689637184143,
0.03358166292309761,
-0.13766269385814667,
-0.2156456708908081,
-0.0430288091301918,
0.043902039527893066,
0.004637292120605707,
0.06893377751111984,
0.0908796489238739,
0.009886275045573711,
0.03651933744549751,
-0.05278649553656578,
-0.016945116221904755,
-0.16621345281600952,
-0.06385431438684464,
-0.09997095912694931,
-0.09246046841144562,
0.017705345526337624,
-0.039673224091529846,
-0.024821901693940163,
0.03401292487978935,
0.06557337939739227,
-0.062298733741045,
0.06484554708003998,
-0.006162852514535189,
0.0013467546086758375,
0.01756233163177967,
0.0050670490600168705,
-0.010809245519340038,
-0.0038332566618919373,
-0.03155558556318283,
-0.19792859256267548,
-0.0020624431781470776,
-0.06680745631456375,
0.017848435789346695,
-0.09242182970046997,
0.007360225077718496,
-0.11882893741130829,
-0.11388389766216278,
-0.020694613456726074,
0.032417166978120804,
-0.05591730400919914,
0.08648976683616638,
-0.03550132364034653,
0.038113903254270554,
0.011762780137360096,
0.17190726101398468,
-0.06932131946086884,
-0.06117447465658188,
-0.031157679855823517,
0.23836812376976013,
0.03568531572818756,
0.13365277647972107,
-0.03453025221824646,
0.01994147337973118,
-0.08931102603673935,
0.2965608835220337,
0.2826343774795532,
-0.031191963702440262,
0.05637151747941971,
0.023857155814766884,
0.021968532353639603,
0.0711701288819313,
0.09498167037963867,
0.0929773822426796,
0.20383574068546295,
-0.08021513372659683,
-0.02829308807849884,
-0.05191989988088608,
0.018051808699965477,
-0.11114205420017242,
0.08119910955429077,
0.05078539252281189,
-0.027734387665987015,
-0.0779368132352829,
0.04326765984296799,
-0.2169463187456131,
0.1460457295179367,
0.021897103637456894,
-0.21945783495903015,
-0.07023286074399948,
-0.021584227681159973,
0.11698152869939804,
0.020614709705114365,
0.050038762390613556,
0.006201825104653835,
-0.07722686976194382,
0.03224457800388336,
0.028164273127913475,
-0.22368861734867096,
-0.053519509732723236,
0.05870307609438896,
-0.043208759278059006,
0.07573892921209335,
-0.017659390345215797,
-0.01913662627339363,
0.09053095430135727,
0.06246627867221832,
-0.039296478033065796,
0.01776166632771492,
0.020526211708784103,
-0.037481632083654404,
-0.04506540298461914,
0.03539389744400978,
0.028919387608766556,
-0.07367279380559921,
0.04664115235209465,
-0.16296660900115967,
0.042235199362039566,
-0.07359538972377777,
-0.017969438806176186,
-0.01596948318183422,
0.06203071400523186,
-0.03910883143544197,
0.04858998581767082,
0.06391781568527222,
0.019815683364868164,
-0.04402141645550728,
-0.046524062752723694,
-0.0350087471306324,
0.0188580509275198,
-0.11477411538362503,
-0.14824460446834564,
-0.12408174574375153,
-0.059207335114479065,
0.06891012191772461,
-0.005253707990050316,
-0.06321454793214798,
-0.034614671021699905,
-0.08397065848112106,
0.04314541816711426,
-0.11227713525295258,
0.07630158960819244,
0.056686945259571075,
0.039795875549316406,
-0.013803495094180107,
-0.03960544243454933,
0.03436680883169174,
0.07656688988208771,
-0.12503498792648315,
-0.08998812735080719
] |
null | null |
transformers
|
## Cyclone Chinese NER
This model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)
Currently, we only support 8 general type of entities ("address", "company", "government", "name", "organization", "position", "scene", "time")
### Usage
from transformers import BertConfig
config = BertConfig.from_pretrained("bert-base-chinese", num_labels=num_labels)
model_path = "cyclone/cyclone-ner"
tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)
model = BertCrfForNer.from_pretrained(model_path, config=config)
|
{}
| null |
cyclone/cyclone-ner
|
[
"transformers",
"pytorch",
"bert",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #endpoints_compatible #region-us
|
## Cyclone Chinese NER
This model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)
Currently, we only support 8 general type of entities ("address", "company", "government", "name", "organization", "position", "scene", "time")
### Usage
from transformers import BertConfig
config = BertConfig.from_pretrained("bert-base-chinese", num_labels=num_labels)
model_path = "cyclone/cyclone-ner"
tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)
model = BertCrfForNer.from_pretrained(model_path, config=config)
|
[
"## Cyclone Chinese NER\r\n\r\nThis model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)\r\nCurrently, we only support 8 general type of entities (\"address\", \"company\", \"government\", \"name\", \"organization\", \"position\", \"scene\", \"time\")",
"### Usage\r\n from transformers import BertConfig\r\n\r\n config = BertConfig.from_pretrained(\"bert-base-chinese\", num_labels=num_labels)\r\n\r\n model_path = \"cyclone/cyclone-ner\"\r\n\r\n tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)\r\n model = BertCrfForNer.from_pretrained(model_path, config=config)"
] |
[
"TAGS\n#transformers #pytorch #bert #endpoints_compatible #region-us \n",
"## Cyclone Chinese NER\r\n\r\nThis model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)\r\nCurrently, we only support 8 general type of entities (\"address\", \"company\", \"government\", \"name\", \"organization\", \"position\", \"scene\", \"time\")",
"### Usage\r\n from transformers import BertConfig\r\n\r\n config = BertConfig.from_pretrained(\"bert-base-chinese\", num_labels=num_labels)\r\n\r\n model_path = \"cyclone/cyclone-ner\"\r\n\r\n tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)\r\n model = BertCrfForNer.from_pretrained(model_path, config=config)"
] |
[
23,
74,
110
] |
[
"passage: TAGS\n#transformers #pytorch #bert #endpoints_compatible #region-us \n## Cyclone Chinese NER\r\n\r\nThis model provides simplified Chinese NER model based on pretrained model BERT (specifically BERT + CRF)\r\nCurrently, we only support 8 general type of entities (\"address\", \"company\", \"government\", \"name\", \"organization\", \"position\", \"scene\", \"time\")### Usage\r\n from transformers import BertConfig\r\n\r\n config = BertConfig.from_pretrained(\"bert-base-chinese\", num_labels=num_labels)\r\n\r\n model_path = \"cyclone/cyclone-ner\"\r\n\r\n tokenizer = CNerTokenizer.from_pretrained(model_path, do_lower_case=True)\r\n model = BertCrfForNer.from_pretrained(model_path, config=config)"
] |
[
-0.00883391685783863,
0.16348592936992645,
-0.0021209942642599344,
0.07804320752620697,
0.1171920970082283,
-0.04894353821873665,
0.06639860570430756,
-0.09048741310834885,
-0.04604075849056244,
0.11351130902767181,
0.05356454849243164,
-0.12329444289207458,
-0.012968155555427074,
-0.06474047154188156,
0.0033601305913180113,
-0.3353304862976074,
0.1149047389626503,
0.07340380549430847,
0.09062208235263824,
0.1066189631819725,
0.027814073488116264,
-0.07614495605230331,
0.10974299162626266,
0.1388358622789383,
-0.10443173348903656,
0.03679715469479561,
-0.11491724103689194,
-0.06516018509864807,
0.03823067247867584,
-0.05092986300587654,
0.20471946895122528,
-0.003203411353752017,
0.033864833414554596,
-0.0622108094394207,
0.009578125551342964,
-0.015694817528128624,
-0.035839494317770004,
0.056278567761182785,
0.11570657044649124,
0.06222368776798248,
0.08738869428634644,
0.027467031031847,
0.03628715127706528,
0.044785402715206146,
-0.08763279020786285,
-0.027777371928095818,
-0.0532035231590271,
0.22580306231975555,
0.10338688641786575,
0.11352836340665817,
0.029132230207324028,
0.07875959575176239,
-0.16616208851337433,
0.016675742343068123,
0.0816519483923912,
-0.22252541780471802,
-0.05822927504777908,
0.15204337239265442,
0.015568469651043415,
-0.07558964192867279,
-0.0205996073782444,
-0.035726457834243774,
0.03595423325896263,
0.013455281965434551,
-0.01310708373785019,
-0.05046100541949272,
-0.13950709998607635,
0.016185307875275612,
-0.15553800761699677,
0.07763293385505676,
0.1135963499546051,
0.08694218844175339,
-0.03716200217604637,
0.052429672330617905,
-0.03178425878286362,
-0.06951792538166046,
-0.05738508701324463,
-0.09031131118535995,
-0.030628742650151253,
-0.036129746586084366,
0.13929763436317444,
-0.021880285814404488,
-0.014240792021155357,
-0.07720693200826645,
-0.14127053320407867,
0.18778881430625916,
0.02091752178966999,
-0.004237720742821693,
-0.1613209992647171,
0.021601341664791107,
-0.13609236478805542,
-0.10934461653232574,
-0.016704287379980087,
-0.0537119135260582,
-0.11225147545337677,
0.08239202201366425,
-0.02741137146949768,
0.0503174290060997,
0.0029329785611480474,
0.14720657467842102,
-0.04524151608347893,
0.035695288330316544,
0.0461336225271225,
-0.010777432471513748,
-0.02755799889564514,
0.07401483505964279,
-0.12579509615898132,
-0.019331924617290497,
0.011030266061425209,
-0.11613353341817856,
0.01577788218855858,
-0.03908366709947586,
-0.13948652148246765,
-0.07092422991991043,
0.03738290071487427,
-0.0308510884642601,
-0.05904029309749603,
0.09481066465377808,
-0.020644094794988632,
-0.09894420951604843,
0.3524816930294037,
-0.00818937923759222,
0.07965563237667084,
0.09280315041542053,
-0.05663742870092392,
0.14428569376468658,
-0.04840005189180374,
0.01698233373463154,
0.033631402999162674,
0.037167906761169434,
-0.10928191989660263,
-0.07108522206544876,
-0.05777784809470177,
-0.09962747991085052,
0.06790393590927124,
-0.1795709729194641,
0.002895439276471734,
-0.15513037145137787,
-0.16947340965270996,
0.08206579834222794,
0.013418465852737427,
-0.07754019647836685,
-0.10086633265018463,
-0.025435298681259155,
-0.0512932650744915,
0.0005449238233268261,
-0.040090810507535934,
-0.18312327563762665,
-0.01444131601601839,
-0.1275091916322708,
-0.03905076906085014,
0.02420087531208992,
-0.13081307709217072,
0.010841581970453262,
-0.08009142428636551,
0.09666513651609421,
-0.12587134540081024,
-0.021499034017324448,
-0.10067395120859146,
0.06693798303604126,
-0.14473176002502441,
-0.0784997045993805,
-0.026866411790251732,
-0.07552085816860199,
0.05872194096446037,
0.14475587010383606,
-0.08986339718103409,
-0.028122445568442345,
-0.025684639811515808,
-0.1304430216550827,
-0.004223653580993414,
0.04902242496609688,
-0.023589691147208214,
-0.06226129084825516,
-0.016673773527145386,
0.09648003429174423,
0.14027345180511475,
0.006683229468762875,
-0.013802730478346348,
-0.001559535856358707,
-0.021564947441220284,
-0.12935499846935272,
0.024583298712968826,
0.07398906350135803,
-0.21222490072250366,
0.06620828807353973,
-0.03153783082962036,
0.032122135162353516,
-0.03512285277247429,
-0.012800252996385098,
-0.008842184208333492,
-0.015660902485251427,
0.04122309386730194,
-0.05087422579526901,
0.08729606866836548,
0.018864529207348824,
0.009317140094935894,
0.11817973107099533,
0.045763663947582245,
-0.013406718149781227,
0.042073484510183334,
-0.045026469975709915,
0.11554250866174698,
-0.0872509554028511,
-0.010915880091488361,
-0.104170061647892,
-0.19831274449825287,
0.10849201679229736,
-0.14496302604675293,
0.05372917652130127,
0.005071135237812996,
0.03516253083944321,
-0.0005857457872480154,
0.024219553917646408,
0.013668647967278957,
0.035281695425510406,
0.06959434598684311,
-0.013203050009906292,
-0.0725175067782402,
-0.07877416163682938,
-0.0671524852514267,
0.025077274069190025,
-0.20447877049446106,
0.0570642463862896,
0.13961753249168396,
0.052317604422569275,
-0.05681491643190384,
0.02646043337881565,
0.08817833662033081,
0.032802119851112366,
-0.03985697403550148,
0.05021775886416435,
0.047316938638687134,
0.04955228045582771,
-0.09757793694734573,
0.145896315574646,
-0.015902426093816757,
-0.12079472094774246,
0.017947230488061905,
0.08768980950117111,
-0.11219435185194016,
0.03714539483189583,
-0.03063557855784893,
-0.060677967965602875,
0.0760369673371315,
0.146461620926857,
0.23030562698841095,
0.035912103950977325,
0.10729829967021942,
-0.01496745366603136,
0.018314702436327934,
0.09884124249219894,
-0.05074106901884079,
-0.04462447389960289,
0.10905405879020691,
0.15365666151046753,
-0.02792956493794918,
0.05522140488028526,
-0.008813739754259586,
-0.14037902653217316,
0.03003675304353237,
0.09379846602678299,
-0.02108561061322689,
-0.01835591159760952,
-0.08007372170686722,
0.057631053030490875,
0.043816372752189636,
-0.08592838048934937,
0.008006539195775986,
0.05212601646780968,
0.03591401129961014,
0.02033975161612034,
-0.10628042370080948,
-0.00946846790611744,
0.06338655203580856,
0.041892196983098984,
0.02940540947020054,
0.055978477001190186,
-0.05315738543868065,
0.02483529783785343,
-0.014353875070810318,
-0.2762748599052429,
0.0008832390885800123,
0.04209081083536148,
-0.04311278834939003,
0.1814439743757248,
-0.01940566673874855,
-0.22433029115200043,
-0.18090112507343292,
-0.18848609924316406,
-0.22870860993862152,
-0.06268364191055298,
0.006729637272655964,
-0.0836026594042778,
-0.0919453576207161,
-0.0862848311662674,
-0.143731027841568,
-0.10098537057638168,
-0.008473072201013565,
-0.02595525048673153,
-0.11237248033285141,
-0.02116020768880844,
-0.06280360370874405,
0.07123667001724243,
-0.028781631961464882,
-0.07482564449310303,
0.04038241133093834,
-0.09099069237709045,
0.14474204182624817,
0.19324196875095367,
-0.048969924449920654,
0.006879524327814579,
0.09563831239938736,
0.16417580842971802,
0.014383723959326744,
0.08638133853673935,
0.02734452486038208,
-0.04255013167858124,
0.05716660991311073,
0.1322832852602005,
0.09902167320251465,
-0.006003084592521191,
-0.029485588893294334,
0.03281179815530777,
-0.062398750334978104,
-0.14226235449314117,
-0.11882056295871735,
-0.07551755756139755,
-0.03670107573270798,
0.03583512827754021,
0.0932830274105072,
0.1331232488155365,
0.09244266152381897,
0.1047682985663414,
-0.019711695611476898,
-0.1172478049993515,
0.14421723783016205,
0.05024704337120056,
0.05729905888438225,
0.0552230067551136,
-0.017807234078645706,
-0.0319703072309494,
0.0324861966073513,
0.06362643837928772,
0.1394285261631012,
0.11014983057975769,
0.12192095071077347,
0.03544853627681732,
0.15413548052310944,
0.2226042002439499,
0.09875956177711487,
-0.1464317888021469,
0.024421794340014458,
-0.0063840229995548725,
-0.01476768497377634,
0.1654629111289978,
-0.020791955292224884,
-0.10788964480161667,
0.017385175451636314,
0.006087517365813255,
-0.10587777942419052,
-0.04945632070302963,
0.0817384347319603,
0.07700443267822266,
-0.256815105676651,
-0.04650413990020752,
-0.07388637214899063,
0.031279392540454865,
-0.023973243311047554,
0.046334199607372284,
0.011212090030312538,
-0.01227999571710825,
0.10524161905050278,
0.00750342383980751,
0.10297038406133652,
0.08345409482717514,
-0.00008558992703910917,
-0.0029944165144115686,
0.07999174296855927,
0.0010234039509668946,
0.055272333323955536,
0.06614978611469269,
0.18908336758613586,
-0.008780206553637981,
-0.09975089132785797,
0.019223535433411598,
-0.06681974232196808,
-0.05813560634851456,
0.2585012912750244,
0.11277859658002853,
0.016729775816202164,
0.031627267599105835,
-0.02018752135336399,
-0.13557876646518707,
0.04000290110707283,
0.02095722407102585,
-0.025769775733351707,
0.05030403658747673,
0.10163144022226334,
-0.060210082679986954,
0.02828136645257473,
0.08102765679359436,
-0.06625525653362274,
-0.10510093718767166,
0.002030544448643923,
0.09787876158952713,
-0.002875476609915495,
-0.004174551460891962,
-0.033325739204883575,
-0.1129191592335701,
0.14514873921871185,
-0.010941362008452415,
0.06696612387895584,
-0.09252320975065231,
-0.18566562235355377,
0.15285395085811615,
-0.0992901474237442,
0.08628367632627487,
-0.10946537554264069,
0.06146196648478508,
0.034815698862075806,
-0.08175361901521683,
0.052268270403146744,
-0.033818915486335754,
-0.009148882701992989,
-0.015993136912584305,
-0.044673021882772446,
0.08585918694734573,
0.01712396740913391,
0.030427319929003716,
0.028792139142751694,
-0.04424772784113884,
-0.14527687430381775,
-0.17082878947257996,
0.01737898588180542,
0.044866226613521576,
0.04400871694087982,
-0.04125792533159256,
-0.04550414904952049,
-0.0025473693385720253,
0.0590326189994812,
0.13606108725070953,
0.033490944653749466,
-0.12638302147388458,
0.06999962776899338,
0.32979679107666016,
0.020566513761878014,
-0.19249576330184937,
-0.013265850022435188,
0.06295210868120193,
-0.00029567867750301957,
-0.14664238691329956,
-0.1150754764676094,
0.17842867970466614,
0.1448170393705368,
-0.033421147614717484,
-0.036965321749448776,
0.01656465046107769,
-0.044453151524066925,
0.13732682168483734,
-0.010792823508381844,
0.15490415692329407,
-0.05035579949617386,
-0.04308201000094414,
0.06829240173101425,
-0.1313418596982956,
0.19983471930027008,
0.03826028108596802,
0.0472685731947422,
-0.042134106159210205,
-0.004968202672898769,
0.022379552945494652,
-0.060708560049533844,
0.13440974056720734,
0.1490289270877838,
0.01558211725205183,
0.021877694875001907,
-0.10088210552930832,
0.08941220492124557,
-0.07426739484071732,
0.11557139456272125,
0.07391120493412018,
0.0741826742887497,
-0.1134292483329773,
-0.07504597306251526,
-0.033457159996032715,
0.05484107881784439,
0.02868897095322609,
0.003151918528601527,
-0.020508727058768272,
0.05456109717488289,
0.04623107984662056,
-0.0020905910059809685,
0.244338721036911,
-0.02203908935189247,
-0.021979084238409996,
0.22520150244235992,
0.007795632816851139,
-0.02610641159117222,
0.015996215865015984,
-0.05972186475992203,
-0.0419321246445179,
0.15492579340934753,
-0.05560680106282234,
-0.013139207847416401,
0.0742802619934082,
0.042630236595869064,
0.029709452763199806,
0.01794574409723282,
-0.019390858709812164,
-0.026086661964654922,
0.04519408196210861,
-0.15290957689285278,
-0.0881403386592865,
-0.06757918745279312,
-0.05875294655561447,
-0.03351186588406563,
0.15315119922161102,
0.020846083760261536,
-0.1672956794500351,
0.0202204417437315,
0.041985489428043365,
-0.027449283748865128,
-0.02772814966738224,
-0.026643233373761177,
0.005643839947879314,
0.04087342694401741,
-0.09079495072364807,
0.029906947165727615,
0.1376098245382309,
-0.010766222141683102,
-0.026863059028983116,
0.0701095461845398,
-0.1516682356595993,
-0.03261586278676987,
-0.07163142412900925,
0.17141488194465637,
-0.15247975289821625,
-0.06590243428945541,
-0.002415867755189538,
-0.06785690039396286,
0.03430604189634323,
0.21858952939510345,
0.055371325463056564,
-0.04585885629057884,
-0.10874046385288239,
-0.0668044164776802,
-0.1140807643532753,
-0.03056999109685421,
0.048200253397226334,
0.11315734684467316,
-0.1264745444059372,
-0.05933491885662079,
0.09495503455400467,
0.0679112896323204,
-0.04208357632160187,
-0.08787265419960022,
-0.1508154571056366,
-0.027709055691957474,
-0.17947298288345337,
0.054583802819252014,
-0.05480792000889778,
0.04542410373687744,
-0.042057156562805176,
-0.031077438965439796,
-0.04601123929023743,
-0.011377681978046894,
-0.052509699016809464,
0.043884243816137314,
-0.01352686807513237,
0.09320466965436935,
0.005718536209315062,
0.040034063160419464,
0.03064812161028385,
-0.028029127046465874,
0.022371001541614532,
0.061833057552576065,
-0.03823760524392128,
0.04668314382433891,
-0.1580805778503418,
-0.01769396662712097,
0.011000622063875198,
0.08091079443693161,
0.10871273279190063,
0.02363027073442936,
-0.00008447138679912314,
0.0038385901134461164,
-0.03161042928695679,
-0.09283122420310974,
-0.05585474148392677,
-0.08645689487457275,
-0.08528915792703629,
-0.060141921043395996,
-0.11002420634031296,
-0.07767754793167114,
-0.04325411841273308,
0.049629777669906616,
0.134361132979393,
0.07890800386667252,
0.024976028129458427,
0.07208441942930222,
0.038760531693696976,
-0.045194290578365326,
-0.013643646612763405,
-0.08465485274791718,
-0.06250950694084167,
-0.062591552734375,
0.014725103043019772,
-0.014986323192715645,
0.150737002491951,
-0.07082123309373856,
-0.04721999540925026,
-0.06720330566167831,
0.01608899049460888,
-0.0044603548012673855,
-0.02260289341211319,
0.19038543105125427,
0.12319029122591019,
0.01927514374256134,
0.038534827530384064,
0.06951585412025452,
-0.029075900092720985,
0.09037573635578156,
0.05678071454167366,
0.043425388634204865,
-0.2070557177066803,
0.06778737157583237,
0.12371062487363815,
-0.0273590087890625,
0.05523725971579552,
-0.005736070219427347,
-0.17287957668304443,
0.07884339988231659,
0.028389638289809227,
0.1412317007780075,
0.15556001663208008,
-0.0047406964004039764,
0.05472429841756821,
0.12730394303798676,
-0.08700279146432877,
-0.11734253168106079,
-0.29605230689048767,
-0.08300289511680603,
-0.09476065635681152,
0.011762529611587524,
-0.05492977797985077,
-0.025360632687807083,
0.07518624514341354,
0.06170244142413139,
0.056990575045347214,
0.09026072919368744,
0.01870608516037464,
0.09164175391197205,
0.020346952602267265,
0.0038824304938316345,
-0.035737767815589905,
0.0569143071770668,
0.04389292374253273,
-0.020261110737919807,
0.005099548492580652,
-0.010210114531219006,
0.022040409967303276,
0.14700709283351898,
0.03381533920764923,
0.04749048873782158,
-0.08962485939264297,
-0.033549949526786804,
-0.046185944229364395,
-0.05184585601091385,
0.04692189767956734,
0.04133811220526695,
-0.01255352795124054,
-0.01081941556185484,
-0.021265635266900063,
-0.034188076853752136,
0.0015658069169148803,
-0.020458156242966652,
0.29926443099975586,
0.026995215564966202,
0.009420214220881462,
0.05748005956411362,
-0.017368579283356667,
-0.09893328696489334,
0.2229585498571396,
0.13357260823249817,
-0.028109634295105934,
0.014176450669765472,
0.035654548555612564,
0.0047594825737178326,
0.016413556411862373,
0.06816580891609192,
0.0035045649856328964,
0.23892654478549957,
0.007355669979006052,
-0.06955362111330032,
-0.10695943981409073,
-0.01650063320994377,
-0.09652462601661682,
-0.008220906369388103,
0.10057258605957031,
-0.13457763195037842,
-0.14856290817260742,
0.051710546016693115,
-0.23665127158164978,
-0.01587158627808094,
0.042537931352853775,
0.058480944484472275,
-0.09229299426078796,
-0.03863649442791939,
-0.017020218074321747,
0.0020889306906610727,
0.02523707039654255,
-0.05601124465465546,
0.026427602395415306,
0.034208767116069794,
0.02163274586200714,
-0.08796228468418121,
-0.045942433178424835,
0.115231953561306,
0.1375461220741272,
0.1398613005876541,
0.03999631479382515,
-0.012331875041127205,
0.03788674622774124,
0.05617082864046097,
-0.14915959537029266,
0.07152508199214935,
-0.0017424849793314934,
-0.11887053400278091,
-0.05440588295459747,
-0.009779517538845539,
-0.029525049030780792,
-0.02327602542936802,
-0.07597402483224869,
-0.13780422508716583,
0.044316940009593964,
0.07142841815948486,
0.015319714322686195,
-0.02246369607746601,
0.043814852833747864,
0.028538979589939117,
0.09589656442403793,
0.15342313051223755,
-0.015491630882024765,
-0.09868676215410233,
-0.05329442769289017,
-0.011209186166524887,
0.005977707449346781,
-0.03644721955060959,
0.03505965694785118,
-0.07836025953292847,
-0.010563082993030548,
-0.11734850704669952,
0.09071695059537888,
-0.15666712820529938,
-0.028867915272712708,
0.019595175981521606,
0.048655200749635696,
0.007279754616320133,
-0.020533209666609764,
0.06519631296396255,
0.056114450097084045,
-0.01895919255912304,
-0.03386515751481056,
0.0011861514067277312,
0.04531729593873024,
-0.14098328351974487,
-0.1133762001991272
] |
null | null |
transformers
|
## Cyclone SIMCSE RoBERTa WWM Ext Chinese
This model provides simplified Chinese sentence embeddings encoding based on [Simple Contrastive Learning](https://arxiv.org/abs/2104.08821).
The pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.
### Usage
Please use [SentenceTransformer](https://github.com/UKPLab/sentence-transformers) to load the model.
from sentence_transformers import SentenceTransformer
encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')
|
{}
|
feature-extraction
|
cyclone/simcse-chinese-roberta-wwm-ext
|
[
"transformers",
"pytorch",
"bert",
"feature-extraction",
"arxiv:2104.08821",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2104.08821"
] |
[] |
TAGS
#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #has_space #region-us
|
## Cyclone SIMCSE RoBERTa WWM Ext Chinese
This model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.
The pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.
### Usage
Please use SentenceTransformer to load the model.
from sentence_transformers import SentenceTransformer
encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')
|
[
"## Cyclone SIMCSE RoBERTa WWM Ext Chinese\r\n\r\nThis model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.\r\nThe pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.",
"### Usage\r\nPlease use SentenceTransformer to load the model.\r\n\r\n from sentence_transformers import SentenceTransformer\r\n \r\n encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')"
] |
[
"TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #has_space #region-us \n",
"## Cyclone SIMCSE RoBERTa WWM Ext Chinese\r\n\r\nThis model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.\r\nThe pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.",
"### Usage\r\nPlease use SentenceTransformer to load the model.\r\n\r\n from sentence_transformers import SentenceTransformer\r\n \r\n encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')"
] |
[
42,
63,
54
] |
[
"passage: TAGS\n#transformers #pytorch #bert #feature-extraction #arxiv-2104.08821 #endpoints_compatible #has_space #region-us \n## Cyclone SIMCSE RoBERTa WWM Ext Chinese\r\n\r\nThis model provides simplified Chinese sentence embeddings encoding based on Simple Contrastive Learning.\r\nThe pretrained model(Chinese RoBERTa WWM Ext) is used for token encoding.### Usage\r\nPlease use SentenceTransformer to load the model.\r\n\r\n from sentence_transformers import SentenceTransformer\r\n \r\n encoder = SentenceTransformer('cyclone/simcse-chinese-roberta-wwm-ext')"
] |
[
-0.028554080054163933,
-0.05265853926539421,
-0.0018137121805921197,
0.03653391823172569,
0.12859664857387543,
-0.018765108659863472,
0.004289775621145964,
-0.011611961759626865,
-0.0632324144244194,
0.012004598043859005,
0.10620573908090591,
-0.0015284570399671793,
0.046450722962617874,
-0.07722622901201248,
-0.025195837020874023,
-0.336261510848999,
0.09935060888528824,
0.05692271888256073,
-0.029211755841970444,
0.12336000055074692,
0.11886604875326157,
-0.07289236783981323,
0.11483168601989746,
0.0844116061925888,
-0.07990805804729462,
0.10707252472639084,
-0.0997481644153595,
-0.08387737721204758,
0.03641323372721672,
0.011655351147055626,
0.13959354162216187,
0.042524319142103195,
0.02819414995610714,
-0.08177418261766434,
-0.013230511918663979,
-0.05726635456085205,
-0.07030005753040314,
-0.007794570177793503,
0.007728426717221737,
-0.04961279407143593,
0.12478499859571457,
-0.06071997806429863,
0.03727329149842262,
0.03151329234242439,
-0.09592186659574509,
-0.05395086482167244,
0.01061064563691616,
0.018184617161750793,
0.18632420897483826,
0.1643422394990921,
-0.027103623375296593,
0.07203170657157898,
-0.1276538223028183,
0.07829044759273529,
0.0941571369767189,
-0.2709347903728485,
0.01792575605213642,
0.06402808427810669,
0.14312276244163513,
-0.056575559079647064,
-0.06324461102485657,
0.0003469485091045499,
0.06056957319378853,
0.047981638461351395,
-0.03507796302437782,
-0.10393670201301575,
-0.04862479120492935,
0.06670326739549637,
-0.14761702716350555,
0.08613467961549759,
0.2490658462047577,
-0.0012308473233133554,
-0.07120515406131744,
-0.01768689602613449,
-0.014781669713556767,
-0.09888111054897308,
-0.01838277094066143,
-0.09485369920730591,
-0.007439131382852793,
0.05624143406748772,
0.04316550865769386,
0.0028199716471135616,
-0.09182246029376984,
-0.03975829482078552,
-0.17681579291820526,
0.2278364896774292,
-0.016466546803712845,
0.012392746284604073,
-0.20823612809181213,
0.04418637976050377,
0.03664528951048851,
-0.1414201259613037,
-0.038434166461229324,
-0.047473516315221786,
-0.056076791137456894,
0.1010257825255394,
-0.08074494451284409,
-0.1453487128019333,
-0.06695928424596786,
0.08865798264741898,
-0.0631721094250679,
0.05931340903043747,
0.14814415574073792,
0.03158188983798027,
0.028389565646648407,
0.2059755176305771,
-0.057035382837057114,
-0.10625506937503815,
-0.05387764051556587,
-0.09393737465143204,
-0.022329088300466537,
-0.059798695147037506,
-0.18364140391349792,
-0.022669104859232903,
-0.0778726190328598,
0.02347579225897789,
-0.059100788086652756,
0.051952484995126724,
-0.07081833481788635,
-0.08205163478851318,
0.24861067533493042,
-0.07975571602582932,
0.03486604616045952,
-0.006547190714627504,
-0.015851693227887154,
0.20657138526439667,
0.014810599386692047,
0.025620901957154274,
-0.0060340105555951595,
0.0047826748341321945,
-0.07942261546850204,
0.0014673139667138457,
-0.03708310052752495,
-0.1331237107515335,
-0.05597119778394699,
-0.047285642474889755,
0.05817592889070511,
-0.1441667079925537,
-0.1476108282804489,
0.02506878413259983,
-0.05086884647607803,
-0.027116307988762856,
-0.029271485283970833,
-0.10783398151397705,
-0.07893465459346771,
0.043636124581098557,
-0.016029750928282738,
-0.1429409235715866,
0.02519022673368454,
-0.06644728034734726,
-0.02345479279756546,
0.06210560351610184,
-0.1985948532819748,
0.014726425521075726,
-0.15611398220062256,
0.04860301315784454,
-0.02490077167749405,
0.08104516565799713,
-0.024987520650029182,
0.09768649935722351,
-0.08799136430025101,
-0.06805640459060669,
-0.022040506824851036,
0.03811465948820114,
0.007087404374033213,
0.16585247218608856,
-0.23564167320728302,
-0.03096463903784752,
0.07353729009628296,
-0.1546444445848465,
-0.02583317644894123,
0.07294319570064545,
-0.018440159037709236,
0.06416332721710205,
0.05775143951177597,
0.06766245514154434,
-0.011863128282129765,
0.015588336624205112,
0.003053799271583557,
0.10569481551647186,
-0.09212550520896912,
0.07409712672233582,
0.0882255882024765,
0.052207693457603455,
-0.08877768367528915,
0.01578286848962307,
0.01281706802546978,
0.04446591064333916,
0.011585998348891735,
-0.0418853722512722,
0.008697301149368286,
-0.02007039450109005,
0.07072444260120392,
-0.034251514822244644,
0.05878858268260956,
-0.03143363073468208,
-0.07551507651805878,
0.07463924586772919,
0.08876726776361465,
-0.05019879713654518,
0.07194610685110092,
-0.14354026317596436,
-0.027039041742682457,
0.012370524927973747,
-0.020934708416461945,
-0.11880430579185486,
0.0626557469367981,
0.002972130896523595,
0.09432724118232727,
0.013447299599647522,
0.04685625061392784,
0.0577484667301178,
-0.016605990007519722,
0.03702881932258606,
-0.048437654972076416,
0.060705721378326416,
0.07131627202033997,
-0.07351797819137573,
-0.03576355427503586,
-0.04745744913816452,
-0.0696883350610733,
0.03193676471710205,
-0.08218434453010559,
0.004894951358437538,
0.1435016393661499,
-0.029737655073404312,
-0.029690783470869064,
0.03174235671758652,
-0.012122434563934803,
0.04217135161161423,
-0.06172099709510803,
0.027768082916736603,
0.006299990229308605,
0.02106938511133194,
-0.2046656459569931,
0.2743980288505554,
-0.15841926634311676,
-0.1504749059677124,
0.08902309834957123,
-0.09196034073829651,
-0.10030601918697357,
0.05243365466594696,
0.028545653447508812,
0.0023706667125225067,
0.05711603909730911,
-0.01004449836909771,
0.30580443143844604,
-0.0616888701915741,
0.14076317846775055,
-0.04595628380775452,
0.03153489530086517,
-0.0014744950458407402,
-0.03238334879279137,
-0.009770937263965607,
0.15129221975803375,
-0.01875651627779007,
-0.18552260100841522,
0.03336269035935402,
0.15125030279159546,
-0.037926916033029556,
0.03542780876159668,
0.045411963015794754,
-0.011871443130075932,
-0.07054996490478516,
-0.04961884394288063,
-0.000028735888918163255,
0.03191017359495163,
-0.11057586222887039,
-0.019851673394441605,
0.028595618903636932,
0.029796605929732323,
0.04230821877717972,
-0.1080506443977356,
-0.05131703242659569,
0.13623537123203278,
0.056160785257816315,
0.006220954470336437,
0.026697786524891853,
-0.057960379868745804,
0.0043800813145935535,
-0.06712402403354645,
-0.11862947791814804,
-0.010847237892448902,
0.016649411991238594,
-0.09160031378269196,
0.23083670437335968,
-0.106538325548172,
-0.2970646321773529,
-0.12933214008808136,
-0.127070352435112,
0.012471036985516548,
-0.01653752289712429,
0.08066755533218384,
-0.1580963432788849,
-0.10652310401201248,
-0.12501975893974304,
-0.009687747806310654,
-0.1872607171535492,
0.001706346869468689,
0.03904769569635391,
-0.060417115688323975,
-0.059557538479566574,
-0.0870155543088913,
0.00479641230776906,
0.005613026209175587,
-0.08638027310371399,
0.0358906053006649,
-0.06581546366214752,
0.09599330276250839,
0.18519079685211182,
-0.0006188054103404284,
0.01355703640729189,
0.017238087952136993,
0.2193848192691803,
-0.018834659829735756,
0.03041720949113369,
0.08937172591686249,
-0.11138813942670822,
0.028958017006516457,
0.003573603928089142,
0.023446472361683846,
-0.06837836652994156,
-0.004410624969750643,
-0.05958646163344383,
-0.058587461709976196,
-0.1452970802783966,
-0.11607512831687927,
-0.07276449352502823,
0.018615519627928734,
-0.02500985562801361,
0.03873768448829651,
0.09215667843818665,
0.10144171118736267,
0.0688202828168869,
0.004301843233406544,
-0.09094186127185822,
0.06628037244081497,
-0.02529226616024971,
-0.01595231704413891,
0.10607825219631195,
-0.06981340795755386,
-0.09109243005514145,
0.02810552343726158,
-0.030084017664194107,
0.08024221658706665,
0.10857710242271423,
0.14462095499038696,
0.030204184353351593,
0.06946121901273727,
0.13509753346443176,
0.10931218415498734,
-0.1279791295528412,
-0.05241833254694939,
-0.037718575447797775,
-0.020788423717021942,
0.11157722026109695,
0.09828516840934753,
0.03357618674635887,
0.023432152345776558,
-0.08205505460500717,
0.027725564315915108,
0.01924886181950569,
-0.050149571150541306,
0.08973336964845657,
-0.14888058602809906,
-0.03232160955667496,
0.05593320354819298,
-0.005431205965578556,
-0.010573669336736202,
0.10648512840270996,
0.06792301684617996,
-0.04916700720787048,
0.08757100999355316,
0.014724217355251312,
0.11016914248466492,
0.12237569689750671,
0.004065375775098801,
-0.07963571697473526,
0.022781288251280785,
0.03808610886335373,
0.07327651977539062,
-0.11530778557062149,
0.11667045205831528,
0.012852412648499012,
-0.05921176075935364,
-0.04763185232877731,
-0.00004634155266103335,
-0.03576110303401947,
0.17809416353702545,
0.18168413639068604,
0.010568063706159592,
-0.12784844636917114,
-0.02469017170369625,
-0.06744131445884705,
0.013821878470480442,
0.12990069389343262,
-0.026604261249303818,
0.05219230428338051,
0.029933864250779152,
0.007591355126351118,
0.03351060673594475,
0.09711827337741852,
0.01522031705826521,
-0.1851930320262909,
0.009077007882297039,
0.056083496659994125,
0.012388921342790127,
0.007132034283131361,
0.020573195070028305,
-0.08857641369104385,
0.10833170264959335,
-0.09139471501111984,
0.008584904484450817,
-0.1025598794221878,
-0.11149031668901443,
0.133989617228508,
-0.09990469366312027,
0.03184884414076805,
0.021654248237609863,
-0.018279055133461952,
-0.02584991417825222,
-0.06818464398384094,
0.11018829792737961,
-0.05169561132788658,
0.014567111618816853,
-0.005228639114648104,
-0.017514174804091454,
0.07076042890548706,
0.06108803302049637,
0.051532745361328125,
0.044428735971450806,
-0.10162658244371414,
-0.1666816771030426,
-0.12140217423439026,
-0.087103933095932,
-0.046067334711551666,
0.12329791486263275,
0.020518776029348373,
-0.013722213916480541,
-0.05480894073843956,
0.0029430733993649483,
0.2777245342731476,
0.08762418478727341,
-0.11529836803674698,
0.12056899070739746,
0.09672082960605621,
-0.024853000417351723,
-0.27188441157341003,
-0.08189088851213455,
-0.027901871129870415,
0.09741096198558807,
0.05826409161090851,
-0.023992864415049553,
0.1111774817109108,
-0.01057279109954834,
0.010985243134200573,
-0.07081631571054459,
-0.198762446641922,
-0.0897345170378685,
0.13268648087978363,
-0.05525638535618782,
0.1980586051940918,
-0.11382845789194107,
-0.03600718826055527,
0.005016077309846878,
-0.127798393368721,
0.1573350876569748,
0.02642645128071308,
0.0704907700419426,
0.019485624507069588,
-0.04403137415647507,
0.021316412836313248,
0.006437281612306833,
0.19769945740699768,
0.07388767600059509,
0.02109670266509056,
-0.004503563046455383,
-0.10134641826152802,
0.13139723241329193,
0.03188442438840866,
0.10773658752441406,
-0.1030254140496254,
0.04505109786987305,
-0.2481520175933838,
-0.056376587599515915,
0.0013444715877994895,
0.008372259326279163,
0.015007957816123962,
-0.008538810536265373,
-0.05351082608103752,
0.018764007836580276,
0.08724860101938248,
-0.01476339902728796,
0.19827066361904144,
-0.03862372785806656,
-0.09327656775712967,
0.15161004662513733,
0.13713997602462769,
-0.007364931516349316,
-0.05112944170832634,
-0.022585105150938034,
-0.03206939995288849,
0.11950196325778961,
-0.13194715976715088,
0.005323527380824089,
0.0015040163416415453,
0.020258383825421333,
0.09056442975997925,
0.06383555382490158,
0.0011089193867519498,
0.06282540410757065,
0.059457313269376755,
-0.1051386147737503,
-0.039845313876867294,
-0.12983925640583038,
0.028257964178919792,
0.06808996945619583,
0.0597837008535862,
0.05755084753036499,
-0.12000826001167297,
0.009038764983415604,
-0.03681228309869766,
0.039911966770887375,
-0.10567508637905121,
0.029260458424687386,
0.014384759590029716,
0.03007824905216694,
-0.1201416552066803,
0.035273630172014236,
0.0661335214972496,
-0.14891234040260315,
0.07775920629501343,
0.10447096079587936,
-0.06324685364961624,
-0.05790479853749275,
-0.1499793529510498,
0.09509885311126709,
-0.1722574234008789,
-0.049012184143066406,
-0.03530192747712135,
-0.1585608869791031,
-0.0020063663832843304,
0.06785351037979126,
0.10528763383626938,
0.0016525924438610673,
-0.11814191937446594,
-0.03947191312909126,
-0.07613006979227066,
-0.03451929986476898,
0.11910589039325714,
0.029985418543219566,
-0.06537564098834991,
0.09379979968070984,
0.09637060761451721,
0.12572512030601501,
-0.05392558500170708,
-0.06278687715530396,
-0.05255541205406189,
0.014549506828188896,
-0.10639381408691406,
0.017303315922617912,
-0.0836903303861618,
-0.027040237560868263,
0.0019835636485368013,
-0.042488500475883484,
-0.042058251798152924,
0.006421484984457493,
-0.05836177244782448,
0.02304888516664505,
-0.03458524867892265,
0.02109677530825138,
0.07761142402887344,
-0.0036607992369681597,
0.048368457704782486,
-0.035062480717897415,
0.00538516603410244,
0.08575626462697983,
-0.09509007632732391,
0.047820672392845154,
-0.1352050006389618,
-0.020805705338716507,
0.05103360861539841,
0.03457112982869148,
0.08598483353853226,
0.044723328202962875,
-0.02341841347515583,
0.024430986493825912,
-0.0029801891651004553,
-0.058883920311927795,
-0.0014586274046450853,
-0.04521581530570984,
-0.05211729556322098,
-0.10447581112384796,
-0.08270181715488434,
-0.07622569799423218,
-0.04873146116733551,
0.04442548006772995,
0.03336840122938156,
0.15338212251663208,
0.003795547178015113,
0.127537339925766,
-0.036449532955884933,
-0.02483913116157055,
-0.03406333178281784,
-0.04557514563202858,
-0.05321895331144333,
-0.08960004150867462,
0.0360647477209568,
-0.03977464511990547,
0.1665407121181488,
-0.09224707633256912,
-0.024264980107545853,
-0.02254582569003105,
0.04323096200823784,
0.10038851201534271,
0.03284026309847832,
0.26751598715782166,
0.09816057980060577,
0.01994429901242256,
-0.12759289145469666,
0.04874265566468239,
0.037268657237291336,
0.10797896236181259,
-0.009253021329641342,
0.03444815054535866,
-0.09033303707838058,
0.22483383119106293,
-0.010294448584318161,
0.034903720021247864,
0.0424368716776371,
-0.05996602401137352,
-0.10859231650829315,
0.021566815674304962,
0.0268521960824728,
0.06506744027137756,
0.16005998849868774,
0.00408919295296073,
0.09532929956912994,
0.10713735222816467,
-0.09572161734104156,
-0.1395399570465088,
-0.10632681846618652,
-0.06639260053634644,
-0.10496421903371811,
-0.0034818986896425486,
-0.12106571346521378,
0.04064825549721718,
0.03073231875896454,
0.02456234022974968,
0.03268531337380409,
0.1320710927248001,
-0.021972794085741043,
-0.01482109073549509,
0.07792782783508301,
-0.03162577003240585,
0.06200997158885002,
0.15181033313274384,
0.053919825702905655,
0.031089046970009804,
-0.017142940312623978,
0.02317793481051922,
0.017793014645576477,
0.02024920843541622,
0.031181026250123978,
-0.03685006871819496,
-0.11704324930906296,
-0.041150324046611786,
0.03336191922426224,
-0.030050378292798996,
0.11827144026756287,
0.0618823766708374,
0.007910322397947311,
-0.030415041372179985,
0.002732502529397607,
-0.07113370299339294,
-0.07490155100822449,
-0.08051985502243042,
0.16381649672985077,
0.11241519451141357,
0.11187484860420227,
0.026086006313562393,
-0.046532705426216125,
-0.04802589863538742,
0.24728617072105408,
0.13325995206832886,
-0.10686445981264114,
0.017686622217297554,
0.04459148645401001,
0.023245971649885178,
0.06552821397781372,
0.06002194806933403,
0.017792677506804466,
0.3122994601726532,
-0.054981350898742676,
-0.061166033148765564,
-0.1696951538324356,
-0.05731309577822685,
-0.05504193902015686,
0.03212856501340866,
0.09968579560518265,
-0.10518928617238998,
-0.023353353142738342,
0.08642531931400299,
-0.16564400494098663,
0.07223304361104965,
0.033861156553030014,
-0.12324117869138718,
-0.04159051179885864,
0.009412085637450218,
0.15891246497631073,
0.1234574019908905,
0.07621511816978455,
-0.0031577274203300476,
0.010425662621855736,
0.06940478086471558,
0.03653779998421669,
-0.14718684554100037,
0.1104818657040596,
0.07711056619882584,
-0.1248268112540245,
-0.024740777909755707,
0.007425549440085888,
0.04336364194750786,
0.042017970234155655,
0.10589101910591125,
-0.021945960819721222,
0.1228252574801445,
-0.03262008726596832,
-0.021187057718634605,
0.08536650985479355,
0.06643002480268478,
-0.004966977518051863,
-0.019930534064769745,
-0.018473999574780464,
-0.17470063269138336,
0.11451516300439835,
0.11301539838314056,
-0.04108269140124321,
-0.012447907589375973,
-0.01188506931066513,
-0.0040296283550560474,
0.10312604159116745,
0.1407899707555771,
-0.026900891214609146,
-0.03061670809984207,
0.0012754366034641862,
-0.004361702594906092,
-0.04622119665145874,
-0.05718992277979851,
-0.049828317016363144,
-0.09782666712999344,
-0.03861868381500244,
0.013082960620522499,
0.06213219463825226,
-0.1631295531988144,
-0.02962542697787285,
-0.058445531874895096,
-0.013389287516474724,
-0.009974230080842972,
0.02627825178205967,
0.12528014183044434,
0.010740464553236961,
-0.02097354643046856,
0.04808202385902405,
0.07183939218521118,
0.03521266579627991,
-0.12103113532066345,
-0.13070207834243774
] |
null | null |
transformers
|
# About
This is a sample repo.
|
{}
|
fill-mask
|
cylee/tutorial
|
[
"transformers",
"tf",
"bert",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us
|
# About
This is a sample repo.
|
[
"# About\n\nThis is a sample repo."
] |
[
"TAGS\n#transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"# About\n\nThis is a sample repo."
] |
[
35,
8
] |
[
"passage: TAGS\n#transformers #tf #bert #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# About\n\nThis is a sample repo."
] |
[
-0.020638464018702507,
-0.04526073858141899,
-0.004770228173583746,
0.0014838868519291282,
0.10005791485309601,
0.0338241346180439,
0.09835878014564514,
0.08801118284463882,
0.029648128896951675,
-0.03379072993993759,
0.1531848907470703,
0.13872165977954865,
-0.028896372765302658,
0.21480605006217957,
-0.06427262723445892,
-0.2564944922924042,
0.05779002979397774,
0.04405888542532921,
-0.10111920535564423,
0.0830889493227005,
0.07054884731769562,
-0.064909428358078,
0.08429653942584991,
-0.036377567797899246,
-0.12653031945228577,
0.08280526101589203,
0.07343019545078278,
-0.10839973390102386,
0.134135901927948,
0.10209453850984573,
0.20319393277168274,
0.04592067003250122,
-0.03285146504640579,
-0.06954608112573624,
0.05843556672334671,
0.01912878453731537,
-0.06747812777757645,
0.04090375453233719,
0.020071323961019516,
-0.010681383311748505,
0.056457579135894775,
0.1526065319776535,
0.017002910375595093,
0.08558686077594757,
-0.15771658718585968,
-0.11301999539136887,
-0.02559477649629116,
-0.02004312165081501,
0.018359532579779625,
0.04852477088570595,
0.0020172931253910065,
0.26829463243484497,
-0.15308307111263275,
0.10696929693222046,
0.20043031871318817,
-0.23793333768844604,
-0.0028316196985542774,
0.0904601514339447,
0.07612587511539459,
-0.08559839427471161,
-0.02544979378581047,
0.04741762951016426,
0.09985996037721634,
0.03825804591178894,
-0.007233008276671171,
-0.07028944790363312,
-0.12097861617803574,
0.004248871933668852,
-0.08399491757154465,
-0.03823775425553322,
0.21030612289905548,
0.028088543564081192,
-0.004382567014545202,
0.0464450865983963,
-0.09857963770627975,
0.002242234768345952,
0.0037664391566067934,
-0.04362202808260918,
-0.07796148210763931,
0.04794059321284294,
-0.09586690366268158,
-0.0039565712213516235,
-0.08636413514614105,
-0.03661603853106499,
-0.2586638629436493,
0.22413961589336395,
-0.012860853224992752,
0.06729239225387573,
-0.15108740329742432,
0.053095199167728424,
-0.053203754127025604,
-0.1053212583065033,
0.018302472308278084,
-0.0981588363647461,
-0.025783687829971313,
-0.06239490211009979,
-0.09151691198348999,
-0.12490551173686981,
0.11915712803602219,
0.24423404037952423,
0.0712323933839798,
0.00542096234858036,
-0.08940641582012177,
0.03892061114311218,
0.03064960613846779,
0.03359363600611687,
-0.014226265251636505,
-0.004457438830286264,
0.0351538360118866,
-0.13886019587516785,
-0.04682942107319832,
-0.0552871972322464,
-0.12201213091611862,
-0.009864922612905502,
0.0020369342528283596,
0.0680900439620018,
0.06029688939452171,
0.0766008049249649,
-0.05104264244437218,
-0.019329437986016273,
0.05552689731121063,
-0.078093983232975,
-0.047489702701568604,
-0.031451862305402756,
0.045884061604738235,
0.05736669525504112,
0.04202933982014656,
0.014117380604147911,
0.027191750705242157,
0.02204611524939537,
-0.07637125253677368,
-0.09899304062128067,
-0.05424252897500992,
-0.09563472121953964,
0.03187945485115051,
-0.062764473259449,
0.05283556133508682,
-0.2242327779531479,
-0.10372186452150345,
0.04945734143257141,
0.04565800353884697,
-0.01933915726840496,
-0.011303355917334557,
0.018657684326171875,
-0.011733991093933582,
0.029940424486994743,
-0.029993772506713867,
0.05706217885017395,
-0.07331940531730652,
0.09153001010417938,
-0.012133529409766197,
0.1242019385099411,
-0.10899204015731812,
0.019415685907006264,
-0.06786307692527771,
0.0035835690796375275,
-0.16919364035129547,
-0.02000764198601246,
-0.04618607088923454,
0.1735881268978119,
-0.06334850937128067,
-0.016138944774866104,
-0.1644633263349533,
0.031540412455797195,
-0.030858565121889114,
0.171010360121727,
-0.1372336745262146,
-0.10122884809970856,
0.21048006415367126,
-0.12789253890514374,
-0.11565631628036499,
0.09264744818210602,
-0.01495369989424944,
0.05486151948571205,
0.03852052986621857,
0.08585153520107269,
0.0543067567050457,
-0.13131888210773468,
0.08173246681690216,
0.1321915090084076,
-0.12771795690059662,
-0.03297298029065132,
0.005200146231800318,
-0.0017359136836603284,
-0.1657947599887848,
0.029110342264175415,
0.10621645301580429,
0.14434485137462616,
-0.09255307912826538,
-0.022239647805690765,
-0.020990025252103806,
-0.06182858720421791,
0.08548519760370255,
0.03181752189993858,
0.09763478487730026,
-0.07307291030883789,
-0.04520689696073532,
-0.015608086250722408,
0.015269991010427475,
0.05225202441215515,
0.00806522835046053,
-0.13513454794883728,
0.11748164892196655,
-0.052085794508457184,
0.0019670245237648487,
-0.17492525279521942,
-0.12753723561763763,
0.0019191503524780273,
0.10624560713768005,
-0.01465243473649025,
0.23093606531620026,
0.11888088285923004,
-0.06675355136394501,
-0.015736235305666924,
0.023919323459267616,
0.08064392954111099,
0.04809902235865593,
-0.0014421007363125682,
-0.11875361949205399,
0.02627851814031601,
-0.10671203583478928,
-0.06991101056337357,
-0.004614669363945723,
-0.04211418330669403,
0.03621794655919075,
0.21150550246238708,
0.05531386286020279,
0.035803135484457016,
-0.02918100357055664,
-0.0023681097663939,
-0.05380008742213249,
-0.03105071745812893,
0.022643905133008957,
0.010222511366009712,
-0.053125474601984024,
0.14990423619747162,
-0.06784448772668839,
0.28559738397598267,
0.18189503252506256,
-0.22580625116825104,
-0.08306197077035904,
0.040265511721372604,
-0.036275144666433334,
0.006764399353414774,
0.026825202628970146,
-0.04138622805476189,
0.06762538105249405,
-0.020113732665777206,
0.11926358938217163,
-0.0045202868059277534,
-0.010695114731788635,
0.028493741527199745,
-0.038538459688425064,
-0.04043921083211899,
0.022380029782652855,
0.11302640289068222,
-0.1906518191099167,
0.13263672590255737,
0.20236220955848694,
0.05764956772327423,
0.1346011757850647,
-0.040712181478738785,
-0.06637027114629745,
0.018114572390913963,
-0.0322340652346611,
0.003096441039815545,
0.06543438136577606,
-0.12667813897132874,
-0.0026154578663408756,
0.07583598047494888,
-0.03307846933603287,
0.04512356221675873,
-0.06946029514074326,
-0.024867791682481766,
0.03357677161693573,
0.0174387339502573,
-0.07151643931865692,
0.11887408047914505,
0.011668744497001171,
0.07698667794466019,
0.020764408633112907,
-0.0873173251748085,
0.08963430672883987,
0.01864396035671234,
-0.08302129805088043,
0.1729007214307785,
-0.11461217701435089,
-0.2905164957046509,
-0.09630388021469116,
-0.1613900363445282,
0.044441163539886475,
0.06808650493621826,
0.05368661880493164,
-0.11668174713850021,
-0.1049017533659935,
0.053734190762043,
0.05478675663471222,
0.015502508729696274,
0.10014822334051132,
-0.038901135325431824,
0.02781647816300392,
0.0025857798755168915,
-0.06986098736524582,
-0.034762393683195114,
0.01434298511594534,
-0.029876770451664925,
0.14037871360778809,
-0.17076954245567322,
0.0825507864356041,
0.1335289478302002,
0.03020785003900528,
0.07823547720909119,
-0.02060268074274063,
0.2283136546611786,
-0.0937584713101387,
0.007528198417276144,
0.14289237558841705,
-0.05993090197443962,
0.012894549407064915,
0.16230887174606323,
-0.0005658446461893618,
-0.0620291493833065,
0.07352832704782486,
-0.008656895719468594,
-0.12492774426937103,
-0.13076554238796234,
-0.06121625751256943,
-0.09274028241634369,
-0.01192986499518156,
0.0684160515666008,
0.03786822035908699,
0.1650833785533905,
0.06022380292415619,
0.04315483197569847,
0.026021292433142662,
-0.10334702581167221,
0.012695228680968285,
0.08549732714891434,
-0.057250864803791046,
0.12792626023292542,
-0.06719677150249481,
-0.15088196098804474,
0.10027607530355453,
-0.0635102167725563,
0.13224521279335022,
0.11084934324026108,
-0.008262407034635544,
0.02277718484401703,
0.10222909599542618,
0.13635724782943726,
0.17997793853282928,
0.05685535818338394,
-0.09208337217569351,
-0.026208745315670967,
-0.0179587509483099,
-0.03988543152809143,
0.06694933772087097,
0.11399666965007782,
-0.11942823976278305,
-0.050956618040800095,
-0.10861597955226898,
0.030256908386945724,
0.15666210651397705,
0.11109250783920288,
-0.26186686754226685,
-0.019157709553837776,
0.008513352833688259,
-0.04765915870666504,
-0.0645303726196289,
0.048812881112098694,
0.02741658128798008,
-0.09513694792985916,
0.033470772206783295,
-0.025694791227579117,
0.0964333638548851,
0.10321322828531265,
0.054411448538303375,
-0.008284061215817928,
-0.0842454805970192,
-0.03354126587510109,
0.0566696934401989,
-0.2782317101955414,
0.27515000104904175,
-0.01016558799892664,
-0.04593979939818382,
-0.09881193190813065,
-0.016508307307958603,
0.06206893548369408,
0.1079106256365776,
0.09999005496501923,
-0.006494046654552221,
-0.14250190556049347,
-0.15325872600078583,
-0.013277361169457436,
0.06332948058843613,
0.10027562826871872,
-0.00243855407461524,
0.018110990524291992,
-0.04434330016374588,
-0.018266664817929268,
0.04309516400098801,
0.14265723526477814,
-0.05457423999905586,
-0.045975372195243835,
0.0641486644744873,
0.004787772428244352,
0.03683453053236008,
-0.04889858886599541,
-0.06182840093970299,
-0.04244852066040039,
0.15385058522224426,
0.002186993369832635,
-0.007641430012881756,
-0.1446448415517807,
0.010328131727874279,
0.13259302079677582,
-0.06688529253005981,
0.12302999198436737,
-0.030522217974066734,
0.028021007776260376,
-0.03566089645028114,
-0.21988730132579803,
0.1290862113237381,
-0.1446457952260971,
-0.015765009447932243,
-0.060969993472099304,
0.06644421070814133,
-0.050436556339263916,
0.05359421297907829,
0.037106774747371674,
0.0032880280632525682,
-0.02391975000500679,
-0.02058585360646248,
-0.002986230654641986,
-0.12060830742120743,
0.04130121320486069,
0.03277308866381645,
-0.030752582475543022,
-0.03232359513640404,
0.012583254836499691,
0.009571905247867107,
0.18893033266067505,
0.18657660484313965,
-0.07027825713157654,
0.1039186343550682,
0.07866182923316956,
-0.009840846061706543,
-0.3097914159297943,
-0.06118353456258774,
-0.07759015262126923,
0.01277625560760498,
0.010034519247710705,
-0.12072498351335526,
0.13402318954467773,
-0.048304833471775055,
-0.01080978661775589,
0.11147766560316086,
-0.11387280374765396,
-0.11328095942735672,
0.29053398966789246,
0.010719874873757362,
0.38116803765296936,
-0.10856204479932785,
-0.06997020542621613,
-0.03777197375893593,
-0.13801439106464386,
0.0843663364648819,
-0.020650748163461685,
0.05087010934948921,
0.023384878411889076,
0.06996983289718628,
0.043651893734931946,
-0.06657912582159042,
0.08374866843223572,
-0.07241640985012054,
0.01520271971821785,
-0.15506552159786224,
-0.08848509192466736,
-0.023774081841111183,
0.011511479504406452,
-0.036034148186445236,
0.0721561461687088,
0.028976531699299812,
-0.025008656084537506,
-0.020763549953699112,
-0.06965016573667526,
0.11772362887859344,
0.04865168407559395,
-0.08418596535921097,
-0.028577685356140137,
-0.0018436623504385352,
-0.02190971188247204,
-0.0247767586261034,
0.13460059463977814,
-0.07623261958360672,
0.24662993848323822,
0.09712981432676315,
0.08616308122873306,
-0.18647433817386627,
-0.00853667501360178,
-0.003607884282246232,
-0.08599067479372025,
0.07983455061912537,
-0.05240541696548462,
0.07452008128166199,
0.08955749869346619,
0.01981748640537262,
0.08886528015136719,
0.11080878973007202,
-0.008541845716536045,
-0.024832597002387047,
0.18267756700515747,
-0.18377229571342468,
0.05420858412981033,
-0.07339087873697281,
-0.036655131727457047,
0.011988529935479164,
-0.014469988644123077,
0.10823020339012146,
0.012440677732229233,
0.013977005146443844,
-0.007391748484224081,
-0.04270554706454277,
-0.07903293520212173,
0.06440825015306473,
0.11268201470375061,
0.04197923094034195,
-0.12098194658756256,
-0.013280150480568409,
-0.0034102106001228094,
-0.1626819521188736,
0.012065340764820576,
-0.015798499807715416,
-0.09504900872707367,
-0.10021037608385086,
0.06259552389383316,
0.10644295811653137,
-0.1291263997554779,
-0.040158115327358246,
-0.11430942267179489,
-0.11904085427522659,
0.09068558365106583,
0.31644853949546814,
0.08677760511636734,
0.10012417286634445,
0.031366318464279175,
-0.005782293621450663,
-0.001768449554219842,
-0.005622320808470249,
0.006916365120559931,
0.02220650389790535,
-0.10641536116600037,
0.11033403873443604,
-0.055655498057603836,
0.1333911418914795,
-0.13713371753692627,
-0.030170608311891556,
-0.18058747053146362,
0.02135332114994526,
-0.11590111255645752,
-0.0422915481030941,
-0.07284621894359589,
-0.05581505596637726,
0.05785138159990311,
-0.09088589251041412,
-0.07888615131378174,
-0.034667741507291794,
-0.13626335561275482,
0.04524976760149002,
0.05795403569936752,
-0.005972633138298988,
-0.02274378575384617,
-0.012868447229266167,
0.13148395717144012,
-0.03458724543452263,
0.05368667095899582,
0.1135440394282341,
-0.06553711742162704,
0.11817150563001633,
-0.09726561605930328,
-0.10096146166324615,
0.07600395381450653,
0.043523259460926056,
0.0619838684797287,
0.034587033092975616,
0.01229378767311573,
0.039380524307489395,
0.04963087663054466,
0.043830014765262604,
0.024640554562211037,
-0.08749712258577347,
0.021984189748764038,
-0.012316819280385971,
-0.13999585807323456,
-0.03081514500081539,
-0.060124751180410385,
0.09107375890016556,
0.019467653706669807,
0.07518960535526276,
-0.07167627662420273,
0.08419368416070938,
-0.042876288294792175,
0.01660837046802044,
-0.030686190351843834,
-0.10572521388530731,
0.022983277216553688,
-0.026656627655029297,
0.01251986064016819,
-0.014156930148601532,
0.2132682204246521,
-0.00628779549151659,
0.05454711616039276,
0.04149198904633522,
0.01970505528151989,
0.10290651023387909,
0.0024518154095858335,
0.27581292390823364,
0.09299074113368988,
-0.016475440934300423,
-0.13366058468818665,
0.0931694284081459,
-0.0031840309966355562,
-0.10320062935352325,
0.14420278370380402,
0.07799676805734634,
-0.07678233087062836,
0.06233298406004906,
-0.0012125575449317694,
-0.05739833042025566,
0.06437022984027863,
-0.15090450644493103,
-0.04037998989224434,
0.09464351087808609,
0.03541694954037666,
-0.029542040079832077,
0.14420783519744873,
-0.04129832237958908,
0.03785943612456322,
-0.012674197554588318,
0.006610057316720486,
-0.18569014966487885,
-0.10675752907991409,
-0.10199112445116043,
-0.13667325675487518,
0.032409824430942535,
-0.06317730247974396,
-0.017300376668572426,
0.0803971216082573,
0.04962587356567383,
-0.0005138966371305287,
0.18775111436843872,
-0.0956338495016098,
-0.07236050069332123,
0.04231424629688263,
-0.009723044000566006,
-0.04206345975399017,
-0.08007223159074783,
-0.035000160336494446,
-0.06545595079660416,
0.028011104092001915,
-0.1088850274682045,
-0.04262785241007805,
-0.05196830630302429,
0.028862489387392998,
-0.09196444600820541,
-0.08700177818536758,
-0.06893384456634521,
0.022616561502218246,
-0.07622350752353668,
0.07738178223371506,
0.0010606672149151564,
0.06625279039144516,
0.014977718703448772,
0.11302351206541061,
-0.0710948035120964,
-0.13553334772586823,
-0.17081569135189056,
0.17477484047412872,
-0.015586421824991703,
0.09106175601482391,
-0.030886072665452957,
-0.0023205445613712072,
-0.07019472867250443,
0.3485925495624542,
0.24955683946609497,
-0.026121770963072777,
0.05311055853962898,
0.04206584766507149,
0.0401461161673069,
0.045228634029626846,
0.1494004875421524,
0.058998625725507736,
0.2065235674381256,
-0.05751710757613182,
-0.15357810258865356,
-0.00707464013248682,
-0.021719615906476974,
-0.06663738191127777,
-0.008245185017585754,
0.08680449426174164,
-0.046184051781892776,
-0.032751306891441345,
0.11718534678220749,
-0.14824621379375458,
0.06742644309997559,
0.015025177039206028,
-0.13683275878429413,
-0.05389387533068657,
-0.06299667805433273,
0.04276658967137337,
0.04083248972892761,
0.0849292203783989,
-0.0768495500087738,
-0.03941527381539345,
0.05695216357707977,
0.014200659468770027,
-0.2243623286485672,
-0.10158994048833847,
0.12319184094667435,
-0.046445563435554504,
0.08290817588567734,
-0.019328314810991287,
0.05993558093905449,
0.08491167426109314,
0.03692527487874031,
-0.02249038778245449,
0.023084858432412148,
0.03654235973954201,
-0.12671387195587158,
-0.040198683738708496,
0.04083153232932091,
0.014049685560166836,
-0.03471129387617111,
0.009849331341683865,
-0.1937323808670044,
0.06009393185377121,
-0.030061572790145874,
-0.03268570452928543,
-0.029644256457686424,
0.0792255625128746,
-0.10289949923753738,
0.11105287075042725,
0.1167161837220192,
0.002029495080932975,
-0.002681157784536481,
-0.06176215037703514,
0.05465411767363548,
0.07483833283185959,
-0.11109264940023422,
-0.14462910592556,
-0.1179095134139061,
-0.05808859318494797,
-0.07839802652597427,
-0.08100049197673798,
-0.17128737270832062,
-0.043028078973293304,
-0.08405254036188126,
0.016455641016364098,
-0.058819618076086044,
0.07526671141386032,
0.08395033329725266,
0.03977546468377113,
0.007307989057153463,
-0.00710875540971756,
0.038840632885694504,
0.0497129000723362,
-0.12808674573898315,
-0.07670791447162628
] |
null | null |
transformers
|
# Description:
This is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.
# How to Use:
The model can be used directly with a pipeline for masked language modeling:
```python
>>> from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline
>>> tokenizer = AutoTokenizer.from_pretrained("d42kw01f/Sinhala-RoBERTa")
>>> model = AutoModelForMaskedLM.from_pretrained("d42kw01f/Sinhala-RoBERTa")
>>> fill_mask = pipeline('fill-mask', model=model, tokenizer=tokenizer)
>>> fill_mask("මම ගෙදර <mask>.")
[{'score': 0.1822454035282135,
'sequence': 'මම ගෙදර ආව.',
'token': 701,
'token_str': ' ආව'},
{'score': 0.10513380169868469,
'sequence': 'මම ගෙදර ය.',
'token': 310,
'token_str': ' ය'},
{'score': 0.06417194753885269,
'sequence': 'මම ගෙදර එක.',
'token': 328,
'token_str': ' එක'},
{'score': 0.05026362091302872,
'sequence': 'මම ගෙදර ඇත.',
'token': 330,
'token_str': ' ඇත'},
{'score': 0.029960114508867264,
'sequence': 'මම ගෙදර යනව.',
'token': 834,
'token_str': ' යනව'}]
```
|
{}
|
fill-mask
|
d42kw01f/Sinhala-RoBERTa
|
[
"transformers",
"pytorch",
"roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
|
# Description:
This is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.
# How to Use:
The model can be used directly with a pipeline for masked language modeling:
|
[
"# Description:\n\nThis is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.",
"# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:"
] |
[
"TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"# Description:\n\nThis is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.",
"# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:"
] |
[
37,
39,
22
] |
[
"passage: TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# Description:\n\nThis is a smaller per-trained model on Sinhalese Language using Masked Language Modeling(MLM). And the model is trained on Oscar Sinhala dataset.# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:"
] |
[
-0.06981303542852402,
-0.013871806673705578,
-0.002937681507319212,
0.06777280569076538,
0.16878779232501984,
0.00494441669434309,
0.12198388576507568,
0.04897332563996315,
-0.018968595191836357,
-0.03944399207830429,
0.13632826507091522,
0.04446830973029137,
0.013817831873893738,
0.16273614764213562,
0.053546447306871414,
-0.4239024221897125,
0.00979489367455244,
-0.0225876085460186,
-0.04638388007879257,
0.10715381056070328,
0.12413541227579117,
-0.02528974786400795,
0.0771169438958168,
0.04932312294840813,
-0.07936476916074753,
0.029843905940651894,
-0.022205574437975883,
-0.09870719164609909,
0.06832592934370041,
0.020515939220786095,
0.0986592248082161,
-0.057686697691679,
0.07245297729969025,
-0.08628211915493011,
0.04990250617265701,
-0.05078044906258583,
0.02937581017613411,
-0.04104109853506088,
-0.03275354206562042,
-0.05463511496782303,
0.25868353247642517,
0.019529519602656364,
0.023048100993037224,
0.03822454810142517,
-0.1305742710828781,
-0.10555802285671234,
0.07398613542318344,
-0.120207279920578,
0.04613327607512474,
0.1433180272579193,
-0.03589393198490143,
0.16687226295471191,
-0.15382301807403564,
0.08084648102521896,
-0.04510420933365822,
-0.17193610966205597,
-0.011939623393118382,
0.16038918495178223,
0.1320858895778656,
-0.060887206345796585,
-0.01902216486632824,
0.09474441409111023,
0.02161748707294464,
0.07923702895641327,
0.05464504659175873,
-0.0870261937379837,
0.12107778340578079,
-0.09012101590633392,
-0.10114505141973495,
0.021397756412625313,
0.15463288128376007,
-0.05901305004954338,
-0.030826086178421974,
-0.09283945709466934,
-0.027226822450757027,
0.08314958959817886,
-0.030422393232584,
-0.0005212990217842162,
-0.04937286302447319,
0.04753800109028816,
0.050784628838300705,
-0.044220682233572006,
-0.05687043070793152,
-0.014370834454894066,
-0.09971508383750916,
0.16880248486995697,
0.011891392059624195,
0.04939929023385048,
-0.24299417436122894,
-0.03704894334077835,
-0.05800619721412659,
-0.11846154928207397,
0.01020907424390316,
-0.06858015805482864,
0.023423422127962112,
0.08615496009588242,
-0.009800590574741364,
-0.10411541908979416,
0.06254685670137405,
0.03461705893278122,
0.07326367497444153,
0.028334802016615868,
0.06767719238996506,
0.030377497896552086,
0.05321294069290161,
0.15920300781726837,
-0.11225254088640213,
-0.03405619040131569,
0.08450593799352646,
-0.08550221472978592,
0.06575038284063339,
-0.05711361765861511,
-0.1318221241235733,
-0.041712261736392975,
-0.0856914222240448,
0.10619644820690155,
-0.003600445343181491,
0.11152949184179306,
-0.05085936188697815,
-0.07302562147378922,
-0.11953794211149216,
-0.14181064069271088,
-0.052827611565589905,
-0.02966049313545227,
-0.020601026713848114,
0.05161895602941513,
0.06776269525289536,
0.0018765988061204553,
-0.07259766757488251,
-0.04061496630311012,
-0.018799640238285065,
0.024129245430231094,
-0.13681986927986145,
-0.15489868819713593,
-0.011783597059547901,
-0.12194237112998962,
0.0465032160282135,
-0.15152134001255035,
-0.1304561197757721,
0.013891668058931828,
0.050511132925748825,
-0.03283613175153732,
-0.022140907123684883,
-0.09171362221240997,
0.029308870434761047,
-0.0015938583528622985,
0.012761018238961697,
-0.03052178956568241,
-0.03874775767326355,
-0.031876496970653534,
0.03341059014201164,
0.149908185005188,
-0.0628058984875679,
0.03604450821876526,
-0.024017399176955223,
0.08518582582473755,
-0.09855884313583374,
0.07131461799144745,
-0.04432553052902222,
0.07287783175706863,
-0.026319727301597595,
-0.07450217008590698,
-0.07329607754945755,
0.06993042677640915,
0.006653843447566032,
0.19128815829753876,
-0.19464631378650665,
-0.04908447340130806,
0.2176087200641632,
-0.13732534646987915,
-0.06714663654565811,
0.09961587190628052,
-0.0014596089022234082,
0.13176384568214417,
0.0161573588848114,
0.08668382465839386,
0.02228495106101036,
-0.12477706372737885,
0.17551299929618835,
0.08307643979787827,
-0.11099719256162643,
-0.04715033620595932,
0.10792861878871918,
0.01342079322785139,
-0.04786711931228638,
0.04617442190647125,
0.03429707884788513,
0.10083793103694916,
-0.061471253633499146,
-0.05933103337883949,
0.028813069686293602,
-0.07108630239963531,
0.00556509755551815,
0.024656744673848152,
0.06972961872816086,
-0.03682412952184677,
-0.002290786011144519,
0.03967765346169472,
0.10976827889680862,
-0.04583033546805382,
0.045886922627687454,
-0.19151130318641663,
0.04099065065383911,
-0.025300942361354828,
-0.013313822448253632,
-0.1543540209531784,
-0.005938657559454441,
-0.03954683616757393,
0.04529602825641632,
0.101474329829216,
-0.006020621862262487,
0.04775169864296913,
0.015370815061032772,
-0.02497561275959015,
0.08216515928506851,
0.08079473674297333,
0.027773968875408173,
-0.021387208253145218,
-0.08774126321077347,
0.03415834903717041,
-0.06915759295225143,
0.15077495574951172,
-0.016075564548373222,
-0.005308669991791248,
-0.14509235322475433,
0.05549193546175957,
-0.017700249329209328,
0.020356709137558937,
0.10053727775812149,
0.07830825448036194,
0.0012671208241954446,
-0.048678360879421234,
0.07464785873889923,
-0.010385793633759022,
-0.14266328513622284,
0.1906343698501587,
-0.18866412341594696,
0.1193932294845581,
0.13800960779190063,
-0.15194469690322876,
0.0442998930811882,
0.11703355610370636,
0.03890621289610863,
0.01901061274111271,
0.019539937376976013,
0.13116420805454254,
0.16899169981479645,
0.016003932803869247,
0.14410416781902313,
-0.031709685921669006,
0.07953306287527084,
0.0012117731384932995,
-0.14844220876693726,
-0.04220759496092796,
0.07287084311246872,
0.11312824487686157,
-0.189293771982193,
0.07846080511808395,
0.0672534853219986,
-0.027889426797628403,
0.21223007142543793,
0.026852065697312355,
-0.0030595725402235985,
-0.061612989753484726,
-0.08369507640600204,
0.020523060113191605,
0.06546889990568161,
-0.21103490889072418,
-0.053424060344696045,
0.027621697634458542,
-0.034523651003837585,
0.08203525096178055,
-0.019154522567987442,
-0.05070457234978676,
0.05294349044561386,
0.010886368341743946,
-0.08545072376728058,
0.13888613879680634,
-0.08097920566797256,
0.013315419666469097,
-0.019067032262682915,
-0.11093518882989883,
0.01257253997027874,
-0.008372005075216293,
-0.07319387048482895,
0.18970900774002075,
-0.16260787844657898,
-0.37465378642082214,
-0.13550755381584167,
-0.12635695934295654,
0.005812535528093576,
0.031675126403570175,
0.006254440639168024,
-0.11712143570184708,
-0.08527054637670517,
0.006639304105192423,
-0.016973422840237617,
-0.10826610773801804,
-0.0563894547522068,
0.026127852499485016,
0.002840016968548298,
-0.018789339810609818,
-0.10269014537334442,
-0.02262425608932972,
-0.012410328723490238,
-0.030902186408638954,
0.09841323643922806,
-0.20094063878059387,
0.054279014468193054,
0.18174642324447632,
0.013270239345729351,
0.08451477438211441,
-0.006945273373275995,
0.20843784511089325,
-0.06812519580125809,
-0.04624941945075989,
0.22769108414649963,
-0.015416008420288563,
0.00309136719442904,
0.10818295180797577,
-0.012097490020096302,
-0.0747574046254158,
0.017006976529955864,
-0.023321153596043587,
-0.13402345776557922,
-0.16680952906608582,
-0.1280718445777893,
-0.09152710437774658,
-0.006352666765451431,
-0.044666580855846405,
0.0029424382373690605,
-0.007380107883363962,
0.07333163917064667,
0.06910490244626999,
-0.04687570035457611,
0.003559195902198553,
0.033822350203990936,
0.09370304644107819,
-0.05874676629900932,
0.0963868722319603,
-0.05147299915552139,
-0.03388892486691475,
0.0285227932035923,
-0.07867797464132309,
0.08715210855007172,
0.12635083496570587,
0.04671206697821617,
0.10385799407958984,
0.12830978631973267,
0.13540245592594147,
0.08221876621246338,
-0.023542679846286774,
-0.009952718392014503,
-0.020009977743029594,
-0.025224508717656136,
-0.05320194363594055,
0.0037842877209186554,
0.08443736284971237,
-0.05407571792602539,
-0.038943514227867126,
0.22783365845680237,
-0.018308676779270172,
0.047492288053035736,
0.0672779232263565,
-0.16845257580280304,
-0.02930283360183239,
0.04018695652484894,
0.056013524532318115,
-0.08576422184705734,
0.05692620202898979,
0.045474570244550705,
-0.11907589435577393,
0.0859183818101883,
0.0018999370513483882,
0.0879063606262207,
0.05377911031246185,
0.044471390545368195,
-0.14334458112716675,
0.04405646771192551,
0.014069139957427979,
0.10192739218473434,
-0.2842293083667755,
0.2967989444732666,
0.02129185199737549,
0.03779105097055435,
-0.06817659735679626,
-0.02865181863307953,
0.06220215559005737,
0.19240248203277588,
0.14305540919303894,
0.044083304703235626,
-0.005308435298502445,
-0.08963818848133087,
-0.10364580154418945,
0.03129034489393234,
0.08174624294042587,
0.060059256851673126,
0.015850959345698357,
0.01797373592853546,
-0.003488609567284584,
-0.0015732628526166081,
0.031241364777088165,
-0.10044240951538086,
-0.10045665502548218,
0.03455459699034691,
0.07342387735843658,
-0.09436299651861191,
-0.03396255895495415,
-0.10102007538080215,
-0.027765469625592232,
0.1761481910943985,
0.04563947767019272,
-0.055925291031599045,
-0.10733362287282944,
-0.01583012193441391,
0.10303926467895508,
-0.11534187942743301,
0.0859491378068924,
-0.07057876139879227,
0.07162322849035263,
-0.057468730956315994,
-0.07035384327173233,
0.1068655475974083,
-0.07607413828372955,
0.07442024350166321,
0.020959671586751938,
0.082126185297966,
0.03496561199426651,
0.02508598193526268,
0.09536189585924149,
-0.011157013475894928,
-0.020121775567531586,
-0.0842025950551033,
-0.08510537445545197,
0.14193648099899292,
-0.025056567043066025,
0.12740880250930786,
-0.08947454392910004,
-0.11093201488256454,
-0.017328087240457535,
-0.07610005140304565,
0.25269782543182373,
0.08380085229873657,
-0.05947375297546387,
0.08079514652490616,
0.17392116785049438,
-0.10820655524730682,
-0.30985620617866516,
-0.11265663802623749,
-0.05654701590538025,
0.15244024991989136,
0.034610189497470856,
-0.12355618178844452,
0.033762846142053604,
-0.07905113697052002,
0.005360300187021494,
-0.1272306740283966,
-0.17537814378738403,
-0.16753803193569183,
0.22839690744876862,
0.08033040910959244,
0.2753322422504425,
-0.04144240543246269,
-0.0078846774995327,
-0.14361068606376648,
-0.05595748871564865,
-0.02754274383187294,
-0.05664630979299545,
0.12739421427249908,
-0.010751232504844666,
0.06433285772800446,
0.03157836198806763,
-0.020144013687968254,
0.13382719457149506,
-0.0300921443849802,
-0.007155481725931168,
-0.11458905786275864,
-0.09206024557352066,
0.07440536469221115,
0.006017157342284918,
0.10585229843854904,
0.0384911447763443,
-0.017833037301898003,
-0.12470695376396179,
-0.08807612210512161,
-0.014424207620322704,
0.040718916803598404,
0.035002194344997406,
-0.049592744559049606,
-0.0019469661638140678,
0.07561645656824112,
0.009402700699865818,
0.013668196275830269,
-0.03202653303742409,
-0.1368483006954193,
-0.006628453731536865,
0.0076835667714476585,
0.24811510741710663,
-0.09351256489753723,
0.02898881398141384,
-0.02161608450114727,
-0.008783620782196522,
0.008379669860005379,
-0.0022153714671730995,
-0.029264526441693306,
0.055540360510349274,
-0.001413254882209003,
0.08502465486526489,
0.03232283145189285,
-0.031914424151182175,
0.06298459321260452,
0.09697291254997253,
0.05013052001595497,
-0.12497101724147797,
-0.02513756975531578,
-0.07694441080093384,
0.058687470853328705,
-0.039932981133461,
0.06700453907251358,
-0.0396491140127182,
-0.02542102336883545,
-0.05117995664477348,
-0.0029914104379713535,
-0.06612729281187057,
0.06443246454000473,
0.06918802112340927,
0.02804923988878727,
-0.07959523797035217,
0.06521640717983246,
-0.019621916115283966,
-0.006398398894816637,
0.04086456075310707,
0.10619234293699265,
-0.10648805648088455,
-0.07938257604837418,
0.03589673712849617,
0.196670800447464,
-0.04694245010614395,
-0.073465995490551,
-0.020500024780631065,
-0.10703641176223755,
-0.021488474681973457,
0.15994028747081757,
0.0773802101612091,
-0.0227949321269989,
-0.13549935817718506,
-0.04319974035024643,
-0.06494088470935822,
0.018689285963773727,
0.13675346970558167,
-0.10095765441656113,
-0.09054374694824219,
0.1403590440750122,
0.07777746766805649,
0.15021046996116638,
-0.09057047218084335,
-0.11753034591674805,
-0.06771652400493622,
0.08718051016330719,
-0.12744997441768646,
0.06747966259717941,
-0.10013841837644577,
-0.014785357750952244,
-0.020686564967036247,
-0.06577055156230927,
-0.06959710270166397,
0.03175614774227142,
-0.08804670721292496,
0.033095620572566986,
-0.04168369621038437,
-0.0202584657818079,
0.04335080087184906,
-0.058555010706186295,
0.05663291737437248,
0.010091989301145077,
0.03339652344584465,
0.115078866481781,
-0.09049095213413239,
0.07406463474035263,
-0.12174998223781586,
-0.05892553552985191,
0.06757483631372452,
0.02956615388393402,
0.056378643959760666,
-0.09290260076522827,
-0.009314191527664661,
0.047689665108919144,
0.06394411623477936,
0.03629130870103836,
0.07573705911636353,
-0.03789807856082916,
-0.02614927478134632,
0.026711298152804375,
-0.04772467166185379,
0.0017705532955005765,
0.02259417437016964,
0.0338849201798439,
0.06664168834686279,
0.10988550633192062,
-0.0647491067647934,
0.0982324630022049,
-0.05597507208585739,
0.023574290797114372,
-0.09655457735061646,
-0.010422144085168839,
-0.013406053185462952,
-0.0621713325381279,
0.05822232365608215,
-0.03247874230146408,
0.18895785510540009,
0.09050537645816803,
0.07911795377731323,
-0.018011586740612984,
0.05070146918296814,
0.031050967052578926,
-0.0325736440718174,
0.1869802325963974,
0.09277777373790741,
0.07548048347234726,
0.0745464339852333,
0.025229545310139656,
-0.007974823005497456,
0.1707540601491928,
0.0059904782101511955,
0.019961170852184296,
0.09936022013425827,
0.1498953253030777,
0.033338773995637894,
0.10847322642803192,
-0.061034440994262695,
-0.23265819251537323,
-0.04999305680394173,
0.058327566832304,
-0.023955088108778,
-0.05058005824685097,
0.16566181182861328,
-0.04174375534057617,
0.08521850407123566,
-0.01863657869398594,
-0.12663014233112335,
-0.23415566980838776,
-0.2629638612270355,
-0.07853604108095169,
0.02835944853723049,
0.02417578734457493,
-0.11699226498603821,
-0.030252784490585327,
0.05915513634681702,
0.1038769856095314,
-0.04004363343119621,
0.1618667095899582,
-0.1528978794813156,
-0.05505026876926422,
0.04162309318780899,
-0.0120856249704957,
0.04972366616129875,
0.021495800465345383,
0.05522678419947624,
-0.11506564915180206,
-0.007906110025942326,
0.013088681735098362,
-0.03737097606062889,
-0.048528190702199936,
0.03407822549343109,
-0.09849001467227936,
-0.04345886781811714,
-0.09345710277557373,
0.024389799684286118,
0.05918976664543152,
0.1555771380662918,
0.009825484827160835,
-0.08015049248933792,
-0.011753348633646965,
0.08923321217298508,
0.025620151311159134,
-0.2840748429298401,
-0.1341552734375,
0.2841452360153198,
-0.0006231832085177302,
0.03215399757027626,
-0.10542232543230057,
0.05108438804745674,
-0.01697799190878868,
0.31983521580696106,
0.3449276089668274,
-0.10300271958112717,
0.009785987436771393,
-0.039946265518665314,
0.029502110555768013,
-0.040291495621204376,
0.14634636044502258,
0.033985935151576996,
0.16231882572174072,
-0.05498005077242851,
0.016917120665311813,
-0.1287693828344345,
-0.0634557232260704,
-0.19240877032279968,
-0.0717082992196083,
0.07472455501556396,
-0.04718128964304924,
0.016637951135635376,
0.13772791624069214,
-0.1737225502729416,
0.13639529049396515,
-0.07779930531978607,
-0.07929279655218124,
-0.09532133489847183,
-0.07971932739019394,
-0.0780882015824318,
0.12131232768297195,
0.07760221511125565,
-0.0401923805475235,
-0.0002456881629768759,
0.01023957785218954,
0.04437786340713501,
-0.16234251856803894,
-0.05833618715405464,
0.13829457759857178,
0.02636900544166565,
0.0517725795507431,
0.029007500037550926,
0.047565240412950516,
0.05981863662600517,
0.05394461750984192,
-0.005608620122075081,
0.15336838364601135,
-0.03615490719676018,
0.09731926023960114,
0.05202067270874977,
0.06769012659788132,
-0.09578811377286911,
-0.09437751024961472,
-0.030333684757351875,
0.0022147540003061295,
0.07850459218025208,
-0.09615244716405869,
-0.03816070035099983,
0.014258679002523422,
0.10375243425369263,
-0.10241423547267914,
0.09992170333862305,
0.12646184861660004,
-0.009689854457974434,
-0.076097272336483,
-0.03708403557538986,
0.058505069464445114,
-0.033339302986860275,
-0.09849344193935394,
-0.1286320835351944,
-0.07759249210357666,
-0.08330436795949936,
-0.10009738802909851,
-0.018204843625426292,
-0.17317692935466766,
-0.03585680574178696,
-0.10608850419521332,
-0.026195736601948738,
-0.014424741268157959,
0.06844572722911835,
0.062000572681427,
0.04333117976784706,
-0.014479577541351318,
-0.050344452261924744,
0.06213580071926117,
0.07075780630111694,
-0.19198167324066162,
-0.12232889235019684
] |
null | null |
transformers
|
# Description:
This is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.
# How to Use:
The model can be used directly with a pipeline for masked language modeling:
```python
>>> from transformers import AutoTokenizer, AutoModelForMaskedLM, pipeline
>>> tokenizer = AutoTokenizer.from_pretrained("d42kw01f/Tamil-RoBERTa")
>>> model = AutoModelForMaskedLM.from_pretrained("d42kw01f/Tamil-RoBERTa")
>>> fill_mask = pipeline('fill-mask', model=model, tokenizer=tokenizer)
>>> fill_mask("நான் வீட்டு <mask>.")
```
|
{}
|
fill-mask
|
d42kw01f/Tamil-RoBERTa
|
[
"transformers",
"pytorch",
"roberta",
"fill-mask",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us
|
# Description:
This is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.
# How to Use:
The model can be used directly with a pipeline for masked language modeling:
|
[
"# Description:\n\nThis is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.",
"# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:"
] |
[
"TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n",
"# Description:\n\nThis is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.",
"# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:"
] |
[
37,
37,
22
] |
[
"passage: TAGS\n#transformers #pytorch #roberta #fill-mask #autotrain_compatible #endpoints_compatible #region-us \n# Description:\n\nThis is a smaller per-trained model on Tamil Language using Masked Language Modeling(MLM). And the model is trained on Oscar Tamil dataset.# How to Use:\nThe model can be used directly with a pipeline for masked language modeling:"
] |
[
-0.05470907688140869,
-0.05960835888981819,
-0.0018824213184416294,
0.06922741234302521,
0.1661115437746048,
-0.017173541709780693,
0.16198600828647614,
0.052911125123500824,
0.03793082758784294,
-0.03748409450054169,
0.12512698769569397,
0.11311330646276474,
0.0008931029005907476,
0.15146887302398682,
0.03670521453022957,
-0.4382869601249695,
-0.0016318145208060741,
-0.0216978769749403,
-0.05161536857485771,
0.0969729796051979,
0.1340651959180832,
-0.03427586704492569,
0.08342784643173218,
0.03197740763425827,
-0.05772296339273453,
0.002233500126749277,
-0.018273701891303062,
-0.08558827638626099,
0.0671960860490799,
-0.011822703294456005,
0.08210847526788712,
-0.06231231614947319,
0.07855113595724106,
-0.07380597293376923,
0.053546734154224396,
-0.049269355833530426,
0.05751131847500801,
-0.00848042406141758,
0.027612505480647087,
-0.02117978036403656,
0.32032328844070435,
0.022043563425540924,
0.059950657188892365,
0.02203219383955002,
-0.08594223111867905,
-0.00498528266325593,
0.052581027150154114,
-0.11161057651042938,
0.05600805580615997,
0.16920772194862366,
-0.032987676560878754,
0.23485156893730164,
-0.10586287826299667,
0.11652710288763046,
-0.05148641765117645,
-0.2313985675573349,
-0.02533225528895855,
0.18577007949352264,
0.141518235206604,
-0.03833044692873955,
-0.0004793820553459227,
0.11216562241315842,
0.005106575321406126,
0.08826375752687454,
0.04255671799182892,
-0.053260210901498795,
0.005597040057182312,
-0.0713481679558754,
-0.09578956663608551,
0.04906518757343292,
0.11872253566980362,
-0.09924400597810745,
0.009651332162320614,
-0.09273356199264526,
-0.049998246133327484,
0.02407675050199032,
-0.030407272279262543,
0.0006808124599047005,
-0.03702101483941078,
0.0627845823764801,
0.07366258651018143,
-0.01833275519311428,
-0.06142260506749153,
-0.012135868892073631,
-0.08868918567895889,
0.14109130203723907,
0.02668546698987484,
0.06436646729707718,
-0.2722378969192505,
-0.024872275069355965,
-0.0666184350848198,
-0.08976247906684875,
0.032113008201122284,
-0.07063554227352142,
0.026589589193463326,
0.051795121282339096,
-0.035061728209257126,
-0.13686594367027283,
0.01243554800748825,
0.011246302165091038,
0.04592601954936981,
-0.009500158950686455,
0.055931925773620605,
0.03926053270697594,
0.05225742235779762,
0.1649980992078781,
-0.17684471607208252,
0.042661987245082855,
0.10008089989423752,
-0.055537380278110504,
0.09357725083827972,
-0.058145295828580856,
-0.11653009057044983,
0.00010735169053077698,
-0.07217703759670258,
0.10839906334877014,
-0.008310756646096706,
0.14391666650772095,
-0.04451220855116844,
-0.08656120300292969,
-0.07378077507019043,
-0.13339774310588837,
-0.03369242697954178,
-0.05964156985282898,
-0.022208398208022118,
0.023993775248527527,
0.06734819710254669,
0.015429181046783924,
-0.05234583467245102,
-0.05064805597066879,
-0.060958340764045715,
0.06755348294973373,
-0.13854171335697174,
-0.17337308824062347,
-0.012320440262556076,
-0.1624804437160492,
0.04091868922114372,
-0.1350836455821991,
-0.12184115499258041,
0.06131918728351593,
0.04719647020101547,
-0.07199189811944962,
-0.020836228504776955,
-0.07197000831365585,
0.009832327254116535,
-0.009213629178702831,
0.02849096804857254,
-0.03244148567318916,
-0.018144771456718445,
-0.04380837455391884,
0.010624543763697147,
0.15644162893295288,
-0.011263446882367134,
0.03310434892773628,
-0.045585617423057556,
0.05232040956616402,
-0.05976366624236107,
0.07339473068714142,
-0.03805239871144295,
0.11347883939743042,
-0.001873405883088708,
-0.047380901873111725,
-0.030809126794338226,
0.05074906349182129,
0.01743883080780506,
0.1730729341506958,
-0.20529970526695251,
-0.026651300489902496,
0.1307196021080017,
-0.12780368328094482,
-0.12306808680295944,
0.06412270665168762,
-0.0072016241028904915,
0.11726126819849014,
0.01873469166457653,
0.04846130311489105,
-0.03647361323237419,
-0.09393813461065292,
0.10772547870874405,
0.0820944607257843,
-0.11752520501613617,
-0.023431673645973206,
0.10385570675134659,
0.03871864452958107,
-0.06846389174461365,
0.035075198858976364,
0.01615588180720806,
0.06839988380670547,
-0.09486784785985947,
-0.05676453560590744,
0.003678669687360525,
-0.08837009221315384,
0.010239314287900925,
0.014936699531972408,
0.11968083679676056,
-0.021168813109397888,
0.002192472107708454,
-0.02296590805053711,
0.10726247727870941,
-0.029091045260429382,
0.030402444303035736,
-0.22280986607074738,
0.04509250819683075,
-0.06738890707492828,
0.010760589502751827,
-0.21285323798656464,
-0.010582932271063328,
0.009867189452052116,
0.09920988231897354,
0.0805811956524849,
-0.038387883454561234,
0.05254087597131729,
0.04172882065176964,
0.004649993032217026,
0.09357380121946335,
0.08531971275806427,
0.04244087636470795,
-0.06568510830402374,
-0.11918900161981583,
0.013146919198334217,
-0.08921844512224197,
0.11132655292749405,
-0.030752940103411674,
0.0009372485219500959,
-0.10042306035757065,
0.06139775738120079,
0.01882277987897396,
0.011669035069644451,
0.09600738435983658,
0.06283106654882431,
-0.005345492158085108,
-0.06425618380308151,
0.06683772802352905,
-0.016582563519477844,
-0.10110731422901154,
0.1629505604505539,
-0.1562114655971527,
0.0994575098156929,
0.11934725940227509,
-0.13983675837516785,
0.05175141245126724,
0.2010537087917328,
0.014175018295645714,
0.02152564562857151,
0.02199883759021759,
0.15309593081474304,
0.18038342893123627,
0.03463246300816536,
0.1663077175617218,
-0.027172105386853218,
0.0782204121351242,
-0.006506034638732672,
-0.14421814680099487,
-0.04551391676068306,
0.06875989586114883,
0.18055371940135956,
-0.18671490252017975,
0.06675016134977341,
-0.0016811555251479149,
-0.010444141924381256,
0.16641713678836823,
0.069273941218853,
-0.006078590638935566,
-0.04627438634634018,
-0.12808829545974731,
0.01999475061893463,
0.018668048083782196,
-0.16872020065784454,
-0.0534202866256237,
0.028000440448522568,
-0.05176851153373718,
0.08970082551240921,
-0.06678209453821182,
-0.05043603852391243,
0.05276208370923996,
0.03353898599743843,
-0.11055230349302292,
0.17110243439674377,
-0.07610686123371124,
0.04360020533204079,
-0.03093993104994297,
-0.08706190437078476,
0.00542545085772872,
0.007926680147647858,
-0.12476666271686554,
0.18095719814300537,
-0.1503944993019104,
-0.39813628792762756,
-0.1443590372800827,
-0.12273567914962769,
-0.009066280908882618,
0.03629279509186745,
-0.01998179219663143,
-0.128207728266716,
-0.10604387521743774,
-0.02379244565963745,
0.0013893023133277893,
-0.06909549981355667,
-0.033664774149656296,
0.03187307342886925,
0.015014891512691975,
-0.02118060737848282,
-0.09267962723970413,
-0.028702717274427414,
-0.02046694979071617,
-0.07742297649383545,
0.0869424119591713,
-0.15517011284828186,
0.041564665734767914,
0.16822278499603271,
-0.028378179296851158,
0.08042066544294357,
0.018262360244989395,
0.21476441621780396,
-0.05835657939314842,
-0.009393049404025078,
0.23334026336669922,
0.027489008381962776,
-0.009977507404983044,
0.16735440492630005,
0.015971677377820015,
-0.060625847429037094,
0.020696556195616722,
-0.028628723695874214,
-0.10505901277065277,
-0.20048055052757263,
-0.14726419746875763,
-0.0724082961678505,
-0.027887897565960884,
-0.021650927141308784,
0.018794596195220947,
0.028393881395459175,
0.07762712240219116,
0.06966081261634827,
0.002358559286221862,
-0.0007221364066936076,
0.024753529578447342,
0.027688756585121155,
-0.04378032684326172,
0.118248850107193,
-0.06938058137893677,
-0.03068488836288452,
0.019398372620344162,
-0.09559008479118347,
0.16733859479427338,
0.12140040099620819,
-0.0124610411003232,
0.09341533482074738,
0.060780927538871765,
0.16136018931865692,
0.10476436465978622,
-0.037302806973457336,
-0.04072825610637665,
-0.0014047938166186213,
-0.029794931411743164,
-0.026749135926365852,
0.04081879183650017,
0.04002274572849274,
-0.0811694785952568,
-0.036639824509620667,
0.242766335606575,
-0.0018147295340895653,
0.0910331979393959,
0.04826464131474495,
-0.23325997591018677,
-0.026663530617952347,
0.03683413565158844,
0.050756510347127914,
-0.06731332838535309,
0.02720138244330883,
0.008171254768967628,
-0.11390470713376999,
0.07529740780591965,
-0.0362260602414608,
0.10274410247802734,
0.021822582930326462,
0.04639635235071182,
-0.10271140187978745,
0.028450652956962585,
0.04233722761273384,
0.1010417565703392,
-0.22346346080303192,
0.3088914155960083,
0.02438518963754177,
0.03930164501070976,
-0.048229657113552094,
-0.014858769252896309,
0.07704320549964905,
0.25367555022239685,
0.1221834123134613,
0.017090747132897377,
0.020732345059514046,
-0.14607270061969757,
-0.09807131439447403,
0.01668313890695572,
0.05396176874637604,
0.030108265578746796,
0.051039084792137146,
-0.01881863735616207,
-0.02964039333164692,
-0.006298616528511047,
0.00410460727289319,
-0.13209199905395508,
-0.0931176245212555,
0.07767115533351898,
0.059787701815366745,
-0.10409180819988251,
-0.04113946855068207,
-0.13636983931064606,
-0.028895972296595573,
0.17957332730293274,
0.05121477693319321,
-0.07007133960723877,
-0.05636760592460632,
0.013145795091986656,
0.09719406068325043,
-0.10545169562101364,
0.06779941916465759,
-0.0785512775182724,
0.0922354981303215,
-0.06162041053175926,
-0.08822758495807648,
0.10471931099891663,
-0.0792984813451767,
0.08371701091527939,
0.03768949955701828,
0.035246215760707855,
-0.006200993433594704,
0.004893151111900806,
0.051283542066812515,
0.02106725238263607,
-0.01412571594119072,
-0.06602934002876282,
-0.1037839949131012,
0.21046972274780273,
-0.05704593285918236,
0.10890568792819977,
-0.1578439474105835,
-0.08706425875425339,
-0.05289813503623009,
-0.06547291576862335,
0.2459067702293396,
0.06599624454975128,
-0.08718974888324738,
0.08300311118364334,
0.11829529702663422,
-0.10427607595920563,
-0.2977031469345093,
-0.09408951550722122,
-0.057912860065698624,
0.13995979726314545,
0.07665543258190155,
-0.1416826844215393,
0.024131832644343376,
-0.09704822301864624,
-0.0003389979829080403,
-0.09147077798843384,
-0.17133642733097076,
-0.15646867454051971,
0.18613915145397186,
0.08867257088422775,
0.2246176153421402,
-0.04760761559009552,
-0.0050115566700696945,
-0.10256312042474747,
0.022693870589137077,
-0.03182420879602432,
0.012759603559970856,
0.1252053678035736,
-0.009845070540904999,
0.0416535884141922,
0.03459608182311058,
-0.028185658156871796,
0.11960572004318237,
-0.06448767334222794,
-0.007938604801893234,
-0.10906194895505905,
-0.11775942146778107,
0.0814235582947731,
0.02609417587518692,
0.1726374626159668,
0.0677919015288353,
0.00804061908274889,
-0.08898533135652542,
-0.08613542467355728,
0.008143571205437183,
-0.01792503520846367,
0.06198234483599663,
-0.026161406189203262,
-0.003882448421791196,
0.07328367978334427,
-0.0044461991637945175,
0.010525495745241642,
-0.03434258699417114,
-0.10778944194316864,
0.0006252369494177401,
-0.011345941573381424,
0.23107992112636566,
-0.1386377364397049,
0.039691027253866196,
-0.017895476892590523,
-0.0011833836324512959,
0.009180154651403427,
-0.015772070735692978,
-0.06031882390379906,
0.061062898486852646,
-0.02788521721959114,
0.08874355256557465,
0.04307638108730316,
0.004839784000068903,
0.03142441436648369,
0.11896596848964691,
0.048598531633615494,
-0.1415558159351349,
-0.043278809636831284,
0.02576340176165104,
0.05438574403524399,
-0.029399719089269638,
0.06887590140104294,
-0.08019950985908508,
-0.02382521890103817,
-0.057635121047496796,
0.020634155720472336,
-0.0436147004365921,
0.03840181604027748,
0.03300134092569351,
0.03224819526076317,
-0.09371161460876465,
0.06165500730276108,
0.01094481348991394,
-0.008821483701467514,
0.033532194793224335,
0.11835259199142456,
-0.1432919055223465,
-0.09383539110422134,
-0.004076181445270777,
0.1691485345363617,
-0.024595364928245544,
-0.026416154578328133,
-0.050277624279260635,
-0.10046710073947906,
-0.004105022177100182,
0.12006903439760208,
0.0747578889131546,
0.006321767345070839,
-0.19471293687820435,
-0.0748923048377037,
-0.06706461310386658,
0.030234016478061676,
0.11164754629135132,
-0.09732991456985474,
-0.12688224017620087,
0.098687544465065,
0.09348946809768677,
0.13393887877464294,
-0.09394941478967667,
-0.09773881733417511,
0.008119504898786545,
0.05483822897076607,
-0.07741593569517136,
0.04114016890525818,
-0.09103123843669891,
-0.03341449424624443,
-0.012335197068750858,
-0.051570385694503784,
-0.06162216141819954,
0.02016761712729931,
-0.07093937695026398,
0.022241219878196716,
-0.08940389007329941,
-0.01255261804908514,
0.03381747379899025,
-0.07289872318506241,
0.06080833449959755,
-0.005969781428575516,
0.04109444469213486,
0.08186545968055725,
-0.10731445997953415,
0.04772002249956131,
-0.09986725449562073,
-0.08211629092693329,
0.03451176732778549,
0.04378829523921013,
0.06429600715637207,
-0.10776594281196594,
0.00364294508472085,
0.047343913465738297,
0.06461150199174881,
0.0542486272752285,
0.08940652012825012,
-0.03988073021173477,
-0.05547548830509186,
0.032934125512838364,
-0.009042522870004177,
-0.010830452665686607,
0.03100249543786049,
0.009135144762694836,
0.08027292042970657,
0.1098882406949997,
-0.07149089872837067,
0.09734955430030823,
-0.054544903337955475,
0.0460338369011879,
-0.10274019092321396,
-0.02142753079533577,
-0.042923711240291595,
-0.05816500261425972,
0.0672895684838295,
-0.05962096527218819,
0.19202573597431183,
0.13290423154830933,
0.06170928478240967,
0.001944023766554892,
0.13265395164489746,
0.060607973486185074,
-0.01588711328804493,
0.14458538591861725,
0.10672629624605179,
0.07872603088617325,
0.02178521454334259,
0.04610838741064072,
-0.006738552823662758,
0.12065097689628601,
0.01929732970893383,
0.0021590429823845625,
0.09680673480033875,
0.1155959814786911,
0.05965888872742653,
0.0991581380367279,
-0.07916926592588425,
-0.24029624462127686,
-0.08411019295454025,
0.040987495332956314,
-0.04715872183442116,
-0.04448936879634857,
0.21106581389904022,
-0.038809794932603836,
0.08042740821838379,
-0.025360815227031708,
-0.11709784716367722,
-0.20076583325862885,
-0.23171910643577576,
-0.1031196191906929,
0.007877098396420479,
0.023358702659606934,
-0.12015574425458908,
-0.04092352092266083,
0.053498148918151855,
0.09491460025310516,
-0.060069579631090164,
0.18100538849830627,
-0.16168765723705292,
-0.020975207909941673,
0.030173353850841522,
-0.019834673032164574,
0.04240437597036362,
0.014735506847500801,
0.04884396120905876,
-0.13155025243759155,
-0.037675920873880386,
0.0163456741720438,
-0.013651524670422077,
-0.06703027337789536,
0.05677202716469765,
-0.06693022698163986,
-0.07033014297485352,
-0.08201046288013458,
-0.003498905338346958,
0.006890883669257164,
0.17361851036548615,
0.017633160576224327,
-0.06654192507266998,
-0.01823173090815544,
0.09366928040981293,
0.02194373682141304,
-0.2344917505979538,
-0.14181962609291077,
0.3148316442966461,
0.01497019175440073,
0.044251225888729095,
-0.05854544788599014,
0.054376598447561264,
-0.0972670167684555,
0.3220164179801941,
0.3822595477104187,
-0.10919993370771408,
-0.012421404011547565,
-0.054995931684970856,
0.026529822498559952,
-0.06727895140647888,
0.1466422975063324,
0.027039680629968643,
0.19114597141742706,
-0.09544327855110168,
0.009551765397191048,
-0.16460226476192474,
-0.027596892789006233,
-0.18674972653388977,
-0.12213287502527237,
0.10238183289766312,
-0.02843421883881092,
0.00970312487334013,
0.1377798616886139,
-0.1451220065355301,
0.20307034254074097,
-0.0973057970404625,
-0.11064338684082031,
-0.0838763639330864,
-0.06061172857880592,
-0.05831148102879524,
0.07210878282785416,
0.11835763603448868,
-0.03170957416296005,
0.017184114083647728,
0.006176309660077095,
0.05678124353289604,
-0.13912534713745117,
-0.046293534338474274,
0.17125077545642853,
0.053332291543483734,
0.02395331673324108,
-0.002228943631052971,
0.07690773159265518,
0.06129154562950134,
0.044584207236766815,
-0.022030437365174294,
0.16790436208248138,
-0.054892707616090775,
0.11412505805492401,
0.04063874110579491,
0.06762026250362396,
-0.07903081178665161,
-0.10690672695636749,
-0.02685241959989071,
-0.010762439109385014,
0.07963492721319199,
-0.16988950967788696,
-0.008411156944930553,
0.011996966786682606,
0.11914410442113876,
-0.08325222879648209,
0.08733963966369629,
0.1527208834886551,
-0.0163130946457386,
-0.0600128136575222,
-0.017280222848057747,
0.0455138124525547,
-0.0498654805123806,
-0.1399865597486496,
-0.13342638313770294,
-0.1232331246137619,
-0.05624934658408165,
-0.11700336635112762,
0.015404727309942245,
-0.1885865330696106,
-0.0600074902176857,
-0.08385811001062393,
-0.03402049466967583,
-0.04960833862423897,
0.05852606147527695,
0.10655941069126129,
0.03948606923222542,
-0.022349946200847626,
-0.06415057182312012,
0.08333013206720352,
0.05577815696597099,
-0.20567737519741058,
-0.14850792288780212
] |
null | null |
transformers
|
## About the Model
An English sequence classification model, trained on MBAD Dataset to detect bias and fairness in sentences (news articles). This model was built on top of distilbert-base-uncased model and trained for 30 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.
- Dataset : MBAD Data
- Carbon emission 0.319355 Kg
| Train Accuracy | Validation Accuracy | Train loss | Test loss |
|---------------:| -------------------:| ----------:|----------:|
| 76.97 | 62.00 | 0.45 | 0.96 |
## Usage
The easiest way is to load the inference api from huggingface and second method is through the pipeline object offered by transformers library.
```python
from transformers import AutoTokenizer, TFAutoModelForSequenceClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("d4data/bias-detection-model")
model = TFAutoModelForSequenceClassification.from_pretrained("d4data/bias-detection-model")
classifier = pipeline('text-classification', model=model, tokenizer=tokenizer) # cuda = 0,1 based on gpu availability
classifier("The irony, of course, is that the exhibit that invites people to throw trash at vacuuming Ivanka Trump lookalike reflects every stereotype feminists claim to stand against, oversexualizing Ivanka’s body and ignoring her hard work.")
```
## Author
This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at:
> Bias & Fairness in AI, (2022), GitHub repository, <https://github.com/dreji18/Fairness-in-AI>
|
{"language": ["en"], "tags": ["Text Classification"], "co2_eq_emissions": 0.319355, "widget": [{"text": "Nevertheless, Trump and other Republicans have tarred the protests as havens for terrorists intent on destroying property.", "example_title": "Biased example 1"}, {"text": "Billie Eilish issues apology for mouthing an anti-Asian derogatory term in a resurfaced video.", "example_title": "Biased example 2"}, {"text": "Christians should make clear that the perpetuation of objectionable vaccines and the lack of alternatives is a kind of coercion.", "example_title": "Biased example 3"}, {"text": "There have been a protest by a group of people", "example_title": "Non-Biased example 1"}, {"text": "While emphasizing he\u2019s not singling out either party, Cohen warned about the danger of normalizing white supremacist ideology.", "example_title": "Non-Biased example 2"}]}
|
text-classification
|
d4data/bias-detection-model
|
[
"transformers",
"tf",
"distilbert",
"text-classification",
"Text Classification",
"en",
"co2_eq_emissions",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us
|
About the Model
---------------
An English sequence classification model, trained on MBAD Dataset to detect bias and fairness in sentences (news articles). This model was built on top of distilbert-base-uncased model and trained for 30 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.
* Dataset : MBAD Data
* Carbon emission 0.319355 Kg
Usage
-----
The easiest way is to load the inference api from huggingface and second method is through the pipeline object offered by transformers library.
Author
------
This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at:
>
> Bias & Fairness in AI, (2022), GitHub repository, <URL
>
>
>
|
[] |
[
"TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
56
] |
[
"passage: TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n"
] |
[
0.0030447407625615597,
0.03561275824904442,
-0.003967563156038523,
0.002684403443709016,
0.09756896644830704,
0.03305675834417343,
0.048824556171894073,
0.12801195681095123,
0.004523457959294319,
-0.019541088491678238,
0.12156841158866882,
0.12630324065685272,
-0.06759809702634811,
0.1782236546278,
-0.126469686627388,
-0.2627699673175812,
0.09936679899692535,
0.0029011210426688194,
-0.09142903238534927,
0.09592537581920624,
0.13408708572387695,
-0.08788714557886124,
0.07706662267446518,
-0.02140623889863491,
-0.14296159148216248,
0.04155931994318962,
0.005567420739680529,
-0.1451527625322342,
0.12652148306369781,
0.14326654374599457,
0.17194248735904694,
0.09678243845701218,
-0.05411718785762787,
-0.15076880156993866,
0.039128534495830536,
0.012777740135788918,
-0.15161707997322083,
0.07291942089796066,
0.006335156038403511,
-0.10108063369989395,
0.05753665044903755,
-0.021811941638588905,
0.026889801025390625,
0.04806036129593849,
-0.14684799313545227,
-0.04409598559141159,
0.019228627905249596,
-0.05425991863012314,
0.05815134942531586,
0.049683649092912674,
-0.021508971229195595,
0.1576307862997055,
-0.19024507701396942,
0.07559267431497574,
0.09761729091405869,
-0.18009695410728455,
0.002716513117775321,
0.08809498697519302,
0.01801876537501812,
0.031610842794179916,
-0.08693847060203552,
0.09159732609987259,
0.08004726469516754,
-0.023128308355808258,
0.06726734340190887,
-0.050405681133270264,
-0.13078929483890533,
0.06456542760133743,
-0.11197783052921295,
-0.011037250980734825,
0.26318371295928955,
0.011916805058717728,
0.03963487222790718,
-0.020938511937856674,
-0.1258339285850525,
-0.10700535774230957,
0.017341673374176025,
-0.006899161264300346,
-0.021580180153250694,
0.04723193496465683,
0.046958763152360916,
0.028744667768478394,
-0.17049813270568848,
0.05253027006983757,
-0.22258786857128143,
0.1832467019557953,
-0.06044076383113861,
-0.00702377874404192,
-0.13008683919906616,
0.07022372633218765,
-0.046192120760679245,
-0.10573878139257431,
0.014253904111683369,
-0.09648748487234116,
-0.05518491938710213,
-0.03634701296687126,
-0.08533494174480438,
-0.008581886067986488,
0.06512907147407532,
0.149742990732193,
0.024938706308603287,
0.010544976219534874,
0.038877684623003006,
0.07558374106884003,
0.06862390786409378,
0.12674999237060547,
-0.015392402186989784,
-0.07372575253248215,
-0.000713519228156656,
-0.14053243398666382,
-0.0033616607543081045,
-0.1158163994550705,
-0.18509183824062347,
-0.06771791726350784,
0.03348708152770996,
0.038883455097675323,
0.025843270123004913,
0.057730235159397125,
-0.07846831530332565,
0.020994039252400398,
0.12722396850585938,
-0.03778969496488571,
0.03465494140982628,
-0.013876206241548061,
0.07767874747514725,
0.045427385717630386,
0.03367210924625397,
0.03809591010212898,
0.018858063966035843,
0.11091658473014832,
-0.08701818436384201,
-0.04121566191315651,
0.010687236674129963,
-0.10727278143167496,
0.036869049072265625,
-0.1689022332429886,
0.06902115792036057,
-0.22945454716682434,
0.05547450855374336,
0.0321432426571846,
-0.0120092099532485,
-0.024373561143875122,
-0.03712425380945206,
0.00556494016200304,
-0.06527160853147507,
0.09767980128526688,
-0.02088155411183834,
-0.03304490074515343,
-0.09601449966430664,
0.013633857481181622,
-0.06739090383052826,
0.08921010047197342,
-0.21043157577514648,
0.03741598129272461,
-0.04572655260562897,
-0.01640142872929573,
-0.14512185752391815,
0.052970338612794876,
-0.018074296414852142,
0.12357403337955475,
-0.040601711720228195,
-0.032524507492780685,
-0.09457866102457047,
0.04140644147992134,
-0.05257413163781166,
0.1856638640165329,
-0.15960383415222168,
-0.08001638948917389,
0.051270488649606705,
-0.06918253004550934,
-0.10409118980169296,
0.08524760603904724,
-0.0023367037065327168,
0.041258517652750015,
0.02013900689780712,
0.13877539336681366,
0.09745793789625168,
-0.029503101482987404,
0.016821812838315964,
0.18721766769886017,
-0.05442248657345772,
0.0009334777714684606,
0.06431844830513,
0.008717769756913185,
-0.0342990905046463,
-0.01926446333527565,
0.13445231318473816,
0.06892441213130951,
-0.04398589953780174,
-0.07752135396003723,
-0.03521163761615753,
0.021037979051470757,
0.13122549653053284,
-0.03826991096138954,
0.09934379160404205,
-0.08157957345247269,
-0.07137472927570343,
0.010078543797135353,
-0.01136819552630186,
0.024818211793899536,
0.05282682925462723,
-0.06836940348148346,
0.09304479509592056,
-0.017679596319794655,
-0.0044711134396493435,
-0.16164633631706238,
-0.11458364129066467,
-0.03337197005748749,
0.1500028669834137,
-0.011598884128034115,
0.21495462954044342,
0.01009464357048273,
-0.0667935386300087,
-0.042998433113098145,
-0.007789708208292723,
0.08050685375928879,
0.07069359719753265,
-0.10684817284345627,
-0.140066996216774,
0.0818798691034317,
-0.08685595542192459,
0.09800353646278381,
-0.08677232265472412,
0.025231724604964256,
0.22852303087711334,
0.1222604364156723,
0.016582446172833443,
0.12246667593717575,
0.015323867090046406,
0.0042299251072108746,
-0.12013501673936844,
-0.0002012588083744049,
0.0369831882417202,
-0.026491466909646988,
-0.1997324675321579,
0.18632403016090393,
-0.13404302299022675,
0.29689866304397583,
0.2191237062215805,
-0.19935975968837738,
-0.06336968392133713,
0.01684598997235298,
-0.009815037250518799,
0.044170476496219635,
-0.026177553460001945,
-0.08391810953617096,
0.03871358186006546,
-0.06773178279399872,
0.11119136214256287,
-0.047754764556884766,
-0.06005491316318512,
0.021524429321289062,
-0.006712525617331266,
-0.05856316164135933,
0.08771272003650665,
0.05071290209889412,
-0.2232901006937027,
0.1669207215309143,
0.3063679039478302,
0.03229040279984474,
0.14505162835121155,
0.018494969233870506,
0.010525359772145748,
0.07153989374637604,
-0.07009445875883102,
-0.0227091945707798,
0.02983943559229374,
-0.16274456679821014,
-0.0012598182074725628,
0.08289109915494919,
0.025264602154493332,
0.05292427912354469,
-0.13099448382854462,
-0.04610361158847809,
0.05480000004172325,
0.037074074149131775,
-0.0657232254743576,
0.09920943528413773,
0.015757234767079353,
0.13252784311771393,
0.003881894052028656,
-0.10069962590932846,
0.08822001516819,
0.016515975818037987,
-0.10068555176258087,
0.2102946639060974,
-0.16477222740650177,
-0.357083261013031,
-0.11167983710765839,
-0.03467564657330513,
0.04270555078983307,
0.0764370933175087,
0.1251467615365982,
-0.13060159981250763,
-0.08809252083301544,
-0.03781290352344513,
0.04193385690450668,
-0.06754764169454575,
0.14487019181251526,
-0.07124979794025421,
0.023656301200389862,
-0.07252122461795807,
-0.09728502482175827,
-0.03742346540093422,
0.020209575071930885,
0.06277142465114594,
0.17064312100410461,
-0.12329518049955368,
0.1238238736987114,
0.21825431287288666,
-0.03671414405107498,
-0.009631489403545856,
-0.03552759066224098,
0.26180997490882874,
-0.15508738160133362,
0.06448381394147873,
0.10212166607379913,
-0.04427342116832733,
0.02132003754377365,
0.17291635274887085,
-0.0013729145284742117,
-0.08028256148099899,
0.06548487395048141,
-0.014059865847229958,
-0.11808150261640549,
-0.1972600817680359,
-0.12737207114696503,
-0.11811349540948868,
0.004956601653248072,
0.00017553767247591168,
0.08646691590547562,
0.15346549451351166,
0.06403986364603043,
-0.004370761103928089,
0.019686032086610794,
-0.0355847105383873,
0.06621620804071426,
0.1664443165063858,
-0.0006922989850863814,
0.1394825130701065,
-0.08571051061153412,
-0.13044239580631256,
0.13673073053359985,
0.009259702637791634,
0.06393202394247055,
0.12305624037981033,
0.07061677426099777,
-0.004863244481384754,
-0.0020029842853546143,
0.18924979865550995,
0.10957886278629303,
0.06180952861905098,
-0.045447930693626404,
-0.06153848022222519,
-0.028836525976657867,
0.03836716338992119,
0.108271025121212,
0.12745170295238495,
-0.1560208797454834,
-0.04737381264567375,
-0.10080402344465256,
0.14455485343933105,
-0.019931213930249214,
0.10663782805204391,
-0.19318506121635437,
0.011806810274720192,
0.12121964991092682,
-0.031210802495479584,
-0.05931861326098442,
0.018893906846642494,
0.08111871033906937,
-0.11036419868469238,
0.08813466876745224,
0.04075728356838226,
0.09476571530103683,
0.03681543096899986,
0.07068170607089996,
-0.07353818416595459,
-0.15256674587726593,
-0.03304918855428696,
0.06361258029937744,
-0.28317368030548096,
0.22242747247219086,
0.028691334649920464,
-0.13894613087177277,
-0.07815413177013397,
-0.010876333341002464,
0.020346343517303467,
0.2730488181114197,
0.10298702865839005,
0.018780041486024857,
-0.05589989200234413,
-0.07048134505748749,
-0.03989514335989952,
0.011784877628087997,
0.12061217427253723,
-0.07507295161485672,
-0.00958926323801279,
-0.007205376401543617,
-0.015852494165301323,
0.06555011868476868,
0.09080211818218231,
0.0016149356961250305,
-0.18333517014980316,
0.10431662946939468,
0.05640332028269768,
0.01732015050947666,
0.009756878949701786,
-0.07158699631690979,
-0.11494877189397812,
0.16323480010032654,
-0.1392003446817398,
-0.034571465104818344,
-0.17626701295375824,
-0.07591486722230911,
0.08612381666898727,
-0.04979781433939934,
0.08264794945716858,
-0.03998352587223053,
0.06358560919761658,
-0.06412811577320099,
-0.14935876429080963,
0.19121810793876648,
-0.09406676888465881,
-0.08808605372905731,
-0.048479728400707245,
0.11985311657190323,
-0.089913010597229,
0.08472684025764465,
0.018271273002028465,
0.041836533695459366,
-0.11856237798929214,
-0.13168443739414215,
-0.004295655060559511,
-0.04938856512308121,
0.06499574333429337,
0.04349742829799652,
-0.03333562985062599,
-0.06113317981362343,
0.041255708783864975,
-0.005855597089976072,
0.24694912135601044,
0.17752128839492798,
-0.10136143863201141,
0.13553497195243835,
0.028455577790737152,
-0.01752307452261448,
-0.3080705404281616,
0.006983303930610418,
-0.1178988441824913,
0.015510481782257557,
-0.05478774011135101,
-0.09479895979166031,
0.07188025116920471,
0.017448028549551964,
-0.027463743463158607,
0.06794582307338715,
-0.13677293062210083,
-0.139754056930542,
0.14790430665016174,
-0.08356745541095734,
0.3058595657348633,
-0.10392998158931732,
-0.03309420868754387,
-0.0963994711637497,
-0.1358947455883026,
0.1964462548494339,
-0.1476818025112152,
0.08089374005794525,
0.04437883570790291,
-0.0002093646617140621,
0.03678297623991966,
0.004117813427001238,
0.20785404741764069,
0.020674627274274826,
0.003784669330343604,
-0.060718901455402374,
-0.15030832588672638,
0.07562144100666046,
-0.01713521219789982,
-0.02440984733402729,
-0.044905804097652435,
0.0037609809078276157,
-0.15431824326515198,
-0.01961950771510601,
-0.028383711352944374,
0.09763988107442856,
-0.00565706891939044,
-0.08623595535755157,
-0.08437298238277435,
-0.009945720434188843,
0.04380376264452934,
-0.04918806254863739,
0.33725568652153015,
-0.07437264174222946,
0.13758453726768494,
0.08907350152730942,
0.18406818807125092,
-0.14467453956604004,
0.07122263312339783,
0.015623977407813072,
-0.058786433190107346,
0.06980963796377182,
-0.2144833505153656,
0.07477838546037674,
0.09643726050853729,
-0.06015926972031593,
0.06263947486877441,
0.12511736154556274,
0.027230622246861458,
-0.03552500158548355,
0.21577046811580658,
-0.18429940938949585,
-0.0004304990288801491,
-0.06221403554081917,
-0.07365942746400833,
0.028211914002895355,
0.05859215557575226,
0.13802923262119293,
0.03212606906890869,
-0.0037002742756158113,
0.02950635552406311,
-0.0451701395213604,
-0.053093887865543365,
0.08158683776855469,
0.0961579978466034,
0.06430510431528091,
-0.11676790565252304,
0.03829602152109146,
0.06230535730719566,
-0.09062439948320389,
-0.019098615273833275,
0.07243547588586807,
-0.13427524268627167,
-0.14091700315475464,
0.005657303147017956,
0.038999609649181366,
-0.1935570389032364,
-0.0175799410790205,
-0.03651077672839165,
-0.1312839388847351,
0.059568677097558975,
0.31135818362236023,
0.1110461950302124,
0.10825495421886444,
-0.012085487134754658,
-0.04572417587041855,
0.03515525534749031,
-0.042650721967220306,
-0.0748450756072998,
0.04857912287116051,
-0.10559840500354767,
0.058037545531988144,
-0.03077862598001957,
0.15600989758968353,
-0.11488500237464905,
-0.011271881870925426,
-0.1685502529144287,
-0.020894385874271393,
-0.023338738828897476,
-0.012252120301127434,
-0.032317329198122025,
-0.044740986078977585,
-0.0025991806760430336,
-0.08191374689340591,
-0.059559416025877,
-0.047004468739032745,
-0.11367763578891754,
0.01688387431204319,
-0.03636320307850838,
0.06981465965509415,
-0.04162442684173584,
-0.03376287221908569,
0.06220081076025963,
-0.049075596034526825,
0.12682709097862244,
0.06035744771361351,
-0.048887692391872406,
0.08230532705783844,
-0.11096751689910889,
-0.06472818553447723,
0.15636053681373596,
0.004200085531920195,
0.09521651268005371,
0.016190115362405777,
0.0030111323576420546,
0.0023956119548529387,
0.030156200751662254,
0.07145892828702927,
0.0013793717371299863,
-0.08792637288570404,
0.07917313277721405,
-0.11488862335681915,
-0.17333751916885376,
-0.049340520054101944,
-0.05015132948756218,
0.09904062747955322,
-0.025797497481107712,
0.10273797065019608,
0.002979304175823927,
0.030520223081111908,
-0.02644912153482437,
0.04976718872785568,
-0.04649771749973297,
-0.16655005514621735,
-0.06230216845870018,
-0.05695953220129013,
0.02447040192782879,
-0.04736072197556496,
0.26963868737220764,
0.0264127254486084,
0.06341100484132767,
0.07939191162586212,
0.05949743837118149,
0.027587521821260452,
0.042532969266176224,
0.17856641113758087,
0.12926986813545227,
-0.042389366775751114,
-0.08185673505067825,
0.04509705677628517,
0.07694289833307266,
-0.044307611882686615,
0.16385933756828308,
-0.004745738115161657,
-0.028212793171405792,
0.1619081199169159,
-0.05489213019609451,
-0.06615646928548813,
-0.0660858079791069,
-0.06144105643033981,
-0.02606404386460781,
0.030002271756529808,
0.002137415576726198,
-0.05653412267565727,
0.15304559469223022,
-0.08933818340301514,
0.0289126206189394,
-0.020964285358786583,
-0.03956152871251106,
-0.20027826726436615,
-0.006761083845049143,
-0.10913234204053879,
-0.10122769325971603,
-0.039936549961566925,
-0.11713333427906036,
-0.021713102236390114,
0.023173460736870766,
0.05882539227604866,
-0.007694044150412083,
0.09698870778083801,
-0.03018326312303543,
-0.06540556997060776,
0.07066706568002701,
0.01935192197561264,
0.07416235655546188,
-0.0801234245300293,
-0.03490458428859711,
-0.07598550617694855,
0.029604848474264145,
-0.07977797091007233,
0.007197295781224966,
-0.01153017207980156,
0.003562238300219178,
-0.11733362823724747,
-0.09972032159566879,
-0.04984261095523834,
0.03845733031630516,
-0.05379603058099747,
0.07126396149396896,
0.007890703156590462,
-0.005018569063395262,
0.0641038790345192,
0.202226459980011,
-0.09354110062122345,
-0.031231531873345375,
-0.08412358909845352,
0.13310369849205017,
0.03651483356952667,
0.15110063552856445,
-0.009868825785815716,
-0.07745393365621567,
-0.1194368228316307,
0.16777972877025604,
0.2668342888355255,
-0.09160645306110382,
0.09658266603946686,
0.023202989250421524,
0.03739716112613678,
0.14158862829208374,
0.053392939269542694,
0.08685697615146637,
0.15230131149291992,
-0.0682278648018837,
-0.08057594299316406,
-0.002351611154153943,
-0.0032608509063720703,
-0.039066724479198456,
0.12082313001155853,
0.06041831150650978,
-0.05469052121043205,
-0.050314761698246,
0.05663662776350975,
-0.14111703634262085,
0.02875680848956108,
0.008586859330534935,
-0.26574620604515076,
-0.07403090596199036,
-0.00399219011887908,
0.02810146100819111,
-0.03931616246700287,
0.07404740899801254,
-0.03388611227273941,
-0.1369231641292572,
-0.01792093552649021,
0.054841794073581696,
-0.2116173356771469,
0.018501151353120804,
0.07188821583986282,
-0.04803735390305519,
-0.02235432341694832,
-0.03459685295820236,
0.009584855288267136,
0.10308630764484406,
0.08371817320585251,
-0.005952516105026007,
0.0967588722705841,
0.04329187795519829,
-0.03360048308968544,
0.019538938999176025,
0.008373145945370197,
-0.005120305344462395,
-0.020674817264080048,
0.10427214205265045,
-0.17236445844173431,
0.0754123106598854,
-0.012249289080500603,
-0.05253284052014351,
-0.022776944562792778,
-0.029909860342741013,
-0.10842640697956085,
0.08185844123363495,
0.09163805097341537,
-0.03337012603878975,
-0.0452372282743454,
0.0033983378671109676,
-0.0004710262001026422,
0.016706841066479683,
-0.13970206677913666,
-0.1083764135837555,
-0.05217093974351883,
-0.01956351287662983,
0.11455786973237991,
0.02631353959441185,
-0.09206452965736389,
-0.021382393315434456,
-0.07599169760942459,
0.08621811866760254,
-0.10824082046747208,
0.08129555732011795,
0.04368048161268234,
-0.007121002767235041,
-0.031804971396923065,
-0.1065235584974289,
0.04479262977838516,
0.04491589963436127,
-0.1251758188009262,
-0.09748663008213043
] |
null | null |
spacy
|
## About the Model
This model is trained on MBAD Dataset to recognize the biased word/phrases in a sentence. This model was built on top of roberta-base offered by Spacy transformers.
This model is in association with https://huggingface.co/d4data/bias-detection-model
| Feature | Description |
| --- | --- |
| **Name** | `Bias Recognizer Model` |
| **Version** | `1.0` |
| **spaCy** | `>=3.2.1,<3.3.0` |
| **Default Pipeline** | `transformer`, `ner` |
| **Components** | `transformer`, `ner` |
## Author
This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at:
> Bias & Fairness in AI, (2022), GitHub repository, <https://github.com/dreji18/Fairness-in-AI>
|
{"language": ["en"], "tags": ["spacy", "token-classification"], "widget": [{"text": "Billie Eilish issues apology for mouthing an anti-Asian derogatory term in a resurfaced video.", "example_title": "Biased example 1"}, {"text": "Christians should make clear that the perpetuation of objectionable vaccines and the lack of alternatives is a kind of coercion.", "example_title": "Biased example 2"}, {"text": "But, whether this switch constitutes a true win for the racist right or not, it\u2019s clear that MAGA conservatives are highly attuned to how decisions are made in the White House and which positions they want to control.", "example_title": "Biased example 3"}, {"text": "The fact that the abortion rate among American blacks is far higher than the rate for whites is routinely chronicled and mourned.", "example_title": "Biased example 4"}]}
|
token-classification
|
d4data/en_pipeline
|
[
"spacy",
"token-classification",
"en",
"model-index",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#spacy #token-classification #en #model-index #region-us
|
About the Model
---------------
This model is trained on MBAD Dataset to recognize the biased word/phrases in a sentence. This model was built on top of roberta-base offered by Spacy transformers.
This model is in association with URL
Author
------
This model is part of the Research topic "Bias and Fairness in AI" conducted by Deepak John Reji, Shaina Raza. If you use this work (code, model or dataset), please star at:
>
> Bias & Fairness in AI, (2022), GitHub repository, <URL
>
>
>
|
[] |
[
"TAGS\n#spacy #token-classification #en #model-index #region-us \n"
] |
[
21
] |
[
"passage: TAGS\n#spacy #token-classification #en #model-index #region-us \n"
] |
[
-0.004639077931642532,
0.08970917761325836,
-0.011319623328745365,
-0.04456602409482002,
0.12429428845643997,
0.08450141549110413,
0.12109387665987015,
0.02920302376151085,
0.2465081661939621,
-0.04450484365224838,
0.1075228676199913,
0.05212441086769104,
-0.012718789279460907,
0.10401767492294312,
-0.04577607288956642,
-0.28677794337272644,
0.08648546040058136,
-0.036799605935811996,
-0.02966843545436859,
0.08118105679750443,
-0.0011295424774289131,
-0.09963174164295197,
0.04967942088842392,
-0.05129582807421684,
-0.05529729276895523,
0.05331800878047943,
0.02475965954363346,
-0.05978570133447647,
0.07705297321081161,
-0.05064434930682182,
0.20552772283554077,
0.03326057270169258,
-0.013609850779175758,
-0.23744425177574158,
0.026697730645537376,
0.013615494593977928,
-0.05840002000331879,
0.07731287181377411,
0.09350038319826126,
-0.06338433176279068,
0.0036208995152264833,
0.0380440317094326,
0.04783717170357704,
0.04987424239516258,
-0.21513299643993378,
-0.11451633274555206,
-0.009212800301611423,
-0.04623694717884064,
0.0597614087164402,
-0.038287725299596786,
0.011695273220539093,
0.1609041839838028,
-0.20021720230579376,
-0.006134488619863987,
0.02089294046163559,
-0.16215701401233673,
0.024579916149377823,
0.22164592146873474,
-0.012725578621029854,
0.10176205635070801,
-0.01484577264636755,
0.10037428885698318,
0.049146804958581924,
-0.032992903143167496,
-0.037102993577718735,
-0.03915470466017723,
0.0082062017172575,
0.10865304619073868,
-0.09575759619474411,
-0.045462388545274734,
0.2971842288970947,
0.06617727875709534,
0.04645711928606033,
0.04059813916683197,
-0.03266412392258644,
-0.015032447874546051,
0.02249535731971264,
-0.016940826550126076,
0.04289291054010391,
0.10213420540094376,
0.22040718793869019,
-0.001251775654964149,
-0.09906679391860962,
0.0016000795876607299,
-0.19151072204113007,
0.2202644795179367,
-0.015638789162039757,
0.10112698376178741,
-0.25754842162132263,
-0.0145059023052454,
-0.12953519821166992,
-0.05042881518602371,
0.06190278008580208,
-0.14057298004627228,
-0.061582960188388824,
-0.048871591687202454,
0.03981878608465195,
0.04016279801726341,
0.07962732017040253,
0.09794528782367706,
-0.047820210456848145,
0.08675193786621094,
0.00498546939343214,
0.12064965069293976,
0.1089821457862854,
0.09155845642089844,
0.0026791237760335207,
0.03822578489780426,
-0.04822041466832161,
-0.09568440169095993,
-0.03676281496882439,
-0.09267649054527283,
-0.08337756991386414,
0.00406569242477417,
-0.050524722784757614,
0.09551965445280075,
0.01818021945655346,
-0.035483673214912415,
-0.1067454144358635,
-0.016831988468766212,
0.04870854690670967,
-0.022051312029361725,
-0.00019913226424250752,
0.020300492644309998,
-0.013283330015838146,
0.0780373215675354,
-0.13371358811855316,
0.042349860072135925,
0.0634743794798851,
0.06287611275911331,
-0.1173018366098404,
-0.011305668391287327,
-0.04903577268123627,
-0.024753671139478683,
0.023300256580114365,
-0.14344896376132965,
0.05461607873439789,
-0.0731770396232605,
-0.09850872308015823,
0.033152904361486435,
0.04465179890394211,
-0.04716750979423523,
0.06489233672618866,
-0.0030849126633256674,
-0.0019773407839238644,
-0.0427132211625576,
-0.027221566066145897,
-0.06601788848638535,
-0.11597322672605515,
0.01127252820879221,
-0.08940979093313217,
0.025642456486821175,
-0.11787502467632294,
0.010194002650678158,
-0.08979786187410355,
0.08618950098752975,
-0.13546229898929596,
0.02606111951172352,
-0.05601274594664574,
0.17304077744483948,
-0.025323886424303055,
0.022280439734458923,
-0.13649441301822662,
0.018360543996095657,
-0.09457273781299591,
0.21585491299629211,
-0.15958209335803986,
-0.08008117973804474,
0.08282697945833206,
-0.12750215828418732,
-0.06157277524471283,
0.008730960078537464,
-0.008921684697270393,
0.033555518835783005,
0.06292471289634705,
0.36872267723083496,
-0.02239096909761429,
0.028824705630540848,
0.05541718378663063,
0.12799608707427979,
-0.15313154458999634,
-0.09711054712533951,
0.09516551345586777,
-0.03142009302973747,
-0.12918882071971893,
0.00782070867717266,
0.016833297908306122,
0.03994018957018852,
-0.0839722752571106,
-0.037850163877010345,
0.0670732706785202,
-0.011009217239916325,
0.13354110717773438,
0.06073354184627533,
0.09307627379894257,
-0.03658507019281387,
0.06287404894828796,
0.04370417073369026,
0.0390278659760952,
0.10642722249031067,
-0.07067322731018066,
-0.055674344301223755,
0.10760679841041565,
-0.05291156470775604,
-0.016980277374386787,
-0.17015394568443298,
-0.1337946057319641,
0.02425357885658741,
0.08296997100114822,
-0.007017239462584257,
0.13200882077217102,
0.08337831497192383,
-0.09470215439796448,
0.015831025317311287,
0.004940938204526901,
0.10201597213745117,
0.056256361305713654,
-0.012591895647346973,
-0.14045414328575134,
0.013452124781906605,
-0.08189942687749863,
-0.03761918470263481,
-0.13976970314979553,
0.0022865014616400003,
0.09205739945173264,
0.09351968765258789,
0.058606091886758804,
0.04304945841431618,
0.048122573643922806,
0.004158638883382082,
0.02548038959503174,
-0.020121674984693527,
0.0848366916179657,
-0.09043246507644653,
-0.0689830556511879,
0.05973832681775093,
-0.0394069068133831,
0.25615110993385315,
0.13620048761367798,
-0.21579033136367798,
0.033656034618616104,
-0.05090385302901268,
0.022952359169721603,
0.04566274583339691,
0.02530262991786003,
0.020270781591534615,
0.05855812877416611,
0.013455922715365887,
0.036737214773893356,
-0.030585456639528275,
-0.001125612878240645,
-0.021057093515992165,
-0.028102826327085495,
-0.12819987535476685,
0.12933386862277985,
0.17221443355083466,
-0.2023140788078308,
0.13022613525390625,
0.26860547065734863,
0.15136657655239105,
0.1639244109392166,
-0.05466530844569206,
-0.0021928204223513603,
0.029909050092101097,
-0.05984814465045929,
-0.07768815755844116,
0.11841558665037155,
-0.16073152422904968,
-0.028791308403015137,
0.034306325018405914,
0.040091656148433685,
0.06906545907258987,
-0.12994487583637238,
-0.11471781134605408,
0.003427002113312483,
0.0026880980003625154,
-0.16119849681854248,
0.021833928301930428,
-0.013583927415311337,
0.0749448612332344,
0.037585701793432236,
-0.10483847558498383,
0.06550408154726028,
-0.04793604835867882,
-0.06524862349033356,
0.14126069843769073,
-0.10965844988822937,
-0.15786713361740112,
-0.04928029701113701,
-0.04247290641069412,
0.008099687285721302,
0.03575556352734566,
-0.041765131056308746,
-0.18511976301670074,
-0.024054406210780144,
0.04566461965441704,
0.025784475728869438,
-0.19681578874588013,
0.028612909838557243,
-0.03096773475408554,
0.11105889081954956,
-0.11762578785419464,
-0.007232296280562878,
-0.05351674184203148,
-0.10171082615852356,
0.1137947291135788,
0.08427844941616058,
-0.16281627118587494,
0.03662322089076042,
0.22456440329551697,
0.021948300302028656,
0.06868069618940353,
0.0017532092751935124,
0.17346027493476868,
-0.12610088288784027,
-0.03955264762043953,
0.11882825940847397,
-0.03700146824121475,
0.039294857531785965,
0.20534248650074005,
0.0891692191362381,
-0.09006188064813614,
-0.06507369130849838,
-0.03757299482822418,
-0.11004941910505295,
-0.1988985389471054,
-0.10496654361486435,
-0.06669024378061295,
0.06219172105193138,
0.009807383641600609,
0.030124573037028313,
0.08064200729131699,
0.03555005043745041,
0.09797866642475128,
-0.0818861871957779,
-0.04048384726047516,
-0.013325564563274384,
0.044969938695430756,
-0.02216709777712822,
0.023401616141200066,
-0.05851900950074196,
-0.07130870223045349,
0.0966208353638649,
0.03682345896959305,
0.12775343656539917,
0.24501833319664001,
0.009995551779866219,
0.022519417107105255,
0.03113345056772232,
0.18061673641204834,
0.08198492228984833,
0.06316947191953659,
-0.026204030960798264,
-0.033824242651462555,
-0.012286003679037094,
0.0019295847741886973,
0.04636871814727783,
0.10154115408658981,
-0.18859173357486725,
-0.030302047729492188,
-0.1503351628780365,
0.07390563935041428,
-0.08441091328859329,
0.13698714971542358,
-0.12238597124814987,
0.10013417899608612,
0.12533064186573029,
0.03783004358410835,
-0.03326953575015068,
0.09254541248083115,
0.01604602299630642,
-0.10436208546161652,
0.06574864685535431,
0.05468543991446495,
0.12417032569646835,
-0.07954820990562439,
0.08784982562065125,
-0.05414898321032524,
-0.06979965418577194,
-0.0021282124798744917,
0.06781022250652313,
-0.03815697133541107,
0.34460777044296265,
0.03927534818649292,
-0.1322224885225296,
-0.0680960938334465,
-0.07073813676834106,
0.04339425638318062,
0.24450771510601044,
0.13076695799827576,
0.045797597616910934,
-0.1860412210226059,
-0.1423356831073761,
-0.011661691591143608,
0.01112963818013668,
0.09137939661741257,
-0.08778875321149826,
-0.09258304536342621,
0.012080238200724125,
0.038386955857276917,
-0.02470325492322445,
-0.02574688009917736,
0.013626595959067345,
-0.010869069024920464,
0.0028962241485714912,
-0.0024593377020210028,
-0.09494844079017639,
0.04063175991177559,
-0.008970282040536404,
-0.0426364429295063,
0.027255645021796227,
0.015760518610477448,
-0.0477132610976696,
-0.0862787738442421,
-0.11995428055524826,
0.12820406258106232,
-0.05647685378789902,
-0.031126320362091064,
-0.04853896051645279,
-0.0242247823625803,
-0.008871877565979958,
-0.17192263901233673,
0.1168031245470047,
-0.04483760893344879,
0.06089157983660698,
-0.07858055830001831,
0.08671921491622925,
-0.06555803120136261,
0.03196974843740463,
-0.006401605438441038,
0.06907150149345398,
-0.026392577216029167,
-0.10749984532594681,
0.13508334755897522,
-0.0888948142528534,
0.00805789977312088,
0.15703195333480835,
0.009936179034411907,
0.01910584792494774,
0.022369487211108208,
0.01389949582517147,
0.1242658719420433,
0.3219046890735626,
-0.050361569970846176,
0.10433299839496613,
0.19669955968856812,
-0.055160827934741974,
-0.23867636919021606,
-0.004658831749111414,
-0.18856526911258698,
-0.026906609535217285,
0.08217744529247284,
-0.207087442278862,
0.12991946935653687,
0.07129856199026108,
-0.06354179233312607,
0.1380489021539688,
-0.15917861461639404,
-0.01845533214509487,
0.19913698732852936,
-0.06235973909497261,
0.4086400270462036,
-0.09474170953035355,
-0.11853466928005219,
-0.02225828729569912,
-0.03049946017563343,
0.13720571994781494,
-0.0674581304192543,
0.01762324571609497,
0.01025628112256527,
-0.06496995687484741,
0.05425776168704033,
-0.022221975028514862,
0.23985937237739563,
0.015817392617464066,
0.09306802600622177,
-0.050412751734256744,
-0.26644742488861084,
0.09111132472753525,
-0.020904146134853363,
-0.05758862569928169,
0.10420601814985275,
0.007044668309390545,
-0.18158294260501862,
0.009055105037987232,
-0.04800739884376526,
-0.0020014476031064987,
0.05716131255030632,
-0.07191526144742966,
-0.05414356663823128,
0.045388806611299515,
-0.08195486664772034,
0.0025779185816645622,
0.28824156522750854,
-0.04370611533522606,
0.11045469343662262,
0.016353119164705276,
0.10725726932287216,
-0.14538061618804932,
0.014544065110385418,
-0.05307826027274132,
-0.028697047382593155,
0.07712770253419876,
-0.1837805211544037,
0.011848709546029568,
0.13038933277130127,
-0.026828432455658913,
0.04118085280060768,
0.1095285639166832,
-0.019497469067573547,
-0.047768451273441315,
0.18089185655117035,
-0.135270893573761,
-0.12138840556144714,
-0.04492122307419777,
-0.11153297126293182,
0.08224985003471375,
0.038569483906030655,
0.04619360342621803,
0.08390957117080688,
0.008496037684381008,
0.030369482934474945,
-0.016053883358836174,
-0.0688047856092453,
-0.007411373779177666,
0.08882346004247665,
0.030600685626268387,
-0.09679323434829712,
0.14854951202869415,
0.09886171668767929,
0.043697789311409,
-0.0855928435921669,
0.10585042834281921,
-0.11350740492343903,
-0.07684192806482315,
-0.10012117028236389,
0.15077625215053558,
-0.03609536588191986,
-0.043035950511693954,
-0.026238704100251198,
-0.09582386910915375,
-0.014815926551818848,
0.1275714486837387,
0.07660308480262756,
0.0581178292632103,
-0.001612194231711328,
-0.07449325919151306,
0.12076675146818161,
-0.08139422535896301,
-0.10948988050222397,
-0.05866561084985733,
-0.1813315600156784,
-0.08627350628376007,
-0.026198316365480423,
0.1304904967546463,
-0.09884200990200043,
-0.10558386147022247,
-0.20356711745262146,
0.06893017143011093,
-0.08658204972743988,
-0.05228438228368759,
-0.04287975654006004,
-0.04149717837572098,
0.017714638262987137,
-0.06983821094036102,
-0.0608670711517334,
-0.07395796477794647,
-0.16240860521793365,
0.0937165841460228,
0.043504081666469574,
0.09737872332334518,
0.008909969590604305,
-0.034158989787101746,
0.12932902574539185,
-0.00030033348593860865,
0.09372983872890472,
0.07952453941106796,
0.00713045010343194,
0.12905433773994446,
-0.10803424566984177,
-0.06866864860057831,
0.09711814671754837,
-0.001475788070820272,
0.06869221478700638,
0.06021283194422722,
-0.04274054616689682,
-0.023911623284220695,
-0.008821733295917511,
0.10030528903007507,
-0.109192855656147,
-0.06137405335903168,
-0.022272862493991852,
-0.016095686703920364,
-0.18659444153308868,
0.02772687003016472,
-0.09598401933908463,
0.18725505471229553,
-0.006836308166384697,
0.043538156896829605,
0.12028458714485168,
0.07629495859146118,
-0.06461116671562195,
0.01902562938630581,
-0.0019335830584168434,
-0.13006001710891724,
0.05873112380504608,
-0.015072413720190525,
0.022259483113884926,
-0.04525398835539818,
0.3661469519138336,
0.08485667407512665,
-0.019175685942173004,
0.033395882695913315,
0.1745140701532364,
0.035372499376535416,
0.09521686285734177,
0.1510782241821289,
0.13052740693092346,
-0.05577104538679123,
-0.038019370287656784,
0.04543108865618706,
-0.009173748083412647,
0.008390256203711033,
0.13104186952114105,
0.11407414078712463,
-0.0039073084481060505,
0.039457883685827255,
0.03247899189591408,
0.009522458538413048,
0.0145067498087883,
-0.04954832047224045,
0.019522642716765404,
0.01868300326168537,
0.027216164395213127,
-0.008242812938988209,
0.145968496799469,
-0.06971205025911331,
0.09183305501937866,
-0.046941161155700684,
-0.021801888942718506,
-0.17487819492816925,
-0.1290074586868286,
-0.022331450134515762,
-0.09879794716835022,
0.038592275232076645,
-0.04632844775915146,
-0.0481463223695755,
0.2544984519481659,
0.04947135969996452,
0.013621468096971512,
-0.01711283251643181,
-0.053839243948459625,
-0.03156109154224396,
0.058983925729990005,
-0.006124243605881929,
-0.0014668542426079512,
-0.08417309075593948,
-0.05869605019688606,
-0.05446097254753113,
-0.0335603803396225,
-0.12128662317991257,
-0.026858549565076828,
-0.09391071647405624,
-0.033354151993989944,
-0.12941251695156097,
-0.1083754375576973,
-0.035169802606105804,
0.017676521092653275,
-0.1223810613155365,
0.017707597464323044,
-0.017873268574476242,
0.017050711438059807,
0.029801826924085617,
0.19557435810565948,
-0.0014115041121840477,
0.06454538553953171,
-0.057563018053770065,
0.12050914764404297,
-0.06442353874444962,
0.1215941533446312,
-0.03651757165789604,
-0.06381290405988693,
-0.054175443947315216,
0.2126113474369049,
0.31652209162712097,
-0.09200555086135864,
-0.004950030241161585,
-0.019949940964579582,
0.024400673806667328,
0.09764713048934937,
0.08890847116708755,
-0.035175107419490814,
0.12033504247665405,
-0.08252035826444626,
0.03625717759132385,
-0.00736491521820426,
-0.002577448496595025,
-0.0006376832607202232,
0.06541309505701065,
0.15297411382198334,
-0.03292388096451759,
-0.14801739156246185,
0.16778747737407684,
-0.20297187566757202,
0.1639130413532257,
0.05840076506137848,
-0.19709526002407074,
-0.09321457147598267,
-0.05363927036523819,
0.02414136379957199,
-0.049636438488960266,
0.09114091843366623,
-0.0883140116930008,
-0.13986757397651672,
-0.18235722184181213,
0.03678620606660843,
-0.3016493618488312,
-0.1992723047733307,
0.09568007290363312,
0.13831616938114166,
0.1138654351234436,
0.0008787679835222661,
0.028696678578853607,
0.014323273673653603,
-0.0016361117595806718,
-0.01864772103726864,
-0.04252839460968971,
0.018433915451169014,
0.038329001516103745,
-0.14162324368953705,
-0.008960560895502567,
0.004876873455941677,
-0.11105360835790634,
0.1013193354010582,
-0.12032376974821091,
0.007551189512014389,
-0.004324350506067276,
-0.11722799390554428,
0.028249593451619148,
0.022816652432084084,
-0.11798198521137238,
0.03353245183825493,
0.0774940699338913,
0.014099461026489735,
-0.028579184785485268,
-0.002242022193968296,
0.013397528789937496,
0.009455411694943905,
-0.13984155654907227,
-0.13611678779125214,
0.07265457510948181,
-0.06825058162212372,
0.13692006468772888,
-0.033100854605436325,
-0.031059378758072853,
0.0257404875010252,
-0.08474133163690567,
0.0929730013012886,
-0.03160342574119568,
0.044930122792720795,
0.1370825320482254,
0.02326572872698307,
-0.023153288289904594,
-0.13521593809127808,
0.09936809539794922,
0.0066788457334041595,
-0.04449629783630371,
-0.07006555795669556
] |
null | null |
transformers
|
## About the Model
An Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .
This model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.
- Dataset : Open Source News data + Custom data
- Carbon emission 0.1069 Kg
## Usage
The easiest way is to load through the pipeline object offered by transformers library.
```python
from transformers import AutoTokenizer, TFAutoModelForSequenceClassification
from transformers import pipeline
tokenizer = AutoTokenizer.from_pretrained("d4data/environmental-due-diligence-model")
model = TFAutoModelForSequenceClassification.from_pretrained("d4data/environmental-due-diligence-model")
classifier = pipeline('text-classification', model=model, tokenizer=tokenizer) # cuda = 0,1 based on gpu availability
classifier("At the every month post-injection monitoring event, TCE, carbon tetrachloride, and chloroform concentrations were above CBSGs in three of the wells")
```
## Author
This model is part of the Research topic "Environmental Due Diligence" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:
> Environmental Due Diligence, (2020), https://www.sciencedirect.com/science/article/pii/S2665963822001117
## You can support me here :)
<a href="https://www.buymeacoffee.com/deepakjohnreji" target="_blank"><img src="https://cdn.buymeacoffee.com/buttons/v2/default-yellow.png" alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
|
{"language": ["en"], "tags": ["Text Classification"], "co2_eq_emissions": 0.1069, "widget": [{"text": "At the every month post-injection monitoring event, TCE, carbon tetrachloride, and chloroform concentrations were above CBSGs in three of the wells", "example_title": "Remediation Standards"}, {"text": "TRPH exceedances were observed in the subsurface soils immediately above the water table and there are no TRPH exceedances in surface soils.", "example_title": "Extent of Contamination"}, {"text": "weathered shale was encountered below the surface area with fluvial deposits. Sediments in the coastal plain region are found above and below the bedrock with sandstones and shales that form the basement rock", "example_title": "Geology"}]}
|
text-classification
|
d4data/environmental-due-diligence-model
|
[
"transformers",
"tf",
"distilbert",
"text-classification",
"Text Classification",
"en",
"co2_eq_emissions",
"autotrain_compatible",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"en"
] |
TAGS
#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us
|
## About the Model
An Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .
This model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.
- Dataset : Open Source News data + Custom data
- Carbon emission 0.1069 Kg
## Usage
The easiest way is to load through the pipeline object offered by transformers library.
## Author
This model is part of the Research topic "Environmental Due Diligence" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:
> Environmental Due Diligence, (2020), URL
## You can support me here :)
<a href="URL target="_blank"><img src="URL alt="Buy Me A Coffee" style="height: 60px !important;width: 217px !important;" ></a>
|
[
"## About the Model\nAn Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .\n\nThis model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.\n\n- Dataset : Open Source News data + Custom data\n- Carbon emission 0.1069 Kg",
"## Usage\nThe easiest way is to load through the pipeline object offered by transformers library.",
"## Author\nThis model is part of the Research topic \"Environmental Due Diligence\" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:\n> Environmental Due Diligence, (2020), URL",
"## You can support me here :)\n<a href=\"URL target=\"_blank\"><img src=\"URL alt=\"Buy Me A Coffee\" style=\"height: 60px !important;width: 217px !important;\" ></a>"
] |
[
"TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n",
"## About the Model\nAn Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .\n\nThis model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.\n\n- Dataset : Open Source News data + Custom data\n- Carbon emission 0.1069 Kg",
"## Usage\nThe easiest way is to load through the pipeline object offered by transformers library.",
"## Author\nThis model is part of the Research topic \"Environmental Due Diligence\" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:\n> Environmental Due Diligence, (2020), URL",
"## You can support me here :)\n<a href=\"URL target=\"_blank\"><img src=\"URL alt=\"Buy Me A Coffee\" style=\"height: 60px !important;width: 217px !important;\" ></a>"
] |
[
56,
170,
23,
64,
57
] |
[
"passage: TAGS\n#transformers #tf #distilbert #text-classification #Text Classification #en #co2_eq_emissions #autotrain_compatible #endpoints_compatible #has_space #region-us \n## About the Model\nAn Environmental due diligence classification model, trained on customized environmental Dataset to detect contamination and remediation activities (both prevailing as well as planned) as a part of site assessment process. This model can identify the source of contamination, the extent of contamination, the types of contaminants present at the site, the flow of contaminants and their interaction with ground water, surface water and other surrounding water bodies .\n\nThis model was built on top of distilbert-base-uncased model and trained for 10 epochs with a batch size of 16, a learning rate of 5e-5, and a maximum sequence length of 512.\n\n- Dataset : Open Source News data + Custom data\n- Carbon emission 0.1069 Kg## Usage\nThe easiest way is to load through the pipeline object offered by transformers library.## Author\nThis model is part of the Research topic \"Environmental Due Diligence\" conducted by Deepak John Reji, Afreen Aman. If you use this work (code, model or dataset), please cite as:\n> Environmental Due Diligence, (2020), URL## You can support me here :)\n<a href=\"URL target=\"_blank\"><img src=\"URL alt=\"Buy Me A Coffee\" style=\"height: 60px !important;width: 217px !important;\" ></a>"
] |
[
-0.04574066027998924,
0.15251892805099487,
-0.006146837025880814,
-0.02539285644888878,
0.010114723816514015,
-0.03171269968152046,
0.07111353427171707,
0.031913161277770996,
0.02520304173231125,
0.05894694849848747,
-0.001611691084690392,
0.06166495010256767,
0.048466041684150696,
0.12457770854234695,
-0.017021093517541885,
-0.19002367556095123,
0.040786392986774445,
-0.07400886714458466,
0.010709427297115326,
0.08649180829524994,
0.12043632566928864,
-0.12918032705783844,
0.10134986788034439,
-0.017200924456119537,
-0.017663389444351196,
0.044435981661081314,
-0.06212648004293442,
-0.04561123251914978,
0.07775898277759552,
0.07616012543439865,
0.15775227546691895,
0.09954197704792023,
0.022251907736063004,
-0.12729273736476898,
0.03173838555812836,
0.03867589682340622,
-0.018665794283151627,
0.0692150890827179,
0.06976564228534698,
-0.014325466006994247,
0.014644082635641098,
-0.08255019038915634,
0.04428892955183983,
0.03770134970545769,
-0.07370244711637497,
-0.07344599813222885,
-0.020073657855391502,
0.06924056261777878,
0.1501713991165161,
0.06084741652011871,
-0.026481354609131813,
0.0690721645951271,
-0.12640956044197083,
0.041200730949640274,
0.10648705810308456,
-0.1054440438747406,
-0.02060776948928833,
0.007930640131235123,
-0.006658782251179218,
0.09703019261360168,
-0.07099346071481705,
0.0025130247231572866,
-0.033968452364206314,
-0.02034963294863701,
0.01953071914613247,
-0.06726916879415512,
0.04353869706392288,
-0.004493996500968933,
-0.09410225600004196,
0.0326760858297348,
0.22215057909488678,
0.08564687520265579,
-0.06492218375205994,
-0.06726871430873871,
-0.0899399146437645,
0.06730841845273972,
0.018037809059023857,
-0.06647762656211853,
0.0482199601829052,
-0.014450157061219215,
0.012517480179667473,
-0.028959831222891808,
-0.1480647623538971,
0.027323376387357712,
-0.1069309189915657,
0.13914118707180023,
-0.0025813847314566374,
0.03781161829829216,
-0.020754894241690636,
0.12187818437814713,
-0.009065881371498108,
-0.09741107374429703,
-0.029819482937455177,
-0.06096413731575012,
-0.06979299336671829,
-0.03918179124593735,
0.031328748911619186,
-0.06386405974626541,
0.03465050086379051,
0.1280697137117386,
0.03689050301909447,
0.030269978567957878,
0.06449033319950104,
0.00514121912419796,
0.09584126621484756,
0.12409459799528122,
-0.11737014353275299,
-0.0852200910449028,
0.051549870520830154,
-0.04653343930840492,
-0.0005668335361406207,
-0.03660234063863754,
-0.05486835911870003,
-0.029879910871386528,
-0.01031565759330988,
0.06290937960147858,
0.06494821608066559,
-0.028829192742705345,
-0.014769148081541061,
-0.08349083364009857,
0.27164244651794434,
-0.12325001507997513,
0.0375208705663681,
0.06494524329900742,
-0.03894604369997978,
0.11188089102506638,
0.0184820294380188,
0.06169426068663597,
-0.04122185707092285,
0.09696130454540253,
-0.04201162979006767,
-0.06632892042398453,
-0.06422460824251175,
-0.10219153016805649,
-0.016350626945495605,
-0.05285278335213661,
-0.00043635134352371097,
-0.14088381826877594,
-0.08640777319669724,
-0.028470244258642197,
0.07183897495269775,
-0.014800316654145718,
-0.06914765387773514,
0.01483566127717495,
0.017257312312722206,
-0.018448470160365105,
0.04258464276790619,
-0.02698396146297455,
-0.012081420980393887,
-0.01993568055331707,
-0.09696568548679352,
0.029056215658783913,
-0.053903114050626755,
0.023735998198390007,
-0.051555510610342026,
0.02959594503045082,
-0.21352297067642212,
0.08912523090839386,
-0.011630861088633537,
0.09032371640205383,
-0.1902482956647873,
-0.02115107886493206,
-0.08390575647354126,
-0.046786919236183167,
-0.019199313595891,
0.13468122482299805,
-0.10097543150186539,
-0.050236184149980545,
-0.022790540009737015,
-0.05940592288970947,
0.010742787271738052,
0.02676469087600708,
-0.0007185001741163433,
0.07457577437162399,
0.1174575462937355,
-0.005151744931936264,
0.16804596781730652,
-0.12917518615722656,
-0.041161153465509415,
0.02503793314099312,
-0.0628400668501854,
0.07715930789709091,
0.09035305678844452,
-0.0664013996720314,
0.0869283601641655,
-0.02635079063475132,
-0.004896120168268681,
-0.008622057735919952,
0.015523879788815975,
-0.0836298018693924,
0.018857410177588463,
-0.0075299786403775215,
-0.027405578643083572,
-0.0811968669295311,
0.0004702853912021965,
0.03481777757406235,
-0.05366146191954613,
0.011975564062595367,
0.05836804583668709,
-0.023889854550361633,
0.018601810559630394,
-0.08265522867441177,
0.05628732964396477,
-0.020410221070051193,
-0.01545432023704052,
-0.1248493567109108,
-0.029396001249551773,
0.061826884746551514,
-0.10618068277835846,
0.009311976842582226,
0.011413322761654854,
0.040766820311546326,
0.010221855714917183,
-0.016518693417310715,
-0.04491233453154564,
-0.04672187939286232,
-0.01390603743493557,
-0.03711352497339249,
-0.1567113697528839,
0.044903941452503204,
0.012681938707828522,
0.09290537983179092,
-0.08235049992799759,
0.02005711756646633,
0.1566419154405594,
0.09120216965675354,
0.01958981156349182,
0.01709512248635292,
0.026851294562220573,
-0.0580076240003109,
0.0062921661883592606,
-0.01967637799680233,
-0.004842888563871384,
0.0003874910471495241,
-0.10239338874816895,
0.08150386065244675,
-0.09225215017795563,
-0.057316020131111145,
0.06550699472427368,
0.034971754997968674,
-0.046534519642591476,
0.030860159546136856,
-0.04444156587123871,
-0.051112789660692215,
-0.04772012308239937,
-0.07291451841592789,
0.16583245992660522,
0.058491095900535583,
0.019586583599448204,
-0.09697184711694717,
-0.06764780730009079,
0.0030652021523565054,
0.012540707364678383,
-0.008537637069821358,
0.09213162958621979,
0.07055628299713135,
-0.1919882893562317,
0.008316610008478165,
0.13165847957134247,
-0.012346128933131695,
0.11933065205812454,
0.008573206141591072,
-0.11763907968997955,
-0.033383190631866455,
0.023902250453829765,
-0.017126012593507767,
0.08658153563737869,
0.004207942634820938,
0.08018642663955688,
0.05204325541853905,
-0.01178748533129692,
0.02835139073431492,
-0.05615681782364845,
0.04825574904680252,
0.03334176540374756,
-0.02369464375078678,
-0.01218377985060215,
0.019213739782571793,
0.018231110647320747,
0.08686993271112442,
0.05067873373627663,
-0.0028709617909044027,
-0.005374182481318712,
-0.040234122425317764,
-0.11363348364830017,
0.21122588217258453,
-0.07736280560493469,
-0.19958265125751495,
-0.09226103127002716,
0.09569676965475082,
0.022098137065768242,
-0.04690944403409958,
0.03901106119155884,
-0.052972640842199326,
-0.11511509865522385,
-0.08548660576343536,
0.00008481266559101641,
-0.0038988685701042414,
-0.02141997590661049,
-0.08132874965667725,
0.0034935097210109234,
-0.04034121707081795,
-0.14254197478294373,
0.017313683405518532,
0.029784230515360832,
-0.07917699962854385,
0.07133261114358902,
-0.027561526745557785,
0.09087400138378143,
0.19381265342235565,
-0.059069402515888214,
-0.03287933021783829,
0.06563223153352737,
0.23536084592342377,
-0.07497134059667587,
0.2157038003206253,
0.10634487867355347,
-0.008686527609825134,
0.1232362762093544,
0.20019158720970154,
0.027471203356981277,
0.009610607288777828,
0.04202979430556297,
0.07293891161680222,
-0.007507536560297012,
-0.2683899998664856,
-0.02862696535885334,
-0.0012949573574587703,
-0.08959247916936874,
-0.022261030972003937,
0.0677807405591011,
0.04820893332362175,
0.0442638024687767,
-0.06739676743745804,
-0.009233905002474785,
0.06617529690265656,
0.07289603352546692,
0.012121142819523811,
0.025808095932006836,
0.05086859315633774,
-0.07525299489498138,
-0.046837083995342255,
0.11182648688554764,
0.02886536903679371,
0.2442592978477478,
0.05535789579153061,
0.16330991685390472,
0.08159395307302475,
-0.013076916337013245,
0.06544455885887146,
0.08031322807073593,
-0.03187863901257515,
0.010486578568816185,
-0.023013664409518242,
-0.09519580006599426,
0.12841014564037323,
0.09065964818000793,
0.1043134555220604,
-0.03936121612787247,
0.12869194149971008,
0.00521129509434104,
0.0686698704957962,
0.22204025089740753,
0.0108240507543087,
-0.119538813829422,
0.0015416473615914583,
0.06489486247301102,
-0.019384892657399178,
-0.06016543135046959,
-0.07444785535335541,
0.0900929719209671,
-0.12171205133199692,
0.012087303213775158,
0.0047754873521625996,
0.08157021552324295,
-0.04616151005029678,
-0.03432614356279373,
0.0007393746054731309,
0.03884008154273033,
-0.03382151201367378,
0.06785280257463455,
-0.1569320261478424,
0.1032557412981987,
0.040970079600811005,
0.019302034750580788,
-0.04555712267756462,
0.014035450294613838,
-0.0045867967419326305,
0.14293953776359558,
0.0908348560333252,
0.07396339625120163,
0.020994629710912704,
-0.028138628229498863,
-0.2258029580116272,
0.06273671239614487,
0.03907780349254608,
-0.13910338282585144,
0.017343567684292793,
0.056232523173093796,
-0.05111941322684288,
-0.0006445308099500835,
0.049238964915275574,
-0.2156134843826294,
-0.12308091670274734,
0.07585311681032181,
-0.06977231055498123,
0.04393265023827553,
-0.038636136800050735,
-0.013456580229103565,
0.045305974781513214,
0.15741804242134094,
-0.22934263944625854,
-0.12460778653621674,
-0.12553048133850098,
-0.16021373867988586,
0.10946594923734665,
-0.11719147861003876,
0.04789116233587265,
-0.04924233257770538,
0.09830271452665329,
0.001972418511286378,
-0.05344719812273979,
0.09598664939403534,
-0.03328228369355202,
-0.1299998164176941,
-0.048164427280426025,
0.09086762368679047,
0.15272800624370575,
0.03565841168165207,
-0.00528474897146225,
0.012492253445088863,
-0.09612254053354263,
-0.1478116661310196,
-0.014768947847187519,
0.12430638074874878,
0.043759435415267944,
0.037143684923648834,
-0.09714917093515396,
-0.1340739130973816,
-0.08316044509410858,
-0.03370065987110138,
0.1172633096575737,
0.10212422162294388,
-0.05392861366271973,
0.08161238580942154,
0.23922491073608398,
-0.05886942893266678,
-0.16763414442539215,
0.0014363809023052454,
0.009329890832304955,
0.04581621289253235,
-0.07136193662881851,
-0.2504485547542572,
0.0982336550951004,
0.10204783827066422,
-0.03136717528104782,
0.09421618282794952,
-0.1962154358625412,
-0.16765202581882477,
0.10345232486724854,
0.02862033061683178,
0.038599979132413864,
-0.10490567982196808,
-0.05683719739317894,
-0.07332302629947662,
-0.13308298587799072,
0.08350306749343872,
-0.12249304354190826,
0.01563127152621746,
0.03922140598297119,
0.06681521236896515,
-0.00869763270020485,
-0.03408339247107506,
0.14909771084785461,
0.06965603679418564,
0.025404363870620728,
0.00014469541201833636,
-0.1426829695701599,
0.11935272067785263,
-0.061062414199113846,
0.11300837248563766,
0.058016009628772736,
0.06001058593392372,
-0.1116458848118782,
-0.019978094846010208,
-0.0549047514796257,
0.09347530454397202,
-0.08643282204866409,
-0.09446042776107788,
-0.05264439433813095,
0.05033815652132034,
0.10081378370523453,
-0.03129594400525093,
0.03620050102472305,
-0.05933946371078491,
-0.07424496859312057,
0.055428918451070786,
0.151398703455925,
0.057856690138578415,
-0.0828607976436615,
-0.023678651079535484,
-0.025805428624153137,
0.059286948293447495,
-0.1450657993555069,
0.08991673588752747,
0.0907568410038948,
-0.000841206987388432,
0.0796288475394249,
-0.02464873157441616,
-0.18418927490711212,
-0.03888631984591484,
0.06895077973604202,
-0.09531193971633911,
-0.003635724075138569,
-0.0531240813434124,
0.08743557333946228,
-0.1295793205499649,
-0.011847114190459251,
0.0540454275906086,
0.00477920426055789,
-0.029164498671889305,
0.00736280158162117,
0.024162115529179573,
0.018646080046892166,
0.07742896676063538,
0.07681622356176376,
0.04976319521665573,
-0.04623602330684662,
0.08618722856044769,
0.09175354987382889,
0.003280655248090625,
-0.060200612992048264,
0.05967264622449875,
-0.09912072867155075,
-0.06044817343354225,
0.04447346553206444,
0.0739363431930542,
-0.03644746541976929,
-0.044421277940273285,
-0.049585338681936264,
-0.08071491122245789,
0.03914749622344971,
0.14668381214141846,
-0.02129938267171383,
0.0800461396574974,
0.028539586812257767,
-0.05144008994102478,
0.023256175220012665,
0.06817099452018738,
-0.033774007111787796,
0.0029193032532930374,
-0.06847559660673141,
0.009608691558241844,
0.013508190400898457,
0.008242862299084663,
-0.021357446908950806,
-0.046136584132909775,
-0.05297013372182846,
-0.06032828986644745,
-0.06087201461195946,
0.02455926686525345,
-0.014890638180077076,
-0.010266530327498913,
-0.05102602392435074,
-0.017351699993014336,
-0.031503718346357346,
0.020724624395370483,
-0.05731847137212753,
-0.023901930078864098,
-0.040154002606868744,
0.06450407952070236,
-0.09365025162696838,
-0.03330626338720322,
0.06853468716144562,
-0.09424836188554764,
0.07697246223688126,
0.026264110580086708,
-0.009858968667685986,
-0.017242813482880592,
-0.11005990207195282,
0.021619873121380806,
0.010850885882973671,
0.05153895914554596,
0.026651842519640923,
-0.16255483031272888,
-0.017425114288926125,
-0.02693134732544422,
0.001291681663133204,
-0.014970771968364716,
0.02322136051952839,
-0.10822205245494843,
0.1170588955283165,
-0.0860414057970047,
-0.12131989747285843,
-0.08595510572195053,
0.002372802933678031,
0.04463030770421028,
0.028173260390758514,
0.10719000548124313,
-0.014050728641450405,
0.060646865516901016,
-0.07790093123912811,
-0.005945276468992233,
0.03739519044756889,
0.027210399508476257,
-0.04505256563425064,
-0.034068480134010315,
0.0009864072781056166,
-0.03456839546561241,
0.12502487003803253,
-0.05894922465085983,
-0.03268537297844887,
0.0679025948047638,
0.005026038736104965,
-0.07048060745000839,
0.05263952538371086,
0.03813190758228302,
0.021109094843268394,
0.03549251705408096,
0.010933861136436462,
-0.022892769426107407,
-0.07126249372959137,
-0.0868583396077156,
0.12188215553760529,
0.10870911926031113,
0.14408831298351288,
0.05676070600748062,
0.07822773605585098,
-0.07495313137769699,
-0.037532102316617966,
0.028093649074435234,
-0.06761477887630463,
0.010241417214274406,
-0.01612755097448826,
0.09092718362808228,
0.16335336863994598,
-0.1759549230337143,
0.04777468740940094,
0.067558154463768,
-0.07082164287567139,
-0.14947204291820526,
-0.1456877738237381,
-0.08511488139629364,
0.028979957103729248,
-0.01427305955439806,
-0.059240031987428665,
-0.0011258147424086928,
-0.050441544502973557,
0.02276228927075863,
-0.002438324736431241,
0.07120219618082047,
-0.009289076551795006,
-0.09958529472351074,
0.08858134597539902,
0.05381350591778755,
0.04432683810591698,
0.05483195185661316,
0.057817794382572174,
0.047271691262722015,
0.07248552143573761,
0.015762703493237495,
0.05404166132211685,
0.026815542951226234,
0.00994402077049017,
-0.04090019688010216,
-0.08463434129953384,
-0.02962556853890419,
-0.020518017932772636,
-0.03571147471666336,
0.021150454878807068,
0.0340644046664238,
-0.03969152644276619,
0.029361680150032043,
0.09902383387088776,
-0.04655895009636879,
-0.05976750701665878,
-0.1007181778550148,
0.0637160986661911,
-0.0011146360775455832,
0.059652168303728104,
0.0683554857969284,
-0.1450282484292984,
-0.0006361271953210235,
0.09253506362438202,
0.09167180210351944,
0.007030678912997246,
0.03139358386397362,
0.030607158318161964,
0.012384295463562012,
0.011382769793272018,
-0.0010600588284432888,
0.0325106717646122,
0.17962783575057983,
-0.07232392579317093,
0.0629541203379631,
-0.01855127140879631,
-0.019972141832113266,
-0.09885431826114655,
0.055900558829307556,
-0.010112597607076168,
-0.029871173202991486,
-0.08556077629327774,
0.08887435495853424,
-0.02783805876970291,
-0.19700294733047485,
0.023700812831521034,
-0.04047194495797157,
-0.08066028356552124,
-0.05473269522190094,
-0.04844019562005997,
-0.016404690220952034,
0.018493561074137688,
0.00529532739892602,
-0.04030222073197365,
0.14892658591270447,
0.05441297963261604,
-0.11767860502004623,
-0.01008913479745388,
0.10839762538671494,
-0.07563642412424088,
0.10397191345691681,
0.05216757208108902,
0.08680932223796844,
0.083571657538414,
-0.014480441808700562,
-0.07667176425457001,
0.11860280483961105,
0.01000902522355318,
-0.09088478982448578,
-0.015942227095365524,
0.06915884464979172,
0.02334674634039402,
0.08767998963594437,
0.0773192048072815,
0.02619030326604843,
0.06464049220085144,
-0.011132905259728432,
-0.04356740042567253,
-0.11879797279834747,
0.10462812334299088,
-0.11498306691646576,
0.0818413496017456,
0.11326054483652115,
-0.05780686065554619,
-0.012690950185060501,
-0.04206295683979988,
0.013061556033790112,
0.06296265870332718,
-0.036291684955358505,
-0.016661055386066437,
-0.06368979066610336,
0.04306502640247345,
-0.07411643862724304,
0.15654104948043823,
-0.08139654248952866,
-0.0495588555932045,
-0.0007024703081697226,
-0.0032828717958182096,
-0.04319816455245018,
0.07629024237394333,
-0.005902636330574751,
0.03860817849636078,
-0.012678115628659725,
-0.08206554502248764,
-0.013408195227384567,
0.048258841037750244,
-0.1439141035079956,
-0.04474255442619324
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# xlm-roberta-base-finetuned-marc-en
This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on the amazon_reviews_multi dataset.
It achieves the following results on the evaluation set:
- Loss: 0.8976
- Mae: 0.4268
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Mae |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 1.092 | 1.0 | 235 | 0.9514 | 0.5122 |
| 0.9509 | 2.0 | 470 | 0.8976 | 0.4268 |
### Framework versions
- Transformers 4.11.3
- Pytorch 1.9.0+cu111
- Datasets 1.14.0
- Tokenizers 0.10.3
|
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["amazon_reviews_multi"], "base_model": "xlm-roberta-base", "model-index": [{"name": "xlm-roberta-base-finetuned-marc-en", "results": []}]}
|
text-classification
|
d4niel92/xlm-roberta-base-finetuned-marc-en
|
[
"transformers",
"pytorch",
"tensorboard",
"xlm-roberta",
"text-classification",
"generated_from_trainer",
"dataset:amazon_reviews_multi",
"base_model:xlm-roberta-base",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #xlm-roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
xlm-roberta-base-finetuned-marc-en
==================================
This model is a fine-tuned version of xlm-roberta-base on the amazon\_reviews\_multi dataset.
It achieves the following results on the evaluation set:
* Loss: 0.8976
* Mae: 0.4268
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 2e-05
* train\_batch\_size: 16
* eval\_batch\_size: 16
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 2
### Training results
### Framework versions
* Transformers 4.11.3
* Pytorch 1.9.0+cu111
* Datasets 1.14.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #xlm-roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
79,
98,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #xlm-roberta #text-classification #generated_from_trainer #dataset-amazon_reviews_multi #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.11.3\n* Pytorch 1.9.0+cu111\n* Datasets 1.14.0\n* Tokenizers 0.10.3"
] |
[
-0.09812021255493164,
0.10352690517902374,
-0.001907684258185327,
0.12290644645690918,
0.1674380898475647,
0.03168686106801033,
0.1575489491224289,
0.11984023451805115,
-0.058105770498514175,
0.004223859403282404,
0.13002419471740723,
0.13833214342594147,
0.01715509593486786,
0.1447533220052719,
-0.06018156558275223,
-0.2425287365913391,
-0.0029156955424696207,
0.049298930913209915,
-0.03127361834049225,
0.1479398012161255,
0.10505940020084381,
-0.12610356509685516,
0.10435055196285248,
0.0037258209194988012,
-0.1818360835313797,
-0.010994154959917068,
0.03205085173249245,
-0.05964047834277153,
0.12747356295585632,
0.036169201135635376,
0.11701266467571259,
0.01050691120326519,
0.0660516768693924,
-0.16916663944721222,
0.020428365096449852,
0.035147905349731445,
0.009059338830411434,
0.1040973961353302,
0.024519264698028564,
-0.028808608651161194,
0.07744082808494568,
-0.058070696890354156,
0.06718502193689346,
0.017978480085730553,
-0.13217800855636597,
-0.24712109565734863,
-0.07379481941461563,
0.044259656220674515,
0.05524620786309242,
0.0917683020234108,
-0.01527134794741869,
0.15148352086544037,
-0.043198853731155396,
0.0969441682100296,
0.21268121898174286,
-0.2841886281967163,
-0.07461445778608322,
0.03160625323653221,
0.04387715458869934,
0.10989709198474884,
-0.0937267318367958,
-0.02421688847243786,
0.04592442885041237,
0.04450565576553345,
0.1321316808462143,
-0.03711600601673126,
-0.021549640223383904,
0.01321747899055481,
-0.12804503738880157,
-0.03180558606982231,
0.21304011344909668,
0.0591527558863163,
-0.05360223352909088,
-0.05201129987835884,
-0.03803767263889313,
-0.1456279456615448,
-0.02930956706404686,
0.01058904454112053,
0.038216665387153625,
-0.05502567067742348,
-0.10964832454919815,
-0.03312109038233757,
-0.09813667088747025,
-0.05130113288760185,
-0.03835471719503403,
0.14890140295028687,
0.021664323285222054,
0.01881633698940277,
-0.020792227238416672,
0.1086689829826355,
0.005186812020838261,
-0.130142480134964,
0.019770849496126175,
0.001205006497912109,
-0.01922762393951416,
-0.04256071895360947,
-0.059533651918172836,
-0.08207199722528458,
-0.005260268691927195,
0.13816562294960022,
-0.006730278022587299,
0.03263462707400322,
0.029495157301425934,
0.04845919460058212,
-0.07001308351755142,
0.21546289324760437,
-0.04511220380663872,
-0.04219766706228256,
-0.0002896675141528249,
0.08963412791490555,
0.028395041823387146,
-0.020405536517500877,
-0.14326611161231995,
0.008495262823998928,
0.10691153258085251,
0.0003075152635574341,
-0.045838259160518646,
0.05366884917020798,
-0.07306834310293198,
-0.07121150940656662,
0.01563962548971176,
-0.07442989200353622,
0.017976202070713043,
-0.024269405752420425,
-0.06692063808441162,
-0.04273039475083351,
0.012124831788241863,
0.028101090341806412,
0.0075922803953289986,
0.10486268997192383,
-0.10410065948963165,
0.020699841901659966,
-0.08116897195577621,
-0.10778643935918808,
-0.00022651022300124168,
-0.07472724467515945,
0.03882025182247162,
-0.1064079999923706,
-0.18410973250865936,
-0.026412202045321465,
0.06071587651968002,
-0.03473951667547226,
-0.08254385739564896,
-0.038923557847738266,
-0.04353710636496544,
0.01611519791185856,
-0.012438597157597542,
0.1264721155166626,
-0.06884513050317764,
0.1065962016582489,
0.05314483866095543,
0.0545429103076458,
-0.057804789394140244,
0.043849628418684006,
-0.09878509491682053,
0.01811285689473152,
-0.1646735519170761,
0.05127790942788124,
-0.032390009611845016,
0.07116829603910446,
-0.09387228637933731,
-0.10144060850143433,
0.012676175683736801,
-0.0026084501296281815,
0.04583049938082695,
0.07100346684455872,
-0.1528351604938507,
-0.0758446529507637,
0.13790775835514069,
-0.04834872484207153,
-0.1352408230304718,
0.12820465862751007,
-0.07910808175802231,
0.0651589184999466,
0.07837335020303726,
0.17916326224803925,
0.0777864009141922,
-0.04574386775493622,
0.035024162381887436,
-0.021069610491394997,
0.06634294241666794,
-0.061811257153749466,
0.1131177693605423,
0.01081791426986456,
-0.03710639476776123,
0.02510187029838562,
-0.05857214331626892,
0.047613855451345444,
-0.08469528704881668,
-0.08893166482448578,
-0.02836320735514164,
-0.10854913294315338,
0.07260706275701523,
0.057163309305906296,
0.06783141195774078,
-0.10898226499557495,
-0.08911730349063873,
0.059539347887039185,
0.08375418931245804,
-0.05782730504870415,
0.012237601913511753,
-0.06102912127971649,
0.07865146547555923,
-0.0875755324959755,
-0.027446014806628227,
-0.1719016432762146,
-0.013196619227528572,
0.01573120430111885,
0.016925813630223274,
0.04338356852531433,
0.022445863112807274,
0.06309623271226883,
0.04790724813938141,
-0.07823556661605835,
-0.023563319817185402,
-0.04988884925842285,
-0.0025884294882416725,
-0.11524929106235504,
-0.17892742156982422,
-0.04058213531970978,
-0.02493293024599552,
0.15957015752792358,
-0.20539827644824982,
0.027443308383226395,
-0.05687323212623596,
0.06554590165615082,
0.04896894097328186,
-0.0265037901699543,
-0.015622194856405258,
0.07002788037061691,
-0.03115350566804409,
-0.04752243682742119,
0.07251476496458054,
0.01747770793735981,
-0.1277834177017212,
-0.014317176304757595,
-0.10614078491926193,
0.18912449479103088,
0.11936116218566895,
-0.04920954629778862,
-0.06957253068685532,
0.017118031159043312,
-0.03793710097670555,
-0.029705340042710304,
-0.05631444603204727,
0.01111234724521637,
0.1715090572834015,
0.011354546062648296,
0.1495170295238495,
-0.0915621891617775,
-0.04248065501451492,
0.02530520223081112,
-0.039175376296043396,
0.019209325313568115,
0.13292847573757172,
0.07611285895109177,
-0.13196781277656555,
0.13387610018253326,
0.1607392430305481,
-0.0736284926533699,
0.15296362340450287,
-0.029240358620882034,
-0.05728021636605263,
-0.0464419424533844,
-0.04359692707657814,
0.0004992583999410272,
0.11432897299528122,
-0.09205528348684311,
0.003642542287707329,
0.037298522889614105,
0.007559128571301699,
0.005296600982546806,
-0.194528266787529,
-0.04211641848087311,
0.04960338771343231,
-0.03238385170698166,
-0.0195712111890316,
-0.005746605806052685,
0.005445985123515129,
0.09943103045225143,
0.026033973321318626,
-0.06495804339647293,
0.041400786489248276,
0.006346049718558788,
-0.07763025164604187,
0.20318613946437836,
-0.06127871572971344,
-0.18176858127117157,
-0.15809854865074158,
-0.0608394481241703,
-0.06807000190019608,
0.017409272491931915,
0.04937880113720894,
-0.05477774143218994,
-0.025062300264835358,
-0.08172362297773361,
-0.0028053519781678915,
-0.01773783005774021,
0.01095736213028431,
-0.0006763212732039392,
0.0013714163796976209,
0.06442662328481674,
-0.09177060425281525,
-0.011626158840954304,
-0.03314218297600746,
-0.014734343625605106,
0.0438874289393425,
0.023591991513967514,
0.10905226320028305,
0.13539284467697144,
-0.015037366189062595,
0.004823619499802589,
-0.020374136045575142,
0.2483033537864685,
-0.08111010491847992,
-0.040294963866472244,
0.13866938650608063,
-0.01982552371919155,
0.042179569602012634,
0.1332840472459793,
0.0634307712316513,
-0.08422891050577164,
0.01686006225645542,
0.01620476320385933,
-0.041330400854349136,
-0.22779057919979095,
-0.024201340973377228,
-0.05223262310028076,
-0.008042591623961926,
0.10177147388458252,
0.017685074359178543,
0.01614299975335598,
0.07274973392486572,
0.03827188163995743,
0.07503646612167358,
-0.0291079580783844,
0.08061981946229935,
0.09651633352041245,
0.05387271195650101,
0.14345185458660126,
-0.03671906143426895,
-0.06028701364994049,
0.04848731309175491,
0.0108230821788311,
0.21189790964126587,
0.01326752733439207,
0.17466220259666443,
0.04744992032647133,
0.12500524520874023,
0.014298989437520504,
0.05333346128463745,
0.015036297030746937,
-0.020865047350525856,
-0.03146560490131378,
-0.022346103563904762,
-0.04086718708276749,
0.02209469862282276,
-0.016497479751706123,
0.06340839713811874,
-0.11195199191570282,
-0.02905680611729622,
0.050002921372652054,
0.24331700801849365,
0.037391938269138336,
-0.3439534902572632,
-0.11366234719753265,
0.01883978210389614,
-0.03631078079342842,
-0.025519123300909996,
0.005277425050735474,
0.08607487380504608,
-0.11822685599327087,
0.0291055329144001,
-0.08441028743982315,
0.0970635935664177,
-0.0894991010427475,
0.03904876112937927,
0.05950405076146126,
0.06615041941404343,
-0.012868217192590237,
0.07652463018894196,
-0.26785486936569214,
0.2654654383659363,
0.0038434755988419056,
0.04138108715415001,
-0.0528125986456871,
-0.026833364740014076,
0.018785936757922173,
0.03856952488422394,
0.05312851443886757,
0.002691675443202257,
-0.031036008149385452,
-0.1918506771326065,
-0.054761793464422226,
0.02141708694398403,
0.058805324137210846,
-0.06886237114667892,
0.10347776114940643,
-0.03732399269938469,
0.002364743035286665,
0.046839240938425064,
0.007273396942764521,
-0.04171981289982796,
-0.1025942862033844,
0.006834252271801233,
0.022444799542427063,
-0.03227124363183975,
-0.0657159686088562,
-0.11993240565061569,
-0.05423363298177719,
0.13757771253585815,
0.013785678893327713,
-0.051518190652132034,
-0.10434737056493759,
0.06923169642686844,
0.08733152598142624,
-0.08491255342960358,
0.025313682854175568,
0.0014104947913438082,
0.09964048862457275,
0.02253890223801136,
-0.04694976285099983,
0.09993849694728851,
-0.03903364762663841,
-0.17351704835891724,
-0.054366543889045715,
0.1256658434867859,
0.020731644704937935,
0.06427878141403198,
-0.020925333723425865,
0.022260025143623352,
-0.06550727039575577,
-0.07281044870615005,
0.029962480068206787,
-0.016384465619921684,
0.06836926192045212,
0.037778619676828384,
-0.01038407627493143,
0.02294698916375637,
-0.08377562463283539,
-0.04059524089097977,
0.18432417511940002,
0.23314684629440308,
-0.08502037823200226,
0.013911966234445572,
0.017628196626901627,
-0.058133941143751144,
-0.13472311198711395,
0.01633010432124138,
0.06756463646888733,
0.01927938126027584,
0.07207529246807098,
-0.14412008225917816,
0.09311923384666443,
0.08463025093078613,
-0.017407385632395744,
0.10958670824766159,
-0.28891894221305847,
-0.13729993999004364,
0.09376998990774155,
0.14318998157978058,
0.13833874464035034,
-0.13397923111915588,
-0.03197073936462402,
-0.04658104479312897,
-0.1487695574760437,
0.13397598266601562,
-0.08040919154882431,
0.13804513216018677,
-0.025544079020619392,
0.096004918217659,
0.01279239822179079,
-0.04526284709572792,
0.1393755078315735,
0.007535177282989025,
0.08931449800729752,
-0.05096963420510292,
-0.04220948740839958,
0.032130055129528046,
-0.04924353212118149,
0.013439087197184563,
-0.09285589307546616,
0.03006870299577713,
-0.11729402840137482,
-0.039012353867292404,
-0.07519809901714325,
0.014598245732486248,
-0.02858785353600979,
-0.05347359552979469,
-0.03161230683326721,
0.04778921604156494,
0.025237634778022766,
-0.012278441339731216,
0.16444189846515656,
-0.0027924254536628723,
0.14722470939159393,
0.10666793584823608,
0.08577708154916763,
-0.046120233833789825,
-0.08052004873752594,
-0.03603455796837807,
-0.033500637859106064,
0.04649407044053078,
-0.14823013544082642,
0.031162351369857788,
0.12681978940963745,
0.007066360209137201,
0.16054615378379822,
0.06166496500372887,
-0.031135370954871178,
0.01827199198305607,
0.06477196514606476,
-0.15788358449935913,
-0.11600645631551743,
-0.03486217185854912,
-0.07604873180389404,
-0.14656369388103485,
0.020384302362799644,
0.13379241526126862,
-0.05541857331991196,
-0.036075398325920105,
-0.008475614711642265,
0.004952962044626474,
-0.04686073958873749,
0.17704343795776367,
0.07624934613704681,
0.05941781401634216,
-0.09578975290060043,
0.08243342489004135,
0.07437931001186371,
-0.04332021623849869,
0.0032680847216397524,
0.042980778962373734,
-0.08939282596111298,
-0.04395003616809845,
0.021134288981556892,
0.17765511572360992,
-0.07725949585437775,
-0.029049687087535858,
-0.16312871873378754,
-0.10595784336328506,
0.06462317705154419,
0.11505185812711716,
0.10770709812641144,
0.005066306795924902,
-0.03954630717635155,
-0.0202241949737072,
-0.0869995504617691,
0.11391117423772812,
0.07455001026391983,
0.07691191881895065,
-0.15757641196250916,
0.08190253376960754,
0.00984792597591877,
0.052459716796875,
-0.015612490475177765,
0.023781491443514824,
-0.1068887934088707,
0.01018856093287468,
-0.1314578652381897,
-0.00713737728074193,
-0.01886388659477234,
0.017276203259825706,
-0.00975011009722948,
-0.0737651139497757,
-0.06535785645246506,
0.005364119540899992,
-0.12144448608160019,
-0.03146880492568016,
0.041686635464429855,
0.059771209955215454,
-0.08568035811185837,
-0.032555531710386276,
0.04000971466302872,
-0.054124634712934494,
0.07259633392095566,
0.034383878111839294,
0.016957225278019905,
0.044215552508831024,
-0.09063462167978287,
0.034417614340782166,
0.021733084693551064,
0.006982472259551287,
0.04584461823105812,
-0.13501961529254913,
0.002426386345177889,
-0.006235599052160978,
0.06819358468055725,
0.022565463557839394,
0.08897408097982407,
-0.14996711909770966,
-0.00550141092389822,
0.002148654079064727,
-0.06669206917285919,
-0.06453054398298264,
0.03369293734431267,
0.06227828189730644,
0.05516096204519272,
0.21624374389648438,
-0.07163715362548828,
0.03457832708954811,
-0.21061794459819794,
0.003025998128578067,
-0.018195707350969315,
-0.12717124819755554,
-0.11335045844316483,
-0.08173266798257828,
0.04758157581090927,
-0.06282927840948105,
0.14101436734199524,
0.025031045079231262,
0.07818511128425598,
0.023037582635879517,
0.00678018294274807,
0.022690627723932266,
0.014421696774661541,
0.1840703934431076,
0.008179683238267899,
-0.0434228889644146,
0.0778246745467186,
0.04037131741642952,
0.09507386386394501,
0.13757115602493286,
0.1634574979543686,
0.16562770307064056,
0.01898784749209881,
0.06936065107584,
0.03429622948169708,
-0.006622704677283764,
-0.13671477138996124,
0.03102089837193489,
-0.008023800328373909,
0.1043572947382927,
-0.0066113825887441635,
0.2086031138896942,
0.07871729135513306,
-0.17107908427715302,
0.037778399884700775,
-0.055767543613910675,
-0.08584146201610565,
-0.09015718847513199,
-0.09161543101072311,
-0.09997442364692688,
-0.14406391978263855,
-0.0019013544078916311,
-0.12244853377342224,
-0.00808026734739542,
0.10944617539644241,
-0.0033328398130834103,
-0.04050534591078758,
0.11308660358190536,
0.020677482709288597,
-0.0025744130834937096,
0.07750289142131805,
0.0031741566490381956,
-0.036672383546829224,
-0.0736335963010788,
-0.06324776262044907,
-0.012521742843091488,
-0.003947390243411064,
0.032419171184301376,
-0.056489042937755585,
-0.06048688292503357,
0.018643613904714584,
-0.023840805515646935,
-0.12456218898296356,
0.012879862450063229,
0.024760475382208824,
0.07430610060691833,
0.04327857121825218,
0.008459833450615406,
0.007238071411848068,
-0.00772605137899518,
0.2739601135253906,
-0.06460370123386383,
-0.046303700655698776,
-0.12886743247509003,
0.21538853645324707,
0.015485062263906002,
-0.03713414818048477,
0.02819807268679142,
-0.07591656595468521,
0.02228512614965439,
0.22270861268043518,
0.21857169270515442,
-0.09196339547634125,
-0.012035978958010674,
0.008699605241417885,
-0.006352963391691446,
0.0012919970322400331,
0.1007501557469368,
0.08472951501607895,
-0.014648787677288055,
-0.07979421317577362,
-0.006122843828052282,
-0.055351871997117996,
-0.007488004397600889,
-0.008557664230465889,
0.06689897924661636,
0.044714320451021194,
0.008150093257427216,
-0.053536467254161835,
0.07880229502916336,
-0.07123448699712753,
-0.11125939339399338,
0.025823740288615227,
-0.21135656535625458,
-0.1810917854309082,
-0.03228117153048515,
0.07168364524841309,
-0.0003595422313082963,
0.06783684343099594,
-0.025290708988904953,
0.004122728016227484,
0.04975440353155136,
-0.011912780813872814,
-0.08325883001089096,
-0.09511129558086395,
0.10446051508188248,
-0.09058190882205963,
0.18431179225444794,
-0.05186767131090164,
0.03716496005654335,
0.12413962930440903,
0.04805907607078552,
-0.07977709174156189,
0.0754096657037735,
0.03829657658934593,
-0.008729022927582264,
0.0597822479903698,
0.11527280509471893,
-0.024547936394810677,
0.09938172250986099,
0.052748117595911026,
-0.11229771375656128,
0.00046200439101085067,
-0.08963160216808319,
-0.030114740133285522,
-0.054946474730968475,
-0.0371994748711586,
-0.06871096044778824,
0.1453767567873001,
0.2193399965763092,
-0.05803428590297699,
-0.02145204320549965,
-0.05205347016453743,
0.02569247968494892,
0.08520379662513733,
0.03857872262597084,
-0.06332465261220932,
-0.22206559777259827,
-0.006284010596573353,
0.07150046527385712,
-0.01069146953523159,
-0.2947491705417633,
-0.07019023597240448,
-0.011928042396903038,
-0.06109558045864105,
-0.049017444252967834,
0.104152150452137,
0.07352162152528763,
0.04259505122900009,
-0.0640808492898941,
-0.03738879784941673,
-0.07639223337173462,
0.15060356259346008,
-0.15641680359840393,
-0.09290610998868942
] |
null | null |
transformers
|
# Harry
|
{"tags": ["conversational"]}
|
text-generation
|
d4rk/harry
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
|
# Harry
|
[
"# Harry"
] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n",
"# Harry"
] |
[
51,
2
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #conversational #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Harry"
] |
[
0.008150327950716019,
0.07997918874025345,
-0.007178797852247953,
0.03567842021584511,
0.13429023325443268,
0.04860587418079376,
0.13002704083919525,
0.15211623907089233,
0.032173920422792435,
-0.021888386458158493,
0.1481919139623642,
0.20967841148376465,
0.002699901582673192,
0.027984336018562317,
-0.07098018378019333,
-0.2619108259677887,
0.04580087959766388,
0.05127737671136856,
-0.03388937562704086,
0.11372090876102448,
0.08055732399225235,
-0.07797889411449432,
0.10050316154956818,
-0.04493314027786255,
-0.1697816699743271,
-0.002110888483002782,
0.03408896550536156,
-0.11597410589456558,
0.14719758927822113,
0.06921592354774475,
0.048300813883543015,
-0.00850814301520586,
-0.076728954911232,
-0.15931929647922516,
0.03596772626042366,
0.009679432958364487,
-0.05316237732768059,
0.07768061757087708,
0.07793281972408295,
-0.06378860771656036,
0.09473442286252975,
0.12092399597167969,
0.012448522262275219,
0.05755617842078209,
-0.16281144320964813,
-0.03889242932200432,
-0.025217758491635323,
0.05875825509428978,
0.02882077358663082,
0.08777906000614166,
-0.025393502786755562,
0.1150071993470192,
-0.06023528426885605,
0.10495458543300629,
0.15590788424015045,
-0.37643104791641235,
-0.01927122101187706,
0.08297953754663467,
0.02421913668513298,
0.07742053270339966,
-0.05841423198580742,
0.05105055496096611,
0.0291367769241333,
0.010136914439499378,
-0.034277502447366714,
-0.08226783573627472,
-0.0691092237830162,
0.056069597601890564,
-0.08630520105361938,
-0.03791604936122894,
0.2868892550468445,
-0.06861335039138794,
0.06846558302640915,
-0.04295716807246208,
-0.1125124916434288,
0.010618508793413639,
-0.023355290293693542,
-0.0318555124104023,
-0.07080025225877762,
0.07488057017326355,
0.010736515745520592,
-0.05902828648686409,
-0.10992886871099472,
-0.00898059457540512,
-0.1566953957080841,
0.18897034227848053,
0.04560272395610809,
0.053054329007864,
-0.2046297937631607,
0.08976330608129501,
0.03162526339292526,
-0.06762029975652695,
0.03180040046572685,
-0.10093335807323456,
0.07211042940616608,
0.01263956818729639,
-0.03548488765954971,
-0.0617138035595417,
0.08190520852804184,
0.09648655354976654,
0.018170509487390518,
0.03031209670007229,
-0.018015019595623016,
0.07860899716615677,
0.07015649229288101,
0.03260401636362076,
0.020849935710430145,
-0.10083217918872833,
0.031029850244522095,
-0.12741801142692566,
-0.004848834127187729,
-0.07212943583726883,
-0.15646961331367493,
-0.032632630318403244,
0.07265544682741165,
0.04057692736387253,
0.04629788547754288,
0.0918584018945694,
-0.017937956377863884,
-0.031683288514614105,
0.005309654865413904,
-0.03881150111556053,
-0.0016185828717425466,
0.010565927252173424,
0.004867391660809517,
0.17569318413734436,
-0.01173572801053524,
0.005573826376348734,
-0.13824132084846497,
0.06755446642637253,
-0.07359027862548828,
0.003291269764304161,
-0.009804779663681984,
-0.043432384729385376,
0.00019013567361980677,
0.02275741659104824,
0.009434896521270275,
-0.1306990087032318,
-0.09456845372915268,
0.005049117375165224,
0.001683380571193993,
-0.030414460226893425,
-0.0880187600851059,
-0.06016523391008377,
0.0005792967858724296,
0.03403691574931145,
-0.03788115829229355,
-0.02053765580058098,
-0.05491064488887787,
0.13973355293273926,
-0.039907243102788925,
0.0680219903588295,
-0.08543722331523895,
0.07607127726078033,
-0.08393777906894684,
-0.03660587593913078,
-0.08674334734678268,
0.04517173394560814,
0.023118294775485992,
0.0952034592628479,
0.009009120985865593,
-0.035681288689374924,
-0.02597375586628914,
0.06909751147031784,
-0.08201564103364944,
0.22704581916332245,
-0.06106000021100044,
-0.13297651708126068,
0.27533063292503357,
-0.06215481460094452,
-0.1545005589723587,
0.12745322287082672,
-0.0022543161176145077,
0.04219355061650276,
0.11488337814807892,
0.1701890528202057,
-0.023590579628944397,
-0.012874958105385303,
0.086909219622612,
0.08520808070898056,
-0.08411797881126404,
0.01972302235662937,
0.021819185465574265,
-0.03323185071349144,
-0.18226908147335052,
0.05368303880095482,
0.06166396662592888,
0.004056228790432215,
-0.03960279002785683,
-0.03434339165687561,
-0.008246045559644699,
-0.002105156658217311,
0.11326885968446732,
-0.0030808148439973593,
0.11007247120141983,
-0.07832269370555878,
-0.02927342616021633,
-0.061117690056562424,
0.013490136712789536,
-0.03389749675989151,
0.04241752624511719,
-0.030486803501844406,
0.104023277759552,
0.030795544385910034,
0.06876599043607712,
-0.1441798061132431,
-0.018419278785586357,
-0.01465595792979002,
0.1655028760433197,
0.03634936735033989,
0.11940139532089233,
0.0469493567943573,
-0.015638815239071846,
-0.021675750613212585,
0.012381220236420631,
0.15536922216415405,
-0.032593898475170135,
-0.08762139081954956,
-0.1258864551782608,
0.06891770660877228,
-0.04575640708208084,
0.0731445699930191,
-0.10150167346000671,
0.015967585146427155,
0.01882955990731716,
0.09948527812957764,
0.008797373622655869,
0.03413883224129677,
-0.02743394672870636,
-0.005820179358124733,
-0.07495683431625366,
0.00875744316726923,
0.0919453501701355,
-0.003998344298452139,
-0.04996733367443085,
0.13806571066379547,
-0.17212142050266266,
0.20840395987033844,
0.20267750322818756,
-0.32493656873703003,
0.009374254383146763,
-0.08949611335992813,
-0.031872112303972244,
0.004958184435963631,
0.05969046801328659,
-0.02312314696609974,
0.13331562280654907,
-0.018370969220995903,
0.17138724029064178,
-0.03461623564362526,
-0.07161927968263626,
-0.03657282143831253,
-0.047394730150699615,
-0.004794360604137182,
0.0676620677113533,
0.05765331909060478,
-0.13297857344150543,
0.19050554931163788,
0.1610855609178543,
0.0399690680205822,
0.1481526792049408,
0.054114822298288345,
-0.012512989342212677,
0.079967200756073,
-0.012349419295787811,
-0.044576261192560196,
-0.044705942273139954,
-0.2649945914745331,
-0.059904683381319046,
0.08162160217761993,
-0.016532359644770622,
0.08187162131071091,
-0.10637134313583374,
-0.027939267456531525,
-0.02658247947692871,
0.010447057895362377,
0.04430146887898445,
0.06908783316612244,
0.04134810343384743,
0.13113145530223846,
0.009810316376388073,
-0.01703023351728916,
0.08961160480976105,
0.013629522174596786,
-0.10744574666023254,
0.16339001059532166,
-0.15301619470119476,
-0.31661707162857056,
-0.1222880631685257,
-0.21792426705360413,
-0.047991618514060974,
0.047343697398900986,
0.11204830557107925,
-0.12909647822380066,
-0.018080465495586395,
0.010634495876729488,
0.08049434423446655,
-0.16831138730049133,
0.014104016125202179,
-0.10683790594339371,
0.050810858607292175,
-0.13064362108707428,
-0.08536121249198914,
-0.03668107092380524,
-0.02152377925813198,
-0.055290527641773224,
0.15841889381408691,
-0.10820674896240234,
0.014260992407798767,
0.20680807530879974,
0.056190866976976395,
0.06024555489420891,
-0.04419902339577675,
0.1893211007118225,
-0.09929632395505905,
-0.024943780153989792,
0.13006441295146942,
-0.07264143228530884,
0.07614395767450333,
0.11094104498624802,
0.010037418454885483,
-0.08044503629207611,
0.017071880400180817,
-0.04862537235021591,
-0.03726791590452194,
-0.26204052567481995,
-0.11292284727096558,
-0.11419184505939484,
0.14296068251132965,
0.07535058259963989,
0.06098230555653572,
0.19169019162654877,
0.05194266512989998,
-0.041768621653318405,
0.060846250504255295,
0.062060531228780746,
0.11330720782279968,
0.22340160608291626,
-0.06009332090616226,
0.1037193313241005,
-0.026680653914809227,
-0.12193924933671951,
0.08004987239837646,
0.06268240511417389,
0.08123214542865753,
0.0726436972618103,
0.029904432594776154,
-0.014741581864655018,
0.11558777838945389,
0.12028518319129944,
0.07602997124195099,
0.03023551218211651,
0.0044477335177361965,
-0.0442391112446785,
-0.027413515374064445,
-0.06894954293966293,
0.05173788592219353,
0.07534655928611755,
-0.13301824033260345,
-0.02288123592734337,
-0.08863618224859238,
0.0733361765742302,
0.08733070641756058,
0.0683031901717186,
-0.15697596967220306,
-0.032764140516519547,
0.08770224452018738,
-0.06571709364652634,
-0.14380478858947754,
0.110891193151474,
0.005665468517690897,
-0.17159228026866913,
0.03885698318481445,
-0.03743424639105797,
0.11335831135511398,
-0.07454986870288849,
0.08642100542783737,
-0.08794202655553818,
-0.09411060065031052,
0.01720825955271721,
0.10373841971158981,
-0.3548857271671295,
0.1734755039215088,
0.007529132533818483,
-0.05844740569591522,
-0.10466589778661728,
-0.0047827474772930145,
0.014064757153391838,
0.0863635241985321,
0.09708419442176819,
-0.0039138952270150185,
0.036004625260829926,
-0.030379468575119972,
0.00469856895506382,
0.023473018780350685,
0.13513074815273285,
-0.048097193241119385,
-0.00026968956808559597,
-0.0625312551856041,
0.0006561221671290696,
-0.05067351087927818,
-0.030396992340683937,
0.056280605494976044,
-0.18914814293384552,
0.10904168337583542,
-0.03988165035843849,
0.11060917377471924,
0.028634263202548027,
-0.0037388638593256474,
-0.11019498109817505,
0.20333942770957947,
-0.12370184063911438,
-0.07889536768198013,
-0.08892582356929779,
-0.045901790261268616,
0.025939425453543663,
-0.04573778435587883,
0.014072186313569546,
-0.05558013170957565,
0.04509544372558594,
-0.07747072726488113,
-0.19148176908493042,
0.10901689529418945,
-0.07811402529478073,
-0.06913676112890244,
-0.044795822352170944,
0.2102656215429306,
-0.04631190374493599,
0.046137742698192596,
0.015599299222230911,
0.02182438224554062,
-0.15843187272548676,
-0.06383797526359558,
0.06230679899454117,
-0.0033062263391911983,
0.041726429015398026,
0.03776678442955017,
-0.042808715254068375,
-0.019449908286333084,
-0.08892440795898438,
-0.026357483118772507,
0.32922229170799255,
0.17607098817825317,
-0.023973498493433,
0.17977000772953033,
0.1232294961810112,
-0.0830162838101387,
-0.26702624559402466,
-0.12107967585325241,
-0.14790202677249908,
-0.07327283173799515,
-0.0888335257768631,
-0.22277477383613586,
0.10721941292285919,
0.009712551720440388,
0.008863099850714207,
0.15568271279335022,
-0.22837719321250916,
-0.07790812104940414,
0.16020281612873077,
0.020157743245363235,
0.4183369576931,
-0.14134015142917633,
-0.1039331778883934,
-0.052355751395225525,
-0.21405909955501556,
0.12258481979370117,
-0.05944736301898956,
0.10431770980358124,
-0.028683850541710854,
0.16617581248283386,
0.0383528396487236,
-0.0013101908843964338,
0.09088904410600662,
-0.0016231819754466414,
-0.04292507842183113,
-0.09273933619260788,
-0.11997798085212708,
0.015495941042900085,
0.022062698379158974,
-0.02430347166955471,
-0.048273127526044846,
0.005268789827823639,
-0.11974101513624191,
-0.03329618647694588,
-0.07172343134880066,
0.023735133931040764,
0.00563058303669095,
-0.0678010806441307,
-0.022930899634957314,
-0.06926808506250381,
-0.0010349624790251255,
0.02622888796031475,
0.22198224067687988,
-0.044733624905347824,
0.20377303659915924,
0.036108262836933136,
0.13506002724170685,
-0.10838258266448975,
-0.035650696605443954,
-0.06348668783903122,
-0.07678893208503723,
0.09647718816995621,
-0.09834396094083786,
0.0431787483394146,
0.11039052158594131,
-0.01863744854927063,
0.07989563792943954,
0.11678334325551987,
0.004321983549743891,
-0.008723610080778599,
0.07342078536748886,
-0.3068614900112152,
-0.0681929960846901,
-0.05015511438250542,
0.02493871934711933,
0.0664868876338005,
0.07411752641201019,
0.18184885382652283,
-0.003767226357012987,
-0.07080593705177307,
0.018064318224787712,
0.036588672548532486,
-0.03494609147310257,
0.06030922010540962,
0.01670447550714016,
0.02351362630724907,
-0.15695813298225403,
0.07625272125005722,
0.012057237327098846,
-0.1320752650499344,
0.015901941806077957,
0.18004119396209717,
-0.10726824402809143,
-0.12882904708385468,
-0.061185527592897415,
0.13282352685928345,
-0.08478786051273346,
-0.0034728446044027805,
-0.03434884920716286,
-0.12939453125,
0.06734540313482285,
0.08906140923500061,
0.05634807422757149,
0.09411884844303131,
-0.07482397556304932,
-0.013990667648613453,
-0.028864599764347076,
-0.004492250271141529,
0.007072563748806715,
0.009635094553232193,
-0.060111153870821,
0.11391019076108932,
-0.04253385215997696,
0.13694676756858826,
-0.08699533343315125,
-0.10300180315971375,
-0.15517710149288177,
0.03107241541147232,
-0.05667778477072716,
-0.07424800097942352,
-0.10900390893220901,
-0.045500848442316055,
0.005280633922666311,
-0.01999524235725403,
-0.025221945717930794,
-0.06647715717554092,
-0.12245137244462967,
0.034579820930957794,
-0.016108276322484016,
0.02834996208548546,
-0.0634443461894989,
0.02798200026154518,
0.09315855801105499,
-0.05594483017921448,
0.1529751867055893,
0.17014192044734955,
-0.11237859725952148,
0.12391463667154312,
-0.08101852238178253,
-0.11565970629453659,
0.06840144097805023,
0.01605931855738163,
0.035929542034864426,
0.08430468291044235,
-0.0229821614921093,
0.023709028959274292,
0.03800255060195923,
0.055617764592170715,
0.07614754885435104,
-0.0866733193397522,
0.06365855783224106,
-0.07000017911195755,
-0.16882477700710297,
-0.05267470329999924,
-0.0494660921394825,
0.051722779870033264,
-0.0029551072511821985,
0.11207453906536102,
-0.06083528697490692,
0.10881359875202179,
-0.03841792419552803,
0.0372023805975914,
0.027518853545188904,
-0.18248122930526733,
-0.047389376908540726,
-0.06853744387626648,
0.04167060926556587,
0.006292351987212896,
0.24125605821609497,
-0.013376148417592049,
-0.014788520522415638,
0.04252627119421959,
0.05376771464943886,
-0.04661921411752701,
0.004615034908056259,
0.22921457886695862,
0.1206577941775322,
-0.0850074514746666,
-0.10838250815868378,
0.06876291334629059,
0.03255823254585266,
0.07629062235355377,
0.13475686311721802,
0.02012047916650772,
0.07085832208395004,
0.06813158839941025,
-0.04799406975507736,
0.016531076282262802,
-0.08833154290914536,
-0.11859961599111557,
0.006483956705778837,
0.04762992635369301,
-0.022459015250205994,
0.20377089083194733,
0.16443926095962524,
-0.02984069101512432,
0.01711403764784336,
-0.04578550159931183,
-0.061951711773872375,
-0.16234396398067474,
-0.07885928452014923,
-0.07652788609266281,
-0.14542102813720703,
-0.0046520838513970375,
-0.1117999479174614,
0.03598939999938011,
0.03860808536410332,
0.06861793249845505,
-0.07057531923055649,
0.022871319204568863,
0.09203927218914032,
-0.10495421290397644,
0.07121753692626953,
-0.017546001821756363,
0.05341155081987381,
-0.043894652277231216,
-0.01867021806538105,
-0.10766304284334183,
-0.01766769029200077,
0.017951572313904762,
0.06088048964738846,
-0.06718584895133972,
0.025688791647553444,
-0.13271553814411163,
-0.11694573611021042,
-0.02835421822965145,
0.06212622672319412,
-0.061673637479543686,
0.14369596540927887,
-0.0004113703325856477,
-0.012999610044062138,
0.03280693292617798,
0.2208307683467865,
-0.09422335773706436,
-0.036435529589653015,
-0.04681945592164993,
0.21503198146820068,
0.029408808797597885,
0.0896737203001976,
0.0074201179668307304,
0.00048709698603488505,
-0.0701461061835289,
0.3298131227493286,
0.31687265634536743,
-0.08254922181367874,
0.008481976576149464,
0.0518588088452816,
0.05116882547736168,
0.12785591185092926,
0.08292554318904877,
0.09929842501878738,
0.2886435091495514,
-0.07870082557201385,
-0.005829294677823782,
-0.01997615210711956,
-0.01728188246488571,
-0.0901113823056221,
0.05832153558731079,
0.05863518640398979,
-0.09249113500118256,
-0.03095107525587082,
0.07863713800907135,
-0.25041717290878296,
0.052830521017313004,
-0.06248736381530762,
-0.18158715963363647,
-0.05754063278436661,
0.004798607900738716,
0.0962035059928894,
0.02952558547258377,
0.10111293941736221,
0.010615319944918156,
-0.09381245076656342,
0.1065032109618187,
0.037793561816215515,
-0.24601826071739197,
-0.01933654025197029,
0.10729585587978363,
-0.10565974563360214,
0.004856930114328861,
-0.015523990616202354,
0.06057559698820114,
0.05267098918557167,
0.058994874358177185,
-0.024261750280857086,
-0.02421385981142521,
0.01983087882399559,
-0.0751492828130722,
-0.02050439454615116,
0.059041865170001984,
0.03624526411294937,
-0.08570602536201477,
0.10611453652381897,
-0.10178600996732712,
0.019064752385020256,
0.04799460619688034,
-0.04521530494093895,
0.005907328333705664,
0.013932397589087486,
-0.057289909571409225,
0.04465124011039734,
0.10357904434204102,
-0.01479580532759428,
-0.02200905978679657,
-0.05266686901450157,
-0.04017287865281105,
-0.0017535027582198381,
-0.07938819378614426,
-0.11415660381317139,
-0.15015143156051636,
-0.12062826007604599,
0.010497049428522587,
-0.016855400055646896,
-0.1783480942249298,
0.0055547296069562435,
-0.08804512023925781,
0.07472650706768036,
-0.19379238784313202,
0.07896775752305984,
0.10117519646883011,
0.005683623719960451,
-0.005208167247474194,
-0.026319080963730812,
0.04194845259189606,
0.11999927461147308,
-0.10750556737184525,
-0.0698806643486023
] |
null | null |
transformers
|
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# opus-mt-zh-en-ep1-renri-zh-to-en
This model is a fine-tuned version of [Helsinki-NLP/opus-mt-zh-en](https://huggingface.co/Helsinki-NLP/opus-mt-zh-en) on an unkown dataset.
It achieves the following results on the evaluation set:
- Loss: 2.2192
- Bleu: 18.2579
- Gen Len: 28.4817
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
|:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|
| 2.2194 | 1.0 | 59472 | 2.2192 | 18.2579 | 28.4817 |
### Framework versions
- Transformers 4.9.2
- Pytorch 1.9.0+cu102
- Datasets 1.11.0
- Tokenizers 0.10.3
|
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["bleu"], "model_index": [{"name": "opus-mt-zh-en-ep1-renri-zh-to-en", "results": [{"task": {"name": "Sequence-to-sequence Language Modeling", "type": "text2text-generation"}, "metric": {"name": "Bleu", "type": "bleu", "value": 18.2579}}]}]}
|
text2text-generation
|
dadada/opus-mt-zh-en-ep1-renri-zh-to-en
|
[
"transformers",
"pytorch",
"tensorboard",
"marian",
"text2text-generation",
"generated_from_trainer",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #tensorboard #marian #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
|
opus-mt-zh-en-ep1-renri-zh-to-en
================================
This model is a fine-tuned version of Helsinki-NLP/opus-mt-zh-en on an unkown dataset.
It achieves the following results on the evaluation set:
* Loss: 2.2192
* Bleu: 18.2579
* Gen Len: 28.4817
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 1e-05
* train\_batch\_size: 32
* eval\_batch\_size: 32
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* num\_epochs: 1
* mixed\_precision\_training: Native AMP
### Training results
### Framework versions
* Transformers 4.9.2
* Pytorch 1.9.0+cu102
* Datasets 1.11.0
* Tokenizers 0.10.3
|
[
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3"
] |
[
"TAGS\n#transformers #pytorch #tensorboard #marian #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP",
"### Training results",
"### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3"
] |
[
58,
113,
4,
34
] |
[
"passage: TAGS\n#transformers #pytorch #tensorboard #marian #text2text-generation #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 1\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.9.2\n* Pytorch 1.9.0+cu102\n* Datasets 1.11.0\n* Tokenizers 0.10.3"
] |
[
-0.08791595697402954,
0.06778696924448013,
-0.002774764085188508,
0.10437888652086258,
0.15418370068073273,
0.012980547733604908,
0.13824374973773956,
0.13198433816432953,
-0.123493492603302,
0.015357204712927341,
0.11528561264276505,
0.1572904884815216,
0.022327890619635582,
0.1280318647623062,
-0.04694313183426857,
-0.26569893956184387,
-0.001036170986481011,
0.03903571888804436,
-0.06751322001218796,
0.13706661760807037,
0.09391605108976364,
-0.1256972998380661,
0.0703691616654396,
0.014480474404990673,
-0.19491037726402283,
0.012739951722323895,
0.007706410717219114,
-0.05720984563231468,
0.15665139257907867,
0.02630198746919632,
0.12842091917991638,
0.01715974695980549,
0.09253322333097458,
-0.20467893779277802,
0.014006925746798515,
0.05626577138900757,
0.0173840019851923,
0.08599337935447693,
0.07135938853025436,
0.0044000945053994656,
0.13464853167533875,
-0.06941051036119461,
0.05643445625901222,
0.020942287519574165,
-0.1284196525812149,
-0.24048922955989838,
-0.09883879870176315,
0.01696871966123581,
0.05995500832796097,
0.10928311198949814,
-0.004749820102006197,
0.13870421051979065,
-0.08293918520212173,
0.09648739546537399,
0.236740380525589,
-0.2777051031589508,
-0.06298667192459106,
0.008424305357038975,
0.03986828401684761,
0.08094967156648636,
-0.09246138483285904,
-0.026077786460518837,
0.03836773708462715,
0.04884025454521179,
0.13770347833633423,
-0.030582552775740623,
-0.12681959569454193,
0.008900257758796215,
-0.1434907168149948,
-0.037285882979631424,
0.1253456473350525,
0.032955851405858994,
-0.026019761338829994,
-0.038010645657777786,
-0.06117980182170868,
-0.14217792451381683,
-0.03621850907802582,
-0.017577173188328743,
0.0491693839430809,
-0.021336745470762253,
-0.07405712455511093,
-0.02605539932847023,
-0.10656807571649551,
-0.0637538805603981,
-0.07205202430486679,
0.10730307549238205,
0.04318959265947342,
0.005975916516035795,
-0.03774913027882576,
0.09780632704496384,
0.0067856283858418465,
-0.1330709606409073,
0.029071807861328125,
0.029501214623451233,
-0.005510215647518635,
-0.04391882196068764,
-0.07265711575746536,
-0.07976450771093369,
0.008835804648697376,
0.11209110170602798,
-0.0746183767914772,
0.05527203530073166,
0.0029112931806594133,
0.046050816774368286,
-0.10201818495988846,
0.17773960530757904,
-0.04022593796253204,
-0.0022444119676947594,
0.006159793585538864,
0.05227687209844589,
0.009077019058167934,
-0.017360802739858627,
-0.10797537118196487,
0.015214648097753525,
0.1084587350487709,
0.01941302977502346,
-0.05710422247648239,
0.061585813760757446,
-0.048781219869852066,
-0.022617818787693977,
-0.01391123328357935,
-0.09246427565813065,
0.03627125173807144,
-0.0005189222283661366,
-0.08581742644309998,
-0.01034771092236042,
0.03210000693798065,
0.03004133328795433,
-0.028537388890981674,
0.09852080047130585,
-0.0703904777765274,
0.043697696179151535,
-0.10688280314207077,
-0.12720033526420593,
0.024808792397379875,
-0.05858444422483444,
0.011153876781463623,
-0.09752579778432846,
-0.1748233586549759,
-0.026145027950406075,
0.05995228886604309,
-0.024947302415966988,
-0.05120270699262619,
-0.05053867772221565,
-0.06109324470162392,
0.012851240113377571,
-0.02651648037135601,
0.14930644631385803,
-0.06233479827642441,
0.10883718729019165,
0.023567747324705124,
0.05881255865097046,
-0.043135132640600204,
0.060773249715566635,
-0.09613388031721115,
0.0010936775943264365,
-0.16715404391288757,
0.049999114125967026,
-0.04385833814740181,
0.061092808842659,
-0.09791086614131927,
-0.10250996053218842,
0.014000436291098595,
-0.0025066849775612354,
0.09485126286745071,
0.08405997604131699,
-0.181447371840477,
-0.07206132262945175,
0.1752375066280365,
-0.06884466111660004,
-0.10394755005836487,
0.11733759194612503,
-0.06311088800430298,
0.057391341775655746,
0.07269296795129776,
0.1771620213985443,
0.05070186033844948,
-0.07557624578475952,
0.05122784897685051,
-0.03185207024216652,
0.061406657099723816,
-0.049615275114774704,
0.05973168835043907,
-0.0029296674765646458,
0.013962152414023876,
0.02333492785692215,
-0.020313812419772148,
0.07495976984500885,
-0.09990449994802475,
-0.09113284945487976,
-0.03881802037358284,
-0.09297861158847809,
0.029516909271478653,
0.06637145578861237,
0.07810681313276291,
-0.10207845270633698,
-0.08077336847782135,
0.07025942206382751,
0.07737163454294205,
-0.06835421919822693,
0.03965305536985397,
-0.055174365639686584,
0.05420743301510811,
-0.019964849576354027,
-0.01201464794576168,
-0.18831293284893036,
-0.009081282652914524,
0.013170500285923481,
-0.023948630318045616,
0.04152487963438034,
0.014624431729316711,
0.07023242115974426,
0.06408316642045975,
-0.04767356440424919,
-0.016442248597741127,
-0.03095763362944126,
-0.0033107686322182417,
-0.11979462951421738,
-0.20715628564357758,
-0.022137418389320374,
-0.018478205427527428,
0.13247568905353546,
-0.1996348351240158,
0.03547767922282219,
-0.022009534761309624,
0.0769844651222229,
0.00915440358221531,
-0.009203110821545124,
-0.04112536087632179,
0.1006089597940445,
-0.033683616667985916,
-0.0494849719107151,
0.08460493385791779,
0.017445456236600876,
-0.08974425494670868,
-0.010026541538536549,
-0.1260373294353485,
0.1414521187543869,
0.1333712786436081,
-0.125062495470047,
-0.07347418367862701,
-0.012869348749518394,
-0.05217713490128517,
-0.03989126533269882,
-0.040883537381887436,
0.04535810276865959,
0.18881762027740479,
0.0006998516037128866,
0.15854158997535706,
-0.07079102098941803,
-0.042756374925374985,
0.017096158117055893,
-0.03485364094376564,
0.034560464322566986,
0.11619038134813309,
0.11101613193750381,
-0.07945899665355682,
0.13235031068325043,
0.14979423582553864,
-0.09426113218069077,
0.1329599916934967,
-0.04455898329615593,
-0.07883826643228531,
-0.01883394829928875,
-0.01687617599964142,
0.003037179820239544,
0.0787491425871849,
-0.12060259282588959,
-0.0006215069442987442,
0.018558118492364883,
0.030713599175214767,
0.02561996318399906,
-0.23275397717952728,
-0.03194938972592354,
0.03767895698547363,
-0.05134912580251694,
-0.019224142655730247,
-0.024468233808875084,
0.01668376661837101,
0.1060754582285881,
-0.005780664272606373,
-0.07749290019273758,
0.025561144575476646,
0.0007406333461403847,
-0.07727126777172089,
0.20280075073242188,
-0.0948910340666771,
-0.16438893973827362,
-0.11776162683963776,
-0.08450033515691757,
-0.033972568809986115,
0.004737980663776398,
0.07276551425457001,
-0.08331981301307678,
-0.023370802402496338,
-0.06590715050697327,
0.03247026354074478,
-0.0031086390372365713,
0.013540412299335003,
-0.0034215173218399286,
0.0018788741435855627,
0.0756073147058487,
-0.11442475765943527,
-0.004830104298889637,
-0.04441869258880615,
-0.06595362722873688,
0.054533060640096664,
0.03986917436122894,
0.1186375841498375,
0.1548691987991333,
-0.024842580780386925,
0.006325635127723217,
-0.031116966158151627,
0.21491365134716034,
-0.06524353474378586,
-0.02686886489391327,
0.14510713517665863,
-0.005088303703814745,
0.059220824390649796,
0.11008308082818985,
0.06758126616477966,
-0.08248697966337204,
0.011944560334086418,
0.031643837690353394,
-0.02990834228694439,
-0.2331797033548355,
-0.04152102395892143,
-0.052740249782800674,
-0.025021586567163467,
0.08995120972394943,
0.021543648093938828,
0.04736454412341118,
0.05815931782126427,
0.037492845207452774,
0.05937453359365463,
-0.019225820899009705,
0.06135893985629082,
0.1426125019788742,
0.04367760941386223,
0.1375243067741394,
-0.040608856827020645,
-0.06634660065174103,
0.04076625406742096,
-0.003692042315378785,
0.2279617339372635,
0.0077956500463187695,
0.14590901136398315,
0.0642782673239708,
0.17274583876132965,
0.004441489465534687,
0.0701407641172409,
0.011776848696172237,
-0.029379138723015785,
-0.02203819528222084,
-0.03802997246384621,
-0.03693774342536926,
0.015029317699372768,
-0.05740036442875862,
0.03606141358613968,
-0.11669307947158813,
-0.021242037415504456,
0.04223807156085968,
0.27904757857322693,
0.022345468401908875,
-0.3133646547794342,
-0.07816466689109802,
-0.0009326481376774609,
-0.04115397483110428,
-0.017890676856040955,
0.021369846537709236,
0.0849958136677742,
-0.0993187204003334,
0.03384238854050636,
-0.07123329490423203,
0.11041678488254547,
-0.048446159809827805,
0.04948434606194496,
0.05271048843860626,
0.0968284085392952,
0.011829374358057976,
0.07941898703575134,
-0.3274395167827606,
0.28487005829811096,
-0.001580566051416099,
0.06564512848854065,
-0.07053191214799881,
0.0022476811427623034,
0.037615302950143814,
0.03592175617814064,
0.036667272448539734,
-0.019895948469638824,
-0.07371751219034195,
-0.19453710317611694,
-0.055536698549985886,
0.030776238068938255,
0.09176211804151535,
-0.0007979092188179493,
0.10239112377166748,
-0.03666335344314575,
0.017148464918136597,
0.07342343777418137,
-0.012709137052297592,
-0.08865892142057419,
-0.09956417977809906,
-0.00429221335798502,
0.02834678627550602,
-0.015119197778403759,
-0.07138216495513916,
-0.11358734965324402,
-0.10876870900392532,
0.15254294872283936,
0.016884198412299156,
-0.02014952339231968,
-0.11183442175388336,
0.0836537554860115,
0.08581627160310745,
-0.08483729511499405,
0.035657159984111786,
0.011061514727771282,
0.06761960685253143,
0.02494577132165432,
-0.0650143101811409,
0.11309472471475601,
-0.05951014161109924,
-0.15685008466243744,
-0.06122062727808952,
0.09008099883794785,
0.03560839220881462,
0.0689372569322586,
-0.011277029290795326,
0.018840542063117027,
-0.04258321225643158,
-0.08181310445070267,
0.015118648298084736,
-0.012172387912869453,
0.055099546909332275,
0.017290908843278885,
-0.06719359010457993,
0.016430864110589027,
-0.06926757097244263,
-0.05912899971008301,
0.20074689388275146,
0.24205602705478668,
-0.09248695522546768,
0.03700857609510422,
0.05333452671766281,
-0.07941559702157974,
-0.1849460005760193,
0.029092177748680115,
0.061317164450883865,
0.0046570925042033195,
0.05488380044698715,
-0.19388100504875183,
0.0865083560347557,
0.11168190836906433,
-0.013554811477661133,
0.0893917977809906,
-0.34763574600219727,
-0.12917818129062653,
0.11600849777460098,
0.15014195442199707,
0.10085119307041168,
-0.16059812903404236,
-0.022164834663271904,
-0.0261512640863657,
-0.11945224553346634,
0.1045541912317276,
-0.10116157680749893,
0.1277056336402893,
-0.021617058664560318,
0.0948023647069931,
0.004748436156660318,
-0.054928869009017944,
0.1132442057132721,
-0.013233612291514874,
0.09570200741291046,
-0.06914517283439636,
0.015387725085020065,
0.04824453219771385,
-0.037530042231082916,
0.008845441974699497,
-0.08513923734426498,
0.025145044550299644,
-0.07850625365972519,
-0.019873928278684616,
-0.08095719665288925,
0.03225196525454521,
-0.03423912823200226,
-0.05450601130723953,
-0.027476105839014053,
0.023596445098519325,
0.05833632871508598,
-0.010757312178611755,
0.12349937111139297,
-0.001488467794843018,
0.16653169691562653,
0.11668171733617783,
0.06793302297592163,
-0.06292713433504105,
-0.04252048209309578,
-0.018778536468744278,
-0.02046940103173256,
0.04895153269171715,
-0.12951982021331787,
0.030880825594067574,
0.14470389485359192,
0.012232869863510132,
0.14289192855358124,
0.07359272241592407,
-0.03637785091996193,
0.021347787231206894,
0.05786600708961487,
-0.14822165668010712,
-0.09396645426750183,
-0.0001800787722459063,
-0.014987001195549965,
-0.08235423266887665,
0.02869078516960144,
0.11042387783527374,
-0.06410638988018036,
-0.01557688508182764,
-0.006117052864283323,
0.006403953768312931,
-0.052458904683589935,
0.2050091177225113,
0.04262058809399605,
0.04175184294581413,
-0.09979036450386047,
0.07624100893735886,
0.07054848223924637,
-0.08902089297771454,
0.01619834080338478,
0.10298233479261398,
-0.07173021882772446,
-0.04470965638756752,
0.10295794159173965,
0.18799014389514923,
-0.0716748759150505,
-0.054219335317611694,
-0.14754971861839294,
-0.1280144602060318,
0.08343587070703506,
0.16131415963172913,
0.09313396364450455,
0.007549448404461145,
-0.05327737703919411,
0.009944615885615349,
-0.1177566647529602,
0.08184123784303665,
0.05848805978894234,
0.06333864480257034,
-0.11996172368526459,
0.17138898372650146,
0.017737073823809624,
0.03105694241821766,
-0.01722029596567154,
0.01946418359875679,
-0.09705539792776108,
0.019925499334931374,
-0.15876075625419617,
-0.029074162244796753,
-0.020942136645317078,
0.0009028622880578041,
-0.008452793583273888,
-0.05427321046590805,
-0.05392909422516823,
0.015804003924131393,
-0.12008407711982727,
-0.030495576560497284,
0.011729372665286064,
0.053578976541757584,
-0.12258152663707733,
-0.04335549473762512,
0.027214687317609787,
-0.05901143327355385,
0.05841781198978424,
0.037109021097421646,
0.0103760976344347,
0.05331910401582718,
-0.1557071954011917,
-0.0023270107340067625,
0.055356234312057495,
0.015400429256260395,
0.05510358512401581,
-0.10649142414331436,
-0.011317585594952106,
0.01198115386068821,
0.06615757942199707,
0.011833176016807556,
0.07519668340682983,
-0.1325407326221466,
-0.01523352786898613,
-0.019456129521131516,
-0.09154514223337173,
-0.05939340218901634,
0.033159635961055756,
0.0717712789773941,
0.02496091090142727,
0.18945281207561493,
-0.08803152292966843,
0.049699075520038605,
-0.21472428739070892,
0.0038321211468428373,
-0.014470343478024006,
-0.11019500344991684,
-0.1173701286315918,
-0.07452867925167084,
0.06759625673294067,
-0.05251024663448334,
0.13292908668518066,
0.021037813276052475,
0.057140760123729706,
0.0304748322814703,
-0.038406092673540115,
0.00577499158680439,
0.019265905022621155,
0.20587655901908875,
0.03435846418142319,
-0.03219500556588173,
0.06764813512563705,
0.05572960898280144,
0.08600502461194992,
0.1285698413848877,
0.19800247251987457,
0.1590302437543869,
0.021843446418642998,
0.07918921858072281,
0.04061319679021835,
-0.05178162455558777,
-0.13923367857933044,
0.047158997505903244,
-0.02539406530559063,
0.10414725542068481,
-0.03412703052163124,
0.2301473766565323,
0.07381507754325867,
-0.16665711998939514,
0.060820817947387695,
-0.05856316536664963,
-0.08320159465074539,
-0.10873010754585266,
-0.04351791739463806,
-0.0879126712679863,
-0.14928403496742249,
-0.008818564005196095,
-0.10968198627233505,
0.04137770086526871,
0.09679077565670013,
0.012256140820682049,
-0.024670546874403954,
0.12618574500083923,
0.041815049946308136,
0.004109008237719536,
0.04980763792991638,
-0.0038069842848926783,
-0.017407245934009552,
-0.10616111010313034,
-0.07752761244773865,
-0.002031184732913971,
-0.00035701438901014626,
0.03581426292657852,
-0.03844163566827774,
-0.06307931244373322,
0.036188434809446335,
-0.039466917514801025,
-0.10083185881376266,
0.018350370228290558,
0.016830839216709137,
0.07902336120605469,
0.06808411329984665,
0.013597323559224606,
0.006269005127251148,
-0.00786199513822794,
0.23300839960575104,
-0.07073984295129776,
-0.10419963300228119,
-0.09793763607740402,
0.26484137773513794,
0.033428456634283066,
-0.01740160398185253,
0.025313038378953934,
-0.057432834059000015,
-0.002584831090644002,
0.24964624643325806,
0.19588227570056915,
-0.09458409249782562,
-0.015881314873695374,
0.00861701462417841,
-0.010505443438887596,
-0.02420739084482193,
0.11767400801181793,
0.1486971229314804,
0.03899753838777542,
-0.10414429008960724,
-0.03259214013814926,
-0.0585840567946434,
-0.011406639590859413,
-0.05947783589363098,
0.06826688349246979,
0.032769493758678436,
0.001064255484379828,
-0.029770469292998314,
0.06272461265325546,
-0.05815092474222183,
-0.08344247192144394,
0.01808124966919422,
-0.20400370657444,
-0.16124103963375092,
-0.016623472794890404,
0.11785387992858887,
0.0013436462031677365,
0.05341049283742905,
-0.026841215789318085,
0.017995573580265045,
0.0729396715760231,
-0.026499157771468163,
-0.07101095467805862,
-0.08631309121847153,
0.09626344591379166,
-0.13334153592586517,
0.18573424220085144,
-0.043481599539518356,
0.05581410974264145,
0.12740737199783325,
0.06772097200155258,
-0.0697949156165123,
0.07902620732784271,
0.03936053067445755,
-0.07416538894176483,
0.03008655644953251,
0.0955444723367691,
-0.03669044002890587,
0.0620080940425396,
0.04833829775452614,
-0.12803326547145844,
0.0352826789021492,
-0.08036347478628159,
-0.050406068563461304,
-0.020906364545226097,
-0.043841950595378876,
-0.05225076898932457,
0.12574462592601776,
0.21636784076690674,
-0.027452873066067696,
0.019237477332353592,
-0.08286303281784058,
0.0032589060720056295,
0.05018598213791847,
0.04007723182439804,
-0.07625171542167664,
-0.2329988181591034,
0.003988604061305523,
0.06324660778045654,
-0.008761893957853317,
-0.24502845108509064,
-0.10347574204206467,
0.0017178135458379984,
-0.07943646609783173,
-0.1006859615445137,
0.09682606160640717,
0.08097605407238007,
0.04936699941754341,
-0.05090503767132759,
-0.0955188050866127,
-0.07206091284751892,
0.16388894617557526,
-0.14609870314598083,
-0.07747668027877808
] |
null | null |
transformers
|
# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)
## Usage (Amazon SageMaker inference applicable)
It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.
### inference_korsts.py
```python
import json
import sys
import logging
import torch
from torch import nn
from transformers import ElectraConfig
from transformers import ElectraModel, AutoTokenizer, ElectraTokenizer, ElectraForSequenceClassification
logging.basicConfig(
level=logging.INFO,
format='[{%(filename)s:%(lineno)d} %(levelname)s - %(message)s',
handlers=[
logging.FileHandler(filename='tmp.log'),
logging.StreamHandler(sys.stdout)
]
)
logger = logging.getLogger(__name__)
max_seq_length = 128
tokenizer = AutoTokenizer.from_pretrained("daekeun-ml/koelectra-small-v3-korsts")
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
# Huggingface pre-trained model: 'monologg/koelectra-small-v3-discriminator'
def model_fn(model_path):
####
# If you have your own trained model
# Huggingface pre-trained model: 'monologg/koelectra-small-v3-discriminator'
####
#config = ElectraConfig.from_json_file(f'{model_path}/config.json')
#model = ElectraForSequenceClassification.from_pretrained(f'{model_path}/model.pth', config=config)
model = ElectraForSequenceClassification.from_pretrained('daekeun-ml/koelectra-small-v3-korsts')
model.to(device)
return model
def input_fn(input_data, content_type="application/jsonlines"):
data_str = input_data.decode("utf-8")
jsonlines = data_str.split("\n")
transformed_inputs = []
for jsonline in jsonlines:
text = json.loads(jsonline)["text"]
logger.info("input text: {}".format(text))
encode_plus_token = tokenizer.encode_plus(
text,
max_length=max_seq_length,
add_special_tokens=True,
return_token_type_ids=False,
padding="max_length",
return_attention_mask=True,
return_tensors="pt",
truncation=True,
)
transformed_inputs.append(encode_plus_token)
return transformed_inputs
def predict_fn(transformed_inputs, model):
predicted_classes = []
for data in transformed_inputs:
data = data.to(device)
output = model(**data)
prediction_dict = {}
prediction_dict['score'] = output[0].squeeze().cpu().detach().numpy().tolist()
jsonline = json.dumps(prediction_dict)
logger.info("jsonline: {}".format(jsonline))
predicted_classes.append(jsonline)
predicted_classes_jsonlines = "\n".join(predicted_classes)
return predicted_classes_jsonlines
def output_fn(outputs, accept="application/jsonlines"):
return outputs, accept
```
### test.py
```python
>>> from inference_korsts import model_fn, input_fn, predict_fn, output_fn
>>> with open('./samples/korsts.txt', mode='rb') as file:
>>> model_input_data = file.read()
>>> model = model_fn()
>>> transformed_inputs = input_fn(model_input_data)
>>> predicted_classes_jsonlines = predict_fn(transformed_inputs, model)
>>> model_outputs = output_fn(predicted_classes_jsonlines)
>>> print(model_outputs[0])
[{inference_korsts.py:44} INFO - input text: ['맛있는 라면을 먹고 싶어요', '후루룩 쩝쩝 후루룩 쩝쩝 맛좋은 라면']
[{inference_korsts.py:44} INFO - input text: ['뽀로로는 내친구', '머신러닝은 러닝머신이 아닙니다.']
[{inference_korsts.py:71} INFO - jsonline: {"score": 4.786738872528076}
[{inference_korsts.py:71} INFO - jsonline: {"score": 0.2319069355726242}
{"score": 4.786738872528076}
{"score": 0.2319069355726242}
```
### Sample data (samples/korsts.txt)
```
{"text": ["맛있는 라면을 먹고 싶어요", "후루룩 쩝쩝 후루룩 쩝쩝 맛좋은 라면"]}
{"text": ["뽀로로는 내친구", "머신러닝은 러닝머신이 아닙니다."]}
```
## References
- KoELECTRA: https://github.com/monologg/KoELECTRA
- KorNLI and KorSTS Dataset: https://github.com/kakaobrain/KorNLUDatasets
|
{"language": ["ko"], "license": "cc-by-4.0", "tags": ["sentence-similarity", "transformers"], "datasets": ["korsts"], "metrics": ["accuracy", "f1", "precision", "recall"], "pipeline_tag": "sentence-similarity"}
|
sentence-similarity
|
daekeun-ml/koelectra-small-v3-korsts
|
[
"transformers",
"pytorch",
"electra",
"text-classification",
"sentence-similarity",
"ko",
"dataset:korsts",
"license:cc-by-4.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ko"
] |
TAGS
#transformers #pytorch #electra #text-classification #sentence-similarity #ko #dataset-korsts #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us
|
# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)
## Usage (Amazon SageMaker inference applicable)
It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.
### inference_korsts.py
### URL
### Sample data (samples/URL)
## References
- KoELECTRA: URL
- KorNLI and KorSTS Dataset: URL
|
[
"# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)",
"## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.",
"### inference_korsts.py",
"### URL",
"### Sample data (samples/URL)",
"## References\n- KoELECTRA: URL\n- KorNLI and KorSTS Dataset: URL"
] |
[
"TAGS\n#transformers #pytorch #electra #text-classification #sentence-similarity #ko #dataset-korsts #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us \n",
"# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)",
"## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.",
"### inference_korsts.py",
"### URL",
"### Sample data (samples/URL)",
"## References\n- KoELECTRA: URL\n- KorNLI and KorSTS Dataset: URL"
] |
[
62,
31,
49,
10,
3,
11,
22
] |
[
"passage: TAGS\n#transformers #pytorch #electra #text-classification #sentence-similarity #ko #dataset-korsts #license-cc-by-4.0 #autotrain_compatible #endpoints_compatible #region-us \n# Similarity between two sentences (fine-tuning with KoELECTRA-Small-v3 model and KorSTS dataset)## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.### inference_korsts.py### URL### Sample data (samples/URL)## References\n- KoELECTRA: URL\n- KorNLI and KorSTS Dataset: URL"
] |
[
-0.0047347466461360455,
0.06698337197303772,
-0.0034649649169296026,
0.03054649755358696,
0.11573665589094162,
0.024143967777490616,
0.06707829236984253,
0.12878170609474182,
0.04050864651799202,
0.009739414788782597,
0.07820677757263184,
0.17552989721298218,
0.04694357514381409,
0.06502728909254074,
-0.10635994374752045,
-0.3591364026069641,
0.06702136248350143,
0.016598263755440712,
0.09348224848508835,
0.10162772983312607,
0.15506428480148315,
-0.054790496826171875,
0.0952109768986702,
0.026718780398368835,
-0.020608941093087196,
-0.026242922991514206,
0.06476584821939468,
-0.13329625129699707,
0.10557737946510315,
0.027673371136188507,
0.08154527842998505,
0.0254717618227005,
-0.002081052865833044,
-0.2279682606458664,
0.031787388026714325,
-0.024102959781885147,
0.037753842771053314,
0.021855339407920837,
-0.003984875977039337,
-0.20847927033901215,
0.12718260288238525,
-0.04636581242084503,
0.07798516750335693,
-0.011021039448678493,
-0.10081597417593002,
0.03733521327376366,
-0.07793368399143219,
0.06660270690917969,
0.1535687893629074,
0.04583572968840599,
0.032869063317775726,
0.04321423918008804,
-0.07960789650678635,
0.09384794533252716,
0.13016077876091003,
-0.1840221881866455,
-0.04823864996433258,
0.08110043406486511,
0.0010659644613042474,
0.030622737482190132,
-0.029681464657187462,
0.04665793851017952,
-0.016974058002233505,
0.03591008484363556,
-0.01858513429760933,
-0.09455232322216034,
-0.1471092253923416,
0.05300560221076012,
-0.07636331021785736,
-0.10084305703639984,
0.2835908532142639,
-0.03372899442911148,
0.024038629606366158,
-0.02275840938091278,
-0.03667881339788437,
0.057394422590732574,
-0.04198109731078148,
0.030336866155266762,
-0.08690455555915833,
-0.012847846373915672,
-0.07601921260356903,
0.01795308105647564,
-0.10957812517881393,
-0.022847551852464676,
-0.09131006896495819,
0.16773216426372528,
0.029936296865344048,
0.0419977605342865,
-0.09686123579740524,
0.08896957337856293,
-0.10706678777933121,
-0.1401028335094452,
-0.046550218015909195,
-0.11809735745191574,
0.01889953389763832,
-0.040724221616983414,
-0.08290157467126846,
-0.027615023776888847,
0.0985589548945427,
0.24734316766262054,
0.09044591337442398,
0.04442659392952919,
0.06372855603694916,
0.03930523619055748,
0.04060497507452965,
0.14790977537631989,
-0.017625711858272552,
-0.09329340606927872,
0.03682626038789749,
0.0008369545103050768,
-0.008333789184689522,
-0.03479443117976189,
-0.09258126467466354,
-0.1131964847445488,
0.07100503891706467,
0.007299143821001053,
0.03272382542490959,
0.1145659014582634,
-0.004818365443497896,
-0.03675428032875061,
0.012663561850786209,
-0.07788670808076859,
-0.0026487845461815596,
-0.027539394795894623,
-0.048811428248882294,
0.09620541334152222,
-0.0013260728446766734,
0.08587871491909027,
-0.05820903927087784,
0.1318535953760147,
-0.033025726675987244,
0.05327574163675308,
0.012741033919155598,
-0.06498956680297852,
0.028794562444090843,
-0.13031615316867828,
0.05637364089488983,
-0.2128443717956543,
-0.19766803085803986,
-0.022274114191532135,
0.0786905512213707,
-0.019139518961310387,
-0.058186691254377365,
-0.06133083999156952,
0.01603388600051403,
0.0692712739109993,
-0.016172679141163826,
-0.1240840032696724,
-0.09123723208904266,
0.021221475675702095,
-0.08880855143070221,
0.069794662296772,
-0.044349510222673416,
0.03370276838541031,
-0.1464269906282425,
-0.005079986993223429,
-0.06594758480787277,
0.028122151270508766,
-0.0701879933476448,
0.04013703018426895,
-0.13636989891529083,
-0.031493909657001495,
-0.03477768227458,
-0.016343610361218452,
0.0021623920183628798,
0.09788310527801514,
-0.1359030306339264,
0.0012926706112921238,
0.044969502836465836,
-0.05896329507231712,
-0.1190456673502922,
0.11680924892425537,
-0.04114580526947975,
0.03707657381892204,
0.10500764101743698,
0.1058821976184845,
0.18054896593093872,
-0.002578007522970438,
0.029525861144065857,
0.04436030611395836,
-0.0386032834649086,
0.056355901062488556,
0.027537574991583824,
0.014164774678647518,
-0.10867151618003845,
0.0594293549656868,
0.04574683681130409,
0.05342676118016243,
-0.036610864102840424,
-0.04707600548863411,
-0.04601868614554405,
0.009590329602360725,
0.02757030539214611,
-0.04460293799638748,
0.001990234712138772,
-0.00813654251396656,
-0.007960189133882523,
0.03975597769021988,
0.08135986328125,
-0.05554728955030441,
0.03506964445114136,
-0.05718698725104332,
0.14690831303596497,
-0.09786899387836456,
-0.008885247632861137,
-0.12564340233802795,
-0.015747999772429466,
-0.019216695800423622,
0.1432356834411621,
0.020975954830646515,
-0.035364437848329544,
0.020245637744665146,
-0.050497423857450485,
-0.016525186598300934,
0.01811681129038334,
0.09831929951906204,
0.06381603330373764,
-0.04247910901904106,
-0.06151831895112991,
0.033495888113975525,
-0.011487774550914764,
-0.0011394458124414086,
-0.024601496756076813,
-0.022517437115311623,
0.08067267388105392,
0.14624866843223572,
-0.04694090411067009,
0.07153759151697159,
0.04039362445473671,
0.03008866123855114,
-0.07035024464130402,
0.01813081093132496,
0.06603799015283585,
-0.02687647007405758,
-0.055146921426057816,
0.21593059599399567,
-0.10926887392997742,
-0.10090675204992294,
0.07466711103916168,
-0.05146265774965286,
-0.014075319282710552,
-0.10754796862602234,
-0.02391841448843479,
-0.011808446608483791,
-0.08884567022323608,
-0.02479969896376133,
0.19303607940673828,
0.02603609301149845,
0.10398738831281662,
-0.07134929299354553,
-0.02641408145427704,
0.030620869249105453,
-0.10614041984081268,
0.003890802152454853,
0.05519390106201172,
0.028778119012713432,
-0.18245956301689148,
0.09254466742277145,
0.09362498670816422,
-0.0808616653084755,
0.06530854851007462,
-0.03128194436430931,
-0.038553688675165176,
-0.00008924516441766173,
-0.015624449588358402,
-0.015844881534576416,
0.012938220985233784,
-0.04051800072193146,
-0.0013284505112096667,
0.12480886280536652,
0.06790155172348022,
0.0033283792436122894,
-0.11112239211797714,
0.004745683167129755,
0.02383899688720703,
0.01764102466404438,
-0.012023720890283585,
0.09021028131246567,
0.08010226488113403,
0.10574827343225479,
0.015115800313651562,
-0.08780558407306671,
0.02556663565337658,
0.014809314161539078,
-0.05196920409798622,
0.21710854768753052,
-0.06904268264770508,
-0.1560797542333603,
-0.1171770840883255,
-0.07013901323080063,
-0.06696134060621262,
-0.0026401132345199585,
0.061644598841667175,
0.031209217384457588,
-0.05773366615176201,
-0.06615184992551804,
0.01138634979724884,
0.04746332764625549,
-0.008994297124445438,
-0.16440294682979584,
0.0028075878508388996,
-0.09015470743179321,
-0.06607478857040405,
-0.0730118453502655,
0.03814960643649101,
-0.008454028517007828,
0.132660374045372,
-0.08826610445976257,
0.05384072661399841,
0.11824871599674225,
-0.06667730957269669,
0.021299075335264206,
-0.02919897437095642,
0.18274137377738953,
-0.03430095687508583,
0.07659627497196198,
0.19104690849781036,
-0.07248382270336151,
0.0771467387676239,
0.20081181824207306,
-0.010332014411687851,
-0.008540507405996323,
0.05736943706870079,
-0.015466660261154175,
-0.07924813032150269,
-0.14250971376895905,
-0.09916990250349045,
-0.0741472914814949,
0.07111512869596481,
0.11465084552764893,
-0.005645128898322582,
0.07468266040086746,
0.1460951268672943,
-0.06931548565626144,
0.037902481853961945,
0.10206445306539536,
0.14787845313549042,
0.1948300302028656,
0.07458969205617905,
0.11542508006095886,
0.0032028353307396173,
-0.08436284214258194,
0.018628671765327454,
0.11525445431470871,
0.13860784471035004,
-0.024605561047792435,
0.027958540245890617,
0.0825359970331192,
-0.013789626769721508,
0.12659163773059845,
0.05612264573574066,
-0.02669862098991871,
0.03061514161527157,
-0.025306299328804016,
-0.04269169643521309,
-0.049560461193323135,
0.10885930806398392,
-0.035389360040426254,
-0.03067903034389019,
-0.032764460891485214,
0.010890682227909565,
0.07840508222579956,
0.1526631861925125,
0.12943609058856964,
-0.2499615103006363,
-0.06466992199420929,
0.05711686983704567,
-0.05146053060889244,
-0.05732480436563492,
0.08361682295799255,
-0.02486833557486534,
-0.16373465955257416,
0.09590458124876022,
-0.044703159481287,
0.12171551585197449,
-0.10072144865989685,
-0.022969935089349747,
-0.06808657944202423,
-0.044528715312480927,
-0.020767943933606148,
0.11286257952451706,
-0.17903868854045868,
0.17581471800804138,
0.013670187443494797,
-0.02859630435705185,
-0.09505024552345276,
-0.030122527852654457,
0.049085695296525955,
0.13617222011089325,
0.04900659620761871,
0.026307910680770874,
-0.20467129349708557,
-0.13605576753616333,
-0.06242190673947334,
0.05721459165215492,
0.038968123495578766,
-0.09793967753648758,
0.0773635134100914,
-0.09148816764354706,
-0.02434861846268177,
-0.04877699539065361,
-0.06265369802713394,
-0.05520952120423317,
-0.16063667833805084,
0.02957036904990673,
0.040427159518003464,
0.09493216872215271,
0.005810496862977743,
-0.013845046982169151,
-0.020243311300873756,
0.020127352327108383,
-0.04833899810910225,
-0.09718294441699982,
-0.10873610526323318,
-0.017401957884430885,
0.10564718395471573,
-0.09924653172492981,
0.04439390078186989,
-0.05166979879140854,
0.06383433938026428,
-0.006911689881235361,
-0.13142766058444977,
0.0016416975995525718,
-0.06476225703954697,
-0.10809382796287537,
0.0005319464835338295,
0.13058793544769287,
-0.02622077614068985,
0.0315069779753685,
0.06032245233654976,
0.09964428842067719,
-0.07464274764060974,
-0.05698271468281746,
-0.09677543491125107,
0.23291191458702087,
0.10774855315685272,
0.04343578964471817,
-0.0007272930815815926,
0.023972835391759872,
-0.10081593692302704,
0.015704168006777763,
0.19455617666244507,
0.05054980143904686,
-0.05172601342201233,
0.08483245968818665,
0.011247311718761921,
-0.062053464353084564,
-0.24973522126674652,
-0.11767413467168808,
-0.011349168606102467,
-0.00004544588955468498,
-0.07851158082485199,
0.0013691209023818374,
0.12632596492767334,
-0.0040120952762663364,
0.006229415535926819,
-0.059041742235422134,
-0.207528218626976,
-0.11996662616729736,
0.10177972167730331,
-0.001314297434873879,
0.1708412766456604,
-0.019580908119678497,
-0.025016143918037415,
-0.06228463351726532,
-0.18050846457481384,
0.17034892737865448,
0.05075979605317116,
0.08136950433254242,
-0.03735990449786186,
0.16268689930438995,
0.04100663959980011,
-0.019745321944355965,
0.12817302346229553,
0.027777165174484253,
0.03292261064052582,
-0.10049614310264587,
-0.09970618784427643,
-0.08059626817703247,
-0.07754673808813095,
0.11599671840667725,
-0.05627911910414696,
0.07178729772567749,
-0.10516287386417389,
-0.02294999733567238,
-0.08401598036289215,
0.08642921596765518,
0.04452158510684967,
-0.07071185857057571,
-0.09144745022058487,
0.008736158721148968,
0.10409601032733917,
-0.02113913744688034,
0.2978004217147827,
-0.017782825976610184,
0.023674875497817993,
0.04828701168298721,
0.12035909295082092,
-0.04376678541302681,
-0.0000375343079213053,
-0.040719639509916306,
-0.052203476428985596,
0.06972648203372955,
-0.18295533955097198,
0.059771209955215454,
0.12584635615348816,
-0.0007742413436062634,
0.08752132207155228,
0.030344625934958458,
-0.013708295300602913,
0.028277361765503883,
0.0678737461566925,
-0.1379864513874054,
-0.04028946906328201,
-0.0701940730214119,
0.03341656178236008,
0.002701857592910528,
-0.06182738021016121,
0.2008303552865982,
-0.112788125872612,
-0.06249849125742912,
-0.010240797884762287,
0.016884448006749153,
-0.030297618359327316,
0.1038190945982933,
0.043900832533836365,
0.05235017463564873,
-0.05591624602675438,
0.07384736835956573,
0.028866160660982132,
-0.15106546878814697,
0.047681279480457306,
-0.013696850277483463,
-0.1798030585050583,
-0.07290016114711761,
-0.021919790655374527,
0.03867126628756523,
-0.11491549015045166,
-0.06272778660058975,
-0.08546964079141617,
-0.05440026894211769,
0.01754038594663143,
0.26310229301452637,
0.07064638286828995,
0.06616653501987457,
-0.015819698572158813,
-0.04610447213053703,
-0.0499856099486351,
0.11068268865346909,
0.0801640972495079,
0.07269149273633957,
-0.13611233234405518,
-0.011610066518187523,
-0.06003193184733391,
0.09697762131690979,
-0.03542131558060646,
-0.06113288924098015,
-0.09681657701730728,
-0.002543934155255556,
-0.17329226434230804,
0.06583770364522934,
-0.19124391674995422,
-0.0146690234541893,
-0.008201994001865387,
-0.07730387151241302,
-0.026386970654129982,
-0.017039431259036064,
-0.0699889287352562,
0.020559048280119896,
-0.08522441983222961,
0.10399357229471207,
-0.0066791316494345665,
-0.05822314694523811,
0.09705992043018341,
-0.03706233948469162,
0.04381147027015686,
0.08670688420534134,
-0.06898224353790283,
0.03320098668336868,
-0.18770034611225128,
-0.047302138060331345,
0.0685649961233139,
0.037804849445819855,
0.03260720893740654,
-0.14315827190876007,
0.05502789095044136,
0.01958625018596649,
0.023919573053717613,
0.006578782573342323,
0.07972851395606995,
-0.1308744102716446,
-0.04521145299077034,
0.01997970975935459,
-0.10257163643836975,
-0.09267932921648026,
-0.06278030574321747,
0.0797363668680191,
0.10620792210102081,
0.18000546097755432,
-0.06776890903711319,
0.08586819469928741,
-0.06019099801778793,
0.00802935753017664,
-0.03433075174689293,
-0.1618131697177887,
-0.07327158004045486,
-0.10998503118753433,
0.008108248934149742,
-0.025836054235696793,
0.09860865771770477,
0.048317596316337585,
-0.012455254793167114,
0.02782413735985756,
0.07568619400262833,
0.1480725109577179,
0.06667952239513397,
0.08527477085590363,
0.06806197762489319,
-0.056923750787973404,
-0.09170914441347122,
0.050047848373651505,
0.0517127588391304,
0.001207811408676207,
0.006005277391523123,
0.12321988493204117,
0.061931923031806946,
0.07083293795585632,
0.034842077642679214,
0.060097772628068924,
0.03873613476753235,
-0.20612744987010956,
-0.030887609347701073,
0.07087163627147675,
-0.0239862073212862,
0.023396093398332596,
0.2147020846605301,
-0.1340850442647934,
0.02039594016969204,
-0.06484381854534149,
-0.08815586566925049,
-0.1295931190252304,
-0.3253213167190552,
-0.136638805270195,
-0.10740829259157181,
0.024640418589115143,
-0.127376526594162,
-0.006660169456154108,
0.14267998933792114,
0.012675908394157887,
0.002425525104627013,
0.07537668943405151,
-0.04466095194220543,
-0.08130687475204468,
0.09743896871805191,
-0.022760892286896706,
0.02442992851138115,
0.0988377258181572,
-0.041378796100616455,
-0.01671222411096096,
-0.09076816588640213,
-0.01748596504330635,
0.03795266151428223,
0.09094393253326416,
-0.03035118617117405,
-0.11668092757463455,
-0.09070106595754623,
0.017284706234931946,
-0.02157043106853962,
0.01052123960107565,
-0.03413090854883194,
0.08679433912038803,
0.029859639704227448,
0.07145636528730392,
0.32008370757102966,
-0.029655376449227333,
-0.19362281262874603,
-0.18223567306995392,
0.15175648033618927,
0.08199354261159897,
0.04329860210418701,
0.029489092528820038,
-0.05374959856271744,
-0.03846396505832672,
0.16031131148338318,
0.10079900175333023,
0.06540550291538239,
0.013679973781108856,
-0.04031546413898468,
0.02943997085094452,
0.06876830756664276,
0.041694946587085724,
0.0008601690060459077,
0.09930795431137085,
-0.06957045197486877,
-0.05109734088182449,
-0.022592982277274132,
-0.03198257088661194,
-0.06881614774465561,
0.07278982549905777,
0.018242783844470978,
-0.08260339498519897,
-0.04561510309576988,
0.1345381885766983,
-0.17034970223903656,
0.17451101541519165,
-0.10364153236150742,
-0.1771167367696762,
-0.11966773122549057,
-0.022080525755882263,
0.09588928520679474,
-0.0002662861661519855,
0.06859806180000305,
-0.017479099333286285,
-0.04662610962986946,
0.012261820025742054,
0.03096942976117134,
-0.12421666830778122,
0.030777614563703537,
0.056639038026332855,
0.04994822293519974,
-0.02273593097925186,
-0.014569342136383057,
0.12482354789972305,
0.11352427303791046,
0.061579033732414246,
-0.050142403692007065,
0.042101986706256866,
0.08091483265161514,
0.03684598207473755,
0.08040231466293335,
0.0838453471660614,
0.014814468100667,
0.038095902651548386,
0.12070711702108383,
-0.09384725242853165,
0.0716153085231781,
-0.018951235339045525,
0.03379041701555252,
-0.09398557990789413,
0.047419045120477676,
-0.02369660511612892,
0.11583822220563889,
0.09435641020536423,
-0.07216659188270569,
-0.0008860358502715826,
-0.05256715416908264,
-0.006230961065739393,
0.03302246704697609,
-0.07336342334747314,
-0.05973894149065018,
-0.09758800268173218,
-0.09193554520606995,
-0.0006679300568066537,
0.02428250014781952,
-0.058878093957901,
-0.02775461971759796,
-0.047869253903627396,
-0.024115782231092453,
-0.01947021484375,
0.14644072949886322,
0.015592047944664955,
-0.007385212928056717,
-0.033241450786590576,
-0.09923451393842697,
0.03018229454755783,
0.09124084562063217,
-0.1459631472826004,
-0.11965306103229523
] |
null | null |
transformers
|
# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)
## Usage (Amazon SageMaker inference applicable)
It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.
### inference_nsmc.py
```python
import json
import sys
import logging
import torch
from torch import nn
from transformers import ElectraConfig
from transformers import ElectraModel, AutoTokenizer, ElectraTokenizer, ElectraForSequenceClassification
logging.basicConfig(
level=logging.INFO,
format='[{%(filename)s:%(lineno)d} %(levelname)s - %(message)s',
handlers=[
logging.FileHandler(filename='tmp.log'),
logging.StreamHandler(sys.stdout)
]
)
logger = logging.getLogger(__name__)
max_seq_length = 128
classes = ['Neg', 'Pos']
tokenizer = AutoTokenizer.from_pretrained("daekeun-ml/koelectra-small-v3-nsmc")
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
def model_fn(model_path=None):
####
# If you have your own trained model
# Huggingface pre-trained model: 'monologg/koelectra-small-v3-discriminator'
####
#config = ElectraConfig.from_json_file(f'{model_path}/config.json')
#model = ElectraForSequenceClassification.from_pretrained(f'{model_path}/model.pth', config=config)
# Download model from the Huggingface hub
model = ElectraForSequenceClassification.from_pretrained('daekeun-ml/koelectra-small-v3-nsmc')
model.to(device)
return model
def input_fn(input_data, content_type="application/jsonlines"):
data_str = input_data.decode("utf-8")
jsonlines = data_str.split("\n")
transformed_inputs = []
for jsonline in jsonlines:
text = json.loads(jsonline)["text"][0]
logger.info("input text: {}".format(text))
encode_plus_token = tokenizer.encode_plus(
text,
max_length=max_seq_length,
add_special_tokens=True,
return_token_type_ids=False,
padding="max_length",
return_attention_mask=True,
return_tensors="pt",
truncation=True,
)
transformed_inputs.append(encode_plus_token)
return transformed_inputs
def predict_fn(transformed_inputs, model):
predicted_classes = []
for data in transformed_inputs:
data = data.to(device)
output = model(**data)
softmax_fn = nn.Softmax(dim=1)
softmax_output = softmax_fn(output[0])
_, prediction = torch.max(softmax_output, dim=1)
predicted_class_idx = prediction.item()
predicted_class = classes[predicted_class_idx]
score = softmax_output[0][predicted_class_idx]
logger.info("predicted_class: {}".format(predicted_class))
prediction_dict = {}
prediction_dict["predicted_label"] = predicted_class
prediction_dict['score'] = score.cpu().detach().numpy().tolist()
jsonline = json.dumps(prediction_dict)
logger.info("jsonline: {}".format(jsonline))
predicted_classes.append(jsonline)
predicted_classes_jsonlines = "\n".join(predicted_classes)
return predicted_classes_jsonlines
def output_fn(outputs, accept="application/jsonlines"):
return outputs, accept
```
### test.py
```python
>>> from inference_nsmc import model_fn, input_fn, predict_fn, output_fn
>>> with open('samples/nsmc.txt', mode='rb') as file:
>>> model_input_data = file.read()
>>> model = model_fn()
>>> transformed_inputs = input_fn(model_input_data)
>>> predicted_classes_jsonlines = predict_fn(transformed_inputs, model)
>>> model_outputs = output_fn(predicted_classes_jsonlines)
>>> print(model_outputs[0])
[{inference_nsmc.py:47} INFO - input text: 이 영화는 최고의 영화입니다
[{inference_nsmc.py:47} INFO - input text: 최악이에요. 배우의 연기력도 좋지 않고 내용도 너무 허접합니다
[{inference_nsmc.py:77} INFO - predicted_class: Pos
[{inference_nsmc.py:84} INFO - jsonline: {"predicted_label": "Pos", "score": 0.9619030952453613}
[{inference_nsmc.py:77} INFO - predicted_class: Neg
[{inference_nsmc.py:84} INFO - jsonline: {"predicted_label": "Neg", "score": 0.9994170665740967}
{"predicted_label": "Pos", "score": 0.9619030952453613}
{"predicted_label": "Neg", "score": 0.9994170665740967}
```
### Sample data (samples/nsmc.txt)
```
{"text": ["이 영화는 최고의 영화입니다"]}
{"text": ["최악이에요. 배우의 연기력도 좋지 않고 내용도 너무 허접합니다"]}
```
## References
- KoELECTRA: https://github.com/monologg/KoELECTRA
- Naver Sentiment Movie Corpus Dataset: https://github.com/e9t/nsmc
|
{"language": ["ko"], "license": "mit", "tags": ["classification"], "datasets": ["nsmc"], "metrics": ["accuracy", "f1", "precision", "recall- accuracy"], "widget": [{"text": "\ubd88\ud6c4\uc758 \uba85\uc791\uc785\ub2c8\ub2e4! \uc774\ub807\uac8c \uac10\ub3d9\uc801\uc778 \ub0b4\uc6a9\uc740 \ucc98\uc74c\uc774\uc5d0\uc694", "example_title": "Positive"}, {"text": "\uc2dc\uac04\uc774 \uc815\ub9d0 \uc544\uae5d\uc2b5\ub2c8\ub2e4. 10\uc810 \ub9cc\uc810\uc5d0 1\uc810\ub3c4 \uc544\uae4c\uc6cc\uc694..", "example_title": "Negative"}]}
|
text-classification
|
daekeun-ml/koelectra-small-v3-nsmc
|
[
"transformers",
"pytorch",
"electra",
"text-classification",
"classification",
"ko",
"dataset:nsmc",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[
"ko"
] |
TAGS
#transformers #pytorch #electra #text-classification #classification #ko #dataset-nsmc #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)
## Usage (Amazon SageMaker inference applicable)
It uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.
### inference_nsmc.py
### URL
### Sample data (samples/URL)
## References
- KoELECTRA: URL
- Naver Sentiment Movie Corpus Dataset: URL
|
[
"# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)",
"## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.",
"### inference_nsmc.py",
"### URL",
"### Sample data (samples/URL)",
"## References\n- KoELECTRA: URL\n- Naver Sentiment Movie Corpus Dataset: URL"
] |
[
"TAGS\n#transformers #pytorch #electra #text-classification #classification #ko #dataset-nsmc #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)",
"## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.",
"### inference_nsmc.py",
"### URL",
"### Sample data (samples/URL)",
"## References\n- KoELECTRA: URL\n- Naver Sentiment Movie Corpus Dataset: URL"
] |
[
54,
34,
49,
10,
3,
11,
21
] |
[
"passage: TAGS\n#transformers #pytorch #electra #text-classification #classification #ko #dataset-nsmc #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# Sentiment Binary Classification (fine-tuning with KoELECTRA-Small-v3 model and Naver Sentiment Movie Corpus dataset)## Usage (Amazon SageMaker inference applicable)\nIt uses the interface of the SageMaker Inference Toolkit as is, so it can be easily deployed to SageMaker Endpoint.### inference_nsmc.py### URL### Sample data (samples/URL)## References\n- KoELECTRA: URL\n- Naver Sentiment Movie Corpus Dataset: URL"
] |
[
-0.014786467887461185,
0.23098772764205933,
-0.0036969019565731287,
0.043974995613098145,
0.19116652011871338,
0.026381278410553932,
0.024669088423252106,
0.1588892936706543,
0.03526761382818222,
-0.005228316877037287,
0.0304886307567358,
0.14527076482772827,
0.05437978729605675,
0.08582165837287903,
-0.10416846722364426,
-0.3227616548538208,
-0.01728934980928898,
0.006590794771909714,
0.14946317672729492,
0.11417071521282196,
0.11091817170381546,
-0.07812947034835815,
0.13666318356990814,
-0.03172285109758377,
-0.06569137424230576,
-0.009367120452225208,
0.056371502578258514,
-0.10664635896682739,
0.09169881790876389,
0.012304577976465225,
0.046281736344099045,
-0.015329360961914062,
0.051237553358078,
-0.1811957061290741,
0.042533837258815765,
-0.039392128586769104,
0.03795422613620758,
0.017011111602187157,
0.05365443229675293,
-0.09442657232284546,
0.23869279026985168,
-0.0022331050131469965,
0.08739997446537018,
0.0023992094211280346,
-0.07222570478916168,
-0.05168024078011513,
-0.05181548744440079,
0.009890105575323105,
0.0014564723242074251,
0.10038107633590698,
0.0029059213120490313,
0.10647036880254745,
-0.08367267996072769,
0.08004851639270782,
0.11414217203855515,
-0.053483542054891586,
-0.062066301703453064,
0.045293353497982025,
0.027928823605179787,
0.09809951484203339,
-0.11523560434579849,
0.02320566214621067,
0.002267599105834961,
0.018671337515115738,
0.042476072907447815,
-0.09217848628759384,
-0.09051650762557983,
-0.033928725868463516,
-0.0652826577425003,
-0.11386865377426147,
0.28046083450317383,
0.00933451671153307,
-0.03818202391266823,
-0.0196248609572649,
0.0183839313685894,
-0.034560125321149826,
-0.08008775860071182,
0.04839877039194107,
-0.030686816200613976,
-0.03935310244560242,
0.01601087674498558,
0.06253810226917267,
-0.074776791036129,
-0.0019074787851423025,
-0.021523647010326385,
0.1210978776216507,
-0.0027828537859022617,
0.07793346047401428,
-0.14864975214004517,
0.056301772594451904,
-0.04611250385642052,
-0.12227950245141983,
-0.024495583027601242,
-0.0895082950592041,
0.0689246878027916,
-0.02250717207789421,
-0.007382403127849102,
0.04455043002963066,
0.009525391273200512,
0.1853436380624771,
0.050949517637491226,
0.050600290298461914,
0.004629630129784346,
0.061313629150390625,
0.11959389597177505,
0.11586995422840118,
-0.034441206604242325,
-0.1121017262339592,
0.07980175316333771,
-0.014027148485183716,
0.06617607176303864,
-0.01205277070403099,
-0.08080190420150757,
-0.03997378051280975,
0.020236141979694366,
0.04109319671988487,
0.00852673128247261,
0.08891777694225311,
-0.06397167593240738,
-0.07176671177148819,
0.04971950501203537,
-0.07102464884519577,
0.018057415261864662,
0.007768373936414719,
-0.10334452986717224,
0.1569424569606781,
-0.014089702628552914,
-0.006978312041610479,
-0.022347241640090942,
0.10470224171876907,
-0.03991285711526871,
0.056124258786439896,
-0.061078015714883804,
-0.06690812855958939,
0.11003544926643372,
-0.13919106125831604,
0.04884723573923111,
-0.19139941036701202,
-0.1646198332309723,
-0.010904600843787193,
0.030261967331171036,
-0.056530267000198364,
-0.04375704750418663,
-0.05320020765066147,
0.02929370477795601,
0.06283441185951233,
-0.014886650256812572,
-0.09688594937324524,
-0.08956416696310043,
0.02074241451919079,
-0.030373310670256615,
0.10341959446668625,
-0.06934195011854172,
0.03121611289680004,
-0.12863564491271973,
-0.00448768911883235,
-0.09640952199697495,
0.04089111089706421,
-0.07335107773542404,
0.08373302966356277,
-0.05496721342206001,
-0.06580646336078644,
-0.01001786533743143,
-0.015914592891931534,
-0.07118751108646393,
0.1373443752527237,
-0.20217573642730713,
-0.046861786395311356,
-0.028281623497605324,
-0.10239576548337936,
-0.14513595402240753,
0.13106723129749298,
-0.039178162813186646,
0.01235207449644804,
0.03497537598013878,
0.11990445852279663,
0.11705966293811798,
-0.03967606648802757,
-0.027553966268897057,
0.042570870369672775,
-0.08456195890903473,
-0.025789543986320496,
0.08167219161987305,
0.14871737360954285,
-0.04368972405791283,
0.03367901220917702,
-0.03292591869831085,
0.10548259317874908,
-0.037426333874464035,
-0.05085908994078636,
0.0035873963497579098,
0.011518720537424088,
-0.04178987070918083,
0.01580202579498291,
0.03008280321955681,
-0.00925477035343647,
-0.012116505764424801,
-0.03160947933793068,
0.1221301406621933,
-0.011837703175842762,
0.07135732471942902,
-0.09303443133831024,
0.1651427149772644,
-0.02724090963602066,
-0.0036860560066998005,
-0.14710354804992676,
0.017668763175606728,
-0.010094530880451202,
0.07047919929027557,
-0.00962329562753439,
-0.07307813316583633,
-0.01465658750385046,
-0.017254244536161423,
-0.031033474951982498,
0.0008580565336160362,
0.06378711014986038,
0.03154159337282181,
0.0540018193423748,
-0.1273265928030014,
-0.004456499591469765,
-0.06729898601770401,
0.000811090343631804,
-0.07384240627288818,
-0.01874241605401039,
0.132750004529953,
0.12240689992904663,
-0.006369890179485083,
-0.00783210713416338,
0.049473654478788376,
0.030738942325115204,
-0.05728776752948761,
-0.051307063549757004,
0.064403235912323,
-0.025377925485372543,
-0.03319472819566727,
0.12164656817913055,
-0.13363812863826752,
0.02177814580500126,
0.12502504885196686,
-0.11848224699497223,
-0.03276360407471657,
0.02552836947143078,
-0.029884299263358116,
0.03510993346571922,
-0.07254363596439362,
0.03736807033419609,
0.10348746180534363,
0.02615479938685894,
0.12066330760717392,
-0.04073816165328026,
-0.018006693571805954,
0.0507483147084713,
-0.08097687363624573,
-0.057009413838386536,
0.06834255903959274,
0.11465656012296677,
-0.10597960650920868,
0.12131784856319427,
0.084998719394207,
-0.08477777987718582,
0.15689244866371155,
0.030655760318040848,
-0.018503382802009583,
0.014791764318943024,
-0.06454315036535263,
0.011920582503080368,
0.03260134533047676,
-0.13153737783432007,
-0.035835497081279755,
0.07768581807613373,
-0.033894408494234085,
-0.03219524770975113,
-0.08948591351509094,
-0.0553504079580307,
0.006901813670992851,
0.04877856373786926,
0.019178541377186775,
0.07600676268339157,
0.03861811012029648,
0.13310495018959045,
0.022772567346692085,
-0.08138955384492874,
0.04758000373840332,
-0.031007003039121628,
-0.094902902841568,
0.1557968407869339,
-0.06803297251462936,
-0.29631781578063965,
-0.0711941346526146,
-0.05073485150933266,
0.014166298322379589,
0.0035305542405694723,
0.02276644855737686,
-0.03782098367810249,
-0.056603677570819855,
-0.08126793056726456,
-0.07534082233905792,
-0.0190375167876482,
-0.033699121326208115,
-0.09326005727052689,
0.03846671059727669,
-0.046766284853219986,
-0.037568021565675735,
-0.050626106560230255,
0.026703765615820885,
-0.05044163390994072,
0.10584945976734161,
-0.04423787444829941,
0.09241622686386108,
0.1309640109539032,
-0.08259017020463943,
0.03870631009340286,
-0.025226671248674393,
0.16189369559288025,
-0.08904612809419632,
0.03801584616303444,
0.20988872647285461,
-0.03143417090177536,
0.03776179999113083,
0.19987154006958008,
0.031123625114560127,
-0.03131222724914551,
0.049705274403095245,
-0.036030225455760956,
-0.08795103430747986,
-0.12180621922016144,
-0.09042128175497055,
-0.0922357589006424,
0.07835268974304199,
0.0872722938656807,
0.04816441237926483,
0.05605848506093025,
0.09430871903896332,
-0.04367098957300186,
-0.004400657955557108,
0.041434623301029205,
0.16593541204929352,
0.1631113886833191,
0.05736257880926132,
0.044560834765434265,
-0.01537561696022749,
-0.05304631590843201,
0.07876890897750854,
0.07296626269817352,
0.09604833275079727,
0.005352878011763096,
0.0928976759314537,
0.010334356687963009,
0.04836827889084816,
0.1373593658208847,
-0.030325090512633324,
0.016015619039535522,
0.01429983600974083,
-0.05249711498618126,
-0.018572034314274788,
-0.11972501873970032,
0.09557142853736877,
-0.022887766361236572,
-0.02819870598614216,
-0.03204416111111641,
0.035432539880275726,
0.0765092670917511,
0.04480331018567085,
0.016723861917853355,
-0.3110915422439575,
-0.029058784246444702,
0.09191053360700607,
0.011903975158929825,
-0.02918325550854206,
0.06207652390003204,
0.023702183738350868,
-0.18390966951847076,
0.11956043541431427,
-0.03245660662651062,
0.10681954771280289,
-0.15609285235404968,
-0.00995816569775343,
-0.02982153184711933,
-0.054476674646139145,
0.01972065679728985,
0.09647791087627411,
-0.15545733273029327,
0.12181592732667923,
-0.03986474126577377,
-0.002573393750935793,
-0.047919537872076035,
-0.05653129890561104,
0.08628342300653458,
0.14367470145225525,
0.16286471486091614,
0.039471160620450974,
-0.053029343485832214,
-0.14049936830997467,
0.005156225524842739,
0.04320294037461281,
-0.010403729975223541,
-0.0644695907831192,
0.03897703066468239,
-0.039867885410785675,
-0.027814829722046852,
-0.06708173453807831,
-0.04060749709606171,
-0.022815313190221786,
-0.13119754195213318,
0.004938212689012289,
0.07570550590753555,
0.049442317336797714,
0.009141473099589348,
-0.057906415313482285,
0.03911169245839119,
0.07864822447299957,
-0.09696775674819946,
-0.04998773708939552,
-0.13919226825237274,
0.038068223744630814,
0.001504798186942935,
-0.07123038172721863,
0.05631377175450325,
-0.044743649661540985,
0.12764042615890503,
-0.026508672162890434,
-0.13455730676651,
0.06253505498170853,
-0.07400516420602798,
-0.06157023087143898,
-0.03412189334630966,
0.07935240119695663,
-0.019478144124150276,
0.02893778868019581,
-0.022508971393108368,
0.03113342821598053,
-0.003281483892351389,
-0.08218267560005188,
-0.04332635551691055,
0.23220466077327728,
0.05083479359745979,
0.0308883897960186,
0.006852016784250736,
-0.09315716475248337,
-0.07964806258678436,
-0.009223462082445621,
0.1824350357055664,
0.07503814250230789,
-0.05245130881667137,
0.05619673430919647,
-0.024100812152028084,
-0.08530927449464798,
-0.24879080057144165,
-0.06541472673416138,
0.032622888684272766,
-0.06257314234972,
0.008126972243189812,
-0.079521544277668,
0.1191362664103508,
0.023088522255420685,
-0.0011491029290482402,
-0.04640189930796623,
-0.17867591977119446,
-0.09726769477128983,
0.09559180587530136,
0.11874502897262573,
0.173162579536438,
-0.05153924599289894,
-0.005980140995234251,
-0.05099638178944588,
-0.1322740912437439,
0.22506025433540344,
0.06786258518695831,
0.06760526448488235,
-0.06772254407405853,
0.16385787725448608,
-0.0031384630128741264,
-0.029496554285287857,
0.09183067828416824,
0.010700803250074387,
0.04521052539348602,
-0.07269557565450668,
-0.188497856259346,
-0.009021544829010963,
-0.04559504985809326,
0.12529966235160828,
-0.029790280386805534,
-0.020125404000282288,
-0.1261119246482849,
-0.027478961274027824,
-0.09274829924106598,
0.06811977922916412,
0.03727315366268158,
-0.05704324692487717,
-0.07082168757915497,
0.08009348809719086,
0.04946598410606384,
0.0025263880379498005,
0.2668389081954956,
-0.030607981607317924,
-0.033143866807222366,
0.10479908436536789,
0.1555032581090927,
0.035138119012117386,
-0.024531781673431396,
-0.03841047361493111,
-0.009145020507276058,
0.0764373242855072,
-0.12063004821538925,
0.022044280543923378,
0.13774782419204712,
0.0003969144308939576,
0.15088137984275818,
0.04101712629199028,
0.014169160276651382,
0.06289204955101013,
0.0947217047214508,
-0.07807566970586777,
-0.0571761354804039,
-0.05763532966375351,
0.11010994017124176,
-0.009399296715855598,
-0.10032239556312561,
0.09735772758722305,
-0.16167287528514862,
-0.05335620790719986,
-0.021309897303581238,
0.03953997418284416,
0.007975631393492222,
0.05481638014316559,
0.03743136301636696,
0.03387037664651871,
-0.11173570901155472,
0.10907525569200516,
-0.0036016006488353014,
-0.2169782519340515,
0.038497939705848694,
0.1441386193037033,
-0.16604465246200562,
-0.08114401996135712,
-0.03994854539632797,
0.09082093089818954,
-0.1583503931760788,
-0.0587514191865921,
-0.03980281576514244,
-0.06608689576387405,
0.047132838517427444,
0.14270298182964325,
0.04959120973944664,
0.07600488513708115,
-0.1158699169754982,
-0.030137954279780388,
-0.056378256529569626,
0.10941847413778305,
0.0634043961763382,
0.01950450800359249,
-0.1822422593832016,
-0.04500950127840042,
0.029273824766278267,
0.04198943451046944,
-0.0650462657213211,
-0.08770410716533661,
-0.03860621899366379,
0.02225322276353836,
-0.13650648295879364,
0.10901317745447159,
-0.17019037902355194,
0.02299000695347786,
-0.0316770002245903,
-0.018465714529156685,
-0.03802873566746712,
-0.017109045758843422,
-0.08351346850395203,
-0.005621641408652067,
-0.05213651806116104,
0.06288959085941315,
-0.02945508435368538,
-0.05376019328832626,
0.05875194072723389,
-0.04813392087817192,
0.050861623138189316,
0.12331602722406387,
-0.0334407277405262,
0.013660218566656113,
-0.21975450217723846,
-0.10679036378860474,
0.13727311789989471,
0.018242355436086655,
0.015714997425675392,
-0.056771282106637955,
0.043886441737413406,
0.055208127945661545,
-0.04339152202010155,
-0.025372449308633804,
0.07022681087255478,
-0.13117288053035736,
0.006901632063090801,
0.05656321346759796,
-0.10165068507194519,
-0.0882052332162857,
0.015734555199742317,
0.07635773718357086,
0.04053354263305664,
0.11111979186534882,
-0.0789937973022461,
0.06885860860347748,
-0.10345108807086945,
-0.003868721891194582,
-0.02096717059612274,
-0.10016138106584549,
-0.2332092523574829,
-0.039405278861522675,
0.028639476746320724,
-0.007526098750531673,
0.10084090381860733,
0.07173269987106323,
-0.06609849631786346,
0.03380194306373596,
0.07838063687086105,
0.09073581546545029,
0.023493321612477303,
0.08670688420534134,
0.08614721894264221,
-0.07485716789960861,
-0.00706704193726182,
-0.0027126066852360964,
0.05721583589911461,
0.07097835838794708,
0.016048923134803772,
0.12245520204305649,
0.09497840702533722,
0.04178434982895851,
0.009059916250407696,
0.06986033171415329,
0.024675630033016205,
-0.024894770234823227,
-0.049866192042827606,
0.07921014726161957,
-0.01722046174108982,
0.051554154604673386,
0.2042725682258606,
-0.06198057904839516,
0.03942282497882843,
-0.044406019151210785,
-0.059847284108400345,
-0.0870075449347496,
-0.3584827780723572,
-0.09696727246046066,
-0.1245264858007431,
0.031727660447359085,
-0.1506877988576889,
0.00655622873455286,
0.023536767810583115,
0.08007577806711197,
-0.08762671798467636,
0.07082799077033997,
-0.04342757537961006,
-0.09662953019142151,
0.15459978580474854,
0.03746257349848747,
0.007783498615026474,
0.033360451459884644,
0.035297419875860214,
0.001332932384684682,
-0.07772250473499298,
0.025722691789269447,
0.06348886340856552,
0.04566781222820282,
0.037181466817855835,
-0.050998471677303314,
-0.11010127514600754,
0.014773055911064148,
-0.01741762086749077,
-0.03971125930547714,
0.05199752002954483,
0.026144281029701233,
0.03823786973953247,
0.05858095362782478,
0.2553360164165497,
-0.04472222551703453,
-0.00649634562432766,
-0.0969150960445404,
0.21759743988513947,
0.024053269997239113,
0.03558950126171112,
0.0013247766764834523,
-0.05480372905731201,
-0.047108352184295654,
0.09435366839170456,
0.17692914605140686,
-0.005356321576982737,
-0.02801627665758133,
-0.05842481553554535,
0.028908487409353256,
0.020263560116291046,
0.0254106055945158,
0.03512173891067505,
0.13744257390499115,
-0.10185080766677856,
0.021580101922154427,
-0.04332271218299866,
-0.02771809510886669,
0.041409555822610855,
-0.032691553235054016,
0.05581994354724884,
-0.039984263479709625,
-0.14047420024871826,
0.08237512409687042,
-0.21469727158546448,
0.02788342721760273,
0.06989103555679321,
-0.1450483351945877,
-0.0662655457854271,
-0.0204871017485857,
0.03774838522076607,
0.07926451414823532,
0.05496862158179283,
0.0001118204090744257,
-0.056788574904203415,
0.027802549302577972,
0.06220212206244469,
-0.22225652635097504,
-0.05413636937737465,
0.07582198828458786,
-0.024560902267694473,
0.043195728212594986,
-0.08023624867200851,
0.07877560704946518,
0.05837387964129448,
0.03727540746331215,
-0.09071105718612671,
0.05120844021439552,
0.031071241945028305,
0.03232815861701965,
0.011723791249096394,
0.2005050927400589,
-0.0038877467159181833,
0.1197550892829895,
0.07168354839086533,
-0.0780734047293663,
0.004551077727228403,
-0.06502371281385422,
0.010169672779738903,
-0.07687671482563019,
0.029501907527446747,
-0.04932163283228874,
0.10986675322055817,
0.06903962790966034,
-0.04284125939011574,
-0.012502774596214294,
-0.07517591118812561,
-0.01890263520181179,
0.08637060225009918,
-0.06347278505563736,
0.027434589341282845,
-0.08509484678506851,
-0.05959940701723099,
-0.04201098904013634,
0.024801725521683693,
-0.09538348764181137,
-0.012066279537975788,
-0.14108064770698547,
-0.005874872673302889,
-0.03362392261624336,
0.052253611385822296,
-0.025058800354599953,
0.04349764063954353,
-0.039022840559482574,
-0.0920834019780159,
0.08402001857757568,
0.11899985373020172,
-0.10500457137823105,
-0.07568652927875519
] |
null | null |
transformers
|
# DALL·E Mini Model Card
This model card focuses on the model associated with the DALL·E mini space on Hugging Face, available [here](https://huggingface.co/spaces/dalle-mini/dalle-mini). The app is called “dalle-mini”, but incorporates “[DALL·E Mini](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy)’’ and “[DALL·E Mega](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mega-Training-Journal--VmlldzoxODMxMDI2)” models (further details on this distinction forthcoming).
The DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the [DALL·E Mega model card](https://huggingface.co/dalle-mini/dalle-mega).
## Model Details
* **Developed by:** Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh
* **Model type:** Transformer-based text-to-image generation model
* **Language(s):** English
* **License:** Apache 2.0
* **Model Description:** This is a model that can be used to generate images based on text prompts. As the model developers wrote in the [project report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy) about DALL·E mini, “OpenAI had the first impressive model for generating images with [DALL·E](https://openai.com/blog/dall-e/). DALL·E mini is an attempt at reproducing those results with an open-source model.”
* **Resources for more information:** See OpenAI’s website for more information about [DALL·E](https://openai.com/blog/dall-e/), including the [DALL·E model card](https://github.com/openai/DALL-E/blob/master/model_card.md). See the [project report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy) for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega [training journal](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mega-Training--VmlldzoxODMxMDI2#training-parameters).
* **Cite as:**
```bib text
@misc{Dayma_DALL·E_Mini_2021,
author = {Dayma, Boris and Patil, Suraj and Cuenca, Pedro and Saifullah, Khalid and Abraham, Tanishq and Lê Khắc, Phúc and Melas, Luke and Ghosh, Ritobrata},
doi = {10.5281/zenodo.5146400},
month = {7},
title = {DALL·E Mini},
url = {https://github.com/borisdayma/dalle-mini},
year = {2021}
}
```
## Uses
### Direct Use
The model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the [Misuse and Out-of-Scope Use](#misuse-malicious-use-and-out-of-scope-use) section.
### Downstream Use
The model could also be used for downstream use cases, including:
* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science
* Development of educational or creative tools
* Generation of artwork and use in design and artistic processes.
* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!
Downstream uses exclude the uses described in [Misuse and Out-of-Scope Use](#misuse-malicious-use-and-out-of-scope-use).
### Misuse, Malicious Use, and Out-of-Scope Use
The model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.
#### Out-of-Scope Use
The model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.
#### Misuse and Malicious Use
Using the model to generate content that is cruel to individuals is a misuse of this model. This includes:
* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.
* Intentionally promoting or propagating discriminatory content or harmful stereotypes.
* Impersonating individuals without their consent.
* Sexual content without consent of the people who might see it.
* Mis- and disinformation
* Representations of egregious violence and gore
* Sharing of copyrighted or licensed material in violation of its terms of use.
* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.
## Limitations and Bias
### Limitations
The model developers discuss the limitations of the model further in the DALL·E Mini [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA):
* Faces and people in general are not generated properly.
* Animals are usually unrealistic.
* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.
* The model has only been trained with English descriptions and will not perform as well in other languages
### Bias
**CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.**
The model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.
While the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.
Our current analyses demonstrate that:
* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented.
* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.
* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.
The [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA) discusses these issues in more detail, and also highlights potential sources of bias in the model development process.
### Limitations and Bias Recommendations
* Users (both direct and downstream) should be made aware of the biases and limitations.
* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.
* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.
## Training
### Training Data
The model developers used 3 datasets for the model:
* [Conceptual Captions Dataset](https://aclanthology.org/P18-1238/), which contains 3 million image and caption pairs.
* [Conceptual 12M](https://arxiv.org/abs/2102.08981), which contains 12 million image and caption pairs.
* The [OpenAI subset](https://github.com/openai/CLIP/blob/main/data/yfcc100m.md) of [YFCC100M](https://multimediacommons.wordpress.com/yfcc100m-core-dataset/), which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.
For fine-tuning the image encoder, a subset of 2 million images were used.
All images (about 15 million) were used for training the Seq2Seq model.
### Training Procedure
As described further in the [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA#our-dall-e-model-architecture) for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:
* Images are encoded through a [VQGAN](https://arxiv.org/abs/2012.09841) encoder, which turns images into a sequence of tokens.
* Descriptions are encoded through a [BART](https://arxiv.org/abs/1910.13461) encoder.
* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.
* Loss is the [softmax cross-entropy](https://wandb.ai/sauravm/Activation-Functions/reports/Activation-Functions-Softmax--VmlldzoxNDU1Njgy#%F0%9F%93%A2-softmax-+-cross-entropy-loss-(caution:-math-alert)) between the model prediction logits and the actual image encodings from the VQGAN.
The simplified training procedure for DALL·E Mega is as follows:
* **Hardware:** 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3
* **Optimizer:** Distributed Shampoo
* **Model Partition Specificiations:** 8 model parallel x 32 data parallel
* **Batch:** 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update
* **Learning rate:** warmup to 0.0001 for 10,000 steps and then kept constant until plateau
* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)
* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model.
* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.
There is more information about the full procedure and technical material in the DALL·E Mega [training journal](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mega-Training--VmlldzoxODMxMDI2#training-parameters).
## Evaluation Results
The model developers discuss their results extensively in their [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-Mini-Explained-with-Demo--Vmlldzo4NjIxODA#the-results-of-our-dall-e-experiment) for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with [DALL·E-pytorch](https://github.com/lucidrains/DALLE-pytorch), OpenAI’s [DALL·E](https://openai.com/blog/dall-e/), and models consisting of a generator coupled with the [CLIP neural network model](https://openai.com/blog/clip/).
For evaluation results related to DALL·E Mega, see this [technical report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini-Generate-images-from-any-text-prompt--VmlldzoyMDE4NDAy).
## Environmental Impact
### DALL·E Mini Estimated Emissions
*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*
Based on that information, we estimate the following CO2 emissions using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.
* **Hardware Type:** TPU v3-8
* **Hours used:** 72 (3 days)
* **Cloud Provider:** GCP (as mentioned in the technical report)
* **Compute Region:** us-east1 (provided by model developers)
* **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** 30.16 kg CO2 eq.
### DALL·E Mega Estimated Emissions
DALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.
* **Hardware Type:** TPU v3-256
* **Hours used:** 960 - 1080 hours (40-45 days)
* **Cloud Provider:** Unknown
* **Compute Region:** Unknown
* **Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid):** Unknown
## Citation
```bibtext
@misc{Dayma_DALL·E_Mini_2021,
author = {Dayma, Boris and Patil, Suraj and Cuenca, Pedro and Saifullah, Khalid and Abraham, Tanishq and Lê Khắc, Phúc and Melas, Luke and Ghosh, Ritobrata},
doi = {10.5281/zenodo.5146400},
month = {7},
title = {DALL·E Mini},
url = {https://github.com/borisdayma/dalle-mini},
year = {2021}
}
```
*This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*
|
{"language": "en", "license": "apache-2.0", "tags": ["text-to-image"], "inference": false, "co2_eq_emissions": {"emissions": 7540, "source": "MLCo2 Machine Learning Impact calculator", "geographical_location": "East USA", "hardware_used": "TPU v3-8"}, "model-index": [{"name": "dalle-mini", "results": []}]}
|
text-to-image
|
dalle-mini/dalle-mini
|
[
"transformers",
"jax",
"dallebart",
"text-to-image",
"en",
"arxiv:2102.08981",
"arxiv:2012.09841",
"arxiv:1910.13461",
"arxiv:1910.09700",
"license:apache-2.0",
"co2_eq_emissions",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[
"2102.08981",
"2012.09841",
"1910.13461",
"1910.09700"
] |
[
"en"
] |
TAGS
#transformers #jax #dallebart #text-to-image #en #arxiv-2102.08981 #arxiv-2012.09841 #arxiv-1910.13461 #arxiv-1910.09700 #license-apache-2.0 #co2_eq_emissions #has_space #region-us
|
# DALL·E Mini Model Card
This model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).
The DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.
## Model Details
* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh
* Model type: Transformer-based text-to-image generation model
* Language(s): English
* License: Apache 2.0
* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”
* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.
* Cite as:
## Uses
### Direct Use
The model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.
### Downstream Use
The model could also be used for downstream use cases, including:
* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science
* Development of educational or creative tools
* Generation of artwork and use in design and artistic processes.
* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!
Downstream uses exclude the uses described in Misuse and Out-of-Scope Use.
### Misuse, Malicious Use, and Out-of-Scope Use
The model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.
#### Out-of-Scope Use
The model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.
#### Misuse and Malicious Use
Using the model to generate content that is cruel to individuals is a misuse of this model. This includes:
* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.
* Intentionally promoting or propagating discriminatory content or harmful stereotypes.
* Impersonating individuals without their consent.
* Sexual content without consent of the people who might see it.
* Mis- and disinformation
* Representations of egregious violence and gore
* Sharing of copyrighted or licensed material in violation of its terms of use.
* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.
## Limitations and Bias
### Limitations
The model developers discuss the limitations of the model further in the DALL·E Mini technical report:
* Faces and people in general are not generated properly.
* Animals are usually unrealistic.
* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.
* The model has only been trained with English descriptions and will not perform as well in other languages
### Bias
CONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.
The model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.
While the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.
Our current analyses demonstrate that:
* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.
* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented.
* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.
* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.
The technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.
### Limitations and Bias Recommendations
* Users (both direct and downstream) should be made aware of the biases and limitations.
* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.
* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.
## Training
### Training Data
The model developers used 3 datasets for the model:
* Conceptual Captions Dataset, which contains 3 million image and caption pairs.
* Conceptual 12M, which contains 12 million image and caption pairs.
* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.
For fine-tuning the image encoder, a subset of 2 million images were used.
All images (about 15 million) were used for training the Seq2Seq model.
### Training Procedure
As described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:
* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.
* Descriptions are encoded through a BART encoder.
* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.
* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.
The simplified training procedure for DALL·E Mega is as follows:
* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3
* Optimizer: Distributed Shampoo
* Model Partition Specificiations: 8 model parallel x 32 data parallel
* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update
* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau
* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)
* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model.
* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.
There is more information about the full procedure and technical material in the DALL·E Mega training journal.
## Evaluation Results
The model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model.
For evaluation results related to DALL·E Mega, see this technical report.
## Environmental Impact
### DALL·E Mini Estimated Emissions
*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*
Based on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.
* Hardware Type: TPU v3-8
* Hours used: 72 (3 days)
* Cloud Provider: GCP (as mentioned in the technical report)
* Compute Region: us-east1 (provided by model developers)
* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq.
### DALL·E Mega Estimated Emissions
DALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.
* Hardware Type: TPU v3-256
* Hours used: 960 - 1080 hours (40-45 days)
* Cloud Provider: Unknown
* Compute Region: Unknown
* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): Unknown
*This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*
|
[
"# DALL·E Mini Model Card\n\nThis model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).\n\nThe DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.",
"## Model Details\n\n* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh\n* Model type: Transformer-based text-to-image generation model\n* Language(s): English\n* License: Apache 2.0\n* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”\n* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.\n* Cite as:",
"## Uses",
"### Direct Use\n\nThe model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.",
"### Downstream Use\n\nThe model could also be used for downstream use cases, including:\n* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science\n* Development of educational or creative tools\n* Generation of artwork and use in design and artistic processes. \n* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!\n\nDownstream uses exclude the uses described in Misuse and Out-of-Scope Use.",
"### Misuse, Malicious Use, and Out-of-Scope Use\n\nThe model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.",
"#### Out-of-Scope Use\n\nThe model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.",
"#### Misuse and Malicious Use \n\nUsing the model to generate content that is cruel to individuals is a misuse of this model. This includes:\n* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.\n* Intentionally promoting or propagating discriminatory content or harmful stereotypes.\n* Impersonating individuals without their consent.\n* Sexual content without consent of the people who might see it.\n* Mis- and disinformation\n* Representations of egregious violence and gore\n* Sharing of copyrighted or licensed material in violation of its terms of use.\n* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.",
"## Limitations and Bias",
"### Limitations\n\nThe model developers discuss the limitations of the model further in the DALL·E Mini technical report:\n* Faces and people in general are not generated properly.\n* Animals are usually unrealistic.\n* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.\n* The model has only been trained with English descriptions and will not perform as well in other languages",
"### Bias \n\nCONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.\n\nThe model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.\n\nWhile the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.\n\nOur current analyses demonstrate that:\n* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.\n* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. \n* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.\n* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.\n\nThe technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.",
"### Limitations and Bias Recommendations\n\n* Users (both direct and downstream) should be made aware of the biases and limitations.\n* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.\n* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.",
"## Training",
"### Training Data\n\nThe model developers used 3 datasets for the model:\n* Conceptual Captions Dataset, which contains 3 million image and caption pairs.\n* Conceptual 12M, which contains 12 million image and caption pairs.\n* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.\n\nFor fine-tuning the image encoder, a subset of 2 million images were used.\nAll images (about 15 million) were used for training the Seq2Seq model.",
"### Training Procedure\n\nAs described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:\n* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.\n* Descriptions are encoded through a BART encoder.\n* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.\n* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.\n\nThe simplified training procedure for DALL·E Mega is as follows: \n\n* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3\n* Optimizer: Distributed Shampoo\n* Model Partition Specificiations: 8 model parallel x 32 data parallel\n* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update\n* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau\n* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)\n* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. \n* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.\n\nThere is more information about the full procedure and technical material in the DALL·E Mega training journal.",
"## Evaluation Results\n\nThe model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. \n\nFor evaluation results related to DALL·E Mega, see this technical report.",
"## Environmental Impact",
"### DALL·E Mini Estimated Emissions\n\n*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*\n\nBased on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-8\n* Hours used: 72 (3 days)\n* Cloud Provider: GCP (as mentioned in the technical report)\n* Compute Region: us-east1 (provided by model developers)\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq.",
"### DALL·E Mega Estimated Emissions\n\nDALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-256\n* Hours used: 960 - 1080 hours (40-45 days)\n* Cloud Provider: Unknown\n* Compute Region: Unknown\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): Unknown\n\n*This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*"
] |
[
"TAGS\n#transformers #jax #dallebart #text-to-image #en #arxiv-2102.08981 #arxiv-2012.09841 #arxiv-1910.13461 #arxiv-1910.09700 #license-apache-2.0 #co2_eq_emissions #has_space #region-us \n",
"# DALL·E Mini Model Card\n\nThis model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).\n\nThe DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.",
"## Model Details\n\n* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh\n* Model type: Transformer-based text-to-image generation model\n* Language(s): English\n* License: Apache 2.0\n* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”\n* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.\n* Cite as:",
"## Uses",
"### Direct Use\n\nThe model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.",
"### Downstream Use\n\nThe model could also be used for downstream use cases, including:\n* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science\n* Development of educational or creative tools\n* Generation of artwork and use in design and artistic processes. \n* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!\n\nDownstream uses exclude the uses described in Misuse and Out-of-Scope Use.",
"### Misuse, Malicious Use, and Out-of-Scope Use\n\nThe model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.",
"#### Out-of-Scope Use\n\nThe model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.",
"#### Misuse and Malicious Use \n\nUsing the model to generate content that is cruel to individuals is a misuse of this model. This includes:\n* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.\n* Intentionally promoting or propagating discriminatory content or harmful stereotypes.\n* Impersonating individuals without their consent.\n* Sexual content without consent of the people who might see it.\n* Mis- and disinformation\n* Representations of egregious violence and gore\n* Sharing of copyrighted or licensed material in violation of its terms of use.\n* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.",
"## Limitations and Bias",
"### Limitations\n\nThe model developers discuss the limitations of the model further in the DALL·E Mini technical report:\n* Faces and people in general are not generated properly.\n* Animals are usually unrealistic.\n* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.\n* The model has only been trained with English descriptions and will not perform as well in other languages",
"### Bias \n\nCONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.\n\nThe model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.\n\nWhile the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.\n\nOur current analyses demonstrate that:\n* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.\n* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. \n* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.\n* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.\n\nThe technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.",
"### Limitations and Bias Recommendations\n\n* Users (both direct and downstream) should be made aware of the biases and limitations.\n* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.\n* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.",
"## Training",
"### Training Data\n\nThe model developers used 3 datasets for the model:\n* Conceptual Captions Dataset, which contains 3 million image and caption pairs.\n* Conceptual 12M, which contains 12 million image and caption pairs.\n* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.\n\nFor fine-tuning the image encoder, a subset of 2 million images were used.\nAll images (about 15 million) were used for training the Seq2Seq model.",
"### Training Procedure\n\nAs described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:\n* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.\n* Descriptions are encoded through a BART encoder.\n* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.\n* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.\n\nThe simplified training procedure for DALL·E Mega is as follows: \n\n* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3\n* Optimizer: Distributed Shampoo\n* Model Partition Specificiations: 8 model parallel x 32 data parallel\n* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update\n* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau\n* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)\n* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. \n* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.\n\nThere is more information about the full procedure and technical material in the DALL·E Mega training journal.",
"## Evaluation Results\n\nThe model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. \n\nFor evaluation results related to DALL·E Mega, see this technical report.",
"## Environmental Impact",
"### DALL·E Mini Estimated Emissions\n\n*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*\n\nBased on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-8\n* Hours used: 72 (3 days)\n* Cloud Provider: GCP (as mentioned in the technical report)\n* Compute Region: us-east1 (provided by model developers)\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq.",
"### DALL·E Mega Estimated Emissions\n\nDALL·E Mega is still training. So far, as on June 9, 2022, the model developers report that DALL·E Mega has been training for about 40-45 days on a TPU v3-256. Using those numbers, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-256\n* Hours used: 960 - 1080 hours (40-45 days)\n* Cloud Provider: Unknown\n* Compute Region: Unknown\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): Unknown\n\n*This model card was written by: Boris Dayma, Margaret Mitchell, Ezi Ozoani, Marissa Gerchick, Irene Solaiman, Clémentine Fourrier, Sasha Luccioni, Emily Witko, Nazneen Rajani, and Julian Herrera.*"
] |
[
80,
113,
217,
3,
81,
183,
79,
51,
163,
6,
95,
421,
104,
2,
155,
380,
92,
3,
181,
236
] |
[
"passage: TAGS\n#transformers #jax #dallebart #text-to-image #en #arxiv-2102.08981 #arxiv-2012.09841 #arxiv-1910.13461 #arxiv-1910.09700 #license-apache-2.0 #co2_eq_emissions #has_space #region-us \n# DALL·E Mini Model Card\n\nThis model card focuses on the model associated with the DALL·E mini space on Hugging Face, available here. The app is called “dalle-mini”, but incorporates “DALL·E Mini’’ and “DALL·E Mega” models (further details on this distinction forthcoming).\n\nThe DALL·E Mega model is the largest version of DALLE Mini. For more information specific to DALL·E Mega, see the DALL·E Mega model card.## Model Details\n\n* Developed by: Boris Dayma, Suraj Patil, Pedro Cuenca, Khalid Saifullah, Tanishq Abraham, Phúc Lê, Luke, Luke Melas, Ritobrata Ghosh\n* Model type: Transformer-based text-to-image generation model\n* Language(s): English\n* License: Apache 2.0\n* Model Description: This is a model that can be used to generate images based on text prompts. As the model developers wrote in the project report about DALL·E mini, “OpenAI had the first impressive model for generating images with DALL·E. DALL·E mini is an attempt at reproducing those results with an open-source model.”\n* Resources for more information: See OpenAI’s website for more information about DALL·E, including the DALL·E model card. See the project report for more information from the model’s developers. To learn more about DALL·E Mega, see the DALL·E Mega training journal.\n* Cite as:## Uses### Direct Use\n\nThe model is intended to be used to generate images based on text prompts for research and personal consumption. Intended uses include supporting creativity, creating humorous content, and providing generations for people curious about the model’s behavior. Intended uses exclude those described in the Misuse and Out-of-Scope Use section.",
"passage: ### Downstream Use\n\nThe model could also be used for downstream use cases, including:\n* Research efforts, such as probing and better understanding the limitations and biases of generative models to further improve the state of science\n* Development of educational or creative tools\n* Generation of artwork and use in design and artistic processes. \n* Other uses that are newly discovered by users. This currently includes poetry illustration (give a poem as prompt), fan art (putting a character in various other visual universes), visual puns, fairy tale illustrations (give a fantasy situation as prompt), concept mashups (applying a texture to something completely different), style transfers (portraits in the style of), … We hope you will find your own application!\n\nDownstream uses exclude the uses described in Misuse and Out-of-Scope Use.### Misuse, Malicious Use, and Out-of-Scope Use\n\nThe model should not be used to intentionally create or disseminate images that create hostile or alienating environments for people. This includes generating images that people would foreseeably find disturbing, distressing, or offensive; or content that propagates historical or current stereotypes.#### Out-of-Scope Use\n\nThe model was not trained to be factual or true representations of people or events, and therefore using the model to generate such content is out-of-scope for the abilities of this model.#### Misuse and Malicious Use \n\nUsing the model to generate content that is cruel to individuals is a misuse of this model. This includes:\n* Generating demeaning, dehumanizing, or otherwise harmful representations of people or their environments, cultures, religions, etc.\n* Intentionally promoting or propagating discriminatory content or harmful stereotypes.\n* Impersonating individuals without their consent.\n* Sexual content without consent of the people who might see it.\n* Mis- and disinformation\n* Representations of egregious violence and gore\n* Sharing of copyrighted or licensed material in violation of its terms of use.\n* Sharing content that is an alteration of copyrighted or licensed material in violation of its terms of use.## Limitations and Bias### Limitations\n\nThe model developers discuss the limitations of the model further in the DALL·E Mini technical report:\n* Faces and people in general are not generated properly.\n* Animals are usually unrealistic.\n* It is hard to predict where the model excels or falls short…Good prompt engineering will lead to the best results.\n* The model has only been trained with English descriptions and will not perform as well in other languages",
"passage: ### Bias \n\nCONTENT WARNING: Readers should be aware this section contains content that is disturbing, offensive, and can propagate historical and current stereotypes.\n\nThe model was trained on unfiltered data from the Internet, limited to pictures with English descriptions. Text and images from communities and cultures using other languages were not utilized. This affects all output of the model, with white and Western culture asserted as a default, and the model’s ability to generate content using non-English prompts is observably lower quality than prompts in English.\n\nWhile the capabilities of image generation models are impressive, they may also reinforce or exacerbate societal biases. The extent and nature of the biases of DALL·E Mini and DALL·E Mega models have yet to be fully documented, but initial testing demonstrates that they may generate images that contain negative stereotypes against minoritized groups. Work to analyze the nature and extent of the models’ biases and limitations is ongoing.\n\nOur current analyses demonstrate that:\n* Images generated by the model can include disturbing and harmful stereotypes across protected classes; identity characteristics; and sensitive, social, and occupational groups.\n* When the model generates images with people in them, it tends to output people who we perceive to be white, while people of color are underrepresented. \n* Images generated by the model can contain biased content that depicts power differentials between people of color and people who are white, with white people in positions of privilege.\n* The model is generally only usable for generating images based on text in English, limiting accessibility of the model for non-English speakers and potentially contributing to the biases in images generated by the model.\n\nThe technical report discusses these issues in more detail, and also highlights potential sources of bias in the model development process.### Limitations and Bias Recommendations\n\n* Users (both direct and downstream) should be made aware of the biases and limitations.\n* Content that is potentially problematic should be filtered out, e.g., via automated models that detect violence or pornography.\n* Further work on this model should include methods for balanced and just representations of people and cultures, for example, by curating the training dataset to be both diverse and inclusive.## Training### Training Data\n\nThe model developers used 3 datasets for the model:\n* Conceptual Captions Dataset, which contains 3 million image and caption pairs.\n* Conceptual 12M, which contains 12 million image and caption pairs.\n* The OpenAI subset of YFCC100M, which contains about 15 million images and that we further sub-sampled to 2 million images due to limitations in storage space. They used both title and description as caption and removed html tags, new lines and extra spaces.\n\nFor fine-tuning the image encoder, a subset of 2 million images were used.\nAll images (about 15 million) were used for training the Seq2Seq model.",
"passage: ### Training Procedure\n\nAs described further in the technical report for DALL·E Mini, during training, images and descriptions are both available and pass through the system as follows:\n* Images are encoded through a VQGAN encoder, which turns images into a sequence of tokens.\n* Descriptions are encoded through a BART encoder.\n* The output of the BART encoder and encoded images are fed through the BART decoder, which is an auto-regressive model whose goal is to predict the next token.\n* Loss is the softmax cross-entropy) between the model prediction logits and the actual image encodings from the VQGAN.\n\nThe simplified training procedure for DALL·E Mega is as follows: \n\n* Hardware: 1 pod TPU v3-256 = 32 nodes of TPU VM v3-8 (8 TPU per node) = 256 TPU v3\n* Optimizer: Distributed Shampoo\n* Model Partition Specificiations: 8 model parallel x 32 data parallel\n* Batch: 44 samples per model x 32 data parallel x 3 gradient accumulation steps = 4224 increasing samples per update\n* Learning rate: warmup to 0.0001 for 10,000 steps and then kept constant until plateau\n* Gradient checkpointing used on each Encoder/Decoder layer (ie, MHA + FFN)\n* Distributed Shampoo + Normformer Optimizations have proved to be effective and efficiently scaling this model. \n* It should also be noted that the learning rate and other parameters are sometimes adjusted on the fly, and batch size increased over time as well.\n\nThere is more information about the full procedure and technical material in the DALL·E Mega training journal.## Evaluation Results\n\nThe model developers discuss their results extensively in their technical report for DALL·E Mini, which provides comparisons between DALL·E Mini’s results with DALL·E-pytorch, OpenAI’s DALL·E, and models consisting of a generator coupled with the CLIP neural network model. \n\nFor evaluation results related to DALL·E Mega, see this technical report.## Environmental Impact### DALL·E Mini Estimated Emissions\n\n*The model is 27 times smaller than the original DALL·E and was trained on a single TPU v3-8 for only 3 days.*\n\nBased on that information, we estimate the following CO2 emissions using the Machine Learning Impact calculator presented in Lacoste et al. (2019). The hardware, runtime, cloud provider, and compute region were utilized to estimate the carbon impact.\n\n* Hardware Type: TPU v3-8\n* Hours used: 72 (3 days)\n* Cloud Provider: GCP (as mentioned in the technical report)\n* Compute Region: us-east1 (provided by model developers)\n* Carbon Emitted (Power consumption x Time x Carbon produced based on location of power grid): 30.16 kg CO2 eq."
] |
[
-0.028254972770810127,
0.03553052246570587,
-0.004067465662956238,
0.049623727798461914,
0.06288625299930573,
-0.008076719008386135,
0.08081356436014175,
0.034917984157800674,
-0.02770288847386837,
0.043668024241924286,
0.0014277361333370209,
-0.07090413570404053,
0.0627814382314682,
-0.02154352515935898,
0.062331415712833405,
-0.22897672653198242,
0.027930893003940582,
-0.056684572249650955,
0.0005141850560903549,
0.06901216506958008,
0.10026480257511139,
-0.0770077258348465,
0.08852638304233551,
0.01946571096777916,
-0.042475152760744095,
-0.04042445868253708,
-0.03032928891479969,
-0.015530581586062908,
0.04920753091573715,
0.08684322237968445,
0.05095826834440231,
-0.005609058775007725,
0.0387713648378849,
-0.1473565697669983,
0.011568363755941391,
0.0707334578037262,
-0.019062168896198273,
0.023666251450777054,
0.07249437272548676,
0.03721291571855545,
0.2478632628917694,
-0.04619394987821579,
0.055587947368621826,
0.05173077434301376,
-0.10144958645105362,
-0.10645191371440887,
-0.06079705059528351,
0.04926029592752457,
0.10882747918367386,
0.07318136096000671,
-0.034097954630851746,
0.09291748702526093,
-0.03491460159420967,
0.029399586841464043,
0.08232337981462479,
-0.07411438971757889,
-0.01771201193332672,
0.013882206752896309,
0.06669515371322632,
0.039162635803222656,
-0.06075152009725571,
0.00817346666008234,
0.0069837504997849464,
0.033988527953624725,
0.05505349859595299,
-0.05015191435813904,
0.2091364711523056,
-0.07313799113035202,
-0.1134466826915741,
-0.04481673985719681,
0.10675753653049469,
0.03127077966928482,
-0.10067102313041687,
-0.1812151074409485,
-0.037602417171001434,
0.09761438518762589,
0.00042806053534150124,
-0.023337190970778465,
0.008568367920815945,
-0.011262997053563595,
0.059470467269420624,
-0.06772677600383759,
-0.09050371497869492,
-0.004825189709663391,
-0.04385785013437271,
0.11731384694576263,
0.021325889974832535,
0.022427991032600403,
-0.07191671431064606,
0.04677552729845047,
0.029860379174351692,
-0.08833827078342438,
-0.06031983345746994,
-0.06868395209312439,
-0.059143178164958954,
-0.02178228460252285,
-0.03351263701915741,
-0.05831075832247734,
-0.03292738273739815,
0.10511258244514465,
-0.05111349746584892,
0.027061428874731064,
0.0024063875898718834,
0.04261112958192825,
0.14077980816364288,
0.09049506485462189,
-0.024822315201163292,
-0.01004154421389103,
0.015713516622781754,
0.023020923137664795,
0.0514226071536541,
-0.0317879319190979,
-0.028848174959421158,
0.021274447441101074,
-0.025584641844034195,
0.04199764132499695,
0.06183002516627312,
0.03949528932571411,
-0.040642499923706055,
-0.05593950301408768,
0.14764374494552612,
-0.07204604893922806,
-0.006800378672778606,
-0.014963438734412193,
-0.02434675768017769,
0.050483979284763336,
0.014719635248184204,
-0.0008798157796263695,
-0.07039676606655121,
0.04281219467520714,
-0.055165309458971024,
-0.029596570879220963,
-0.10530353337526321,
-0.07354016602039337,
0.002370290458202362,
0.04515904188156128,
-0.024029551073908806,
-0.12368092685937881,
-0.14114633202552795,
-0.06599659472703934,
0.0413743332028389,
-0.04968554526567459,
-0.012931989505887032,
0.020917214453220367,
-0.026565685868263245,
-0.035824522376060486,
0.041345492005348206,
-0.035178959369659424,
-0.01451888121664524,
0.01812676340341568,
-0.055077023804187775,
0.04741940647363663,
0.004489791579544544,
0.021693943068385124,
-0.07072308659553528,
0.06197230890393257,
-0.15173320472240448,
0.0929078608751297,
-0.026153195649385452,
-0.024275565519928932,
-0.0450371615588665,
-0.03753272444009781,
-0.032213978469371796,
0.07454928755760193,
-0.022696223109960556,
0.116361103951931,
-0.20337817072868347,
-0.007742009125649929,
0.06977975368499756,
-0.1483323723077774,
-0.022403698414564133,
0.10106468945741653,
-0.036246828734874725,
0.07688304036855698,
0.08305046707391739,
0.036925122141838074,
-0.0020695943385362625,
-0.025012217462062836,
-0.01153053529560566,
-0.04870731383562088,
-0.062164828181266785,
0.17251208424568176,
0.06693437695503235,
-0.04911139979958534,
0.06920278072357178,
-0.018502071499824524,
-0.009777560830116272,
-0.02227686159312725,
0.008205408230423927,
-0.022113990038633347,
0.03777501732110977,
-0.0027716532349586487,
0.007867496460676193,
-0.016110830008983612,
-0.057321034371852875,
0.007892405614256859,
-0.09775183349847794,
-0.02275952510535717,
0.08690518140792847,
-0.027565091848373413,
0.03292963281273842,
-0.09944845736026764,
0.05250699073076248,
0.020833153277635574,
-0.005148601718246937,
-0.14252260327339172,
-0.02017667517066002,
0.03046880103647709,
-0.08808696269989014,
0.03742404282093048,
0.059281185269355774,
0.0067289830185472965,
0.07056766748428345,
-0.011954881250858307,
0.0026678312569856644,
-0.07645005732774734,
-0.0060551525093615055,
-0.058328427374362946,
-0.09309162199497223,
0.0014808042906224728,
-0.028311999514698982,
0.05253634974360466,
-0.1274254471063614,
0.0010249641491100192,
0.06342525035142899,
0.05306694284081459,
0.05651799589395523,
-0.03859721124172211,
0.02156982384622097,
0.01850355789065361,
-0.03478720784187317,
-0.05338340252637863,
0.0368497408926487,
-0.012614390812814236,
-0.017246492207050323,
0.08421136438846588,
-0.16252359747886658,
-0.10591991245746613,
0.062445297837257385,
-0.026578843593597412,
-0.08782987296581268,
-0.015330689027905464,
0.006832335144281387,
-0.011905509978532791,
-0.045760687440633774,
-0.06502213329076767,
0.20284192264080048,
0.011342409998178482,
0.04700389876961708,
-0.08265016973018646,
-0.05355772003531456,
0.01115522999316454,
-0.027509599924087524,
-0.05203269422054291,
0.03487518057227135,
0.08564162999391556,
-0.18022382259368896,
0.04727400466799736,
0.07712538540363312,
0.02383784018456936,
0.1252809315919876,
0.0259387344121933,
-0.07890556752681732,
-0.03962235152721405,
-0.011019477620720863,
0.01223394088447094,
0.11317340284585953,
0.01215303409844637,
0.017476949840784073,
0.016512854024767876,
0.01794200949370861,
0.0327441468834877,
-0.04873259365558624,
0.035352662205696106,
0.0667865127325058,
0.03654003515839577,
-0.03536738455295563,
-0.009528130292892456,
-0.03636325150728226,
0.06997880339622498,
0.012568825855851173,
0.049166239798069,
0.00017326511442661285,
-0.05551569163799286,
-0.10173077136278152,
0.1352764070034027,
-0.058849312365055084,
-0.22323130071163177,
-0.09861410409212112,
0.03589071333408356,
-0.015951819717884064,
0.026546331122517586,
0.014165805652737617,
-0.03896123543381691,
-0.07032185792922974,
-0.11490955203771591,
0.06354472786188126,
-0.031060636043548584,
-0.05945001542568207,
0.0015988685190677643,
0.025862351059913635,
-0.004441087134182453,
-0.08530192077159882,
-0.005646780598908663,
0.011179406195878983,
-0.058748986572027206,
0.03651847317814827,
0.01139405183494091,
0.09869913756847382,
0.059338077902793884,
0.020532462745904922,
-0.054640598595142365,
-0.011880309320986271,
0.22668468952178955,
-0.08934766054153442,
0.09054826200008392,
0.17836976051330566,
-0.10897806286811829,
0.07825756072998047,
0.08518539369106293,
-0.004636690486222506,
-0.06597138941287994,
0.037353768944740295,
0.04858146607875824,
-0.053951188921928406,
-0.10588184744119644,
-0.06042233109474182,
-0.021527068689465523,
-0.07084983587265015,
0.029493208974599838,
0.04446162283420563,
0.010762091726064682,
0.0269143208861351,
-0.0894990935921669,
0.01688317395746708,
0.04011239856481552,
0.08663161844015121,
0.04454725980758667,
-0.02348225563764572,
0.04329780861735344,
-0.03495113551616669,
0.000022210588213056326,
0.07992084324359894,
-0.06985311210155487,
0.26336443424224854,
-0.012084297835826874,
0.1265924572944641,
0.08194877952337265,
-0.036599621176719666,
0.06675823032855988,
0.0014846008270978928,
-0.04959314316511154,
0.0007254760712385178,
-0.05866572633385658,
-0.05423877015709877,
-0.008063754998147488,
0.06840036064386368,
0.025193601846694946,
-0.03208202123641968,
-0.039588361978530884,
-0.0018940456211566925,
0.04813583195209503,
0.11805759370326996,
-0.03873157873749733,
-0.08269913494586945,
-0.04491004720330238,
0.07159694284200668,
-0.033708326518535614,
-0.06928415596485138,
0.010913601145148277,
0.14027845859527588,
-0.11137884855270386,
0.007270073983818293,
-0.0032869079150259495,
0.07233675569295883,
-0.13551637530326843,
0.0071943155489861965,
0.0011359783820807934,
0.021979784592986107,
-0.008933153934776783,
0.04316481202840805,
-0.13815392553806305,
0.09047337621450424,
0.008687290363013744,
0.056146007031202316,
-0.0850694477558136,
0.007569180801510811,
0.0006364388391375542,
0.04126792773604393,
0.11525162309408188,
0.04153529927134514,
-0.07946962118148804,
-0.056585464626550674,
0.0032431408762931824,
0.028699912130832672,
0.07434144616127014,
-0.05389831215143204,
0.09108325839042664,
-0.01798831857740879,
0.023831937462091446,
-0.02646324597299099,
0.012910842895507812,
-0.13566648960113525,
-0.176319420337677,
0.04301288723945618,
-0.12183986604213715,
-0.03633871674537659,
-0.07857294380664825,
0.0010197460651397705,
-0.004415360279381275,
0.09473318606615067,
-0.12726230919361115,
-0.09807215631008148,
-0.10268238931894302,
-0.074149951338768,
0.05964420363306999,
-0.06631974130868912,
0.043368808925151825,
0.02423635497689247,
0.14521607756614685,
-0.040510646998882294,
-0.061803922057151794,
0.017293665558099747,
-0.07009810209274292,
-0.1672673225402832,
-0.05681026726961136,
0.0615042969584465,
0.13682357966899872,
0.05068415403366089,
0.011426086537539959,
0.03253782168030739,
0.0072747101075947285,
-0.10684018582105637,
-0.010879680514335632,
0.16073298454284668,
0.03374543786048889,
-0.003926137927919626,
0.0009258128702640533,
-0.023555630818009377,
-0.06896185874938965,
-0.05595635250210762,
0.047547418624162674,
0.1831439733505249,
-0.008493962697684765,
0.09611589461565018,
0.17328378558158875,
-0.08459242433309555,
-0.186487078666687,
-0.019887955859303474,
0.02539883367717266,
0.011104135774075985,
0.11181462556123734,
-0.16512854397296906,
-0.010743679478764534,
0.04467933624982834,
0.027011357247829437,
0.06291723251342773,
-0.2347528040409088,
-0.09806735068559647,
0.029611662030220032,
0.06993131339550018,
0.03655482828617096,
-0.0708540827035904,
-0.01600913517177105,
-0.05113302916288376,
-0.06620535999536514,
0.09347490221261978,
-0.03533082827925682,
0.03588863089680672,
0.00538367684930563,
0.01825021021068096,
0.02594309113919735,
-0.006132916547358036,
0.15769094228744507,
-0.05109522491693497,
0.0729566439986229,
-0.11589480936527252,
-0.0852176696062088,
0.02390149235725403,
-0.027406781911849976,
0.03099888749420643,
-0.03756943717598915,
0.014072074554860592,
-0.06434018909931183,
-0.06413149833679199,
-0.05263620242476463,
0.006362117361277342,
-0.06739330291748047,
-0.07019312679767609,
-0.07778807729482651,
0.08532857149839401,
0.0671629011631012,
-0.010375987738370895,
-0.037961240857839584,
-0.0735565721988678,
-0.05837365612387657,
0.045114077627658844,
0.17235277593135834,
0.05043085664510727,
-0.12761260569095612,
-0.029124373570084572,
0.006755275651812553,
0.1039365828037262,
-0.09552700072526932,
0.009892976842820644,
0.04434242472052574,
-0.017265576869249344,
0.10081908106803894,
0.009294233284890652,
-0.1340622454881668,
0.0487697497010231,
0.04260619729757309,
-0.023452933877706528,
-0.15278008580207825,
-0.02933608554303646,
0.07298611849546432,
-0.07325984537601471,
-0.09022702276706696,
0.08628445863723755,
-0.04538871347904205,
-0.004823296330869198,
-0.009522996842861176,
0.06333762407302856,
0.023358602076768875,
0.024976138025522232,
0.0028527677059173584,
0.0302369873970747,
-0.03408993035554886,
0.08393940329551697,
0.054362453520298004,
-0.042776189744472504,
0.04139868542551994,
0.07357312738895416,
-0.07243712991476059,
-0.064208023250103,
-0.07890606671571732,
0.04653486981987953,
-0.005285482853651047,
-0.05212779715657234,
0.04838936775922775,
-0.07191093266010284,
0.0017336253076791763,
0.0533301942050457,
0.012732119299471378,
0.023290159180760384,
-0.03371123969554901,
0.015377596952021122,
-0.03579647094011307,
0.03916158527135849,
-0.020676905289292336,
0.002309344708919525,
-0.04743212088942528,
0.032694797962903976,
0.05079290643334389,
-0.0027627190575003624,
-0.02251817099750042,
-0.04504823684692383,
-0.08618436008691788,
-0.0047812615521252155,
-0.13337156176567078,
-0.007625746540725231,
-0.09619545191526413,
-0.02372058853507042,
-0.0005544261075556278,
0.018877064809203148,
-0.007224884815514088,
0.01956903375685215,
-0.027327407151460648,
-0.02873135358095169,
-0.00702556362375617,
0.044896624982357025,
-0.05840778350830078,
0.011147970333695412,
0.07285720109939575,
-0.05937982350587845,
0.07405173033475876,
-0.059571970254182816,
-0.042355410754680634,
-0.011622674763202667,
-0.07043963670730591,
0.052913691848516464,
-0.03563066944479942,
0.009132534265518188,
-0.0056062377989292145,
-0.12580133974552155,
-0.00755306426435709,
-0.03438004106283188,
-0.043141745030879974,
-0.01551198773086071,
0.05328216776251793,
-0.06578578054904938,
0.06367267668247223,
0.04630739986896515,
-0.03355802595615387,
-0.07394486665725708,
0.01726502552628517,
0.03311347961425781,
0.035739585757255554,
0.1184537410736084,
-0.0021206100936979055,
0.041592374444007874,
-0.10370165854692459,
-0.0005483492277562618,
0.04291285201907158,
0.053148772567510605,
0.04874550178647041,
-0.054888710379600525,
0.0162661112844944,
-0.03041190654039383,
0.12889668345451355,
0.007716603577136993,
-0.03280327469110489,
0.04245828464627266,
0.021763024851679802,
-0.049131445586681366,
0.028598498553037643,
0.004202160984277725,
-0.01807636208832264,
-0.004112003371119499,
-0.04956802725791931,
-0.015185982920229435,
-0.0360383614897728,
-0.05825777351856232,
0.10740920901298523,
0.08184842765331268,
0.09030456840991974,
0.03267945349216461,
0.0066353571601212025,
-0.0509047731757164,
-0.04715346544981003,
0.0016539925709366798,
0.030112003907561302,
0.011715339496731758,
-0.058775369077920914,
0.027856627479195595,
0.1601923406124115,
-0.08058802038431168,
0.08788246661424637,
-0.013036543503403664,
-0.028034061193466187,
-0.018286975100636482,
-0.23193016648292542,
-0.006117557175457478,
0.040568433701992035,
-0.012143414467573166,
-0.05696731060743332,
0.028236743062734604,
0.06608431041240692,
-0.00397786358371377,
-0.038200631737709045,
0.06065433472394943,
-0.11099310219287872,
-0.1268375962972641,
0.026680264621973038,
0.006194673478603363,
0.03170626610517502,
0.05601657181978226,
0.04785269498825073,
0.018475983291864395,
0.036732178181409836,
0.04482865333557129,
0.07156772911548615,
0.005309471860527992,
0.01584353856742382,
-0.05377034470438957,
-0.06344524025917053,
0.00503256730735302,
-0.001949157565832138,
0.01793571561574936,
0.19221460819244385,
0.044266775250434875,
-0.007715856656432152,
-0.0005565970204770565,
0.10910259932279587,
0.0050235651433467865,
-0.07127586007118225,
-0.10156755149364471,
0.11592833697795868,
-0.009269963949918747,
0.007596705108880997,
-0.004099417477846146,
-0.09853032231330872,
0.035631634294986725,
0.14595124125480652,
0.11793772876262665,
-0.06497637927532196,
-0.010544723831117153,
-0.024871796369552612,
0.01234507467597723,
-0.04233910143375397,
0.08465882390737534,
-0.0002979077398777008,
0.27663376927375793,
-0.042093705385923386,
0.08295948803424835,
-0.048427827656269073,
-0.014403104782104492,
-0.020525701344013214,
0.09552521258592606,
0.025097861886024475,
0.01702967658638954,
-0.05797044560313225,
0.07815112173557281,
-0.043247804045677185,
-0.18040023744106293,
0.02749481052160263,
0.017609108239412308,
-0.0531073622405529,
0.0482812263071537,
-0.00957708340138197,
0.004043959081172943,
0.0781099796295166,
0.010732724331319332,
-0.007306508719921112,
0.10703416913747787,
0.023312224075198174,
-0.023071173578500748,
0.015670571476221085,
0.08158473670482635,
0.00038789771497249603,
0.17934975028038025,
0.018467240035533905,
0.11872272193431854,
0.07088959962129593,
0.023334555327892303,
-0.09702812135219574,
0.04668901488184929,
0.019254200160503387,
-0.047453783452510834,
0.013965137302875519,
0.16018974781036377,
0.0009419107809662819,
0.04975389689207077,
0.08270584791898727,
0.03995238617062569,
0.08036544173955917,
-0.02389020100235939,
-0.007218390237540007,
-0.03987351059913635,
0.09136070311069489,
-0.09668055176734924,
0.14175905287265778,
0.0936768427491188,
0.00007086072582751513,
-0.03070146217942238,
-0.018281973898410797,
0.02458261512219906,
0.010043719783425331,
0.0709032416343689,
-0.008239459246397018,
-0.07810752093791962,
0.00790075771510601,
-0.0020623058080673218,
0.033364538103342056,
-0.12251053005456924,
-0.049385227262973785,
0.035902246832847595,
-0.029955897480249405,
0.02067720890045166,
0.051837995648384094,
0.049863290041685104,
0.00032302155159413815,
-0.03971381485462189,
-0.010713146068155766,
-0.020958060398697853,
0.05297427996993065,
-0.06335750222206116,
-0.02293112501502037
] |
null | null |
transformers
|
## VQGAN-f16-16384
### Model Description
This is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in [Taming Transformers for High-Resolution Image Synthesis](https://compvis.github.io/taming-transformers/) ([CVPR paper](https://openaccess.thecvf.com/content/CVPR2021/html/Esser_Taming_Transformers_for_High-Resolution_Image_Synthesis_CVPR_2021_paper.html)).
The model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.
This version of the model uses a reduction factor `f=16` and a vocabulary of `16,384` tokens.
As an example of how the reduction factor works, images of size `256x256` are encoded to sequences of `256` tokens: `256/16 * 256/16`. Images of `512x512` would result in sequences of `1024` tokens.
This model was ported to JAX using [a checkpoint trained on ImageNet](https://heibox.uni-heidelberg.de/d/a7530b09fed84f80a887/).
### How to Use
The checkpoint can be loaded using [Suraj Patil's implementation](https://github.com/patil-suraj/vqgan-jax) of `VQModel`.
### Other
This model can be used as part of the implementation of [DALL·E mini](https://github.com/borisdayma/dalle-mini). Our [report](https://wandb.ai/dalle-mini/dalle-mini/reports/DALL-E-mini--Vmlldzo4NjIxODA) contains more details on how to leverage it in an image encoding / generation pipeline.
|
{}
| null |
dalle-mini/vqgan_imagenet_f16_16384
|
[
"transformers",
"jax",
"endpoints_compatible",
"has_space",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #jax #endpoints_compatible #has_space #region-us
|
## VQGAN-f16-16384
### Model Description
This is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).
The model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.
This version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.
As an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.
This model was ported to JAX using a checkpoint trained on ImageNet.
### How to Use
The checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.
### Other
This model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline.
|
[
"## VQGAN-f16-16384",
"### Model Description\n\nThis is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).\n\nThe model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.\n\nThis version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.\n\nAs an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.\n\nThis model was ported to JAX using a checkpoint trained on ImageNet.",
"### How to Use\n\nThe checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.",
"### Other\n\nThis model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline."
] |
[
"TAGS\n#transformers #jax #endpoints_compatible #has_space #region-us \n",
"## VQGAN-f16-16384",
"### Model Description\n\nThis is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).\n\nThe model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.\n\nThis version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.\n\nAs an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.\n\nThis model was ported to JAX using a checkpoint trained on ImageNet.",
"### How to Use\n\nThe checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.",
"### Other\n\nThis model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline."
] |
[
24,
9,
208,
27,
43
] |
[
"passage: TAGS\n#transformers #jax #endpoints_compatible #has_space #region-us \n## VQGAN-f16-16384### Model Description\n\nThis is a Flax/JAX implementation of VQGAN, which learns a codebook of context-rich visual parts by leveraging both the use of convolutional methods and transformers. It was introduced in Taming Transformers for High-Resolution Image Synthesis (CVPR paper).\n\nThe model allows the encoding of images as a fixed-length sequence of tokens taken from the codebook.\n\nThis version of the model uses a reduction factor 'f=16' and a vocabulary of '16,384' tokens.\n\nAs an example of how the reduction factor works, images of size '256x256' are encoded to sequences of '256' tokens: '256/16 * 256/16'. Images of '512x512' would result in sequences of '1024' tokens.\n\nThis model was ported to JAX using a checkpoint trained on ImageNet.### How to Use\n\nThe checkpoint can be loaded using Suraj Patil's implementation of 'VQModel'.### Other\n\nThis model can be used as part of the implementation of DALL·E mini. Our report contains more details on how to leverage it in an image encoding / generation pipeline."
] |
[
-0.006597475148737431,
0.06075231358408928,
-0.004456229507923126,
0.04626632481813431,
0.13018949329853058,
-0.0017908165464177728,
0.10405097901821136,
0.12290865182876587,
-0.14265818893909454,
0.0680757537484169,
0.05414735898375511,
0.021042343229055405,
0.06811831146478653,
0.10318948328495026,
0.08109906315803528,
-0.24791069328784943,
0.04128504544496536,
-0.02032165229320526,
-0.0578073225915432,
0.04168806970119476,
0.09300445765256882,
-0.1249479278922081,
0.09109122306108475,
0.03338020294904709,
-0.16884322464466095,
0.09058205038309097,
-0.004405213985592127,
-0.04228891059756279,
0.08763128519058228,
0.105554960668087,
0.04212513938546181,
-0.03425802290439606,
0.0791950672864914,
-0.16392023861408234,
0.010320218279957771,
0.10594967007637024,
-0.013712236657738686,
0.053026873618364334,
0.019013753160834312,
0.10305416584014893,
0.14872924983501434,
-0.07393950968980789,
0.025447476655244827,
0.01763094589114189,
-0.047375500202178955,
-0.2571936249732971,
-0.05998358502984047,
0.0817716047167778,
0.05016034096479416,
0.04075494781136513,
0.014871301129460335,
0.12323004007339478,
0.06085049733519554,
0.0879000648856163,
0.18393535912036896,
-0.258868008852005,
-0.01534380204975605,
0.12966981530189514,
0.05181879177689552,
0.06185908243060112,
-0.028509553521871567,
0.020341485738754272,
0.020377730950713158,
0.030974650755524635,
0.14051689207553864,
-0.05164024606347084,
0.014032313600182533,
-0.01648372784256935,
-0.08758914470672607,
-0.11567975580692291,
0.07135719805955887,
-0.026942329481244087,
-0.059363286942243576,
-0.13615205883979797,
-0.07512452453374863,
-0.030233802273869514,
-0.023930590599775314,
-0.1090991199016571,
0.032140638679265976,
0.022404339164495468,
0.026487985625863075,
-0.10340812802314758,
-0.11289744824171066,
-0.03791520372033119,
-0.028171759098768234,
0.042469050735235214,
0.03576323762536049,
0.06014057248830795,
-0.14054431021213531,
0.042105112224817276,
-0.07394247502088547,
-0.04886261001229286,
-0.056407660245895386,
-0.06822925806045532,
-0.018442349508404732,
0.03717406466603279,
0.009981624782085419,
-0.11268740147352219,
-0.024045506492257118,
0.05856532230973244,
-0.05484084412455559,
0.0833575576543808,
-0.008255742490291595,
0.020815057680010796,
0.04935887083411217,
0.15444493293762207,
-0.02487511932849884,
0.06773091852664948,
0.026983026415109634,
-0.034935593605041504,
0.0204461682587862,
-0.08993024379014969,
-0.10091729462146759,
-0.0000751412080717273,
0.04272305965423584,
0.04678112268447876,
-0.07348236441612244,
0.015170237980782986,
0.03322199732065201,
-0.019192127510905266,
0.2181636542081833,
-0.10131433606147766,
0.03327342867851257,
-0.011691084131598473,
-0.0007019525510258973,
0.1201300248503685,
0.04993820935487747,
-0.05283498018980026,
-0.09633058309555054,
-0.042896125465631485,
-0.07582768052816391,
-0.04715600237250328,
-0.10648223757743835,
-0.1572473794221878,
-0.02822878025472164,
-0.08625566214323044,
0.014249906875193119,
-0.1635131537914276,
-0.09891855716705322,
0.021731648594141006,
0.018696719780564308,
-0.020777670666575432,
0.04252914711833,
0.014576028101146221,
-0.04842452332377434,
0.0018831976922228932,
0.02426300384104252,
-0.029093675315380096,
0.0022141761146485806,
0.0033505542669445276,
0.04526962712407112,
0.08727675676345825,
-0.026030369102954865,
-0.017985988408327103,
-0.046722494065761566,
0.07640711218118668,
-0.0996888130903244,
0.08612365275621414,
-0.02746998891234398,
-0.008499612100422382,
-0.08770707994699478,
-0.06252073496580124,
0.009874527342617512,
0.005788901820778847,
0.06697894632816315,
0.0740143209695816,
-0.18879136443138123,
-0.012639903463423252,
0.21226376295089722,
-0.14473704993724823,
-0.04311424493789673,
0.08136896044015884,
-0.02050139009952545,
0.00577427726238966,
0.06739732623100281,
0.11653833091259003,
0.13528983294963837,
-0.11646295338869095,
-0.038024045526981354,
0.08854041248559952,
-0.006200303323566914,
-0.010076467879116535,
0.04961676523089409,
0.004704251419752836,
0.0714205875992775,
0.009145119227468967,
0.05470747873187065,
-0.05887226387858391,
-0.022057505324482918,
-0.032411523163318634,
-0.025771135464310646,
0.009696109220385551,
0.0036381424870342016,
-0.004393335897475481,
0.009960193186998367,
-0.021631227806210518,
-0.08955947309732437,
-0.00195832597091794,
0.09833954274654388,
-0.1251336932182312,
0.04001585766673088,
-0.09963786602020264,
0.05185620114207268,
-0.1132979616522789,
0.026814347133040428,
-0.16703198850154877,
-0.005378156900405884,
0.043721772730350494,
-0.08561054617166519,
-0.030888544395565987,
0.06687929481267929,
0.040966808795928955,
0.03634021803736687,
0.06186843663454056,
0.039850231260061264,
-0.06892046332359314,
-0.03140026330947876,
-0.0678844153881073,
-0.024940405040979385,
-0.11805476993322372,
-0.06634949892759323,
-0.05006800591945648,
-0.14766615629196167,
0.01235931646078825,
0.12429475784301758,
0.031140774488449097,
0.04423638433218002,
-0.02016276866197586,
-0.012890681624412537,
-0.024266621097922325,
-0.014412510208785534,
-0.08429989218711853,
0.014282654039561749,
0.04672420769929886,
0.040734656155109406,
0.04981568455696106,
-0.20533928275108337,
-0.004526039585471153,
0.0733381137251854,
0.07315460592508316,
-0.007910169661045074,
-0.016951825469732285,
-0.007829505950212479,
0.02091672644019127,
-0.027646712958812714,
-0.02911241166293621,
0.15080216526985168,
0.00466998340561986,
0.12674598395824432,
-0.09203733503818512,
-0.038568466901779175,
0.11060196161270142,
-0.0323791429400444,
-0.05549416318535805,
0.011546962894499302,
-0.020403606817126274,
-0.09077993780374527,
0.013790734112262726,
0.025763211771845818,
0.043240610510110855,
0.12315165996551514,
0.01666579209268093,
-0.08409300446510315,
-0.061058081686496735,
-0.05177260562777519,
0.02807004377245903,
0.09256446361541748,
-0.03530491888523102,
-0.05356869846582413,
0.03532037511467934,
0.054990701377391815,
0.04022848606109619,
-0.09337830543518066,
0.06275992840528488,
0.10194756090641022,
-0.041075292974710464,
-0.0023498975206166506,
0.007090691477060318,
-0.0664815679192543,
0.02529526688158512,
0.06247000768780708,
0.11698020994663239,
0.0023803780786693096,
-0.049999549984931946,
-0.09919466823339462,
0.16966071724891663,
-0.12142793089151382,
-0.2606508433818817,
-0.1457156091928482,
0.005495865363627672,
0.01880567893385887,
0.005322332493960857,
0.03676815703511238,
-0.0627497136592865,
-0.09325451403856277,
-0.10142833739519119,
-0.013901646248996258,
-0.13209189474582672,
-0.013541982509195805,
0.06323250383138657,
-0.020253533497452736,
0.029269423335790634,
-0.15135237574577332,
-0.003396119922399521,
0.060459356755018234,
-0.03442729264497757,
-0.014017732813954353,
0.013907987624406815,
0.08593756705522537,
0.05869848281145096,
-0.09817526489496231,
0.04889814183115959,
-0.012063420377671719,
0.25734779238700867,
-0.04158835485577583,
0.03772962465882301,
0.19707587361335754,
-0.0643192008137703,
0.05201348289847374,
0.01932741515338421,
-0.008330519311130047,
-0.04560014232993126,
0.009362342767417431,
-0.00662068510428071,
-0.09569109231233597,
-0.1392700970172882,
0.015375569462776184,
-0.05292893573641777,
0.01994434930384159,
0.07249003648757935,
0.045524463057518005,
0.005793232470750809,
0.05838911980390549,
0.006608649156987667,
0.13250994682312012,
-0.06137358397245407,
0.08221586048603058,
0.16668537259101868,
0.027526775375008583,
0.028263969346880913,
-0.09543631970882416,
0.00855342485010624,
0.07282500714063644,
0.08445308357477188,
0.15255744755268097,
-0.02411399967968464,
0.15503886342048645,
0.04605429619550705,
0.09253711253404617,
0.07632394880056381,
0.09708815068006516,
-0.09284227341413498,
0.0043264818377792835,
-0.031210321933031082,
-0.06493503600358963,
-0.053058214485645294,
0.026106497272849083,
-0.04852234572172165,
-0.02268843725323677,
-0.0718969851732254,
0.0593363493680954,
0.005808097776025534,
0.12530788779258728,
0.01945807971060276,
-0.19315417110919952,
-0.09573047608137131,
0.024855302646756172,
0.014949682168662548,
-0.09338223189115524,
0.02405383437871933,
0.20658162236213684,
-0.11168184131383896,
-0.029361600056290627,
-0.05385259911417961,
0.07018579542636871,
-0.12810710072517395,
0.018515532836318016,
-0.005681218113750219,
0.10133697837591171,
0.032182879745960236,
0.1119547039270401,
-0.212228462100029,
0.06661240011453629,
0.023691875860095024,
0.08967912197113037,
-0.0917946994304657,
0.09227713197469711,
-0.034604329615831375,
0.007272906601428986,
0.13490265607833862,
-0.012805657461285591,
-0.11157333850860596,
-0.09284047037363052,
-0.009131581522524357,
0.044310424476861954,
0.0792350172996521,
0.003048093756660819,
0.09695817530155182,
-0.04963739216327667,
-0.00020549895998556167,
-0.020748862996697426,
0.06001126393675804,
-0.08139175921678543,
-0.22103972733020782,
0.026732521131634712,
-0.016249913722276688,
-0.07797889411449432,
-0.04177949205040932,
-0.0050312429666519165,
0.0004254453524481505,
0.10854334384202957,
-0.018195275217294693,
-0.08033880591392517,
-0.11139972507953644,
-0.001542100333608687,
0.055060893297195435,
-0.06872950494289398,
0.0857735425233841,
-0.0012561575276777148,
0.22845135629177094,
-0.028693117201328278,
-0.14935676753520966,
0.04444728046655655,
-0.086384616792202,
-0.06287693977355957,
0.011671624146401882,
0.033289022743701935,
0.0040389783680438995,
0.0053517804481089115,
0.05004841089248657,
0.02189982496201992,
-0.01815970055758953,
-0.09972747415304184,
0.005550415720790625,
0.09419801831245422,
0.02809162065386772,
-0.08013885468244553,
0.030777225270867348,
0.022404231131076813,
-0.003428179770708084,
0.025596708059310913,
0.13273324072360992,
0.18982960283756256,
-0.10918436199426651,
0.07709953933954239,
0.14891357719898224,
-0.11561719328165054,
-0.1807917207479477,
-0.04807161167263985,
-0.00025641193496994674,
0.07081091403961182,
0.07354515045881271,
-0.14198927581310272,
0.005407281219959259,
0.03167736902832985,
-0.0030958608258515596,
0.016827527433633804,
-0.3455412983894348,
-0.06075385957956314,
0.006283250171691179,
0.12176771461963654,
0.007726958021521568,
-0.09619499742984772,
-0.027506904676556587,
-0.027833834290504456,
-0.16414155066013336,
0.12659214437007904,
-0.12525399029254913,
0.026031112298369408,
0.0020261602476239204,
-0.01911516673862934,
0.032257769256830215,
-0.03817703574895859,
0.06458020955324173,
-0.07834397256374359,
0.07468651235103607,
-0.0668681189417839,
-0.014223086647689342,
0.15718825161457062,
-0.05069514364004135,
0.10355378687381744,
-0.03157993033528328,
0.10741634666919708,
-0.13223323225975037,
-0.04418587684631348,
-0.0068824090994894505,
-0.016070617362856865,
-0.06778328120708466,
-0.06885010749101639,
-0.09452197700738907,
-0.004750771913677454,
0.0843408927321434,
-0.010388900525867939,
0.08726832270622253,
-0.009407827630639076,
0.061569783836603165,
0.24702170491218567,
0.07448070496320724,
0.003772465046495199,
-0.0968116894364357,
0.009650221094489098,
0.008484577760100365,
0.09505777806043625,
-0.17656753957271576,
0.05040058121085167,
0.07832375168800354,
0.012504837475717068,
0.04480051249265671,
0.03847331926226616,
-0.11980484426021576,
0.05803735554218292,
0.027435265481472015,
-0.11127851903438568,
-0.11785715073347092,
-0.015646275132894516,
0.06931867450475693,
-0.053927477449178696,
0.02001003362238407,
0.1374770849943161,
-0.08543037623167038,
-0.029841013252735138,
-0.012725061737000942,
0.0015431044157594442,
-0.01536548137664795,
0.09658888727426529,
0.09057372808456421,
0.06130669638514519,
-0.0581083782017231,
0.0878656879067421,
0.06505430489778519,
-0.12993492186069489,
0.0748833417892456,
0.11734998971223831,
-0.09341247379779816,
-0.04442073777318001,
-0.014730908907949924,
0.09598558396100998,
-0.004352628253400326,
-0.06942406296730042,
-0.03553401678800583,
-0.05442747101187706,
0.006994308438152075,
0.013820101507008076,
0.02109253779053688,
0.02499021776020527,
-0.0893305242061615,
0.03774488717317581,
-0.17469793558120728,
0.10036244988441467,
-0.05669589713215828,
0.018628878518939018,
-0.15796402096748352,
0.11058587580919266,
0.013718505389988422,
0.12749408185482025,
-0.026879139244556427,
-0.03250686451792717,
-0.06825758516788483,
-0.02350478246808052,
-0.09902729094028473,
0.03509525582194328,
-0.055099762976169586,
-0.021361948922276497,
-0.009451759979128838,
0.0593080036342144,
-0.03397646173834801,
0.061375875025987625,
-0.03203617408871651,
-0.03825535625219345,
-0.03777103126049042,
-0.007951532490551472,
-0.0702953115105629,
0.017413616180419922,
-0.014654004015028477,
-0.05552372708916664,
0.07954970002174377,
-0.041637446731328964,
-0.026718569919466972,
0.03670267388224602,
0.08424276858568192,
0.03195495903491974,
0.02637110836803913,
0.02550472691655159,
0.04410451278090477,
0.0027906291652470827,
0.001845145015977323,
0.008112234994769096,
-0.034366805106401443,
-0.04744882136583328,
0.013782696798443794,
-0.06374825537204742,
-0.003988170996308327,
-0.0711117535829544,
0.042406100779771805,
-0.0447426363825798,
0.08148878067731857,
0.04901253432035446,
0.035465240478515625,
0.055809393525123596,
-0.058375775814056396,
0.06057257577776909,
-0.12482760846614838,
-0.05911221355199814,
0.0314796045422554,
-0.030342796817421913,
-0.010087205097079277,
-0.06973439455032349,
0.035862065851688385,
-0.07887274026870728,
0.15780657529830933,
0.04261070489883423,
0.11324318498373032,
0.014319085516035557,
-0.1634550392627716,
-0.08779416978359222,
0.007092856336385012,
0.18563377857208252,
0.03199489414691925,
-0.010736756026744843,
0.03260720148682594,
0.033558931201696396,
-0.0020708914380520582,
0.11565103381872177,
0.1250579059123993,
-0.03536168113350868,
0.02845638617873192,
0.12242060154676437,
0.026716085150837898,
-0.035879477858543396,
-0.13783526420593262,
0.01163066178560257,
-0.05714265629649162,
0.11408662796020508,
-0.06558507680892944,
-0.027836283668875694,
0.21363282203674316,
-0.0946296826004982,
0.10407939553260803,
0.002776293782517314,
-0.07000042498111725,
-0.08789616078138351,
-0.15721938014030457,
-0.04738462716341019,
-0.12313863635063171,
-0.030244335532188416,
-0.13286641240119934,
0.06727684289216995,
0.07897362858057022,
0.004011448938399553,
-0.031941525638103485,
0.11826067417860031,
0.013859299942851067,
-0.06373712420463562,
0.034589286893606186,
-0.009034894406795502,
0.026351328939199448,
0.03934279829263687,
0.08221487700939178,
0.06607542186975479,
0.018375547602772713,
0.06798091530799866,
0.05738638713955879,
0.12599711120128632,
0.041117943823337555,
-0.03357372060418129,
-0.05379308760166168,
-0.0221824012696743,
0.010575284250080585,
-0.004123662132769823,
0.06506295502185822,
0.04956422746181488,
-0.07565885782241821,
-0.034462738782167435,
0.09981994330883026,
-0.06497890502214432,
-0.02921859733760357,
-0.09181968122720718,
0.11621768772602081,
0.02331257238984108,
0.07822538912296295,
0.01085728220641613,
-0.09844338148832321,
-0.04450950399041176,
0.22150084376335144,
0.11270467191934586,
0.030686188489198685,
0.011965623125433922,
0.05013643950223923,
0.0018154382705688477,
-0.049349017441272736,
0.170135498046875,
0.03877163678407669,
0.23721516132354736,
-0.04799175262451172,
-0.001933518098667264,
-0.023950425907969475,
-0.007442439906299114,
-0.06679996103048325,
0.07730937749147415,
-0.013000392355024815,
-0.012245827354490757,
0.015115676447749138,
0.006956877652555704,
-0.030422406271100044,
-0.16831789910793304,
0.033524464815855026,
-0.06351705640554428,
-0.043520230799913406,
0.03693699091672897,
0.002903989516198635,
0.008957106620073318,
0.07417849451303482,
-0.025139357894659042,
0.020773641765117645,
0.24488048255443573,
0.01650667004287243,
-0.07464957982301712,
-0.033127978444099426,
0.041611023247241974,
-0.047933436930179596,
0.16033388674259186,
0.0018052976811304688,
0.05183035880327225,
0.07483185082674026,
0.013185892254114151,
-0.1254570335149765,
-0.012934296391904354,
0.02052290551364422,
0.00007073237065924332,
-0.0352497361600399,
0.08517999202013016,
-0.013035633601248264,
-0.028518186882138252,
0.022692952305078506,
-0.08160614967346191,
-0.0008438150398433208,
-0.012770135886967182,
0.06884321570396423,
-0.047352395951747894,
-0.025135094299912453,
-0.08898429572582245,
0.127938911318779,
0.08836307376623154,
-0.008296037092804909,
-0.02270093932747841,
-0.006957540288567543,
0.0747096836566925,
0.0004607860173564404,
0.12509967386722565,
-0.00608002208173275,
-0.06737674027681351,
-0.013234317302703857,
-0.05201323702931404,
0.0066835917532444,
-0.1706818789243698,
-0.029993275180459023,
0.046318527311086655,
-0.031917065382003784,
-0.034376516938209534,
0.07591889798641205,
0.10142166912555695,
0.04014519229531288,
-0.07242535054683685,
-0.015774894505739212,
-0.03859666734933853,
0.1052083820104599,
-0.09445430338382721,
-0.07853350043296814
] |
null | null |
transformers
|
# HIV_BERT model
## Table of Contents
- [Summary](#model-summary)
- [Model Description](#model-description)
- [Intended Uses & Limitations](#intended-uses-&-limitations)
- [How to Use](#how-to-use)
- [Training Data](#training-data)
- [Training Procedure](#training-procedure)
- [Preprocessing](#preprocessing)
- [Training](#training)
- [Evaluation Results](#evaluation-results)
- [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info)
## Summary
The HIV-BERT model was trained as a refinement of the [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd) for HIV centric tasks. It was refined with whole viral genomes from the [Los Alamos HIV Sequence Database](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).
## Model Description
Like the original [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd), this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.
## Intended Uses & Limitations
As a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.
## How to use
As this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position.
```python
from transformers import pipeline
unmasker = pipeline("fill-mask", model="damlab/HIV_FLT")
unmasker(f"C T R P N [MASK] N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C")
[
{
"score": 0.9581968188285828,
"token": 17,
"token_str": "N",
"sequence": "C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"
},
{
"score": 0.022986575961112976,
"token": 12,
"token_str": "K",
"sequence": "C T R P N K N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"
},
{
"score": 0.003997281193733215,
"token": 14,
"token_str": "D",
"sequence": "C T R P N D N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"
},
{
"score": 0.003636382520198822,
"token": 15,
"token_str": "T",
"sequence": "C T R P N T N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"
},
{
"score": 0.002701344434171915,
"token": 10,
"token_str": "S",
"sequence": "C T R P N S N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"
}
]
```
## Training Data
The dataset [damlab/HIV_FLT](https://huggingface.co/datasets/damlab/HIV_FLT) was used to refine the original [rostlab/Prot-bert-bfd](https://huggingface.co/Rostlab/prot_bert_bfd). This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens.
## Training Procedure
### Preprocessing
As with the [rostlab/Prot-bert-bfd](https://huggingface.co/Rostlab/prot_bert_bfd) model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
Training was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset.
## BibTeX Entry and Citation Info
[More Information Needed]
|
{"license": "mit", "datasets": ["damlab/HIV_FLT"], "metrics": ["accuracy"], "widget": [{"text": "C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C", "example_title": "V3"}, {"text": "M E P V D P R L E P W K H P G S Q P K T A C T N C Y C K K C C F H C Q V C F I T K A L G I S Y G R K K R R Q R R R A H Q N S Q T H Q A S L S K Q P T S Q P R G D P T G P K E S K K K V E R E T E T D P F D", "example_title": "Tat"}, {"text": "P Q I T L W Q R P L V T I K I G G Q L K E A L L D T G A D D T V L E E M N L P G R W K P K M I G G I G G F I K V R Q Y D Q I L I E I C G H K A I G T V L V G P T P V N I I G R N L L T Q I G C T L N F", "example_title": "PR"}]}
|
fill-mask
|
damlab/HIV_BERT
|
[
"transformers",
"pytorch",
"bert",
"fill-mask",
"dataset:damlab/HIV_FLT",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #fill-mask #dataset-damlab/HIV_FLT #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# HIV_BERT model
## Table of Contents
- Summary
- Model Description
- Intended Uses & Limitations
- How to Use
- Training Data
- Training Procedure
- Preprocessing
- Training
- Evaluation Results
- BibTeX Entry and Citation Info
## Summary
The HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).
## Model Description
Like the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.
## Intended Uses & Limitations
As a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.
## How to use
As this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position.
## Training Data
The dataset damlab/HIV_FLT was used to refine the original rostlab/Prot-bert-bfd. This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens.
## Training Procedure
### Preprocessing
As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
Training was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset.
## BibTeX Entry and Citation Info
|
[
"# HIV_BERT model",
"## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).",
"## Model Description\r\n\r\nLike the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.",
"## Intended Uses & Limitations\r\n\r\nAs a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.",
"## How to use\r\n\r\nAs this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position.",
"## Training Data\r\n\r\nThe dataset damlab/HIV_FLT was used to refine the original rostlab/Prot-bert-bfd. This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nTraining was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset.",
"## BibTeX Entry and Citation Info"
] |
[
"TAGS\n#transformers #pytorch #bert #fill-mask #dataset-damlab/HIV_FLT #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# HIV_BERT model",
"## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).",
"## Model Description\r\n\r\nLike the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.",
"## Intended Uses & Limitations\r\n\r\nAs a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.",
"## How to use\r\n\r\nAs this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position.",
"## Training Data\r\n\r\nThe dataset damlab/HIV_FLT was used to refine the original rostlab/Prot-bert-bfd. This dataset contains 1790 full HIV genomes from across the globe. When translated, these genomes contain approximately 3.9 million amino-acid tokens.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nTraining was performed with the HuggingFace training module using the MaskedLM data loader with a 15% masking rate. The learning rate was set at E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset.",
"## BibTeX Entry and Citation Info"
] |
[
53,
6,
52,
109,
93,
96,
35,
69,
4,
80,
82,
11
] |
[
"passage: TAGS\n#transformers #pytorch #bert #fill-mask #dataset-damlab/HIV_FLT #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# HIV_BERT model## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT model was trained as a refinement of the ProtBert-BFD model for HIV centric tasks. It was refined with whole viral genomes from the Los Alamos HIV Sequence Database. This pretraining is important for HIV related tasks as the original BFD database contains few viral proteins making it sub-optimal when used as the basis for transfer learning tasks. This model and other related HIV prediction tasks have been published (link).## Model Description\r\n\r\nLike the original ProtBert-BFD model, this model encodes each amino acid as an individual token. This model was trained using Masked Language Modeling: a process in which a random set of tokens are masked with the model trained on their prediction. This model was trained using the damlab/hiv-flt dataset with 256 amino acid chunks and a 15% mask rate.## Intended Uses & Limitations\r\n\r\nAs a masked language model this tool can be used to predict expected mutations using a masking approach. This could be used to identify highly mutated sequences, sequencing artifacts, or other contexts. As a BERT model, this tool can also be used as the base for transfer learning. This pretrained model could be used as the base when developing HIV-specific classification tasks.## How to use\r\n\r\nAs this is a BERT-style Masked Language learner, it can be used to determine the most likely amino acid at a masked position."
] |
[
-0.03160731494426727,
0.08184020221233368,
-0.0024370141327381134,
-0.0017657733988016844,
0.1079944297671318,
0.006299123167991638,
0.11540769785642624,
0.10264714062213898,
0.07042404264211655,
0.08010247349739075,
0.013361725956201553,
-0.08970969170331955,
0.07429111748933792,
0.17727544903755188,
0.10472066700458527,
-0.2174062728881836,
0.06442227214574814,
-0.05051722005009651,
-0.04334679991006851,
0.06222013011574745,
0.08324696868658066,
-0.0671345591545105,
0.032846543937921524,
0.0032635671086609364,
0.03448178246617317,
0.013491138815879822,
-0.03732994571328163,
-0.039532002061605453,
0.0732407197356224,
0.0759970024228096,
0.07817909866571426,
0.07251995801925659,
0.10258915275335312,
-0.24010802805423737,
0.005486767273396254,
0.07223475724458694,
-0.019493550062179565,
0.027336424216628075,
0.08258278667926788,
0.011854149401187897,
0.15314346551895142,
-0.030468327924609184,
0.09030371904373169,
0.08182527124881744,
-0.11814821511507034,
-0.1487777978181839,
-0.06202559545636177,
0.11534034460783005,
-0.007193629164248705,
0.08529489487409592,
-0.0072442227974534035,
0.03316739201545715,
0.03165663778781891,
0.032895952463150024,
0.07277265191078186,
-0.13950832188129425,
0.0044417791068553925,
0.18499457836151123,
0.07112042605876923,
0.032196447253227234,
-0.06537136435508728,
0.0008379845530726016,
-0.01113959588110447,
0.06624946743249893,
0.18331147730350494,
-0.04116291552782059,
0.014031779021024704,
-0.0869210734963417,
-0.16838794946670532,
-0.09083826839923859,
0.074562206864357,
-0.012571455910801888,
-0.07326241582632065,
-0.11875617504119873,
-0.007513594813644886,
0.08611667901277542,
-0.02506449446082115,
-0.04385041072964668,
-0.030140668153762817,
0.054710716009140015,
-0.006755960639566183,
-0.10260245203971863,
-0.015962377190589905,
-0.09109465777873993,
0.049168769270181656,
0.0693390816450119,
0.015662826597690582,
0.02303341031074524,
-0.06443852186203003,
0.04439757019281387,
-0.13740015029907227,
-0.06974027305841446,
-0.03589406609535217,
-0.05470375344157219,
-0.05780988559126854,
-0.029501045122742653,
0.0006595068261958659,
-0.23916099965572357,
0.009552656672894955,
0.07025258243083954,
-0.03208720311522484,
0.013530377298593521,
-0.0480194017291069,
0.008104556240141392,
0.07704471796751022,
0.10271189361810684,
-0.09971120208501816,
0.007560933008790016,
0.019337201490998268,
-0.0014732856070622802,
-0.01306725014001131,
0.0016453496646136045,
-0.021655114367604256,
0.01043171901255846,
0.040306538343429565,
0.0770094096660614,
-0.01743638701736927,
0.05049917474389076,
0.03415177762508392,
-0.0168002862483263,
0.1082007884979248,
-0.07711682468652725,
-0.03906136006116867,
-0.014817952178418636,
-0.015302197076380253,
-0.013202599249780178,
0.11944244801998138,
-0.07413462549448013,
-0.03427067771553993,
0.06302405893802643,
-0.054763101041316986,
-0.0794227123260498,
-0.0979824811220169,
-0.1480003446340561,
-0.014773267321288586,
0.06731698662042618,
-0.06953515112400055,
-0.07586141675710678,
-0.16603928804397583,
-0.08370038866996765,
-0.0031973354052752256,
0.011809129267930984,
0.010481659322977066,
-0.002352563664317131,
-0.002000609878450632,
0.007760423235595226,
0.011679306626319885,
-0.00511739170178771,
-0.017887115478515625,
0.009769155643880367,
-0.1315736621618271,
0.08200034499168396,
0.0767325833439827,
0.023149242624640465,
-0.09360368549823761,
0.08319010585546494,
-0.25641965866088867,
0.09263823181390762,
-0.14606024324893951,
0.013222502544522285,
-0.13882675766944885,
-0.06467430293560028,
-0.05345821753144264,
-0.023891186341643333,
0.0452410951256752,
0.10663195699453354,
-0.10864375531673431,
-0.025458596646785736,
0.3334011137485504,
-0.13566696643829346,
-0.008723828941583633,
0.06359130889177322,
-0.03992639482021332,
0.14630448818206787,
0.08286789059638977,
0.07404768466949463,
0.10201387107372284,
-0.029866553843021393,
-0.05626515671610832,
-0.0029493430629372597,
-0.021125486120581627,
0.1304551362991333,
0.003975489642471075,
-0.07363470643758774,
-0.016918087378144264,
-0.004502100870013237,
-0.028936713933944702,
0.013756237924098969,
-0.029400087893009186,
0.0018798175733536482,
0.00503338361158967,
-0.0561564601957798,
0.023376064375042915,
-0.018961075693368912,
-0.04946143180131912,
0.023546185344457626,
-0.0778389573097229,
0.04167269542813301,
0.13663369417190552,
-0.111199289560318,
0.03823857009410858,
-0.068715900182724,
0.022078987210989,
-0.014179069548845291,
-0.0015426711179316044,
-0.1750757396221161,
-0.08655544370412827,
0.04300263896584511,
-0.09966649115085602,
0.07571746408939362,
-0.11990191042423248,
0.014994307421147823,
0.049916334450244904,
-0.06765154749155045,
0.007487924303859472,
-0.07730565965175629,
-0.013647693209350109,
-0.07051810622215271,
-0.129505917429924,
-0.08810960501432419,
-0.05179751664400101,
0.10057520866394043,
-0.12918749451637268,
-0.0020136674866080284,
-0.05913009122014046,
0.05756545811891556,
0.07480531185865402,
-0.05125879868865013,
0.12508521974086761,
0.04679304361343384,
-0.00485282763838768,
-0.04968849569559097,
-0.007092348299920559,
0.017676912248134613,
-0.06722059845924377,
-0.004071415867656469,
-0.054200902581214905,
-0.22943422198295593,
-0.002058328129351139,
0.05710954591631889,
-0.10711275041103363,
0.02745118737220764,
0.007747008930891752,
0.0004480364441405982,
-0.0932864099740982,
-0.015478672459721565,
0.12693412601947784,
0.0074186441488564014,
0.07410623133182526,
-0.02657615765929222,
0.04684929549694061,
0.026029640808701515,
0.017007220536470413,
-0.009218432940542698,
0.03400113806128502,
0.09209150820970535,
-0.15974603593349457,
0.05429435521364212,
-0.04989403113722801,
0.03709767758846283,
0.11756640672683716,
0.015987396240234375,
-0.09369159489870071,
0.0410664901137352,
-0.024875590577721596,
0.044307123869657516,
0.07563639432191849,
-0.06674271821975708,
0.004805553238838911,
0.03249132260680199,
-0.005684540141373873,
0.011619137600064278,
0.010314682498574257,
0.0743369609117508,
0.026360981166362762,
-0.0753566101193428,
-0.07903289794921875,
-0.05104093253612518,
-0.03921645134687424,
0.07496732473373413,
0.04149901494383812,
-0.0035801606718450785,
-0.03663461655378342,
-0.030806943774223328,
-0.13376981019973755,
0.13642382621765137,
-0.08264297991991043,
-0.12520423531532288,
-0.16399821639060974,
0.0393928661942482,
-0.02282705530524254,
0.012948415242135525,
-0.006458152085542679,
0.03005753643810749,
-0.07008805125951767,
-0.11811563372612,
0.11437851935625076,
-0.06783881783485413,
-0.04900658130645752,
-0.050251949578523636,
0.03875108063220978,
0.06965424120426178,
-0.08301272988319397,
0.030640767887234688,
-0.006817636080086231,
-0.028605429455637932,
0.02017347887158394,
-0.04948161542415619,
0.1089276373386383,
0.10561530292034149,
0.06171383708715439,
-0.02672586217522621,
-0.06609233468770981,
0.19854214787483215,
-0.06727918237447739,
0.11289804428815842,
0.10889790207147598,
-0.03343617543578148,
0.03395366296172142,
0.08421114832162857,
0.0023107193410396576,
-0.023346057161688805,
0.07187540829181671,
0.04792638123035431,
-0.06971272826194763,
-0.18097008764743805,
-0.0976305678486824,
-0.020187528803944588,
-0.010687149129807949,
0.042016346007585526,
0.03795215114951134,
0.06717193126678467,
0.025557825341820717,
-0.08396553248167038,
0.017423806712031364,
0.0007197275408543646,
0.06256408989429474,
-0.016830705106258392,
0.019698625430464745,
0.04722112789750099,
-0.03720737248659134,
0.013955305330455303,
0.082737497985363,
-0.0956791415810585,
0.20778913795948029,
0.006416297983378172,
0.1454060971736908,
0.07481274753808975,
0.08124256134033203,
0.05509515106678009,
0.07210695743560791,
-0.03684117645025253,
0.033437591046094894,
0.006032864097505808,
-0.055803775787353516,
0.020915597677230835,
0.03014427237212658,
-0.028910603374242783,
0.0010487191611900926,
0.0033973711542785168,
-0.10603979229927063,
-0.019858913496136665,
0.19137801229953766,
0.06879045814275742,
-0.13722455501556396,
-0.09666093438863754,
0.023248231038451195,
-0.044597748667001724,
-0.13211849331855774,
-0.008966930210590363,
0.08480949699878693,
-0.16601990163326263,
0.05876415595412254,
-0.02526981383562088,
0.1181800588965416,
-0.0379619374871254,
-0.03588990494608879,
-0.07319354265928268,
0.0526764802634716,
-0.09817440062761307,
0.09795933961868286,
-0.23473189771175385,
0.09108690917491913,
0.009039143100380898,
0.07738828659057617,
-0.07091183960437775,
0.007512554991990328,
0.07903625816106796,
0.1357707977294922,
0.1511835753917694,
0.01604856364428997,
-0.0828404575586319,
-0.060985926538705826,
-0.12917175889015198,
-0.006699876859784126,
0.0529341958463192,
-0.10671232640743256,
0.1350230723619461,
0.027585729956626892,
0.019079705700278282,
-0.02510324865579605,
0.10989196598529816,
-0.20259182155132294,
-0.08384209126234055,
0.02859257161617279,
-0.041456107050180435,
0.13330771028995514,
-0.03827917203307152,
-0.038814838975667953,
-0.0118956184014678,
0.07642661035060883,
-0.08309301733970642,
-0.09962732344865799,
-0.14849746227264404,
0.020433329045772552,
0.10089107602834702,
-0.08184989541769028,
0.041065577417612076,
0.015195704065263271,
0.16021110117435455,
-0.05906122177839279,
-0.1313001811504364,
0.04946618899703026,
-0.08779840171337128,
-0.15352584421634674,
-0.0683356299996376,
0.04100200906395912,
0.15021930634975433,
0.10053034126758575,
0.031091948971152306,
0.05869051814079285,
0.08793776482343674,
-0.07756544649600983,
0.05476096272468567,
0.1467912346124649,
-0.010593698360025883,
0.06717412918806076,
-0.11423752456903458,
-0.0022176317870616913,
-0.1154656857252121,
-0.0679335743188858,
0.10547921061515808,
0.07683044672012329,
-0.06638002395629883,
0.07459528744220734,
0.17174145579338074,
-0.14396347105503082,
-0.26237061619758606,
-0.021390384063124657,
0.07517693191766739,
0.05139570310711861,
0.04520454630255699,
-0.31111499667167664,
-0.013472672551870346,
0.012048818171024323,
-0.04515145719051361,
0.030407149344682693,
-0.30475470423698425,
-0.11408619582653046,
0.09906263649463654,
0.021627308800816536,
-0.04134950041770935,
-0.10673472285270691,
-0.05366339161992073,
-0.05232163891196251,
0.005613545887172222,
0.09186619520187378,
-0.003874091897159815,
0.0290667787194252,
-0.0011168023338541389,
0.020208969712257385,
0.022507863119244576,
0.007088452577590942,
0.06697382032871246,
-0.004702885635197163,
0.05286511778831482,
-0.05073634162545204,
0.09919659793376923,
0.08743195235729218,
-0.06430305540561676,
0.09296002984046936,
0.09006441384553909,
0.02909923531115055,
-0.14122337102890015,
-0.06018795818090439,
-0.07830724865198135,
0.04160580411553383,
-0.0311388298869133,
-0.07857856899499893,
-0.10017159581184387,
0.07978589832782745,
0.07483614236116409,
-0.014357876032590866,
0.028417523950338364,
-0.10843921452760696,
0.051863063126802444,
0.14290058612823486,
0.10864768177270889,
0.09554078429937363,
-0.11520737409591675,
0.06326237320899963,
-0.02823018841445446,
0.06830411404371262,
-0.04200564697384834,
0.08296621590852737,
0.04463620111346245,
0.044546496123075485,
0.13018985092639923,
0.024470621719956398,
-0.19115090370178223,
0.00866314209997654,
0.038987815380096436,
-0.11250537633895874,
-0.15485775470733643,
-0.012188976630568504,
0.01697447896003723,
-0.144207164645195,
-0.06209023296833038,
0.11877616494894028,
-0.09773363173007965,
-0.02378847263753414,
-0.06136301904916763,
0.07287470251321793,
-0.006168958730995655,
0.11332488805055618,
0.0769299641251564,
0.010792993009090424,
-0.033668018877506256,
0.11311119794845581,
0.10772878676652908,
-0.023058723658323288,
0.017621925100684166,
0.103319451212883,
-0.12781469523906708,
-0.04093518480658531,
0.013308694586157799,
0.17064911127090454,
-0.031489789485931396,
-0.08550266176462173,
0.010269268415868282,
-0.0732814148068428,
0.03020956926047802,
0.2873147130012512,
-0.012263068929314613,
-0.010687115602195263,
-0.04333498328924179,
-0.005429246928542852,
-0.1097390204668045,
0.047697339206933975,
-0.024259664118289948,
0.03277336433529854,
-0.00228871894069016,
0.12751010060310364,
0.03488048538565636,
0.0032585631124675274,
-0.05147862434387207,
-0.045694150030612946,
-0.07938101887702942,
-0.0007907294202595949,
-0.07521766424179077,
0.0083607267588377,
-0.05946226045489311,
-0.06398316472768784,
-0.021153632551431656,
0.014610878191888332,
0.03417413309216499,
0.037653207778930664,
-0.05514027923345566,
-0.06387418508529663,
-0.0534047931432724,
0.08635072410106659,
-0.104946568608284,
-0.024999504908919334,
0.026352720335125923,
-0.06258925050497055,
0.07679574936628342,
0.022163791581988335,
0.03532474488019943,
0.004434044472873211,
0.013533213175833225,
0.028438910841941833,
-0.044944312423467636,
0.009288948960602283,
-0.019555632025003433,
-0.13018125295639038,
-0.04048657417297363,
-0.09097623825073242,
-0.08578509837388992,
0.012447060085833073,
0.028310224413871765,
-0.022109124809503555,
0.041516974568367004,
0.006526718381792307,
0.07814847677946091,
-0.0106589924544096,
0.0719728097319603,
0.04976297914981842,
0.06030280888080597,
0.0757729634642601,
-0.052615225315093994,
0.05220349133014679,
-0.16826477646827698,
-0.01869014836847782,
-0.011319675482809544,
0.031503405421972275,
-0.0016281139105558395,
-0.004496106877923012,
0.055898621678352356,
0.019526282325387,
0.1037278026342392,
0.015969248488545418,
0.05092817172408104,
-0.0072476621717214584,
-0.04169609025120735,
0.02931950055062771,
-0.02954006753861904,
0.07024990022182465,
0.06208176538348198,
-0.01353778038173914,
0.07201679050922394,
0.03634333238005638,
-0.03969399258494377,
0.11280103772878647,
0.12077973037958145,
0.027768276631832123,
0.17382313311100006,
0.02325740084052086,
-0.0022627986036241055,
-0.05406462773680687,
-0.041376009583473206,
-0.13184191286563873,
0.07027874886989594,
0.06977260857820511,
0.0015727955615147948,
0.0796724185347557,
0.091609425842762,
-0.16743120551109314,
0.17163099348545074,
0.04555178061127663,
-0.08702503144741058,
-0.0664086788892746,
-0.17916418612003326,
-0.0180982518941164,
-0.0658874437212944,
-0.018318844959139824,
-0.1344824731349945,
-0.004736015107482672,
0.23141685128211975,
-0.016429012641310692,
0.02095138654112816,
0.04612371698021889,
-0.11564882099628448,
-0.13097380101680756,
0.07957284152507782,
0.010798080824315548,
0.013363288715481758,
-0.06348557025194168,
0.05876903235912323,
0.061607252806425095,
0.04946037009358406,
0.03263687714934349,
0.05961637198925018,
0.0638686940073967,
0.028516510501503944,
-0.0000816223764559254,
-0.04278930649161339,
0.0031329705379903316,
-0.007779599167406559,
-0.014379450120031834,
0.2180221974849701,
0.05310342460870743,
-0.023029213771224022,
0.0006738431402482092,
0.19974569976329803,
-0.0627751350402832,
0.001479700906202197,
-0.14140281081199646,
0.285643607378006,
0.02874072454869747,
-0.011257763020694256,
-0.022850453853607178,
-0.10173040628433228,
-0.007690343074500561,
0.2491970956325531,
0.05874577537178993,
0.007352062501013279,
-0.013507021591067314,
-0.039145778864622116,
-0.00336787779815495,
0.05735045671463013,
0.07761713862419128,
0.00730143254622817,
0.1860969364643097,
-0.0791102796792984,
0.1288548707962036,
-0.07933977246284485,
-0.011456173844635487,
-0.09948314726352692,
0.06998540461063385,
-0.034847237169742584,
-0.023341134190559387,
-0.04567922279238701,
0.10097071528434753,
-0.09982161223888397,
-0.25055545568466187,
0.0066065252758562565,
-0.02080710604786873,
-0.0699877142906189,
-0.03633860498666763,
-0.15775062143802643,
0.07641047239303589,
0.03961900621652603,
0.009179677814245224,
0.08007105439901352,
0.16303613781929016,
0.06013350561261177,
0.027892109006643295,
-0.14789946377277374,
0.06910606473684311,
-0.10954555124044418,
0.258358895778656,
0.010404652915894985,
0.004275194369256496,
0.05665592476725578,
-0.03424082696437836,
-0.11772129684686661,
0.05173727497458458,
0.027354545891284943,
0.05828693509101868,
-0.0024237572215497494,
0.19032509624958038,
-0.02321215346455574,
0.13119223713874817,
0.012200595811009407,
-0.0574658066034317,
0.06812827289104462,
-0.029567541554570198,
-0.07447396963834763,
-0.03271806985139847,
0.05477922037243843,
-0.08938704431056976,
0.11554192006587982,
0.16245220601558685,
-0.08159726858139038,
-0.016407953575253487,
-0.044470228254795074,
0.005118480417877436,
-0.002644595690071583,
0.05763623118400574,
-0.0526757538318634,
-0.08004289120435715,
0.016442440450191498,
-0.09890700876712799,
-0.0070437039248645306,
-0.29775190353393555,
-0.013063786551356316,
0.0060995640233159065,
-0.015924198552966118,
0.02456398867070675,
0.07617245614528656,
-0.0017113048816099763,
-0.016144322231411934,
-0.037635937333106995,
-0.014578529633581638,
0.013219890184700489,
0.0966174378991127,
-0.12027067691087723,
-0.04590489715337753
] |
null | null |
transformers
|
# HIV_PR_resist model
## Table of Contents
- [Summary](#model-summary)
- [Model Description](#model-description)
- [Intended Uses & Limitations](#intended-uses-&-limitations)
- [How to Use](#how-to-use)
- [Training Data](#training-data)
- [Training Procedure](#training-procedure)
- [Preprocessing](#preprocessing)
- [Training](#training)
- [Evaluation Results](#evaluation-results)
- [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info)
## Summary
The HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the [Stanford HIV Genotype-Phenotype Database](https://hivdb.stanford.edu/pages/genotype-phenotype.html), allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.
## Model Description
The HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.
## Intended Uses & Limitations
This tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.
## How to use
*Prediction example of protease sequences*
## Training Data
This model was trained using the [damlab/HIV-PI dataset](https://huggingface.co/datasets/damlab/HIV_PI) using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database.
## Training Procedure
### Preprocessing
As with the [rostlab/Prot-bert-bfd model](https://huggingface.co/Rostlab/prot_bert_bfd), the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
The [damlab/HIV-BERT model](https://huggingface.co/damlab/HIV_BERT) was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.
## Evaluation Results
*Need to add*
## BibTeX Entry and Citation Info
[More Information Needed]
|
{"license": "mit"}
|
text-classification
|
damlab/HIV_PR_resist
|
[
"transformers",
"pytorch",
"bert",
"text-classification",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# HIV_PR_resist model
## Table of Contents
- Summary
- Model Description
- Intended Uses & Limitations
- How to Use
- Training Data
- Training Procedure
- Preprocessing
- Training
- Evaluation Results
- BibTeX Entry and Citation Info
## Summary
The HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.
## Model Description
The HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.
## Intended Uses & Limitations
This tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.
## How to use
*Prediction example of protease sequences*
## Training Data
This model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database.
## Training Procedure
### Preprocessing
As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.
## Evaluation Results
*Need to add*
## BibTeX Entry and Citation Info
|
[
"# HIV_PR_resist model",
"## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.",
"## Model Description\r\n\r\nThe HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.",
"## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.",
"## How to use\r\n\r\n*Prediction example of protease sequences*",
"## Training Data\r\n\r\nThis model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.",
"## Evaluation Results\r\n\r\n*Need to add*",
"## BibTeX Entry and Citation Info"
] |
[
"TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# HIV_PR_resist model",
"## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.",
"## Model Description\r\n\r\nThe HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.",
"## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.",
"## How to use\r\n\r\n*Prediction example of protease sequences*",
"## Training Data\r\n\r\nThis model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be resistant to multiple drugs) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.",
"## Evaluation Results\r\n\r\n*Need to add*",
"## BibTeX Entry and Citation Info"
] |
[
41,
8,
52,
140,
145,
45,
16,
60,
4,
80,
147,
10,
11
] |
[
"passage: TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# HIV_PR_resist model## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT-Protease-Resistance model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict whether an HIV protease sequence will be resistant to certain protease inhibitors. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV protease sequences from the Stanford HIV Genotype-Phenotype Database, allowing even more precise prediction protease inhibitor resistance than the HIV-BERT model can provide.## Model Description\r\n\r\nThe HIV-BERT-Protease-Resistance model is intended to predict the likelihood that an HIV protease sequence will be resistant to protease inhibitors. The protease gene is responsible for cleaving viral proteins into their active states, and as such is an ideal target for antiretroviral therapy. Annotation programs designed to predict and identify protease resistance using known mutations already exist, however with varied results. The HIV-BERT-Protease-Resistance model is designed to provide an alternative, NLP-based mechanism for predicting resistance mutations when provided with an HIV protease sequence.## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of protease resistance mutations within an HIV genomic sequence. It should not be considered a clinical diagnostic tool.## How to use\r\n\r\n*Prediction example of protease sequences*## Training Data\r\n\r\nThis model was trained using the damlab/HIV-PI dataset using the 0th fold. The dataset consists of 1959 sequences (approximately 99 tokens each) extracted from the Stanford HIV Genotype-Phenotype Database."
] |
[
0.03759939223527908,
0.1811004877090454,
-0.005275247152894735,
-0.0016653244383633137,
0.06647336483001709,
-0.02797546423971653,
0.07238733768463135,
0.09313631057739258,
0.08026246726512909,
0.09454163908958435,
-0.022272001951932907,
-0.053498901426792145,
0.07700856775045395,
0.14030224084854126,
0.11999727785587311,
-0.1583612561225891,
0.021882042288780212,
-0.07187645137310028,
0.08002239465713501,
0.05448950082063675,
0.05009039491415024,
-0.0752778947353363,
0.0388336181640625,
-0.01012490689754486,
0.17756116390228271,
0.08721281588077545,
-0.032658182084560394,
-0.0684254914522171,
0.06963898986577988,
0.029028626158833504,
0.0370267853140831,
0.0573788620531559,
0.0966794416308403,
-0.31986719369888306,
-0.01749889738857746,
0.05986238643527031,
0.023922357708215714,
0.01625056378543377,
0.05478396266698837,
-0.01305504608899355,
-0.058195777237415314,
-0.10853391885757446,
0.0759270042181015,
0.02460586652159691,
-0.09214690327644348,
-0.049027614295482635,
-0.09103091806173325,
0.061085738241672516,
-0.02033831551671028,
0.10036751627922058,
0.008781551383435726,
0.06017056852579117,
0.0696747675538063,
-0.02498628944158554,
0.12560752034187317,
-0.19922679662704468,
0.024764124304056168,
0.07729950547218323,
0.033481746912002563,
0.07083631306886673,
-0.05794446915388107,
-0.011227642185986042,
-0.04463041201233864,
0.014577838592231274,
0.19673016667366028,
-0.03273582458496094,
0.09773971885442734,
-0.05113109201192856,
-0.1341041475534439,
-0.11683572828769684,
0.12823784351348877,
-0.05819873511791229,
-0.06788526475429535,
-0.17247320711612701,
-0.0010948420967906713,
0.08283565938472748,
-0.0028273488860577345,
-0.07561152428388596,
0.05030634254217148,
0.03513815999031067,
-0.03077141009271145,
-0.13899745047092438,
-0.03295093774795532,
-0.009500078856945038,
-0.041425447911024094,
0.06321065127849579,
0.03332485258579254,
-0.010962124913930893,
-0.009678233414888382,
0.10489029437303543,
-0.13822005689144135,
-0.04670137166976929,
-0.06343226134777069,
-0.06876541674137115,
-0.06350399553775787,
-0.045724645256996155,
0.017421245574951172,
-0.15425941348075867,
-0.001517034717835486,
0.14616365730762482,
0.0031839213334023952,
0.028689850121736526,
-0.009004035964608192,
-0.03304310142993927,
0.08482928574085236,
0.17576980590820312,
-0.06782957911491394,
-0.08227598667144775,
-0.005507292225956917,
-0.009776064194738865,
0.07551927864551544,
0.00890791229903698,
0.03580969572067261,
-0.03999567776918411,
0.019512062892317772,
0.08640681207180023,
-0.03331591188907623,
0.021170027554035187,
-0.0037219570949673653,
0.0006380330305546522,
0.031259745359420776,
-0.10808227956295013,
-0.009877776727080345,
-0.06388071179389954,
0.009706777520477772,
0.042752064764499664,
0.03746505081653595,
-0.04442311078310013,
-0.07136335968971252,
0.0412011593580246,
-0.022126296535134315,
-0.060908928513526917,
-0.10125567018985748,
-0.06278789043426514,
-0.005686821881681681,
0.06507955491542816,
-0.07500436902046204,
-0.05149383470416069,
-0.08016815036535263,
-0.03620055317878723,
-0.06018335744738579,
-0.020619908347725868,
-0.0057435669004917145,
0.029504727572202682,
0.04645809158682823,
-0.014163045212626457,
-0.05207303166389465,
-0.13491271436214447,
-0.008516674861311913,
-0.023376157507300377,
-0.03553251549601555,
0.08672550320625305,
0.14190787076950073,
0.0015477915294468403,
-0.12378039211034775,
0.07251696288585663,
-0.312050998210907,
0.08705675601959229,
-0.13963238894939423,
-0.03205489367246628,
-0.13100355863571167,
-0.02807345613837242,
-0.007766786962747574,
-0.06336680054664612,
0.03532043844461441,
0.01419304870069027,
-0.1384613811969757,
0.00836973637342453,
0.27244964241981506,
-0.09564736485481262,
0.006154497154057026,
-0.0070581091567873955,
-0.03632959723472595,
0.09271364659070969,
0.08387318253517151,
0.0694236159324646,
0.06896570324897766,
-0.07470788061618805,
-0.13361191749572754,
-0.040645260363817215,
-0.043732449412345886,
0.19100266695022583,
0.018884822726249695,
-0.0801161378622055,
0.04775583744049072,
0.04335174709558487,
-0.060410164296627045,
-0.06926874816417694,
0.016057610511779785,
-0.008313636295497417,
-0.017341557890176773,
-0.0050749084912240505,
-0.03172372654080391,
-0.07324813306331635,
-0.09037307649850845,
0.03587707504630089,
-0.09270653873682022,
0.04492701590061188,
0.10338139533996582,
-0.097435861825943,
0.05097056180238724,
-0.09203905612230301,
-0.03374597057700157,
0.00649010855704546,
-0.03952035307884216,
-0.14833825826644897,
-0.0851527601480484,
0.0947282686829567,
-0.19074124097824097,
0.07122239470481873,
-0.10220426321029663,
0.026657545939087868,
0.030098918825387955,
-0.029873989522457123,
0.006681392900645733,
-0.05475631728768349,
0.000668159918859601,
-0.044532015919685364,
-0.035881832242012024,
-0.0659172385931015,
-0.006512963213026524,
0.12988300621509552,
0.00792597234249115,
-0.0030877566896378994,
-0.10502265393733978,
0.10357020050287247,
0.04787364602088928,
-0.014810997992753983,
0.08366970717906952,
0.02277916669845581,
-0.012709617614746094,
-0.014826282858848572,
-0.008842221461236477,
-0.05102645605802536,
0.0006726623978465796,
0.03935408592224121,
-0.07918965816497803,
-0.2221839725971222,
0.016947265714406967,
0.11526995152235031,
-0.057258911430835724,
-0.03404441475868225,
0.016513053327798843,
0.0074379644356667995,
-0.10275450348854065,
0.01089879684150219,
0.18504828214645386,
0.06840232014656067,
0.07905574887990952,
-0.027851048856973648,
-0.05271309241652489,
-0.01441711001098156,
0.0139690563082695,
-0.03298703208565712,
-0.003460625885054469,
0.124714195728302,
-0.06294621527194977,
0.03458625078201294,
0.012200135737657547,
0.09262026101350784,
0.02443649247288704,
0.030827391892671585,
-0.13604114949703217,
-0.012728985399007797,
-0.07606378197669983,
0.029514286667108536,
0.012785453349351883,
0.06849613785743713,
0.046066246926784515,
0.020651958882808685,
-0.04153621941804886,
-0.028905903920531273,
-0.015369297005236149,
0.064766526222229,
0.04846344143152237,
-0.04698481410741806,
-0.033767785876989365,
-0.08754422515630722,
0.023052990436553955,
0.0541781410574913,
0.024164851754903793,
0.08100020885467529,
-0.07884149253368378,
-0.008584176190197468,
-0.0635836273431778,
0.039979238063097,
-0.07380315661430359,
-0.2706591486930847,
-0.1736638844013214,
0.02013574168086052,
-0.002409926615655422,
0.005408197175711393,
-0.03597337007522583,
0.029153980314731598,
-0.07886345684528351,
-0.1492493450641632,
0.08341369032859802,
-0.013178545981645584,
-0.12553085386753082,
-0.02026027999818325,
0.10003422945737839,
0.08914697170257568,
-0.06733259558677673,
0.009895531460642815,
-0.005878643598407507,
-0.07189464569091797,
0.021841909736394882,
0.05086406320333481,
0.10593097656965256,
-0.01905815303325653,
0.03232419863343239,
-0.0713084414601326,
-0.05403566360473633,
0.11207576096057892,
-0.04962191358208656,
0.08423410356044769,
0.1685478538274765,
-0.04984462633728981,
0.06545472145080566,
0.07483260333538055,
-0.030343621969223022,
-0.00303398328833282,
0.06539268791675568,
0.063563272356987,
0.008387415669858456,
-0.1520681381225586,
-0.0407644584774971,
0.026288751512765884,
-0.0022619592491537333,
0.08520010113716125,
0.020384911447763443,
0.016820218414068222,
0.04035505652427673,
-0.08212796598672867,
0.028108257800340652,
0.026777375489473343,
0.07921186089515686,
0.18859553337097168,
-0.03249955177307129,
0.07380087673664093,
-0.04385247826576233,
-0.07986612617969513,
0.08481666445732117,
0.06506460160017014,
0.1401170790195465,
-0.011412478052079678,
0.1483108103275299,
0.021862756460905075,
0.12464991211891174,
0.042023781687021255,
0.0314248725771904,
0.02337219938635826,
0.03774510696530342,
-0.006389522459357977,
-0.07185009121894836,
-0.052570175379514694,
0.008461512625217438,
-0.003190064337104559,
-0.0074364920146763325,
0.07872216403484344,
-0.05678480491042137,
-0.0061725773848593235,
0.16314932703971863,
0.014008614234626293,
-0.02947455830872059,
-0.026647068560123444,
0.05906763672828674,
-0.11058709025382996,
-0.09346568584442139,
0.035319019109010696,
0.06496059894561768,
-0.16121035814285278,
0.07342901825904846,
0.032305583357810974,
0.09486091136932373,
-0.15488353371620178,
-0.02228248491883278,
0.0025402952451258898,
0.057733092457056046,
-0.03541799634695053,
0.11765548586845398,
-0.15791404247283936,
-0.06339266896247864,
0.004578631836920977,
0.05556744337081909,
-0.05548364669084549,
0.04722217842936516,
-0.0370306521654129,
0.021431710571050644,
0.1475292444229126,
0.0033288474660366774,
0.031824659556150436,
-0.06494658440351486,
-0.19097405672073364,
-0.019413381814956665,
0.06244737654924393,
-0.16081535816192627,
0.17074409127235413,
-0.05187441408634186,
-0.004669396672397852,
-0.06761005520820618,
0.026811186224222183,
-0.15285727381706238,
-0.20637941360473633,
0.04397650435566902,
-0.056215040385723114,
0.1372886300086975,
-0.03868239372968674,
-0.008260798640549183,
0.005108225625008345,
0.06520271301269531,
-0.15427598357200623,
-0.08898352086544037,
-0.12270700931549072,
-0.031082777306437492,
0.17381176352500916,
-0.05810571834445,
0.02394566312432289,
0.034921564161777496,
0.20335090160369873,
-0.04102395474910736,
-0.15232068300247192,
-0.02884870208799839,
-0.03996500372886658,
-0.12436136603355408,
-0.06956590712070465,
0.11134812980890274,
0.03589320927858353,
0.10841906070709229,
0.02668655477464199,
0.08365277945995331,
0.0629497766494751,
-0.057779014110565186,
0.03994737192988396,
0.22221556305885315,
0.009785555303096771,
0.01050223782658577,
-0.1030258983373642,
-0.03467562794685364,
-0.12832535803318024,
-0.06445877254009247,
0.1747114062309265,
0.13407352566719055,
-0.05581485852599144,
0.1269439458847046,
0.06776411831378937,
-0.13774912059307098,
-0.20570212602615356,
-0.08163122832775116,
0.0879523754119873,
-0.02847713977098465,
0.09350083768367767,
-0.3371870815753937,
0.03778991848230362,
0.06769061088562012,
0.0006754352943971753,
-0.07756920903921127,
-0.1849459707736969,
-0.12298566848039627,
0.018717538565397263,
0.006286133546382189,
-0.11018836498260498,
-0.03562384471297264,
-0.03705140948295593,
-0.05089316889643669,
-0.032666660845279694,
0.1327630877494812,
0.022651463747024536,
0.010037020780146122,
-0.005573086440563202,
0.10131679475307465,
0.035189077258110046,
0.0034852323587983847,
0.02730105072259903,
-0.05083545297384262,
0.03323967754840851,
-0.044979095458984375,
0.09806713461875916,
0.09106766432523727,
-0.028752246871590614,
0.0945800319314003,
0.13296785950660706,
0.0061467853374779224,
-0.059247761964797974,
-0.09029071033000946,
-0.005557273980230093,
-0.025198932737112045,
-0.03789623826742172,
-0.07522541284561157,
-0.08176995813846588,
0.06209201365709305,
0.04889515042304993,
-0.07742760330438614,
0.03534708917140961,
-0.06802807748317719,
-0.05819077789783478,
0.11488523334264755,
0.1370917111635208,
0.16688239574432373,
-0.0789678692817688,
0.03504548594355583,
-0.015266964212059975,
-0.021080389618873596,
0.0298110730946064,
0.097935751080513,
0.043939635157585144,
0.015425147488713264,
0.05543963983654976,
-0.00025826715864241123,
-0.16025932133197784,
-0.017884396016597748,
0.02569904364645481,
-0.12419571727514267,
-0.18786364793777466,
0.00536001892760396,
0.07122281193733215,
-0.1181166023015976,
-0.082621268928051,
0.1459631770849228,
-0.06228185072541237,
-0.06968487799167633,
-0.03565364331007004,
0.06677500158548355,
0.08385327458381653,
0.0804404690861702,
-0.0030789305455982685,
-0.04232468456029892,
-0.06873565912246704,
0.12887097895145416,
0.129071444272995,
-0.05559881031513214,
0.004469600506126881,
0.06196723133325577,
-0.09761014580726624,
-0.017652496695518494,
-0.07085452973842621,
0.07986007630825043,
-0.09518279880285263,
-0.048345476388931274,
0.03516027331352234,
-0.058493807911872864,
0.047562241554260254,
0.19755050539970398,
-0.03967907279729843,
-0.01644454523921013,
-0.017222687602043152,
-0.010730108246207237,
-0.09492789208889008,
0.06691862642765045,
-0.08225513994693756,
0.06157761067152023,
-0.005548554938286543,
0.08087248355150223,
0.032446905970573425,
-0.0357397124171257,
-0.030051447451114655,
-0.06825295090675354,
-0.040125392377376556,
-0.03623053804039955,
-0.08979237079620361,
-0.006579010747373104,
-0.046404413878917694,
-0.07661603391170502,
0.024807341396808624,
0.028376556932926178,
0.027941405773162842,
0.02333184890449047,
-0.011040539480745792,
-0.06538629531860352,
-0.05979723855853081,
0.09290378540754318,
-0.1571292281150818,
0.03457926958799362,
0.08653062582015991,
-0.07237344980239868,
0.08966808021068573,
0.03050336241722107,
0.07629293948411942,
-0.031047234311699867,
0.08682555705308914,
-0.017902277410030365,
-0.04124719649553299,
0.09740127623081207,
-0.00970284640789032,
-0.13471093773841858,
-0.008250370621681213,
-0.028333835303783417,
-0.14069518446922302,
-0.017407046630978584,
0.04344906657934189,
-0.05070403218269348,
0.006401121616363525,
0.06992943584918976,
0.03042948991060257,
-0.039441049098968506,
0.02695251628756523,
0.10999053716659546,
-0.0154475849121809,
0.07093559950590134,
-0.02988836169242859,
0.029247445985674858,
-0.1580517590045929,
-0.024078261107206345,
0.001577793387696147,
0.030457347631454468,
-0.018856598064303398,
-0.05827171728014946,
0.046446919441223145,
0.04718555510044098,
0.19375720620155334,
0.027657831087708473,
0.09311176091432571,
0.01098061352968216,
-0.05215401202440262,
-0.01187965553253889,
0.0027061705477535725,
0.04091552272439003,
0.09907020628452301,
-0.025152741000056267,
0.026153119280934334,
-0.0030253189615905285,
-0.060988135635852814,
-0.03702419251203537,
-0.01833326183259487,
0.01889619044959545,
0.12874072790145874,
-0.06339117139577866,
0.005924062803387642,
-0.02882041223347187,
-0.05775367096066475,
0.004353818949311972,
-0.018385134637355804,
0.003812011331319809,
-0.00735698314383626,
0.1382600963115692,
0.035366982221603394,
-0.12821809947490692,
0.16777533292770386,
-0.011427806690335274,
-0.05010339617729187,
-0.07769612222909927,
-0.1536966860294342,
-0.04494483768939972,
-0.0382576659321785,
-0.00758164469152689,
-0.14694911241531372,
0.025452380999922752,
0.152153879404068,
-0.017203103750944138,
-0.012809069827198982,
0.0057948557659983635,
-0.10290820896625519,
-0.1644412875175476,
0.010107362642884254,
0.044301606714725494,
0.049468718469142914,
-0.047109536826610565,
0.0798170194029808,
0.043427012860774994,
0.0658397227525711,
0.050030417740345,
0.06626547873020172,
0.11816535890102386,
-0.01583992876112461,
-0.021196018904447556,
-0.01810024119913578,
-0.0047297696582973,
-0.057024359703063965,
-0.029013697057962418,
0.13414441049098969,
0.04826921597123146,
0.029293011873960495,
0.0027151750400662422,
0.31061697006225586,
-0.04727781563997269,
-0.017880911007523537,
-0.09975673258304596,
0.24604973196983337,
0.06904648244380951,
0.061678461730480194,
0.01742108352482319,
-0.0796351507306099,
-0.01785106211900711,
0.1940334439277649,
-0.03901670128107071,
-0.010308751836419106,
0.0038644627202302217,
-0.01250983402132988,
0.010589761659502983,
0.05745153874158859,
0.045966751873493195,
0.03238805755972862,
0.17104360461235046,
-0.0903128832578659,
0.13837242126464844,
-0.061429329216480255,
0.018846873193979263,
-0.09426000714302063,
0.13641102612018585,
-0.08204039931297302,
0.014797335490584373,
-0.038325317203998566,
0.04486700892448425,
-0.012459568679332733,
-0.26624271273612976,
0.010427931323647499,
-0.050739601254463196,
-0.06692800670862198,
0.022182591259479523,
0.001386547926813364,
-0.016342846676707268,
-0.003273446811363101,
0.09379615634679794,
0.07814252376556396,
0.2623385488986969,
0.06187966838479042,
-0.013552546501159668,
0.01925036683678627,
0.046031299978494644,
-0.07766826450824738,
0.1698407232761383,
0.01158224232494831,
0.004491983447223902,
0.036291927099227905,
-0.05160602927207947,
-0.13909533619880676,
0.08550453186035156,
0.0442168191075325,
-0.02016105130314827,
0.01336743589490652,
0.15764692425727844,
0.03137124329805374,
0.19532187283039093,
0.060700953006744385,
-0.006959350313991308,
0.0877094715833664,
0.02735014446079731,
-0.04875956475734711,
-0.01390833593904972,
0.0586174838244915,
-0.09777700901031494,
0.09242677688598633,
0.10537789016962051,
-0.05395539104938507,
0.02350202202796936,
-0.08035305142402649,
0.005768085364252329,
-0.014955663122236729,
0.07118266820907593,
0.00983533076941967,
0.0033639161847531796,
-0.019128110259771347,
-0.08082140982151031,
0.025024354457855225,
-0.26988399028778076,
-0.0382113978266716,
0.07463887333869934,
-0.011230136267840862,
-0.002007109811529517,
0.09312419593334198,
0.030261240899562836,
0.026292182505130768,
-0.024744225665926933,
-0.021597329527139664,
0.053426846861839294,
0.03971724584698677,
-0.07576552033424377,
0.018197868019342422
] |
null | null |
transformers
|
# HIV_V3_coreceptor model
## Table of Contents
- [Summary](#model-summary)
- [Model Description](#model-description)
- [Intended Uses & Limitations](#intended-uses-&-limitations)
- [How to Use](#how-to-use)
- [Training Data](#training-data)
- [Training Procedure](#training-procedure)
- [Preprocessing](#preprocessing)
- [Training](#training)
- [Evaluation Results](#evaluation-results)
- [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info)
## Summary
The HIV-BERT-Coreceptor model was trained as a refinement of the [HIV-BERT model](https://huggingface.co/damlab/HIV_BERT) and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the [ProtBert-BFD model](https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the [Los Alamos HIV Sequence Database](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html), allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.
## Model Description
The HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors.
## Intended Uses & Limitations
This tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool.
This tool was trained using the [Los Alamos HIV sequence dataset](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.
## How to use
*Need to add*
## Training Data
This model was trained using the [damlab/HIV_V3_coreceptor dataset](https://huggingface.co/datasets/damlab/HIV_V3_coreceptor) using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the [Los Alamos HIV Sequence database](https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html).
## Training Procedure
### Preprocessing
As with the [rostlab/Prot-bert-bfd model](https://huggingface.co/Rostlab/prot_bert_bfd), the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
The [damlab/HIV-BERT model](https://huggingface.co/damlab/HIV_BERT) was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.
## Evaluation Results
*Need to add*
## BibTeX Entry and Citation Info
[More Information Needed]
|
{"license": "mit", "widget": [{"text": "C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C"}, {"text": "C T R P N N N T R K S I H I G P G R A F Y T T G Q I I G D I R Q A Y C"}, {"text": "C T R P N N N T R R S I R I G P G Q A F Y A T G D I I G D I R Q A H C"}, {"text": "C G R P N N H R I K G L R I G P G R A F F A M G A I G G G E I R Q A H C"}]}
|
text-classification
|
damlab/HIV_V3_Coreceptor
|
[
"transformers",
"pytorch",
"bert",
"text-classification",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us
|
# HIV_V3_coreceptor model
## Table of Contents
- Summary
- Model Description
- Intended Uses & Limitations
- How to Use
- Training Data
- Training Procedure
- Preprocessing
- Training
- Evaluation Results
- BibTeX Entry and Citation Info
## Summary
The HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.
## Model Description
The HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors.
## Intended Uses & Limitations
This tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool.
This tool was trained using the Los Alamos HIV sequence dataset. Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.
## How to use
*Need to add*
## Training Data
This model was trained using the damlab/HIV_V3_coreceptor dataset using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.
## Training Procedure
### Preprocessing
As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.
## Evaluation Results
*Need to add*
## BibTeX Entry and Citation Info
|
[
"# HIV_V3_coreceptor model",
"## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.",
"## Model Description\r\n\r\nThe HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors.",
"## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset. Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.",
"## How to use\r\n\r\n*Need to add*",
"## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_coreceptor dataset using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.",
"## Evaluation Results\r\n\r\n*Need to add*",
"## BibTeX Entry and Citation Info"
] |
[
"TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n",
"# HIV_V3_coreceptor model",
"## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.",
"## Model Description\r\n\r\nThe HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors.",
"## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of HIV tropism from the Env-V3 loop. It can recognize both R5, X4, and dual tropic viruses natively. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset. Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.",
"## How to use\r\n\r\n*Need to add*",
"## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_coreceptor dataset using the 0th fold. The dataset consists of 2935 V3 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can bind to CCR5, CXCR4, neither, or both) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.",
"## Evaluation Results\r\n\r\n*Need to add*",
"## BibTeX Entry and Citation Info"
] |
[
41,
10,
52,
122,
178,
166,
10,
67,
4,
80,
157,
10,
11
] |
[
"passage: TAGS\n#transformers #pytorch #bert #text-classification #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# HIV_V3_coreceptor model## Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT-Coreceptor model was trained as a refinement of the HIV-BERT model and serves to better predict HIV V3 coreceptor tropism. HIV-BERT is a model refined from the ProtBert-BFD model to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database, allowing even more precise prediction of V3 coreceptor tropism than the HIV-BERT model can provide.## Model Description\r\n\r\nThe HIV-BERT-Coreceptor model is intended to predict the Co-receptor tropism of HIV from a segment of the envelope protein. These envelope proteins encapsulate the virus and interact with the host cell through the human CD4 receptor. HIV then requires the interaction of one, of two, co-receptors: CCR5 or CXCR4. The availability of these co-receptors on different cell types allows the virus to invade different areas of the body and evade antiretroviral therapy. The 3rd variable loop of the envelope protein, the V3 loop, is responsible for this interaction. Given a V3 loop sequence, the HIV-BERT-Coreceptor model will predict the likelihood of binding to each of these co-receptors."
] |
[
0.06304185092449188,
0.038277871906757355,
-0.005255136173218489,
0.011456751264631748,
0.05832764133810997,
0.01965138129889965,
0.06130440533161163,
0.059777140617370605,
0.07188352942466736,
0.08445162326097488,
0.02141999453306198,
-0.13655461370944977,
0.11485499143600464,
0.11875556409358978,
0.11328098922967911,
-0.17541760206222534,
0.017055882140994072,
-0.0301374364644289,
-0.07306236773729324,
0.047906726598739624,
0.05424671620130539,
-0.0775853618979454,
0.016127677634358406,
0.012562540359795094,
0.14036977291107178,
0.04261400178074837,
-0.06231396272778511,
-0.025280790403485298,
0.0826248973608017,
0.03791312128305435,
0.07303430885076523,
0.11607526242733002,
0.13070784509181976,
-0.2738986015319824,
-0.013374436646699905,
0.04605238884687424,
-0.03385396674275398,
0.0027804195415228605,
0.06675755232572556,
-0.045513514429330826,
-0.024566659703850746,
-0.06545241922140121,
0.11289360374212265,
0.02610521763563156,
-0.078685462474823,
-0.10580453276634216,
-0.09225580096244812,
0.10832826793193817,
0.07367023825645447,
0.08334372937679291,
-0.003148880787193775,
0.06594443321228027,
0.060694094747304916,
-0.011583025567233562,
0.11060794442892075,
-0.2684257924556732,
0.0417213998734951,
0.13195350766181946,
0.10475316643714905,
0.0325782373547554,
-0.0344257615506649,
-0.005889280699193478,
-0.005764179863035679,
0.04434531554579735,
0.23558297753334045,
-0.03120933659374714,
0.1304090917110443,
-0.06521786749362946,
-0.18202099204063416,
-0.13945253193378448,
0.09942317008972168,
0.0029369089752435684,
-0.0411800742149353,
-0.018591908738017082,
0.011966670863330364,
-0.0017762277275323868,
0.02636871114373207,
-0.12912605702877045,
0.014965733513236046,
0.037421129643917084,
-0.03313160315155983,
-0.1453290730714798,
-0.06925689429044724,
-0.034851182252168655,
-0.06483589112758636,
-0.014762414619326591,
0.032990697771310806,
-0.0030665576923638582,
-0.04735502973198891,
0.1114123985171318,
-0.13629299402236938,
-0.0730876624584198,
-0.03158801421523094,
-0.06353745609521866,
-0.11720213294029236,
-0.0205636415630579,
0.03513287380337715,
-0.12049135565757751,
0.015110020525753498,
0.07598813623189926,
-0.06479258835315704,
-0.0058450279757380486,
-0.044877681881189346,
-0.011358481831848621,
0.0758550763130188,
0.07122500985860825,
-0.10512079298496246,
-0.019925301894545555,
-0.025464538484811783,
-0.05369582399725914,
0.013921063393354416,
-0.008144865743815899,
-0.0165707990527153,
0.0012589811813086271,
0.001352486084215343,
0.09475207328796387,
-0.020759284496307373,
0.010765650309622288,
0.056437619030475616,
0.013524743728339672,
0.02619141899049282,
-0.050621725618839264,
-0.004649277310818434,
-0.05001932382583618,
0.04791996628046036,
0.07179481536149979,
0.07484856247901917,
-0.011866617947816849,
-0.048270974308252335,
0.02597421035170555,
-0.014187011867761612,
-0.08982949703931808,
-0.02509119361639023,
-0.1434115767478943,
0.04200616106390953,
0.001062069903127849,
-0.07476682215929031,
-0.0981450080871582,
0.039714351296424866,
-0.05778617784380913,
-0.04408406466245651,
-0.030377229675650597,
-0.010950113646686077,
0.003428337862715125,
0.05056367442011833,
0.012823468074202538,
-0.030861124396324158,
-0.03982600197196007,
-0.005648248828947544,
-0.04418274387717247,
-0.11474456638097763,
0.11387906223535538,
0.14653600752353668,
0.016151832416653633,
-0.019368644803762436,
0.08030670136213303,
-0.15906256437301636,
-0.02269372157752514,
-0.141703262925148,
-0.0316464826464653,
-0.16859391331672668,
-0.0683484673500061,
0.03348802402615547,
-0.09503242373466492,
0.04929175600409508,
0.04025646671652794,
-0.05260516703128815,
-0.021593736484646797,
0.32618772983551025,
-0.0729360282421112,
0.004837557673454285,
-0.019868839532136917,
0.019788723438978195,
0.08223002403974533,
0.03054015524685383,
-0.0018649415578693151,
0.14938129484653473,
-0.09599261730909348,
-0.07157807797193527,
-0.02511782944202423,
-0.05676668509840965,
0.1672838032245636,
0.07907973229885101,
-0.08840636163949966,
0.034039195626974106,
-0.006876127794384956,
-0.02262250706553459,
-0.02177908830344677,
-0.027521761134266853,
0.021832099184393883,
0.004905417561531067,
0.016735803335905075,
0.08150217682123184,
-0.10590943694114685,
-0.07635287940502167,
0.0004211771010886878,
-0.15232639014720917,
-0.0482960119843483,
0.10598732531070709,
-0.11164949089288712,
0.07850310951471329,
-0.09995913505554199,
-0.02309507131576538,
-0.037483882158994675,
0.005992199759930372,
-0.07179397344589233,
-0.05110282078385353,
0.04660904407501221,
-0.08148818463087082,
0.06149788945913315,
0.02868681214749813,
0.03719834238290787,
0.04128427430987358,
-0.0259479321539402,
0.024950483813881874,
-0.021803267300128937,
0.011342105455696583,
-0.020574655383825302,
-0.09605486690998077,
-0.10301151126623154,
-0.057718340307474136,
0.22426657378673553,
-0.02834719605743885,
-0.021654529497027397,
-0.014526769518852234,
0.09868931025266647,
0.0021537637803703547,
0.04237409308552742,
0.09961269050836563,
0.03243880718946457,
-0.005038838367909193,
-0.010977767407894135,
0.001113676349632442,
-0.038338176906108856,
-0.018898451700806618,
0.03995632007718086,
-0.07929535955190659,
-0.28655126690864563,
0.009290214627981186,
0.07118827849626541,
-0.08192501217126846,
-0.0379057377576828,
0.0551086887717247,
0.007499190047383308,
-0.06613204628229141,
-0.0313890241086483,
0.20230650901794434,
0.08687224984169006,
0.06374077498912811,
0.028014304116368294,
-0.03572671860456467,
0.05107852816581726,
-0.004121359437704086,
-0.048383209854364395,
-0.0025889407843351364,
0.09656529128551483,
-0.07325196266174316,
-0.014419058337807655,
0.12971480190753937,
0.059261586517095566,
-0.007417609449476004,
0.033718500286340714,
-0.09270873665809631,
-0.023544564843177795,
-0.06898431479930878,
0.032838787883520126,
-0.047804977744817734,
-0.0572294183075428,
0.024990664795041084,
0.05409792438149452,
-0.0029779369942843914,
-0.0548243373632431,
-0.015034960582852364,
0.056535571813583374,
0.09979942440986633,
-0.04659482464194298,
-0.06763483583927155,
-0.07434526830911636,
0.0020132220815867186,
0.03781338036060333,
0.04328995570540428,
-0.029249366372823715,
-0.04640140011906624,
0.004631000570952892,
-0.12200570106506348,
0.05236726254224777,
-0.09863657504320145,
-0.2765951454639435,
-0.1120847538113594,
-0.10583920031785965,
0.012647552415728569,
0.009440762922167778,
0.048752039670944214,
-0.04113860800862312,
-0.05399787798523903,
-0.10969893634319305,
0.12205632776021957,
-0.05777646601200104,
-0.0347459577023983,
0.0496504046022892,
0.08351383358240128,
0.026896104216575623,
-0.05024004355072975,
0.014179032295942307,
0.005243465304374695,
-0.11105494946241379,
0.09598727524280548,
-0.02685021422803402,
0.08549114316701889,
0.14944300055503845,
0.045886971056461334,
-0.021046141162514687,
-0.08322375267744064,
0.04890952259302139,
-0.029875174164772034,
0.10896429419517517,
0.1566115915775299,
-0.026005355641245842,
0.03198747709393501,
0.08636834472417831,
-0.04137278348207474,
0.05372130125761032,
0.05555278807878494,
0.01089609507471323,
-0.045836761593818665,
-0.20544400811195374,
-0.04724711924791336,
0.04146306589245796,
-0.0031864899210631847,
0.058058883994817734,
-0.0026699507143348455,
0.03713122755289078,
0.06226210296154022,
-0.04082668945193291,
-0.04384096711874008,
0.04495798051357269,
0.07372234016656876,
0.1654563695192337,
-0.059835284948349,
0.0831209123134613,
0.0027721773367375135,
-0.0257403701543808,
0.08603367954492569,
0.05738729238510132,
0.1873473823070526,
0.06160099804401398,
0.0646575540304184,
0.06981179863214493,
0.0785248801112175,
0.014019660651683807,
0.1237616315484047,
0.05058739706873894,
0.026031184941530228,
-0.016380442306399345,
-0.05235745385289192,
0.045278631150722504,
-0.016949832439422607,
0.06565877050161362,
-0.08839518576860428,
-0.030551359057426453,
-0.04619378224015236,
-0.06370609253644943,
0.09487385302782059,
0.1242203339934349,
-0.046852875500917435,
-0.01088802795857191,
0.049867983907461166,
-0.06811561435461044,
-0.08465846627950668,
0.004262722097337246,
0.056823037564754486,
-0.18137530982494354,
0.08067908138036728,
0.032502759248018265,
0.09726283699274063,
-0.0886302962899208,
-0.016787763684988022,
-0.07661639153957367,
-0.04813670739531517,
-0.05851779133081436,
0.10212942212820053,
-0.05801958218216896,
0.09469825774431229,
-0.010536075569689274,
-0.03665851801633835,
-0.04785652086138725,
0.02855571359395981,
-0.053140223026275635,
0.1541147232055664,
0.14416712522506714,
0.005398011766374111,
0.08278685063123703,
-0.05517059937119484,
-0.16197219491004944,
-0.011432445608079433,
0.05011100694537163,
-0.15613631904125214,
0.15725071728229523,
-0.01887868344783783,
-0.01784617453813553,
-0.04338131472468376,
0.12509246170520782,
-0.12660053372383118,
-0.10954976081848145,
0.0802057534456253,
-0.07003729045391083,
0.15521842241287231,
-0.011015117168426514,
-0.05835800990462303,
0.020828772336244583,
0.04451818764209747,
-0.1413705199956894,
-0.03828517347574234,
-0.12314609438180923,
0.004024547524750233,
0.1496194303035736,
-0.0934191644191742,
0.019255924969911575,
0.003828669199720025,
0.17677581310272217,
-0.08922654390335083,
-0.11051526665687561,
-0.012163834646344185,
-0.06383049488067627,
-0.11407368630170822,
-0.05333055555820465,
0.1633380651473999,
-0.0009648537961766124,
0.1245366707444191,
0.052181169390678406,
0.03906460106372833,
0.04601620137691498,
-0.040530189871788025,
0.08610653132200241,
0.17122265696525574,
0.021884877234697342,
-0.0367249995470047,
-0.04043189808726311,
-0.016488738358020782,
-0.10849010944366455,
-0.07525812834501266,
0.25780677795410156,
0.20693102478981018,
-0.07523921132087708,
0.1333887130022049,
0.05168937146663666,
-0.13050861656665802,
-0.18076229095458984,
-0.12448012083768845,
0.10860200971364975,
-0.019582344219088554,
0.07587889581918716,
-0.2439144402742386,
-0.05081578716635704,
0.014813202433288097,
-0.0012409311020746827,
-0.10266922414302826,
-0.16276507079601288,
-0.08745688199996948,
0.08941500633955002,
-0.05698670074343681,
-0.03694053366780281,
-0.00953042134642601,
-0.020108159631490707,
-0.07911936938762665,
-0.20620031654834747,
0.08701407164335251,
-0.040902767330408096,
0.014856589958071709,
0.013640847988426685,
0.07488173246383667,
0.026014532893896103,
0.02450283244252205,
0.058709513396024704,
-0.037056658416986465,
-0.04650451987981796,
-0.06184700131416321,
0.06162923946976662,
0.04463960602879524,
-0.04459527134895325,
0.05470937490463257,
0.13293029367923737,
-0.031214579939842224,
-0.09644337743520737,
-0.0720161497592926,
-0.0038323646876960993,
-0.03335823863744736,
-0.03758377581834793,
-0.07516448944807053,
-0.0681750625371933,
0.008401933126151562,
0.09172461926937103,
-0.0755753219127655,
0.02262652851641178,
-0.0704009011387825,
-0.042772326618433,
0.2065475583076477,
0.10661358386278152,
0.05373998358845711,
-0.12913952767848969,
0.05661953613162041,
-0.04890184476971626,
-0.03763178735971451,
-0.07610036432743073,
0.09916724264621735,
0.04724298417568207,
-0.00048201228491961956,
0.04658383131027222,
0.029229160398244858,
-0.15197037160396576,
-0.0008786373073235154,
0.05023971199989319,
-0.09994117170572281,
-0.11050596833229065,
-0.00620501721277833,
0.04906553775072098,
-0.13080033659934998,
-0.055501751601696014,
0.18776562809944153,
-0.0773831233382225,
-0.046732157468795776,
-0.04193071648478508,
0.06726586073637009,
-0.0018346133874729276,
0.10329953581094742,
0.011979230679571629,
0.00871298648416996,
-0.038688432425260544,
0.06962879747152328,
0.14304883778095245,
-0.026209479197859764,
0.02003627084195614,
0.042441099882125854,
-0.10191446542739868,
-0.08200469613075256,
-0.02938578836619854,
0.048647332936525345,
-0.13232110440731049,
-0.07231497764587402,
0.11401233822107315,
-0.08589986711740494,
0.03567283973097801,
0.2359609454870224,
-0.0576138012111187,
-0.030248748138546944,
-0.0009561756742186844,
-0.04373728111386299,
-0.11285912245512009,
0.04005805775523186,
-0.038870103657245636,
0.04827633500099182,
-0.017275074496865273,
0.16283348202705383,
0.04089611396193504,
0.08429885655641556,
-0.05325634777545929,
-0.07728512585163116,
-0.049083709716796875,
-0.025054845958948135,
-0.14265528321266174,
-0.05127342790365219,
-0.05496468394994736,
-0.10414478182792664,
0.03820716589689255,
0.02523139864206314,
0.03242756798863411,
0.022152388468384743,
-0.027517085894942284,
-0.022041451185941696,
-0.07923436164855957,
0.09213250130414963,
-0.10168766230344772,
-0.0007751139928586781,
0.08668641000986099,
-0.04746590927243233,
0.10150681436061859,
-0.027933571487665176,
0.08749864250421524,
-0.014271017163991928,
0.05583091825246811,
-0.016906317323446274,
-0.03234570473432541,
0.056694645434617996,
0.03376379609107971,
-0.18050982058048248,
-0.03237343579530716,
-0.08500757068395615,
-0.17610931396484375,
0.005044750869274139,
0.052656397223472595,
0.020985621958971024,
-0.0030218723695725203,
-0.03244859352707863,
0.018944961950182915,
0.0019031489500775933,
0.007913932204246521,
0.04810767620801926,
0.01788301020860672,
0.04541456699371338,
-0.022417431697249413,
0.10082297027111053,
-0.15900923311710358,
-0.030022811144590378,
-0.03335358574986458,
0.041458435356616974,
0.06307192891836166,
-0.07363659888505936,
0.045667655766010284,
0.045111119747161865,
0.05913393944501877,
0.026986001059412956,
0.175282284617424,
-0.024613594636321068,
-0.07280116528272629,
0.05333014577627182,
-0.018102627247571945,
-0.03421558812260628,
0.09990638494491577,
-0.044693849980831146,
-0.020567605271935463,
0.001525377156212926,
-0.008814873173832893,
-0.0517512746155262,
0.02796175889670849,
-0.10379165410995483,
0.11819389462471008,
-0.03892535716295242,
0.041308771818876266,
-0.07702023535966873,
-0.0940595269203186,
-0.158336341381073,
0.08430630713701248,
0.045421820133924484,
-0.020004330202937126,
0.12952034175395966,
0.05257087200880051,
-0.13810716569423676,
0.11156881600618362,
-0.006082148756831884,
-0.050174612551927567,
-0.07194142788648605,
-0.1375245451927185,
-0.047754235565662384,
-0.16307185590267181,
-0.0109166344627738,
-0.140285462141037,
0.004929568152874708,
0.14871491491794586,
0.0275628250092268,
0.041758839040994644,
-0.046950992196798325,
-0.1348588466644287,
-0.12898586690425873,
-0.015132687985897064,
0.047615841031074524,
0.0745978131890297,
-0.02981341816484928,
0.11998172104358673,
0.0399327352643013,
0.08236002922058105,
0.021400844678282738,
0.060576051473617554,
0.10936515778303146,
0.001344833755865693,
-0.02010592445731163,
-0.0479782335460186,
-0.03850357607007027,
-0.014355931431055069,
-0.0148639976978302,
0.20363420248031616,
0.05810800939798355,
0.027991382405161858,
-0.010875318199396133,
0.2232297956943512,
-0.04765631631016731,
-0.041434139013290405,
-0.11074668914079666,
0.2421865165233612,
0.06861910223960876,
0.07472074031829834,
-0.001283062039874494,
-0.08484840393066406,
-0.09068533033132553,
0.20498539507389069,
0.02784121036529541,
0.06517013162374496,
-0.001685642870143056,
0.018674984574317932,
0.019500769674777985,
0.07384303212165833,
0.025525497272610664,
0.07930118590593338,
0.14074470102787018,
-0.14401187002658844,
0.1033245176076889,
-0.06155822053551674,
0.0347275547683239,
-0.10232886672019958,
0.013716918416321278,
-0.044855717569589615,
-0.004370691254734993,
0.0756082609295845,
0.1079501286149025,
0.025035370141267776,
-0.19509050250053406,
-0.01826980896294117,
-0.06230061128735542,
-0.032054975628852844,
0.028220098465681076,
-0.010632585734128952,
0.0244760625064373,
0.0378497838973999,
0.046843092888593674,
0.028017748147249222,
0.27924153208732605,
0.06676096469163895,
0.012742509134113789,
-0.02761792205274105,
0.10039977729320526,
-0.13041366636753082,
0.17278729379177094,
0.01243556383997202,
0.06207701936364174,
0.01920948550105095,
-0.00760053563863039,
-0.14095449447631836,
0.06774874776601791,
0.0627657100558281,
0.0063465009443461895,
0.037174418568611145,
0.12625809013843536,
-0.018857285380363464,
0.08823982626199722,
0.04358869418501854,
-0.09473904967308044,
0.0791286900639534,
0.03211073577404022,
0.051379963755607605,
0.07334025949239731,
0.05727555975317955,
-0.09651822596788406,
0.08398137241601944,
0.11305423080921173,
-0.08340760320425034,
-0.03028222918510437,
-0.07216130942106247,
0.0020380974747240543,
-0.008609727956354618,
0.1337393969297409,
0.0031060189940035343,
-0.027202390134334564,
0.0197328832000494,
-0.08333741128444672,
0.03189343214035034,
-0.23835958540439606,
-0.05579927936196327,
0.03661378100514412,
0.014116104692220688,
-0.000055254087783396244,
0.11352834105491638,
0.03810524195432663,
0.027585996314883232,
-0.03415660560131073,
0.01794571615755558,
0.03528251126408577,
0.05738217011094093,
-0.14821724593639374,
0.013086756691336632
] |
null | null |
transformers
|
# Model Card for [HIV_V3_bodysite]
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Summary](#model-summary)
- [Model Description](#model-description)
- [Intended Uses & Limitations](#intended-uses-&-limitations)
- [How to Use](#how-to-use)
- [Training Data](#training-data)
- [Training Procedure](#training-procedure)
- [Preprocessing](#preprocessing)
- [Training](#training)
- [Evaluation Results](#evaluation-results)
- [BibTeX Entry and Citation Info](#bibtex-entry-and-citation-info)
## Summary
The HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (https://huggingface.co/Rostlab/prot_bert_bfd) to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html), allowing even more precise prediction of body site location than the HIV-BERT model can provide.
## Model Description
The HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from.
## Intended Uses & Limitations
This tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool.
This tool was trained using the Los Alamos HIV sequence dataset (https://www.hiv.lanl.gov/content/sequence/HIV/mainpage.html). Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.
## How to use
This model is able to predict the likely bodysite from a V3 sequence.
This may be use for surveillance of cells that are emerging from latent reservoirs.
Remember, a sequence can come from multiple sites, they are not mutually exclusive.
```python
from transformers import pipeline
predictor = pipeline("text-classification", model="damlab/HIV_V3_bodysite")
predictor(f"C T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C")
[
[
{
"label": "periphery-tcell",
"score": 0.29097115993499756
},
{
"label": "periphery-monocyte",
"score": 0.014322502538561821
},
{
"label": "CNS",
"score": 0.06870711594820023
},
{
"label": "breast-milk",
"score": 0.002785981632769108
},
{
"label": "female-genitals",
"score": 0.024997007101774216
},
{
"label": "male-genitals",
"score": 0.01040483545511961
},
{
"label": "gastric",
"score": 0.06872137635946274
},
{
"label": "lung",
"score": 0.04432062804698944
},
{
"label": "organ",
"score": 0.47476938366889954
}
]
]
```
## Training Data
This model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.
## Training Procedure
### Preprocessing
As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.
## Evaluation Results
*Need to add*
## BibTeX Entry and Citation Info
[More Information Needed]
|
{"datasets": ["damlab/HIV_V3_bodysite"], "metrics": ["accuracy"], "licence": "mit", "widget": [{"text": "T R P N N N T R K S I R I Q R G P G R A F V T I G K I G N M R Q A H C", "example_title": "V3 Macrophage"}, {"text": "C T R P N N N T R K S I H I G P G R A F Y T T G Q I I G D I R Q A Y C", "example_title": "V3 T-cell"}]}
|
text-classification
|
damlab/HIV_V3_bodysite
|
[
"transformers",
"pytorch",
"bert",
"text-classification",
"dataset:damlab/HIV_V3_bodysite",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #bert #text-classification #dataset-damlab/HIV_V3_bodysite #autotrain_compatible #endpoints_compatible #region-us
|
# Model Card for [HIV_V3_bodysite]
## Table of Contents
- Table of Contents
- Summary
- Model Description
- Intended Uses & Limitations
- How to Use
- Training Data
- Training Procedure
- Preprocessing
- Training
- Evaluation Results
- BibTeX Entry and Citation Info
## Summary
The HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.
## Model Description
The HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from.
## Intended Uses & Limitations
This tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool.
This tool was trained using the Los Alamos HIV sequence dataset (URL Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.
## How to use
This model is able to predict the likely bodysite from a V3 sequence.
This may be use for surveillance of cells that are emerging from latent reservoirs.
Remember, a sequence can come from multiple sites, they are not mutually exclusive.
## Training Data
This model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.
## Training Procedure
### Preprocessing
As with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.
### Training
The damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.
## Evaluation Results
*Need to add*
## BibTeX Entry and Citation Info
|
[
"# Model Card for [HIV_V3_bodysite]",
"## Table of Contents\r\n- Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.",
"## Model Description\r\n\r\nThe HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from.",
"## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset (URL Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.",
"## How to use\r\n\r\nThis model is able to predict the likely bodysite from a V3 sequence.\r\nThis may be use for surveillance of cells that are emerging from latent reservoirs.\r\nRemember, a sequence can come from multiple sites, they are not mutually exclusive.",
"## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.",
"## Evaluation Results\r\n\r\n*Need to add*",
"## BibTeX Entry and Citation Info"
] |
[
"TAGS\n#transformers #pytorch #bert #text-classification #dataset-damlab/HIV_V3_bodysite #autotrain_compatible #endpoints_compatible #region-us \n",
"# Model Card for [HIV_V3_bodysite]",
"## Table of Contents\r\n- Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info",
"## Summary\r\n\r\nThe HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.",
"## Model Description\r\n\r\nThe HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from.",
"## Intended Uses & Limitations\r\n\r\nThis tool can be used as a predictor of which body site an HIV sample was derived from based on its genomic sequence. It should not be considered a clinical diagnostic tool. \r\n \r\nThis tool was trained using the Los Alamos HIV sequence dataset (URL Due to the sampling nature of this database, it is predominantly composed of subtype B sequences from North America and Europe with only minor contributions of Subtype C, A, and D. Currently, there was no effort made to balance the performance across these classes. As such, one should consider refinement with additional sequences to perform well on non-B sequences.",
"## How to use\r\n\r\nThis model is able to predict the likely bodysite from a V3 sequence.\r\nThis may be use for surveillance of cells that are emerging from latent reservoirs.\r\nRemember, a sequence can come from multiple sites, they are not mutually exclusive.",
"## Training Data\r\n\r\nThis model was trained using the damlab/HIV_V3_bodysite dataset using the 0th fold. The dataset consists of 5510 sequences (approximately 35 tokens each) extracted from the Los Alamos HIV Sequence database.",
"## Training Procedure",
"### Preprocessing\r\n\r\nAs with the rostlab/Prot-bert-bfd model, the rare amino acids U, Z, O, and B were converted to X and spaces were added between each amino acid. All strings were concatenated and chunked into 256 token chunks for training. A random 20% of chunks were held for validation.",
"### Training\r\n\r\nThe damlab/HIV-BERT model was used as the initial weights for an AutoModelforClassificiation. The model was trained with a learning rate of 1E-5, 50K warm-up steps, and a cosine_with_restarts learning rate schedule and continued until 3 consecutive epochs did not improve the loss on the held-out dataset. As this is a multiple classification task (a protein can be found in multiple sites) the loss was calculated as the Binary Cross Entropy for each category. The BCE was weighted by the inverse of the class ratio to balance the weight across the class imbalance.",
"## Evaluation Results\r\n\r\n*Need to add*",
"## BibTeX Entry and Citation Info"
] |
[
51,
14,
57,
135,
134,
154,
64,
64,
4,
80,
146,
10,
11
] |
[
"passage: TAGS\n#transformers #pytorch #bert #text-classification #dataset-damlab/HIV_V3_bodysite #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for [HIV_V3_bodysite]## Table of Contents\r\n- Table of Contents\r\n- Summary\r\n- Model Description\r\n- Intended Uses & Limitations\r\n- How to Use\r\n- Training Data\r\n- Training Procedure\r\n - Preprocessing\r\n - Training\r\n- Evaluation Results\r\n- BibTeX Entry and Citation Info## Summary\r\n\r\nThe HIV-BERT-Bodysite-Identification model was trained as a refinement of the HIV-BERT model (insert link) and serves to better predict the location that an HIV V3 loop sample was derived from. HIV-BERT is a model refined from the ProtBert-BFD model (URL to better fulfill HIV-centric tasks. This model was then trained using HIV V3 sequences from the Los Alamos HIV Sequence Database (URL allowing even more precise prediction of body site location than the HIV-BERT model can provide.## Model Description\r\n\r\nThe HIV-BERT-Bodysite-Identification model is intended to predict the location as to where an HIV sequence was most likely derived from. Because HIV infects immune cells, it uses these as a means of rapidly spreading throughout the body. Thus, body site identification can help determine where exactly these HIV particles ultimately end up. This would be helpful when attempting to study HIV treatment strategies. When provided with an HIV genomic sequence, the HIV-BERT-Bodysite-Identification model can predict which tissue it was derived from."
] |
[
0.012887292541563511,
0.06151840463280678,
-0.004258778877556324,
0.006362018175423145,
0.1302095502614975,
-0.024671535938978195,
-0.004384350962936878,
0.11429488658905029,
0.011422245763242245,
0.03229138255119324,
-0.028772104531526566,
-0.06519068777561188,
0.07248427718877792,
0.055942993611097336,
0.11257477104663849,
-0.16200335323810577,
0.06368739902973175,
-0.017705049365758896,
-0.006136989686638117,
0.09631022065877914,
0.07635916024446487,
-0.06845550984144211,
0.022369256243109703,
0.010594753548502922,
0.02978496067225933,
0.10197686403989792,
-0.05623631924390793,
-0.00017819467757362872,
0.06551805138587952,
0.030090991407632828,
0.10532616823911667,
0.04674935340881348,
0.11340585350990295,
-0.3081340491771698,
-0.025692861527204514,
0.07937849313020706,
-0.01604086346924305,
0.0011388957500457764,
0.08173859119415283,
-0.058101508766412735,
0.06873352825641632,
-0.011925478465855122,
0.05739499628543854,
0.03818191960453987,
-0.10923519730567932,
-0.17877399921417236,
-0.0551244355738163,
0.07483562082052231,
0.01519174687564373,
0.131434366106987,
-0.04492952302098274,
0.016592277213931084,
0.005909188184887171,
0.020080238580703735,
0.11836729943752289,
-0.17855849862098694,
0.05196421965956688,
0.18489761650562286,
0.07832688838243484,
0.05058107525110245,
-0.056761402636766434,
-0.0059006838127970695,
-0.021662579849362373,
0.04394694045186043,
0.23263123631477356,
0.0018252730369567871,
0.06676644086837769,
-0.024941299110651016,
-0.18229180574417114,
-0.15425047278404236,
0.10358088463544846,
0.0033026011660695076,
-0.032439131289720535,
-0.10920988768339157,
0.010087123140692711,
-0.025654785335063934,
-0.01183370128273964,
-0.13304631412029266,
0.014272408559918404,
0.03525824099779129,
0.012636978179216385,
-0.06621433794498444,
-0.05822742357850075,
-0.026632189750671387,
-0.0086868517100811,
0.030498631298542023,
0.033026158809661865,
0.03190593793988228,
-0.08317345380783081,
0.12327539175748825,
-0.06723096966743469,
-0.06601660698652267,
-0.01748684048652649,
-0.06785068660974503,
-0.07212547957897186,
-0.02455214597284794,
0.04695058986544609,
-0.08990133553743362,
-0.03787358105182648,
0.09198123961687088,
0.01898838020861149,
-0.01421631034463644,
0.0482005774974823,
-0.03928138688206673,
0.06909767538309097,
0.10551634430885315,
-0.09267488121986389,
-0.006819534581154585,
-0.02529069036245346,
-0.014049369841814041,
-0.02892865613102913,
0.03596708923578262,
-0.02421029657125473,
0.004155692644417286,
0.09699009358882904,
0.09782527387142181,
-0.043391916900873184,
0.03700830787420273,
0.04033190384507179,
-0.04208134859800339,
0.016642510890960693,
-0.06719732284545898,
-0.03839370980858803,
-0.05494130775332451,
-0.026601940393447876,
0.04783088341355324,
0.1254914402961731,
-0.0036604467313736677,
-0.0830947756767273,
0.05232146754860878,
0.01617281697690487,
-0.0775766521692276,
-0.11344961076974869,
-0.1401938498020172,
-0.034438036382198334,
0.03904158994555473,
-0.06022157147526741,
-0.12735344469547272,
-0.1033669114112854,
-0.10631021857261658,
-0.023626644164323807,
-0.0006380007253028452,
0.021520348265767097,
-0.004056849516928196,
-0.006118765566498041,
-0.010859561152756214,
-0.01939123310148716,
0.005538190249353647,
-0.019273169338703156,
-0.01122094877064228,
-0.09847062826156616,
0.06950119882822037,
0.15504896640777588,
0.05901160463690758,
-0.10700967907905579,
0.05687057226896286,
-0.17613112926483154,
0.08629896491765976,
-0.12810148298740387,
0.009037183597683907,
-0.186017706990242,
-0.057714540511369705,
-0.028785046190023422,
-0.017580820247530937,
0.019781963899731636,
0.06242162734270096,
-0.07480897009372711,
-0.01670507900416851,
0.26055651903152466,
-0.10492757707834244,
-0.03641529753804207,
0.011913743801414967,
-0.04415324702858925,
0.17137327790260315,
0.08708783239126205,
0.06571190804243088,
0.08601225167512894,
-0.07179263979196548,
-0.06348806619644165,
-0.015124248340725899,
-0.009700821712613106,
0.17370325326919556,
0.01970459148287773,
-0.02283773384988308,
0.013318299315869808,
0.01953735388815403,
-0.05121449753642082,
0.03341323137283325,
-0.01104023028165102,
-0.012370200827717781,
-0.017936265096068382,
-0.009650399908423424,
-0.03275647386908531,
-0.006768854334950447,
-0.02116566337645054,
0.051190461963415146,
-0.10584553331136703,
0.12361074984073639,
0.09477990120649338,
-0.12892483174800873,
0.04084309935569763,
-0.08253917843103409,
-0.014403436332941055,
-0.02766610123217106,
-0.009005319327116013,
-0.17579568922519684,
-0.11831831932067871,
0.0559016652405262,
-0.11227898299694061,
-0.012260925956070423,
-0.04504537954926491,
0.04670209810137749,
0.00149160111322999,
-0.07747117429971695,
-0.036389730870723724,
-0.03766532614827156,
0.007084935903549194,
-0.042842548340559006,
-0.057931117713451385,
-0.12029367685317993,
-0.03248000517487526,
0.052269063889980316,
-0.10253547877073288,
-0.015847498551011086,
-0.0721929743885994,
0.12910272181034088,
0.04262322187423706,
-0.009643194265663624,
0.09273183345794678,
0.04947248473763466,
-0.017650719732046127,
-0.05320250988006592,
0.02441239170730114,
-0.026526102796196938,
0.0003528133674990386,
0.028487198054790497,
-0.05207078903913498,
-0.19223234057426453,
0.030003782361745834,
0.03475531190633774,
-0.06790245324373245,
-0.060199517756700516,
-0.023165108636021614,
0.013188173063099384,
-0.07556750625371933,
-0.07282821834087372,
0.10334975272417068,
0.016429772600531578,
0.07627879828214645,
-0.029541490599513054,
-0.05829354003071785,
-0.001734513440169394,
0.04460212588310242,
-0.02732153795659542,
0.04426313191652298,
0.1370672881603241,
-0.11487479507923126,
0.04899001121520996,
0.0075276135466992855,
0.08519256114959717,
0.03238862007856369,
-0.009216784499585629,
-0.10901125520467758,
0.0440395288169384,
-0.044030338525772095,
0.012349446304142475,
-0.03579283133149147,
-0.08520973473787308,
-0.02304900623857975,
0.03592671826481819,
-0.000748879392631352,
0.006143375765532255,
-0.01275670062750578,
0.06575703620910645,
0.07190787047147751,
-0.05295172333717346,
-0.02355881780385971,
-0.06206244230270386,
-0.04233216866850853,
0.06893905252218246,
0.0028262303676456213,
-0.0016713906079530716,
-0.03335881233215332,
-0.010941363871097565,
-0.13292723894119263,
0.1067304015159607,
-0.03768392279744148,
-0.22099265456199646,
-0.12338364124298096,
-0.0629907175898552,
0.011318654753267765,
-0.004643134772777557,
0.022639263421297073,
-0.03684712573885918,
-0.07750391960144043,
-0.13269321620464325,
0.12285836040973663,
-0.09458009153604507,
-0.0565369576215744,
-0.10928758978843689,
0.09623570740222931,
0.058799196034669876,
-0.06304925680160522,
0.00017854604811873287,
-0.020906833931803703,
-0.02614372782409191,
0.03272782266139984,
-0.08536405116319656,
0.13291022181510925,
0.0662907138466835,
0.06842458993196487,
-0.02777400240302086,
-0.06689327955245972,
0.18134362995624542,
-0.057362694293260574,
0.08085675537586212,
0.19916005432605743,
-0.017099332064390182,
0.01981503702700138,
0.12397634983062744,
-0.017399000003933907,
-0.007241521030664444,
0.08217058330774307,
0.06209579482674599,
-0.031000999733805656,
-0.20324131846427917,
-0.06931722164154053,
0.0049181291833519936,
-0.16222703456878662,
0.09595222771167755,
0.032315827906131744,
-0.006610723212361336,
0.03306223452091217,
-0.06776181608438492,
0.00699925422668457,
0.013976983726024628,
0.05983612313866615,
0.0893184021115303,
-0.00276610953733325,
0.08657602965831757,
0.0019032452255487442,
-0.06248388811945915,
0.05629395693540573,
0.030935153365135193,
0.20945386588573456,
-0.001673980848863721,
0.07227261364459991,
0.12206456810235977,
0.15035219490528107,
0.04996349290013313,
0.05584040284156799,
0.014599867165088654,
0.026299644261598587,
-0.002185997786000371,
-0.0478147491812706,
0.06596379727125168,
0.01321526151150465,
-0.03144485503435135,
-0.01850949600338936,
0.013846066780388355,
-0.09899923205375671,
-0.003309585154056549,
0.17481017112731934,
0.03241235390305519,
-0.12584620714187622,
-0.05763930082321167,
0.06275447458028793,
-0.04912459850311279,
-0.1055913195014,
0.0050491453148424625,
0.04196362942457199,
-0.1433035433292389,
0.04547049105167389,
-0.004558056592941284,
0.11272744834423065,
-0.07493476569652557,
-0.01478924322873354,
0.03175162523984909,
-0.012731343507766724,
-0.047391269356012344,
0.14609898626804352,
-0.22803665697574615,
0.09898733347654343,
0.00033433354110457003,
0.03176498785614967,
-0.05315471068024635,
0.024522388353943825,
-0.02653498202562332,
0.11462824791669846,
0.2059725970029831,
0.019247684627771378,
0.023520953953266144,
-0.10064761340618134,
-0.06939798593521118,
0.021397346630692482,
0.049318280071020126,
-0.18099994957447052,
0.17395354807376862,
-0.007085139863193035,
0.017034703865647316,
-0.03359585255384445,
0.028349166736006737,
-0.1249225065112114,
-0.1403205841779709,
0.03333389386534691,
-0.12390125542879105,
0.20293672382831573,
-0.031239043921232224,
-0.028091520071029663,
-0.044373735785484314,
0.10086078941822052,
-0.05470593273639679,
-0.09342528879642487,
-0.15313206613063812,
0.04486581310629845,
0.14021526277065277,
-0.0571739487349987,
0.045809660106897354,
0.05553135275840759,
0.1785009205341339,
-0.03245517238974571,
-0.15005147457122803,
0.022948548197746277,
-0.0818212702870369,
-0.10615754127502441,
-0.056420598179101944,
0.10059263557195663,
0.09197677671909332,
0.05252281203866005,
0.042138662189245224,
0.03685314208269119,
0.04096671938896179,
-0.05430780351161957,
0.1147325411438942,
0.22266031801700592,
-0.014379465021193027,
0.008723573759198189,
-0.11295894533395767,
-0.03396187722682953,
-0.14748001098632812,
-0.08626767247915268,
0.12662777304649353,
0.042685989290475845,
-0.05854545906186104,
0.07588420808315277,
0.10798167437314987,
-0.16585427522659302,
-0.20671062171459198,
-0.09265704452991486,
0.13029901683330536,
0.02723987214267254,
0.028375715017318726,
-0.26500633358955383,
0.0595407709479332,
0.01729048602283001,
-0.016695799306035042,
-0.1207960769534111,
-0.24037370085716248,
-0.1237783208489418,
0.08034287393093109,
0.04341192543506622,
-0.09733373671770096,
-0.03666049987077713,
-0.027385011315345764,
-0.06260029226541519,
-0.03809675946831703,
0.12410107254981995,
0.02254951186478138,
0.004474301356822252,
-0.013225039467215538,
0.06845452636480331,
0.03478122502565384,
0.011129920370876789,
0.07252702116966248,
0.03305421397089958,
0.049718499183654785,
-0.04216274246573448,
0.09056010097265244,
0.06302536278963089,
-0.05838059261441231,
0.10484781861305237,
0.16260194778442383,
0.005040259100496769,
-0.1991230696439743,
-0.08767138421535492,
-0.051085952669382095,
-0.002138921059668064,
-0.02608432061970234,
-0.10336363315582275,
-0.06830218434333801,
0.04514848440885544,
0.1220749095082283,
-0.05401338264346123,
0.020827215164899826,
-0.09426812082529068,
-0.00478715542703867,
0.13374412059783936,
0.12320049852132797,
0.0859149694442749,
-0.16011343896389008,
0.049271415919065475,
-0.012033580802381039,
0.04514162614941597,
-0.07821802794933319,
0.08531176298856735,
0.09915575385093689,
0.002508297562599182,
0.10409419238567352,
0.011874858289957047,
-0.1845703125,
-0.029974836856126785,
0.05471672862768173,
-0.10237817466259003,
-0.11054009199142456,
-0.058045707643032074,
0.033616598695516586,
-0.1297440230846405,
-0.08827165514230728,
0.108632393181324,
-0.10154467076063156,
-0.0580969974398613,
-0.05878300964832306,
0.07619377970695496,
0.02401014044880867,
0.1056516170501709,
-0.003983214497566223,
0.03334534168243408,
-0.051888566464185715,
0.10907109826803207,
0.09734705835580826,
-0.06694486737251282,
-0.021579276770353317,
0.04510325938463211,
-0.14465919137001038,
-0.041397109627723694,
-0.009048184379935265,
0.10939307510852814,
-0.08192956447601318,
-0.07514877617359161,
0.049719586968421936,
-0.09085695445537567,
0.04781816527247429,
0.20221541821956635,
-0.005057462956756353,
-0.025093775242567062,
-0.06609842926263809,
-0.04964565858244896,
-0.10709129273891449,
0.06694014370441437,
0.028527235612273216,
0.024805227294564247,
-0.05772258713841438,
0.0926315188407898,
0.058225229382514954,
-0.024552948772907257,
-0.045788440853357315,
-0.020091241225600243,
-0.08378448337316513,
0.028142817318439484,
-0.11609061062335968,
-0.03419467434287071,
-0.01969808340072632,
-0.06242380291223526,
0.0031408267095685005,
0.01789969950914383,
0.014275576919317245,
0.0537700355052948,
-0.07911156117916107,
-0.05483726039528847,
-0.026508398354053497,
0.04984379932284355,
-0.06793858110904694,
-0.02009834162890911,
0.04744263365864754,
-0.08470276743173599,
0.07587844878435135,
0.026316415518522263,
0.08204741775989532,
0.011742666363716125,
-0.047058794647455215,
-0.023241743445396423,
0.05985129624605179,
0.06168239936232567,
-0.03680925816297531,
-0.14892318844795227,
-0.01217821054160595,
-0.0188972856849432,
-0.1631450206041336,
-0.014167898334562778,
-0.02359636127948761,
-0.03122730739414692,
-0.05007247254252434,
0.03510918468236923,
0.10810108482837677,
-0.03274676203727722,
0.07678423821926117,
0.05437168478965759,
0.0346166156232357,
0.09152014553546906,
-0.027301503345370293,
0.0387318953871727,
-0.1836395114660263,
-0.037817928940057755,
-0.04206538200378418,
0.01727692037820816,
-0.07361917942762375,
-0.07898217439651489,
0.04874015599489212,
0.004404996056109667,
0.13077573478221893,
0.07732725143432617,
0.12684805691242218,
-0.004598189145326614,
-0.059280868619680405,
0.05126990005373955,
-0.02342306077480316,
0.04850829392671585,
0.09297559410333633,
-0.0003294401103630662,
0.06699632108211517,
0.004898841492831707,
-0.03690188750624657,
0.13181185722351074,
-0.040133558213710785,
0.06722390651702881,
0.13568946719169617,
-0.0156870037317276,
0.02448122762143612,
-0.0793728455901146,
-0.07304597645998001,
-0.10195977240800858,
0.008597454987466335,
0.06382250785827637,
-0.028246955946087837,
0.10790937393903732,
0.05368416756391525,
-0.14168837666511536,
0.1382240206003189,
0.02025742270052433,
-0.06004124507308006,
-0.07230665534734726,
-0.19975151121616364,
-0.01872202940285206,
-0.10035151243209839,
-0.011819781735539436,
-0.14206039905548096,
-0.03383460268378258,
0.15635524690151215,
0.015712503343820572,
0.01657671108841896,
-0.008531395345926285,
-0.05436290428042412,
-0.1921509951353073,
0.04118581861257553,
-0.021324949339032173,
0.03407832607626915,
-0.037715714424848557,
0.12888126075267792,
0.10040091723203659,
0.027437033131718636,
-0.0017360521014779806,
0.03923061117529869,
0.06351657211780548,
0.018736934289336205,
0.011803303845226765,
-0.02619599550962448,
-0.0016353518003597856,
-0.010711491107940674,
-0.018739717081189156,
0.2128135710954666,
0.08643458038568497,
0.010859278962016106,
-0.013327307067811489,
0.2719590961933136,
-0.013744792900979519,
0.00038731511449441314,
-0.10017619282007217,
0.2847927510738373,
0.051864150911569595,
0.05363902449607849,
-0.024702399969100952,
-0.06802988052368164,
-0.00497882766649127,
0.2589443027973175,
-0.04052933678030968,
0.05868498235940933,
-0.03245920315384865,
0.010165945626795292,
-0.00025562828523106873,
0.08680397272109985,
0.05918252468109131,
0.0788346454501152,
0.16360385715961456,
-0.12562677264213562,
0.1111917644739151,
-0.06995449215173721,
-0.04773583635687828,
0.010547695681452751,
0.13325881958007812,
-0.01633138582110405,
0.012655765749514103,
-0.015622206032276154,
0.10090653598308563,
-0.055712636560201645,
-0.21698783338069916,
0.033841684460639954,
-0.00856885313987732,
-0.03553998842835426,
-0.01651792973279953,
-0.07828269898891449,
0.07462149113416672,
-0.005076433066278696,
0.013068592175841331,
0.04322120174765587,
0.24629251658916473,
0.09069105237722397,
-0.0426977202296257,
-0.07326193153858185,
0.08207601308822632,
-0.05020296573638916,
0.13780580461025238,
-0.0005999592249281704,
0.008184017613530159,
0.04595213383436203,
-0.03715120628476143,
-0.16037903726100922,
0.06034229323267937,
0.020304184406995773,
0.012375769205391407,
0.0346093513071537,
0.16013646125793457,
0.030511880293488503,
0.1392693668603897,
0.023589445278048515,
-0.04924188181757927,
0.0775790587067604,
-0.03543220832943916,
-0.07890307903289795,
0.009469885379076004,
0.028317265212535858,
-0.050621990114450455,
0.06648676842451096,
0.17520247399806976,
-0.06480361521244049,
-0.00023298927408177406,
-0.10357184708118439,
0.04084021598100662,
-0.011726520024240017,
0.04468474164605141,
0.04073478654026985,
-0.05806577950716019,
0.018614808097481728,
0.0017798752523958683,
-0.03336693346500397,
-0.18242084980010986,
-0.021990565583109856,
0.09258150309324265,
-0.02295256406068802,
-0.0020227113272994757,
0.1114959716796875,
0.03674369677901268,
0.041227374225854874,
-0.04708845540881157,
-0.016768645495176315,
0.03697868436574936,
0.07144790887832642,
-0.12090978026390076,
-0.006146871950477362
] |
null | null |
transformers
|
#dialogue
|
{"tags": ["text-generation"]}
|
text-generation
|
danchang11/GPT2-TraditionalChat
|
[
"transformers",
"pytorch",
"gpt2",
"text-generation",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] |
2022-03-02T23:29:05+00:00
|
[] |
[] |
TAGS
#transformers #pytorch #gpt2 #text-generation #endpoints_compatible #text-generation-inference #region-us
|
#dialogue
|
[] |
[
"TAGS\n#transformers #pytorch #gpt2 #text-generation #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
39
] |
[
"passage: TAGS\n#transformers #pytorch #gpt2 #text-generation #endpoints_compatible #text-generation-inference #region-us \n"
] |
[
-0.0177617147564888,
0.00413000863045454,
-0.008743596263229847,
-0.014220969751477242,
0.164886936545372,
0.03891453891992569,
0.046693552285432816,
0.13523150980472565,
0.015874169766902924,
-0.004716119728982449,
0.1357993483543396,
0.1721617579460144,
0.006274322979152203,
0.005896460264921188,
-0.05455324426293373,
-0.2677989900112152,
0.07517701387405396,
0.08140098303556442,
-0.042099032551050186,
0.12790833413600922,
0.048993777483701706,
-0.04671265929937363,
0.08353178203105927,
-0.018136927857995033,
-0.14306217432022095,
0.012910505756735802,
0.0358387716114521,
-0.0987507775425911,
0.11372382193803787,
0.02934994176030159,
0.10152171552181244,
0.0027315847110003233,
-0.11149000376462936,
-0.16700994968414307,
0.029015103355050087,
0.05075612664222717,
-0.07201661914587021,
0.05401792749762535,
0.08914339542388916,
-0.12092448025941849,
0.12759658694267273,
0.029667116701602936,
-0.051347605884075165,
0.049732182174921036,
-0.1655282825231552,
-0.09687618911266327,
-0.02854047901928425,
0.025650938972830772,
0.03822527453303337,
0.10048377513885498,
-0.016911430284380913,
0.06471045315265656,
-0.09987933188676834,
0.08688510209321976,
0.22739006578922272,
-0.32783472537994385,
0.00570092024281621,
0.11302487552165985,
0.05605051666498184,
0.053175803273916245,
-0.02404518984258175,
0.05229756981134415,
0.01313769444823265,
0.01957249827682972,
0.00137265061493963,
-0.07302742451429367,
-0.0565350204706192,
0.08518718928098679,
-0.10417694598436356,
-0.10550206154584885,
0.2108500748872757,
-0.06404201686382294,
0.06052133068442345,
-0.024971894919872284,
-0.10919266194105148,
-0.058717526495456696,
-0.0136891920119524,
0.047699883580207825,
-0.05458793044090271,
0.09626783430576324,
0.04053015634417534,
-0.08627340197563171,
-0.13180461525917053,
-0.06005355343222618,
-0.17273692786693573,
0.20520804822444916,
0.024591796100139618,
0.07910293340682983,
-0.2197970300912857,
0.11058394610881805,
-0.038688383996486664,
-0.06404577195644379,
0.002036143559962511,
-0.10641399770975113,
0.05131516978144646,
0.03840160742402077,
-0.07174071669578552,
-0.05681565776467323,
0.09885678440332413,
0.0772608071565628,
-0.08083102107048035,
0.02651338279247284,
-0.03876886144280434,
0.10964981466531754,
0.013784559443593025,
0.07848335802555084,
-0.028710849583148956,
0.03977745771408081,
0.01966063678264618,
-0.17050744593143463,
-0.0029470184817910194,
-0.057985570281744,
-0.11522450298070908,
-0.0807778462767601,
0.03479090705513954,
0.09325509518384933,
0.010769971646368504,
0.0810101106762886,
-0.03773309290409088,
-0.025342658162117004,
0.0269398745149374,
-0.05964063107967377,
-0.030446602031588554,
0.032796863466501236,
0.012962838634848595,
0.1671743243932724,
-0.0137644587084651,
-0.0017080423422157764,
-0.15164265036582947,
0.05012911558151245,
-0.07729051262140274,
0.017933771014213562,
-0.041089978069067,
-0.03279638662934303,
0.022797340527176857,
-0.11531727015972137,
0.004681426100432873,
-0.11883518099784851,
-0.17856267094612122,
0.027148963883519173,
-0.00009361533011542633,
-0.015542703680694103,
-0.0018064614851027727,
-0.028014300391077995,
-0.052028052508831024,
0.024566810578107834,
-0.06845179945230484,
0.001042075571604073,
-0.06374721974134445,
0.1192232072353363,
-0.06242113187909126,
0.06873898208141327,
-0.11195021122694016,
0.07659032940864563,
-0.10372360050678253,
0.01668623648583889,
-0.12242420017719269,
0.0736471563577652,
-0.005726009141653776,
0.09933587908744812,
-0.0381762757897377,
-0.07277093082666397,
-0.11802409589290619,
0.052166618406772614,
-0.052433934062719345,
0.19637678563594818,
-0.05141520872712135,
-0.12538684904575348,
0.3230677843093872,
-0.07854989171028137,
-0.12571749091148376,
0.08480606228113174,
0.017596984282135963,
-0.009740262292325497,
0.08438948541879654,
0.22746822237968445,
0.013743794523179531,
0.005568898748606443,
0.05719493329524994,
0.1417950987815857,
-0.13118228316307068,
-0.0950319841504097,
0.04732563719153404,
-0.054628971964120865,
-0.10563048720359802,
0.030682945623993874,
-0.007876500487327576,
0.0630379170179367,
-0.0447998009622097,
-0.012814326211810112,
-0.04133640602231026,
0.02648276835680008,
0.08373766392469406,
0.001436202903278172,
0.10575487464666367,
-0.043722447007894516,
-0.04275013506412506,
-0.01118161529302597,
-0.01528346911072731,
-0.06143542379140854,
0.06491949409246445,
-0.014812358655035496,
0.14447778463363647,
-0.03690091893076897,
0.04783803969621658,
-0.18692760169506073,
-0.08071625232696533,
-0.023726923391222954,
0.10554835200309753,
-0.03628280386328697,
0.07664146274328232,
0.06979329884052277,
-0.06121150404214859,
-0.011977508664131165,
-0.00036841665860265493,
0.10460905730724335,
-0.04247349128127098,
-0.024623505771160126,
-0.040476251393556595,
0.04367184638977051,
-0.050804685801267624,
-0.03764241188764572,
-0.043807633221149445,
0.026319099590182304,
0.12130927294492722,
0.08805370330810547,
-0.011063527315855026,
0.01108147669583559,
0.0008715765434317291,
0.002451015170663595,
-0.04143768176436424,
-0.01616770774126053,
0.1087322011590004,
-0.01118565909564495,
-0.09968315064907074,
0.20184315741062164,
-0.12686163187026978,
0.18463411927223206,
0.1917361617088318,
-0.28694963455200195,
0.044258002191782,
-0.08406686037778854,
-0.04173731803894043,
0.03883618488907814,
0.0610942617058754,
-0.054276976734399796,
0.1691325157880783,
0.01757095381617546,
0.1580827832221985,
-0.07692365348339081,
-0.06435651332139969,
-0.009723078459501266,
-0.027760373428463936,
0.008332083001732826,
0.08378587663173676,
0.08490931242704391,
-0.09190400689840317,
0.18039464950561523,
0.17116224765777588,
0.05264123156666756,
0.1655624955892563,
-0.02682124264538288,
-0.04416370019316673,
0.08961658924818039,
0.05326202139258385,
-0.04856514930725098,
-0.07251864671707153,
-0.2637195289134979,
-0.02071956731379032,
0.08327561616897583,
0.07756919413805008,
0.14509862661361694,
-0.14547564089298248,
-0.05993083491921425,
-0.02641592174768448,
-0.05574978142976761,
-0.05129653960466385,
0.0821082666516304,
0.061170388013124466,
0.1182895228266716,
0.011390134692192078,
0.04968147352337837,
0.10705231875181198,
0.0072124917060136795,
-0.0920310989022255,
0.22349299490451813,
-0.11133892834186554,
-0.34016597270965576,
-0.14436322450637817,
-0.11788070946931839,
-0.030494745820760727,
0.04577368497848511,
0.10457808524370193,
-0.136713445186615,
0.003679262939840555,
0.02427663654088974,
0.14769063889980316,
-0.13262353837490082,
0.017893647775053978,
-0.041815612465143204,
0.04790092632174492,
-0.10780538618564606,
-0.09514795243740082,
-0.05651114135980606,
-0.036214686930179596,
-0.0907103642821312,
0.13784471154212952,
-0.12632136046886444,
0.03253214433789253,
0.1913776844739914,
0.06095456704497337,
0.06541487574577332,
-0.040304239839315414,
0.17827682197093964,
-0.1092667505145073,
-0.034401800483465195,
0.21323581039905548,
-0.02538575418293476,
0.09009216725826263,
0.05674639344215393,
0.0030026263557374477,
-0.08552411198616028,
-0.005033917725086212,
-0.023024432361125946,
-0.11021243780851364,
-0.26796042919158936,
-0.10801071673631668,
-0.12731461226940155,
0.0727757066488266,
0.022434573620557785,
0.06192222982645035,
0.1212952733039856,
0.0660119503736496,
-0.02397187612950802,
-0.007011115085333586,
0.027425037696957588,
0.07903468608856201,
0.1664983183145523,
-0.029877230525016785,
0.09351065754890442,
-0.05715025216341019,
-0.09589941799640656,
0.08371074497699738,
0.07919108867645264,
0.20126491785049438,
0.03149360418319702,
0.07982219010591507,
0.045319534838199615,
0.058093857020139694,
0.13299427926540375,
0.1143735721707344,
-0.02366695925593376,
-0.00476102065294981,
-0.02725902758538723,
-0.023289548233151436,
-0.04717805236577988,
0.0179450660943985,
0.007347718812525272,
-0.1569279283285141,
-0.058918919414281845,
-0.1493380069732666,
0.11772475391626358,
0.0877038910984993,
0.05193497985601425,
-0.18433049321174622,
-0.0027360362000763416,
0.09183459728956223,
-0.016695870086550713,
-0.11040494590997696,
0.10688318312168121,
0.025785457342863083,
-0.13458241522312164,
0.05990788713097572,
-0.062118303030729294,
0.12250654399394989,
-0.060343630611896515,
0.08080611377954483,
-0.04021589085459709,
-0.09235090017318726,
0.02796080708503723,
0.1174207478761673,
-0.2523665726184845,
0.22567634284496307,
-0.004532721359282732,
-0.05051229149103165,
-0.09785845875740051,
-0.004584147594869137,
-0.006866491865366697,
0.10252957046031952,
0.15214580297470093,
0.026027202606201172,
-0.014887488447129726,
-0.0593542717397213,
-0.0008516228408552706,
0.03857841715216637,
0.1469660848379135,
-0.06455042958259583,
-0.01702277362346649,
-0.021815843880176544,
0.005529934074729681,
-0.0365385003387928,
-0.03495530039072037,
0.0736987516283989,
-0.1503359079360962,
0.06277604401111603,
0.008920188993215561,
0.09208177030086517,
0.001672430313192308,
0.009417003951966763,
-0.07299439609050751,
0.20857995748519897,
-0.13122130930423737,
-0.13205280900001526,
-0.09646400064229965,
-0.0619744136929512,
0.08571894466876984,
-0.05828427895903587,
0.0448489636182785,
-0.06629502773284912,
0.007595032919198275,
-0.05393918231129646,
-0.21739354729652405,
0.10785721987485886,
-0.06963614374399185,
-0.023533938452601433,
0.0036003789864480495,
0.246476411819458,
-0.06690777838230133,
0.006640743464231491,
0.01160342339426279,
0.020129498094320297,
-0.10629308968782425,
-0.12935012578964233,
0.026942672207951546,
0.016331583261489868,
0.08289532363414764,
0.08407847583293915,
-0.04428640007972717,
0.055529527366161346,
-0.020418979227542877,
0.02320215106010437,
0.31711724400520325,
0.10947677493095398,
-0.03180767595767975,
0.18130561709403992,
0.08401691168546677,
-0.08147493004798889,
-0.2822689414024353,
-0.06397486478090286,
-0.1189005896449089,
-0.05097424238920212,
-0.09442377835512161,
-0.2242605984210968,
0.08713490515947342,
0.06292454153299332,
0.008994982577860355,
0.15863020718097687,
-0.31459149718284607,
-0.05044165253639221,
0.08400531858205795,
-0.004015425220131874,
0.4111485481262207,
-0.14122027158737183,
-0.1208438128232956,
-0.02601119875907898,
-0.21869809925556183,
0.15852734446525574,
-0.0737798884510994,
0.107994444668293,
-0.03963931277394295,
0.10422861576080322,
0.03994893655180931,
-0.0722108855843544,
0.11042612791061401,
0.05572624132037163,
0.0006772224442102015,
-0.07959822565317154,
-0.03933039307594299,
0.07553410530090332,
0.01759473606944084,
0.01572279818356037,
-0.027233067899942398,
0.031986165791749954,
-0.15855683386325836,
-0.04007037356495857,
-0.11417832225561142,
0.04086349159479141,
0.047451384365558624,
-0.05472583696246147,
-0.022158317267894745,
-0.06267859041690826,
0.0036529593635350466,
0.02783399075269699,
0.2299688458442688,
-0.055845387279987335,
0.14678794145584106,
0.0181744322180748,
0.08333985507488251,
-0.12178298830986023,
-0.07275176048278809,
-0.06875777989625931,
-0.028424803167581558,
0.0988553985953331,
-0.16970506310462952,
0.053557995706796646,
0.1198544055223465,
-0.024564003571867943,
0.046731337904930115,
0.13614340126514435,
0.005411188583821058,
0.008843154646456242,
0.11319591850042343,
-0.24978069961071014,
-0.06546348333358765,
-0.06806614249944687,
-0.06450063735246658,
0.09459757059812546,
0.11480807512998581,
0.1617022007703781,
0.044600822031497955,
-0.034621790051460266,
-0.015353205613791943,
0.019293226301670074,
-0.06277140974998474,
0.03454780951142311,
0.001771117327734828,
0.029499240219593048,
-0.14590995013713837,
0.08422111719846725,
-0.01743590086698532,
-0.15481045842170715,
0.0021370314061641693,
0.1372750699520111,
-0.141866534948349,
-0.10142909735441208,
-0.062354519963264465,
0.05168379843235016,
-0.11357450485229492,
-0.023666704073548317,
-0.03506386652588844,
-0.12296982854604721,
0.09337733685970306,
0.15561185777187347,
0.06566707044839859,
0.11923837661743164,
-0.026205087080597878,
-0.024233929812908173,
-0.00927067268639803,
-0.06355437636375427,
-0.04055394232273102,
-0.006395004689693451,
-0.06335654109716415,
0.06671498715877533,
-0.021831858903169632,
0.14458994567394257,
-0.07311990857124329,
-0.07258106768131256,
-0.15852877497673035,
0.05074010789394379,
-0.11197909712791443,
-0.0547679141163826,
-0.10145534574985504,
-0.061023611575365067,
-0.02561028115451336,
-0.003428260562941432,
-0.04455279931426048,
-0.03193459287285805,
-0.11993534862995148,
0.03322478011250496,
-0.05789078772068024,
0.013840819709002972,
-0.07794932276010513,
0.013500452041625977,
0.09164173901081085,
-0.043710220605134964,
0.15036717057228088,
0.16742666065692902,
-0.09713615477085114,
0.14140373468399048,
-0.15944255888462067,
-0.09699242562055588,
0.1065526008605957,
-0.004710288718342781,
0.01055878959596157,
0.08062771707773209,
0.03611044958233833,
0.04957300424575806,
0.00015067339700181037,
0.0604633130133152,
-0.036007050424814224,
-0.12734907865524292,
0.03170613944530487,
-0.042643286287784576,
-0.10766422003507614,
-0.06020660325884819,
-0.049973342567682266,
0.06957629323005676,
0.05035904794931412,
0.06711549311876297,
-0.011089643463492393,
0.11557795107364655,
-0.050442613661289215,
0.018985580652952194,
0.026638763025403023,
-0.16677306592464447,
0.008527749218046665,
-0.061151642352342606,
0.03204059973359108,
0.010763505473732948,
0.28714069724082947,
-0.00013240271073300391,
-0.007132190745323896,
0.014635466039180756,
0.0900757908821106,
0.04493294656276703,
0.017144059762358665,
0.2578983008861542,
0.10303528606891632,
-0.06955546885728836,
-0.08077843487262726,
0.06792601943016052,
0.012195531278848648,
0.02784431166946888,
0.17747752368450165,
0.08463454991579056,
0.02749769017100334,
0.09605445712804794,
-0.040658168494701385,
0.021314077079296112,
-0.10011026263237,
-0.11307407170534134,
0.013244640082120895,
0.04699188470840454,
0.0012895361287519336,
0.1374855935573578,
0.14933034777641296,
-0.040645238012075424,
0.06967397034168243,
0.016282886266708374,
-0.05199767276644707,
-0.15802720189094543,
-0.14379066228866577,
-0.04298508167266846,
-0.14894913136959076,
0.021038243547081947,
-0.13705667853355408,
0.03882404416799545,
0.13195720314979553,
0.06392589956521988,
-0.03431929647922516,
0.1295498013496399,
0.07832565158605576,
-0.11752352863550186,
0.08415429294109344,
-0.032199811190366745,
0.08165641874074936,
0.03971846029162407,
-0.017476707696914673,
-0.03917689621448517,
-0.04564012959599495,
0.011646753177046776,
0.06824102252721786,
-0.030940555036067963,
0.012069804593920708,
-0.16889329254627228,
-0.08907386660575867,
-0.0529489628970623,
0.09230942279100418,
-0.05592195317149162,
0.13175910711288452,
0.0032278846483677626,
-0.04776560515165329,
0.0334765650331974,
0.2324390858411789,
-0.0619046650826931,
-0.018577080219984055,
-0.018306629732251167,
0.17200686037540436,
0.05329083278775215,
0.08580727875232697,
-0.00465694535523653,
-0.006655924487859011,
-0.07144376635551453,
0.3501553535461426,
0.2633618414402008,
-0.03886903077363968,
0.012465479783713818,
0.041434697806835175,
0.04495251551270485,
0.17578192055225372,
0.1147269532084465,
0.11233104765415192,
0.29954424500465393,
-0.07917296886444092,
-0.0670166090130806,
-0.009023794904351234,
-0.004121111705899239,
-0.10456884652376175,
0.10345932841300964,
0.039856355637311935,
-0.09799449145793915,
-0.04581701010465622,
0.11436376720666885,
-0.25572752952575684,
0.09593817591667175,
-0.048779647797346115,
-0.1697508841753006,
-0.047806475311517715,
-0.013967693783342838,
0.10703971982002258,
0.010679148137569427,
0.10505855083465576,
-0.006022031884640455,
-0.13563650846481323,
0.04811090603470802,
0.06031722202897072,
-0.26589345932006836,
-0.01062775868922472,
0.05762946978211403,
-0.02625512331724167,
-0.010761603713035583,
-0.020649395883083344,
0.04956243932247162,
0.06090697646141052,
0.03886969015002251,
-0.02475808933377266,
0.01703202910721302,
-0.004924232140183449,
-0.040182795375585556,
-0.02096102386713028,
0.05230327695608139,
0.006438924930989742,
-0.15112970769405365,
0.06094548851251602,
-0.135308176279068,
0.02098478563129902,
0.028334662318229675,
-0.032334886491298676,
-0.019326696172356606,
-0.044271863996982574,
-0.08218339085578918,
0.022670846432447433,
0.08878248929977417,
-0.0012725733686238527,
-0.01129397377371788,
-0.08182375878095627,
-0.022717300802469254,
-0.019387206062674522,
-0.05968303605914116,
-0.09699249267578125,
-0.09970560669898987,
-0.10814160108566284,
0.1260090470314026,
-0.020027613267302513,
-0.1781155914068222,
0.031491469591856,
-0.05561300739645958,
0.07841652631759644,
-0.1629980504512787,
0.0618479959666729,
0.04849570244550705,
0.02506282366812229,
-0.00006693792238365859,
-0.01461757067590952,
0.06004534289240837,
0.09022362530231476,
-0.08918830007314682,
-0.08250809460878372
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.